CN103699660A - Large-scale network streaming data cache-write method - Google Patents

Large-scale network streaming data cache-write method Download PDF

Info

Publication number
CN103699660A
CN103699660A CN201310741116.6A CN201310741116A CN103699660A CN 103699660 A CN103699660 A CN 103699660A CN 201310741116 A CN201310741116 A CN 201310741116A CN 103699660 A CN103699660 A CN 103699660A
Authority
CN
China
Prior art keywords
data
distributed
loaded
file system
different
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201310741116.6A
Other languages
Chinese (zh)
Other versions
CN103699660B (en
Inventor
汪东升
王丽婧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CERTUSNET CORP
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201310741116.6A priority Critical patent/CN103699660B/en
Publication of CN103699660A publication Critical patent/CN103699660A/en
Application granted granted Critical
Publication of CN103699660B publication Critical patent/CN103699660B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • G06F16/2219Large Object storage; Management thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2453Query optimisation
    • G06F16/24534Query rewriting; Transformation
    • G06F16/24539Query rewriting; Transformation using cached or materialised query results
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/25Integrating or interfacing systems involving database management systems
    • G06F16/256Integrating or interfacing systems involving database management systems in federated or virtual databases

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a large-scale network streaming data cache-write method including that a client side packages collected data and sent the same to a server side; the server side analyzes the received data to determine data resources and unifies data formats; the data with uniform formats are written into cache regions of different inner memories according the data resources; different cache strategies are made according to the data which are written into the cache regions of the different inner memories, and the data meeting strategy trigger conditions are written into a local file system from the inner memories; the data written into the local file system are loaded to a distributed file system or distributed database according to the upper application demand; and to the data which are loaded to the distributed file system or distributed database, small data blocks are combined into large data blocks at fixed time. The large-scale network streaming data cache-write method is applicable to write of large-scale network data that are of different resources and different inflow velocity by the aid of a multi-stage cache mechanism.

Description

A kind of method that large scale network stream data buffer memory writes
Technical field
The present invention relates to networking technology area, be specifically related to a kind of method that large scale network stream data buffer memory writes.
Background technology
In recent years, along with the fast development of internet, the rapid growth of data become the opportunities and challenges of many industry facings.At current net environment, mass data source is real-time continual, and requiring is also real-time to user's response time.These data are collected with the form of streaming, calculate and inquire about.For example Network anomaly detection system, by data such as collection network bag, network logs, analyzes, and guarantee to return analysis result within the scope of certain hour, ensures the high availability of network.The feature of this system is: each has the network data of magnanimity of all kinds to flow into system constantly, and inflow velocity is different, data structure complexity various (comprising binary file, text, compressed file etc.), and Network anomaly detection is a kind of application.For this type of application, need bottom storage system to support: the data that flow into are stored with consolidation form, to upper layer application, provide unified interface, convenient search, and real-time be there are certain requirements.
For large data trend now, emerged in large numbers a collection of large data processing platform (DPP), application comprises more widely and adopts the Hadoop distributed system of MapReduce parallel processing framework to process framework, comprises again HDFS(Hadoop Distributed File System in this Open Source Framework), Hbase(Hadoop Database), Hive(Tool for Data Warehouse) etc. sub-project.The design of HDFS can be stored the data (normally PB is even higher) of magnanimity, and the mode of application program file reading is assumed that streaming reads, and has done much optimization in the performance that HDFS reads in streaming; HBase is distributed non-relational Database Systems, HBase builds on HDFS distributed memory system, that is to say that HDFS supports for HBase provides the distributed bottom storage of high reliability, Hadoop MapReduce provides high performance Distributed Calculation engine for HBase; Hive is a Tool for Data Warehouse based on Hadoop bottom, structurized data file can be mapped as to a database table, and high-rise SQL query function is provided, the very applicable operation of the statistical study for massive structured data; In addition, at present application more distributed streaming processing platform has the S4 of Yahoo, the Storm of Twitter, the two be all freely increase income, distributed, high fault-tolerant real time computation system.
But the batch mode of Hadoop framework can not meet the requirement of real-time calculating, system processing speed slows down, and is not suitable for data and flows directly into; Hbase and Hive belong to distributed data base, yet it is different from the different pieces of information source and course of network, to enter the data layout of system, and the not treated distributed data base that cannot directly offer is used; The S4 of Yahoo, this class of the Storm of Twitter are for the problem of distributed streaming processing platform specially, they are more to provide a kind of computing power of stream data, the data of all arrival directly enter in internal memory and calculate after treatment, the data that flow into are not carried out to persistent storage, can not meet the demand of application.
Summary of the invention
(1) technical matters solving
For the deficiencies in the prior art, the invention provides a kind of large scale network stream data buffer memory wiring method, can utilize writing of the different large scale network data in multi-level buffer mechanism reply different pieces of information source, inflow velocity.
(2) technical scheme
In order to reach above object, the present invention is achieved by the following technical programs:
The method that large scale network stream data buffer memory writes, the method comprises:
Client, by after the data encapsulation gathering, sends to server end;
Server end is resolved the data that receive, and confirms the source of data, and data layout is unified;
By the data after uniform format, according to its source, write the buffer zone of different internal memories;
To writing the data in different memory caches region, formulate different cache policies, and the data that meet tactful trigger condition are write to local file system from internal memory;
The data based upper layer application demand that writes local file system is loaded in distributed file system or distributed data base;
For the data that are loaded into distributed file system or distributed data base, regularly small data is merged into large data.
Wherein, described data layout after reunification, should be Key-Value form or relational data form.
Preferably, the method further comprises: for the data that will be loaded in distributed data base, further judge whether it is relational data, if relational data, be loaded into distributed relation database, if not relational data is loaded into distributed No-SQL database.
Preferably, the method also comprises: for the data that are loaded into distributed data base, the storage format of the large data after merging is converted to the ranks storage format for the special optimization of distributed data base.
The system that large scale network stream data buffer memory writes, this system comprises:
Data transmission blocks, for by after the data encapsulation gathering, sends to server end;
Providing data formatting module, for the data that receive are resolved, confirms the source of data, and data layout is unified;
Data cache module, for by the data after uniform format, writes the buffer zone of different internal memories according to its source;
Data persistence module, for to writing the different cache policy of data customization of different buffer zones, and writes local file system by the data that meet tactful trigger condition from internal memory;
Data load-on module, for being loaded into distributed file system or distributed data base by the data based upper layer application demand that writes local file system;
The regular module of data, for to being loaded into the data of distributed file system or distributed data base, is regularly merged into small data large data.
Wherein, described providing data formatting module, by data layout after reunification, the form of its data should be Key-Value form or relational data form.
Preferably, described data load-on module, whether the data that further judgement is loaded in distributed data base are relational data, if relational data, be loaded into distributed relation database, if not relational data is loaded into distributed No-SQL database.
Preferably, the regular module of described data, further, by the storage format of the large data after merging in distributed data base, converts the ranks storage format for the special optimization of distributed data base to.
(3) beneficial effect
The present invention has following beneficial effect at least:
Method provided by the invention utilizes multi-level buffer mechanism to carry out buffer memory to data, when server receives after data, the form of data is unified, and formulate different cache policies according to the separate sources of data, and can tackle writing of the different large scale network data in different pieces of information source, inflow velocity, scrappy small data is merged into long data block, improved data processing speed, reduce data space, reduced data management cost, meet large data processing demand.
Method provided by the invention, data layout is unified, and data layout is Key-Value form or relational data form after reunification, facilitates subsequent calculations directly to call.
Method provided by the invention, except the small data being loaded in the data of distributed data base is merged into long data block, also data storing format conversion is become to the ranks storage format for the special optimization of distributed data base, when saving storage area, optimized search efficiency.
Accompanying drawing explanation
In order to be illustrated more clearly in the embodiment of the present invention or technical scheme of the prior art, to the accompanying drawing of required use in embodiment or description of the Prior Art be briefly described below, apparently, accompanying drawing in the following describes is only some embodiments of the present invention, for those of ordinary skills, do not paying under the prerequisite of creative work, can also obtain according to these figure other accompanying drawing.
Fig. 1 is the process flow diagram of the method that writes of a kind of large scale network stream data buffer memory that the embodiment of the present invention provides;
Fig. 2 is the process flow diagram of the method that writes of a kind of extensive stream data buffer memory that a preferred embodiment of the present invention provides;
Fig. 3 is illustrating about cache policy.
Embodiment
Below in conjunction with the accompanying drawing in the embodiment of the present invention, the technical scheme in the embodiment of the present invention is clearly and completely described, obviously, described embodiment is only the present invention's part embodiment, rather than whole embodiment.Embodiment based in the present invention, those of ordinary skills, not making the every other embodiment obtaining under creative work prerequisite, belong to the scope of protection of the invention.
Referring to Fig. 1, a kind of method that the embodiment of the present invention provides large scale network stream data buffer memory to write, the step of the method is as follows:
Step 101: client, by after the data encapsulation gathering, sends to server end;
Step 102: server end is resolved the data that receive, the source of confirmation data, and data layout is unified;
Step 103: by the data after uniform format, write the buffer zone of different internal memories according to its source;
Step 104: formulate different cache policies to writing the data in different memory caches region, and the data that meet tactful trigger condition are write to local file system from internal memory;
Step 105: the data based upper layer application demand that writes local file system is loaded in distributed file system or distributed data base;
Step 106: for the data that are loaded into distributed file system distributed data base, according to set strategy, small data is merged into large data.
The method that the embodiment of the present invention provides utilizes multi-level buffer mechanism to carry out buffer memory to data, when server receives after data, the form of data is unified, and formulate different cache policies according to the separate sources of data, can tackle writing of the different large scale network data in different pieces of information source, inflow velocity, scrappy small data is merged into long data block, improved data processing speed, reduce data space, reduce data management cost, met large data processing demand.
Below by an example more specifically, the implementation procedure of a preferred embodiment of the present invention is described, referring to Fig. 2, the step of the method is as follows:
Step 201: client, by after the data encapsulation gathering, sends to server end.
In this step, client is only responsible for the data simplified package that each is gathered constantly, indicates its Data Source, and the mode with POST, sends to server end by http protocol.
Step 202: server end is resolved the data that receive, the source of confirmation data, and data layout is unified.
In this step, by data layout data layout after reunification, be Key-Value form or relational data form, facilitate subsequent calculations directly to call.
Step 203: by the data after uniform format, write the buffer zone of different internal memories according to its source, and formulate different cache policies to writing the data in different memory caches region.
In this step, for dissimilar data, according to its inflow velocity, data volume and upper layer application to data ageing requirement formulate different cache policies.As shown in Figure 3, illustrate concrete cache policy, for upper layer application, to monitor in real time, belong to time-sensitive class, guarantee the ageing of data, if just can accumulate mass data within the very short time interval, as the first row in Fig. 2 simultaneously, within one minute, can accumulate 400M data, comprehensive above-mentioned several conditions set cache policies be every 1 minute persistence once; And for another kind of data, as the second row in Fig. 2,10 minutes ability accumulative total 50M, and drawing according to application demand, not high to requirement of real-time, can set cache policy and be every accumulation 128M data persistence once;
Step 204: accordingly whether decision data the threshold condition of cache policy, if meet, goes to step 205, if do not meet, goes to step 203.
Step 205: data are write local file system from internal memory.
Step 206: according to upper layer application demand, whether decision data will be loaded in distributed data base, if so, goes to step 207, if not, go to step 210.
Step 207: for the data that are loaded into distributed data base, judge whether it is relational data, if so, goes to step 208, if not, go to step 209.
Step 208: data are loaded into distributed relational database.
Step 209: data are loaded into distributed No-SQL database.
Step 210: data are loaded in distributed file system.
Step 211: judge whether the data that are loaded in distributed file system meet the condition of set strategy, if met, go to step 214, if do not met, go to step 210.
In this step, the set policy condition need meeting is formulated according to the inflow velocity of data etc., as not being very fast data for flowing velocity, as the data of 1 day ability accumulation 64M, can regularly to it, carry out the merging of small data piece, to meet the demand of distributed file system to large data processing, if the default tile size of HDFS file system storing documents is 64M, data for not enough 64M are stored with 64M, if being less than 64M, file can cause space waste, metadata is too much, cause system operation slowly, this method is regularly merged into the multiple of 64M to small data piece, reduce disk fragments, improve storage space utilization factor and promote recall precision.
Step 212: judge whether the data that are loaded in distributed No-SQL database meet the condition of set strategy, if met, go to step 214, if do not met, go to step 210.
Step 213: judge whether the data that are loaded in distributed relation database meet the condition of set strategy, if met, go to step 214, if do not met, go to step 210.
Step 214: the small data satisfying condition in distributed file system is merged into long data block.
Step 215: the small data satisfying condition in distributed No-SQL database is merged into long data block.
In this step, except small data is merged into long data block, also data storing format conversion is become as RCFile etc. for the ranks storage format of the special optimization of distributed data base, in saving storage area, optimized search efficiency.
Step 216: the small data piece satisfying condition in distributed relation database is merged into long data block.
In this step, except small data is merged into long data block, also data storing format conversion is become as RCFile etc. for the ranks storage format of the special optimization of distributed data base, in saving storage area, optimized search efficiency.
Step 217: end data buffer memory writes.
Above embodiment only, in order to technical scheme of the present invention to be described, is not intended to limit; Although the present invention is had been described in detail with reference to previous embodiment, those of ordinary skill in the art is to be understood that; Its technical scheme that still can record aforementioned each embodiment is modified, or part technical characterictic is wherein equal to replacement; And these modifications or replacement do not make the essence of appropriate technical solution depart from the spirit and scope of various embodiments of the present invention technical scheme.

Claims (8)

1. the method that large scale network stream data buffer memory writes, is characterized in that, the method comprises:
Client, by after the data encapsulation gathering, sends to server end;
Server end is resolved the data that receive, and confirms the source of data, and data layout is unified;
By the data after uniform format, according to its source, write the buffer zone of different internal memories;
To writing the data in different memory caches region, formulate different cache policies, and the data that meet tactful trigger condition are write to local file system from internal memory;
The data based upper layer application demand that writes local file system is loaded in distributed file system or distributed data base;
Data for being loaded into distributed file system or distributed data base, are regularly merged into long data block by small data.
2. method according to claim 1, is characterized in that, described data layout after reunification, should be Key-Value form or relational data form.
3. method according to claim 1, is characterized in that, the method further comprises:
For the data that will be loaded in distributed data base, further judge whether it is relational data, if relational data is loaded into distributed relation database, if not relational data is loaded into distributed No-SQL database.
4. method according to claim 1, is characterized in that, the method also comprises:
For the data that are loaded into distributed data base, the storage format of the large data after merging is converted to the ranks storage format for the special optimization of distributed data base.
5. the system that large scale network stream data buffer memory writes, is characterized in that, this system comprises:
Data transmission blocks, for by after the data encapsulation gathering, sends to server end;
Providing data formatting module, for the data that receive are resolved, confirms the source of data, and data layout is unified;
Data cache module, for by the data after uniform format, writes the buffer zone of different internal memories according to its source;
Data persistence module, for to writing the different cache policy of data customization of different buffer zones, and writes local file system by the data that meet tactful trigger condition from internal memory;
Data load-on module, for being loaded into distributed file system or distributed data base by the data based upper layer application demand that writes local file system;
The regular module of data, for to being loaded into the data of distributed file system or distributed data base, is regularly merged into small data large data.
6. system according to claim 5, is characterized in that, described providing data formatting module, and by data layout after reunification, the form of its data should be Key-Value form or relational data form.
7. system according to claim 5, it is characterized in that, described data load-on module, whether the data that further judgement is loaded in distributed data base are relational data, if relational data, be loaded into distributed relation database, if not relational data is loaded into distributed No-SQL database.
8. system according to claim 5, is characterized in that, the regular module of described data further, by the storage format of the large data after merging in distributed data base, converts the ranks storage format for the special optimization of distributed data base to.
CN201310741116.6A 2013-12-26 2013-12-26 A kind of method of large scale network stream data caching write Active CN103699660B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310741116.6A CN103699660B (en) 2013-12-26 2013-12-26 A kind of method of large scale network stream data caching write

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310741116.6A CN103699660B (en) 2013-12-26 2013-12-26 A kind of method of large scale network stream data caching write

Publications (2)

Publication Number Publication Date
CN103699660A true CN103699660A (en) 2014-04-02
CN103699660B CN103699660B (en) 2016-10-12

Family

ID=50361188

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310741116.6A Active CN103699660B (en) 2013-12-26 2013-12-26 A kind of method of large scale network stream data caching write

Country Status (1)

Country Link
CN (1) CN103699660B (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104536699A (en) * 2014-12-11 2015-04-22 中国科学院声学研究所 Streamed data write-in method based on embedded file system
CN104598563A (en) * 2015-01-08 2015-05-06 北京京东尚科信息技术有限公司 High concurrency data storage method and device
CN104657502A (en) * 2015-03-12 2015-05-27 浪潮集团有限公司 System and method for carrying out real-time statistics on mass data based on Hadoop
WO2015180070A1 (en) * 2014-05-28 2015-12-03 北京大学深圳研究生院 Data caching method and device for distributed storage system
CN105205084A (en) * 2014-06-30 2015-12-30 清华大学 Method, device and system for processing data
CN105512168A (en) * 2015-11-16 2016-04-20 天津南大通用数据技术股份有限公司 Cluster database composite data loading method and apparatus
CN105975521A (en) * 2016-04-28 2016-09-28 乐视控股(北京)有限公司 Stream data uploading method and device
CN106484329A (en) * 2016-09-26 2017-03-08 浪潮电子信息产业股份有限公司 Big data transmission integrity protection mechanism based on multi-level storage
CN106909641A (en) * 2017-02-16 2017-06-30 青岛高校信息产业股份有限公司 A kind of real-time data memory device
CN107180082A (en) * 2017-05-03 2017-09-19 珠海格力电器股份有限公司 Data updating system and method based on multi-level cache mechanism
CN107943802A (en) * 2016-10-12 2018-04-20 北京京东尚科信息技术有限公司 A kind of log analysis method and system
CN111367979A (en) * 2020-03-05 2020-07-03 广州快决测信息科技有限公司 Data collection method and system
CN114430351A (en) * 2022-04-06 2022-05-03 北京快立方科技有限公司 Distributed database node secure communication method and system
CN117827979A (en) * 2024-03-05 2024-04-05 数翊科技(北京)有限公司武汉分公司 Data batch import method and device, electronic equipment and storage medium
US11960497B2 (en) 2020-03-05 2024-04-16 Guangzhou Quick Decision Information Technology Co., Ltd. Method and system for automatically generating data determining result

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050235108A1 (en) * 2004-04-20 2005-10-20 Hitachi Global Storage Technologies Disk device and control method for cache
CN103279429A (en) * 2013-05-24 2013-09-04 浪潮电子信息产业股份有限公司 Application-aware distributed global shared cache partition method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050235108A1 (en) * 2004-04-20 2005-10-20 Hitachi Global Storage Technologies Disk device and control method for cache
CN103279429A (en) * 2013-05-24 2013-09-04 浪潮电子信息产业股份有限公司 Application-aware distributed global shared cache partition method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
秦秀磊等: "云计算环境下分布式缓存技术的现状与挑战", 《中国期刊全文数据库 软件学报》 *

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015180070A1 (en) * 2014-05-28 2015-12-03 北京大学深圳研究生院 Data caching method and device for distributed storage system
CN105205084A (en) * 2014-06-30 2015-12-30 清华大学 Method, device and system for processing data
CN105205084B (en) * 2014-06-30 2018-10-16 清华大学 A kind of data processing method, apparatus and system
CN104536699B (en) * 2014-12-11 2017-10-17 中国科学院声学研究所 A kind of stream data wiring method based on embedded file system
CN104536699A (en) * 2014-12-11 2015-04-22 中国科学院声学研究所 Streamed data write-in method based on embedded file system
CN104598563A (en) * 2015-01-08 2015-05-06 北京京东尚科信息技术有限公司 High concurrency data storage method and device
CN104598563B (en) * 2015-01-08 2018-09-04 北京京东尚科信息技术有限公司 High concurrent date storage method and device
CN104657502A (en) * 2015-03-12 2015-05-27 浪潮集团有限公司 System and method for carrying out real-time statistics on mass data based on Hadoop
CN105512168A (en) * 2015-11-16 2016-04-20 天津南大通用数据技术股份有限公司 Cluster database composite data loading method and apparatus
CN105975521A (en) * 2016-04-28 2016-09-28 乐视控股(北京)有限公司 Stream data uploading method and device
CN106484329A (en) * 2016-09-26 2017-03-08 浪潮电子信息产业股份有限公司 Big data transmission integrity protection mechanism based on multi-level storage
CN106484329B (en) * 2016-09-26 2019-01-08 浪潮电子信息产业股份有限公司 Big data transmission integrity protection method based on multi-level storage
CN107943802A (en) * 2016-10-12 2018-04-20 北京京东尚科信息技术有限公司 A kind of log analysis method and system
CN106909641A (en) * 2017-02-16 2017-06-30 青岛高校信息产业股份有限公司 A kind of real-time data memory device
CN106909641B (en) * 2017-02-16 2020-09-29 青岛高校信息产业股份有限公司 Real-time data memory
CN107180082B (en) * 2017-05-03 2020-12-18 珠海格力电器股份有限公司 Data updating system and method based on multi-level cache mechanism
CN107180082A (en) * 2017-05-03 2017-09-19 珠海格力电器股份有限公司 Data updating system and method based on multi-level cache mechanism
CN111367979A (en) * 2020-03-05 2020-07-03 广州快决测信息科技有限公司 Data collection method and system
US11960497B2 (en) 2020-03-05 2024-04-16 Guangzhou Quick Decision Information Technology Co., Ltd. Method and system for automatically generating data determining result
CN114430351A (en) * 2022-04-06 2022-05-03 北京快立方科技有限公司 Distributed database node secure communication method and system
CN117827979A (en) * 2024-03-05 2024-04-05 数翊科技(北京)有限公司武汉分公司 Data batch import method and device, electronic equipment and storage medium
CN117827979B (en) * 2024-03-05 2024-05-17 数翊科技(北京)有限公司武汉分公司 Data batch import method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN103699660B (en) 2016-10-12

Similar Documents

Publication Publication Date Title
CN103699660A (en) Large-scale network streaming data cache-write method
CN104820670B (en) A kind of acquisition of power information big data and storage method
Ji et al. Big data processing in cloud computing environments
US8819335B1 (en) System and method for executing map-reduce tasks in a storage device
US10013440B1 (en) Incremental out-of-place updates for index structures
US20170155707A1 (en) Multi-level data staging for low latency data access
Ji et al. Big data processing: Big challenges and opportunities
US20140214752A1 (en) Data stream splitting for low-latency data access
Lai et al. Towards a framework for large-scale multimedia data storage and processing on Hadoop platform
CN107766402A (en) A kind of building dictionary cloud source of houses big data platform
CN105139281A (en) Method and system for processing big data of electric power marketing
CN103268336A (en) Fast data and big data combined data processing method and system
Devakunchari Analysis on big data over the years
CN112000703B (en) Data warehousing processing method and device, computer equipment and storage medium
Caldarola et al. Big data: A survey-the new paradigms, methodologies and tools
CN105630810A (en) Method for uploading mass small files in distributed storage system
Ji et al. Ibdp: An industrial big data ingestion and analysis platform and case studies
CN105095255A (en) Data index creating method and device
CN108763562A (en) A kind of construction method based on big data skill upgrading data exchange efficiency
Shen et al. Meteorological sensor data storage mechanism based on timescaledb and kafka
Anusha et al. Big data techniques for efficient storage and processing of weather data
CN109542960B (en) Data analysis domain system
Dong et al. Research on Architecture of Power Big Data High-Speed Storage System for Energy Interconnection
Chen et al. Analysis of plant breeding on hadoop and spark
CN107766452B (en) Indexing system suitable for high-speed access of power dispatching data and indexing method thereof

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20180111

Address after: 210042 Xuanwu District, Xuanwu District, Jiangsu, Nanjing, No. 699-22, building 18

Patentee after: CERTUSNET CORP.

Address before: 100084 Beijing Haidian District Tsinghua Yuan 100084-82 mailbox

Patentee before: Tsinghua University

TR01 Transfer of patent right