CN110689245A - Method and system for analyzing call relation of big data workflow - Google Patents

Method and system for analyzing call relation of big data workflow Download PDF

Info

Publication number
CN110689245A
CN110689245A CN201910877666.8A CN201910877666A CN110689245A CN 110689245 A CN110689245 A CN 110689245A CN 201910877666 A CN201910877666 A CN 201910877666A CN 110689245 A CN110689245 A CN 110689245A
Authority
CN
China
Prior art keywords
workflow
file
name
work
big data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910877666.8A
Other languages
Chinese (zh)
Other versions
CN110689245B (en
Inventor
徐涛
吴峰
郭伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wheel Interconnection Technology Shanghai Co ltd
Original Assignee
Shanghai Yidianshikong Network Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Yidianshikong Network Co Ltd filed Critical Shanghai Yidianshikong Network Co Ltd
Priority to CN201910877666.8A priority Critical patent/CN110689245B/en
Publication of CN110689245A publication Critical patent/CN110689245A/en
Application granted granted Critical
Publication of CN110689245B publication Critical patent/CN110689245B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0633Workflow analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • G06F16/2282Tablespace storage structures; Management thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/242Query formulation
    • G06F16/2433Query languages
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

A method for analyzing call relations of big data workflows comprises the following steps: storing metadata of a workflow scheduling engine, and obtaining the name of a workflow and the storage position of a description file; reading a first file according to the storage position, wherein the first file describes all job types and execution sequences of the current workflow; reading the first file, and obtaining the operation type and the execution file; writing the name of the workflow and the execution file into a first table, wherein the first table records the workflow attribution of the current operation; and when the job type is hive, continuously reading the execution file, and obtaining a target table and a source table in a data table. By the method for analyzing the call relation of the big data workflow, the relation between the big data workflow scheduling engine workflow and the operation and the relation between the operation and the operation are analyzed. A global invocation relationship chain may be established to assist in querying and analyzing dependencies between them.

Description

Method and system for analyzing call relation of big data workflow
Technical Field
The application relates to the field of databases, in particular to a method and a system for analyzing call relation of big data workflow.
Background
The existing big data workflow scheduling engines mainly comprise Oozie, Azkaban, AirFlow and the like, a DAG graph is used for providing a group of operations and flows which are run in a workflow in a specific sequence, and the supported operation types comprise java, map-reduce, hive, shell, python and the like. However, none of the existing workflow scheduling engines provides a method for querying which workflow a certain scheduling job belongs to, and the existing workflow scheduling engines need to manually traverse all workflows one by one to search. The invention provides a method for analyzing the scheduling relation between jobs and workflows, which helps to quickly find the workflow to which the jobs belong.
The disadvantages are as follows:
1. the existing big data workflow scheduling engine does not provide a method for inquiring the workflow to which the job belongs
2. The existing big data workflow scheduling engine does not provide the dependency relationship between internal tables of hive jobs
Disclosure of Invention
The main purpose of the present application is to provide a method for analyzing a big data workflow invocation relationship, which includes:
storing metadata of a workflow scheduling engine, and obtaining the name of a workflow and the storage position of a description file;
reading a first file according to the storage position, wherein the first file describes all job types and execution sequences of the current workflow;
reading the first file, and obtaining the operation type and the execution file;
writing the name of the workflow and the execution file into a first table, wherein the first table records the workflow attribution of the current operation;
and when the job type is hive, continuously reading the execution file, and obtaining a target table and a source table in a data table.
Optionally, the workflow attribution of the job execution file is queried by executing the file table, and the workflow attribution of the data table generated by querying the data warehouse is queried by executing the file table and the word _ table join.
Optionally, the metadata of ozie is stored through mysql, and WF _ JOBS table is searched through sql below, so as to obtain the name app _ name of the workflow and the saving position app _ path of the description file in hdfs.
Optionally, the first file is a workflow file that reads the lower side of the folder according to the location of the app _ path.
Optionally, the job type work _ type is obtained through a name attribute of an action tag, the execution file work _ script is obtained through a value of a script, and the app _ name and the work _ script are written into an app _ work table of mysql.
Optionally, when the work _ type is live, continuing to read the work _ script file, where the work _ script file records sql that needs to be executed by the job, and analyzing the sql by the following regular expressions to obtain a target table and a source table in a data table:
target table name dst _ table: insert [ \\ n \ t ] + overlay [ \\ n \ t ] + table [ \\ n \ t ] + ([ ^ n \ n ^ t ] +) source table name src _ table: [ \\ n \ t ] from [ \ n \ t ] + ([ ^ n \ n ^ t ] +) or join [ \ n \ t ] + ([ ^ n \ n ^ t ] +)
And writing the work _ script, the dst _ table and the src _ table into a work _ table of mysql, wherein the work _ table records the dependency relationship of the hive operation table.
The application also provides a system for analyzing the call relation of the big data workflow, which comprises the following steps:
the storage module is used for storing the metadata of the workflow scheduling engine and acquiring the name of the workflow and the storage position of the description file;
the reading module is used for reading a first file according to the storage position, wherein the first file describes all job types and execution sequences of the current workflow, and the job types and the execution files are obtained;
the writing module is used for writing the name of the workflow and the execution file into a first table, and the first table records the workflow attribution of the current operation;
and the obtaining module is used for continuously reading the execution file when the operation type is hive, and obtaining a target table and a source table in a data table.
The application also discloses a computer device, which comprises a memory, a processor and a computer program stored in the memory and capable of being executed by the processor, wherein the processor realizes the method of any one of the above items when executing the computer program.
The application also discloses a computer-readable storage medium, a non-volatile readable storage medium, having stored therein a computer program which, when executed by a processor, implements the method of any of the above.
The present application also discloses a computer program product comprising computer readable code which, when executed by a computer device, causes the computer device to perform the method of any of the above.
By the method for analyzing the call relation of the big data workflow, the relation between the big data workflow scheduling engine workflow and the operation and the relation between the operation and the operation are analyzed. A global invocation relationship chain may be established to assist in querying and analyzing dependencies between them.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this application, serve to provide a further understanding of the application and to enable other features, objects, and advantages of the application to be more apparent. The drawings and their description illustrate the embodiments of the invention and do not limit it. In the drawings:
FIG. 1 is a schematic flow diagram of a method of analyzing big data workflow invocation relationships, according to one embodiment of the present application;
FIG. 2 is a flowchart illustrating a method for analyzing call relations of a big data workflow according to an embodiment of the present application
FIG. 3 is a schematic diagram of a computer device according to one embodiment of the present application; and
FIG. 4 is a schematic diagram of a computer-readable storage medium according to one embodiment of the present application.
Detailed Description
In order to make the technical solutions better understood by those skilled in the art, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only partial embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
It should be noted that the terms "comprises" and "comprising," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
Furthermore, the terms "mounted," "disposed," "provided," "connected," and "sleeved" are to be construed broadly. For example, it may be a fixed connection, a removable connection, or a unitary construction; can be a mechanical connection, or an electrical connection; may be directly connected, or indirectly connected through intervening media, or may be in internal communication between two devices, elements or components. The specific meaning of the above terms in the present application can be understood by those of ordinary skill in the art as appropriate.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
As shown in fig. 1, an embodiment of the present application provides a method for analyzing a call relationship of a big data workflow, including:
s2: storing metadata of a workflow scheduling engine, and obtaining the name of a workflow and the storage position of a description file;
s4: reading a first file according to the storage position, wherein the first file describes all job types and execution sequences of the current workflow;
s6: reading the first file, and obtaining the operation type and the execution file;
s8: writing the name of the workflow and the execution file into a first table, wherein the first table records the workflow attribution of the current operation;
s10: and when the job type is hive, continuously reading the execution file, and obtaining a target table and a source table in a data table.
In an embodiment of the application, the workflow attribution of the job execution file is queried through the execution file table, and the workflow attribution of the data table generated by the data warehouse is queried through the execution file table and the word _ table join.
In an embodiment of the present application, the mysql is used to store metadata of Oozie, and the WF _ JOBS table is searched by the following sql to obtain the name app _ name of the workflow and the storage location app _ path of the description file in hdfs.
In an embodiment of the present application, the first file is a workflow.xml file that reads the bottom of the folder according to the location of the app _ path.
In an embodiment of the application, the job type work _ type is obtained through a name attribute of an action tag, the execution file work _ script is obtained through a value of a script, and the app _ name and the work _ script are written into an app _ work table of mysql.
In an embodiment of the application, when work _ type is live, continuing to read the work _ script file, where the work _ script file records sql to be executed for a job change, and analyzing the sql through the following regular expressions to obtain a target table and a source table in a data table:
target table name dst _ table: insert [ \\ n \ t ] + overlay [ \\ n \ t ] + table [ \\ n \ t ] + ([ ^ n \ n ^ t ] +) source table name src _ table: [ \\ n \ t ] from [ \ n \ t ] + ([ ^ n \ n ^ t ] +) or join [ \ n \ t ] + ([ ^ n \ n ^ t ] +)
And writing the work _ script, the dst _ table and the src _ table into a work _ table of mysql, wherein the work _ table records the dependency relationship of the hive operation table.
The application also provides a system for analyzing the call relation of the big data workflow, which comprises the following steps:
the storage module is used for storing the metadata of the workflow scheduling engine and acquiring the name of the workflow and the storage position of the description file;
the reading module is used for reading a first file according to the storage position, wherein the first file describes all job types and execution sequences of the current workflow, and the job types and the execution files are obtained;
the writing module is used for writing the name of the workflow and the execution file into a first table, and the first table records the workflow attribution of the current operation;
and the obtaining module is used for continuously reading the execution file when the operation type is hive, and obtaining a target table and a source table in a data table.
Referring to fig. 3, the present application further provides a computer device including a memory, a processor, and a computer program stored in the memory and executable by the processor, wherein the processor implements the method of any one of the above methods when executing the computer program.
Referring to fig. 4, the present application also provides a computer-readable storage medium, a non-volatile readable storage medium, having stored therein a computer program, which when executed by a processor implements the method of any of the above.
The present application also provides a computer program product comprising computer readable code which, when executed by a computer device, causes the computer device to perform the method of any of the above.
Referring to fig. 2, in an embodiment of the present application, a process of analyzing a call relationship of a big data workflow scheduling engine (Oozie, Azkaban, AirFlow, etc.) is shown in the above diagram, and specific analysis steps are described in detail below with Oozie as a column:
1. the Oozie metadata is stored by using mysql, and the WF _ JOBS table is searched by using the following sql, so that the names app _ name of all workflows and the position app _ path of the workflow description file stored in hdfs can be obtained.
SELECT max(id)id,app_name,max(app_path)app_path FROM WF_JOBS GROUP BYapp_name
2. Reading a workflow file below the folder according to the position of the app _ path, wherein the workflow file describes all job types and job execution sequences in the current workflow
3. Xml file is read, the job type work _ type can be obtained by the name attribute of < action > tag, and the job execution file work _ script can be obtained by the value of < script >. And writing the app _ name and the work _ script in the step 1 into an app _ work table of mysql, wherein the table records which workflow the job belongs to.
4. If work _ type is live, continuing to read a work _ script file, recording the sql which needs to be executed by the job, and analyzing the sql by using the following regular expression to find out a target table and a source table in the data table:
target table name dst _ table: insert [ \\ n \ t ] + overlay [ \\ n \ t ] + table [ \ n \ t ] + ([ ^ n \ n ^ t ] +)
Source table name src _ table: [ \\ n \ t ] from [ \ n \ t ] + ([ ^ n \ n ^ t ] +) or join [ \ n \ t ] + ([ ^ n \ n ^ t ] +)
Writing work _ script, dst _ table and src _ table into mysql work _ table, wherein the table records the dependency relationship of the table in hive operation
5. The workflow to which the job execution file belongs can be inquired by using the work _ script table, and the workflow in which the data table generated by the data warehouse is generated can be inquired by performing join on the work _ script table and the work _ table.
The basic concepts that may be used in this application are as follows:
big data:
https://wiki.mbalib.com/wiki/%E5%A4%A7%E6%95%B0%E6%8D%AE
Hadoop:https://hadoop.apache.org/
big data scheduling engine:
https://www.cnblogs.com/barneywill/p/10109820.html
Oozie:https://oozie.apache.org/
Hive:https://www.yiibai.com/hive/
and (3) data warehouse:
library https:// baike.baidu.com/% E6% 95% B0% E6% 8D% AE% E4% BB% 93% E5% BA% 93
mysql: is an open source code relational database management system (RDBMS) using the most common databases
Management language-Structured Query Language (SQL) for database management
The regular expression is as follows: also known as regular expressions, is a concept of computer science. Regular expressions are typically used to retrieve, replace, text that conforms to a certain pattern (rule).
A DAG: directed acyclic graphs, which are graphs in which any one side has a direction and no loop exists
Oozie: is a workflow scheduler for managing Apache Hadoop jobs. Oozie combines multiple jobs sequentially into one logical unit of work as a Directed Acyclic Graph (DAG) of operations
hive, a data warehouse tool based on Hadoop can map the structured data file into a database table, provide a simple sql query function and convert the sql statement into a MapReduce task to run
hdfs: HDFS Hadoop distributtered FILE SYSTEM is short for, and is a DISTRIBUTED file system
sql: is a standard database language
By the method for analyzing the call relation of the big data workflow, the relation between the big data workflow scheduling engine workflow and the operation and the relation between the operation and the operation are analyzed. A global invocation relationship chain may be established to assist in querying and analyzing dependencies between them.
The above description is only a preferred embodiment of the present application and is not intended to limit the present application, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (10)

1. A method for analyzing call relation of big data workflow is characterized by comprising the following steps:
storing metadata of a workflow scheduling engine, and obtaining the name of a workflow and the storage position of a description file;
reading a first file according to the storage position, wherein the first file describes all job types and execution sequences of the current workflow;
reading the first file, and obtaining the operation type and the execution file;
writing the name of the workflow and the execution file into a first table, wherein the first table records the workflow attribution of the current operation;
and when the job type is hive, continuously reading the execution file, and obtaining a target table and a source table in a data table.
2. The method for analyzing call relation of big data workflow according to claim 1, wherein workflow attribution of job execution file is inquired by executing file table, and workflow attribution of data table generated by data warehouse is inquired by executing file table and word _ table join.
3. The method for analyzing the call relation of the big data workflow as claimed in claim 2, wherein the name app _ name of the workflow and the saving location app _ path of the description file in hdfs are obtained by storing the metadata of Oozie through mysql and looking up the WF _ JOBS table through the following sql.
4. The method for analyzing big data workflow invocation relationship according to claim 3, wherein said first file is a workflow.xml file under the folder read according to the location of said app _ path.
5. The method for analyzing call relation of big data workflow as claimed in claim 4, wherein the job type work _ type is obtained by name attribute of action tag, the execution file work _ script is obtained by value of script, and the app _ name and the work _ script are written into app _ work table of mysql.
6. The method for analyzing call relation of big data workflow as claimed in claim 5, wherein when work _ type is live, continuing to read the work _ script file, the work _ script file records sql that needs to be executed for job change, and the target table and the source table in the data table are obtained by analyzing sql through the following regular expressions:
target table name dst _ table: insert [ \\ n \ t ] + overlay [ \\ n \ t ] + table [ \ n \ t ] + ([ ^ n \ n ^ t ] +)
Source table name src _ table: [ \\ n \ t ] from [ \ n \ t ] + ([ ^ n \ n ^ t ] +) or join [ \ n \ t ] + ([ ^ n \ n ^ t ] +)
And writing the work _ script, the dst _ table and the src _ table into a work _ table of mysql, wherein the work _ table records the dependency relationship of the hive operation table.
7. A system for analyzing call relations of a big data workflow, comprising:
the storage module is used for storing the metadata of the workflow scheduling engine and acquiring the name of the workflow and the storage position of the description file;
the reading module is used for reading a first file according to the storage position, wherein the first file describes all job types and execution sequences of the current workflow, and the job types and the execution files are obtained;
the writing module is used for writing the name of the workflow and the execution file into a first table, and the first table records the workflow attribution of the current operation;
and the obtaining module is used for continuously reading the execution file when the operation type is hive, and obtaining a target table and a source table in a data table.
8. A computer device comprising a memory, a processor and a computer program stored in the memory and executable by the processor, the processor implementing the method of any one of claims 1-6 when executing the computer program.
9. A computer-readable storage medium, a non-transitory readable storage medium, having stored therein a computer program, which when executed by a processor, implements the method of any one of claims 1-6.
10. A computer program product comprising computer readable code which, when executed by a computer device, causes the computer device to perform the method of any of claims 1-6.
CN201910877666.8A 2019-09-17 2019-09-17 Method and system for analyzing call relation of big data workflow Active CN110689245B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910877666.8A CN110689245B (en) 2019-09-17 2019-09-17 Method and system for analyzing call relation of big data workflow

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910877666.8A CN110689245B (en) 2019-09-17 2019-09-17 Method and system for analyzing call relation of big data workflow

Publications (2)

Publication Number Publication Date
CN110689245A true CN110689245A (en) 2020-01-14
CN110689245B CN110689245B (en) 2022-07-12

Family

ID=69109305

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910877666.8A Active CN110689245B (en) 2019-09-17 2019-09-17 Method and system for analyzing call relation of big data workflow

Country Status (1)

Country Link
CN (1) CN110689245B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111651460A (en) * 2020-06-11 2020-09-11 上海德易车信息科技有限公司 Data management method and device, electronic equipment and readable storage medium
CN112506957A (en) * 2020-12-18 2021-03-16 杭州数梦工场科技有限公司 Method and device for determining workflow dependency relationship
CN115525680A (en) * 2022-09-21 2022-12-27 京信数据科技有限公司 Data processing job scheduling method and device, computer equipment and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101645180A (en) * 2009-05-31 2010-02-10 北京龙德时代科技发展有限公司 Intelligent data acquisition unit
CN106020955A (en) * 2016-05-12 2016-10-12 深圳市傲天科技股份有限公司 Infinite big data workflow dispatching platform
US20170193069A1 (en) * 2015-12-31 2017-07-06 Dropbox, Inc. Propagating Computable Dependencies Within Synchronized Content Items Between First and Third-Party Applications
CN107105009A (en) * 2017-03-22 2017-08-29 北京荣之联科技股份有限公司 Job scheduling method and device based on Kubernetes system docking workflow engines
CN107689982A (en) * 2017-06-25 2018-02-13 平安科技(深圳)有限公司 Multi-data source method of data synchronization, application server and computer-readable recording medium
CN109670780A (en) * 2018-12-03 2019-04-23 中国建设银行股份有限公司 Workflow processing method, equipment and storage medium under complex scene
US20190361718A1 (en) * 2018-05-25 2019-11-28 Vmware, Inc. 3D API Redirection for Virtual Desktop Infrastructure

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101645180A (en) * 2009-05-31 2010-02-10 北京龙德时代科技发展有限公司 Intelligent data acquisition unit
US20170193069A1 (en) * 2015-12-31 2017-07-06 Dropbox, Inc. Propagating Computable Dependencies Within Synchronized Content Items Between First and Third-Party Applications
CN106020955A (en) * 2016-05-12 2016-10-12 深圳市傲天科技股份有限公司 Infinite big data workflow dispatching platform
CN107105009A (en) * 2017-03-22 2017-08-29 北京荣之联科技股份有限公司 Job scheduling method and device based on Kubernetes system docking workflow engines
CN107689982A (en) * 2017-06-25 2018-02-13 平安科技(深圳)有限公司 Multi-data source method of data synchronization, application server and computer-readable recording medium
US20190361718A1 (en) * 2018-05-25 2019-11-28 Vmware, Inc. 3D API Redirection for Virtual Desktop Infrastructure
CN109670780A (en) * 2018-12-03 2019-04-23 中国建设银行股份有限公司 Workflow processing method, equipment and storage medium under complex scene

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111651460A (en) * 2020-06-11 2020-09-11 上海德易车信息科技有限公司 Data management method and device, electronic equipment and readable storage medium
CN112506957A (en) * 2020-12-18 2021-03-16 杭州数梦工场科技有限公司 Method and device for determining workflow dependency relationship
CN115525680A (en) * 2022-09-21 2022-12-27 京信数据科技有限公司 Data processing job scheduling method and device, computer equipment and storage medium

Also Published As

Publication number Publication date
CN110689245B (en) 2022-07-12

Similar Documents

Publication Publication Date Title
CN110689245B (en) Method and system for analyzing call relation of big data workflow
CN107463632B (en) Distributed NewSQL database system and data query method
Hagedorn et al. The STARK framework for spatio-temporal data analytics on spark
US11423082B2 (en) Methods and apparatus for subgraph matching in big data analysis
US8898181B2 (en) Subscription for integrating external data from external system
JP6964384B2 (en) Methods, programs, and systems for the automatic discovery of relationships between fields in a mixed heterogeneous data source environment.
CN105630881A (en) Data storage method and query method for RDF (Resource Description Framework)
US9406018B2 (en) Systems and methods for semantic data integration
Santos et al. Data models in NoSQL databases for big data contexts
CN110555035A (en) Method and device for optimizing query statement
CN104156385B (en) A kind of method and apparatus of processing time sequence data
US20200159722A1 (en) Presenting updated data using persisting views
US10095737B2 (en) Information storage system
US9773003B2 (en) Computer implemented system and method for investigative data analytics
CN114564482A (en) Multi-entity-oriented label system and processing method
JPWO2017170459A6 (en) Method, program, and system for automatic discovery of relationships between fields in a heterogeneous data source mixed environment
RU2015112157A (en) SYSTEM AND METHOD OF DATA SEARCH IN THE DATABASE OF GRAPHS
US8825596B2 (en) Systems and methods for robust data source access
CN109471904B (en) Method and system for organizing labels
JP2020160494A (en) Information processing apparatus, document management system and program
US10324927B2 (en) Data-driven union pruning in a database semantic layer
JP2004192657A (en) Information retrieval system, and recording medium recording information retrieval method and program for information retrieval
JP5998835B2 (en) Information processing apparatus and program
US11126604B2 (en) Aggregation apparatus, aggregation method, and storage medium
GB2596769A (en) Extensible data skipping

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP03 Change of name, title or address
CP03 Change of name, title or address

Address after: 200125, Room 325, 3rd Floor, Unit 2, No. 231, Shibocun Road, China (Shanghai) Pilot Free Trade Zone, Pudong New Area, Shanghai

Patentee after: Wheel interconnection technology (Shanghai) Co.,Ltd.

Address before: 200125 Room 501, 5 / F, building 3, 3601 Dongfang Road, Pudong New Area, Shanghai

Patentee before: SHANGHAI YIDIAN SPACE NETWORK Co.,Ltd.