CN112579685A - State monitoring and health degree evaluation method and device for big data operation - Google Patents

State monitoring and health degree evaluation method and device for big data operation Download PDF

Info

Publication number
CN112579685A
CN112579685A CN202011644280.1A CN202011644280A CN112579685A CN 112579685 A CN112579685 A CN 112579685A CN 202011644280 A CN202011644280 A CN 202011644280A CN 112579685 A CN112579685 A CN 112579685A
Authority
CN
China
Prior art keywords
data
job
analysis
monitoring
parameters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202011644280.1A
Other languages
Chinese (zh)
Inventor
温秋荣
沈鹏
王圣玉
张程
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Wiwide Network Technology Co ltd
Original Assignee
Shanghai Wiwide Network Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Wiwide Network Technology Co ltd filed Critical Shanghai Wiwide Network Technology Co ltd
Priority to CN202011644280.1A priority Critical patent/CN112579685A/en
Publication of CN112579685A publication Critical patent/CN112579685A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/25Integrating or interfacing systems involving database management systems
    • G06F16/254Extract, transform and load [ETL] procedures, e.g. ETL data flows in data warehouses
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/3003Monitoring arrangements specially adapted to the computing system or computing system component being monitored
    • G06F11/3034Monitoring arrangements specially adapted to the computing system or computing system component being monitored where the computing system component is a storage system, e.g. DASD based or network based
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/3051Monitoring arrangements for monitoring the configuration of the computing system or of the computing system component, e.g. monitoring the presence of processing resources, peripherals, I/O links, software programs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/34Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
    • G06F11/3409Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment for performance assessment

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Quality & Reliability (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Hardware Design (AREA)
  • Mathematical Physics (AREA)
  • Debugging And Monitoring (AREA)

Abstract

The method and the device for monitoring the state and evaluating the health degree of big data operation comprise the following steps: configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis projects in advance; collecting operation running data related to the operation according to the monitoring parameters in the operation running process; ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation; and analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated. The invention establishes a unified log rule, log management configuration and log format for each big data job, thereby realizing the purposes of rapidly acquiring the operation state of the job, finding problems in time and rapidly and accurately positioning abnormity.

Description

State monitoring and health degree evaluation method and device for big data operation
Technical Field
The invention relates to the technical field of big data operation, in particular to a method and a device for monitoring the state and evaluating the health degree of big data operation.
Background
In large data jobs, when a job has a problem, it is often difficult to find and process the job in a timely manner. Specifically, due to factors such as a data source, program logic, and a running environment, it is often difficult for a server to comprehensively know the working state of a running job, and information related to the job is often recorded in each scheduling link in a fragmented manner, so that a maintainer often cannot find a job problem in advance.
Even after a problem occurs, a maintainer needs to log in a scheduling node to screen a corresponding log file, and then screen related abnormal information from a large amount of log information recorded in the log file, so that the problem troubleshooting process is complicated and consumes a long time.
Therefore, how to quickly acquire the operation state of the operation, find problems in time and quickly and accurately locate the abnormality in the big data operation is a problem to be solved urgently in the field.
The ETL (Extract-Transform-Load) refers to extracting (Extract) data of a business system (or other source end), converting (Transform), and loading (Load) the data to a data warehouse (or other destination end), so as to integrate scattered, disordered, and non-uniform data in an enterprise, and provide an analysis basis for enterprise decision making.
Disclosure of Invention
The technical problem solved by the invention is as follows: in big data operation, how to quickly acquire operation running state, find problems in time and quickly and accurately locate abnormity.
In order to solve the above technical problem, an embodiment of the present invention provides a method for monitoring a status and evaluating a health degree of big data job, including:
configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis projects in advance;
when the operation is started, acquiring operation information of the operation, and acquiring a monitoring parameter corresponding to the operation type of the operation according to the operation information;
collecting operation running data related to the operation according to the monitoring parameters in the operation running process;
ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation;
and analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated.
Optionally, the method for monitoring the status and evaluating the health degree of the big data job further includes: and generating identifiers of the jobs in the running process according to the job information and the monitoring parameters/analysis parameters, wherein different jobs have different identifiers, and the identifiers act on the starting position and the ending position of the job scheduling node to realize the association between the jobs and the scheduling environment.
Optionally, the job information includes: basic information and supplementary information of the job; wherein the supplemental information is determined by the job type and the business needs of the job.
Optionally, the monitoring parameters include: the status reporting frequency, the fault tolerance mechanism and the alarm mechanism of the operation type.
Optionally, after the obtaining of the monitoring parameter corresponding to the job type of the job, the method further includes: and creating an information acquisition thread related to the job according to the acquired job information of the job and the monitoring parameters corresponding to the job type.
Optionally, in the process of job running, collecting job running data about the job according to the monitoring parameter includes: the method comprises the steps of starting a data acquisition module according to preset starting parameters and monitoring parameters, periodically monitoring whether each operation runs or not by the data acquisition module, if the operation running is found, acquiring the monitoring parameters corresponding to the operation type of the operation according to the operation type of the operation, and acquiring operation running data related to the operation according to the monitoring parameters.
Optionally, the performing ETL processing on the job running data to obtain data to be evaluated about the job includes: and acquiring parameters related to the relevant index model, and carrying out ETL processing on the job operation data according to the parameters related to the relevant index model.
Optionally, the analysis includes a plurality of analysis items, different analysis items respectively have corresponding index data, weighting coefficients and analysis rules, and after analysis of a certain item, a subentry score of an analyzed node with respect to the analysis item is obtained.
Optionally, the total score of the analyzed node is calculated according to the itemized scores of the analyzed node on each analysis item, the weighting coefficients of each analysis item, and the analysis rules of each analysis item, and the calculation formula is as follows:
Score=A1W1X1+A2W2X2+…+AnWnXn
wherein Score denotes the overall Score, AnAnalysis rule parameters, W, representing individual analysis itemsnWeight coefficient, X, representing each analysis itemnThe itemized scores of the respective analysis items are represented.
Optionally, the job running data and the data to be evaluated of a certain node are bound with the node identifier of the corresponding node for common transmission, and the node identifier of each node adopts a secret code.
In order to solve the above technical problem, an embodiment of the present invention further provides a device for monitoring status and evaluating health degree of big data job, including:
a processor adapted to load and execute instructions of a software program;
a memory adapted to store a software program comprising instructions for performing the steps of:
configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis projects in advance;
when the operation is started, acquiring operation information of the operation, and acquiring a monitoring parameter corresponding to the operation type of the operation according to the operation information;
collecting operation running data related to the operation according to the monitoring parameters in the operation running process;
ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation;
and analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated.
Optionally, the device for monitoring status and evaluating health degree of big data job further includes: the system comprises a configuration module 201, a data acquisition module 202, an extraction service module 203 and an analysis service module 204; wherein:
the configuration module 201 is suitable for configuring monitoring parameters for various types of operation in advance and configuring analysis parameters for various analysis projects in advance;
the data acquisition module 202 is suitable for acquiring the operation data of the operation according to the monitoring parameters in the operation process of the operation;
the extraction service module 203 is suitable for performing ETL processing on the job running data to obtain data to be evaluated about the job;
the analysis service module 204 is adapted to analyze the data to be evaluated according to the analysis parameters, so as to obtain a score of a job corresponding to the analyzed data to be evaluated.
In order to solve the above technical problem, an embodiment of the present invention further provides a server, including the above status monitoring and health degree evaluating device for big data job.
Compared with the prior art, the technical scheme of the invention has the following beneficial effects:
configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis projects in advance; collecting operation running data related to the operation according to the monitoring parameters in the operation running process; ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation; and analyzing the data to be evaluated according to the analysis parameters to obtain scores of the jobs corresponding to the analyzed data to be evaluated, so that a uniform log rule, log management configuration and log format are established for each big data job, and further, the operation state of the job is quickly obtained, the problem is timely found, and the abnormality is quickly and accurately positioned.
Furthermore, a specific scoring rule is disclosed, and a total score is further obtained according to the fractional score of each analysis project, so that a uniform running state evaluation mechanism is established for each big data job.
Drawings
FIG. 1 is a flow chart of a method for status monitoring and health assessment of big data jobs according to an embodiment of the present invention;
FIG. 2 is a block diagram of a status monitoring and health assessment apparatus for big data jobs according to an embodiment of the present invention.
Detailed Description
As can be seen from the analysis in the background art, in a big data job, when a job problem occurs, it is often difficult to find and process the job in time. For example, when a problem with a software compiler causes an error in the application software, problem location is very time consuming.
How to quickly acquire the operation state, find problems in time and quickly and accurately locate the abnormality is a problem which needs to be solved urgently in the field.
After research, the inventor finds that big data operation mainly relates to aspects such as data acquisition, data preprocessing, data storage and the like. Wherein:
the data acquisition mainly comprises database acquisition and file acquisition. Generally, scripts need to be written, corresponding tools need to be called to extract and convert data, and the data are transmitted to a designated big data storage system, the process is generally scheduled and executed through a scheduling system, and job failure may be caused by problems of data rules, network connection and the like in the scheduling process. Error information about data acquisition is usually saved in a log file of a node where a script runs, and when abnormal information is checked, the node needs to be logged in, and a system command is used for checking the log information of a corresponding file.
The data preprocessing mainly comprises data cleaning, data conversion, data integration and data specification. The process is mainly completed by a computing engine, data needs to be analyzed, filtered, converted, classified and the like in the job execution process, and a job execution error may be caused due to insufficient node resource allocation, disordered data sources and the like in the operation, and the error may affect a single piece of data or multiple pieces of data, even cause the whole job to fail. Exception information regarding data preprocessing is typically stored in a log directory of a compute engine cluster, requiring a command and job ID provided by the compute engine to be used to view log information corresponding to the job.
The data storage mainly encapsulates the storage interface capability of the bottom layer so as to realize the storage and the update of data, and operation execution errors are caused by the size of data volume, the data type, the node network state and other reasons in the operation process. The abnormal information about data storage is usually stored in a log directory of a storage engine cluster, the storage engines are of different types and are suitable for different service scenes, the log directories are also different, the abnormal information for troubleshooting needs to be logged in the storage cluster, and a system command is used for checking the log information of a corresponding file.
Through the analysis, the log information is important information of the job, and whether the running logic and the job state of the job meet expectations or not can be known through the log information.
In the prior art, a program scheduled in a job can provide a log information recording and configuring method, even a log information display interface, and can check job running information more conveniently.
However, in practical applications, log records and configuration methods provided by the programs of the job scheduling are not uniform, log context information cannot be effectively associated in each scheduling link in the same job, the recorded log information is staggered and fragmented, maintenance personnel often need to repeatedly screen logs in log information of multiple job scheduling nodes, the job problem screening process is complicated, the consumed time is long, and further additional consumption of system maintenance cost is caused.
For example, in the prior art, for the case of software error, the conventional way is to locate the problem by using a debugger manually, but the problem locating efficiency of the compiler is low.
The method comprises the steps of configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis items in advance; collecting operation running data related to the operation according to the monitoring parameters in the operation running process; ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation; and analyzing the data to be evaluated according to the analysis parameters to obtain scores of the jobs corresponding to the analyzed data to be evaluated, so that a uniform log rule, log management configuration and log format are established for each big data job, and further, the operation state of the job is quickly obtained, the problem is timely found, and the abnormality is quickly and accurately positioned.
In order that those skilled in the art will better understand and realize the present invention, the following detailed description is given by way of specific embodiments with reference to the accompanying drawings.
Example one
As described below, embodiments of the present invention provide a method for monitoring status and evaluating health of big data job.
Referring to the flow chart of the status monitoring and health degree evaluation method of big data job shown in fig. 1, the following detailed description is made through specific steps:
s101, monitoring parameters are configured for various types of jobs in advance, and analysis parameters are configured for various analysis items in advance.
Further, a data acquisition module, an extraction service module and a determination service module (server as environment) are deployed on the server of the big data cluster based on the pre-configured information such as the configuration of the designated big data cluster, the configuration of job acquisition, the configuration of extraction, the configuration of calculation model and the like.
S102, when the operation is started, the operation information of the operation is obtained, and the monitoring parameters corresponding to the operation type of the operation are obtained according to the operation information.
With respect to job information, in some embodiments, the job information may include: basic information and supplementary information of the job.
Wherein the supplemental information is determined by the job type and the business needs of the job.
With respect to monitoring parameters, in some embodiments, the monitoring parameters may include: the status reporting frequency, the fault tolerance mechanism and the alarm mechanism of the operation type.
In some embodiments, after the obtaining of the monitoring parameter corresponding to the job type of the job, the method may further include: and creating an information acquisition thread related to the job according to the acquired job information of the job and the monitoring parameters corresponding to the job type.
Further, in some embodiments, the method for monitoring the status and evaluating the health of the big data job may further include: and generating identifiers of the jobs in the running process according to the job information and the monitoring parameters/analysis parameters, wherein different jobs have different identifiers, and the identifiers act on the starting position and the ending position of the job scheduling node to realize the association between the jobs and the scheduling environment.
And S103, collecting the operation data of the operation according to the monitoring parameters in the operation process of the operation.
And establishing an acquisition thread according to the acquired configuration information, collecting the operation data of the corresponding operation according to the configured operation information, and periodically submitting the acquired operation data to an acquisition server according to the monitoring parameters.
Specifically, in some embodiments, the collecting, according to the monitoring parameter, job running data about the job during the job running may include: the method comprises the steps of starting a data acquisition module according to preset starting parameters and monitoring parameters, periodically monitoring whether each operation runs or not by the data acquisition module, if the operation running is found, acquiring the monitoring parameters corresponding to the operation type of the operation according to the operation type of the operation, and acquiring operation running data related to the operation according to the monitoring parameters.
When the operation is started, the operation is initialized first, and the operation information and the monitoring parameters configured corresponding to the current operation type are acquired first in the initialization process, so that the configuration information of the latest version can be acquired.
Furthermore, when the data acquisition module is started, an initialization operation is performed first, and the starting module on each node is started based on the preset starting parameters and the preset monitoring parameters. After the starting module is started, the operation running state can be periodically monitored, if the running big data operation exists, corresponding data acquisition is carried out according to the monitoring parameters, and the data are synchronously reported to the preprocessing module.
Furthermore, because the current big data clusters are all in a lamb mode, a plurality of service nodes are usually used for executing the operation, the module generates a unique key for each operation based on the configuration of the index model and the operation related information of the service node, collects the related index data during the operation of the operation, maps and associates the key of the operation and the index data of the operation, and synchronizes the data to the extraction module, so that the problem that a plurality of index data exist when the operation is split and operated by the plurality of service nodes is solved, and even the plurality of service nodes can take the index data of each operation.
The "key" is understood to mean that in a distributed system, since a combination of a plurality of jobs calculated by a plurality of nodes is combined to form a final result, it is necessary to know which job the final result belongs to, and the "key" has a role in enabling the result of multi-node processing to be mapped onto one job, thereby realizing distributed calculation. The "key" is then reported as part of the data.
With respect to the "key" described above, in some embodiments, a "key" for a job may be formed based on the identity of the job, the server node ID associated with the job, and the serial number at which the job runs.
Further, on this basis, a "key" for a certain job may be formed based on information such as a machine room ID associated with the job, a rack ID associated with the job, and a time at which the job program is executed.
For example, the key may be:
"key" JobName _ roommid _ racldd _ NodeId _ Time _ No.
Wherein, JobName represents the identification of the operation, roommid represents the machine room ID related to the operation, RackId represents the rack ID related to the operation, NodeId represents the server node ID related to the operation, Time represents the Time of obtaining execution when the operation program runs, and No represents the serial number when the operation runs.
S104, ETL processing is carried out on the job running data, and data to be evaluated about the job are obtained.
Specifically, in some embodiments, the performing ETL processing on the job running data to obtain data to be evaluated about the job may include: and acquiring parameters related to the relevant index model, and carrying out ETL processing on the job operation data according to the parameters related to the relevant index model.
After the extraction service module receives the operation running data reported by the data acquisition module, the extraction service module loads data of the index model, ETL processing is carried out on the synchronized operation running data based on the index model data, finally, calculation data which can be directly used by the determination service module is achieved, and operation health scoring is synchronously carried out on the determination service module.
The extraction service module is used for solving the problem that the change of metadata in the data warehouse design cannot be directly consumed and used by downstream services, so that the data needs to be processed hierarchically. In some embodiments, the metadata is preprocessed, and the preprocessing method includes logic of structuring, screening and the like of the health data, and is finally provided for a data model (data to be evaluated) which can be directly analyzed at the downstream. The specific pretreatment method is not the focus of the present invention, and is not described herein.
And S105, analyzing the data to be evaluated according to the analysis parameters to obtain a score of the job corresponding to the analyzed data to be evaluated.
After the service module receives the data after the ETL of the service module is extracted, the service module is determined to load the configuration data, the analysis is performed based on the configuration data, the analysis service module has multiple analysis modes, the analysis of different analysis service modules evaluates the health score of a certain node of the cluster based on different index data, corresponding weighting coefficients and analysis parameters, and for convenience of understanding, for example, the health score may be:
Score=A1W1X1+A2W2X2+…+AnWnXn
wherein A isnIs an analysis parameter of a different analysis service module, WnRefers to the corresponding coefficient of each evaluation index parameter, which may also be a weighting coefficient, XnThe values of the analysis parameters have different judgment rules according to different analysis modules, so that the health condition of the cluster nodes can be calculated more accurately and more comprehensively.
And analyzing the health scores 'Score' provided by all the service modules finally obtained by the service modules, classifying the index data of the same job according to the provided 'key', and finally determining the (health state) Score of the job running process.
The above description of the technical solution shows that: in this embodiment, monitoring parameters are configured for various types of jobs in advance, and analysis parameters are configured for various analysis items in advance; collecting operation running data related to the operation according to the monitoring parameters in the operation running process; ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation; and analyzing the data to be evaluated according to the analysis parameters to obtain scores of the jobs corresponding to the analyzed data to be evaluated, so that a uniform log rule, log management configuration and log format are established for each big data job, and further, the operation state of the job is quickly obtained, the problem is timely found, and the abnormality is quickly and accurately positioned.
Specifically, in some embodiments, the analysis may include a plurality of analysis items, and different analysis items respectively have corresponding index data, weighting coefficients and analysis rules, and after a certain item is analyzed, a fractional score of an analyzed node with respect to the analysis item is obtained.
Further, the total score of the analyzed node can be calculated according to the itemized scores of the analyzed node relative to each analysis item, the weighting coefficients of each analysis item and the analysis rules of each analysis item, and the calculation formula is as follows:
Score=A1W1X1+A2W2X2+…+AnWnXn
wherein Score denotes the overall Score, AnAnalysis rule parameters, W, representing individual analysis itemsnWeight coefficient, X, representing each analysis itemnThe itemized scores of the respective analysis items are represented.
The above description of the technical solution shows that: in the embodiment, a specific scoring rule is further disclosed, and a total score is further obtained according to the subentry scores of all analysis items, so that a uniform running state evaluation mechanism is established for all big data jobs.
In some embodiments, job running data and data to be evaluated of a certain node may be bound with a node identifier of a corresponding node for common transmission, and the node identifier of each node adopts a secret code.
Example two
As described below, embodiments of the present invention provide a status monitoring and health evaluation device for big data job.
The state monitoring and health degree evaluation device for big data operation comprises:
a processor adapted to load and execute instructions of a software program;
a memory adapted to store a software program comprising instructions for performing the steps of:
configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis projects in advance;
when the operation is started, acquiring operation information of the operation, and acquiring a monitoring parameter corresponding to the operation type of the operation according to the operation information;
collecting operation running data related to the operation according to the monitoring parameters in the operation running process;
ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation;
and analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated.
In some embodiments, as shown in fig. 2, the status monitoring and health evaluating apparatus for big data job may further include: the system comprises a configuration module, a data acquisition module, an extraction service module and an analysis service module; wherein:
the configuration module is suitable for configuring monitoring parameters for various types of operation in advance and configuring analysis parameters for various analysis projects in advance;
the data acquisition module is suitable for acquiring operation running data related to the operation according to the monitoring parameters in the operation running process;
the extraction service module is suitable for carrying out ETL processing on the operation data to obtain data to be evaluated related to the operation;
and the analysis service module is suitable for analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated.
The above description of the technical solution shows that: in this embodiment, monitoring parameters are configured for various types of jobs in advance, and analysis parameters are configured for various analysis items in advance; collecting operation running data related to the operation according to the monitoring parameters in the operation running process; ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation; and analyzing the data to be evaluated according to the analysis parameters to obtain scores of the jobs corresponding to the analyzed data to be evaluated, so that a uniform log rule, log management configuration and log format are established for each big data job, and further, the operation state of the job is quickly obtained, the problem is timely found, and the abnormality is quickly and accurately positioned.
EXAMPLE III
As described below, embodiments of the present invention provide a server.
The difference from the prior art is that the server includes a status monitoring and health assessment device for big data jobs as provided in the embodiments of the present invention. Therefore, the server can configure monitoring parameters for various types of operation in advance and analysis parameters for various analysis items in advance; collecting operation running data related to the operation according to the monitoring parameters in the operation running process; ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation; and analyzing the data to be evaluated according to the analysis parameters to obtain scores of the jobs corresponding to the analyzed data to be evaluated, so that a uniform log rule, log management configuration and log format are established for each big data job, and further, the operation state of the job is quickly obtained, the problem is timely found, and the abnormality is quickly and accurately positioned.
Those skilled in the art will understand that, in the methods of the embodiments, all or part of the steps can be performed by hardware associated with program instructions, and the program can be stored in a computer-readable storage medium, which can include: ROM, RAM, magnetic or optical disks, and the like.
Although the present invention is disclosed above, the present invention is not limited thereto. Various changes and modifications may be effected therein by one skilled in the art without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (13)

1. A method for monitoring the state and evaluating the health degree of big data operation is characterized by comprising the following steps:
configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis projects in advance;
when the operation is started, acquiring operation information of the operation, and acquiring a monitoring parameter corresponding to the operation type of the operation according to the operation information;
collecting operation running data related to the operation according to the monitoring parameters in the operation running process;
ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation;
and analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated.
2. The method for status monitoring and health assessment of big data jobs according to claim 1, wherein the method for status monitoring and health assessment of big data jobs further comprises: and generating identifiers of the jobs in the running process according to the job information and the monitoring parameters/analysis parameters, wherein different jobs have different identifiers, and the identifiers act on the starting position and the ending position of the job scheduling node to realize the association between the jobs and the scheduling environment.
3. The method for status monitoring and health assessment of big data jobs according to claim 1, wherein the job information comprises: basic information and supplementary information of the job; wherein the supplemental information is determined by the job type and the business needs of the job.
4. The method for status monitoring and health assessment of big data jobs according to claim 1, wherein the monitoring parameters comprise: the status reporting frequency, the fault tolerance mechanism and the alarm mechanism of the operation type.
5. The method for status monitoring and health assessment of big data job according to claim 1, further comprising after obtaining the monitoring parameter corresponding to the job type of the job: and creating an information acquisition thread related to the job according to the acquired job information of the job and the monitoring parameters corresponding to the job type.
6. The method for monitoring the status and evaluating the health of the big data job according to claim 1, wherein the collecting the job running data related to the job according to the monitoring parameters during the running of the job comprises: the method comprises the steps of starting a data acquisition module according to preset starting parameters and monitoring parameters, periodically monitoring whether each operation runs or not by the data acquisition module, if the operation running is found, acquiring the monitoring parameters corresponding to the operation type of the operation according to the operation type of the operation, and acquiring operation running data related to the operation according to the monitoring parameters.
7. The method for status monitoring and health assessment of big data jobs according to claim 1, wherein the ETL processing of the job running data to obtain the data to be assessed about the job comprises: and acquiring parameters related to the relevant index model, and carrying out ETL processing on the job operation data according to the parameters related to the relevant index model.
8. The method according to claim 1, wherein the analysis includes a plurality of analysis items, different analysis items have corresponding index data, weighting coefficients and analysis rules, and after a certain item is analyzed, a sub-item score of the analyzed node with respect to the analysis item is obtained.
9. The method according to claim 8, wherein the total score of the analyzed node is calculated according to the itemized scores of the analyzed nodes with respect to the analysis items, the weighting coefficients of the analysis items, and the analysis rules of the analysis items, and the calculation formula is as follows:
Score=A1W1X1+A2W2X2+...+AnWnXn
wherein Score denotes the overall Score, AnAnalysis rule parameters, W, representing individual analysis itemsnWeight coefficient, X, representing each analysis itemnThe itemized scores of the respective analysis items are represented.
10. The method for monitoring the status and evaluating the health of the big data job according to claim 1, wherein the job running data and the data to be evaluated of a certain node are bound with the node identifier of the corresponding node and transmitted together, and the node identifier of each node adopts a secret code.
11. A status monitoring and health degree evaluation device for big data operation is characterized by comprising:
a processor adapted to load and execute instructions of a software program;
a memory adapted to store a software program comprising instructions for performing the steps of:
configuring monitoring parameters for various types of operation in advance, and configuring analysis parameters for various analysis projects in advance;
when the operation is started, acquiring operation information of the operation, and acquiring a monitoring parameter corresponding to the operation type of the operation according to the operation information;
collecting operation running data related to the operation according to the monitoring parameters in the operation running process;
ETL processing is carried out on the operation data to obtain data to be evaluated related to the operation;
and analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated.
12. The big data job status monitoring and health assessment apparatus according to claim 11, further comprising: the system comprises a configuration module, a data acquisition module, an extraction service module and an analysis service module; wherein:
the configuration module is suitable for configuring monitoring parameters for various types of operation in advance and configuring analysis parameters for various analysis projects in advance;
the data acquisition module is suitable for acquiring operation running data related to the operation according to the monitoring parameters in the operation running process;
the extraction service module is suitable for carrying out ETL processing on the operation data to obtain data to be evaluated related to the operation;
and the analysis service module is suitable for analyzing the data to be evaluated according to the analysis parameters to obtain the scores of the jobs corresponding to the analyzed data to be evaluated.
13. A server, characterized by comprising a status monitoring and health assessment device for big data jobs according to any one of claims 11 to 12.
CN202011644280.1A 2020-12-31 2020-12-31 State monitoring and health degree evaluation method and device for big data operation Withdrawn CN112579685A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011644280.1A CN112579685A (en) 2020-12-31 2020-12-31 State monitoring and health degree evaluation method and device for big data operation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011644280.1A CN112579685A (en) 2020-12-31 2020-12-31 State monitoring and health degree evaluation method and device for big data operation

Publications (1)

Publication Number Publication Date
CN112579685A true CN112579685A (en) 2021-03-30

Family

ID=75145539

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011644280.1A Withdrawn CN112579685A (en) 2020-12-31 2020-12-31 State monitoring and health degree evaluation method and device for big data operation

Country Status (1)

Country Link
CN (1) CN112579685A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113240396A (en) * 2021-05-20 2021-08-10 北京明略昭辉科技有限公司 Method, device and equipment for analyzing working state of employee and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106411609A (en) * 2016-11-08 2017-02-15 上海新炬网络信息技术有限公司 IT software and hardware running state monitoring system
CN110378564A (en) * 2019-06-18 2019-10-25 中国平安财产保险股份有限公司 Monitoring model generation method, device, terminal device and storage medium
CN111274087A (en) * 2020-01-15 2020-06-12 国网湖南省电力有限公司 Health degree evaluation method of IT centralized monitoring business system
CN111552607A (en) * 2020-03-27 2020-08-18 深圳壹账通智能科技有限公司 Health evaluation method, device and equipment of application program and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106411609A (en) * 2016-11-08 2017-02-15 上海新炬网络信息技术有限公司 IT software and hardware running state monitoring system
CN110378564A (en) * 2019-06-18 2019-10-25 中国平安财产保险股份有限公司 Monitoring model generation method, device, terminal device and storage medium
CN111274087A (en) * 2020-01-15 2020-06-12 国网湖南省电力有限公司 Health degree evaluation method of IT centralized monitoring business system
CN111552607A (en) * 2020-03-27 2020-08-18 深圳壹账通智能科技有限公司 Health evaluation method, device and equipment of application program and storage medium

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113240396A (en) * 2021-05-20 2021-08-10 北京明略昭辉科技有限公司 Method, device and equipment for analyzing working state of employee and storage medium

Similar Documents

Publication Publication Date Title
Mayer et al. An approach to extract the architecture of microservice-based software systems
US11968264B2 (en) Systems and methods for operation management and monitoring of bots
US8667334B2 (en) Problem isolation in a virtual environment
CN108521339B (en) Feedback type node fault processing method and system based on cluster log
Ghoshal et al. Provenance from log files: a BigData problem
US8661125B2 (en) System comprising probe runner, monitor, and responder with associated databases for multi-level monitoring of a cloud service
US20070282470A1 (en) Method and system for capturing and reusing intellectual capital in IT management
CN102143008A (en) Method and device for diagnosing fault event in data center
CN111124806B (en) Method and system for monitoring equipment state in real time based on distributed scheduling task
CN111563130A (en) Data credible data management method and system based on block chain technology
CN111858251B (en) Data security audit method and system based on big data computing technology
CN111552556A (en) GPU cluster service management system and method
US7162390B2 (en) Framework for collecting, storing, and analyzing system metrics
CN116009428A (en) Industrial data monitoring system and method based on stream computing engine and medium
CN112579685A (en) State monitoring and health degree evaluation method and device for big data operation
KR101830936B1 (en) Performance Improving System Based Web for Database and Application
CN117422434A (en) Wisdom fortune dimension dispatch platform
CN112579552A (en) Log storage and calling method, device and system
EP4174596A1 (en) System and method for collecting mes information
Iuhasz et al. Monitoring of exascale data processing
CN112181759A (en) Method for monitoring micro-service performance and diagnosing abnormity
US20060167923A1 (en) Method and a system for process discovery
CN112667469A (en) Method, system and readable medium for automatically generating diversified big data statistical report
US20230055902A1 (en) Method for representing a distributed computing system by graph embedding
Ren et al. Anomaly analysis and diagnosis for co-located datacenter workloads in the alibaba cluster

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20210330

WW01 Invention patent application withdrawn after publication