CN117493583A - Method and system for generating flow operation sequence by combining event log and knowledge graph - Google Patents

Method and system for generating flow operation sequence by combining event log and knowledge graph Download PDF

Info

Publication number
CN117493583A
CN117493583A CN202410006733.XA CN202410006733A CN117493583A CN 117493583 A CN117493583 A CN 117493583A CN 202410006733 A CN202410006733 A CN 202410006733A CN 117493583 A CN117493583 A CN 117493583A
Authority
CN
China
Prior art keywords
event
representing
sequence
flow
candidate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202410006733.XA
Other languages
Chinese (zh)
Other versions
CN117493583B (en
Inventor
袁水平
应泽宇
高元新
吴信东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Anhui Sigao Intelligent Technology Co ltd
Original Assignee
Anhui Sigao Intelligent Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Anhui Sigao Intelligent Technology Co ltd filed Critical Anhui Sigao Intelligent Technology Co ltd
Priority to CN202410006733.XA priority Critical patent/CN117493583B/en
Priority claimed from CN202410006733.XA external-priority patent/CN117493583B/en
Publication of CN117493583A publication Critical patent/CN117493583A/en
Application granted granted Critical
Publication of CN117493583B publication Critical patent/CN117493583B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • G06N3/0442Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Animal Behavior & Ethology (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention provides a flow operation sequence generation method and system combining an event log and a knowledge graph, which relate to the technical field of knowledge graphs, utilize historical event log information and related flow knowledge graphs related to a target generation sequence flow, flexibly and comprehensively represent events by combining a transform technology, learn context information existing in the event log by using LSTM, learn evolution mode information existing in the knowledge graph by using SGNN technology, provide next event prediction by combining the context information and the event evolution mode, and iteratively run on the basis, so as to generate a reliable flow operation sequence. The beneficial effects of the invention are as follows: by combining time sequence data and a knowledge graph method, the next event prediction is realized more flexibly and accurately, and the method has better universality and effectiveness.

Description

Method and system for generating flow operation sequence by combining event log and knowledge graph
Technical Field
The invention relates to the technical field of knowledge graphs, in particular to a method and a system for generating a flow operation sequence by combining an event log and a knowledge graph.
Background
With the trend of digital transformation and the convenience of electronic information storage, event logs of business systems are growing on a huge scale. The massive event logs can enable organizations to discover, monitor and improve business through process mining methods, while organizations are actively investing in providing solutions through predictive analysis and automated techniques to gain insight into and improve their performance because of the rapid changes in markets, which organizations want to adapt more quickly and accurately to changes in markets, rather than improving through post-hoc analysis.
Robot process automation (Robotic Process Automation, RPA) is a novel technical concept. The system allows a software robot to simulate and execute a given business process based on interaction action of a certain rule, and can simulate operation behaviors of a person between different systems on a computer to replace the person to automatically execute the business process based on the rule and with high repeatability before the computer. Rather, it is not a real, macroscopic robot, but rather a process automation service. Generating executable code is the last step of RPA, for which it is necessary to know the sequence of flow operations that the target business flow may occur to introduce flow mining techniques. Where predictive business process monitoring (Predictive Business Process Monitoring, PBPM) is an important area of process mining, emphasis is placed on estimating future features of an ongoing business process. Among them, predicting the next event (Predict Next Event) is an important issue in PBPM. By predicting the next event technique, a prediction of the future operation of the operational flow may be obtained, providing a valid reference for the RPA to generate code at runtime.
In the prior art, knowledge graphs lack participation on actual time sequence data, so that the knowledge graphs cannot be flexibly adapted to event prediction in an actual scene, the relation reflected by log data is often linear, but complex information such as branches and the like may exist in event development.
Disclosure of Invention
In order to solve the problems, the invention provides a flow operation sequence generation method combining an event log and a knowledge graph, which combines the event log as a time sequence data carrier and the knowledge graph verified by expert knowledge, so that the advantages of time sequence data prediction are reserved, and meanwhile, an event development rule and an evolution mode are fused to improve the accuracy of operation sequence prediction.
A flow operation sequence generation method combining event logs and knowledge maps comprises the following steps:
s1: acquiring a knowledge graph Q and an event log which are associated with a target flow P of which an operation sequence needs to be generated;
s2: selecting an event from the event log according to the running state of the target process P to obtain a predicted event sequence eQueue;
s3: according to the predicted event sequence eQueue, vector representation of each event and vector representation of a candidate event set of the next event are obtained, and all event representations are obtained;
s4: converting the predicted event sequence obtained in the step S3 into event representation fused with context time sequence information through a standard LSTM;
s5: converting all event representations and knowledge patterns Q obtained in the step S3 into event representations fused with global evolution information through SGNN;
s6: obtaining the event with highest occurrence probability according to the event representation H obtained in the step S4 and the event representation E obtained in the step S5
S7: handle eventAdded to the predicted event sequence eQueue and +.>Make a judgment if the event->If the event is an event representing the end of the flow or a signal of artificial termination is received, outputting a predicted event sequence eQueue as a predicted flow operation sequence; otherwise, go back to step S3 to continue with the above steps.
Further, the step S1 specifically includes:
s1.1: sending a request for acquiring the related knowledge graph to a knowledge graph base according to the flow description of the target flow P, and obtaining a knowledge graph Q associated with the target flow P;
s1.2: selecting event log records log of m executed process instances from the historical event log according to the process description of the target process P 1 ,…,log m Obtaining log of event log 1 Event log record, log, representing flow 1 instance m An event log record representing an mth flow instance.
Further, the step S2 specifically includes:
s2.1: when the target process P is in an operation state, namely the target process P is an executing process, taking out an event log, taking out events in the event log according to time sequence and adding the events into a set predicted event sequence eQueue;
s2.2: when the target flow P is in an unoperated state, the event log is extracted according to the description of the target flow P, and part of the events in the event log are extracted and added into the predicted event sequence eQueue.
Further, the step S3 specifically includes:
s3.1: event content e of event e c Subscript e of co-event e in predicted event sequence eQueue t Learning the respective embeddings by using a transducer encoder to obtain event content embeddingsAnd subscript embedding->
S3.2: according to the formulaObtaining a vector representation of each event, obtaining a vector representation of the predicted event sequence as +.>N is a positive integer;
s3.3: obtaining event logs of m finished process instances from the event logs log, and obtaining vector representations of m candidate events of the predicted event sequence through steps S3.1 and S3.2:
where n represents the number of events,last 1 events representing the 1 st flow instance, +.>Last 1 events representing flow instance 2, +.>The last 1 event representing the mth flow instance;
and further obtaining vector representation of the candidate event set corresponding to the next eventWherein the subscript c represents the next event, i.e., c=n+1;
s3.4: concatenating the vector representation of the predicted event sequence and the vector representation of the candidate event set for the next event as all event representations E 0
Wherein,next event representing flow 1 instance,/->Next event representing flow instance 2,/->Representing the next event of the mth flow instance.
Further, step S4 specifically includes:
s4.1: the predicted event sequence obtained in the step S3 is processedIs input to an LSTM layer by the formula +.>Get the ith event->Hidden representation +.>,/>Hidden representation representing the last event, i=1, 2,..n, resulting in a set of hidden representations corresponding to the predicted event sequence +.>
S4.2: vector representation of candidate event sets based on next eventInputting each candidate event into LSTM layer to obtain hidden representation of current candidate event +.>,/>Next event representing jth flow instance,/->A hidden representation representing an nth event;
s4.3: handleHidden representation and +.>The hidden representations of the candidate events are calculated as event representations H fused with context timing information by the following formula:
where n represents the number of events,hidden representation representing event 1, +.>Hidden representation representing event 2, +.>Hidden representation representing nth event, +.>Next event representing flow 1 instance,/->Next event representing flow instance 2,/->Representing the next event of the mth flow instance.
Further, step S5 specifically includes:
s5.1: all event representations E obtained by step S3 0 Extracting an adjacency matrix A of the subgraph from the knowledge graph Q obtained in the step S1:
wherein,representing the ith event, +.>Represents the j-th event,/->Representation->After occurrence->Probability of occurrence;
s5.2: representing all events E 0 And inputting the adjacency matrix A into the SGNN, and obtaining the event representation E fused with the global evolution information through the development rule and the evolution mode among the SGNN learning events.
Further, the step S6 specifically includes:
s6.1: dividing the event representation H fused with the context timing information into first context events H 1 And a first candidate event H 2 Calculating the weight of each context event for each candidate event to obtain a first weight matrix alpha:
wherein,representing a first candidate event H 2 Is a transpose of (2);
s6.2: obtaining a first representative vector of a sequence of predicted events
Wherein,representing a transpose of the first weight matrix;
S6.3: taking the cosine similarity of the first representative vector of the predicted event sequence and the first candidate event as the first score of the first candidate event
Wherein,representing cosine similarity;
s6.4: dividing the event representation E fused with the global evolution information into second context events E 1 And a second candidate event E 2 Calculating the weight of each context event for each candidate event to obtain a second weight matrix beta:
wherein,representing a second candidate event E 2 Is a transpose of (2);
s6.5: obtaining a second representative vector of the predicted event sequence
Wherein,representing a transpose of the second weight matrix β;
s6.6: taking the cosine similarity of the cosine of the candidate event and the second representative vector of the predicted event sequence as the second score of the candidate event
Wherein,representing cosine similarity;
s6.7: for candidate eventsCandidate event->Is>And a second score->Weighted summation as occurrence probability of the candidate event +.>
Wherein,representing candidate event->Is>Coefficients of (2);
s6.8: calculating the final occurrence probability of each candidate event, and selecting the event e with the highest occurrence probability p
Where softmax () represents the activation function, relu () represents the linear rectification function, T represents the transpose,represents the j-th candidate event->Probability of final occurrence, argmax j p () represents a function for finding the maximum probability, +.>Represents the j-th candidate event->Probability of occurrence of->Represents the kth candidate event->Is a probability of occurrence of (a).
The storage device stores instructions and data for implementing a flow operation sequence generation method combining an event log and a knowledge graph.
A system for generating a sequence of flow operations in combination with an event log and a knowledge graph, comprising: a processor and the storage device; the processor loads and executes the instructions and the data in the storage device to realize a flow operation sequence generation method combining the event log and the knowledge graph.
The technical scheme provided by the invention has the beneficial effects that: the invention uses the historical event log information and the related process knowledge graph related to the target generation sequence process, combines the Transformer technology to flexibly and comprehensively represent the event, uses the context information existing in the LSTM learning event log, learns the evolution mode information existing in the knowledge graph through the SGNN technology, provides the next event prediction by combining the context information and the event evolution mode, and generates a reliable process operation sequence on the basis of the next event prediction. By combining time sequence data and a knowledge graph method, the next event prediction is realized more flexibly and accurately, and the method has better universality and effectiveness.
Drawings
FIG. 1 is a flowchart of a method for generating a sequence of flow operations in combination with an event log and a knowledge graph, in an embodiment of the invention;
FIG. 2 is a schematic diagram of the operation of a hardware device in an embodiment of the invention.
Detailed Description
For a clearer understanding of technical features, objects and effects of the present invention, a detailed description of embodiments of the present invention will be made with reference to the accompanying drawings.
The invention provides a method and a system for generating a flow operation sequence by combining an event log and a knowledge graph.
Referring to fig. 1, a method for generating a flow operation sequence by combining an event log and a knowledge graph specifically includes:
s1: acquiring a knowledge graph Q and an event log which are associated with a target flow P of which an operation sequence needs to be generated; the step S1 specifically comprises the following steps:
s1.1: sending a request for acquiring the related knowledge graph to a knowledge graph base according to the flow description of the target flow P, and obtaining a knowledge graph Q associated with the target flow P;
s1.2: selecting event log records log of m executed process instances from the historical event log according to the process description of the target process P 1 ,…,log m Obtaining log and log of event log 1 Event log record, log, representing flow 1 instance m An event log record representing an mth flow instance;
s2: initializing a set predicted event sequence eQueue: selecting certain event-in predicted event sequences eQueue from event logs according to the running state of the target process P; when the events are selected, the events need to be satisfied, namely the events in the actually completed flow track event logs of the flow described by the target flow P, for example, the target flow is 'payment', and the events need to be selected from the track event logs log1, log2 and the like of which the 'payment' is completed in the past; the step S2 specifically comprises the following steps:
s2.1: if the target flow P is the executing flow, taking out an event log of the current executing flow track, taking out the events in the event log according to time sequence and adding the events into a predicted event sequence eQueue, otherwise, performing step S2.2;
s2.2: the target flow P is not in an operation state, an event log of a history flow track is taken out according to the description of the target flow P, the first few events in the event log are taken out and added into a predicted event sequence eQueue; the first few events can be flexibly selected according to the scene and the actual use condition, and one event can be taken, and multiple events can be taken. Generally, the first event is "beginning", the first events of some processes can be selected directly to avoid multiple calculations if the first events are relatively fixed, but if the event change in the beginning stage is relatively large, the subsequent events of the first event are selected directly and obtained through prediction.
S3: setting an event presentation layer: obtaining a vector representation of each event e in the predicted event sequence eQueue and a vector representation of a candidate event set of a next event; the step S3 specifically comprises the following steps:
s3.1: event content e of event e c Subscript e of co-event e in predicted event sequence eQueue t Learning the respective embeddings by using a transducer encoder to obtain event content embeddingsAnd subscript embedding->
S3.2: according to the formulaObtaining a vector representation of each event e, obtaining a vector representation of the predicted event sequence eQueue +.>,/>Vector representation representing event 1, +.>Vector representation representing event 2, < +.>A vector representation representing an nth event, i.e., a vector representation of the last 1 event, n representing the number of events, being a positive integer;
s3.3: obtaining event logs of m finished process instances from the event logs log, and obtaining vector representations of m candidate event sets of the predicted event sequence through steps S3.1 and S3.2:
wherein,nrepresenting the number of events, m representing the number of flow instances; because the event length of completion of different processes is uncertain, usingnThe representation is made in such a way that,last 1 events representing the 1 st flow instance, +.>Last 1 events representing flow instance 2, +.>The last 1 event representing the mth flow instance; for example, the 1 st flow instance has 3 events, then the last event is e 31 The 4 th flow instance has 5 events, then the last event is e 54 . It can be seen that the number of candidate events is equal to the number of flow instances, and is denoted by m.
And then get the next eventVector representation of corresponding candidate event setsWherein c=n+1, +.>Representing the next event after the 1 st flow instance,representing the next event after the 2 nd flow instance,/->Representing a next event after the mth flow instance; n is the number of the last event in the currently existing predicted event sequence eQueue, i.e. the first n events to occur are already known, and to know what is n+1 events, n+1 events of the event logs of m flow instances need to be selected as candidate event sets to be referred to.
S3.4: concatenating the vector representation of the predicted event sequence and the vector representation of the candidate event set for the next event as all event representations E 0
Wherein the subscript c represents the sequence number of the next event, e cm Is the reference to the completed mth flow instance (a flow instance is a sequence of historically occurring events) at event number c (i.e. n + 1),next event representing flow 1 instance,/->Next event representing flow instance 2,/->The next event (one event after the nth event) representing the mth flow instance.
And step S3, word embedding is carried out on various elements representing the event through a transducer, and then vector representations obtained by word embedding are spliced to obtain the representation of the whole event, so that flexible and comprehensive application of various information in log data is realized, meanwhile, candidate events of the next event are extracted from a historical event log to carry out word embedding, the historical data is fully utilized, and the combination of prediction of the next event and an actual scene is ensured.
S4: a set time sequence representation module: the output of the input event presentation layer (i.e. the predicted event sequence output in step S3) is used to obtain an event presentation H fused with the context timing information using a standard LSTM; the LSTM is used for learning time sequence information in the context event, so that the real continuous relation between the events influences the event prediction result, and the prediction accuracy is improved by combining the use of the candidate event set. The specific steps of step S4 are as follows:
s4.1: inputting the predicted event sequence obtained in the step S3
S4.2: handle sequenceInputting to an LSTM layer to obtain each event +.>Hidden representation +.>I.e. +.>,/>Hidden representation representing the last event, i=1, 2,..n, resulting in a set of hidden representations corresponding to the predicted event sequence +.>
S4.3: vector representation of candidate event sets based on next eventEach candidate event is +.>Inputting into LSTM layer to obtain hidden representation of current candidate event>
S4.4: handleHidden representation of individual context events and +.>Hidden representations of the candidate events the event representation H fused with the context timing information is obtained by the following calculation formula:
wherein n represents the number of times,hidden representation representing event 1, +.>Hidden representation representing event 2, +.>Hidden representation representing nth event, +.>Next event representing flow 1 instance,/->Next event representing flow instance 2,/->Representing the next event of the mth flow instance.
S5: setting an event evolution module: inputting the output of the event representation layer (namely all event representations output in the step S3) and the knowledge graph Q, and obtaining an event representation E fused with global evolution information by using SGNN; firstly, event representation and a knowledge graph are subjected to sub-graph matching to obtain a knowledge graph mode sub-graph related to a current event sequence, and then SGNN is used for event prediction, so that global evolution information contained in the knowledge graph is well fused, and prediction accuracy is further improved. The specific steps of step S5 are as follows:
s5.1: all event representations E obtained by step S3 0 Extracting an adjacency matrix A of the subgraph from the knowledge graph Q obtained in the step S1, wherein the extraction formula is as follows:
wherein the formalization of the map Q is expressed asRepresenting node set, ++>Representing an edge set comprising all events occurring in the event chain of the training set, the edge set comprising all edges connecting between nodes,/->I.e. nodes in the node set, representing an event, event->And event->Weights of edges ∈>Representation->After occurrence->Probability of occurrence, calculation formula: />,/>Representing event->And->And the number of occurrences in the flow instance of the training set.
S5.2: representing all events E 0 The adjacency matrix A is input into the SGNN, and event representation E fused with global evolution information is obtained through the development rule and the evolution mode among the SGNN learning events;
s5.2.1: and carrying out message transmission on the child graph nodes through an adjacency matrix A, wherein the formula is as follows:
wherein A represents an adjacency matrix,node status (activation) representing the current time step t, obtained by passing information from the adjacency matrix a,/->An event representation representing a last time step t-1, b being a bias term (bias);
s5.2.2: updating the state of each node by the information of the last time step and other nodes: the specific formula is as follows:
wherein,is an update gate that controls how much of the current time step information should be retained, a value between 0 and 1, for weighting the event representation of the last time step; />Reset the gate, determine how much the information of the last time step affects the node state of the current time step, add by bit, +.>Representing a sigmoid activation function for mapping the input value between 0 and 1; />、/>、/>Are weight matrixes for learning how to integrate the information of the current time step; />、/>Are weight matrices for learning how to integrate the information of the last time step; />The event representation representing the current time step t integrates the information of the last time step by updating the gate and resetting the gate;
s5.2.3: and iterating the steps for k times to obtain the event representation E which is output by the event evolution module and is fused with global evolution information.
S6: and (3) setting a prediction module: inputting H and E, and outputting event E with highest probability of occurrence of next event in target flow p
S6.1: dividing the event representation H fused with the context time sequence information into context events H 1 And candidate event H 2 And calculating the weight of each event for each candidate event to obtain a weight matrix alpha of the context event and the candidate event, wherein the calculation formula is as follows:
wherein,representing candidate event H 2 Is a transpose of (2);
s6.2: obtaining a first representative vector of a sequence of predicted events
Wherein,representing a transpose of the weight matrix α;
s6.3: taking the cosine similarity of the cosine of the candidate event and the first representative vector of the predicted event sequence as the first score of the candidate event
Wherein,representing cosine similarity;
s6.4: dividing the event representation E fusing global evolution information into context events E 1 And candidate event E 2 And calculating the weight of each event for each candidate event to obtain a context event and a weight matrix beta of the candidate event, wherein the calculation formula is as follows:
s6.5: obtaining a second representative vector of the predicted event sequence
S6.6: taking the cosine similarity of the cosine of the candidate event and the second representative vector of the predicted event sequence as the second score of the candidate event
S6.7: for candidate eventsCandidate event->Is>And a second score->Weighted summation as occurrence probability of the candidate event +.>
Wherein,representing event->Is>Coefficients of (2);
s6.8: calculating the final occurrence probability of each candidate event, and selecting the event e with the highest occurrence probability p
Wherein softmax () represents the activation function, relu () represents the linear rectification function, T represents the transpose, argmax j p () represents an event that maximizes pFunction of->Representing candidate event->Probability of occurrence of->Representing candidate event->K represents the k candidate event;
s7: event e p Adding into predicted event sequence eQueue, judging event e p Whether it is a flow end event or if the predicted process is interrupted by an artificial (i.e., terminated by an artificial) if e p If the event represents the end of the flow or a signal of artificial termination is received, the eQueue is output as a predicted flow operation sequence; otherwise, the current eQueue is input to the step S3 to continue the steps.
Referring to fig. 2, fig. 2 is a schematic working diagram of a hardware device according to an embodiment of the present invention, where the hardware device specifically includes: a sequence of flow operations combining event logs and knowledge maps is generated 201, processor 202, and storage 203.
A sequence of flow operations generation system 201 that combines event logs and knowledge maps: the process operation sequence generating system 201 combining the event log and the knowledge graph realizes the process operation sequence generating method combining the event log and the knowledge graph.
Processor 202: the processor 202 loads and executes the instructions and data in the storage device 203 to implement the method for generating a sequence of flow operations combining event logs and knowledge maps.
Storage device 203: the storage device 203 stores instructions and data; the storage device 203 is configured to implement the method for generating a flow operation sequence by combining an event log and a knowledge graph.
The beneficial effects of the invention are as follows: the method comprises the steps of utilizing historical event log information and related process knowledge patterns related to a target generation sequence process, carrying out flexible and comprehensive event representation on an event by combining a Transformer technology, utilizing context information existing in an LSTM learning event log, learning evolution mode information existing in the knowledge patterns by an SGNN technology, providing next event prediction by combining the context information and an event evolution mode, and carrying out iterative operation on the basis, so as to generate a reliable process operation sequence. According to the invention, by combining the time sequence data and the knowledge graph method, on one hand, the event time sequence information in the actual operation scene is utilized, and on the other hand, the rule and mode of the aggregated event evolution in the knowledge graph are utilized, so that more flexible and accurate prediction of the next event is realized, and the method has better universality and effectiveness.
The foregoing description of the preferred embodiments of the invention is not intended to limit the invention to the precise form disclosed, and any such modifications, equivalents, and alternatives falling within the spirit and scope of the invention are intended to be included within the scope of the invention.

Claims (9)

1. A process operation sequence generation method combining event logs and knowledge maps is characterized in that: comprising the following steps:
s1: acquiring a knowledge graph Q and an event log which are associated with a target flow P of which an operation sequence needs to be generated;
s2: selecting an event from the event log according to the running state of the target process P to obtain a predicted event sequence eQueue;
s3: according to the predicted event sequence eQueue, vector representation of each event and vector representation of a candidate event set of the next event are obtained, and all event representations are obtained;
s4: converting the predicted event sequence obtained in the step S3 into event representation H fused with context time sequence information through a standard LSTM;
s5: converting all event representations and knowledge patterns Q obtained in the step S3 into event representations E fused with global evolution information through SGNN;
s6: obtaining the event with highest occurrence probability according to the event representation H obtained in the step S4 and the event representation E obtained in the step S5
S7: handle eventAdded to the predicted event sequence eQueue and +.>Make a judgment if the event->If the event is an event representing the end of the flow or a signal of artificial termination is received, outputting a predicted event sequence eQueue as a predicted flow operation sequence; otherwise, go back to step S3 to continue with the above steps.
2. The method for generating a sequence of flow operations by combining an event log and a knowledge graph as claimed in claim 1, wherein: the step S1 specifically comprises the following steps:
s1.1: sending a request for acquiring the related knowledge graph to a knowledge graph base according to the flow description of the target flow P, and obtaining a knowledge graph Q associated with the target flow P;
s1.2: selecting event log records log of m executed process instances from the historical event log according to the process description of the target process P 1 ,…,log m Obtaining log of event log 1 Event log record, log, representing flow 1 instance m An event log record representing an mth flow instance.
3. The method for generating a sequence of flow operations combining an event log and a knowledge graph as claimed in claim 2, wherein: the step S2 specifically comprises the following steps:
s2.1: when the target flow P is in an operation state, namely the target flow P is an executing flow, taking out an event log, taking out events in the event log according to time sequence and adding the events into a predicted event sequence eQueue;
s2.2: when the target flow P is in an unoperated state, the event log is extracted according to the description of the target flow P, and part of the events in the event log are extracted and added into the predicted event sequence eQueue.
4. The method for generating a sequence of flow operations combining an event log and a knowledge graph as claimed in claim 3, wherein: the step S3 specifically comprises the following steps:
s3.1: event content e of event e c Subscript e of co-event e in predicted event sequence eQueue t Learning the respective embeddings by using a transducer encoder to obtain event content embeddingsAnd subscript embedding->
S3.2: according to the formulaObtaining a vector representation of each event, obtaining a vector representation of the predicted event sequence as +.>,/>Vector representation representing event 1, +.>Vector representation representing event 2, < +.>A vector representation representing an nth event, n representing the number of events;
s3.3: obtaining event logs of m finished process instances from the event logs log, and obtaining vector representations of m candidate events of the predicted event sequence through steps S3.1 and S3.2:
where n represents the number of events,last 1 events representing the 1 st flow instance, +.>Last 1 events representing flow instance 2, +.>The last 1 event representing the mth flow instance;
and further obtains a vector representation of the candidate event set corresponding to the next eventWherein the subscript c represents the next event, i.e., c=n+1;
s3.4: concatenating the vector representation of the predicted event sequence and the vector representation of the candidate event set for the next event as all event representations E 0
Wherein,representing the next event after the 1 st flow instance,/->Representing the next event after the 2 nd flow instance,/->Representing the next event after the mth flow instance.
5. The method for generating a sequence of flow operations combining an event log and a knowledge graph as claimed in claim 4, wherein: the step S4 specifically comprises the following steps:
s4.1: the predicted event sequence obtained in the step S3 is processedIs input to an LSTM layer by the formula +.>Get the ith event->Hidden representation +.>,/>Hidden representation representing the last event, i=1, 2,..n, resulting in a set of hidden representations corresponding to the predicted event sequence +.>
S4.2: vector representation of candidate event sets based on next eventInputting each candidate event into LSTM layer to obtain hidden representation of current candidate event +.>,/>Next event representing jth flow instance,/->A hidden representation representing an nth event;
s4.3: handleHidden representation and +.>The hidden representations of the candidate events are calculated as event representations H fused with context timing information by the following formula:
where n represents the number of events,hidden representation representing event 1, +.>A hidden representation of the 2 nd event is represented,hidden representation representing nth event, +.>Next event representing flow 1 instance,/->Next event representing flow instance 2,/->Representing the next event of the mth flow instance.
6. The method for generating a sequence of flow operations in combination with an event log and a knowledge graph as claimed in claim 5, wherein: the step S5 specifically comprises the following steps:
s5.1: all event representations E obtained by step S3 0 Extracting an adjacency matrix A of the subgraph from the knowledge graph Q obtained in the step S1:
wherein,representing the ith event, +.>Represents the j-th event,/->Representation->After occurrence->Probability of occurrence;
s5.2: representing all events E 0 And inputting the adjacency matrix A into the SGNN, and obtaining the event representation E fused with the global evolution information through the development rule and the evolution mode among the SGNN learning events.
7. The method for generating a sequence of flow operations in combination with an event log and a knowledge graph as claimed in claim 6, wherein: the step S6 specifically comprises the following steps:
s6.1: dividing the event representation H fused with the context timing information into first context events H 1 And a first candidateEvent H 2 Calculating the weight of each context event for each candidate event to obtain a first weight matrix alpha:
wherein,representing candidate event H 2 Is a transpose of (2);
s6.2: obtaining a first representative vector of a sequence of predicted events
Wherein,representing a transpose of the weight matrix;
s6.3: taking the cosine similarity of the first representative vector of the predicted event sequence and the first candidate event as the first score of the first candidate event
Wherein,representing cosine similarity;
s6.4: dividing the event representation E fused with the global evolution information into second context events E 1 And a second candidate event E 2 Calculating the weight of each event to each candidate event to obtain the firstTwo weight matrix β:
wherein,representing a second candidate event E 2 Is a transpose of (2);
s6.5: obtaining a second representative vector of the predicted event sequence
Wherein,representing a transpose of the second weight matrix β;
s6.6: taking the cosine similarity of the second representative vector of the predicted event sequence and the second candidate event as a second score of the second candidate event
Wherein,representing cosine similarity;
s6.7: for candidate eventsCandidate event->Is>And a second score->Weighted summation as occurrence probability of the candidate event +.>
Wherein,representing candidate event->Is>Coefficients of (2);
s6.8: calculating the final occurrence probability of each candidate event, and selecting the event e with the highest occurrence probability p
Where softmax () represents the activation function, relu () represents the linear rectification function, T represents the transpose,represents the j-th candidate event->Probability of final occurrence, argmax j p () represents a function for finding the maximum probability, +.>Represents the j-th candidate event->Probability of occurrence of->Represents the kth candidate event->Is a probability of occurrence of (a).
8. A memory device, characterized by: the storage device stores instructions and data for implementing the method for generating a flow operation sequence by combining an event log and a knowledge graph according to any one of claims 1 to 7.
9. A flow operation sequence generation system combining event logs and knowledge maps is characterized in that: the system comprises: a processor and a storage device; the processor loads and executes the instructions and data in the storage device to implement the method for generating a flow operation sequence by combining the event log and the knowledge graph according to any one of claims 1 to 7.
CN202410006733.XA 2024-01-03 Method and system for generating flow operation sequence by combining event log and knowledge graph Active CN117493583B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410006733.XA CN117493583B (en) 2024-01-03 Method and system for generating flow operation sequence by combining event log and knowledge graph

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410006733.XA CN117493583B (en) 2024-01-03 Method and system for generating flow operation sequence by combining event log and knowledge graph

Publications (2)

Publication Number Publication Date
CN117493583A true CN117493583A (en) 2024-02-02
CN117493583B CN117493583B (en) 2024-05-14

Family

ID=

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190043500A1 (en) * 2017-08-03 2019-02-07 Nowsportz Llc Voice based realtime event logging
US20190163552A1 (en) * 2017-11-28 2019-05-30 Nec Laboratories America, Inc. System and method for contextual event sequence analysis
US20200372373A1 (en) * 2019-05-21 2020-11-26 Sisense Ltd. System and method for generating organizational memory using semantic knowledge graphs
CN112632223A (en) * 2020-12-29 2021-04-09 天津汇智星源信息技术有限公司 Case and event knowledge graph construction method and related equipment
CN114357197A (en) * 2022-03-08 2022-04-15 支付宝(杭州)信息技术有限公司 Event reasoning method and device
CN116069947A (en) * 2023-01-30 2023-05-05 思必驰科技股份有限公司 Log data event map construction method, device, equipment and storage medium
WO2023115761A1 (en) * 2021-12-20 2023-06-29 北京邮电大学 Event detection method and apparatus based on temporal knowledge graph
CN117236677A (en) * 2023-08-07 2023-12-15 安徽思高智能科技有限公司 RPA process mining method and device based on event extraction

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190043500A1 (en) * 2017-08-03 2019-02-07 Nowsportz Llc Voice based realtime event logging
US20190163552A1 (en) * 2017-11-28 2019-05-30 Nec Laboratories America, Inc. System and method for contextual event sequence analysis
US20200372373A1 (en) * 2019-05-21 2020-11-26 Sisense Ltd. System and method for generating organizational memory using semantic knowledge graphs
CN112632223A (en) * 2020-12-29 2021-04-09 天津汇智星源信息技术有限公司 Case and event knowledge graph construction method and related equipment
WO2023115761A1 (en) * 2021-12-20 2023-06-29 北京邮电大学 Event detection method and apparatus based on temporal knowledge graph
CN114357197A (en) * 2022-03-08 2022-04-15 支付宝(杭州)信息技术有限公司 Event reasoning method and device
CN116069947A (en) * 2023-01-30 2023-05-05 思必驰科技股份有限公司 Log data event map construction method, device, equipment and storage medium
CN117236677A (en) * 2023-08-07 2023-12-15 安徽思高智能科技有限公司 RPA process mining method and device based on event extraction

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
李家炎: "知识图谱技术在金融网络日志告警压降溯源领域的应用研究", 《中国金融电脑》, 7 June 2023 (2023-06-07), pages 73 - 76 *

Similar Documents

Publication Publication Date Title
US11423325B2 (en) Regression for metric dataset
CN115185736B (en) Micro-service call chain abnormity detection method and device based on graph convolution neural network
CN113239702A (en) Intention recognition method and device and electronic equipment
CN115618269A (en) Big data analysis method and system based on industrial sensor production
CN111915086A (en) Abnormal user prediction method and equipment
CN117236677A (en) RPA process mining method and device based on event extraction
WO2021147405A1 (en) Customer-service statement quality detection method and related device
CN117493583B (en) Method and system for generating flow operation sequence by combining event log and knowledge graph
JP2019095599A (en) Acoustic model learning device, speech recognition device, and method and program for them
CN116703466A (en) System access quantity prediction method based on improved wolf algorithm and related equipment thereof
US20200380446A1 (en) Artificial Intelligence Based Job Wages Benchmarks
CN116341720A (en) Multi-fan wind speed and direction prediction method based on dynamic graph convolution and transformation
CN117493583A (en) Method and system for generating flow operation sequence by combining event log and knowledge graph
CN111161238A (en) Image quality evaluation method and device, electronic device, and storage medium
CN115660795A (en) Data processing method, device, equipment, storage medium and program product
CN115062769A (en) Knowledge distillation-based model training method, device, equipment and storage medium
CN113779116A (en) Object sorting method, related equipment and medium
US20230146635A1 (en) Method and Systems for Conditioning Data Sets for Efficient Computational Processing
CN115640336B (en) Business big data mining method, system and cloud platform
CN112307227B (en) Data classification method
CN113469244B (en) Volkswagen app classification system
CN117056452B (en) Knowledge point learning path construction method, device, equipment and storage medium
CN117391713A (en) Information pushing method and device, electronic equipment and storage medium
CN116562904A (en) Target object determining method based on federal learning and electronic equipment
CN114154687A (en) Trajectory prediction method and apparatus, electronic device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant