CN115017368A - Dynamic graph representation learning method based on self-supervision learning - Google Patents

Dynamic graph representation learning method based on self-supervision learning Download PDF

Info

Publication number
CN115017368A
CN115017368A CN202210455958.4A CN202210455958A CN115017368A CN 115017368 A CN115017368 A CN 115017368A CN 202210455958 A CN202210455958 A CN 202210455958A CN 115017368 A CN115017368 A CN 115017368A
Authority
CN
China
Prior art keywords
time
node
space
graph
subgraph
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210455958.4A
Other languages
Chinese (zh)
Inventor
鲍鹏
李家年
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Jiaotong University
Original Assignee
Beijing Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Jiaotong University filed Critical Beijing Jiaotong University
Priority to CN202210455958.4A priority Critical patent/CN115017368A/en
Publication of CN115017368A publication Critical patent/CN115017368A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/901Indexing; Data structures therefor; Storage structures
    • G06F16/9024Graphs; Linked lists
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9537Spatial or temporal dependent retrieval, e.g. spatiotemporal queries

Abstract

The invention relates to a dynamic graph representation learning method based on self-supervision learning, which comprises the following steps: sampling the dynamic graph to respectively obtain a space-time subgraph, a non-time sequence subgraph and a non-space subgraph; respectively obtaining node levels of the three subgraphs and representation of the graph level by using a space-time weight encoder; designing time-space comparison learning, and defining a total loss function from the angles of time and space respectively; and (3) completing the training of a space-time weight encoder through a total loss function, finally sending all nodes in the dynamic graph into the trained space-time weight encoder to obtain the node representation of the dynamic graph, and applying the node representation to tasks such as classification, recommendation, link prediction and the like in the application scene of the dynamic graph model. The invention integrates the time and space information of the dynamic graph and enhances the performance of the dynamic graph representation method; has good interpretability, verifiability, characterization performance and mobility. And the invention can be applied to the e-commerce network in reality to predict the goods which the user may purchase.

Description

Dynamic graph representation learning method based on self-supervision learning
Technical Field
The invention belongs to the field of graph representation learning, and particularly relates to a dynamic graph representation learning method based on self-supervision learning.
Background
With the explosive growth of graph structure-related data, more and more scholars are engaged in research related to the field of graph representation learning. However, most existing supervised or semi-supervised graph representations rely heavily on data labels for learning methods. On the one hand, collecting data tags would cost a lot of manpower and material resources. On the other hand, models trained by using data labels are poorly generalizable and less robust. This has caused a great deal of trouble in current research, and in recent years, there have been many studies on learning of a self-supervision map representation to solve the above-mentioned problems.
The idea of learning represented by the self-supervision diagram is that a specific task is designed, dependency of a model on a data label is relieved, and generalization capability and robustness of the model are further improved. For example: the DGI model, a first self-supervised graph representation learning approach, generates a representation of a node by maximizing the mutual information between the node representation and the global graph representation. MVGRL is a multi-view-angle, self-supervised graph representation model whose goal is to maximize node characterization in the original graph and node characterization in the diffusion graph. The Sub-Con model maximizes mutual information between the node representation and the Sub-graph representation to learn the node representation. Unlike previous work, is the Grace model, which focuses primarily on node-level consistency. The method generates a new visual angle through a disturbance function, and learns node representation by maximizing the representation of the same node under different views.
However, existing self-supervision graphs represent that the learning model ignores the evolution process on the graph. Specifically, the existing self-supervision graph shows that the learning model discards time information on the graph for simplifying the operation, which leads to the failure of the model to capture the evolution process on the graph, and further leads to calculation errors. Through research, we find that the relevant research of learning dynamic graph representation based on an auto-supervision mode is very little.
The method disclosed by the invention integrates the thought of the self-supervision learning, namely, a dynamic graph representation learning method based on the self-supervision learning is researched. The overall idea is to define a comparative learning task which takes time and space into consideration, and the model completes the task to capture the time and space characteristics of the dynamic graph. Specifically, the method firstly defines three subgraphs, namely a spatio-temporal subgraph, a non-time sequence subgraph and a non-space subgraph. Furthermore, the invention comprises a spatio-temporal weight encoder that aggregates the temporal and spatial information of the sub-pictures. Finally, the method is applied to an e-commerce scene to predict the commodities which are possibly purchased by the user.
Disclosure of Invention
The invention provides a dynamic graph representation learning method based on self-supervision learning. Firstly, the method takes any node as a central node in a graph, and respectively carries out space-time random walk and random walk, and the sampled subgraphs are respectively defined as a space-time subgraph and a non-time sequence subgraph. In addition, random space-time walking is carried out by taking any other node as a central node, and the sampled subgraph is called a non-space subgraph; secondly, capturing time and space information in the graph by using a space-time weight encoder to obtain node characteristics and sub-graph characteristics of corresponding nodes; and finally, the characterization of the node is learned by maximizing mutual information between the node characterization and the subgraph characterization through a contrast learning method. The specific technical scheme is as follows:
a dynamic graph representation learning method based on self-supervision learning comprises the following steps:
s1, sampling any node in the graph as a central node to obtain a space-time subgraph, a non-time sequence subgraph and a non-space subgraph respectively;
s2, capturing time and space information in the graph by using a space-time weight encoder to obtain node characteristics of corresponding nodes and graph level characteristics of the subgraph;
s3, learning the node representation by comparing the mutual information between the maximum node representation and the graph level representation of the subgraph in a learning method, and respectively defining a total loss function from the aspects of time and space;
and S4, completing the training of the space-time weight encoder through a total loss function, and finally sending all the nodes in the dynamic graph into the trained space-time weight encoder to obtain the node representation of the dynamic graph.
On the basis of the above technical solution, the specific steps of step S1 are: given a dynamic graph
Figure BDA0003620467090000031
Where V represents the set of edges and,
Figure BDA0003620467090000032
is a set of a series of edges, where there is an edge E ═ (v, w, t) ∈ E t Representing the edges generated by node v and node w at time t,
Figure BDA0003620467090000033
representing a function that can map each edge to a timestamp. The invention hopes to acquire a space-time subgraph G taking a node v as a central node in a dynamic graph G v
For spatio-temporal subgraph G v The sampling is mainly divided into the following two steps: initial edge selection and spatio-temporal random walk.
In the initial edge selection step, sampling is performed with node v as the center node. The samples include unbiased samples and biased samples.
For unbiased sampling, define all edge sets around node v as Ψ v The initial edge is sampled using a uniformly distributed sampling approach. The edge e ∈ Ψ (v, w, t) ∈ Ψ v The probability of being selected is defined as:
Figure BDA0003620467090000034
for biased sampling, it is assumed that all edges around node v are selected with different probabilities. Biased sampling will assign different sampling probabilities to edges depending on when they occur. Specifically, the edge e ═ v, w, t ∈ Ψ v The probability of being selected is defined as:
Figure BDA0003620467090000041
wherein e' is as much as Ψ v Representing any edge connected to the central node v,
Figure BDA0003620467090000042
the time corresponding to the edge e' is,
Figure BDA0003620467090000043
represents the time corresponding to the edge e, and t min Is the minimum time associated with an edge in the dynamic graph. Biased sampling is to encourage sampling of later occurring edges.
The task mainly performed by the step of spatio-temporal random walk is to select the next node. The selection of the next node may be selected from the set of spatio-temporal neighbors Γ. The set of spatiotemporal neighbors may be defined as:
Figure BDA0003620467090000044
wherein, gamma is t (v) Represents the set of spatio-temporal neighbors of node v at time t, w represents the nodes adjacent to node v, and t' represents the time on the edge between node w and node v.
Similarly, the invention has two space-time random walk strategies, which can be defined as biased space-time random walk and unbiased space-time random walk.
For unbiased space-time random walk, given an arbitrary edge e ═ (u, v, t), the space-time neighbor w ∈ Γ of node v at time t t (v) The probability of being selected is as follows:
Figure BDA0003620467090000045
for biased space-time random walk, the space-time neighbor w epsilon Γ of the node v at the time t t (v) The probability of being selected is:
Figure BDA0003620467090000051
wherein w ' represents any space-time neighbor of the node v at the time t, δ (w) is the time corresponding to the node w, and δ (w ') is the time corresponding to the node w '.
Furthermore, we refer to the idea of random walk, i.e. sampling without taking into account the time factor when sampling. Random walk and sampling non-time sequence subgraph with node v as center node
Figure BDA0003620467090000054
And performing space-time random walk by taking the node o (wherein the node o and the node v are different nodes) as a central node to obtain a non-space subgraph G o
On the basis of the above technical solution, the specific steps of step S2 are: by using a space-time weight encoder, a characterization of all nodes is obtained.
Specifically, a spatio-temporal subgraph G is obtained v The spatio-temporal weight encoder is then used to output a representation of each node in the spatio-temporal subgraph. While the main goal of the spatio-temporal weight encoder is to avoid the so-called characterization obsolescence problem. The spatio-temporal weight encoder includes a temporal aggregator and a spatial aggregator.
For the time aggregator, when an event occurs involving the node itself, the timing of node v at time t characterizes z v (t) is expressed as:
z v (t)=LSTM(f (v,w) (t),h v (t - ))
where LSTM is represented as the LSTM model, f (v,w) (t) represents the interaction feature between nodes v and w at time t, h v (t - ) Is the time that node v last updated.
For the spatial aggregator, this component is used to incorporate characterization information of spatio-temporal neighbors, which can be assigned different weights according to different times of occurrence of neighboring nodes. In particular, the spatial characterization of node v at time t
Figure BDA0003620467090000052
The definition is as follows:
Figure BDA0003620467090000053
wherein w ∈ Neighbor representsA node w connected to the node v,
Figure BDA0003620467090000061
representing the calculation of the current time t and the time t w Time coding of the spacing, z w (t) is a time-series representation of node W at time t, W 1 Are learnable parameters.
Finally, we characterize z by aggregation timing v (t) and spatial characterization
Figure BDA0003620467090000062
Obtaining the node characterization h of the node v at the time t v (t):
Figure BDA0003620467090000063
Wherein, W 2 Are learnable parameters.
Notably, the graph-level characterization s of the spatio-temporal subgraph v Is obtained by a readout function.
Similarly, non-time-series subgraphs can be obtained by a space-time weight coder
Figure BDA0003620467090000064
Graph level characterization of
Figure BDA0003620467090000065
And a non-spatial subgraph G o Graph-level characterization s o
The specific steps of step S3 are: and training the space-time weight encoder by comparing positive and negative samples. The specific comparison learning target is to distinguish a spatio-temporal subgraph, a non-time sequence subgraph and a non-spatial subgraph. For the selection of the positive sample, the node characterization of the spatio-temporal subgraph and the graph-level characterization of the spatio-temporal subgraph are formed into the positive sample.
Secondly, the invention maps the non-time sequence
Figure BDA0003620467090000066
Graph level characterization of
Figure BDA0003620467090000067
And spatio-temporal subgraph G v Characterization of nodes of h v Time negative samples are formed, called time negative examples. The loss function for a negative sample of time is defined as:
Figure BDA0003620467090000068
wherein sigma represents a sigmoid function, phi is a marginal value,
Figure BDA0003620467090000069
representing an expectation function, s v For graph-level characterization of spatio-temporal subgraphs, h v For node characterization of node v in the spatio-temporal subgraph,
Figure BDA00036204670900000610
graph-level characterization for non-time-series subgraphs.
Furthermore, a non-spatial subgraph G o Graph-level characterization s o And spatio-temporal subgraph G v Characterization of nodes of h v Constituting a spatial negative example. The loss function for a spatial negative example is defined as:
Figure BDA0003620467090000071
where, σ is the sigmoid function,
Figure BDA0003620467090000072
is the value of the margin at which,
Figure BDA0003620467090000073
representing an expectation function, s v For graph-level characterization of spatio-temporal subgraphs, h v For node characterization of node v in a spatio-temporal subgraph, s o A graph-level characterization of a non-spatial subgraph.
The total loss function of the space-time weight encoder is therefore:
Figure BDA0003620467090000074
where λ is the balance parameter of the total loss function.
The invention has the beneficial effects that:
in order to more efficiently mine time and space information contained in a dynamic graph, the application intends to fuse the time and space information on the graph in the process of learning the dynamic graph representation, and provides a learning method for the dynamic graph representation based on self-supervision learning, wherein the method has good interpretability and excellent characterization performance and has the following advantages:
(1) the interpretability is high: the invention samples the dynamic graph to respectively obtain a space-time subgraph, a non-time sequence subgraph and a non-space subgraph, designs a comparison learning task based on the three subgraphs, and finishes capturing time and space information. The scheme has good interpretability.
(2) Strong mobility: the space-time sampling graph method is relatively independent, and the generated subgraph can be used as an auxiliary graph of other dynamic graph representation learning methods and has strong mobility.
(3) The characterization performance is excellent: the method is used for capturing time and space information on the dynamic graph, and experiments show that the method greatly improves the performance of tasks such as classification, recommendation, prediction and the like in the graph model application scene.
Drawings
The invention has the following drawings:
FIG. 1 is a schematic diagram of a learning framework based on a dynamic graph representation of an auto-supervised learning;
fig. 2 is a schematic diagram of a spatiotemporal weight encoder.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings.
The invention provides a dynamic graph representation learning method (hereinafter referred to as a method) based on self-supervision learning, which is shown in figure 1.
The invention mainly takes the E-commerce field as the background, the commodity and the user are nodes in the graph, and the interaction between the user and the commodity is an edge in the graph. The specific interaction can be understood as the following operations: browsing, collecting and purchasing. Finally, our invention will get the user and commodity characterization and predict the commodities the user may buy based on the characterization.
Firstly, the method takes any node as a central node in a graph and respectively carries out space-time random walk and random walk, and the sampled subgraphs are respectively defined as a space-time subgraph and a non-time sequence subgraph. In addition, random space-time walking is carried out by taking any other node as a central node, and the sampled subgraph is called a non-space subgraph; secondly, capturing time and space information in the graph by using a space-time weight encoder to obtain node characteristics and sub-graph characteristics of corresponding nodes; and finally, the characterization of the node is learned by maximizing mutual information between the node characterization and the subgraph characterization through a contrast learning method. The specific technical scheme is as follows:
(1) generating subgraphs
As shown in FIG. 1, a dynamic graph is given
Figure BDA0003620467090000091
Where V represents a set of edges and,
Figure BDA0003620467090000092
is a set of a series of edges, where there is an edge E ═ (v, w, t) ∈ E t Representing the edges generated by node v and node w at time t,
Figure BDA0003620467090000093
representing a function that can map each edge to a timestamp. The invention hopes to acquire a space-time subgraph G taking a node v as a central node in a dynamic graph G i
For spatio-temporal subgraph G i The sampling is mainly divided into the following two steps: initial edge selection and spatio-temporal random walk.
In the initial edge selection step, sampling is performed by taking the node v as a center node. There are two ways in which this approach can be chosen, one is called unbiased sampling and the other is called biased sampling.
In the initial edge selection step, sampling is performed with node v as the center node. The samples include unbiased samples and biased samples.
For unbiased sampling, define the set of all edges around node v as Ψ v The initial edge is sampled using an evenly distributed sampling approach. The edge e ∈ Ψ (v, w, t) ∈ Ψ v The probability of being selected is defined as:
Figure BDA0003620467090000094
for biased sampling, it is assumed that all edges around node v are selected with different probabilities. Biased sampling will assign different sampling probabilities to edges depending on when they occur. Specifically, the edge e ═ v, w, t ∈ Ψ v The probability of being selected is defined as:
Figure BDA0003620467090000101
wherein e' ∈ Ψ v Representing any edge connected to the central node v,
Figure BDA0003620467090000102
the time corresponding to the edge e' is,
Figure BDA0003620467090000103
represents the time corresponding to the edge e and t min Is the minimum time associated with an edge in the dynamic graph. Biased sampling is to encourage sampling of later occurring edges.
The task mainly completed by the step of space-time random walk is to select the next node. The selection of the next node may be selected from the set of spatio-temporal neighbors Γ. The set of spatiotemporal neighbors may be defined as:
Figure BDA0003620467090000104
wherein, gamma is t (v) Represents the set of spatio-temporal neighbors of node v at time t, w represents the nodes adjacent to node v, and t' represents the time on the edge between node w and node v.
Similarly, the invention also has two space-time random walk strategies, which can be defined as biased space-time random walk and unbiased space-time random walk.
For unbiased spatio-temporal random walk, given an arbitrary edge e ═ u, v, t, the spatio-temporal neighborhood w ∈ Γ of node v at time t t (v) The probability of being selected is as follows:
Figure BDA0003620467090000105
for biased space-time random walk, the space-time neighbor w epsilon Γ of the node v at the time t t (v) The probability of being selected is:
Figure BDA0003620467090000106
wherein w ' represents any space-time neighbor of the node v at the time t, δ (w) is the time corresponding to the node w, and δ (w ') is the time corresponding to the node w '.
Furthermore, we refer to the idea of random walk, i.e. sampling without taking into account the time factor when sampling. Random walk and sampling non-time sequence subgraph with node v as center node
Figure BDA0003620467090000111
And performing space-time random walk by taking the node o (wherein the node o and the node v are different nodes) as a central node to obtain a non-space subgraph G o
(2) Space-time weight encoder
In order to capture the temporal and spatial characterization of the nodes, the present invention designs a spatio-temporal weight encoder, as shown in fig. 2. Specifically, in obtaining G v The spatio-temporal weight encoder is then used to output a representation of each node in the spatio-temporal subgraph. And temporal and spatial weight reprogrammingThe main goal of the coders is to avoid the so-called characterization obsolescence problem. It contains two components, called temporal aggregator and spatial aggregator.
For the time aggregator, when an event occurs involving the node itself, the timing of node v at time t characterizes z v (t) is expressed as:
z v (t)=LSTM(f (v,w) (t),h v (t - ))
where LSTM is represented as the LSTM model, f (v,w) (t) represents the interaction feature between nodes v and w at time t, h v (t - ) Is the time that node v last updated.
For the spatial aggregator, this component is used to incorporate characterization information of spatio-temporal neighbors, which can be assigned different weights according to different times of occurrence of neighboring nodes. In particular, the spatial characterization of node v at time t
Figure BDA0003620467090000112
The definition is as follows:
Figure BDA0003620467090000113
wherein w ∈ Neighbor denotes a node w connected to the node v,
Figure BDA0003620467090000114
representing the calculation of the current time t and the time t w Temporal coding of the spacing, z w (t) is a time-series representation of node W at time t, W 1 Are learnable parameters.
Finally, we characterize z by aggregation timing v (t) and spatial characterization
Figure BDA0003620467090000115
Obtaining the representation h of the node v at the time t v (t):
Figure BDA0003620467090000121
Wherein, W 2 Are learnable parameters. Notably, the graph-level characterization s of the spatio-temporal subgraph v Is obtained by a readout function.
Similarly, non-time-series subgraphs can be obtained by a space-time weight coder
Figure BDA0003620467090000122
Graph level characterization of
Figure BDA0003620467090000123
And a non-spatial subgraph G o Graph-level characterization s o
(3) Spatio-temporal contrast learning
And training the space-time weight encoder by comparing positive and negative samples. The specific contrast learning objective is to distinguish spatio-temporal subgraphs, non-temporal subgraphs and non-spatial subgraphs. For the selection of the positive sample, the invention forms the positive sample by the central node representation of the spatio-temporal subgraph and the graph level representation of the spatio-temporal subgraph.
Secondly, the invention maps the non-time sequence
Figure BDA0003620467090000124
Graph level characterization of
Figure BDA0003620467090000125
And spatio-temporal subgraph G v Characterization of the center node of h v Time negative examples are formed and are called time negative examples. The loss function for a time negative example is defined as:
Figure BDA0003620467090000126
wherein, sigma represents sigmoid function, phi is marginal value,
Figure BDA0003620467090000127
representing an expectation function, s v For graph-level characterization of spatio-temporal subgraphs, h v For the node v characterization in the spatio-temporal subgraph,
Figure BDA0003620467090000128
graph-level characterization for non-time-series subgraphs.
Furthermore, a non-spatial subgraph G o Graph characterization s of o And spatio-temporal subgraph G v Characterization of the center node of h v Constituting a spatial negative example. The loss function for a spatial negative example is defined as:
Figure BDA0003620467090000131
where, σ is the sigmoid function,
Figure BDA0003620467090000132
is the value of the margin at which,
Figure BDA0003620467090000133
representing an expectation function, s v For graph-level characterization of spatio-temporal subgraphs, h v For the characterization of a node v in a spatio-temporal subgraph, s o Is a graph-level characterization of a non-spatial subgraph. The total loss function of the space-time weight encoder is therefore:
Figure BDA0003620467090000134
where λ is the equilibrium parameter of the total loss function.
And S4, completing the training of the space-time weight encoder through a total loss function, and finally sending all nodes in the dynamic graph into the trained space-time weight encoder to obtain the node representation of the dynamic graph. By applying the method to the E-commerce scene, the commodities which are possibly purchased by the user can be predicted.
1. In order to depict the time and space information of the dynamic graph and more efficiently model the original graph, the invention provides a method for utilizing a comparative learning thought and further enriching the time and space information in the learning process of the dynamic graph representation by sampling subgraphs on the dynamic graph as auxiliary graphs, and the method has good interpretability.
2. Aiming at the characteristic that nodes and edges in a subgraph are dynamically changed, the invention provides a space-time weight encoder to capture the time and space characteristics in the dynamic graph.
3. The space-time sampling graph method is relatively independent, and the generated subgraph can be used as an auxiliary graph of other dynamic graph representation learning methods and has strong mobility.
4. The method is used for capturing the time and space information on the dynamic graph, and experiments show that the method can be applied to e-commerce scenes to predict commodities purchased by users.
The examples given herein are given solely for the purpose of illustrating the invention and are not to be construed as limiting the embodiments of the invention, as those skilled in the art will be able to make other variations and modifications based on the above teachings, which are not exhaustive of all embodiments and obvious variations and modifications of the invention as may be within the scope of the appended claims.
Those not described in detail in this specification are within the skill of the art.

Claims (7)

1. A dynamic graph representation learning method based on self-supervision learning is characterized by comprising the following steps:
s1, sampling any node in the graph as a central node to obtain a space-time subgraph, a non-time sequence subgraph and a non-space subgraph respectively;
s2, capturing time and space information in the graph by using a space-time weight encoder to obtain node characteristics of corresponding nodes and graph level characteristics of the subgraph;
s3, learning the node representation by comparing the mutual information between the maximum node representation and the graph level representation of the subgraph in a learning method, and respectively defining a total loss function from the aspects of time and space;
and S4, completing the training of the space-time weight encoder through a total loss function, and finally sending all the nodes in the dynamic graph into the trained space-time weight encoder to obtain the node representation of the dynamic graph.
2. The method for learning a dynamic graph representation based on the self-supervised learning as claimed in claim 1, wherein the step S1 comprises the following steps: given a dynamic graph
Figure FDA0003620467080000011
Where V represents the set of edges and,
Figure FDA0003620467080000012
is a set of a series of edges, where there is an edge E ═ v, w, t ∈ E in the graph t Representing the edges generated by node v and node w at time t,
Figure FDA0003620467080000013
represents a function that can map each edge to a timestamp;
time-space diagram G v The sampling comprises the following two steps: initial edge selection and spatio-temporal random walk.
3. The self-supervised learning based dynamic graph representation learning method as claimed in claim 2, wherein: in the initial edge selection step, sampling is carried out by taking the node v as a central node; the sampling comprises unbiased sampling and biased sampling;
for unbiased sampling, define all edge sets around node v as Ψ v Sampling the initial edge by using a uniformly distributed sampling mode; the edge e ═ v, w, t ∈ Ψ v The probability of being selected is defined as:
Figure FDA0003620467080000021
for biased sampling, all edges around a node v are determined to have different probabilities to be selected; the biased sampling allocates different sampling probabilities to the edge according to the time of the edge; in particular toIn other words, the edge e ═ (v, w, t) ∈ Ψ v The probability of being selected is defined as:
Figure FDA0003620467080000022
wherein e' ∈ Ψ v Representing any edge connected to the central node v,
Figure FDA0003620467080000023
the time corresponding to the edge e' is,
Figure FDA0003620467080000024
indicates the time, t, corresponding to the edge e min Is the minimum time associated with an edge in the dynamic graph.
4. The self-supervised learning based dynamic graph representation learning method as claimed in claim 2, wherein: the main task of the space-time random walk step is to select the next node; the next node is selected from the space-time neighbor set gamma; the set of spatio-temporal neighbors is defined as:
Figure FDA0003620467080000025
wherein, gamma is t (v) Representing a spatio-temporal neighbor set of a node v at time t, w representing a node adjacent to the node v, and t' representing time on an edge between the node w and the node v;
the space-time random walk comprises biased space-time random walk and unbiased space-time random walk;
for unbiased spatio-temporal random walk, given an arbitrary edge e ═ u, v, t, the spatio-temporal neighborhood w ∈ Γ of node v at time t t (v) The probability of being selected is as follows:
Figure FDA0003620467080000031
for biased space-time random walk, the space-time neighbor w epsilon Γ of the node v at the time t t (v) The probability of being selected is:
Figure FDA0003620467080000032
wherein w ' represents any space-time neighbor of the node v at the time t, δ (w) is the time corresponding to the node w, and δ (w ') is the time corresponding to the node w '.
5. The self-supervised learning based dynamic graph representation learning method as claimed in claim 2, wherein: the idea of random walk is introduced, and sampling is carried out without considering time factors during sampling; random walk and sampling non-time sequence subgraph with node v as center node
Figure FDA0003620467080000033
And performing space-time random walk by taking the node o as a central node to obtain a non-space subgraph G o
6. The method for learning a dynamic graph representation based on the self-supervised learning as claimed in claim 1, wherein the step S2 comprises the following steps: the space-time weight encoder comprises a time aggregator and a space aggregator;
for the time aggregator, when an event occurs involving the node itself, the timing of node v at time t characterizes z v (t) is expressed as:
z v (t)=LSTM(f (v,w) (t),h v (t - ))
where LSTM is represented as the LSTM model, f (v,w) (t) represents the interaction feature between nodes v and w at time t, h v (t - ) Is the last update time of node v;
the space aggregator is used for merging the representation information of the space-time neighbors and distributing different weights according to different occurrence times of the neighbor nodes; spatial characterization of node v at time t
Figure FDA0003620467080000034
The definition is as follows:
Figure FDA0003620467080000041
wherein w ∈ Neighbor denotes a node w connected to the node v,
Figure FDA0003620467080000042
represents calculating the current time t and the time t w Temporal coding of the spacing, z w (t) is a time-series representation of node W at time t, W 1 Is a learnable parameter;
finally, characterization of z by polymerization timing v (t) and spatial characterization
Figure FDA0003620467080000043
Obtaining the node characterization h of the node v at the time t v (t):
Figure FDA0003620467080000044
Wherein, W 2 Is a learnable parameter;
non-time series diagram
Figure FDA0003620467080000045
Graph level characterization of
Figure FDA0003620467080000046
And a non-spatial subgraph G o Graph level characterization s o Obtained by a space-time weight encoder; graph-level characterization s of spatio-temporal subgraphs v Obtained by the readout function.
7. The self-supervised learning-based dynamic graph representation learning method as claimed in claim 1, wherein the specific steps of step S3 are as follows: training a space-time weight encoder by comparing positive and negative samples; the comparison learning target is to distinguish a space-time subgraph, a non-time sequence subgraph and a non-space subgraph; for the selection of the positive sample, constructing the node characterization of the spatio-temporal subgraph and the graph level characterization of the spatio-temporal subgraph into the positive sample;
will not be time sequential subgraph
Figure FDA0003620467080000047
Graph level characterization of
Figure FDA0003620467080000048
And spatio-temporal subgraph G v Characterization of nodes of h v Forming a time negative example, the loss function of which is defined as:
Figure FDA0003620467080000049
wherein, sigma represents sigmoid function, phi is marginal value,
Figure FDA00036204670800000410
representing an expectation function, s v For graph-level characterization of spatio-temporal subgraphs, h v For node characterization of node v in the spatio-temporal subgraph,
Figure FDA00036204670800000411
graph-level characterization for non-temporal subgraphs;
non-spatial subgraph G o Graph-level characterization s o And spatio-temporal subgraph G v Characterization of nodes of h v Forming a space negative sample; the loss function for a spatial negative example is defined as:
Figure FDA0003620467080000051
where, σ is the sigmoid function,
Figure FDA0003620467080000052
is a value of a margin, which is,
Figure FDA0003620467080000053
representing an expectation function, s v For graph-level characterization of spatio-temporal subgraphs, h v For node characterization of node v in a spatio-temporal subgraph, s o Graph-level characterization for non-spatial subgraphs;
the total loss function of the space-time weight encoder is:
Figure FDA0003620467080000054
where λ is the equilibrium parameter of the total loss function.
CN202210455958.4A 2022-04-28 2022-04-28 Dynamic graph representation learning method based on self-supervision learning Pending CN115017368A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210455958.4A CN115017368A (en) 2022-04-28 2022-04-28 Dynamic graph representation learning method based on self-supervision learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210455958.4A CN115017368A (en) 2022-04-28 2022-04-28 Dynamic graph representation learning method based on self-supervision learning

Publications (1)

Publication Number Publication Date
CN115017368A true CN115017368A (en) 2022-09-06

Family

ID=83067000

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210455958.4A Pending CN115017368A (en) 2022-04-28 2022-04-28 Dynamic graph representation learning method based on self-supervision learning

Country Status (1)

Country Link
CN (1) CN115017368A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116561688A (en) * 2023-05-09 2023-08-08 浙江大学 Emerging technology identification method based on dynamic graph anomaly detection

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116561688A (en) * 2023-05-09 2023-08-08 浙江大学 Emerging technology identification method based on dynamic graph anomaly detection
CN116561688B (en) * 2023-05-09 2024-03-22 浙江大学 Emerging technology identification method based on dynamic graph anomaly detection

Similar Documents

Publication Publication Date Title
CN109891897B (en) Method for analyzing media content
Moons et al. Distilling optimal neural networks: Rapid search in diverse spaces
US20210174152A1 (en) Video classification method and server
Rottmann et al. Uncertainty measures and prediction quality rating for the semantic segmentation of nested multi resolution street scene images
Cheng et al. Evolutionary support vector machine inference system for construction management
EP3918472B1 (en) Techniques to detect fusible operators with machine learning
Lemonnier et al. Multivariate Hawkes processes for large-scale inference
US11861467B2 (en) Adaptive quantization for execution of machine learning models
Lu et al. Automating deep neural network model selection for edge inference
CN112464057A (en) Network data classification method, device, equipment and readable storage medium
CN115017368A (en) Dynamic graph representation learning method based on self-supervision learning
Li et al. Short-term traffic forecasting using high-resolution traffic data
Hou et al. NeuLens: spatial-based dynamic acceleration of convolutional neural networks on edge
WO2021169478A1 (en) Fusion training method and apparatus for neural network model
Jiang et al. Read: Aggregating reconstruction error into out-of-distribution detection
CN117315331A (en) Dynamic graph anomaly detection method and system based on GNN and LSTM
CN112910680A (en) Network embedding method for fusing multi-granularity community information
CN108596068B (en) Method and device for recognizing actions
Koesdwiady et al. Methods to improve multi-step time series prediction
CN113255899B (en) Knowledge distillation method and system with self-correlation of channels
Wu et al. Forecasting online adaptation methods for energy domain
CN115544307A (en) Directed graph data feature extraction and expression method and system based on incidence matrix
WO2022162839A1 (en) Learning device, learning method, and recording medium
CN116090504A (en) Training method and device for graphic neural network model, classifying method and computing equipment
Peng et al. An extreme learning machine and gene expression programming-based hybrid model for daily precipitation prediction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination