CN116340842A - Common attention-based heterogeneous graph representation learning method - Google Patents

Common attention-based heterogeneous graph representation learning method Download PDF

Info

Publication number
CN116340842A
CN116340842A CN202310309514.4A CN202310309514A CN116340842A CN 116340842 A CN116340842 A CN 116340842A CN 202310309514 A CN202310309514 A CN 202310309514A CN 116340842 A CN116340842 A CN 116340842A
Authority
CN
China
Prior art keywords
target
vector
path
node
meta
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310309514.4A
Other languages
Chinese (zh)
Inventor
张政
季颖生
佴湲坤
沈佳辰
陶振宇
蔡明�
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Zhongyi Turing Digital Technology Co ltd
Original Assignee
Shanghai Zhongyi Turing Digital Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Zhongyi Turing Digital Technology Co ltd filed Critical Shanghai Zhongyi Turing Digital Technology Co ltd
Priority to CN202310309514.4A priority Critical patent/CN116340842A/en
Publication of CN116340842A publication Critical patent/CN116340842A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Abstract

The invention provides a heterogeneous diagram representation learning method based on common attention, which comprises the following steps: determining local attention scores of neighbor nodes of the target element path based on a preset attention parameter vector, a first hidden feature vector, a second hidden feature vector and a semantic fusion vector of the target element path; determining a global attention score of the target element path neighbor node based on the local attention score, the number of other element paths corresponding to the target element path neighbor node and a preset super parameter; determining an updated semantic representation vector of the target node based on the global attention score, the first hidden feature vector, the second hidden feature vector, and the semantic fusion vector; and determining the attention weight of each target element path based on the updated semantic representation vector of the node in each target element path in the heterogeneous graph, and determining the target semantic representation vector of the target node based on the attention weight and the updated semantic representation vector. By the method, high-quality node representation can be obtained.

Description

Common attention-based heterogeneous graph representation learning method
Technical Field
The invention relates to the technical field of heterogeneous diagram representation learning, in particular to a heterogeneous diagram representation learning method based on common attention.
Background
Graph representation learning is a method of mapping graph data to low-dimensional vectors that aims to capture the topology on a graph, learn the representation of nodes or edges in low-dimensional space. In recent years, heterogeneous charts represent learning that has received a great deal of attention.
The graph representation learning method comprises graph representation learning based on a homogeneous graph and graph representation learning based on a heterogeneous graph, wherein the homogeneous graph is composed of nodes and edges of the same type. In the real world, systems are often made up of complex interrelationships of various components with one another. The traditional graph based on the homogeneous graph shows that the learning method can cause a great deal of heterogeneous information loss, thereby affecting the effect of downstream tasks. Thus, more and more methods attempt to abstract a real system into a heterogram, i.e., a heterogram is composed of multiple types of nodes or edges; the advantage of heterogeneous maps is that more information can be integrated than homogeneous maps.
However, heterogeneous graphs have a complex topology, and nodes and edges have heterogeneity. In the related art, semantic information on a heterogeneous graph is generally not efficiently and sufficiently mined, thereby limiting the semantic expression capability of the heterogeneous graph.
Therefore, how to effectively mine the abundant semantic information in the heterogeneous graph is a problem to be solved at present.
Disclosure of Invention
Aiming at the problems existing in the prior art, the embodiment of the invention provides a heterogeneous diagram representation learning method based on common attention.
The invention provides a heterogeneous diagram representation learning method based on common attention, which comprises the following steps:
determining local attention scores of neighbor nodes of any target element path in a heterogeneous graph based on preset attention parameter vectors, first hidden feature vectors, second hidden feature vectors and semantic fusion vectors of all element path examples between the target node and the neighbor nodes of the target element path for at least one target node of the target element path in the heterogeneous graph; the local attention score is used for representing the importance degree of the target element path neighbor node to the target node based on the target element path; the first hidden feature vector is the hidden feature vector of the target node, and the second hidden feature vector is the hidden feature vector of the target element path neighbor node;
determining a global attention score of the target element path neighbor node based on the local attention score, the number of other element paths corresponding to the target element path neighbor node in the heterogeneous graph and a preset super parameter; the global attention score is used for representing the importance degree of all meta-path neighbor nodes in the heterogeneous graph to the target node based on all meta-paths in the heterogeneous graph, and the super-parameters are used for representing the importance degree of all meta-path examples between the target node and the target meta-path neighbor nodes to the target node;
Determining an updated semantic representation vector corresponding to the target node based on the global attention score, the first hidden feature vector, the second hidden feature vector, and the semantic fusion vector;
and determining the attention weight of each target element path based on the updated semantic representation vector of the node in each target element path in the heterogeneous graph, and determining the target semantic representation vector corresponding to the target node based on the attention weight of each target element path and the updated semantic representation vector of the node in each target element path.
Optionally, the determining, based on the global attention score, the first hidden feature vector, the second hidden feature vector, and the semantic fusion vector, an updated semantic representation vector corresponding to the target node includes:
normalizing the global attention score to generate a target attention score;
performing aggregation processing on the target attention score, the second hidden feature vector and the semantic fusion vector to obtain an aggregation message vector corresponding to the target node;
the update semantic representation vector is determined based on the aggregate message vector and the first hidden feature vector.
Optionally, the determining the global attention score of the target meta-path neighboring node based on the local attention score, the number of other meta-paths corresponding to the target meta-path neighboring node in the heterogeneous graph, and a preset super-parameter includes:
determining a global attention score of the target node by using a formula (1) based on the local attention score, the number of other element paths corresponding to the neighbor nodes of the target element path in the heterogeneous graph and a preset super parameter; the formula (1) is:
Figure BDA0004147938870000031
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA0004147938870000032
representing the global attention score; n represents the number of other meta paths corresponding to the target meta path neighbor nodes in the heterogeneous graph; η represents the preset hyper-parameters; />
Figure BDA0004147938870000033
Representing the local attention score.
Optionally, before the determining the local attention score of the target meta-path neighboring node based on the preset attention parameter vector, the first hidden feature vector, the second hidden feature vector, and the semantic fusion vector of all meta-path instances between the target node and the target meta-path neighboring node, the method further includes:
acquiring at least one meta-path instance corresponding to the target meta-path, wherein the data structure of each meta-path instance is the same as the data structure of the target meta-path;
Encoding each meta-path instance to generate a semantic representation vector of each meta-path instance;
and carrying out aggregation processing on the semantic representation vectors of each meta-path instance to generate the semantic fusion vector.
Optionally, the meta-path instance includes l+1 child nodes and L relationship edges, where L is a positive integer;
encoding each meta-path instance to generate a semantic representation vector of each meta-path instance, including:
generating a child node sequence and a relation edge sequence corresponding to each meta-path instance aiming at each meta-path instance; the child node sequence comprises L+1 child nodes, and the relationship edge sequence comprises L relationship edges;
inputting a child node hidden characteristic vector sequence into a first model group, and updating the child node hidden characteristic vector sequence to obtain an updated hidden characteristic vector of the child node; the first model group comprises a first cyclic neural network model (RNN) and a second RNN, and the first RNN and the second RNN are connected in a jumping manner; the child node hidden feature vector sequence is generated based on the child node sequence;
inputting a relation edge hidden characteristic vector sequence into the first model group, and updating the relation edge hidden characteristic vector sequence to obtain an updated hidden characteristic vector of the relation edge; the relation edge hidden feature vector sequence is generated based on the child node sequence;
And generating a semantic representation vector of the target element path instance based on the updated hidden feature vector of the child node and the updated hidden feature vector of the relation edge.
Optionally, after the determining the target semantic representation vector corresponding to the target node, the method further includes:
inputting the target semantic representation vector into an initial target service classification model, and training the initial target service classification model by using a preset loss function until the initial target service classification model converges to obtain a trained target service classification model and a target semantic representation vector which is output by the target service classification model and is suitable for target service;
and inputting the target semantic representation vector suitable for the target service into the trained target service classification model to obtain a classification result aiming at the target service, which is output by the trained target service classification model.
The heterogeneous graph representation learning method based on common attention provided by the invention can determine the local attention score of the neighbor node of the target element path based on the preset attention parameter vector of the target element path, the hidden feature vector of the target node, the hidden feature vector of the neighbor node of the target element path and the semantic fusion vector of all element path examples between the target node and the neighbor node of the target element path aiming at least one target node of any target element path in the heterogeneous graph, wherein the local attention score is used for representing the importance degree of the neighbor node of the target element path to the target node based on the target element path; then, based on the local attention score, the number of other element paths corresponding to the target element path neighbor nodes in the heterogeneous graph and preset super parameters, the global attention score of the target element path neighbor nodes can be determined, wherein the global attention score is used for representing the importance degree of the target element path neighbor nodes to the target nodes based on all element paths in the heterogeneous graph; based on the global attention score, the hidden feature vector of the target node, the hidden feature vector of the neighbor node of the target node and the semantic fusion vector, the complex semantics can be adaptively and deeply mined, so that the semantic feature representation of the target node is enhanced, the updated semantic representation vector corresponding to the target node is obtained, and the defect that the conventional heterogeneous graph mining algorithm excessively relies on element paths to carry out semantic mining is overcome; based on the attention weight of each target element path and the updated semantic representation vector of the nodes in each target element path, the determined target semantic representation vector corresponding to the target node can be fused with the semantic importance on different target element paths, so that the target node can effectively express rich semantic information in the heterogeneous graph, and high-quality node representation is obtained.
Drawings
In order to more clearly illustrate the invention or the technical solutions of the prior art, the following description will briefly explain the drawings used in the embodiments or the description of the prior art, and it is obvious that the drawings in the following description are some embodiments of the invention, and other drawings can be obtained according to the drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic flow chart of a common attention-based heterogeneous diagram representation learning method provided by the invention;
FIG. 2 is a schematic illustration of a co-attention based semantic feature computation provided by the present invention;
FIG. 3 is a schematic diagram of a meta-path instance encoding structure provided by an embodiment of the present invention;
FIG. 4 is a schematic diagram showing a learning method based on a common attention heterogram.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the present invention more apparent, the technical solutions of the present invention will be clearly and completely described below with reference to the accompanying drawings, and it is apparent that the described embodiments are some embodiments of the present invention, not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
In order to facilitate a clearer understanding of the various embodiments of the present application, some relevant background knowledge is first presented below.
Graph representation learning is a technical approach to map graph data to low-dimensional vectors, which aims to capture the topology on a graph, learn the representation of nodes or edges in low-dimensional space, also known as graph embedding. Early graphs represent that the learning algorithm is mainly based on homogeneous graphs, i.e. graphs are made up of nodes and edges of the same type. In the real world, systems are often made up of complex interrelationships of various components with one another. The traditional method based on the homogeneous map can cause a great deal of heterogeneous information loss, thereby affecting the effect of downstream tasks. Thus, more and more methods attempt to abstract a real-world system into a heterogeneous graph, i.e., a graph is composed of multiple types of nodes or edges. The advantage of heterogeneous maps is that more information can be integrated than homogeneous maps.
Heterograms represent a study that has received a great deal of attention in recent years, and this problem can be summarized as: through heterogeneous graph embedding, heterogeneous graph representation is learned in a low-dimensional space, and complex structure and semantic information on the heterogeneous graph are reserved as far as possible and used for downstream machine learning tasks. The biggest challenge in learning heterograms is how to efficiently mine complex structures and rich semantics on the graph. The concrete steps are as follows:
1) Complex topologies. Heterogeneous graphs are abstract representations of complex systems in reality, and contain rich semantics including node objects of different types and interrelationships between objects with different meanings. How to represent the fusion of such information becomes a great challenge;
2) Nodes and edges have heterogeneity. On heterogeneous graphs, different types of nodes and edges represent different meanings, with different characteristics, including dimensions, data types, and numerical ranges, etc. How to aggregate different types of neighbors and their heterogeneous information is another challenge.
The common heterogeneous graph representation learning method has the following technical means.
a) A divide-and-conquer based method.
The main idea of the method is based on a random walk method. The method mainly comprises the following steps: the method comprises the steps of presetting a walk mode (such as a meta path, a meta structure and the like), carrying out walk on a heterogeneous graph according to the mode, sampling to obtain a node sequence containing heterostructure and semantic information, and then carrying out representation learning on the node sequence. Typical algorithms are e.g. metaath 2vec, HIN2Vec, HINE, ESim etc.
b) An automatic encoder-based method.
The method mainly comprises the following steps: and constructing encoder learning node information and graph structure information by adopting a neural network model, and reconstructing original heterogeneous graph information by a decoder. First, the sub-graph structure is extracted by techniques such as meta-path. The design encoder encodes the sub-graph structure and then fuses. Finally, a reconstruction loss function is designed to train the whole network. Typical algorithms are BL-MNE, SHINE, etc.
c) Based on the method of generating the challenge network.
The method mainly comprises the following steps: a robust node representation is generated by a mutual game between the generator and the discriminator. Here the generator and discriminator need to integrate the relationships between the different nodes and capture heterogeneous semantic information. Typical algorithms are e.g. HeGAN, MV-ACM, etc.
Message-based methods. The method mainly comprises the following steps: and designing a message passing mechanism and constructing a graph neural network model aiming at the heterogeneous graph. And collecting node and neighborhood information thereof, then fusing, and delivering a fused result to the neighbor nodes so as to obtain higher-order knowledge. Finally, a node representation is obtained. Typical algorithms are HAN, hetGNN, MAGNN, etc.
However, in the related art, the heterogeneous graph representation learning has the following drawbacks:
a) The method based on division and treatment comprises the following steps:
the method has the main defects that the topological structure relationship of the heterogeneous diagram is complex, and the heterogeneous diagram is decomposed into a plurality of homogeneous diagrams, so that heterogeneous information is lost, and the heterogeneous structure and semantic information on the diagram cannot be effectively and fully mined.
b) A random walk-based method:
the method is mainly used for analyzing the graph structure, namely learning the relation between nodes on the graph, and cannot be used for mining heterogeneous attributes of different types of nodes and edges on the graph.
c) An automatic encoder-based method:
the method based on the automatic encoder mainly relies on an objective function for reconstruction, but the topological structure of the heterogeneous graph is complex, and a comprehensive reconstruction target is difficult to design, so that the heterogeneous structure and semantic information on the graph cannot be fully mined.
d) Based on the method of generating the countermeasure network:
the main defects are that: firstly, the method based on generating the countermeasure network is often high in training cost and is not suitable for a large-scale network; secondly, most of the existing methods generate samples from known nodes, and generalize the unknown nodes with poor capability; finally, existing methods still lack empirical knowledge of the combinations for the complex semantics contained in the heterograms.
e) Message-based methods.
Firstly, because the network structure is complex, the noise is easier to fit; secondly, the method has higher overall space-time complexity, which results in relatively higher training difficulty; finally, most of the existing methods adopt a meta path or a meta graph to mine complex semantics in heterogeneous graphs, which on one hand needs more accurate priori knowledge and is difficult to realize in some scenes; on the other hand, the expression capacity of the model is limited by doing so, and the complex semantics cannot be further deeply mined.
In summary, in order to effectively mine abundant semantic information in the heterogeneous graphs, the embodiment of the invention provides a heterogeneous graph representation learning method based on common attention.
The common-attention-based heterogeneous diagram representation learning method provided by the invention is specifically described below with reference to fig. 1 to 4. Fig. 1 is a schematic flow chart of a learning method represented by a hetero graph based on common attention, and referring to fig. 1, the method includes steps 101-104, in which:
step 101, determining a local attention score of a target element path neighbor node according to a preset attention parameter vector, a first hidden feature vector, a second hidden feature vector and a semantic fusion vector of all element path examples between the target node and the target element path neighbor node of any target element path in a heterogeneous graph; the local attention score is used for representing the importance degree of the target element path neighbor node to the target node based on the target element path; the first hidden feature vector is the hidden feature vector of the target node, and the second hidden feature vector is the hidden feature vector of the target element path neighbor node.
In this embodiment, the heterogeneous graph includes a plurality of target element paths, each of which includes a plurality of target nodes and a relationship edge connecting the target nodes.
Wherein a given heterogeneous information attribute network (i.e., heterogeneous graph) may be represented as
Figure BDA0004147938870000091
The node set comprises a plurality of target nodes, which can be expressed as +.>
Figure BDA0004147938870000092
The relation edge set including a plurality of relation edges connecting the target nodes can be expressed as +.>
Figure BDA0004147938870000093
Note that, node type set
Figure BDA0004147938870000094
Satisfying the node type mapping function phi:>
Figure BDA0004147938870000095
relation edge type set +.>
Figure BDA0004147938870000096
Satisfy the relation edge type mapping function psi:>
Figure BDA0004147938870000097
and satisfy->
Figure BDA0004147938870000098
Wherein the node attribute set comprises the attribute of each target node, and the node attribute set can be expressed as +.>
Figure BDA0004147938870000099
The attribute set of the relationship side comprises the attribute of each relationship side, and the attribute set of the relationship side can be expressed as
Figure BDA00041479388700000910
The node attribute matrix is composed of node attribute matrixes of different types and relation edge attribute matrixes of different types.
Wherein the method comprises the steps of
Figure BDA00041479388700000911
And->
Figure BDA00041479388700000912
Respectively representing the number of different types of target nodes and the number of different types of relation edges in the heterogeneous information attribute network,/->
Figure BDA00041479388700000913
And->
Figure BDA00041479388700000914
Respectively representing the attribute number of the corresponding node type and the attribute number of the corresponding relation edge type.
Further, in the heterogeneous information network, different nodes are generally associated by adopting meta paths, so that rich semantic information contained in the meta paths can be embodied, and the heterogeneous network is facilitated to be mined. The meta-path P is denoted here as
Figure BDA0004147938870000101
Is a path of (a) describing node type A 1 And A l+1 A complex relationship between->
Figure BDA0004147938870000102
Wherein->
Figure BDA0004147938870000103
Representing complex operators between relationships.
In particular, the meta-path instance P of P represents a sequence of nodes and relationship edges that satisfy the meta-path definition. Here, for the input heterogeneous diagram
Figure BDA0004147938870000104
There is a set of pre-set meta-paths +.>
Figure BDA0004147938870000105
In one implementation, the meta-path +.>
Figure BDA0004147938870000106
The head and tail target nodes of (a) are all target type nodes. Thus, for each target type node, a different meta-path sub-graph may be extracted from the set of meta-paths.
In this embodiment, neighbor message aggregation is first required. FIG. 2 is a schematic diagram of the calculation of semantic features based on common attention according to the present invention, as shown in FIG. 2, and in FIG. 2, a plurality of meta-path subgraphs G are included 1 、G 2 、...、G k The method comprises the steps of carrying out a first treatment on the surface of the By introducing other semantic information and sharing its attention parameter a 1 ,W 1 ;A 2 ,W 2 ;...;A k ,W k Obtaining a semantic representation vector h of the local semantics 1 、h 2 、...、h k
Specifically, a target node u and a set of element paths are given
Figure BDA0004147938870000107
For a particular target meta-path P k It is required to base the target meta path P k Is +.>
Figure BDA0004147938870000108
Hidden feature vector h of target node u Hidden target element path neighbor node Feature vector h v And semantic fusion vector +.A semantic fusion vector of all meta-path examples between the target node and the target meta-path neighbor node +.>
Figure BDA0004147938870000109
Determining the importance degree of the target element path neighbor node v to the target node u based on the target element path +.>
Figure BDA00041479388700001010
In particular, the local graph attention scoring function, i.e., the expression (2), may be used:
Figure BDA00041479388700001011
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA00041479388700001012
representing local attention scores of the target element path neighbor nodes; />
Figure BDA0004147938870000111
Is the target meta-path P k Is a vector of attention parameters of (a); />
Figure BDA0004147938870000112
Representing a parameter matrix; />
Figure BDA0004147938870000113
A semantic fusion vector representing a meta-path; h is a u A hidden feature vector representing the target node; h is a v A hidden feature vector representing a target element path neighbor node; sigma (·) represents the activation function; the term "vector" means a vector concatenation.
In the embodiment, the importance degree of the target element path neighbor node to the target node based on the target element path can be determined by calculating the local attention score of the target element path neighbor node; namely, the importance of different neighbor nodes in the target element path can be learned through the local attention score, and the method has strong characterization capability.
102, determining a global attention score of the target element path neighbor node based on the local attention score, the number of other element paths corresponding to the target element path neighbor node in the heterogeneous diagram and a preset super parameter; the global attention score is used for representing the importance degree of all meta-path neighbor nodes in the heterogeneous graph to the target node based on all meta-paths in the heterogeneous graph, and the super-parameters are used for representing the importance degree of all meta-path examples between the target node and the target meta-path neighbor nodes to the target node.
In this embodiment, the local attention score is based on the local importance of a single semantic, where global information is lacking. Therefore, cross-semantic information needs to be introduced to calculate the global attention value of the target element path neighbor node, so that the importance degree of all element path examples between the target node and the target element path neighbor node to the target node can be calculated more accurately, and the noise is reduced.
Optionally, based on the local attention score of each target node, the number of other element paths corresponding to the target element path neighbor nodes in the heterogeneous graph and a preset super parameter, determining the global attention score of the target element path neighbor nodes by using a global graph attention scoring function, namely, a formula (1); the formula (1) is:
Figure BDA0004147938870000114
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA0004147938870000115
representing the global attention score; n represents the number of other element paths corresponding to the neighbor nodes of the target element path in the heterogeneous graph; η represents the preset hyper-parameters; />
Figure BDA0004147938870000121
Representing the local attention score.
It should be noted that the number of the substrates,
Figure BDA0004147938870000122
the importance of other semantics is expressed, and the larger the value is, the higher the importance of other semantics is. Especially when η=0, the global graph attention scoring function is degraded to a general local graph attention scoring function. Notably, the parameters of the global graph attention scoring function are shared across all meta-paths, which can introduce global information without increasing model complexity.
In the embodiment, a graph learning method based on multi-semantic common attention is adopted, and the deep mining is performed on local semantic features by introducing other semantic information and sharing attention parameters, redundant information is filtered, sparse information is enhanced, so that the difference of different semantics is eliminated, and the semantic feature mining is enhanced.
Step 103, determining an updated semantic representation vector corresponding to the target node based on the global attention score, the first hidden feature vector, the second hidden feature vector and the semantic fusion vector.
In this embodiment, the update semantic representation vector corresponding to the target node u may be represented as
Figure BDA0004147938870000123
Updating the semantic representation vector includes enriching semantic features in the heterograms.
Further, a single layer of updated semantic representation vectors may not be sufficient to obtain all of the semantic features in the heterogeneous map, so a multi-layer map neural network model (Graph Neural Network, GNN) may be stacked to capture high-order semantic information. For example, after stacking the L-layer GNN models, the update semantic representation vector corresponding to the target node u can be represented as
Figure BDA0004147938870000124
Step 104, determining the attention weight of each target element path based on the updated semantic representation vector of the node in each target element path in the heterogeneous graph, and determining the target semantic representation vector corresponding to the target node based on the attention weight of each target element path and the updated semantic representation vector of the node in each target element path.
In this embodiment, semantic feature fusion is required, where different semantic features are fused by adopting an attention mechanism, so as to obtain importance of different semantics.
For the target node u, the target semantic representation vector corresponding to the final target node may be determined based on the attention weights of other target element paths and the updated semantic representation vector after stacking the L-layer GNN model, and specifically may be represented by the following formula (3):
Figure BDA0004147938870000131
wherein q u Representing a target semantic representation vector;
Figure BDA0004147938870000132
representing the attention weights of other target meta-paths.
Figure BDA0004147938870000133
Specifically, the method can be calculated by the following formula (4):
Figure BDA0004147938870000134
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA0004147938870000135
is a semantic attention parameter vector; />
Figure BDA0004147938870000136
Representing a parameter matrix.
The heterogeneous graph representation learning method based on common attention provided by the invention can determine the local attention score of the neighbor node of the target element path based on the preset attention parameter vector of the target element path, the hidden feature vector of the target node, the hidden feature vector of the neighbor node of the target element path and the semantic fusion vector of all element path examples between the target node and the neighbor node of the target element path aiming at least one target node of any target element path in the heterogeneous graph, wherein the local attention score is used for representing the importance degree of the neighbor node of the target element path to the target node based on the target element path; then, based on the local attention score, the number of other element paths corresponding to the target element path neighbor nodes in the heterogeneous graph and preset super parameters, the global attention score of the target element path neighbor nodes can be determined, wherein the global attention score is used for representing the importance degree of the target element path neighbor nodes to the target nodes based on all element paths in the heterogeneous graph; based on the global attention score, the hidden feature vector of the target node, the hidden feature vector of the neighbor node of the target node and the semantic fusion vector, the complex semantics can be adaptively and deeply mined, so that the semantic feature representation of the target node is enhanced, the updated semantic representation vector corresponding to the target node is obtained, and the defect that the conventional heterogeneous graph mining algorithm excessively relies on element paths to carry out semantic mining is overcome; based on the attention weight of each target element path and the updated semantic representation vector of the nodes in each target element path, the determined target semantic representation vector corresponding to the target node can be fused with the semantic importance on different target element paths, so that the target node can effectively express rich semantic information in the heterogeneous graph, and high-quality node representation is obtained.
Optionally, before determining the local attention score of the target element path neighbor node based on the preset attention parameter vector, the first hidden feature vector, the second hidden feature vector and the semantic fusion vector of all element path instances between the target node and the target element path neighbor node, encoding each target element path in the heterogeneous graph to generate the semantic fusion vector of all element path instances between the target node and the target element path neighbor node; the method can be realized by the following steps (1) -step (3):
step (1), at least one meta-path instance corresponding to the target meta-path is obtained, and the data structure of each meta-path instance is the same as the data structure of the target meta-path;
step (2), encoding each meta-path instance to generate a semantic representation vector of each meta-path instance;
and (3) carrying out aggregation processing on the semantic representation vectors of the meta-path examples to generate the semantic fusion vector.
One target meta-path in the heterogeneous graph corresponds to a plurality of meta-path instances, and in this embodiment, before encoding each meta-path instance, features of nodes and relationship edges of different types in the heterogeneous graph need to be mapped to the same hidden space through feature transformation.
A linear transformation based on node and relationship edge types is employed herein. Specifically, a given node type
Figure BDA0004147938870000141
Relationship edge type +.>
Figure BDA0004147938870000142
For any node V e V A And an arbitrary relationship edge E R Has h v =W A ·x v H e =W R ·x e
Wherein x is v ∈X A ,x e ∈X R Is the initial feature vector of the feature set,
Figure BDA0004147938870000143
hidden feature vectors of node v and relationship edge e, respectively, and +.>
Figure BDA0004147938870000144
The parameter matrices of node type a and relationship edge type R, respectively.
After the characteristics of nodes and relation edges of different types in the heterogeneous graph are mapped to the same hidden space, encoding each element path instance to generate a semantic representation vector of each element path instance; however, the method is thatThen, the semantic expression vector of each element path instance is aggregated, and then a target semantic fusion vector is generated
Figure BDA0004147938870000151
In the embodiment, the characteristics of the nodes and the relation edges of different types in the heterogeneous graph can be mapped to the same hidden space through the characteristic transformation, so that semantic information in the heterogeneous graph can be conveniently and fully mined.
Optionally, the meta-path instance includes l+1 child nodes and L relationship edges, where L is a positive integer;
the encoding of each meta-path instance to generate a semantic representation vector of each meta-path instance can be realized by the following steps [1] -step [4 ]:
Step [1], aiming at each meta-path instance, generating a sub-node sequence and a relation edge sequence corresponding to the meta-path instance; the child node sequence comprises L+1 child nodes, and the relationship edge sequence comprises L relationship edges.
Specifically, a meta-path is given
Figure BDA0004147938870000152
And path instance thereof
Figure BDA0004147938870000153
It can be split into node sequences +.>
Figure BDA0004147938870000154
And relation edge sequence->
Figure BDA0004147938870000155
Step [2], inputting a child node hidden feature vector sequence into a first model group, and updating the child node hidden feature vector sequence to obtain an updated hidden feature vector of the child node; the first model group comprises a first cyclic neural network model (RNN) and a second RNN, and the first RNN and the second RNN are connected in a jumping manner; the sequence of child node hidden feature vectors is generated based on the sequence of child nodes.
Specifically, a sequence of hidden feature vectors to child nodes
Figure BDA0004147938870000156
The updating can be represented by the following formula (5) and formula (6):
Figure BDA0004147938870000157
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA0004147938870000158
an output hidden vector representing the t-th step of the hidden characteristic vector sequence of the child node; />
Figure BDA0004147938870000159
Updating hidden feature vectors representing corresponding nodes or relationship edges; />
Figure BDA0004147938870000161
Representing a parameter matrix; />
Figure BDA0004147938870000162
Representing the parameter vector.
Then will
Figure BDA0004147938870000163
GRU unit input to RNN to generate output vector of t+1 step of node sequence +. >
Figure BDA0004147938870000164
(i.e., updated hidden feature vector of child node), specifically expressed by the following formula (6):
Figure BDA0004147938870000165
and step [3], inputting the hidden characteristic vector sequence of the relation edge into the first model group, and updating the hidden characteristic vector sequence of the relation edge to obtain the updated hidden characteristic vector of the relation edge.
Specifically, the feature vector sequence is hidden from the relation edge
Figure BDA0004147938870000166
The updating can be represented by the following formulas (7) and (8):
Figure BDA0004147938870000167
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA0004147938870000168
an output hidden vector representing the t-th step of the relation edge hidden feature vector sequence; />
Figure BDA0004147938870000169
Updating hidden feature vectors representing relationship edges; />
Figure BDA00041479388700001610
Representing a parameter matrix; />
Figure BDA00041479388700001611
Representing the parameter vector.
Then will
Figure BDA00041479388700001612
GRU unit input to RNN to generate output vector of t+1 th step of relation edge sequence>
Figure BDA00041479388700001613
(i.e., updated hidden feature vectors of the relationship edges), specifically expressed by the following equation (8):
Figure BDA00041479388700001614
and step [4] determining the semantic representation vector of the meta-path instance based on the updated hidden feature vector of the child node and the updated hidden feature vector of the relation edge.
Specifically, after updating the hidden feature vector of the child node and the hidden feature vector of the relationship edge, the output vector of the last child node (i.e. the (L+1) th child node in the meta-path example) in the meta-path example is used as the semantic representation vector of the meta-path example; wherein the semantic representation vector of the meta-path instance can be represented as:
Figure BDA0004147938870000171
h p All semantic information on the meta-path instance is contained.
FIG. 3 is a schematic diagram of a meta-path instance code structure according to an embodiment of the present invention. As shown in fig. 3, the GRU models in fig. 3 are connected in a jumping manner, v represents a child node in the meta-path instance, e represents a relationship edge for connecting the child node in the meta-path instance, and the child node and the relationship edge are input into the GRU model to finally obtain a semantic representation vector of the meta-path instance.
In this embodiment, the semantic representation vectors of the multiple meta-path instances are aggregated by adopting an attention mechanism, which is specifically implemented by the following manner:
in practical application, there may be multiple meta-path instances between node u and node v, denoted as p uv They are also of different importance.
Therefore, the semantic representation vector of each element path instance needs to be aggregated through learning the attention weight, and finally a semantic fusion vector is generated. Specifically, the expression can be represented by the following formula (5):
Figure BDA0004147938870000172
wherein mu p The importance of an instance of a meta-path can be calculated by the following equation (6):
Figure BDA0004147938870000173
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA0004147938870000174
representing the meta-path instance attention parameter vector.
In the above embodiment, a meta path encoder is designed so as to uniformly mine semantic information in a meta path. The meta-path encoder comprises three modules, namely feature transformation, meta-path instance encoding and meta-path instance fusion. The feature transformation maps heterogeneous information to the same feature space, the meta-path instance coding fully utilizes the node and relation side information in the path to extract semantic features, and the meta-path instance fusion can identify the semantic importance of different instances.
Through the method, through the common attention of multiple semantics and fully utilizing the relationship side information, the rich semantic information in the heterogeneous graph is deeply mined, so that high-quality node representation can be obtained and the method is applied to downstream node classification tasks.
Optionally, the determining, based on the global attention score, the first hidden feature vector, the second hidden feature vector, and the semantic fusion vector, an updated semantic representation vector corresponding to the target node may be specifically implemented by the following steps [ a ] -step [ c ]:
step [ a ], carrying out normalization processing on the global attention score to generate a target attention score;
step [ b ], the target attention score, the second hidden feature vector and the semantic fusion vector are subjected to aggregation processing to obtain an aggregation message vector corresponding to the target node;
step [ c ], determining the updated semantic representation vector based on the aggregate message vector and the first hidden feature vector.
In this embodiment, the update semantic representation vector corresponding to the target node u may be represented as
Figure BDA0004147938870000181
Updating semantic representation vector +.>
Figure BDA0004147938870000182
Including rich semantic features in the heterograms. For calculating- >
Figure BDA0004147938870000183
First, the software is required to pass through a softmax function pair
Figure BDA0004147938870000184
Normalization processing is carried out to generate a target attention score +.>
Figure BDA0004147938870000185
Specifically, the following formula (11) can be used to realize:
Figure BDA0004147938870000186
then will
Figure BDA0004147938870000187
Second hidden feature vector h v Semantic fusion vector +.>
Figure BDA0004147938870000188
Aggregation processing is carried out to obtain an aggregation message vector +.>
Figure BDA0004147938870000189
Specifically, the following formula (12) may be used to realize:
Figure BDA00041479388700001810
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA00041479388700001811
representing a parameter matrix.
After obtaining the aggregate message vector
Figure BDA00041479388700001812
The semantic representation vector of the target node is updated by adopting a pre-activated residual error connection method to obtain an updated semantic representation vector corresponding to the target node>
Figure BDA0004147938870000191
Specifically, for the target meta-path P k The updated semantic representation vector of the target node u can be represented using the following equation (13):
Figure BDA0004147938870000192
further, a single layer of updated semantic representation vectors may not be sufficient to obtain all of the semantic features in the heterogeneous map, so a multi-layer map neural network model (Graph Neural Network, GNN) may be stacked to capture high-order semantic information.
In particular, the update semantic representation vector of the target node u may be updated recursively herein. For example, after stacking L layers of GNN models, the L+1st update semantic representation vector of the target node u
Figure BDA0004147938870000193
Can be obtained by the following formula (14):
Figure BDA0004147938870000194
/>
Wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure BDA0004147938870000195
representing a parameter matrix; />
Figure BDA0004147938870000196
Representing that target node u is in target element path P k The lower layer L aggregates message vectors; />
Figure BDA0004147938870000197
An L-layer semantic representation vector representing the target node u, and a 0-layer semantic representation vector is +.>
Figure BDA0004147938870000198
In the embodiment, the structure and the semantic information of the heterogeneous graph can be mined by updating the semantic representation vector, so that the downstream node classification performance effect is improved. The method has the advantages that redundant information can be filtered and sparse information can be enhanced through a graph learning method based on common attention, the defect that the conventional heterogeneous graph mining algorithm excessively relies on element paths to carry out semantic mining is overcome, and complex semantics can be adaptively and deeply mined, so that semantic feature representation is enhanced.
Optionally, after the target semantic representation vector corresponding to the target node is determined, training the semantic classification model by using the target semantic representation vector, so as to obtain a trained target semantic classification model, where the target semantic classification model can be applied to a downstream semantic classification task. The method can be realized by the following steps:
and step [1], inputting the target semantic representation vector into an initial target service classification model, and training the initial target service classification model by using a preset loss function until the initial target service classification model converges to obtain a trained target service classification model and a target semantic representation vector which is output by the target service classification model and is suitable for target service.
And step 2, inputting the target semantic representation vector suitable for the target service into the trained target service classification model to obtain a classification result aiming at the target service, which is output by the trained target service classification model.
Specifically, in the training stage of the initial target service classification model, the target semantic representation vector is input into the initial target service classification model, and the prediction label is obtained through calculation.
And then, calculating a loss value between the real label and the predicted label according to the loss function, and iteratively updating parameters in the neural network through a back propagation algorithm until convergence. After training is completed, a trained target service classification model and a target semantic representation vector which is output by the target service classification model and is applicable to target service can be obtained.
In the model application stage, the target semantic representation vector suitable for the target service is input into a trained target service classification model, and a classification result aiming at the target service, which is output by the trained target service classification model, can be obtained.
Taking a target service as a node classification task as an example, firstly, inputting a target semantic representation vector into an initial semantic classification model, training the initial semantic classification model by utilizing a pre-designed loss function until the initial semantic classification model converges, obtaining a trained semantic classification model, and outputting the target semantic representation vector suitable for the node classification task by the trained semantic classification model;
And inputting the target semantic representation vector suitable for the node classification task into the trained semantic classification model to obtain a semantic classification result which is output by the trained semantic classification model and aims at any node in the heterogeneous graph.
FIG. 4 is a schematic diagram showing a learning method based on a common attention heterogram; referring to fig. 4, the method includes steps 401-413, wherein:
step 401, obtaining at least one meta-path instance corresponding to a target meta-path in a heterogeneous graph; wherein the data structure of each meta-path instance is the same as the data structure of the target meta-path.
Step 402, generating a sub-node sequence and a relation edge sequence corresponding to each meta-path instance; the child node sequence comprises L+1 child nodes, the relationship edge sequence comprises L relationship edges, and L is a positive integer.
Step 403, inputting the hidden characteristic vector sequence of the child node into a first model group, and updating the hidden characteristic vector sequence of the child node to obtain an updated hidden characteristic vector of the child node; the first model group comprises a first RNN and a second RNN, and the first RNN and the second RNN are connected in a jumping manner; the sequence of child node hidden feature vectors is generated based on the sequence of child nodes.
Step 404, inputting the hidden feature vector sequence of the relationship side into a first model group, and updating the hidden feature vector sequence of the relationship side to obtain an updated hidden feature vector of the relationship side; the sequence of relational edge hidden feature vectors is generated based on the sequence of child nodes.
Step 405, generating a semantic representation vector of the meta-path instance based on the updated hidden feature vector of the child node and the updated hidden feature vector of the relationship edge.
Step 406, aggregating the semantic representation vectors of each element path instance to generate a semantic fusion vector.
Step 407, determining, for at least one target node of any target element path in the heterogeneous graph, a local attention score of a target element path neighbor node based on a preset attention parameter vector, a first hidden feature vector, a second hidden feature vector and semantic fusion vectors of all element path instances between the target node and the target element path neighbor node of the target element path.
Specifically, the local attention score is used for representing the importance degree of the target element path neighbor node to the target node based on the target element path; the first hidden feature vector is the hidden feature vector of the target node, and the second hidden feature vector is the hidden feature vector of the target element path neighbor node.
Step 408, determining a global attention score of the target element path neighbor node based on the local attention score, the number of other element paths corresponding to the target element path neighbor node in the heterogeneous graph and a preset super parameter.
Specifically, the global attention score is used for representing the importance degree of all element paths in the heterogeneous graph to the target node, and the super parameter is used for representing the importance degree of all element path examples between the target node and the target element path neighbor node to the target node.
Step 409, normalizing the global attention score to generate a target attention score.
Step 410, aggregating the target attention score, the second hidden feature vector and the semantic fusion vector to obtain an aggregate message vector corresponding to the target node.
Step 411, determining an updated semantic representation vector based on the aggregate message vector and the first hidden feature vector.
Step 412, determining the attention weight of each target element path based on the updated semantic representation vector of the node in each target element path in the heterogeneous graph, and determining the target semantic representation vector corresponding to the target node based on the attention weight of each target element path and the updated semantic representation vector of the node in each target element path.
Step 413, inputting the target semantic representation vector into an initial target service classification model, and training the initial target service classification model by using a preset loss function until the initial target service classification model converges, so as to obtain a trained target service classification model and a target semantic representation vector which is output by the target service classification model and is suitable for the target service.
Step 414, inputting the target semantic representation vector suitable for the target service into the trained target service classification model to obtain the classification result for the target service output by the trained target service classification model.
The heterogeneous graph representation learning method based on common attention aims at mining the structure and semantic information of the heterogeneous graph, and further improves the downstream node classification performance effect. On the basis of a common meta-path method, redundant information is filtered and sparse information is enhanced by a graph learning method based on common attention, and the method overcomes the defect that the conventional heterogeneous graph mining algorithm excessively relies on meta-paths to carry out semantic mining, and can carry out deep mining on complex semantics in a self-adaptive manner, so that semantic feature representation is enhanced. In addition, for the problem that the prior heterogeneous graph representation learning method based on the meta-path generally lacks to use the relationship side information, thereby causing information loss, the invention provides a meta-path encoder so as to uniformly mine semantic information. Therefore, the method and the device have the advantages that through the common attention of multiple semantics and the full utilization of the relationship side information, the rich semantic information in the heterogeneous graph is deeply mined, so that high-quality node representation is obtained, and the method and the device are applied to downstream node classification tasks.

Claims (6)

1. A method of learning a common-attention-based heterogram representation, comprising:
determining local attention scores of neighbor nodes of any target element path in a heterogeneous graph based on preset attention parameter vectors, first hidden feature vectors, second hidden feature vectors and semantic fusion vectors of all element path examples between the target node and the neighbor nodes of the target element path for at least one target node of the target element path in the heterogeneous graph; the local attention score is used for representing the importance degree of the target element path neighbor node to the target node based on the target element path; the first hidden feature vector is the hidden feature vector of the target node, and the second hidden feature vector is the hidden feature vector of the target element path neighbor node;
determining a global attention score of the target element path neighbor node based on the local attention score, the number of other element paths corresponding to the target element path neighbor node in the heterogeneous graph and a preset super parameter; the global attention score is used for representing the importance degree of all meta-path neighbor nodes in the heterogeneous graph to the target node based on all meta-paths in the heterogeneous graph, and the super-parameters are used for representing the importance degree of all meta-path examples between the target node and the target meta-path neighbor nodes to the target node;
Determining an updated semantic representation vector corresponding to the target node based on the global attention score, the first hidden feature vector, the second hidden feature vector, and the semantic fusion vector;
and determining the attention weight of each target element path based on the updated semantic representation vector of the node in each target element path in the heterogeneous graph, and determining the target semantic representation vector corresponding to the target node based on the attention weight of each target element path and the updated semantic representation vector of the node in each target element path.
2. The method of claim 1, wherein determining the updated semantic representation vector corresponding to the target node based on the global attention score, the first hidden feature vector, the second hidden feature vector, and the semantic fusion vector comprises:
normalizing the global attention score to generate a target attention score;
performing aggregation processing on the target attention score, the second hidden feature vector and the semantic fusion vector to obtain an aggregation message vector corresponding to the target node;
The update semantic representation vector is determined based on the aggregate message vector and the first hidden feature vector.
3. The method for learning a heterogram representation based on common attention according to claim 1, wherein determining the global attention score of the target meta-path neighbor based on the local attention score, the number of other meta-paths corresponding to the target meta-path neighbor in the heterogram, and a preset hyper-parameter includes:
determining a global attention score of the target node by using a formula (1) based on the local attention score, the number of other element paths corresponding to the neighbor nodes of the target element path in the heterogeneous graph and a preset super parameter; the formula (1) is:
Figure FDA0004147938850000021
wherein, the liquid crystal display device comprises a liquid crystal display device,
Figure FDA0004147938850000022
representing the global attention score; n represents the heterogeneous diagramThe number of other meta paths corresponding to the target meta path neighbor nodes; η represents the preset hyper-parameters; />
Figure FDA0004147938850000023
Representing the local attention score.
4. A co-attention based heterogram representation learning method according to any one of claims 1 to 3, wherein prior to said determining a local attention score for said target meta-path neighbor based on a pre-set attention parameter vector for said target meta-path, a first latent feature vector, a second latent feature vector, and a semantic fusion vector of all meta-path instances between said target node and said target meta-path neighbor, said method further comprises:
Acquiring at least one meta-path instance corresponding to the target meta-path, wherein the data structure of each meta-path instance is the same as the data structure of the target meta-path;
encoding each meta-path instance to generate a semantic representation vector of each meta-path instance;
and carrying out aggregation processing on the semantic representation vectors of each meta-path instance to generate the semantic fusion vector.
5. The co-attention based heterogram representation learning method of claim 4 wherein said meta-path instance comprises l+1 children nodes and L relationship edges, L being a positive integer;
encoding each meta-path instance to generate a semantic representation vector of each meta-path instance, including:
generating a child node sequence and a relation edge sequence corresponding to each meta-path instance aiming at each meta-path instance; the child node sequence comprises L+1 child nodes, and the relationship edge sequence comprises L relationship edges;
inputting a child node hidden characteristic vector sequence into a first model group, and updating the child node hidden characteristic vector sequence to obtain an updated hidden characteristic vector of the child node; the first model group comprises a first cyclic neural network model (RNN) and a second RNN, and the first RNN and the second RNN are connected in a jumping manner; the child node hidden feature vector sequence is generated based on the child node sequence;
Inputting a relation edge hidden characteristic vector sequence into the first model group, and updating the relation edge hidden characteristic vector sequence to obtain an updated hidden characteristic vector of the relation edge; the relation edge hidden feature vector sequence is generated based on the child node sequence;
and generating a semantic representation vector of the target element path instance based on the updated hidden feature vector of the child node and the updated hidden feature vector of the relation edge.
6. The co-attention based heterograph representation learning method of claim 1, wherein after said determining a target semantic representation vector for the target node, the method further comprises:
inputting the target semantic representation vector into an initial target service classification model, and training the initial target service classification model by using a preset loss function until the initial target service classification model converges to obtain a trained target service classification model and a target semantic representation vector which is output by the target service classification model and is suitable for target service;
and inputting the target semantic representation vector suitable for the target service into the trained target service classification model to obtain a classification result aiming at the target service, which is output by the trained target service classification model.
CN202310309514.4A 2023-03-27 2023-03-27 Common attention-based heterogeneous graph representation learning method Pending CN116340842A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310309514.4A CN116340842A (en) 2023-03-27 2023-03-27 Common attention-based heterogeneous graph representation learning method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310309514.4A CN116340842A (en) 2023-03-27 2023-03-27 Common attention-based heterogeneous graph representation learning method

Publications (1)

Publication Number Publication Date
CN116340842A true CN116340842A (en) 2023-06-27

Family

ID=86875830

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310309514.4A Pending CN116340842A (en) 2023-03-27 2023-03-27 Common attention-based heterogeneous graph representation learning method

Country Status (1)

Country Link
CN (1) CN116340842A (en)

Similar Documents

Publication Publication Date Title
CN105512289B (en) Image search method based on deep learning and Hash
JP6395158B2 (en) How to semantically label acquired images of a scene
CN111159425A (en) Temporal knowledge graph representation method based on historical relationship and double-graph convolution network
CN111462282A (en) Scene graph generation method
CN111008337B (en) Deep attention rumor identification method and device based on ternary characteristics
CN112200266B (en) Network training method and device based on graph structure data and node classification method
CN112217674B (en) Alarm root cause identification method based on causal network mining and graph attention network
CN111832615A (en) Sample expansion method and system based on foreground and background feature fusion
CN115908908B (en) Remote sensing image aggregation type target recognition method and device based on graph attention network
CN115544264B (en) Knowledge-driven intelligent construction method and system for digital twin scene of bridge construction
CN117237559B (en) Digital twin city-oriented three-dimensional model data intelligent analysis method and system
CN110851654A (en) Industrial equipment fault detection and classification method based on tensor data dimension reduction
CN115114421A (en) Question-answer model training method
CN115051929A (en) Network fault prediction method and device based on self-supervision target perception neural network
CN115660135A (en) Traffic flow prediction method and system based on Bayes method and graph convolution
CN114723037A (en) Heterogeneous graph neural network computing method for aggregating high-order neighbor nodes
CN109542949B (en) Formal vector-based decision information system knowledge acquisition method
CN111460277B (en) Personalized recommendation method based on mobile social network tree-shaped transmission path
CN116525029A (en) Molecular diagram generation method and device based on flow model
CN115204171A (en) Document-level event extraction method and system based on hypergraph neural network
CN116340842A (en) Common attention-based heterogeneous graph representation learning method
CN115860119A (en) Low-sample knowledge graph completion method and system based on dynamic meta-learning
CN113360772B (en) Interpretable recommendation model training method and device
CN115545833A (en) Recommendation method and system based on user social information
CN113361928B (en) Crowd-sourced task recommendation method based on heterogram attention network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination