CN115994560A - Graph neural network method based on multi-scale graph comparison learning - Google Patents

Graph neural network method based on multi-scale graph comparison learning Download PDF

Info

Publication number
CN115994560A
CN115994560A CN202310135024.7A CN202310135024A CN115994560A CN 115994560 A CN115994560 A CN 115994560A CN 202310135024 A CN202310135024 A CN 202310135024A CN 115994560 A CN115994560 A CN 115994560A
Authority
CN
China
Prior art keywords
graph
global
local
neural network
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310135024.7A
Other languages
Chinese (zh)
Inventor
王波
刘彦北
李志胜
徐振宇
国英龙
王伟
贾智洋
杨铭锴
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siteng Heli Tianjin Technology Co ltd
Original Assignee
Siteng Heli Tianjin Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siteng Heli Tianjin Technology Co ltd filed Critical Siteng Heli Tianjin Technology Co ltd
Priority to CN202310135024.7A priority Critical patent/CN115994560A/en
Publication of CN115994560A publication Critical patent/CN115994560A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a graph neural network method based on multi-scale graph comparison learning, which comprises the following steps: collecting original graph data, marking, enhancing the data, controlling the scale of a generated sub graph, further generating global views and local views with different scales, learning global information potential representations and local information potential representations of the original graph data, respectively executing different contrast learning strategies for the global information potential representations and the local information potential representations learned by the graph neural network, improving the distances of different views in a characterization space by minimizing an objective function, and improving the quality of the characterization learned by the network. The graph neural network method based on multi-scale graph comparison learning can obtain good graph level representation by learning only by using self-supervision signals extracted by data per se under the condition of not depending on data tag information, and can be widely applied to the technical fields of chemical molecular property prediction, biological protein function prediction, social network analysis and the like.

Description

Graph neural network method based on multi-scale graph comparison learning
Technical Field
The invention relates to a neural network technology, in particular to a graph neural network method based on multi-scale graph comparison learning.
Background
The graph structure data has been proved to be capable of effectively representing various data, such as social networks, financial networks, chemical molecular graphs and the like, and has wide application value. Along with the increasing popularity of graph data in real-world scenarios, learning the representation of graph data is becoming increasingly important.
Graph representation learning, i.e., extracting high-dimensional structure and attribute information from a graph and encoding it into a low-dimensional representation vector, has been widely used in many graph data analysis and processing tasks. In the information age, the increase in the amount of data has led to the increasing scarcity of marking data, resulting in the need for significant human resources and time to collect data tags. As a result, more and more researchers are beginning to focus on unsupervised or self-supervised graph representation learning.
Conventional methods such as matrix decomposition and random walk cannot be extended to larger-scale graphs and cannot make good use of the attribute information of the nodes, so learning the representation of the graph without relying on the label information of the graph becomes critical.
At present, graphs based on contrast learning represent a feasible method, and many related researches have also achieved promising results. However, most studies do not take into account multi-scale information of the graph data, i.e., contrast learning from a single scale, ignoring more global information or finer local information.
Disclosure of Invention
In order to solve the problems, the invention provides a graph neural network method based on multi-scale graph contrast learning, which overcomes the defect of multi-scale information in the prior method, proposes concepts of global views and local views by controlling the number of random walk nodes, and simultaneously formulates different contrast learning strategies, so that different strategies are executed among different views, rich multi-scale information in graph data can be mined, and different contrast learning strategies are executed, thereby effectively improving the quality of learned graph representation.
In order to achieve the above purpose, the present invention provides a graph neural network method based on multi-scale graph contrast learning, comprising the following steps:
s1, collecting original image data, and marking the collected original image data;
s2, carrying out data enhancement on the collected original image data in a random walk mode, controlling the scale of a generated sub image, and further generating global views and local views with different scales;
s3, learning global information potential representation and local information potential representation of the original graph data through a graph neural network;
s4, respectively executing different comparison learning strategies aiming at the global information potential representation and the local information potential representation learned by the graph neural network;
s5, by minimizing an objective function, the distances of different views in the characterization space are improved, and the quality of the characterization learned by the network is improved.
Preferably, in step S1, the original graph structure data collected from the network is marked as a corresponding tag file according to different data types.
Preferably, in step S1, a graph dataset D is defined, which dataset is made up of N graph data
Figure BDA0004085142630000021
Composition, g= [ V, E }, for each graph in the dataset, where +.>
Figure BDA0004085142630000022
Representing a set of nodes in the graph, E representing a set of edges in the graph, v i Representing the ith node in the graph, if node v i And v j And there is a correlation between them, then e ij =(v i ,v j ) E is the edge in the graph;
thus, the graph data has an adjacency matrix A with dimensions V x V, if e ij E is A ij =1;
In addition, each node of the graph data also has a feature vector x i ∈R d All node features in the graph form a feature matrix X epsilon R |V|×d
Figure BDA0004085142630000023
Is the label to which the graph data corresponds.
Preferably, in step S2, a G ε { G n : n.epsilon.N }, define an enhancement graph for the graph
Figure BDA0004085142630000024
Wherein->
Figure BDA0004085142630000025
Enhancing the method for the used graph; the enhancement method mainly comprises sub-sampling, and controlling the scale of the generated sub-image by controlling the number of nodes, so as to obtain a global view and a local view with different scale information, namely +.>
Figure BDA0004085142630000031
And
Figure BDA0004085142630000032
preferably, in step S3, the potential representation of the view is learned by using a five-layer map isomorphic network for sharing model parameters between the global view and the local view.
Preferably, in step S4, a regression metric is used for selecting the noise contrast estimation function between global views and between global and local views, and for selecting between local views, the regressor is implemented as a multi-layer perceptron with batch normalization and RELU activation functions.
Preferably, in step S4, the graph neural network mainly follows a message passing mechanism, and each node acquires attribute and structure information of a neighboring node through the message passing mechanism, so as to update its own node representation, and through k-layer iteration, the node captures information of its k-hop neighbor, namely:
Figure BDA0004085142630000033
Figure BDA0004085142630000034
AGGREGATE (-) and COMBINE (-) are respectively information of aggregation neighborhood nodes and update own node information; the node potential representation of the global and local views after data enhancement is obtained through a graph isomorphic neural network;
the potential representation of the entire graph is then obtained by pooling, i.e.:
Figure BDA0004085142630000035
READOUT (·) is graph pooling, a sum pooling mechanism;
finally, the representation of the acquired image level is obtained through a nonlinear transformation, namely:
z=g(f(G))
g (·) is a nonlinear transformation, which is a two-layer perceptron with a RELU activation function;
thereby obtaining global and local representations of the graph data, i.e
Figure BDA0004085142630000036
And->
Figure BDA0004085142630000037
Preferably, in step S5, the neural network is optimized by assigning different weight coefficients to three different loss items, so that it can learn good graph level characterization, and thus be used for different downstream tasks.
Preferably, the step S5 specifically includes the following steps:
s51, simultaneously considering global representation and local representation, implementing different contrast learning strategies on the global representation and the local representation, further considering different scale information to improve performance, and defining noise contrast estimation loss
Figure BDA0004085142630000041
Is l s The method comprises the following steps:
Figure BDA0004085142630000042
τ is the temperature coefficient, z + And z - Positive and negative samples, respectively;
s52, considering maximizing the global representation of the same original graph, minimizing the similarity of the global representations of different original graphs, and losing the function l gg Is defined as follows:
Figure BDA0004085142630000043
wherein N is the number of samples in the batch;
s53, in order to establish local and global association, defining a loss function, namely:
Figure BDA0004085142630000044
s54, providing local and global similarity by a measurement mode with a learnable parameter, wherein the local and global similarity is a five-layer perceptron f with RELU activation function θ It is desirable that the local views from the same graph have a higher similarity than the local views from different graphs, and therefore a loss function ψ is defined, namely:
Figure BDA0004085142630000045
s55, by maximizing ψ (θ d ) Training the regressor and regarding its output as its similarity measure by Sigmoid activation function, and training
Figure BDA0004085142630000046
Considered as an estimator i for estimating local view similarity d Thus, for between partial views, the partial view loss function is defined as:
Figure BDA0004085142630000047
s56, defining an overall loss function as follows:
Figure BDA0004085142630000048
s57, minimizing total loss by means of Adam gradient descent algorithm
Figure BDA0004085142630000049
To update the encoder parameters and use the pre-trained encoder for different downstream tasks.
Compared with the prior art, the invention has the following beneficial effects:
1. the multi-scale information in graph comparison learning is considered, different comparison learning strategies are executed for different scale information, the learned characterization quality is better, and the method can be better used for a series of downstream tasks.
2. The method can be used for scenes such as graph data analysis and graph representation learning, and can assist people in better analyzing and using the graph data.
The technical scheme of the invention is further described in detail through the drawings and the embodiments.
Drawings
FIG. 1 is a schematic diagram of a frame of the present invention;
FIG. 2 is a schematic of an algorithm of the present invention;
FIG. 3 is a regressor design of the present invention.
Detailed Description
The present invention will be further described with reference to the accompanying drawings, and it should be noted that, while the present embodiment provides a detailed implementation and a specific operation process on the premise of the present technical solution, the protection scope of the present invention is not limited to the present embodiment.
FIG. 1 is a schematic diagram of a frame of the present invention; FIG. 2 is a schematic of an algorithm of the present invention; fig. 3 is a regressor design diagram of the present invention, as shown in fig. 1-3, a graph neural network method based on multi-scale graph contrast learning, comprising the steps of:
s1, collecting original image data, and marking the collected original image data;
preferably, in step S1, the original graph structure data collected from the network is marked as a corresponding tag file according to different data types.
Preferably, in step S1, a graph dataset D is defined, which dataset is made up of N graph data
Figure BDA0004085142630000051
Composition, g= { V, E }, for each graph in the dataset, where +.>
Figure BDA0004085142630000052
Representing a set of nodes in the graph, E representing a set of edges in the graph, v i Representative graphIf node v i And v j And there is a correlation between them, then e ij =(v i ,v j ) E is the edge in the graph;
thus, the graph data has an adjacency matrix A with dimensions V x V, if e ij E is A ij =1;
In addition, each node of the graph data also has a feature vector x i ∈R d All node features in the graph form a feature matrix X epsilon R |V|×d
Figure BDA0004085142630000061
Is the label to which the graph data corresponds.
S2, graph data enhancement is important to the contrast learning employed herein, and the lack of data enhancement results in a model that is less effective than an untrained model. The data enhancement aims at converting the data to a certain degree under the condition that the semantic information of the original data is not affected as much as possible, and creating some novel data.
In this embodiment, the collected original graph data is enhanced by a random walk mode, and the scale of the generated subgraph is controlled, so that global views and local views with different scales are generated;
preferably, in step S2, a G ε { G n : n.epsilon.N }, define an enhancement graph for the graph
Figure BDA0004085142630000062
Wherein->
Figure BDA0004085142630000063
Enhancing the method for the used graph; the enhancement method mainly comprises sub-sampling, and controlling the scale of the generated sub-image by controlling the number of nodes, so as to obtain a global view and a local view with different scale information, namely +.>
Figure BDA0004085142630000064
And
Figure BDA0004085142630000065
s3, learning global information potential representation and local information potential representation of original image data through the image neural network because the image neural network is an image data analysis method with strong expression capability;
preferably, in step S3, the potential representation of the view is learned by using a five-layer map isomorphic network for sharing model parameters between the global view and the local view.
S4, respectively executing different comparison learning strategies aiming at the global information potential representation and the local information potential representation learned by the graph neural network;
it should be noted that, in the contrast learning stage, since the global view contains most of the information of the graph data and has high semantic similarity, it is desirable to reduce the distance between them in the characterization space; due to the large size of global views, which largely contain the content of local views, it is also desirable to reduce their distance in the characterization space; however, for partial views, the semantic similarity is low because the descriptions are always different, and instead, the distances of the partial views in the characterization space need to be pulled, namely the dissimilarity of the partial views is encouraged.
Preferably, in step S4, a regression metric is used for selecting the noise contrast estimation function between global views and between global and local views, and for selecting between local views, the regressor is implemented as a multi-layer perceptron with batch normalization and RELU activation functions.
Preferably, in step S4, the graph neural network mainly follows a message passing mechanism, and each node acquires attribute and structure information of a neighboring node through the message passing mechanism, so as to update its own node representation, and through k-layer iteration, the node captures information of its k-hop neighbor, namely:
Figure BDA0004085142630000071
Figure BDA0004085142630000072
AGGREGATE (-) and COMBINE (-) are respectively information of aggregation neighborhood nodes and update own node information; the node potential representation of the global and local views after data enhancement is obtained through a graph isomorphic neural network;
the potential representation of the entire graph is then obtained by pooling, i.e.:
Figure BDA0004085142630000073
READOUT (·) is graph pooling, a sum pooling mechanism;
finally, the representation of the acquired image level is obtained through a nonlinear transformation, namely:
z=g(f(G))
g (·) is a nonlinear transformation, which is a two-layer perceptron with a RELU activation function;
thereby obtaining global and local representations of the graph data, i.e
Figure BDA0004085142630000074
And->
Figure BDA0004085142630000075
S5, by minimizing an objective function, the distances of different views in the characterization space are improved, and the quality of the characterization learned by the network is improved.
Preferably, in step S5, the neural network is optimized by assigning different weight coefficients to three different loss items, so that it can learn good graph level characterization, and thus be used for different downstream tasks.
Preferably, the step S5 specifically includes the following steps:
s51, simultaneously considering global representation and local representation, implementing different contrast learning strategies on the global representation and the local representation, further considering different scale information to improve performance, and defining noise contrast estimation lossLoss of function
Figure BDA0004085142630000081
Is l s The method comprises the following steps:
Figure BDA0004085142630000082
τ is the temperature coefficient, z + And z - Positive and negative samples, respectively;
s52, since the global view generally contains most of the contents of the graph, the global view has very similar semantic information to the previous one, so that the global representation of the same original graph is considered to be maximized, the similarity of the global representations of different original graphs is minimized, and the loss function l gg Is defined as follows:
Figure BDA0004085142630000083
wherein N is the number of samples in the batch;
s53, because the global view has a larger size and contains the content of the local view to a large extent, a part of semantic content can be shared between the global view and the content, so that in order to establish local and global association, a loss function is defined, namely:
Figure BDA0004085142630000084
because the partial views from the same original graph generally describe different content, the semantic similarity between them is low, so that their potential representation distance is no longer reduced as before, but their dissimilarity is encouraged. However, given that there is still some semantic similarity between them, their distance is no longer directly pulled away by the noise contrast estimation penalty, so local and global similarity is given by a metric with a learnable parameter, which is a five-layer with RELU activation functionPerceptron f θ It is desirable that the local views from the same graph have a higher similarity than the local views from different graphs, and therefore a loss function ψ is defined, namely:
Figure BDA0004085142630000085
s55, by maximizing ψ (θ d ) Training the regressor and regarding its output as its similarity measure by Sigmoid activation function, and training
Figure BDA0004085142630000086
Considered as an estimator i for estimating local view similarity d Thus, for between partial views, the partial view loss function is defined as:
Figure BDA0004085142630000091
s56, defining an overall loss function as follows:
Figure BDA0004085142630000092
s57, minimizing total loss by means of Adam gradient descent algorithm
Figure BDA0004085142630000093
To update the encoder parameters and use the pre-trained encoder for different downstream tasks.
Therefore, the graph neural network method based on the multi-scale graph comparison learning can obtain good graph level representation by learning only by using the self-supervision signals extracted by the data under the condition of not depending on the data tag information, and can be widely applied to the technical fields of chemical molecular property prediction, biological protein function prediction, social network analysis and the like.
Experimental example
Table 1 is a table of statistics for graph number, node number, and edge number for a dataset
Datasets Category Graph Node Edges
MUTAG Molecules 188 17.93 19.79
NCI1 Molecules 4110 29.87 32.30
PROTEINS Molecules 1113 39.06 72.82
DD Molecules 1178 284.32 715.66
IMDB-B Social Network 1000 19.77 95.63
COLLAB Social Network 5000 74.49 2457.78
RDT-B Social Network 2000 429.63 497.75
RDT-M5K Social Network 5000 508.52 594.87
As can be seen from table 1, the TUDataset series data set is referenced for the experiment, including four chemical molecule data sets and four social network data sets. Where NCI1 is a compound dataset provided by the united states cancer institute (NCI), comprising 4100 graph samples, node feature dimension 37.DD is a data set containing 1178 protein structures, each protein represented by a graph, each node representing an amino acid. The MUTAG dataset consisted of 188 compounds, two classes were categorized according to their mutagenic effect on bacteria. The graph data nodes in the PROTEINS are two-level element structures, and if two nodes are adjacent in amino acid sequence or three-dimensional space, there is an edge between them. Each graph in the REDDIT-BINARY dataset corresponds to an online discussion thread, wherein nodes correspond to users, and edges exist when mutual responses exist between the two nodes. REDDIT-M5K expands the data collection range to five different sub communities based on REDDIT-BINARY. COLLAB is a scientifically collaborative dataset, data from 3 areas, namely high energy physics, condensed physics and celestial physics. IMDB-BINARY is a collaborative dataset of movies, with nodes representing actors for each graph, there being one edge between two actors if they appear in the same movie. Since there is no official division of the data set described above, a random division is employed and 10-fold cross-validation is performed.
The method of the invention is compared with the following method:
WL: combining the WL algorithm with the graph kernel, providing a WL graph kernel method, decomposing the graph into subtrees, and measuring the subtree similarity to obtain the graph similarity;
DGK: decomposing the graph into substructures, and measuring graph similarity through the substructures;
node2vec: depth-first and breadth-first algorithms are considered based on variants of the deep walk algorithm;
sub2vec: learning the sub-graph feature representation by performing a random walk method of sub-graph truncation;
graph2vec: the doc2vec in natural language processing is expanded into graph data, and unsupervised representation learning is performed by generating rooted subgraphs and negative samples.
GAE: the encoder is trained by encoding the graph into a potential feature representation and then restoring it to the original graph.
DGI: the concept of mutual information maximization is introduced into the graph data, so that the mutual information maximization of the nodes and the graph is realized.
ContexPred: the encoder model is trained by maximizing the similarity of the representation of the center node to the representation of the context nodes.
InfoGraph: similar to DGI, the idea is maximized based on mutual information as well, but it only focuses on the representation learning at the graph level.
GraphCL: four different data enhancement methods are proposed and contrast learning is used to optimize the consistency of the enhanced view.
JOAO: the data enhancement scheme can be selected adaptively and dynamically on the basis of the GraphCL.
SimgRACE: disruption of semantic information by data enhancement is avoided by directly perturbing the encoder.
In the experimental example, two tasks are performed to verify the effectiveness of the proposed method, and the two tasks are classified into an unsupervised classification task and a semi-supervised classification task. For unsupervised classification, the encoder is pre-trained using proposed multi-scale contrast learning, and then the representation of the output of the encoder after training is fed into a downstream linear SVM classifier. For semi-supervised classification, a multi-layer perceptron that delivers a representation downstream on an unsupervised basis is used to classify and give a proportion of label information to fine tune the model. Acc was used as an evaluation index in both tasks.
Table 2 is an unsupervised classification experiment result table
Figure BDA0004085142630000111
Figure BDA0004085142630000121
Table 2 shows the performance of the method of the present invention in eight downstream task data sets, and as can be seen from Table 2, the classification performance of the proposed Method (MSSGCL) is best, and is superior to all other baseline models, and the average Acc improvement achieved by the present invention can reach more than 2% compared with GraphCL encouraging small-size view similarity.
Table 3 shows the results of semi-supervised classification experiments
Figure BDA0004085142630000122
Figure BDA0004085142630000131
Table 3 shows the performance of the inventive method for eight data sets under semi-supervised settings, reporting subtasks with label rates of 1% and 10% as seen in table 3. At a label rate of 1% set, the proposed method is higher than all baseline models, even with improvement compared to the previous best model SimGRACE. When the label rate is 10%, the method is superior to the previous limit model, the optimal performance is realized in 6 of 7 data sets, and compared with GraphCL, the average Acc improvement is 2%.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention and not for limiting it, and although the present invention has been described in detail with reference to the preferred embodiments, it will be understood by those skilled in the art that: the technical scheme of the invention can be modified or replaced by the same, and the modified technical scheme cannot deviate from the spirit and scope of the technical scheme of the invention.

Claims (9)

1. A graph neural network method based on multi-scale graph contrast learning is characterized in that: the method comprises the following steps:
s1, collecting original image data, and marking the collected original image data;
s2, carrying out data enhancement on the collected original image data in a random walk mode, controlling the scale of a generated sub image, and further generating global views and local views with different scales;
s3, learning global information potential representation and local information potential representation of the original graph data through a graph neural network;
s4, respectively executing different comparison learning strategies aiming at the global information potential representation and the local information potential representation learned by the graph neural network;
s5, by minimizing an objective function, the distances of different views in the characterization space are improved, and the quality of the characterization learned by the network is improved.
2. The graph neural network method based on multi-scale graph contrast learning of claim 1, wherein the method comprises the following steps: in step S1, the original graph structure data collected from the network is marked as a corresponding tag file according to different data types.
3. The graph neural network method based on multi-scale graph contrast learning according to claim 2, wherein the method is characterized in that: in step S1, a graph dataset D is defined, which dataset is composed of N graph data
Figure FDA0004085142620000011
Composition, g= { V, E }, for each graph in the dataset, where +.>
Figure FDA0004085142620000012
Representing a set of nodes in the graph, E representing a set of edges in the graph, v i Representing the ith node in the graph, if node v i And v j And there is a correlation between them, then e ij =(v i ,v j ) E is the edge in the graph;
thus, the graph data has an adjacency matrix A with dimensions V x V, if e ij E is A ij =1;
In addition, each node of the graph data also has a feature vector x i ∈R d All node features in the graph form a feature matrix X epsilon R |V|×d ,y k E y is the label to which the graph data corresponds.
4. The graph neural network method based on multi-scale graph contrast learning of claim 1, wherein the method comprises the following steps: in step S2, a G ε { G }, is given n : n.epsilon.N }, define an enhancement graph for the graph
Figure FDA0004085142620000013
Wherein->
Figure FDA0004085142620000014
Enhancing the method for the used graph; the enhancement method mainly comprises sub-sampling, and controlling the scale of the generated sub-image by controlling the number of nodes, so as to obtain a global view and a local view with different scale information, namely +.>
Figure FDA0004085142620000021
And->
Figure FDA0004085142620000022
5. The graph neural network method based on multi-scale graph contrast learning of claim 1, wherein the method comprises the following steps: in step S3, the potential representation of the view is learned by using a five-layer map isomorphic network for sharing model parameters between the global view and the local view.
6. The graph neural network method based on multi-scale graph contrast learning of claim 1, wherein the method comprises the following steps: in step S4, a regressor metric is used for selecting the noise contrast estimation function between the global views and between the global and local views, and for selecting between the local views, the regressor is implemented as a multi-layer perceptron with batch normalization and RELU activation functions.
7. The graph neural network method based on multi-scale graph contrast learning of claim 6, wherein the method comprises the following steps: in step S4, the graph neural network mainly follows a message passing mechanism, and each node acquires attribute and structure information of a neighboring node through the message passing mechanism, so as to update its own node representation, and through k-layer iteration, the node captures information of its k-hop neighbor, namely:
Figure FDA0004085142620000023
Figure FDA0004085142620000024
AGGREGATE (-) and COMBINE (-) are respectively information of aggregation neighborhood nodes and update own node information; the node potential representation of the global and local views after data enhancement is obtained through a graph isomorphic neural network;
the potential representation of the entire graph is then obtained by pooling, i.e.:
Figure FDA0004085142620000025
READOUT (·) is graph pooling, a sum pooling mechanism;
finally, the representation of the acquired image level is obtained through a nonlinear transformation, namely:
z=g(f(G))
g (·) is a nonlinear transformation, which is a two-layer perceptron with a RELU activation function;
thereby obtaining global and local representations of the graph data, i.e
Figure FDA0004085142620000026
And->
Figure FDA0004085142620000027
8. The graph neural network method based on multi-scale graph contrast learning of claim 1, wherein the method comprises the following steps: in step S5, the neural network is optimized by assigning different weight coefficients to three different loss items, so that it can learn good graph level representation, and then be used for different downstream tasks.
9. The graph neural network method based on multi-scale graph contrast learning of claim 8, wherein the method comprises the following steps: the step S5 specifically comprises the following steps:
s51, simultaneously considering global representation and local representation, implementing different contrast learning strategies on the global representation and the local representation, further considering different scale information to improve performance, and defining noise contrast estimation loss
Figure FDA0004085142620000031
Is l s The method comprises the following steps:
Figure FDA0004085142620000032
τ is the temperature coefficient, z + And z - Positive and negative samples, respectively;
s52, considering maximizing the global representation of the same original graph, minimizing the similarity of the global representations of different original graphs, and losing the function l gg Is defined as follows:
Figure FDA0004085142620000033
wherein N is the number of samples in the batch;
s53, in order to establish local and global association, defining a loss function, namely:
Figure FDA0004085142620000034
s54, providing local and global similarity by a measurement mode with a learnable parameter, wherein the local and global similarity is a five-layer perceptron f with RELU activation function θ It is desirable that the local views from the same graph have a higher similarity than the local views from different graphs, and therefore a loss function ψ is defined, namely:
Figure FDA0004085142620000035
s55, by maximizing ψ (θ d ) Training the regressor and regarding its output as its similarity measure by Sigmoid activation function, and training
Figure FDA0004085142620000036
Considered as an estimator i for estimating local view similarity d Thus, for between partial views, the partial view loss function is defined as:
Figure FDA0004085142620000041
s56, defining an overall loss function as follows:
Figure FDA0004085142620000042
s57, minimizing total loss by means of Adam gradient descent algorithm
Figure FDA0004085142620000043
To update the encoder parameters and use the pre-trained encoder for different downstream tasks. />
CN202310135024.7A 2023-02-20 2023-02-20 Graph neural network method based on multi-scale graph comparison learning Pending CN115994560A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310135024.7A CN115994560A (en) 2023-02-20 2023-02-20 Graph neural network method based on multi-scale graph comparison learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310135024.7A CN115994560A (en) 2023-02-20 2023-02-20 Graph neural network method based on multi-scale graph comparison learning

Publications (1)

Publication Number Publication Date
CN115994560A true CN115994560A (en) 2023-04-21

Family

ID=85995142

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310135024.7A Pending CN115994560A (en) 2023-02-20 2023-02-20 Graph neural network method based on multi-scale graph comparison learning

Country Status (1)

Country Link
CN (1) CN115994560A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117648623A (en) * 2023-11-24 2024-03-05 成都理工大学 Network classification algorithm based on pooling comparison learning

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117648623A (en) * 2023-11-24 2024-03-05 成都理工大学 Network classification algorithm based on pooling comparison learning

Similar Documents

Publication Publication Date Title
Zhang et al. Consensus one-step multi-view subspace clustering
Wen et al. Low-rank representation with adaptive graph regularization
Shao et al. Multiple incomplete views clustering via weighted nonnegative matrix factorization with regularization
Lee et al. Wasserstein introspective neural networks
CN109753589A (en) A kind of figure method for visualizing based on figure convolutional network
Fatras et al. Minibatch optimal transport distances; analysis and applications
Zhang et al. Non-negative multi-label feature selection with dynamic graph constraints
Li et al. Multi-view graph learning with adaptive label propagation
Fang et al. Confident learning-based domain adaptation for hyperspectral image classification
Zhang et al. Deep learning of path-based tree classifiers for large-scale plant species identification
CN115994560A (en) Graph neural network method based on multi-scale graph comparison learning
Xie et al. TPNE: topology preserving network embedding
CN115983351A (en) Self-supervision graph neural network method based on comparative learning
Chen et al. Variational Graph Embedding and Clustering with Laplacian Eigenmaps.
Liu et al. Self-supervised transformer-based pre-training method using latent semantic masking auto-encoder for pest and disease classification
Nong et al. Hypergraph wavelet neural networks for 3D object classification
Xu et al. Semi-supervised self-growing generative adversarial networks for image recognition
Peng et al. JGSED: An End-to-End Spectral Clustering Model for Joint Graph Construction, Spectral Embedding and Discretization
Ahmadi et al. Deep graph clustering via mutual information maximization and mixture model
CN115392474B (en) Local perception graph representation learning method based on iterative optimization
Wei et al. Sparse representation classifier guided Grassmann reconstruction metric learning with applications to image set analysis
Cen et al. Anae: Learning node context representation for attributed network embedding
CN111126443A (en) Network representation learning method based on random walk
Dhanalakshmi et al. Tomato leaf disease identification by modified inception based sequential convolution neural networks
Kumar et al. A comparative study of machine learning algorithms for image recognition in privacy protection and crime detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination