WO2020114122A1 - Neural network system and method for analyzing relationship network graph - Google Patents

Neural network system and method for analyzing relationship network graph Download PDF

Info

Publication number
WO2020114122A1
WO2020114122A1 PCT/CN2019/112564 CN2019112564W WO2020114122A1 WO 2020114122 A1 WO2020114122 A1 WO 2020114122A1 CN 2019112564 W CN2019112564 W CN 2019112564W WO 2020114122 A1 WO2020114122 A1 WO 2020114122A1
Authority
WO
WIPO (PCT)
Prior art keywords
output
graph
neural network
weighting factor
network
Prior art date
Application number
PCT/CN2019/112564
Other languages
French (fr)
Chinese (zh)
Inventor
常晓夫
宋乐
Original Assignee
阿里巴巴集团控股有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴集团控股有限公司 filed Critical 阿里巴巴集团控股有限公司
Publication of WO2020114122A1 publication Critical patent/WO2020114122A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/288Entity relationship models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Definitions

  • One or more embodiments of this specification relate to a neural network system executed by a computer, and particularly to a neural network system and method for analyzing a relational network graph.
  • the graph is a powerful tool for modeling relational data. Therefore, at present, the data in which the relationship exists is often expressed and modeled in the form of a graph.
  • a graph-based neural network using deep learning methods Graph Neural Network (Graph Neural Network, Graph NN or GNN) is proposed to learn graph information.
  • the graph neural network GNN can effectively use the information transfer on the graph and fuse the feature information of the nodes or edges to complete the machine learning tasks such as the classification or regression of the nodes or edges on the graph.
  • One or more embodiments of this specification describe a neural network system and method executed by a computer for analyzing a relationship network graph, which can learn, analyze, and predict the relationship network graph more effectively.
  • a neural network system for analyzing a relationship network graph executed by a computer including:
  • the feature extraction layer is used to extract the feature vectors of the nodes in the relationship network graph
  • a deep neural network which is used to perform first processing on the feature vector to obtain a first output
  • a graph neural network used to combine the adjacency information of the relational network graph to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent each of the nodes included in the relational network graph Connection between
  • a fusion layer is used to fuse the first output and the second output, and output a prediction result for the node based on the fusion result.
  • each node included in the relationship network diagram corresponds to each user, and the connection relationship between the various nodes includes one or more of the following: social relationship, media relationship, and fund relationship between users .
  • the relationship network graph is a directed graph
  • the adjacency information includes an adjacency list or a cross-linked list corresponding to the directed graph.
  • the adjacency information includes the adjacency matrix of the relationship network graph.
  • the graph neural network is a graph convolutional network
  • the graph convolutional network includes multiple network layers to perform the second processing, and the second processing includes at least using the adjacency matrix.
  • the element is a weighting factor, which performs a weighted sum operation on the feature vectors of the node and its neighbors.
  • the above fusion layer is specifically used for weighted summation of the first output and the second output, where the first output corresponds to a first weighting factor and the second output corresponds to a second Weighting factor.
  • the first weighting factor is a function of the first output
  • the second weighting factor is a function of the second output
  • the sum of the first weighting factor and the second weighting factor is 1, and:
  • the first weighting factor is a function of the first output
  • the second weighting factor is a function of the second output.
  • the sum of the first weighting factor and the second weighting factor is 1, and:
  • the first weighting factor is a function of the first output and the second output; or,
  • the second weighting factor is a function of the first output and the second output.
  • the neural network system is trained in an end-to-end manner.
  • a computer-implemented method for analyzing a relationship network graph including:
  • a deep neural network is used to perform the first processing on the feature vector to obtain a first output
  • a graph neural network is used, combined with the adjacency information of the relational network graph, to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent the relationship between each node included in the relational network graph Connection
  • the first output and the second output are fused, and a prediction result for the node is output based on the fusion result.
  • a computer-readable storage medium on which a computer program is stored, and when the computer program is executed in a computer, the computer is caused to perform the method of the second aspect.
  • a computing device including a memory and a processor, wherein the memory stores executable code, and when the processor executes the executable code, the method of the second aspect is implemented .
  • a deep neural network DNN and a graph neural network GNN are combined, and the single-node feature processing capability of DNN and the relational feature processing capability of GNN are fused to make the combined neural network system Can effectively analyze and learn a variety of relationship network diagrams.
  • the graph neural network GNN can play a major role, and the deep neural network DNN supplements the single node analysis; and if the relationship features are missing or the effect is limited, the deep neural network can still be used
  • the branch of DNN effectively analyzes and processes the nodes in the graph to give a more ideal prediction result.
  • FIG. 1 is a schematic diagram of a relationship network diagram disclosed in this specification
  • FIG. 2 shows a schematic diagram of a neural network system according to an embodiment
  • FIG. 3 shows a schematic diagram of a deep neural network DNN according to an embodiment
  • FIG. 4 shows a schematic diagram of a graph convolutional network GCN according to an embodiment
  • FIG. 5 shows a flowchart of a method for analyzing a relationship network diagram according to an embodiment.
  • a neural network system for processing relational data is proposed.
  • the neural network system can be used for learning and predicting relational network graphs.
  • FIG. 1 is a schematic diagram of a relationship network diagram disclosed in this specification. It can be seen that the relational network graph includes multiple nodes, and the nodes with associated relationships are connected to each other by connecting edges. Nodes that are not associated with other nodes form isolated nodes, such as nodes A, B, and C in Figure 1.
  • the type and/or strength of the association relationship can also be distinguished, so as to assign certain attributes or weights to the connected edges. For example, in FIG. 1, thick lines indicate strong connections, and thin lines indicate weak connections. However, this is not necessary.
  • the relationship network graph can reflect the association relationship between various entities.
  • the nodes may represent sellers or buyers, and the edges between the nodes may represent that a transaction has occurred, thereby reflecting the transaction relationship between entities through a relationship network graph.
  • the nodes represent individual users, and the edges between the nodes represent the association between users. More specifically, in different examples, a connection relationship may be established for nodes based on different types of association relationships between users.
  • the association relationship between users may include a social relationship between users.
  • a social relationship between users In a relationship network formed based on social relationships, if two users have common followers (such as Weibo accounts following the same person together), or they have previous contacts, or join a common group (such as QQ group, WeChat group) Etc.), or interactions in activities such as red envelopes and lottery tickets, then it can be considered that there is a social relationship between the two nodes, and an edge can be established to connect.
  • the association relationship between users may include a media relationship between users.
  • a media relationship between users In the relationship network formed based on the media relationship, if two users have used the same media, such as encrypted bank card, ID card, mailbox, account number, mobile phone number, physical address (such as MAC address), terminal device number ( For example, UMID, TID, UTDID, etc., then there is a media relationship between these two users, you can establish an edge to connect.
  • the association relationship between users may include the financial relationship between users.
  • Fund transactions can include charge, charge, bar code collection, bar code payment, AA collection, C2C mobile phone face-to-face payment, gift money, rent, red envelope, credit card repayment, purchase, intimate payment, subscription service, etc.
  • the nodes in the relationship network graph may also represent other entities, and the connections between the nodes may be based on various types of association relationships.
  • the graph neural network GNN can generally be used for learning and prediction.
  • the learning process generally corresponds to the training process of graph neural network GNN.
  • the graph neural network GNN When training the graph neural network GNN, it is necessary to add labels to at least some nodes in FIG. 1 according to the purpose of the prediction business, so as to perform supervised learning. For example, assuming that the nodes in the relational network graph in FIG. 1 correspond to individual users, if you want to learn and predict the credit risk of each user, at least some nodes need to add "high risk users” (users suspected of fraud) and "normal users” ", provide these labeled nodes, together with the connection relationship of these nodes in the graph, to the graph neural network GNN for GNN to train and learn. After the training is completed, the unknown user can be input to the graph neural network GNN, so that the graph neural network GNN uses the trained network parameters to predict the unknown user.
  • the relationship network graph is constructed based on the association relationship.
  • association relationship there are various association relationships between nodes. If the relationship between the selected association relationship and the business purpose is not high enough when constructing the relationship network graph, then the relationship network graph constructed based on such association relationship Learning, the results are often not ideal. For example, in the case where the business purpose is to learn the user’s purchase intention of a certain type of product, if the association relationship is selected as whether it has a social relationship with a common object of interest, such relationship data may not be effective for determining the business purpose of the purchase intention .
  • a combined neural network system which can learn the relationship network graph more effectively.
  • FIG. 2 shows a schematic diagram of a neural network system according to an embodiment.
  • the neural network system is executed by a computer and is used to learn a relationship network graph and process relationship data.
  • the neural network system includes a feature extraction layer 21 for extracting feature vectors of nodes in the relational network graph; a deep neural network 22 for performing first processing on the feature vectors to obtain An output; graph neural network 23, used to combine the adjacency information of the relational network graph, and perform a second processing on the feature vector to obtain a second output; a fusion layer 24, used for the first output and the The second output is fused, and a prediction result for the node is output based on the fusion result.
  • the following specifically describes the implementation of each of the above network parts.
  • the feature extraction layer 21 is used to extract feature vectors of nodes in the relationship network graph.
  • the relationship network diagram is, for example, the relationship network diagram shown in FIG. 1, which includes multiple nodes, and there is a connection relationship between nodes having an association relationship.
  • the feature extraction layer 21 performs feature extraction on the nodes in the relational network graph, and the extracted features are the features of the relevant nodes themselves, so the extracted features constitute a feature vector.
  • each node in the above relational network diagram corresponds to each user, for example, Alipay user.
  • the feature extraction layer 21 may extract the user's basic attribute features to form a feature vector.
  • the basic attribute features include, for example, user ID, registration duration, gender, age, and so on.
  • the feature extraction layer 21 also extracts features related to the business purpose according to the business purpose. For example, in the case where the business purpose is to predict the user's purchase intention, the feature extraction layer 21 also obtains the user's purchase records, and performs feature extraction based on the purchase records, such as extracting multiple features such as the number of purchases, purchase categories, and purchase amount. For another example, when the business purpose is to predict the borrowing risk of the user, the feature extraction layer 21 also obtains the borrowing record of the user, and extracts the features based on the borrowing record, such as extracting the number of loans, the amount of loan, the amount of repayment, and the number of trustworthiness , Overdue times and other characteristics.
  • the feature extraction layer 21 extracts the feature vectors of the nodes, the feature vectors are input to the deep neural network 22 and the graph neural network 23 in parallel for processing.
  • Deep Neural Network is a multi-layer fully connected forward-structured artificial neural network.
  • Figure 3 shows a schematic diagram of a deep neural network according to one embodiment.
  • the deep neural network DNN contains multiple network layers. These network layers can be divided into an input layer, a hidden layer, and an output layer. Between the input layer and the output layer are hidden layers. Generally, when there are many hidden layers in a neural network, we call it a deep neural network.
  • Each network layer in the DNN contains several neurons, and all neurons except the input layer operate on the input data through the activation function.
  • the network layers are fully connected, that is, any neuron in layer i is connected to any neuron in layer i+1.
  • Deep neural network DNN can be designed and trained for analysis and prediction of various business scenarios.
  • the deep neural network DNN is taken as a branch part of the neural network system of an example.
  • the feature extraction layer 21 extracts the feature vector of the node
  • the feature vector is provided to the input layer of the deep neural network DNN 22, which is processed through the hidden layer, and the processing result is output through the output layer of the DNN.
  • the process of processing the feature vectors by the hidden layer of the DNN is called first processing, and the processing result output by the DNN output layer is called the first output.
  • the node feature vector extracted by the feature extraction layer 21 is also input to the graph neural network GNN23.
  • the graph neural network GNN is used to analyze and learn the relational network graph. Similar to the conventional neural network, the graph neural network GNN also has a multi-layer network structure, which operates and processes input data through the function mapping of neurons.
  • GNN will process the feature vectors of nodes in the process of combining the connection information between the nodes in the relational network graph.
  • the connection relationship information between the nodes in the above relationship network graph is also called adjacency information.
  • the process of processing the feature vectors of the nodes in the GNN in conjunction with the adjacency information is called second processing, and the result of the GNN processing is called the second output.
  • the adjacency information of the relationship network graph is obtained in advance and provided to the graph neural network GNN23.
  • connection relationship between nodes in the relational network graph can be represented by an adjacency matrix.
  • the adjacency matrix is an n*n-dimensional matrix.
  • the above relational network graph is a directed graph, that is, the connection between nodes is directed.
  • an adjacency table is also used to record the adjacency information of the relationship network graph, where the adjacency table may further include a forward adjacency table and a reverse adjacency table.
  • a cross-linked list is generated based on the positive adjacency list and the reverse adjacency list, and the cross-linked list is used to record the connection relationship between the nodes in the directed graph.
  • the adjacency information may also be recorded in other forms.
  • the graph neural network GNN 23 can determine the neighbor nodes of the current node and the connection edge information between the current node and these neighbor nodes based on the adjacency information, and extract the node information of these neighbor nodes and the edge information and features of the connection edges.
  • the feature vectors of the current node input at layer 21 are integrated, and a second output is obtained.
  • the graph neural network GNN 23 is implemented using a graph convolutional network GCN (Graph Convolutional Network).
  • GCN Graph Convolutional Network
  • Fig. 4 shows a schematic diagram of a graph convolution network GCN according to one embodiment.
  • the graph convolutional network GCN contains multiple network layers, and each network layer defines a neural network model f(X,A) through the neurons therein, where X is the input feature vector, that is, the aforementioned features
  • the feature vector of the current node input to the GCN by the extraction layer 21, A is the adjacency matrix of the relational network graph, and the neural network model f(X, A) can be more specifically expressed as:
  • I a degree matrix of A+ ⁇ I
  • is a hyperparameter, used to control the weight of a node relative to its neighbors, and is set to 1 in the original model.
  • H (l+1) represents the output of each network layer.
  • H 0 X, that is, the input layer receives the feature vector X of the node.
  • the d*d dimension vector W (l) and the d*1 dimension parameter b are both trainable network layer parameters, and ⁇ is a nonlinear function.
  • the ⁇ function may be a Relu function, a sigmoid function, a tanh function, a softmax function, and so on.
  • the first layer of the network uses the elements in the adjacency matrix as the weighting factors to sum the feature vectors of the current node and its neighbors (labeled or unlabeled), and then use W (l) and b Perform a linear transformation operation, and then apply a nonlinear activation function ⁇ .
  • the subsequent operations of each network layer also include at least, using the elements in the adjacency matrix as a weighting factor, and performing a weighted sum operation on the node vector output by the previous network layer and its neighbor node vectors; in addition, using W (l ) And b linear transformation operation, and the applied nonlinear activation function ⁇ operation.
  • W l and b l are the trainable network layer parameters of layer l
  • nhood(i) represents the neighbor node of node i.
  • the feature vectors of the nodes are processed by combining the adjacency information expressed by the adjacency matrix.
  • the above adjacency matrix may be a matrix after normalization, so as to avoid the situation that the distribution of elements in some adjacency matrices is too large.
  • some relational network graphs contain some super nodes, which are connected to almost every node in the graph; on the other hand, some nodes are very isolated and have few connections, which will cause different nodes in the adjacency matrix
  • the number of corresponding connecting edges (for example, the sum of elements corresponding to a row or a column in the matrix) is very different.
  • the adjacency matrix can be normalized.
  • the use of normalized adjacency matrix for weighted summation in GCN is equivalent to the average pooling operation of the current node and the adjacent nodes.
  • the graph neural network GNN 23 may also use other network structures and hidden layer algorithms.
  • the common point is that the second processing performed by the graph neural network GNN 23 needs to combine the adjacency information of the relational network graph to comprehensively process the feature vector of the current node to obtain the second output.
  • the first processing performed by the deep neural network DNN 22 is only for the feature vector of the current node, focusing on the analysis of the attribute characteristics of the node itself, that is, the single point feature
  • the second processing performed by the graph neural network GNN 23 needs to incorporate the adjacency information of the relational network graph and introduce the relationship characteristics between the current node and other nodes.
  • the first output of the deep neural network DNN 22 and the second output of the graph neural network 23 are fused through the fusion layer 24, and the prediction result for the current node is output based on the fusion result.
  • the first output be H1 and the second output H2.
  • the fusion layer 24 may fuse H1 and H2 in various ways to obtain a fusion result H.
  • the fusion layer 24 fuses the first output H1 and the second output H2 through a fusion function F:
  • the fusion function F can be various linear or nonlinear functions.
  • the fusion layer 24 performs weighted summation on the first output H1 and the second output H2 (corresponding to the case where the fusion function is linear summation), that is:
  • w1 is the first weighting factor corresponding to the first output
  • w2 is the second weighting factor corresponding to the second output
  • the first output H1 and the second output H2 are both in the form of output vectors; and the weighting factors w1 and w2 may be scalars, vectors, or even matrices.
  • the values of the weighting factors w1 and w2 are optimized and determined through the training process of the neural network system.
  • the above weight factor is set to a function corresponding to the output, for example, the first weight factor w1 is set to a function of the first output H1, and the second weight factor w2 is set to the second output H2 function:
  • the function g is preferably a non-linear function, such as sigmoid function and tanh function.
  • the weighting factors w1 and w2 are trained and determined, that is, the training and determination parameters u1, b1, u2, b2.
  • the weighting factors w1 and w2 are trained independently of each other, and the value range of the final result H is not guaranteed.
  • first the sum of the first weighting factor and the second weighting factor is set to 1, and then only one of them is set and adjusted.
  • set the fusion result H as:
  • the first weighting factor ⁇ may be set as a function of the first output, or a function of the first output and the second output, namely:
  • the second weighting factor ⁇ can also be set and adjusted so that the first weighting factor is (1- ⁇ ), that is:
  • the second weighting factor may also be set as a function of the second output, or a function of the first output and the second output, namely:
  • the fusion layer 24 obtains a fusion result H, and outputs a prediction result for the current node based on the fusion result H.
  • the prediction result is for the predicted value of the labeled node in the training phase; in the use phase, it is the final classification prediction for the unknown result.
  • the following describes the execution process of the neural network system shown in FIG. 2 in the training phase and the use phase.
  • the end-to-end training can be used.
  • labeled node information is input on the input side of the entire neural network system, that is, the feature extraction layer 21 extracts feature vectors of several labeled nodes.
  • the labels may be various types of labels, for example, labels used to represent risk levels, such as 1 for high-risk users, 0 for ordinary users, and so on. Then, the prediction results for each node are obtained on the output side of the entire neural network system.
  • the prediction result is output by the fusion layer 24 according to the fusion result, and may be embodied as a prediction value for each node. Compare the predicted value of each node with its label, according to the comparison result and the preset loss function, get the error of this batch of samples, and then through the error back propagation, adjust the network parameters of the entire neural network system, and finally Identify the network parameters that minimize the error. Once the optimal network parameters are determined, it can be considered that the training of the neural network system is completed, and the neural network system can be used for the prediction of unknown nodes.
  • the node information of the unknown node is input to the input side of the neural network system, that is, the feature extraction layer 21 extracts the feature vector of the unknown node.
  • the network parameters determined by the training stage in the neural network system are used, and the feature vector is processed by the parallel deep neural network DNN and the graph neural network GNN.
  • the fusion layer 24 outputs a prediction result based on the fusion result.
  • the prediction result is The output result of business prediction for the unknown node.
  • the neural network system of FIG. 2 combines the deep neural network DNN and the graph neural network GNN, which fuses the single-node feature processing ability of the DNN and the relational feature processing ability of the GNN, making the combined neural network
  • the system can effectively analyze and learn all kinds of relational network graphs.
  • Graph neural network GNN can play a major role in the case where the relationship features in the relationship network graph are perfect and effective.
  • the deep neural network DNN supplements the analysis of single nodes; and if the relationship features are missing or have limited effects, such as the existence of the relationship network graph A large number of isolated nodes, or the relationship based on the construction of the relationship network graph is not very effective for the business. In such a case, the nodes in the graph can be effectively analyzed and processed through the branch of the deep neural network DNN to give More ideal prediction results.
  • FIG. 5 shows a flowchart of a method for analyzing a relationship network diagram according to an embodiment. It can be understood that this method can be performed by any device, device, computing platform, or computing cluster with computing and processing capabilities. As shown in Figure 5, the method includes:
  • Step 51 Extract the feature vectors of the nodes in the relationship network graph
  • Step 52 Use a deep neural network to perform first processing on the feature vector to obtain a first output
  • Step 53 A graph neural network is used to combine the adjacency information of the relational network graph to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent each of the relations network graph The connection relationship between nodes;
  • Step 54 Fusion the first output and the second output, and output a prediction result for the node based on the fusion result.
  • step 52 and step 53 may be executed in any order, or in parallel, and are not limited herein.
  • each node included in the relationship network diagram corresponds to each user, and the connection relationship between the various nodes includes one or more of the following: social relationship, media relationship, and fund relationship between users .
  • the relationship network graph is a directed graph
  • the adjacency information includes an adjacency list or a cross-linked list corresponding to the directed graph.
  • the adjacency information includes the adjacency matrix of the relationship network graph.
  • the graph neural network described above is a graph convolutional network.
  • the graph convolutional network includes a plurality of network layers to perform the second processing.
  • the second processing includes at least using the adjacency matrix.
  • the element is a weighting factor, which performs a weighted sum operation on the feature vectors of the node and its neighbors.
  • the fusing the first output and the second output in step 54 specifically includes weighting and summing the first output and the second output, wherein the first output Corresponding to the first weighting factor, the second output corresponds to the second weighting factor.
  • the first weighting factor is a function of the first output
  • the second weighting factor is a function of the second output
  • the sum of the first weighting factor and the second weighting factor is 1, and:
  • the first weighting factor is a function of the first output
  • the second weighting factor is a function of the second output.
  • the sum of the first weighting factor and the second weighting factor is 1, and:
  • the first weighting factor is a function of the first output and the second output; or,
  • the second weighting factor is a function of the first output and the second output.
  • the relationship network graph is comprehensively analyzed.
  • a computer-readable storage medium on which a computer program is stored, and when the computer program is executed in a computer, the computer is caused to perform the method described in conjunction with FIG. 5.
  • a computing device including a memory and a processor, where executable code is stored in the memory, and when the processor executes the executable code, the implementation described in conjunction with FIG. 5 is implemented method.

Abstract

Disclosed are a neural network system and method executed by means of a computer and used for analyzing a relationship network graph. The neural network system comprises: a feature extraction layer (21) for extracting a feature vector of a node in a relationship network graph; a deep neural network (22) for carrying out first processing on the feature vector, so as to obtain a first output; a graph neural network (23) for carrying out second processing on the feature vector in combination with adjacency information of the relationship network graph, so as to obtain a second output, wherein the adjacency information is used for representing connection relationships between various nodes included in the relationship network graph; and a fusion layer (24) for fusing the first output and the second output, and outputting, based on a fusion result, a prediction result regarding the node.

Description

用于分析关系网络图的神经网络系统和方法Neural network system and method for analyzing relational network graph 技术领域Technical field
本说明书一个或多个实施例涉及通过计算机执行的神经网络系统,尤其涉及用于分析关系网络图的神经网络系统和方法。One or more embodiments of this specification relate to a neural network system executed by a computer, and particularly to a neural network system and method for analyzing a relational network graph.
背景技术Background technique
图是建模关系数据的有力工具,因此,目前往往是将存在关联关系的数据用图的形式表达和建模。另一方面,提出了基于图的运用深度学习方法的神经网络,图神经网络(Graph Neural Network,Graph NN或GNN),用于对图信息进行学习。图神经网络GNN可以有效地利用图上的信息传递以及融合节点或边的特征信息,来完成图上节点或边的分类或者回归等机器学习任务。The graph is a powerful tool for modeling relational data. Therefore, at present, the data in which the relationship exists is often expressed and modeled in the form of a graph. On the other hand, a graph-based neural network using deep learning methods, Graph Neural Network (Graph Neural Network, Graph NN or GNN) is proposed to learn graph information. The graph neural network GNN can effectively use the information transfer on the graph and fuse the feature information of the nodes or edges to complete the machine learning tasks such as the classification or regression of the nodes or edges on the graph.
然而,在真实的业务场景中,尤其是在业务的初始阶段,例如邀请新用户阶段,关系数据不完整。从图的角度来看,即存在很多的孤立节点。在这样的情况下,图神经网络GNN并不能达到预期的分析效果。However, in real business scenarios, especially at the initial stage of the business, such as inviting new users, the relational data is incomplete. From the perspective of the graph, there are many isolated nodes. Under such circumstances, the graph neural network GNN cannot achieve the expected analysis results.
因此,希望能有改进的方案,更加有效地对关系网络图进行学习、分析和预测。Therefore, it is hoped that there can be an improved solution to learn, analyze and predict the relationship network graph more effectively.
发明内容Summary of the invention
本说明书一个或多个实施例描述了通过计算机执行的、用于分析关系网络图的神经网络系统和方法,可以更加有效地对关系网络图进行学习、分析和预测。One or more embodiments of this specification describe a neural network system and method executed by a computer for analyzing a relationship network graph, which can learn, analyze, and predict the relationship network graph more effectively.
根据第一方面,提供了一种通过计算机执行的、用于分析关系网络图的神经网络系统,包括:According to a first aspect, a neural network system for analyzing a relationship network graph executed by a computer is provided, including:
特征提取层,用于提取关系网络图中的节点的特征向量;The feature extraction layer is used to extract the feature vectors of the nodes in the relationship network graph;
深度神经网络,用于对所述特征向量进行第一处理,得到第一输出;A deep neural network, which is used to perform first processing on the feature vector to obtain a first output;
图神经网络,用于结合所述关系网络图的邻接信息,对所述特征向量进行第二处理,得到第二输出;其中所述邻接信息用于表示所述关系网络图所包含的各个节点之间的连接关系;A graph neural network, used to combine the adjacency information of the relational network graph to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent each of the nodes included in the relational network graph Connection between
融合层,用于对所述第一输出和所述第二输出进行融合,基于融合结果输出针对 所述节点的预测结果。A fusion layer is used to fuse the first output and the second output, and output a prediction result for the node based on the fusion result.
在一个实施例中,关系网络图所包含的各个节点对应于各个用户,所述各个节点之间的连接关系包括以下中的一种或多种:用户之间的社交关系、媒介关系和资金关系。In one embodiment, each node included in the relationship network diagram corresponds to each user, and the connection relationship between the various nodes includes one or more of the following: social relationship, media relationship, and fund relationship between users .
根据一种可能的实现方式,关系网络图为有向图,相应地,所述邻接信息包括,所述有向图对应的邻接表或十字链表。According to a possible implementation manner, the relationship network graph is a directed graph, and accordingly, the adjacency information includes an adjacency list or a cross-linked list corresponding to the directed graph.
根据一种实施方式,邻接信息包括所述关系网络图的邻接矩阵。According to an embodiment, the adjacency information includes the adjacency matrix of the relationship network graph.
根据一个实施例,所述图神经网络为图卷积网络,所述图卷积网络包含多个网络层以进行所述第二处理,所述第二处理至少包括,使用所述邻接矩阵中的元素为权重因子,对所述节点及其邻居节点的特征向量进行加权求和操作。According to an embodiment, the graph neural network is a graph convolutional network, and the graph convolutional network includes multiple network layers to perform the second processing, and the second processing includes at least using the adjacency matrix. The element is a weighting factor, which performs a weighted sum operation on the feature vectors of the node and its neighbors.
根据一种实施方式,上述融合层具体用于,对所述第一输出和所述第二输出进行加权求和,其中所述第一输出对应第一权重因子,所述第二输出对应第二权重因子。According to an embodiment, the above fusion layer is specifically used for weighted summation of the first output and the second output, where the first output corresponds to a first weighting factor and the second output corresponds to a second Weighting factor.
进一步的,在一个实施例中,第一权重因子是第一输出的函数,第二权重因子是第二输出的函数。Further, in one embodiment, the first weighting factor is a function of the first output, and the second weighting factor is a function of the second output.
在另一实施例中,第一权重因子与第二权重因子和为1,并且:In another embodiment, the sum of the first weighting factor and the second weighting factor is 1, and:
所述第一权重因子是所述第一输出的函数;或者,The first weighting factor is a function of the first output; or,
所述第二权重因子是所述第二输出的函数。The second weighting factor is a function of the second output.
在又一实施例中,第一权重因子与第二权重因子和为1,并且:In yet another embodiment, the sum of the first weighting factor and the second weighting factor is 1, and:
所述第一权重因子是所述第一输出和第二输出的函数;或者,The first weighting factor is a function of the first output and the second output; or,
所述第二权重因子是所述第一输出和第二输出的函数。The second weighting factor is a function of the first output and the second output.
根据一种实现方式,所述神经网络系统通过端到端方式训练。According to one implementation, the neural network system is trained in an end-to-end manner.
根据第二方面,提供一种通过计算机执行的、用于分析关系网络图的方法,包括:According to a second aspect, there is provided a computer-implemented method for analyzing a relationship network graph, including:
提取关系网络图中的节点的特征向量;Extract the feature vectors of the nodes in the relationship network graph;
采用深度神经网络,对所述特征向量进行第一处理,得到第一输出;A deep neural network is used to perform the first processing on the feature vector to obtain a first output;
采用图神经网络,结合所述关系网络图的邻接信息,对所述特征向量进行第二处理,得到第二输出;其中所述邻接信息用于表示所述关系网络图所包含的各个节点之间的连接关系;A graph neural network is used, combined with the adjacency information of the relational network graph, to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent the relationship between each node included in the relational network graph Connection
对所述第一输出和所述第二输出进行融合,基于融合结果输出针对所述节点的预测结果。The first output and the second output are fused, and a prediction result for the node is output based on the fusion result.
根据第三方面,提供了一种计算机可读存储介质,其上存储有计算机程序,当所述计算机程序在计算机中执行时,令计算机执行第二方面的方法。According to a third aspect, a computer-readable storage medium is provided on which a computer program is stored, and when the computer program is executed in a computer, the computer is caused to perform the method of the second aspect.
根据第四方面,提供了一种计算设备,包括存储器和处理器,其特征在于,所述存储器中存储有可执行代码,所述处理器执行所述可执行代码时,实现第二方面的方法。According to a fourth aspect, a computing device is provided, including a memory and a processor, wherein the memory stores executable code, and when the processor executes the executable code, the method of the second aspect is implemented .
通过本说明书实施例提供的神经网络系统和方法,组合了深度神经网络DNN和图神经网络GNN,将DNN的单节点特征处理能力和GNN的关系特征处理能力进行了融合,使得组合的神经网络系统可以对各种各样的关系网络图进行有效的分析和学习。在关系网络图中关系特征完善且有效的情况下,图神经网络GNN可以发挥主要作用,深度神经网络DNN对单节点的分析作为补充;而如果关系特征缺失或者效果有限,仍然可以通过深度神经网络DNN的分支,对图中的节点进行有效分析处理,给出较为理想的预测结果。Through the neural network system and method provided by the embodiments of the present specification, a deep neural network DNN and a graph neural network GNN are combined, and the single-node feature processing capability of DNN and the relational feature processing capability of GNN are fused to make the combined neural network system Can effectively analyze and learn a variety of relationship network diagrams. In the case where the relationship features in the relationship network graph are perfect and effective, the graph neural network GNN can play a major role, and the deep neural network DNN supplements the single node analysis; and if the relationship features are missing or the effect is limited, the deep neural network can still be used The branch of DNN effectively analyzes and processes the nodes in the graph to give a more ideal prediction result.
附图说明BRIEF DESCRIPTION
为了更清楚地说明本申请实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to more clearly explain the technical solutions of the embodiments of the present application, the following will briefly introduce the drawings used in the description of the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application. Those of ordinary skill in the art can obtain other drawings based on these drawings without creative efforts.
图1为本说明书披露的一个实施例的关系网络图的示意图;FIG. 1 is a schematic diagram of a relationship network diagram disclosed in this specification;
图2示出根据一个实施例的神经网络系统的示意图;2 shows a schematic diagram of a neural network system according to an embodiment;
图3示出根据一个实施例的深度神经网络DNN的示意图;FIG. 3 shows a schematic diagram of a deep neural network DNN according to an embodiment;
图4示出根据一个实施例的图卷积网络GCN的示意图;4 shows a schematic diagram of a graph convolutional network GCN according to an embodiment;
图5示出根据一个实施例的分析关系网络图的方法流程图。FIG. 5 shows a flowchart of a method for analyzing a relationship network diagram according to an embodiment.
具体实施方式detailed description
下面结合附图,对本说明书提供的方案进行描述。The solution provided in this specification will be described below in conjunction with the drawings.
根据本说明书一个或多个实施例,提出一种用于处理关系数据的神经网络系统, 该神经网络系统可以用于对关系网络图进行学习和预测。According to one or more embodiments of the present specification, a neural network system for processing relational data is proposed. The neural network system can be used for learning and predicting relational network graphs.
下面首先对关系网络图进行描述。图1为本说明书披露的一个实施例的关系网络图的示意图。可以看到,关系网络图中包括多个节点,具有关联关系的节点之间通过连接边互相连接。与其他节点都没有关联关系的节点,形成孤立节点,例如图1中的节点A,B和C。The following first describes the relationship network diagram. FIG. 1 is a schematic diagram of a relationship network diagram disclosed in this specification. It can be seen that the relational network graph includes multiple nodes, and the nodes with associated relationships are connected to each other by connecting edges. Nodes that are not associated with other nodes form isolated nodes, such as nodes A, B, and C in Figure 1.
在一个实施例中,还可以对关联关系的类型和/或强度进行区分,从而为连接边赋予一定的属性或权值。例如图1中用粗线条示出强连接,细线条示出弱连接。然而,这并不是必须的。In one embodiment, the type and/or strength of the association relationship can also be distinguished, so as to assign certain attributes or weights to the connected edges. For example, in FIG. 1, thick lines indicate strong connections, and thin lines indicate weak connections. However, this is not necessary.
根据节点所代表的实体的不同,关系网络图可以反映各种实体之间的关联关系。例如,在一个实施例中,节点可以代表卖家或买家,节点之间的边可以代表发生过交易,由此,通过关系网络图反映实体之间的交易关系。According to the different entities represented by the nodes, the relationship network graph can reflect the association relationship between various entities. For example, in one embodiment, the nodes may represent sellers or buyers, and the edges between the nodes may represent that a transaction has occurred, thereby reflecting the transaction relationship between entities through a relationship network graph.
在另一个实施例中,节点代表各个用户,节点之间的边代表用户之间的关联关系。更具体地,在不同例子中,可以基于用户之间不同类型的关联关系,对节点建立连接关系。In another embodiment, the nodes represent individual users, and the edges between the nodes represent the association between users. More specifically, in different examples, a connection relationship may be established for nodes based on different types of association relationships between users.
在一个例子中,用户之间的关联关系可以包括,用户之间的社交关系。在基于社交关系形成的关系网络中,若两个用户有共同关注对象(例如微博账号共同关注了同一人),或他们之前有来往联系,或加入了共同群组(例如QQ群、微信群等),或在红包、彩票等活动中有互动,那么可以认为这两个节点之间存在社交关系,可以建立一条边进行连接。In one example, the association relationship between users may include a social relationship between users. In a relationship network formed based on social relationships, if two users have common followers (such as Weibo accounts following the same person together), or they have previous contacts, or join a common group (such as QQ group, WeChat group) Etc.), or interactions in activities such as red envelopes and lottery tickets, then it can be considered that there is a social relationship between the two nodes, and an edge can be established to connect.
在另一个例子中,用户之间的关联关系可以包括,用户之间的媒介关系。在基于媒介关系形成的关系网络中,若两个用户使用过同样的媒介,例如加密后的银行卡、身份证、邮箱、户号、手机号、物理地址(例如MAC地址)、终端设备号(例如UMID、TID、UTDID)等,则这两个用户之间存在媒介关系的关联,可以建立一条边进行连接。In another example, the association relationship between users may include a media relationship between users. In the relationship network formed based on the media relationship, if two users have used the same media, such as encrypted bank card, ID card, mailbox, account number, mobile phone number, physical address (such as MAC address), terminal device number ( For example, UMID, TID, UTDID, etc., then there is a media relationship between these two users, you can establish an edge to connect.
在又一例子中,用户之间的关联关系可以包括,用户之间的资金关系。在基于资金关系形成的关系网络中,如果两个用户之间至少存在资金交易中的一种,那么两节点之间存在一条边。资金交易可以包括代充、代付、条码收款、条码付款、AA收款、C2C手机当面付、送礼金、交房租、红包、信用卡代还款、代购、亲密付、代订服务等等。In yet another example, the association relationship between users may include the financial relationship between users. In the relationship network formed based on the fund relationship, if at least one of the fund transactions exists between two users, then there is an edge between the two nodes. Fund transactions can include charge, charge, bar code collection, bar code payment, AA collection, C2C mobile phone face-to-face payment, gift money, rent, red envelope, credit card repayment, purchase, intimate payment, subscription service, etc.
在其他实施例中,关系网络图中的节点还可以代表其他实体,节点之间的连接可以基于各种类型的关联关系。In other embodiments, the nodes in the relationship network graph may also represent other entities, and the connections between the nodes may be based on various types of association relationships.
对于图1的关系网络图,一般可以采用图神经网络GNN进行学习和预测。学习的过程一般对应于图神经网络GNN的训练过程。For the relational network graph of FIG. 1, the graph neural network GNN can generally be used for learning and prediction. The learning process generally corresponds to the training process of graph neural network GNN.
在训练图神经网络GNN时,需要根据预测业务的目的,为图1中至少部分节点添加标签,从而进行监督学习。例如,假定图1中关系网络图中的节点对应于各个用户,如果要学习和预测各个用户的信用风险,那么至少需要为部分节点添加“高风险用户”(涉嫌欺诈的用户)和“正常用户”的标签,将这些有标签的节点,连同这些节点在图中的连接关系提供给图神经网络GNN,以供GNN进行训练和学习。训练完成之后,就可以将未知用户输入到图神经网络GNN,使得图神经网络GNN利用训练好的网络参数,对该未知用户进行预测。When training the graph neural network GNN, it is necessary to add labels to at least some nodes in FIG. 1 according to the purpose of the prediction business, so as to perform supervised learning. For example, assuming that the nodes in the relational network graph in FIG. 1 correspond to individual users, if you want to learn and predict the credit risk of each user, at least some nodes need to add "high risk users" (users suspected of fraud) and "normal users" ", provide these labeled nodes, together with the connection relationship of these nodes in the graph, to the graph neural network GNN for GNN to train and learn. After the training is completed, the unknown user can be input to the graph neural network GNN, so that the graph neural network GNN uses the trained network parameters to predict the unknown user.
然而,在许多情况下,上述方案的效果不够理想。However, in many cases, the effect of the above scheme is not ideal.
一方面,关系网络图是基于关联关系进行构建。如前所述,节点之间存在各种各样的关联关系,如果在构建关系网络图时,所选择的关联关系与业务目的匹配度不够高,那么基于这样的关联关系所构建的关系网络图进行学习,效果往往不够理想。例如,在业务目的是想要学习用户对某类产品的购买意图的情况下,如果关联关系选择为是否具有共同关注对象的社交关系,那么这样的关系数据对于判断购买意图的业务目的不一定有效。On the one hand, the relationship network graph is constructed based on the association relationship. As mentioned earlier, there are various association relationships between nodes. If the relationship between the selected association relationship and the business purpose is not high enough when constructing the relationship network graph, then the relationship network graph constructed based on such association relationship Learning, the results are often not ideal. For example, in the case where the business purpose is to learn the user’s purchase intention of a certain type of product, if the association relationship is selected as whether it has a social relationship with a common object of interest, such relationship data may not be effective for determining the business purpose of the purchase intention .
另一方面,在一些业务场景下,例如拉新、促销等业务阶段中,大量增加新用户,此时新用户的关系数据很不完整,在图中形成大量孤立节点。对这些孤立节点的预测存在一定困难。因为在训练阶段,图神经网络GNN通常是基于连接关系比较完善的关系图进行学习的,训练阶段和预测阶段样本分布差异较大,会影响图神经网络GNN的预测效果。并且,孤立节点本身有关连接关系的特征数据欠缺,也使得图神经网络GNN无法达到其最优效果。On the other hand, in some business scenarios, such as pull-in, promotion, and other business stages, a large number of new users are added. At this time, the relationship data of the new users is very incomplete, forming a large number of isolated nodes in the graph. There are certain difficulties in predicting these isolated nodes. Because in the training phase, the graph neural network GNN is usually learned based on the relationship graph with a relatively perfect connection relationship, the sample distribution between the training phase and the prediction phase is quite different, which will affect the prediction effect of the graph neural network GNN. Moreover, the lack of characteristic data about the connection relationship of the isolated node itself makes the graph neural network GNN unable to achieve its optimal effect.
基于以上的分析和研究,在本说明书的实施例中,提供一种组合式的神经网络系统,可以更有效地对关系网络图进行学习。Based on the above analysis and research, in the embodiments of the present specification, a combined neural network system is provided, which can learn the relationship network graph more effectively.
图2示出根据一个实施例的神经网络系统的示意图,该神经网络系统通过计算机执行,用于对关系网络图进行学习,对关系数据进行处理。如图2所示,该神经网路系统包括,特征提取层21,用于提取关系网络图中的节点的特征向量;深度神经网络22,用于对所述特征向量进行第一处理,得到第一输出;图神经网络23,用于结合所述关系网络图的邻接信息,对所述特征向量进行第二处理,得到第二输出;融合层24,用于对 所述第一输出和所述第二输出进行融合,基于融合结果输出针对所述节点的预测结果。下面具体描述以上各个网络部分的执行方式。FIG. 2 shows a schematic diagram of a neural network system according to an embodiment. The neural network system is executed by a computer and is used to learn a relationship network graph and process relationship data. As shown in FIG. 2, the neural network system includes a feature extraction layer 21 for extracting feature vectors of nodes in the relational network graph; a deep neural network 22 for performing first processing on the feature vectors to obtain An output; graph neural network 23, used to combine the adjacency information of the relational network graph, and perform a second processing on the feature vector to obtain a second output; a fusion layer 24, used for the first output and the The second output is fused, and a prediction result for the node is output based on the fusion result. The following specifically describes the implementation of each of the above network parts.
特征提取层21用于提取关系网络图中的节点的特征向量。该关系网络图例如是图1所示的关系网络图,其中包含多个节点,具有关联关系的节点之间存在连接关系。特征提取层21针对关系网络图中的节点进行特征提取,所提取的特征是有关节点本身的特征,如此提取的特征构成特征向量。The feature extraction layer 21 is used to extract feature vectors of nodes in the relationship network graph. The relationship network diagram is, for example, the relationship network diagram shown in FIG. 1, which includes multiple nodes, and there is a connection relationship between nodes having an association relationship. The feature extraction layer 21 performs feature extraction on the nodes in the relational network graph, and the extracted features are the features of the relevant nodes themselves, so the extracted features constitute a feature vector.
在一个实施例中,上述关系网络图中的各个节点对应于各个用户,例如支付宝用户。针对这样的节点,特征提取层21可以提取用户的基本属性特征,用于构成特征向量,上述基本属性特征例如包括用户ID、注册时长、性别、年龄等等。In one embodiment, each node in the above relational network diagram corresponds to each user, for example, Alipay user. For such a node, the feature extraction layer 21 may extract the user's basic attribute features to form a feature vector. The basic attribute features include, for example, user ID, registration duration, gender, age, and so on.
在一个实施例中,特征提取层21还根据业务目的,提取与业务目的相关的特征。例如,在业务目的为,预测用户的购买意图的情况下,特征提取层21还获取用户的购买记录,基于购买记录进行特征提取,例如提取购买次数、购买品类、购买金额等多项特征。又例如,在业务目的为,预测用户的借贷风险的情况下,特征提取层21还获取用户的借贷记录,基于借贷记录进行特征提取,例如提取借贷次数、借贷金额、已还款金额、守信次数、逾期次数等多项特征。In one embodiment, the feature extraction layer 21 also extracts features related to the business purpose according to the business purpose. For example, in the case where the business purpose is to predict the user's purchase intention, the feature extraction layer 21 also obtains the user's purchase records, and performs feature extraction based on the purchase records, such as extracting multiple features such as the number of purchases, purchase categories, and purchase amount. For another example, when the business purpose is to predict the borrowing risk of the user, the feature extraction layer 21 also obtains the borrowing record of the user, and extracts the features based on the borrowing record, such as extracting the number of loans, the amount of loan, the amount of repayment, and the number of trustworthiness , Overdue times and other characteristics.
在特征提取层21提取得到节点的特征向量后,将该特征向量并行地分别输入到深度神经网络22和图神经网络23,分别进行处理。After the feature extraction layer 21 extracts the feature vectors of the nodes, the feature vectors are input to the deep neural network 22 and the graph neural network 23 in parallel for processing.
深度神经网络(Deep Neural Network,DNN)是一种多层全连接的前向结构的人工神经网络。图3示出根据一个实施例的深度神经网络的示意图。如图3所示,深度神经网络DNN包含多个网络层,这些网络层可以分为输入层,隐藏层和输出层,输入层与输出层之间的都是隐藏层。一般的,当一个神经网络中隐藏层的层数较多时,我们才称之为深度神经网络。DNN中各个网络层均包含若干神经元,除输入层之外的神经元均通过激活函数对输入数据进行运算操作。在DNN中,网络层之间是全连接的,也就是说,第i层的任意一个神经元均与第i+1层的任意一个神经元相连。Deep Neural Network (DNN) is a multi-layer fully connected forward-structured artificial neural network. Figure 3 shows a schematic diagram of a deep neural network according to one embodiment. As shown in Figure 3, the deep neural network DNN contains multiple network layers. These network layers can be divided into an input layer, a hidden layer, and an output layer. Between the input layer and the output layer are hidden layers. Generally, when there are many hidden layers in a neural network, we call it a deep neural network. Each network layer in the DNN contains several neurons, and all neurons except the input layer operate on the input data through the activation function. In DNN, the network layers are fully connected, that is, any neuron in layer i is connected to any neuron in layer i+1.
深度神经网络DNN可以被设计、训练用于各种业务场景的分析和预测。Deep neural network DNN can be designed and trained for analysis and prediction of various business scenarios.
根据本说明书一种实施方式,将深度神经网络DNN作为一个实施例的神经网络系统的一个分支部分。在特征提取层21提取得到节点的特征向量后,该特征向量被提供给深度神经网络DNN 22的输入层,经由隐藏层对其进行处理,并通过DNN的输出层输出处理结果。为了描述的简单,将DNN隐藏层对特征向量的处理过程称为第一处理, 将DNN输出层输出的处理结果称为第一输出。According to an embodiment of the present specification, the deep neural network DNN is taken as a branch part of the neural network system of an example. After the feature extraction layer 21 extracts the feature vector of the node, the feature vector is provided to the input layer of the deep neural network DNN 22, which is processed through the hidden layer, and the processing result is output through the output layer of the DNN. For simplicity of description, the process of processing the feature vectors by the hidden layer of the DNN is called first processing, and the processing result output by the DNN output layer is called the first output.
需要说明的是,本文中的“第一”、“第二”等描述仅仅是位于区分相似概念,以及为了描述的简单和清楚,并不具有顺序等其他方面的限定作用。It should be noted that the descriptions of "first" and "second" in this article are only for distinguishing similar concepts, and for simplicity and clarity of description, they do not have a limiting effect on order and other aspects.
另一方面,特征提取层21提取的节点特征向量还被输入到图神经网络GNN 23。可以理解,图神经网络GNN用于对关系网络图进行分析和学习。与常规神经网络类似的,图神经网络GNN也具有多层的网络结构,通过神经元的函数映射对输入数据进行操作和处理。特别的,作为专用于处理关系网络图的神经网络,GNN在处理过程中,会结合关系网络图中各个节点之间的连接关系信息,对节点的特征向量进行处理。上述关系网络图中各个节点之间的连接关系信息又称为邻接信息。为了描述方便,将GNN中结合邻接信息对节点的特征向量进行处理的过程称为第二处理,将GNN处理的结果称为第二输出。On the other hand, the node feature vector extracted by the feature extraction layer 21 is also input to the graph neural network GNN23. It can be understood that the graph neural network GNN is used to analyze and learn the relational network graph. Similar to the conventional neural network, the graph neural network GNN also has a multi-layer network structure, which operates and processes input data through the function mapping of neurons. In particular, as a neural network dedicated to processing relational network graphs, GNN will process the feature vectors of nodes in the process of combining the connection information between the nodes in the relational network graph. The connection relationship information between the nodes in the above relationship network graph is also called adjacency information. For convenience of description, the process of processing the feature vectors of the nodes in the GNN in conjunction with the adjacency information is called second processing, and the result of the GNN processing is called the second output.
如图2所示,为了GNN处理的需要,预先获取关系网络图的邻接信息,提供给图神经网络GNN 23。As shown in FIG. 2, for the needs of GNN processing, the adjacency information of the relationship network graph is obtained in advance and provided to the graph neural network GNN23.
上述邻接信息可以体现为多种形式。典型的,可以通过邻接矩阵来表示关系网络图中节点之间的连接关系。假定关系网络图中包含n个节点,那么邻接矩阵是一个n*n维的矩阵,在最简单的情况下(节点之间的连接边具有等同权值的情况下),如果节点i和节点j之间有连接关系,则矩阵元素A ij=1,否则,矩阵元素A ij=0。 The aforementioned adjacency information can be embodied in various forms. Typically, the connection relationship between nodes in the relational network graph can be represented by an adjacency matrix. Assuming that the relational network graph contains n nodes, the adjacency matrix is an n*n-dimensional matrix. In the simplest case (in the case where the connecting edges between nodes have equal weights), if node i and node j There is a connection relationship, then the matrix element A ij =1, otherwise, the matrix element A ij =0.
此外,还可以采用度矩阵、拉普拉斯矩阵等形式,表示关系网络图中节点之间的连接关系。In addition, you can also use the degree matrix, Laplace matrix and other forms to represent the connection relationship between nodes in the relationship network graph.
在一个实施例中,上述关系网络图是一个有向图,即,节点之间的连接是有方向的。对于有向图,在一些情况下,还采用邻接表来记录关系网络图的邻接信息,其中邻接表可以进一步包括正向邻接表和逆邻接表。在另一种示例中,基于正邻接表和逆邻接表生成十字链表,采用十字链表记录有向图中节点之间的连接关系。In one embodiment, the above relational network graph is a directed graph, that is, the connection between nodes is directed. For directed graphs, in some cases, an adjacency table is also used to record the adjacency information of the relationship network graph, where the adjacency table may further include a forward adjacency table and a reverse adjacency table. In another example, a cross-linked list is generated based on the positive adjacency list and the reverse adjacency list, and the cross-linked list is used to record the connection relationship between the nodes in the directed graph.
在其他实施例中,还可以采用其他形式记录邻接信息。In other embodiments, the adjacency information may also be recorded in other forms.
于是,图神经网络GNN 23可以基于邻接信息,确定当前节点的邻居节点,以及当前节点与这些邻居节点之间的连接边的信息,将这些邻居节点的节点信息和连接边的边信息与特征提取层21输入进来的当前节点的特征向量进行综合处理,从而得到第二输出。Therefore, the graph neural network GNN 23 can determine the neighbor nodes of the current node and the connection edge information between the current node and these neighbor nodes based on the adjacency information, and extract the node information of these neighbor nodes and the edge information and features of the connection edges. The feature vectors of the current node input at layer 21 are integrated, and a second output is obtained.
具体的,在一个实施例中,图神经网络GNN 23采用图卷积网络GCN(Graph  Convolutional Network)实现。Specifically, in one embodiment, the graph neural network GNN 23 is implemented using a graph convolutional network GCN (Graph Convolutional Network).
图4示出根据一个实施例的图卷积网络GCN的示意图。在一个实施例中,图卷积网络GCN包含多个网络层,每个网络层通过其中的神经元定义了一个神经网络模型f(X,A),其中X为输入特征向量,即前述的特征提取层21输入到GCN的、当前节点的特征向量,A为关系网络图的邻接矩阵,该神经网络模型f(X,A)更具体的可以表示为:Fig. 4 shows a schematic diagram of a graph convolution network GCN according to one embodiment. In one embodiment, the graph convolutional network GCN contains multiple network layers, and each network layer defines a neural network model f(X,A) through the neurons therein, where X is the input feature vector, that is, the aforementioned features The feature vector of the current node input to the GCN by the extraction layer 21, A is the adjacency matrix of the relational network graph, and the neural network model f(X, A) can be more specifically expressed as:
Figure PCTCN2019112564-appb-000001
Figure PCTCN2019112564-appb-000001
Figure PCTCN2019112564-appb-000002
Figure PCTCN2019112564-appb-000002
其中,
Figure PCTCN2019112564-appb-000003
为A+λI的度矩阵,λ为超参数,用于控制节点相对于其邻居节点的权重,在原始模型中被设定为1。
among them,
Figure PCTCN2019112564-appb-000003
Is a degree matrix of A+λI, and λ is a hyperparameter, used to control the weight of a node relative to its neighbors, and is set to 1 in the original model.
H (l+1)表示每一网络层的输出。l=0时,H 0=X,即为输入层接收到节点的特征向量X的情况。d*d维向量W (l)和d*1维的参数b均为可训练的网络层参数,σ为非线性函数。在不同例子中,σ函数可以是Relu函数,sigmoid函数,tanh函数,softmax函数等等。 H (l+1) represents the output of each network layer. When l=0, H 0 =X, that is, the input layer receives the feature vector X of the node. The d*d dimension vector W (l) and the d*1 dimension parameter b are both trainable network layer parameters, and σ is a nonlinear function. In different examples, the σ function may be a Relu function, a sigmoid function, a tanh function, a softmax function, and so on.
通过上式可以看到,网络的第一层使用邻接矩阵中的元素作为权重因子,对当前节点及其邻居节点(有标签或无标签)的特征向量进行求和,然后使用W (l)和b进行线性变换操作,之后施加非线性激活函数σ。后续的每个网络层的操作也至少包括,使用邻接矩阵中的元素为权重因子,对上一网络层输出的节点向量及其邻居节点向量进行加权求和操作;此外还包括,使用W (l)和b进行的线性变换操作,以及施加的非线性激活函数σ操作。 It can be seen from the above equation that the first layer of the network uses the elements in the adjacency matrix as the weighting factors to sum the feature vectors of the current node and its neighbors (labeled or unlabeled), and then use W (l) and b Perform a linear transformation operation, and then apply a nonlinear activation function σ. The subsequent operations of each network layer also include at least, using the elements in the adjacency matrix as a weighting factor, and performing a weighted sum operation on the node vector output by the previous network layer and its neighbor node vectors; in addition, using W (l ) And b linear transformation operation, and the applied nonlinear activation function σ operation.
换而言之,对于节点u i,第l层的输出通过下式计算: In other words, for node u i , the output of layer l is calculated by the following formula:
Figure PCTCN2019112564-appb-000004
Figure PCTCN2019112564-appb-000004
其中W l和b l为l层的可训练的网络层参数,nhood(i)表示节点i的邻居节点。 Where W l and b l are the trainable network layer parameters of layer l, and nhood(i) represents the neighbor node of node i.
如此,在图卷积网络GCN中,结合邻接矩阵所表达的邻接信息,对节点的特征向量进行处理。In this way, in the graph convolutional network GCN, the feature vectors of the nodes are processed by combining the adjacency information expressed by the adjacency matrix.
在一个实施例中,上述邻接矩阵可以是归一化之后的矩阵,如此避免有些邻接矩阵中元素分布差异过大的情况。例如,在有些关系网络图中包含一些超级节点,超级节 点跟图中的几乎每个节点都有连接;另一方面,有一些节点非常孤立,连接极少,这就会造成邻接矩阵中不同节点对应的连接边的数目(例如对应于矩阵中某行或某列的元素之和)差异非常大。为此,可以对邻接矩阵进行归一化。在GCN中使用归一化的邻接矩阵进行加权求和,相当于对当前节点和邻接节点进行了平均池化(average pooling)操作。In an embodiment, the above adjacency matrix may be a matrix after normalization, so as to avoid the situation that the distribution of elements in some adjacency matrices is too large. For example, some relational network graphs contain some super nodes, which are connected to almost every node in the graph; on the other hand, some nodes are very isolated and have few connections, which will cause different nodes in the adjacency matrix The number of corresponding connecting edges (for example, the sum of elements corresponding to a row or a column in the matrix) is very different. To this end, the adjacency matrix can be normalized. The use of normalized adjacency matrix for weighted summation in GCN is equivalent to the average pooling operation of the current node and the adjacent nodes.
在其他实施例中,图神经网络GNN 23还可以采用其他的网络结构和隐藏层算法。然而,共同点是,图神经网络GNN 23所进行的第二处理,需要结合关系网络图的邻接信息,对当前节点的特征向量进行综合处理,从而得到第二输出。In other embodiments, the graph neural network GNN 23 may also use other network structures and hidden layer algorithms. However, the common point is that the second processing performed by the graph neural network GNN 23 needs to combine the adjacency information of the relational network graph to comprehensively process the feature vector of the current node to obtain the second output.
对比深度神经网络DNN 22和图神经网络GNN 23的处理过程可以理解,深度神经网络DNN 22所进行的第一处理仅针对当前节点的特征向量,侧重于分析节点本身的属性特征,即单点特征;图神经网络GNN 23所进行的第二处理需要结合关系网络图的邻接信息,引入了当前节点与其他节点的关系特征。Comparing the processing of the deep neural network DNN 22 and the graph neural network GNN 23, it can be understood that the first processing performed by the deep neural network DNN 22 is only for the feature vector of the current node, focusing on the analysis of the attribute characteristics of the node itself, that is, the single point feature The second processing performed by the graph neural network GNN 23 needs to incorporate the adjacency information of the relational network graph and introduce the relationship characteristics between the current node and other nodes.
接下来,通过融合层24将深度神经网络DNN 22的第一输出和图神经网络23的第二输出进行融合,基于融合结果输出针对当前节点的预测结果。为了表述方便,将第一输出记为H1,将第二输出记为H2。在不同实施例中,融合层24可以通过各种不同方式对H1和H2进行融合,得到融合结果H。Next, the first output of the deep neural network DNN 22 and the second output of the graph neural network 23 are fused through the fusion layer 24, and the prediction result for the current node is output based on the fusion result. For convenience of expression, let the first output be H1 and the second output H2. In different embodiments, the fusion layer 24 may fuse H1 and H2 in various ways to obtain a fusion result H.
在一个实施例中,融合层24通过一个融合函数F对第一输出H1和第二输出H2进行融合:In one embodiment, the fusion layer 24 fuses the first output H1 and the second output H2 through a fusion function F:
H=F(H1,H2)H=F(H1, H2)
其中融合函数F可以是各种线性或非线性函数。The fusion function F can be various linear or nonlinear functions.
在一个实施例中,融合层24对第一输出H1和第二输出H2进行加权求和(对应于融合函数为线性求和的情况),即:In one embodiment, the fusion layer 24 performs weighted summation on the first output H1 and the second output H2 (corresponding to the case where the fusion function is linear summation), that is:
H=w1*H1+w2*H2H=w1*H1+w2*H2
其中w1为第一输出对应的第一权重因子,w2为第二输出对应的第二权重因子。Where w1 is the first weighting factor corresponding to the first output, and w2 is the second weighting factor corresponding to the second output.
可以理解,第一输出H1和第二输出H2均为输出向量的形式;而权重因子w1和w2可以是标量,也可以是向量,甚至是矩阵。权重因子w1和w2的值通过神经网络系统的训练过程而优化、确定。It can be understood that the first output H1 and the second output H2 are both in the form of output vectors; and the weighting factors w1 and w2 may be scalars, vectors, or even matrices. The values of the weighting factors w1 and w2 are optimized and determined through the training process of the neural network system.
在一个实施例中,进一步地,将上述权重因子设置为对应输出的函数,例如,将 第一权重因子w1设置为第一输出H1的函数,将第二权重因子w2设置为第二输出H2的函数:In one embodiment, further, the above weight factor is set to a function corresponding to the output, for example, the first weight factor w1 is set to a function of the first output H1, and the second weight factor w2 is set to the second output H2 function:
w1=f1(H1)w1=f1(H1)
w2=f2(H2)w2=f2(H2)
更具体的,上述函数具体形式可以为:More specifically, the specific form of the above function may be:
w1=g(u1*H1+b1)w1=g(u1*H1+b1)
w2=g(u2*H2+b2)w2=g(u2*H2+b2)
其中函数g优选为非线性函数,例如sigmoid函数,tanh函数。The function g is preferably a non-linear function, such as sigmoid function and tanh function.
如此,训练和确定权重因子w1和w2,即训练和确定参数u1,b1,u2,b2。In this way, the weighting factors w1 and w2 are trained and determined, that is, the training and determination parameters u1, b1, u2, b2.
在以上方式中,权重因子w1和w2互相独立训练,并且不确保最终结果H的取值范围。In the above manner, the weighting factors w1 and w2 are trained independently of each other, and the value range of the final result H is not guaranteed.
在另一实施例中,首先将第一权重因子与第二权重因子之和设为1,然后仅设置和调整其中一个。例如,将融合结果H设定为:In another embodiment, first the sum of the first weighting factor and the second weighting factor is set to 1, and then only one of them is set and adjusted. For example, set the fusion result H as:
H=α*H1+(1-α)*H2H=α*H1+(1-α)*H2
如此,仅需要设置和调整第一权重因子α,第二权重因子随之确定。In this way, only the first weighting factor α needs to be set and adjusted, and the second weighting factor is determined accordingly.
在一个实施例中,该第一权重因子α可以设置为第一输出的函数,或者,第一输出和第二输出的函数,即:In an embodiment, the first weighting factor α may be set as a function of the first output, or a function of the first output and the second output, namely:
α=g(H1)=g(u*H1+b),或者α=g(H1)=g(u*H1+b), or
α=g(H1,H2)=g(u1*H1+u2*H2+b)α=g(H1, H2)=g(u1*H1+u2*H2+b)
当然,也可以设置和调整第二权重因子β,使得第一权重因子为(1-β),即:Of course, the second weighting factor β can also be set and adjusted so that the first weighting factor is (1-β), that is:
H=(1-β)*H1+β*H2H=(1-β)*H1+β*H2
进一步地,也可以将第二权重因子设置为第二输出的函数,或者,第一输出和第二输出的函数,即:Further, the second weighting factor may also be set as a function of the second output, or a function of the first output and the second output, namely:
β=g(H2)=g(u*H2+b),或者β=g(H2)=g(u*H2+b), or
β=g(H1,H2)=g(u1*H1+u2*H2+b)β=g(H1, H2)=g(u1*H1+u2*H2+b)
以上举例了几种典型的融合方式。在以上举例的基础上,本领域技术人员还可以 联想到其他类似的融合方式,这些方式均应涵盖在本说明书的构思之内。The above examples illustrate several typical fusion methods. On the basis of the above examples, those skilled in the art can also think of other similar fusion methods, which should be covered by the concept of this specification.
通过各种融合方式,融合层24得到融合结果H,并基于融合结果H输出针对当前节点的预测结果。该预测结果在训练阶段,是针对有标签节点的预测值;在使用阶段,是针对未知结果的最终分类预测。下面描述图2所示的神经网络系统在训练阶段和使用阶段的执行过程。Through various fusion methods, the fusion layer 24 obtains a fusion result H, and outputs a prediction result for the current node based on the fusion result H. The prediction result is for the predicted value of the labeled node in the training phase; in the use phase, it is the final classification prediction for the unknown result. The following describes the execution process of the neural network system shown in FIG. 2 in the training phase and the use phase.
对于图2所示的组合了深度神经网络DNN和图神经网络GNN的神经网络系统,可以采用端到端方式进行训练。具体而言,在训练阶段,在整个神经网络系统的输入侧输入带标签的节点信息,即特征提取层21提取若干带标签节点的特征向量。如前所述,根据业务目的不同,标签可以是各种类别标签,例如用于表示风险等级的标签,比如1表示高风险用户,0表示普通用户,等等。然后,在整个神经网络系统的输出侧获取针对各个节点的预测结果。具体地,该预测结果是融合层24根据融合结果而输出,可以体现为针对各个节点的预测值。将各个节点的预测值与其标签进行比对,根据比对结果和预设的损失函数,得到这批样本的误差,然后通过误差反向传播,对整个神经网络系统的网络参数进行调整,并最终确定出使得误差达到最小的网络参数。一旦确定出最优的网络参数,可以认为完成了神经网络系统的训练,该神经网络系统可以用于未知节点的预测。For the neural network system shown in FIG. 2 combining the deep neural network DNN and the graph neural network GNN, the end-to-end training can be used. Specifically, in the training stage, labeled node information is input on the input side of the entire neural network system, that is, the feature extraction layer 21 extracts feature vectors of several labeled nodes. As mentioned above, according to different business purposes, the labels may be various types of labels, for example, labels used to represent risk levels, such as 1 for high-risk users, 0 for ordinary users, and so on. Then, the prediction results for each node are obtained on the output side of the entire neural network system. Specifically, the prediction result is output by the fusion layer 24 according to the fusion result, and may be embodied as a prediction value for each node. Compare the predicted value of each node with its label, according to the comparison result and the preset loss function, get the error of this batch of samples, and then through the error back propagation, adjust the network parameters of the entire neural network system, and finally Identify the network parameters that minimize the error. Once the optimal network parameters are determined, it can be considered that the training of the neural network system is completed, and the neural network system can be used for the prediction of unknown nodes.
在使用阶段,将未知节点的节点信息输入到神经网络系统的输入侧,即,即特征提取层21提取未知节点的特征向量。然后采用神经网络系统中经由训练阶段而确定下来的网络参数,通过并列的深度神经网络DNN和图神经网络GNN对该特征向量进行处理,由融合层24基于融合结果输出预测结果,该预测结果即针对该未知节点进行业务预测的输出结果。In the use phase, the node information of the unknown node is input to the input side of the neural network system, that is, the feature extraction layer 21 extracts the feature vector of the unknown node. Then, the network parameters determined by the training stage in the neural network system are used, and the feature vector is processed by the parallel deep neural network DNN and the graph neural network GNN. The fusion layer 24 outputs a prediction result based on the fusion result. The prediction result is The output result of business prediction for the unknown node.
综合以上可以看到,图2的神经网络系统,由于组合了深度神经网络DNN和图神经网络GNN,将DNN的单节点特征处理能力和GNN的关系特征处理能力进行了融合,使得组合的神经网络系统可以对各种各样的关系网络图进行有效的分析和学习。在关系网络图中关系特征完善且有效的情况下,图神经网络GNN可以发挥主要作用,深度神经网络DNN对单节点的分析作为补充;而如果关系特征缺失或者效果有限,例如关系网络图中存在大量孤立节点,或者关系网络图的构建所基于的关联关系对于业务并不十分有效,在这样的情况下,仍然可以通过深度神经网络DNN的分支,对图中的节点进行有效分析处理,给出较为理想的预测结果。It can be seen from the above that the neural network system of FIG. 2 combines the deep neural network DNN and the graph neural network GNN, which fuses the single-node feature processing ability of the DNN and the relational feature processing ability of the GNN, making the combined neural network The system can effectively analyze and learn all kinds of relational network graphs. Graph neural network GNN can play a major role in the case where the relationship features in the relationship network graph are perfect and effective. The deep neural network DNN supplements the analysis of single nodes; and if the relationship features are missing or have limited effects, such as the existence of the relationship network graph A large number of isolated nodes, or the relationship based on the construction of the relationship network graph is not very effective for the business. In such a case, the nodes in the graph can be effectively analyzed and processed through the branch of the deep neural network DNN to give More ideal prediction results.
根据另一方面的实施例,还提供一种通过计算机执行的、用于分析关系网络图的 方法。图5示出根据一个实施例的分析关系网络图的方法流程图。可以理解,该方法可以通过任何具有计算、处理能力的装置、设备、计算平台、计算集群来执行。如图5所示,该方法包括:According to an embodiment of another aspect, a method for analyzing a relationship network graph executed by a computer is also provided. FIG. 5 shows a flowchart of a method for analyzing a relationship network diagram according to an embodiment. It can be understood that this method can be performed by any device, device, computing platform, or computing cluster with computing and processing capabilities. As shown in Figure 5, the method includes:
步骤51,提取关系网络图中的节点的特征向量;Step 51: Extract the feature vectors of the nodes in the relationship network graph;
步骤52,采用深度神经网络,对所述特征向量进行第一处理,得到第一输出;Step 52: Use a deep neural network to perform first processing on the feature vector to obtain a first output;
步骤53,采用图神经网络,结合所述关系网络图的邻接信息,对所述特征向量进行第二处理,得到第二输出;其中所述邻接信息用于表示所述关系网络图所包含的各个节点之间的连接关系;Step 53: A graph neural network is used to combine the adjacency information of the relational network graph to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent each of the relations network graph The connection relationship between nodes;
步骤54,对所述第一输出和所述第二输出进行融合,基于融合结果输出针对所述节点的预测结果。Step 54: Fusion the first output and the second output, and output a prediction result for the node based on the fusion result.
需要说明的是,步骤52和步骤53可以以任意先后顺序执行,或并行执行,在此不做限定。It should be noted that step 52 and step 53 may be executed in any order, or in parallel, and are not limited herein.
在一个实施例中,关系网络图所包含的各个节点对应于各个用户,所述各个节点之间的连接关系包括以下中的一种或多种:用户之间的社交关系、媒介关系和资金关系。In one embodiment, each node included in the relationship network diagram corresponds to each user, and the connection relationship between the various nodes includes one or more of the following: social relationship, media relationship, and fund relationship between users .
在一种可能的实施方式中,关系网络图为有向图,所述邻接信息包括,所述有向图对应的邻接表或十字链表。In a possible implementation manner, the relationship network graph is a directed graph, and the adjacency information includes an adjacency list or a cross-linked list corresponding to the directed graph.
在一种实施方案中,邻接信息包括所述关系网络图的邻接矩阵。In one embodiment, the adjacency information includes the adjacency matrix of the relationship network graph.
根据一种实施方式,上述图神经网络为图卷积网络,所述图卷积网络包含多个网络层以进行所述第二处理,所述第二处理至少包括,使用所述邻接矩阵中的元素为权重因子,对所述节点及其邻居节点的特征向量进行加权求和操作。According to an embodiment, the graph neural network described above is a graph convolutional network. The graph convolutional network includes a plurality of network layers to perform the second processing. The second processing includes at least using the adjacency matrix. The element is a weighting factor, which performs a weighted sum operation on the feature vectors of the node and its neighbors.
在一种实施方式中,步骤54中对所述第一输出和所述第二输出进行融合具体包括,对所述第一输出和所述第二输出进行加权求和,其中所述第一输出对应第一权重因子,所述第二输出对应第二权重因子。In one embodiment, the fusing the first output and the second output in step 54 specifically includes weighting and summing the first output and the second output, wherein the first output Corresponding to the first weighting factor, the second output corresponds to the second weighting factor.
进一步的,在一个实施例中,上述第一权重因子是第一输出的函数,第二权重因子是第二输出的函数。Further, in an embodiment, the first weighting factor is a function of the first output, and the second weighting factor is a function of the second output.
在另一实施例中,第一权重因子与第二权重因子和为1,并且:In another embodiment, the sum of the first weighting factor and the second weighting factor is 1, and:
第一权重因子是第一输出的函数;或者,The first weighting factor is a function of the first output; or,
第二权重因子是第二输出的函数。The second weighting factor is a function of the second output.
或者,在又一实施例中,第一权重因子与第二权重因子和为1,并且:Or, in yet another embodiment, the sum of the first weighting factor and the second weighting factor is 1, and:
第一权重因子是第一输出和第二输出的函数;或者,The first weighting factor is a function of the first output and the second output; or,
第二权重因子是第一输出和第二输出的函数。The second weighting factor is a function of the first output and the second output.
通过以上的方法,结合深度神经网络对单节点特征的处理,以及图神经网络对节点关系特征的处理,综合地对关系网络图进行分析。Through the above methods, combined with the deep neural network processing of single-node features and the graph neural network processing of node relationship features, the relationship network graph is comprehensively analyzed.
根据另一方面的实施例,还提供一种计算机可读存储介质,其上存储有计算机程序,当所述计算机程序在计算机中执行时,令计算机执行结合图5所描述的方法。According to an embodiment of another aspect, there is also provided a computer-readable storage medium on which a computer program is stored, and when the computer program is executed in a computer, the computer is caused to perform the method described in conjunction with FIG. 5.
根据再一方面的实施例,还提供一种计算设备,包括存储器和处理器,所述存储器中存储有可执行代码,所述处理器执行所述可执行代码时,实现结合图5所述的方法。According to an embodiment of still another aspect, a computing device is further provided, including a memory and a processor, where executable code is stored in the memory, and when the processor executes the executable code, the implementation described in conjunction with FIG. 5 is implemented method.
本领域技术人员应该可以意识到,在上述一个或多个示例中,本申请所描述的功能可以用硬件、软件、固件或它们的任意组合来实现。当使用软件实现时,可以将这些功能存储在计算机可读介质中或者作为计算机可读介质上的一个或多个指令或代码进行传输。Those skilled in the art should be aware that in one or more of the above examples, the functions described in this application may be implemented by hardware, software, firmware, or any combination thereof. When implemented in software, these functions can be stored in a computer-readable medium or transmitted as one or more instructions or code on a computer-readable medium.
以上所述的具体实施方式,对本申请的目的、技术方案和有益效果进行了进一步详细说明,所应理解的是,以上所述仅为本申请的具体实施方式而已,并不用于限定本申请的保护范围,凡在本申请的技术方案的基础之上,所做的任何修改、等同替换、改进等,均应包括在本申请的保护范围之内。The specific embodiments described above further describe the purpose, technical solutions, and beneficial effects of this application in detail. It should be understood that the above descriptions are only specific implementations of this application, and are not intended to limit the scope of this application. The scope of protection, any modifications, equivalent replacements, improvements, etc. made on the basis of the technical solutions of this application, shall be included in the scope of protection of this application.

Claims (20)

  1. 一种通过计算机执行的、用于分析关系网络图的神经网络系统,包括:A neural network system executed by a computer and used to analyze a relational network graph, including:
    特征提取层,用于提取关系网络图中的节点的特征向量;The feature extraction layer is used to extract the feature vectors of the nodes in the relationship network graph;
    深度神经网络,用于对所述特征向量进行第一处理,得到第一输出;A deep neural network, which is used to perform first processing on the feature vector to obtain a first output;
    图神经网络,用于结合所述关系网络图的邻接信息,对所述特征向量进行第二处理,得到第二输出;其中所述邻接信息用于表示所述关系网络图所包含的各个节点之间的连接关系;A graph neural network, used to combine the adjacency information of the relational network graph to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent each of the nodes included in the relational network graph Connection between
    融合层,用于对所述第一输出和所述第二输出进行融合,基于融合结果输出针对所述节点的预测结果。A fusion layer is used to fuse the first output and the second output, and output a prediction result for the node based on the fusion result.
  2. 根据权利要求1所述的神经网络系统,其中,所述关系网络图所包含的各个节点对应于各个用户,所述各个节点之间的连接关系包括以下中的一种或多种:用户之间的社交关系、媒介关系和资金关系。The neural network system according to claim 1, wherein each node included in the relational network graph corresponds to each user, and the connection relationship between the various nodes includes one or more of the following: between users Social relations, media relations and financial relations.
  3. 根据权利要求1所述的神经网络系统,其中,所述关系网络图为有向图,所述邻接信息包括,所述有向图对应的邻接表或十字链表。The neural network system according to claim 1, wherein the relational network graph is a directed graph, and the adjacency information includes an adjacency list or a cross list corresponding to the directed graph.
  4. 根据权利要求1所述的神经网络系统,其中所述邻接信息包括所述关系网络图的邻接矩阵。The neural network system according to claim 1, wherein the adjacency information includes an adjacency matrix of the relational network graph.
  5. 根据权利要求4所述的神经网络系统,其中,所述图神经网络为图卷积网络,所述图卷积网络包含多个网络层以进行所述第二处理,所述第二处理至少包括,使用所述邻接矩阵中的元素为权重因子,对所述节点及其邻居节点的特征向量进行加权求和操作。The neural network system according to claim 4, wherein the graph neural network is a graph convolutional network, the graph convolutional network includes a plurality of network layers to perform the second processing, the second processing includes at least , Using the elements in the adjacency matrix as a weighting factor to perform a weighted sum operation on the feature vectors of the node and its neighbors.
  6. 根据权利要求1所述的神经网络系统,所述融合层具体用于,对所述第一输出和所述第二输出进行加权求和,其中所述第一输出对应第一权重因子,所述第二输出对应第二权重因子。The neural network system according to claim 1, the fusion layer is specifically used for weighted summation of the first output and the second output, wherein the first output corresponds to a first weighting factor, the The second output corresponds to the second weighting factor.
  7. 根据权利要求6所述的神经网络系统,其中,所述第一权重因子是所述第一输出的函数,所述第二权重因子是所述第二输出的函数。The neural network system of claim 6, wherein the first weighting factor is a function of the first output and the second weighting factor is a function of the second output.
  8. 根据权利要求6所述的神经网络系统,其中,所述第一权重因子与第二权重因子和为1,并且:The neural network system according to claim 6, wherein the sum of the first weighting factor and the second weighting factor is 1, and:
    所述第一权重因子是所述第一输出的函数;或者,The first weighting factor is a function of the first output; or,
    所述第二权重因子是所述第二输出的函数。The second weighting factor is a function of the second output.
  9. 根据权利要求6所述的神经网络系统,其中,所述第一权重因子与第二权重因子和为1,并且:The neural network system according to claim 6, wherein the sum of the first weighting factor and the second weighting factor is 1, and:
    所述第一权重因子是所述第一输出和第二输出的函数;或者,The first weighting factor is a function of the first output and the second output; or,
    所述第二权重因子是所述第一输出和第二输出的函数。The second weighting factor is a function of the first output and the second output.
  10. 根据权利要求1所述的神经网络系统,所述神经网络系统通过端到端方式训练。The neural network system according to claim 1, the neural network system being trained in an end-to-end manner.
  11. 一种通过计算机执行的、用于分析关系网络图的方法,包括:A computer-implemented method for analyzing relational network diagrams, including:
    提取关系网络图中的节点的特征向量;Extract the feature vectors of the nodes in the relationship network graph;
    采用深度神经网络,对所述特征向量进行第一处理,得到第一输出;A deep neural network is used to perform the first processing on the feature vector to obtain a first output;
    采用图神经网络,结合所述关系网络图的邻接信息,对所述特征向量进行第二处理,得到第二输出;其中所述邻接信息用于表示所述关系网络图所包含的各个节点之间的连接关系;A graph neural network is used, combined with the adjacency information of the relational network graph, to perform a second processing on the feature vector to obtain a second output; wherein the adjacency information is used to represent the relationship between each node included in the relational network graph Connection
    对所述第一输出和所述第二输出进行融合,基于融合结果输出针对所述节点的预测结果。The first output and the second output are fused, and a prediction result for the node is output based on the fusion result.
  12. 根据权利要求11所述的方法,其中,所述关系网络图所包含的各个节点对应于各个用户,所述各个节点之间的连接关系包括以下中的一种或多种:用户之间的社交关系、媒介关系和资金关系。The method according to claim 11, wherein each node included in the relationship network graph corresponds to each user, and the connection relationship between the various nodes includes one or more of the following: social between users Relations, media relations and funding relations.
  13. 根据权利要求11所述的方法,其中,所述关系网络图为有向图,所述邻接信息包括,所述有向图对应的邻接表或十字链表。The method according to claim 11, wherein the relationship network graph is a directed graph, and the adjacency information includes an adjacency list or a cross-linked list corresponding to the directed graph.
  14. 根据权利要求11所述的方法,其中,所述邻接信息包括所述关系网络图的邻接矩阵。The method according to claim 11, wherein the adjacency information includes an adjacency matrix of the relational network graph.
  15. 根据权利要求14所述的方法,其中,所述图神经网络为图卷积网络,所述图卷积网络包含多个网络层以进行所述第二处理,所述第二处理至少包括,使用所述邻接矩阵中的元素为权重因子,对所述节点及其邻居节点的特征向量进行加权求和操作。The method according to claim 14, wherein the graph neural network is a graph convolutional network, the graph convolutional network includes a plurality of network layers to perform the second processing, the second processing includes at least, using The elements in the adjacency matrix are weighting factors, and a weighted sum operation is performed on the feature vectors of the node and its neighbors.
  16. 根据权利要求11所述的方法,其中,所述对所述第一输出和所述第二输出进行融合包括,对所述第一输出和所述第二输出进行加权求和,其中所述第一输出对应第一权重因子,所述第二输出对应第二权重因子。The method of claim 11, wherein the fusing the first output and the second output includes weighted summing the first output and the second output, wherein the first One output corresponds to the first weighting factor, and the second output corresponds to the second weighting factor.
  17. 根据权利要求16所述的方法,其中,所述第一权重因子是所述第一输出的函数,所述第二权重因子是所述第二输出的函数。The method of claim 16, wherein the first weighting factor is a function of the first output and the second weighting factor is a function of the second output.
  18. 根据权利要求16所述的方法,其中,所述第一权重因子与第二权重因子和为1,并且:The method according to claim 16, wherein the sum of the first weighting factor and the second weighting factor is 1, and:
    所述第一权重因子是所述第一输出的函数;或者,The first weighting factor is a function of the first output; or,
    所述第二权重因子是所述第二输出的函数。The second weighting factor is a function of the second output.
  19. 根据权利要求16所述的方法,其中,所述第一权重因子与第二权重因子和为1,并且:The method according to claim 16, wherein the sum of the first weighting factor and the second weighting factor is 1, and:
    所述第一权重因子是所述第一输出和第二输出的函数;或者,The first weighting factor is a function of the first output and the second output; or,
    所述第二权重因子是所述第一输出和第二输出的函数。The second weighting factor is a function of the first output and the second output.
  20. 一种计算设备,包括存储器和处理器,其特征在于,所述存储器中存储有可执行代码,所述处理器执行所述可执行代码时,实现权利要求11-19中任一项所述的方法。A computing device, including a memory and a processor, characterized in that executable code is stored in the memory, and when the processor executes the executable code, it implements any one of claims 11-19 method.
PCT/CN2019/112564 2018-12-07 2019-10-22 Neural network system and method for analyzing relationship network graph WO2020114122A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811497595.0A CN110009093B (en) 2018-12-07 2018-12-07 Neural network system and method for analyzing relational network graph
CN201811497595.0 2018-12-07

Publications (1)

Publication Number Publication Date
WO2020114122A1 true WO2020114122A1 (en) 2020-06-11

Family

ID=67165074

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/112564 WO2020114122A1 (en) 2018-12-07 2019-10-22 Neural network system and method for analyzing relationship network graph

Country Status (3)

Country Link
CN (1) CN110009093B (en)
TW (1) TWI709086B (en)
WO (1) WO2020114122A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112085104A (en) * 2020-09-10 2020-12-15 杭州中奥科技有限公司 Event feature extraction method and device, storage medium and electronic equipment
CN112819152A (en) * 2020-08-14 2021-05-18 腾讯科技(深圳)有限公司 Neural network training method and device
CN113377656A (en) * 2021-06-16 2021-09-10 南京大学 Crowd-sourcing recommendation method based on graph neural network
CN113792089A (en) * 2021-09-16 2021-12-14 平安银行股份有限公司 Illegal behavior detection method, device, equipment and medium based on artificial intelligence
CN114169515A (en) * 2020-08-20 2022-03-11 四川大学 Social relationship identification method based on high-order graph neural network
CN114818973A (en) * 2021-07-15 2022-07-29 支付宝(杭州)信息技术有限公司 Method, device and equipment for training graph model based on privacy protection
US20220407879A1 (en) * 2020-10-16 2022-12-22 Visa International Service Association System, method, and computer program product for user network activity anomaly detection
CN116258504A (en) * 2023-03-16 2023-06-13 广州信瑞泰信息科技有限公司 Bank customer relationship management system and method thereof
WO2023165352A1 (en) * 2022-03-03 2023-09-07 百果园技术(新加坡)有限公司 Service object classification method and apparatus, device and storage medium
WO2024021738A1 (en) * 2022-07-29 2024-02-01 腾讯科技(深圳)有限公司 Data network graph embedding method and apparatus, computer device, and storage medium

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110009093B (en) * 2018-12-07 2020-08-07 阿里巴巴集团控股有限公司 Neural network system and method for analyzing relational network graph
CN110555469B (en) * 2019-08-15 2020-07-24 阿里巴巴集团控股有限公司 Method and device for processing interactive sequence data
US10936950B1 (en) 2019-08-15 2021-03-02 Advanced New Technologies Co., Ltd. Processing sequential interaction data
CN110659723B (en) * 2019-09-03 2023-09-19 腾讯科技(深圳)有限公司 Data processing method and device based on artificial intelligence, medium and electronic equipment
CN110705613B (en) * 2019-09-19 2021-06-11 创新奇智(青岛)科技有限公司 Object classification method
CN110705709B (en) * 2019-10-14 2021-03-23 支付宝(杭州)信息技术有限公司 Method and device for training neural network model of graph
CN110490274B (en) * 2019-10-17 2020-02-07 支付宝(杭州)信息技术有限公司 Method and device for evaluating interaction event
CN111079780B (en) * 2019-11-06 2023-06-23 中国科学院深圳先进技术研究院 Training method for space diagram convolution network, electronic equipment and storage medium
CN111027610B (en) * 2019-12-03 2022-02-25 腾讯医疗健康(深圳)有限公司 Image feature fusion method, apparatus, and medium
CN110991914B (en) * 2019-12-09 2024-04-16 朱递 Facility site selection method based on graph convolution neural network
CN115081589A (en) * 2020-01-09 2022-09-20 支付宝(杭州)信息技术有限公司 Method and device for processing interactive data by using LSTM neural network model
CN111210279B (en) * 2020-01-09 2022-08-16 支付宝(杭州)信息技术有限公司 Target user prediction method and device and electronic equipment
CN111277433B (en) * 2020-01-15 2021-02-12 同济大学 Network service abnormity detection method and device based on attribute network characterization learning
CN111414989B (en) * 2020-02-13 2023-11-07 山东师范大学 User trust relationship network link prediction method and system based on gating mechanism
CN111340611B (en) * 2020-02-20 2024-03-08 中国建设银行股份有限公司 Risk early warning method and device
CN111581488B (en) * 2020-05-14 2023-08-04 上海商汤智能科技有限公司 Data processing method and device, electronic equipment and storage medium
CN111798934B (en) * 2020-06-23 2023-11-14 苏州浦意智能医疗科技有限公司 Molecular property prediction method based on graph neural network
CN112036418A (en) * 2020-09-04 2020-12-04 京东数字科技控股股份有限公司 Method and device for extracting user features
CN113536383B (en) * 2021-01-27 2023-10-27 支付宝(杭州)信息技术有限公司 Method and device for training graph neural network based on privacy protection
CN112801288A (en) * 2021-02-05 2021-05-14 厦门市美亚柏科信息股份有限公司 Vector representation method and device of graph network
CN112766500B (en) * 2021-02-07 2022-05-17 支付宝(杭州)信息技术有限公司 Method and device for training graph neural network
CN113470828A (en) * 2021-06-30 2021-10-01 上海商汤智能科技有限公司 Classification method and device, electronic equipment and storage medium
CN113408706B (en) * 2021-07-01 2022-04-12 支付宝(杭州)信息技术有限公司 Method and device for training user interest mining model and user interest mining
CN113835899B (en) * 2021-11-25 2022-02-22 支付宝(杭州)信息技术有限公司 Data fusion method and device for distributed graph learning
CN114677234B (en) * 2022-04-26 2024-04-30 河南大学 Graph convolution neural network social recommendation method and system integrating multichannel attention mechanisms
CN115545189B (en) * 2022-11-29 2023-04-18 支付宝(杭州)信息技术有限公司 Method and device for generating network of training diagram and method and device for generating neural network of training diagram
CN115994373B (en) * 2023-03-22 2023-05-30 山东中联翰元教育科技有限公司 College entrance examination volunteer filling system data encryption method based on big data

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106708922A (en) * 2016-10-21 2017-05-24 天津海量信息技术股份有限公司 Character relation atlas analysis method based on mass data
CN107145977A (en) * 2017-04-28 2017-09-08 电子科技大学 A kind of method that structured attributes deduction is carried out to online social network user
CN107943967A (en) * 2017-11-28 2018-04-20 华南理工大学 Algorithm of documents categorization based on multi-angle convolutional neural networks and Recognition with Recurrent Neural Network
CN108733792A (en) * 2018-05-14 2018-11-02 北京大学深圳研究生院 A kind of entity relation extraction method
US20180341719A1 (en) * 2017-05-24 2018-11-29 International Business Machines Corporation Neural Bit Embeddings for Graphs
CN110009093A (en) * 2018-12-07 2019-07-12 阿里巴巴集团控股有限公司 For analyzing the nerve network system and method for relational network figure

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10536357B2 (en) * 2015-06-05 2020-01-14 Cisco Technology, Inc. Late data detection in data center
JP6719727B2 (en) * 2016-03-23 2020-07-08 富士ゼロックス株式会社 Purchase behavior analysis device and program
US10489887B2 (en) * 2017-04-10 2019-11-26 Samsung Electronics Co., Ltd. System and method for deep learning image super resolution
CN107808168B (en) * 2017-10-31 2023-08-01 北京科技大学 Social network user behavior prediction method based on strong and weak relation
CN108921566B (en) * 2018-05-03 2021-11-05 创新先进技术有限公司 False transaction identification method and device based on graph structure model
CN108874914B (en) * 2018-05-29 2021-11-02 吉林大学 Information recommendation method based on graph convolution and neural collaborative filtering
CN108805203A (en) * 2018-06-11 2018-11-13 腾讯科技(深圳)有限公司 Image procossing and object recognition methods, device, equipment and storage medium again

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106708922A (en) * 2016-10-21 2017-05-24 天津海量信息技术股份有限公司 Character relation atlas analysis method based on mass data
CN107145977A (en) * 2017-04-28 2017-09-08 电子科技大学 A kind of method that structured attributes deduction is carried out to online social network user
US20180341719A1 (en) * 2017-05-24 2018-11-29 International Business Machines Corporation Neural Bit Embeddings for Graphs
CN107943967A (en) * 2017-11-28 2018-04-20 华南理工大学 Algorithm of documents categorization based on multi-angle convolutional neural networks and Recognition with Recurrent Neural Network
CN108733792A (en) * 2018-05-14 2018-11-02 北京大学深圳研究生院 A kind of entity relation extraction method
CN110009093A (en) * 2018-12-07 2019-07-12 阿里巴巴集团控股有限公司 For analyzing the nerve network system and method for relational network figure

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112819152A (en) * 2020-08-14 2021-05-18 腾讯科技(深圳)有限公司 Neural network training method and device
CN112819152B (en) * 2020-08-14 2024-03-01 腾讯科技(深圳)有限公司 Neural network training method and device
CN114169515A (en) * 2020-08-20 2022-03-11 四川大学 Social relationship identification method based on high-order graph neural network
CN114169515B (en) * 2020-08-20 2023-04-28 四川大学 Social relationship identification method based on high-order graph neural network
CN112085104B (en) * 2020-09-10 2024-04-12 杭州中奥科技有限公司 Event feature extraction method and device, storage medium and electronic equipment
CN112085104A (en) * 2020-09-10 2020-12-15 杭州中奥科技有限公司 Event feature extraction method and device, storage medium and electronic equipment
US11711391B2 (en) * 2020-10-16 2023-07-25 Visa International Service Association System, method, and computer program product for user network activity anomaly detection
US20220407879A1 (en) * 2020-10-16 2022-12-22 Visa International Service Association System, method, and computer program product for user network activity anomaly detection
CN113377656B (en) * 2021-06-16 2023-06-23 南京大学 Public testing recommendation method based on graph neural network
CN113377656A (en) * 2021-06-16 2021-09-10 南京大学 Crowd-sourcing recommendation method based on graph neural network
CN114818973A (en) * 2021-07-15 2022-07-29 支付宝(杭州)信息技术有限公司 Method, device and equipment for training graph model based on privacy protection
CN113792089B (en) * 2021-09-16 2024-03-22 平安银行股份有限公司 Illegal behavior detection method, device, equipment and medium based on artificial intelligence
CN113792089A (en) * 2021-09-16 2021-12-14 平安银行股份有限公司 Illegal behavior detection method, device, equipment and medium based on artificial intelligence
WO2023165352A1 (en) * 2022-03-03 2023-09-07 百果园技术(新加坡)有限公司 Service object classification method and apparatus, device and storage medium
WO2024021738A1 (en) * 2022-07-29 2024-02-01 腾讯科技(深圳)有限公司 Data network graph embedding method and apparatus, computer device, and storage medium
CN116258504A (en) * 2023-03-16 2023-06-13 广州信瑞泰信息科技有限公司 Bank customer relationship management system and method thereof

Also Published As

Publication number Publication date
TW202032422A (en) 2020-09-01
CN110009093B (en) 2020-08-07
CN110009093A (en) 2019-07-12
TWI709086B (en) 2020-11-01

Similar Documents

Publication Publication Date Title
TWI709086B (en) Neural network system and method for analyzing relationship network graph
Bellotti et al. Forecasting recovery rates on non-performing loans with machine learning
CN109102393B (en) Method and device for training and using relational network embedded model
Byanjankar et al. Predicting credit risk in peer-to-peer lending: A neural network approach
WO2019196546A1 (en) Method and apparatus for determining risk probability of service request event
Dahiya et al. A feature selection enabled hybrid‐bagging algorithm for credit risk evaluation
US20100169137A1 (en) Methods and systems to analyze data using a graph
EP4236197A2 (en) Micro-loan system
Suryanarayana et al. Machine learning approaches for credit card fraud detection
Yotsawat et al. A novel method for credit scoring based on cost-sensitive neural network ensemble
WO2023109085A1 (en) Method for training account risk model, and method for determining risk user group
US11823026B2 (en) Artificial intelligence system employing graph convolutional networks for analyzing multi-entity-type multi-relational data
Aphale et al. Predict loan approval in banking system machine learning approach for cooperative banks loan approval
Eddy et al. Credit scoring models: Techniques and issues
Byanjankar et al. Data‐driven optimization of peer‐to‐peer lending portfolios based on the expected value framework
Bouzidi et al. Deep learning-based automated learning environment using smart data to improve corporate marketing, business strategies, fraud detection in financial services, and financial time series forecasting
Orji et al. Machine learning models for predicting bank loan eligibility
Zhou Data mining for individual consumer credit default prediction under e-commence context: a comparative study
Zhou et al. FinBrain 2.0: when finance meets trustworthy AI
CN114240659A (en) Block chain abnormal node identification method based on dynamic graph convolutional neural network
Rahman et al. Nearest neighbor classifier method for making loan decision in commercial bank
Zhou et al. Forecasting credit default risk with graph attention networks
Belhaj Slimene et al. NeuroEvolution of Augmenting Topologies for predicting financial distress: A multicriteria decision analysis
Ye et al. Supply Chain Finance Assistance for Small and Medium-Sized Enterprises Using Cognitive Web Services.
US20240119517A1 (en) Artificial intelligence based methods and systems for predicting creditworthiness of merchants

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19893494

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19893494

Country of ref document: EP

Kind code of ref document: A1