WO2021179838A1 - Prediction method and system based on heterogeneous graph neural network model - Google Patents

Prediction method and system based on heterogeneous graph neural network model Download PDF

Info

Publication number
WO2021179838A1
WO2021179838A1 PCT/CN2021/074479 CN2021074479W WO2021179838A1 WO 2021179838 A1 WO2021179838 A1 WO 2021179838A1 CN 2021074479 W CN2021074479 W CN 2021074479W WO 2021179838 A1 WO2021179838 A1 WO 2021179838A1
Authority
WO
WIPO (PCT)
Prior art keywords
node
predicted
path
heterogeneous graph
prediction
Prior art date
Application number
PCT/CN2021/074479
Other languages
French (fr)
Chinese (zh)
Inventor
胡斌斌
张志强
周俊
杨双红
Original Assignee
支付宝(杭州)信息技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 支付宝(杭州)信息技术有限公司 filed Critical 支付宝(杭州)信息技术有限公司
Publication of WO2021179838A1 publication Critical patent/WO2021179838A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/901Indexing; Data structures therefor; Storage structures
    • G06F16/9024Graphs; Linked lists
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/906Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q40/00Finance; Insurance; Tax strategies; Processing of corporate or income taxes
    • G06Q40/03Credit; Loans; Processing thereof

Definitions

  • One or more embodiments of this specification relate to the field of data processing, and in particular to a method and system for prediction based on a heterogeneous graph neural network model.
  • Heterogeneous graph also known as heterogeneous information network, is a special graph structure that usually contains different types of nodes and different types of paths, and can be used to represent some complex social network relationship data. Nodes in a heterogeneous graph can be used to represent entity objects such as individual users. Usually, a heterogeneous graph containing entity object data can be input into a graph neural network to predict nodes in a heterogeneous graph, for example, users can be predicted to make judgments. Its category, risk level or preference habits. Heterogeneous graphs exhibit high complexity due to their different node types and path types. Therefore, how to predict the entity object data in the heterogeneous graphs is very important.
  • this specification proposes a method and system for prediction based on a heterogeneous graph neural network model.
  • the method for judging the category of an entity object through entity object data includes: obtaining heterogeneous graph data related to predicted content, the heterogeneous graph data including a node to be predicted, neighbor nodes of the node to be predicted, and connecting to the node to be predicted.
  • Predict a path between a node and the neighbor node the path includes at least one type; group the neighbor nodes based on the type of the path, so that the types of paths of the neighbor nodes in the same group are the same Input the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model, to obtain the representation vector of the node to be predicted, and then input the trained prediction model for prediction.
  • the system for judging the category of an entity object based on entity object data includes: a heterogeneous graph acquisition module for acquiring heterogeneous graph data related to predicted content, and the heterogeneous graph data includes the node to be predicted and the information of the node to be predicted.
  • a neighbor node and a path connecting the node to be predicted and the neighbor node, the path includes at least one type; a grouping module is configured to group the neighbor nodes based on the type of the path to Make the path types of the neighbor nodes in the same group the same; a node prediction module for inputting the nodes to be predicted, the neighbor nodes after grouping, and the paths between the nodes into the trained heterogeneous graph neural network model , Get the node to be predicted.
  • One of the embodiments of this specification provides an apparatus for predicting based on a heterogeneous graph neural network model, which includes a processor configured to execute a method for judging the category of an entity object through entity object data.
  • One of the embodiments of this specification provides a computer-readable storage medium that stores computer instructions. After the computer reads the computer instructions in the storage medium, the computer executes a prediction method based on a heterogeneous graph neural network model.
  • Fig. 1 is a schematic diagram of a scenario application of prediction based on a heterogeneous graph neural network model according to some embodiments of this specification;
  • Fig. 2 is a block diagram of a prediction system based on a heterogeneous graph neural network model according to some embodiments of this specification;
  • Fig. 3 is an exemplary flowchart of a method for prediction based on a heterogeneous graph neural network model according to some embodiments of the present specification
  • Fig. 4 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to some embodiments of the present specification
  • Fig. 5 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to some embodiments of this specification.
  • Fig. 6 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to other embodiments of this specification.
  • system is a method for distinguishing different components, elements, parts, parts, or assemblies of different levels.
  • the words can be replaced by other expressions.
  • Fig. 1 is a schematic diagram of a scenario application of prediction based on a heterogeneous graph neural network model according to some embodiments of this specification.
  • Heterogeneous graphs include relational networks describing different types of paths for N nodes.
  • nodes refer to physical objects, such as users.
  • the path represents the connection relationship between nodes.
  • entity objects and paths please refer to Figure 3, which will not be repeated here.
  • FIG. 1 For ease of description, only a part of the heterogeneous graph is shown in FIG. 1.
  • different types of paths may be processed separately to obtain the group aggregation vector of the entity objects in each group of paths; then, according to the importance (weight) corresponding to the node to be predicted and each group of paths, These group aggregation vectors are merged to obtain a representation vector, and finally the representation result is output based on the representation vector.
  • the network of each group of different paths is processed separately to obtain each group's aggregation vector, which can avoid tedious manual feature extraction Further, it is possible to automatically determine the importance (weight) of different types of path relationship networks under the current prediction content, to achieve information fusion in each group of paths, so as to make the evaluation results of the predicted nodes more accurate.
  • the entity object is the user AH
  • the path type can include the interaction between users (indicated by a two-way arrow), loan method (indicated by a solid line), and income level (indicated by a dotted line).
  • the predicted content is that the user is in The level of risk in financial lending (such as the probability of default).
  • the characteristic information of user C and user F in the network can be processed according to the network of the interactive relationship path of user A to obtain the group aggregation vector for user A.
  • the characteristic information of user G and users B and H can be processed separately according to the network of the loan mode path and the income level path to obtain two other group aggregation vectors for user A. Then, the importance of each group of users in different paths relative to user A is determined according to the aggregation vector representation of each group. Then, the group aggregation vectors corresponding to different paths are merged based on different degrees of importance to obtain a representation vector for user A. According to the representation vector, the user's financial loan risk level can be output. According to the risk level, follow-up services can be carried out, such as limiting the user's loan amount, prohibiting the user from lending business, etc.
  • Entity objects can include users A, C, D, F, movies B, H, and director G; path types can include interaction paths between users (indicated by two-way arrows), user satisfaction paths with movies (indicated by dashed lines) ), the path of the user's satisfaction with the director (indicated by a solid line).
  • the prediction content can be whether the user prefers a certain movie, or it can be the rating of the movie or the director.
  • the characteristic information of movie B and movie H in the network can be processed according to the network of user A's movie satisfaction path to obtain a group aggregation vector for user A.
  • the characteristic information of user G and directors B and H can be processed according to the network of interaction path and director satisfaction path, and two other group aggregation vectors for user A can be obtained.
  • the importance of each group of users in different paths relative to user A is determined according to the aggregation vector representation of each group.
  • the group aggregation vectors corresponding to different paths are merged based on different degrees of importance to obtain a representation vector for user A.
  • the user A's preference for a certain movie is predicted.
  • predicting the score of director G may be based on the network of users' satisfaction with the movie, and based on users A, D, and F in the network.
  • Fig. 2 is a block diagram of a system for predicting based on a heterogeneous graph neural network model according to some embodiments of this specification.
  • the system 200 for judging the entity object category based on entity object data may include a heterogeneous graph acquisition module 210, a grouping module 220, and a node prediction module 230.
  • the heterogeneous graph acquisition module 210 can acquire heterogeneous graph data related to the predicted content, the heterogeneous graph data including the node to be predicted, the neighbor nodes of the node to be predicted, and the connection between the node to be predicted and the neighbor node
  • the path between, the path includes at least one type.
  • the grouping module 220 may be configured to group the neighbor nodes based on the types of the paths, so that the types of paths of the neighbor nodes in the same group are the same. For a detailed description of the grouping module 220, refer to step 304 in FIG. 3, which will not be repeated here.
  • the node prediction module 230 may be used to input the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model, obtain the representation vector of the node to be predicted, and then input the trained Predictive models make predictions.
  • the node prediction module is further used to obtain the representation vector of the node to be predicted after fusing the node attention weight and/or the path attention weight.
  • the node prediction module 230 may obtain the group aggregation vector corresponding to the group one by one according to the characteristic information of the neighbor nodes of the same group; for each group aggregation vector, the characteristic information of the node to be predicted Fusion, get the node information to be predicted that is grouped and fused.
  • the node prediction module is further configured to determine the attention weight of the path based on the importance of the path; determine the node to be predicted based on the information of the node to be predicted fused by the group and the attention weight of the path The representation vector.
  • the node prediction module 230 may determine the attention weight vector of the node based on the importance of neighbor nodes, and determine the attention weight vector of the path based on the importance of the path. In some embodiments, the node prediction module 230 may determine the representation vector of the node to be predicted based on the grouped fusion of node information to be predicted, the attention weight vector of the node, and the attention weight vector of the path. For a detailed description of the node prediction module 230, refer to step 306 in FIG. 3, which is not repeated here.
  • the system 100 further includes: a training module for end-to-end training of the heterogeneous graph network neural model and prediction model of the system, specifically the iterative update of the prediction model based on the loss function of the prediction model and the different Compose the model parameters of the neural network model until the iteration cut-off condition is met.
  • a training module for end-to-end training of the heterogeneous graph network neural model and prediction model of the system, specifically the iterative update of the prediction model based on the loss function of the prediction model and the different Compose the model parameters of the neural network model until the iteration cut-off condition is met.
  • several heterogeneous graph data can be used as training data, and the correct result of the node corresponding to the heterogeneous graph data can be used as the label data of the training data.
  • the parameters of the prediction model and the heterogeneous graph neural network The parameters of the model are updated through training iterations using the training data and the label data. Specifically, you can participate in the training process of the heterogen
  • system and its modules shown in FIG. 2 can be implemented in various ways.
  • the system and its modules may be implemented by hardware, software, or a combination of software and hardware.
  • the hardware part can be implemented using dedicated logic;
  • the software part can be stored in a memory and executed by an appropriate instruction execution system, such as a microprocessor or dedicated design hardware.
  • processor control codes for example on a carrier medium such as a disk, CD or DVD-ROM, such as a read-only memory (firmware Such codes are provided on a programmable memory or a data carrier such as an optical or electronic signal carrier.
  • the system and its modules in this specification can not only be implemented by hardware circuits such as very large-scale integrated circuits or gate arrays, semiconductors such as logic chips, transistors, etc., or programmable hardware devices such as field programmable gate arrays, programmable logic devices, etc. It may also be implemented by software executed by various types of processors, or may be implemented by a combination of the foregoing hardware circuit and software (for example, firmware).
  • the above description of the candidate item display, determination system and its modules is only for convenience of description, and does not limit this specification to the scope of the examples mentioned. It can be understood that for those skilled in the art, after understanding the principle of the system, it is possible to arbitrarily combine various modules, or form a subsystem to connect with other modules without departing from this principle.
  • the heterogeneous graph acquisition module 210 and the grouping module 220 may be two modules, or one module may have both heterogeneous graph acquisition and grouping functions.
  • each module may share a storage module, and each module may also have its own storage module. Such deformations are all within the protection scope of this specification.
  • Fig. 3 is an exemplary flowchart of a method for prediction based on a heterogeneous graph neural network model according to some embodiments of the present specification. As shown in FIG. 3, the method 300 for judging the category of an entity object based on entity object data includes:
  • Step 302 Obtain heterogeneous graph data related to the predicted content.
  • the heterogeneous graph data includes a node to be predicted, neighbor nodes of the node to be predicted, and a path connecting the node to be predicted and the neighbor node, and the path includes at least one type.
  • step 302 may be performed by the heterogeneous graph acquisition module 210.
  • Heterogeneous graph data is a data relationship graph including different types of nodes and paths between different types of nodes. For example, a graph of user viewing data. Another example is a graph of user financial relationship data.
  • Nodes refer to physical objects.
  • the types of nodes may be the same.
  • nodes can all be users.
  • the types of nodes can be different.
  • nodes can include users, directors, actors, and so on.
  • the path represents the connection relationship between nodes.
  • different types of paths may represent different connection relationships between nodes.
  • connection relationship may be a connection relationship between entity objects of the same type.
  • connection relationship may be an interaction relationship between entity objects of the same type (such as attention, likes, and comments between users).
  • the path can be expressed as: User A—Like—User B.
  • the connection relationship may also be a common preference between entity objects of the same type. For example, the viewing methods that users commonly prefer (such as online viewing on Youku, viewing in movie theaters).
  • the path may be expressed as: user A—youku online viewing—user B.
  • connection relationship may also be a relationship between different types of entity objects.
  • entity object director K such as like, insensitivity, dislike
  • the path can be expressed as: user A-likes-director K.
  • the heterogeneous graph data may be obtained through data stored in a social network, a review website, a credit information website, etc. authorized by the user, and may also be obtained by calling related interfaces or other methods.
  • This embodiment does not limit the acquisition method.
  • the entity object-user can be obtained through the registration information authorized by the social network, and the path can be obtained through the interactive behavior between the users of the social network.
  • the content to be predicted may be unknown data of the physical object.
  • the prediction content of the prediction model may include the category, risk level, or preference habits of the entity object corresponding to the node to be predicted.
  • the category of the entity object may include the user's income level category (poor, well-off, and rich), director evaluation category (poor, fair, good, very good), and so on.
  • the risk level of the entity object may include the user's financial lending risk level (for example, the probability of default is low, the probability of default is medium, the probability of default is high), the health risk level of the applicant (for example, the probability of compensation is low, The possibility of compensation is medium, the probability of compensation is high), etc.
  • the preference habits of the entity object may include the user's preferred movie type (eg, comedy, action, science fiction film), and the user's preferred loan repayment method (eg, equal principal and interest, equal principal).
  • the user's preferred movie type eg, comedy, action, science fiction film
  • the user's preferred loan repayment method eg, equal principal and interest, equal principal
  • the node to be predicted is a node that outputs the content to be predicted. It is understandable that the node to be predicted can be selected according to the content to be predicted. Take the user’s preferred movie type as an example.
  • the entity object includes a heterogeneous graph of users, directors, and actors. If the content to be predicted is the type of movie that each user likes, the node to be predicted may be the user; the content to be predicted is the rating of each director , The node to be predicted can be the director. For further explanation, the following is an example of the prediction content of the user's loan risk.
  • the neighbor nodes are directly connected to the node to be predicted through the path.
  • a prediction node may be connected to multiple neighbor nodes through multiple paths.
  • the types of multiple paths may be the same or different. For example, if user A of the node to be predicted “likes" 1 user and "comments" 2 users, the number of neighbor nodes of the interactive path is 3; user A of the node to be predicted and the other 2 users both use "Youku” To watch movies online, and to watch movies in cinemas with 3 other users at the same time, the number of neighbor nodes of the path of the movie viewing mode is 5.
  • an application scenario of user loan risk is taken as another embodiment for description.
  • the corresponding node to be predicted is the user, and the content to be predicted is the loan risk level of each user.
  • Path types can include interactions between users, user’s lending methods, income levels, etc.
  • the corresponding paths can be exemplarily corresponding to: user A-comment-user D, user A-microfinance company-user B, user D -Middle class-User C.
  • Step 304 Group the neighboring nodes based on the types of the paths, so that the types of paths of the neighboring nodes in the same group are the same.
  • step 304 may be performed by the grouping module 220.
  • neighbor nodes are classified according to the type of path connecting the node to be predicted and its neighbor nodes. Following the above example, the neighbor nodes connected by the interactive path are grouped into a group, and the neighbor nodes connected by the path of the viewing mode are grouped into a group.
  • each group of neighbor nodes corresponds to a graph neural network based on the path connection structure of the node to be predicted and the neighbor nodes therein.
  • Step 306 Input the node to be predicted, the grouped neighbor nodes, and the path between the nodes into the trained heterogeneous graph neural network model to obtain the representation vector of the node to be predicted and then input the trained prediction model to make predictions .
  • step 306 may be performed by the node prediction module 230.
  • the characteristic information of the node to be predicted, the grouped neighbor nodes, and the path between the nodes are input into the trained heterogeneous graph neural network model, and the output is the representation vector of the node to be predicted.
  • the characteristic information of a node is data that can characterize the characteristics of an entity object.
  • the characteristic information of the path is data that can characterize the characteristic of the path.
  • the measurement result is the value of the content to be predicted.
  • the prediction content of the prediction model may include the category, risk level, or preference habits of the entity object corresponding to the node to be predicted.
  • the prediction model may be a machine learning model with classification capabilities, such as a binary classification model, a logistic regression model, or a neural network, which is not specifically limited in this specification.
  • each output value is used to characterize the confidence of the category corresponding to the output value, that is, the model's Forecast probability. For example, there are three output values, 0.6, 0.2, 0.2, respectively, corresponding to the probability that the user is a comedy fan, an action fan, and a science fiction fan. Since the probability of a comedy fan is 0.6, it can be considered that Users belong to comedy fans.
  • Fig. 4 is an exemplary sub-flow chart of a method for prediction based on a heterogeneous graph neural network model according to some embodiments of the present specification.
  • the heterogeneous graph neural network model may include a group aggregation vector layer, a node information layer to be predicted, and a representation vector layer.
  • the method 400 for inputting the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into a trained heterogeneous graph neural network model to obtain a representation vector of the node to be predicted includes step 402 ⁇ Step 404.
  • Step 402 According to the feature information of the neighbor nodes in the same group, obtain the group aggregation vector corresponding to the group one by one.
  • the group aggregation vector layer may be used to aggregate the input feature information of neighbor nodes of the same group to generate a group aggregation vector.
  • the group aggregation vector layer can first vectorize the feature information of the neighbor nodes to obtain the feature vector X; the feature vectors of the same group of neighbor nodes are then aggregated through the graph neural network to obtain the feature aggregation vector h; The group aggregation vector H is obtained based on the feature aggregation vector.
  • the characteristic information of a node is data that can characterize the characteristics of an entity object.
  • the entity object is a user
  • the characteristic information of the entity object may include the user's identity information data (such as age, occupation), and may also include the user's historical preference data (such as favorite movies, TV series, etc.).
  • the feature information of the entity object can be obtained through the authorization data entered by the user during registration, or through the actual operation of the user (such as likes, comments, etc.), and it can also be obtained by reading stored data and calling related Obtained via interface or other methods. This embodiment does not limit the acquisition method.
  • the feature information may be vectorized through the vectorized representation model in advance to obtain the corresponding feature vector.
  • the vectorized representation model may be a word embedding model, and the word embedding model may include but is not limited to: Word2vec model, term frequency-inverse document frequency model (Term Frequency-Inverse Document Frequency, TF-IDF) or SSWE-C (skip-gram based combined-sentiment word embedding) model, etc.
  • the feature aggregation vector h can be obtained by performing feature aggregation on the feature vector X of each neighbor node. It can be understood that a group of neighbor nodes corresponds to a feature aggregation vector h.
  • neighbor nodes within a predetermined order (such as order 2) of the node to be predicted can be used as feature aggregation nodes, or neighbors within a predetermined order can be sampled, and the neighbor nodes obtained by sampling can be used for feature aggregation.
  • the feature aggregation may aggregate the feature vector X through a preset aggregation method; it may also aggregate the feature vector X using the attention mechanism based on the graph neural network corresponding to each group of neighbor nodes; The feature vector X of each neighbor node can be convolved through the graph neural network.
  • the pre-set aggregation method may be to use a pre-trained parameter matrix to add, average, take the maximum value, calculate the weighted sum, etc., on the feature vector X of the neighbor nodes, etc., which will not be used here. limited.
  • the feature vector X u of the node to be predicted is multiplied by the first parameter matrix to obtain a dimension of 1*5
  • the first parameter matrix can be regarded as a weighted sum operation for each element of the feature vector X u of the node to be predicted, and then the feature information of the node to be predicted is aggregated, so that the node to be predicted
  • the elements of the feature aggregation vector h u can represent richer information.
  • the neighbor node feature vector shown in Figure 5 For example, if its dimension is 1*10, add the feature vectors of multiple neighbor nodes and divide by 3 to obtain an average vector with a dimension of 1*10, and then multiply the average vector by a dimension of 10*5 The second parameter matrix of, obtains the feature aggregation vector of neighbor nodes with dimension 1*5 In the same way, the feature aggregation vector of neighbor nodes can be obtained
  • the second parameter matrix may be different.
  • the heterogeneous graph neural network model can be trained in advance. The training process can be referred to the description elsewhere in this article.
  • the graph neural network can have one or more layers. In some embodiments, there may also be multiple sets of model parameters corresponding to each layer of graph neural network.
  • Attention Mechanism is a deep learning technology that enables graph neural networks to have the ability to focus on a subset of their inputs (or features).
  • the neighbor attention weight (ie, neighbor weight vector) of the neural network can be obtained through the attention mechanism; for multi-layer graph neural networks, the feature attention weight of the hidden layer (ie feature Weight vector), which in turn makes the more important feature vectors in neighbor nodes to be used to a higher degree, and reduces the interference of noise information.
  • each group of model parameters is a weight matrix
  • a layer of graph neural network may correspond to multiple weight matrices.
  • the model parameters can be determined through parameter adjustment during the training process.
  • the weight matrix may be obtained based on the importance of each neighbor node.
  • the importance of each neighbor node can be determined by the feature vector of the node to be predicted and the neighbor node.
  • the neighbor weight vector can be calculated by formula (1). For example, there are N nodes in the j-th group of neighbor nodes of the node u to be predicted. For the node u to be predicted, the neighbor weight of the neighbor node k can be:
  • the matrices V (for example, called the first weight matrix) and W 1 (for example, called the second weight matrix) are the model parameters determined during the training process of the graph neural network
  • b 1 is the constant parameter determined during the training process of the graph neural network
  • X u and X k are the current feature expression vector corresponding to node u and node k respectively.
  • the activation functions softmax and tanh can also be replaced by other activation functions (such as Relu, etc.), which are not limited here.
  • the corresponding neighbor weight can be determined for each neighbor node.
  • the neighbor weight for the corresponding neighbor node is also different. It is worth noting that the node u to be predicted can also be regarded as its own neighbor node, for example, it is called a zero-order neighbor node.
  • the feature aggregation of each neighbor node can be performed by methods such as weighted sum.
  • N u represents the j th group of neighbor nodes of node u set to be predicted, to be predicted neighbor nodes after node u neighbor neural network of FIG polymerization results are:
  • an aggregation result of the current layer can be obtained, for example, the aggregation result of node k (also called a feature expression vector) is
  • the above neighbor aggregation result may be further aggregated with the feature expression vector of the node to be predicted to obtain the aggregation result of the node to be predicted in the current layer of the graph neural network.
  • node 1, node 2, node 3...Node k... are neighbor nodes of the node u to be predicted, and aggregate their corresponding features in the i-1th layer (i ⁇ 2).
  • the feature aggregation result corresponding to the node u to be predicted in the i-1th layer is recorded as Then in the i-th layer, the current feature expression vector of the corresponding node is the feature aggregation result of the i-1th layer (that is, the feature expression vector output by the i-1th layer).
  • the neighbor nodes of node u are aggregated to obtain the neighbor aggregation result Then, the feature aggregation results of the upper network node u As feature vector with Perform aggregation to get the feature expression vector of the node u to be predicted in the i-th layer Therefore, in the graph neural network of the j-th group of neighbor nodes, through the iterative processing of the pre-trained graph neural network, a corresponding feature aggregation vector can finally be obtained.
  • each feature expression vector may also have a feature importance (feature weight).
  • the hidden layer of the graph neural network can determine the feature weight vector formed by the feature weights corresponding to each feature expression vector in the following manner:
  • W 2 (for example, called the third weight matrix) and W 3 (for example, called the fourth weight matrix) are the weight matrices of the i-th layer in the graph neural network, and b 2 and b 3 are constant parameters.
  • W 2 , W 3 , b 2 , and b 3 are constant parameters.
  • the final feature aggregation result in the upper layer of the neural network of the neighbor node u to be predicted can be used Sure.
  • the excitation function Relu can also be replaced by other suitable excitation functions, which will not be repeated here.
  • Each element in the feature weight vector ⁇ i corresponds to the feature weight of each feature.
  • the feature aggregation result of the current node u in the current layer can be obtained.
  • the way to determine the final aggregation result according to the feature weight can be expressed as:
  • means multiplying the corresponding elements of two matrices (such as Hadamard product).
  • the k-th element in ⁇ i is the same as As a result of aggregation
  • the result of the vector (A, B, C) ⁇ (a, b, c) is (Aa, Bb, Cc).
  • the node contribution degree and feature contribution degree can be considered at the same time, and a more accurate feature aggregation result of neighbor nodes can be obtained.
  • the feature aggregation model is a graph neural network
  • the aggregation result obtained in the last layer is the feature aggregation vector used by the j-th group of neighbor node pairs.
  • the feature aggregation vector may also be obtained by convolution based on the graph neural network. Specifically, one or more layers of convolution can be performed on the feature vector of each neighbor node, and then the output convolution results of each neighbor node can be summed, averaged, maximum value, weighted sum, etc. , It is not limited here.
  • the intermediate vector of the neighbor node v in the l-th layer of the graph neural network can be:
  • W l can be a model parameter in the form of a matrix, which can be called a weight matrix.
  • the formula can also consider the feature aggregation of higher-order neighbor nodes of the current node, which is represented by an ellipsis here. The principle is similar to the feature aggregation of first-order neighbor nodes, and will not be repeated here. Among them, different neighbor nodes have different normalization factors and different feature expression vectors, so the product multiplied by the weight matrix is also different, so they have different neighbor weights.
  • the intermediate vector output by the last layer of the graph convolutional neural network is the convolution result of node v.
  • the feature information of neighbor nodes corresponding to different types of paths are quite different, the feature information of neighbor nodes corresponding to the same type of path is relatively close.
  • the neighbor nodes are classified based on the type of path and the information is aggregated to make the final representation learning vector The information contained is richer.
  • feature aggregation is not limited to the above three methods, and other methods can also be used.
  • the feature aggregation vector can be directly used as the group aggregation vector for further processing.
  • the feature aggregation vector can also be reduced in dimension. Dimensionality reduction can further perform feature extraction on the feature aggregation vector, and at the same time make the subsequent steps more efficient.
  • a feature aggregation vector with a dimension of 1 ⁇ 3N Reduce dimensionality to 1 ⁇ 5 dimensional group aggregation vector For example, a feature aggregation vector with a dimension of 1 ⁇ 3N Reduce dimensionality to 1 ⁇ 5 dimensional group aggregation vector.
  • m groups of neighbor nodes can get m group aggregation vectors
  • the above-mentioned dimensionality reduction methods may include, but are not limited to: principal component analysis (PCA), linear discriminant analysis (LDA), multidimensional scaling (MDS), partial Locally Linear Embedding (LLE), Adjacency Map (ISOMAP, Isometric Feature Mapping), and Kernel Principle Component Analysis (KPCA), etc.
  • PCA principal component analysis
  • LDA linear discriminant analysis
  • MDS multidimensional scaling
  • LLE partial Locally Linear Embedding
  • ISOMAP Isometric Feature Mapping
  • KPCA Kernel Principle Component Analysis
  • Step 404 For each group of aggregation vectors, the feature information of the node to be predicted is fused to obtain grouped and fused information of the node to be predicted.
  • the to-be-predicted node information layer may be used to fuse the input feature information of the to-be-predicted node and the group aggregation vector, and output the grouped and fused information of the to-be-predicted node.
  • the grouped and fused information of the node to be predicted is a vector that combines the feature information of the node to be predicted and the feature information of the neighbor node (ie, the adjacent node fusion vector).
  • the information layer of the node to be predicted may vectorize the feature information of the node to be predicted.
  • the dimensionality of the vectorized feature information of the node to be predicted may be reduced to obtain the feature vector of the node to be predicted. For example, H u .
  • the vectorization and dimensionality reduction methods of the feature information of the node to be predicted may refer to step 402, which will not be repeated here.
  • the feature vector of the node to be predicted may be spliced with each group aggregation vector obtained in step 402 to obtain the corresponding group aggregation splicing vector.
  • group aggregation splicing vector For example, H u and Perform splicing to obtain m group aggregation splicing vectors.
  • the splicing method may be direct splicing or polymerization in step 402, which is not limited in this embodiment.
  • the splicing vector may be aggregated based on each group to obtain the corresponding node information to be predicted (ie, the adjacent node fusion vector).
  • the splicing vector may be aggregated based on each group to obtain the corresponding node information to be predicted (ie, the adjacent node fusion vector).
  • the grouped and fused node information to be predicted is input to the next layer of processing of the heterogeneous graph neural network model.
  • the next layer of the heterogeneous graph neural network model represents the vector layer.
  • the information of the node to be predicted can be fused with at least one of the node attention weight or the path attention weight to obtain the final representation learning vector.
  • the node attention weight and The path attention weights are all fused, and for example, in Figure 6, only the path attention weights are fused.
  • the node attention weight and the path attention weight may not be merged, and the final representation vector is obtained directly based on the information of the node to be predicted.
  • Fig. 5 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to some embodiments of the application.
  • the representation vector layer may output the representation vector of the node to be predicted based on the input node information to be predicted, the node attention weight, and the path attention weight.
  • the presentation vector layer can first calculate the attention weight vector of the node through the group aggregation vector of the node to be predicted and the adjacent node fusion vector; then obtain the corresponding attention weight vector based on the node's attention weight vector and the adjacent node fusion vector Information weighted fusion vector; then the attention weight vector of the path is calculated by the information weighted fusion vector and the feature vector of the path; finally, the representation vector of the node to be predicted is obtained from the information weighted fusion vector and the attention weight vector of the path.
  • Step 502 Determine the attention weight vector of the node based on the importance of neighbor nodes.
  • the attention weight vector of neighbor nodes is a vector whose elements are the importance of each group of neighbor nodes.
  • the importance of neighbor nodes represents the degree to which the feature information of each group of neighbor nodes is used in the process of calculating the representation vector of the node to be predicted for different nodes to be predicted.
  • the path importance of the user interaction relationship type is low, and the path importance of the user income level type is high.
  • the attention weight vector of each group of neighbor nodes can be determined by the feature vector of the node to be predicted and the neighbor node fusion vector of each group of neighbor nodes, including: The adjacent node fusion vector is spliced with the feature vector of the node to be predicted, and the vector obtained based on the splicing is multiplied by the preset third weight matrix to obtain the attention weight vector of the node corresponding to the adjacent node fusion vector .
  • the attention weight vector of the node can refer to the way of obtaining the neighbor weight in step 402, which will not be repeated here.
  • Step 504 Obtain a corresponding information weighted fusion vector based on the attention weight vector of the node and the adjacent node fusion vector.
  • the adjacent node fusion vector and the node attention weight vector have the same dimension, and each element of the node attention weight vector is used to represent the same position of the adjacent node fusion vector. The importance of the element.
  • the information weight corresponding to the adjacent node fusion vector is obtained based on the node attention weight vector corresponding to the adjacent node fusion vector and the adjacent node fusion vector
  • the fusion vector includes: for each adjacent node fusion vector, multiplying each element of the adjacent node fusion vector by the element at the same position of the node attention weight vector corresponding to the adjacent node fusion vector, And use the obtained product as the element at the same position of the information weighted fusion vector to obtain the information weighted fusion vector.
  • the fusion vector of adjacent nodes with dimension 1*6 Is [A1, B1, C1, D1, E1, F1], and its corresponding dimension is 1*6 node attention weight vector It is [A2, B2, C2, D2, E2, F2], in a possible embodiment, A2 is used to characterize the importance of A1, B2 is used to characterize the importance of B1, and the rest of the elements are analogously. Then based on the multiplication between the above elements, the fusion vector corresponding to the adjacent node is obtained Information weighted fusion vector It is [A1*A2, B1*B2, C1*C2, D1*D2, E1*E2, F1*F2]. Similarly, the information weighted fusion vector It can be obtained based on the same method.
  • Step 506 Determine the attention weight vector of the path based on the importance of the path.
  • the attention weight vector of the path is obtained based on the information weighted fusion vector.
  • the method for obtaining the attention weight vector of the path may refer to the obtaining of the feature weight vector in step 402.
  • a fourth weight matrix with a dimension of 6*1 to obtain the path attention weight ⁇ j corresponding to the information weighted fusion vector.
  • other path attention weights are obtained based on the same method as described above: ⁇ 1 , ⁇ 2 ,... ⁇ m .
  • Combining all the path attentions can obtain the path attention weight vector ⁇ u as shown in Fig. 5.
  • the parameter vectors of different information weighted fusion vectors can be the same, and the parameter vectors can be obtained by training the model corresponding to the method for judging the entity object category through the entity object data.
  • Step 508 Determine the representation vector of the node to be predicted based on the grouped fusion of the node information to be predicted, the attention weight vector of the node, and the attention weight vector of the path.
  • obtaining the information representation learning vector of the node to be predicted based on the information weighted fusion vector and the path attention weight corresponding to the information weighted fusion vector includes: weighting each information fusion vector Multiply it with the path attention weight ⁇ j in the information-weighted fusion vector ⁇ u , and add all the vectors obtained by the multiplication to obtain the information representation learning vector e u , such as:
  • the attention weight vector of the corresponding path is ⁇ 1 , ⁇ 2 , and the obtained information indicates that the learning vector e u is
  • the information can be directly represented by the learning vector e u for related calculations, for example, directly input into the prediction model for calculation.
  • path attention weight Because of the integration of path attention weight, node attention weight, and characteristic information of neighbor nodes and nodes to be predicted, this information indicates that the learning vector can express richer information and make subsequent prediction results more accurate.
  • the attention weight of the node may not be merged, and the corresponding representation learning vector e u is obtained based on the path attention weight and the feature information of the node to be predicted, ie, see Fig. 6 and Fig. 6
  • Step 602 and step 604 obtain the node information to be predicted based on the grouped neighbor nodes.
  • steps 402 and 404 respectively, which will not be repeated here.
  • Steps 606 and 608 respectively determine the attention weight of the path based on the importance of the path, and determine the representation learning vector e u of the node to be predicted based on the above-mentioned node information to be predicted and the attention weight of the path.
  • the specific method can be seen in Figure 5 Steps 506 and 508 are not repeated here.
  • end-to-end training may be performed on the heterogeneous graph network model and the prediction model based on a large number of training samples with identifications.
  • the end-to-end training refers to multiple models in the training process, according to the model's data processing steps, input data from the input end of the first model, and obtain the result from the output end of the last model, based on the result and the true value The error of all models are adjusted iteratively until the model meets the cut-off condition.
  • the end-to-end training can save the training data required for training each independent model, and at the same time, the training results between the models will not affect each other.
  • the training samples with the logo may be input into the heterogeneous graph network model, and then the representation vector output by the heterogeneous network model is input into the prediction model, and the loss function is constructed based on the output value of the prediction model. Iteratively update the parameters of the heterogeneous graph network model and the prediction model.
  • the training sample may be several pieces of heterogeneous graph data related to the target object (for example, the user's credit rating; or the user's movie preference type), and each piece of heterogeneous graph data includes related node data, Adjacent node data and path information between each node.
  • the heterogeneous graph data can be constructed based on acquired historical data information, for example, based on acquired user personal data, user preference data for movie types, preference data for movie actors, preference data for movie directors, and the like.
  • each piece of heterogeneous graph data in the training sample is labeled, and the labeled data includes the evaluation result of the target content in each piece of heterogeneous graph data (for example, the user’s credit rating is good; or the user likes The movie genre is comedy movie).
  • the evaluation result may also be determined by obtaining historical evaluation information of the target object (for example, the user).
  • the heterogeneous graph data with sample identification will be input into the heterogeneous graph network model
  • the heterogeneous graph data will be input into the heterogeneous graph network model.
  • the representation vector output by the graph network model is used as the input data of the prediction model
  • the labeled data corresponding to the heterogeneous graph data is used as the output data of the prediction model
  • the input data and output data are input to the prediction model for training.
  • a loss function can be constructed based on the actual output value of the prediction model, and the parameters of the heterogeneous graph network model hybrid prediction model can be iteratively updated through the loss function.
  • the parameters of the heterogeneous network model may include a first parameter matrix, a second parameter matrix, a first weight matrix, a second weight matrix, etc.
  • the parameters of a prediction model may include a binary classification model, a logistic regression model, and a neural network. The weights, thresholds, etc. of the network model.
  • the training ends.
  • the iterative cut-off condition can be that the result of the loss function converges or is less than a threshold.
  • the beneficial effects that the embodiments of this specification may bring include, but are not limited to: (1) The method for judging the entity object category through entity object data combines the feature information of the node to be predicted and the neighbor node, as well as the node attention weight and path attention The weight fully extracts all aspects of the information in the heterogeneous graph, so that the obtained information indicates that the information contained in the learning vector is richer, and the prediction result of the input prediction model is more accurate; (2) For the same type of path, different The feature information of neighbor nodes effectively extracts and utilizes the structural information in the heterogeneous graph, so that the obtained information learning representation vector contains richer semantic information. It should be noted that different embodiments may have different beneficial effects. In different embodiments, the possible beneficial effects may be any one or a combination of the above, or any other beneficial effects that may be obtained.
  • the computer storage medium may contain a propagated data signal containing a computer program code, for example on a baseband or as part of a carrier wave.
  • the propagated signal may have multiple manifestations, including electromagnetic forms, optical forms, etc., or a suitable combination.
  • the computer storage medium may be any computer readable medium other than the computer readable storage medium, and the medium may be connected to an instruction execution system, device, or device to realize communication, propagation, or transmission of the program for use.
  • the program code located on the computer storage medium can be transmitted through any suitable medium, including radio, cable, fiber optic cable, RF, or similar medium, or any combination of the above medium.
  • the computer program codes required for the operation of each part of this manual can be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB.NET, Python Etc., conventional programming languages such as C language, Visual Basic, Fortran 2003, Perl, COBOL 2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages.
  • the program code can be run entirely on the user's computer, or run as an independent software package on the user's computer, or partly run on the user's computer and partly run on a remote computer, or run entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any network form, such as a local area network (LAN) or a wide area network (WAN), or connected to an external computer (for example, via the Internet), or in a cloud computing environment, or as a service Use software as a service (SaaS).
  • LAN local area network
  • WAN wide area network
  • SaaS service Use software as a service
  • numbers describing the number of ingredients and attributes are used. It should be understood that such numbers used in the description of the embodiments use the modifier "about”, “approximately” or “substantially” in some examples. Retouch. Unless otherwise stated, “approximately”, “approximately” or “substantially” indicates that the number is allowed to vary by ⁇ 20%.
  • the numerical parameters used in the specification and claims are approximate values, and the approximate values can be changed according to the required characteristics of individual embodiments. In some embodiments, the numerical parameter should consider the prescribed effective digits and adopt the method of general digit retention. Although the numerical ranges and parameters used to confirm the breadth of the ranges in some embodiments of this specification are approximate values, in specific embodiments, the setting of such numerical values is as accurate as possible within the feasible range.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Data Mining & Analysis (AREA)
  • Finance (AREA)
  • Accounting & Taxation (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Development Economics (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Strategic Management (AREA)
  • Technology Law (AREA)
  • General Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

A prediction method and a system based on a heterogeneous graph neural network model. The method comprises: acquiring heterogeneous graph data related to prediction content, wherein the heterogeneous graph data comprises a node to be subjected to prediction, neighbor nodes of the node to be subjected to prediction, and paths connected between the node to be subjected to prediction and the neighbor nodes, wherein the paths comprise at least one type (302); grouping the neighbor nodes on the basis of the type of the paths, such that the types of paths of the neighbor nodes of the same group are the same (304); and inputting the node to be subjected to prediction, the grouped neighbor nodes, and the paths between the nodes into a trained heterogeneous graph neural network model, so as to obtain a representation vector of the node to be subjected to prediction, and then inputting the representation vector into a trained prediction model for prediction (306).

Description

一种基于异构图神经网络模型进行预测的方法和系统Method and system for prediction based on heterogeneous graph neural network model 技术领域Technical field
本说明书一个或多个实施例涉及数据处理领域,特别涉及一种基于异构图神经网络模型进行预测的方法和系统。One or more embodiments of this specification relate to the field of data processing, and in particular to a method and system for prediction based on a heterogeneous graph neural network model.
背景技术Background technique
异构图,又称作异质信息网络,是一种特殊的图结构,其通常包含不同类型的节点以及不同类型的路径,可以用于表示一些复杂的社交网络关系数据。异构图中的节点可以用于表示实体对象例如个人用户,通常可以将包含实体对象数据的异构图输入图神经网络来对异构图中的节点进行预测,例如对用户进行预测预测以判断其类别、风险等级或者偏好习惯。而异构图由于其节点类型以及路径类型的不同,体现出较高的复杂性,因此,如何基于异构图对其中的实体对象数据进行预测至关重要。Heterogeneous graph, also known as heterogeneous information network, is a special graph structure that usually contains different types of nodes and different types of paths, and can be used to represent some complex social network relationship data. Nodes in a heterogeneous graph can be used to represent entity objects such as individual users. Usually, a heterogeneous graph containing entity object data can be input into a graph neural network to predict nodes in a heterogeneous graph, for example, users can be predicted to make judgments. Its category, risk level or preference habits. Heterogeneous graphs exhibit high complexity due to their different node types and path types. Therefore, how to predict the entity object data in the heterogeneous graphs is very important.
基于此,本说明书提出一种基于异构图神经网络模型进行预测的方法和系统。Based on this, this specification proposes a method and system for prediction based on a heterogeneous graph neural network model.
发明内容Summary of the invention
本说明书实施例之一提供一种通过实体对象数据判断实体对象类别的方法。所述通过实体对象数据判断实体对象类别的方法包括:获取与预测内容相关的异构图数据,所述异构图数据包括待预测节点、所述待预测节点的邻居节点、以及连接所述待预测节点与所述邻居节点之间的路径,所述路径包括至少一种类型;基于所述路径的类型,对所述邻居节点进行分组,以使得同一组的所述邻居节点的路径的类型相同;将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量后输入训练好的预测模型进行预测。One of the embodiments of this specification provides a method for judging the type of an entity object through entity object data. The method for judging the category of an entity object through entity object data includes: obtaining heterogeneous graph data related to predicted content, the heterogeneous graph data including a node to be predicted, neighbor nodes of the node to be predicted, and connecting to the node to be predicted. Predict a path between a node and the neighbor node, the path includes at least one type; group the neighbor nodes based on the type of the path, so that the types of paths of the neighbor nodes in the same group are the same Input the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model, to obtain the representation vector of the node to be predicted, and then input the trained prediction model for prediction.
本说明书实施例之一提供一种基于异构图神经网络模型进行预测的系统。所述通过实体对象数据判断实体对象类别的系统包括:异构图获取模块,用于获取与预测内容相关的异构图数据,所述异构图数据包括待预测节点、所述待预测节点的邻居节点、以及连接所述待预测节点与所述邻居节点之间的路径,所述路径包括至少一种类型;分组模块,用于基于所述路径的类型,对所述邻居节点进行分组,以使得同一组的所述邻居节点的路径的类型相同;节点预测模块,用于将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点。One of the embodiments of this specification provides a prediction system based on a heterogeneous graph neural network model. The system for judging the category of an entity object based on entity object data includes: a heterogeneous graph acquisition module for acquiring heterogeneous graph data related to predicted content, and the heterogeneous graph data includes the node to be predicted and the information of the node to be predicted. A neighbor node and a path connecting the node to be predicted and the neighbor node, the path includes at least one type; a grouping module is configured to group the neighbor nodes based on the type of the path to Make the path types of the neighbor nodes in the same group the same; a node prediction module for inputting the nodes to be predicted, the neighbor nodes after grouping, and the paths between the nodes into the trained heterogeneous graph neural network model , Get the node to be predicted.
本说明书实施例之一提供一种基于异构图神经网络模型进行预测的装置,包括处理器,所述处理器用于执行通过实体对象数据判断实体对象类别的方法。One of the embodiments of this specification provides an apparatus for predicting based on a heterogeneous graph neural network model, which includes a processor configured to execute a method for judging the category of an entity object through entity object data.
本说明书实施例之一提供一种计算机可读存储介质,所述存储介质存储计算机指令,当计算机读取存储介质中的计算机指令后,计算机执行基于异构图神经网络模型进行预测的方法。One of the embodiments of this specification provides a computer-readable storage medium that stores computer instructions. After the computer reads the computer instructions in the storage medium, the computer executes a prediction method based on a heterogeneous graph neural network model.
附图说明Description of the drawings
本说明书将以示例性实施例的方式进一步说明,这些示例性实施例将通过附图进行详细描述。这些实施例并非限制性的,在这些实施例中,相同的编号表示相同的结构,其中:This specification will be further described in the form of exemplary embodiments, and these exemplary embodiments will be described in detail with the accompanying drawings. These embodiments are not restrictive. In these embodiments, the same number represents the same structure, in which:
图1是根据本说明书一些实施例所示的一种基于异构图神经网络模型进行预测的场景应用示意图;Fig. 1 is a schematic diagram of a scenario application of prediction based on a heterogeneous graph neural network model according to some embodiments of this specification;
图2是根据本说明书一些实施例所示的一种基于异构图神经网络模型进行预测的系统的模块图;Fig. 2 is a block diagram of a prediction system based on a heterogeneous graph neural network model according to some embodiments of this specification;
图3是根据本说明书一些实施例所示的基于异构图神经网络模型进行预测的方法的示例性流程图;Fig. 3 is an exemplary flowchart of a method for prediction based on a heterogeneous graph neural network model according to some embodiments of the present specification;
图4是根据本说明书一些实施例所示的基于异构图神经网络模型进行预测方法的示例性子流程图;Fig. 4 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to some embodiments of the present specification;
图5是根据本说明书一些实施例所示的基于异构图神经网络模型进行预测方法的示例性子流程图;以及Fig. 5 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to some embodiments of this specification; and
图6是根据本说明书另一些实施例所示的基于异构图神经网络模型进行预测方法的示例性子流程图。Fig. 6 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to other embodiments of this specification.
具体实施方式Detailed ways
为了更清楚地说明本说明书实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单的介绍。显而易见地,下面描述中的附图仅仅是本说明书的一些示例或实施例,对于本领域的普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图将本说明书应用于其它类似情景。除非从语言环境中显而易见或另做说明,图中相同标号代表相同结构或操作。In order to more clearly describe the technical solutions of the embodiments of the present specification, the following will briefly introduce the accompanying drawings used in the description of the embodiments. Obviously, the drawings in the following description are only some examples or embodiments of this specification. For those of ordinary skill in the art, without creative work, this specification can also be applied to these drawings. Other similar scenarios. Unless it is obvious from the language environment or otherwise stated, the same reference numerals in the figures represent the same structure or operation.
应当理解,本文使用的“系统”、“装置”、“单元”和/或“模组”是用于区分不同级别的不同组件、元件、部件、部分或装配的一种方法。然而,如果其他词语可实现相同的目的,则可通过其他表达来替换所述词语。It should be understood that the “system”, “device”, “unit” and/or “module” used herein is a method for distinguishing different components, elements, parts, parts, or assemblies of different levels. However, if other words can achieve the same purpose, the words can be replaced by other expressions.
如本说明书和权利要求书中所示,除非上下文明确提示例外情形,“一”、“一个”、“一种”和/或“该”等词并非特指单数,也可包括复数。一般说来,术语“包括”与“包含”仅提示包括已明确标识的步骤和元素,而这些步骤和元素不构成一个排它性的罗列,方法或者设备也可能包含其它的步骤或元素。As shown in this specification and claims, unless the context clearly indicates exceptions, the words "a", "an", "an" and/or "the" do not specifically refer to the singular, but may also include the plural. Generally speaking, the terms "include" and "include" only suggest that the clearly identified steps and elements are included, and these steps and elements do not constitute an exclusive list, and the method or device may also include other steps or elements.
本说明书中使用了流程图用来说明根据本说明书的实施例的系统所执行的操作。应当理解的是,前面或后面操作不一定按照顺序来精确地执行。相反,可以按照倒序或同时处理各个步骤。同时,也可以将其他操作添加到这些过程中,或从这些过程移除某一步或数步操作。In this specification, a flowchart is used to illustrate the operations performed by the system according to the embodiment of this specification. It should be understood that the preceding or following operations are not necessarily performed exactly in order. Instead, the steps can be processed in reverse order or at the same time. At the same time, other operations can be added to these processes, or a certain step or several operations can be removed from these processes.
图1是根据本说明书一些实施例所示的基于异构图神经网络模型进行预测的场景应用示意图。Fig. 1 is a schematic diagram of a scenario application of prediction based on a heterogeneous graph neural network model according to some embodiments of this specification.
异构图包括针对N个节点的不同类型的路径进行描述的关系网络。其中,节点是指实体对象,例如用户。路径表示节点之间的连接关系。关于实体对象和路径的更多详细描述可以参见图3,在此不再赘述。Heterogeneous graphs include relational networks describing different types of paths for N nodes. Among them, nodes refer to physical objects, such as users. The path represents the connection relationship between nodes. For more detailed descriptions of entity objects and paths, please refer to Figure 3, which will not be repeated here.
为了便于说明,图1中只示出了异构图的一部分。在一些实施例中,可以先针对不同类型的路径分别进行处理,得到实体对象分别在各组路径中的组聚合向量;然后, 根据待预测节点和各组路径分别对应的重要程度(权重),对这些组聚合向量进行融合,从而得到一个表示向量,最后基于表示向量输出表示结果。一方面,由于利用了多个不同类型的路径,可以更加全面的刻画实体对象的特征;另一方面,先针对各组不同路径的网络分别处理得到各个组聚合向量,可以避免繁琐的手工特征抽取;进一步地,可以自动确定当前预测内容下,不同类型路径关系网络的重要程度(权重),实现在各组路径中的信息融合,从而使得对待预测节点的评估结果更加准确。For ease of description, only a part of the heterogeneous graph is shown in FIG. 1. In some embodiments, different types of paths may be processed separately to obtain the group aggregation vector of the entity objects in each group of paths; then, according to the importance (weight) corresponding to the node to be predicted and each group of paths, These group aggregation vectors are merged to obtain a representation vector, and finally the representation result is output based on the representation vector. On the one hand, due to the use of multiple different types of paths, the characteristics of the entity object can be more comprehensively described; on the other hand, the network of each group of different paths is processed separately to obtain each group's aggregation vector, which can avoid tedious manual feature extraction Further, it is possible to automatically determine the importance (weight) of different types of path relationship networks under the current prediction content, to achieve information fusion in each group of paths, so as to make the evaluation results of the predicted nodes more accurate.
以用户金融关系数据图预测用户金融风险等级为应用场景进行说明。在此应用场景下,实体对象为用户A-H,路径类型可以包括用户之间交互关系(以双向箭头表示)、借贷方式(以实线表示)和收入等级(以虚线表示),预测内容为用户在金融借贷时的风险等级(如违约概率)。示例性地,预测用户A的金融风险等级,可以根据用户A的交互关系路径的网络,对网络中用户C和用户F的特征信息进行处理,得到针对用户A的组聚合向量。同理,可以根据借贷方式路径和收入等级路径的网络,分别对其中的用户G和用户B、H的特征信息进行处理,得到另外两个针对用户A的组聚合向量。然后,根据各个组聚合向量表征确定不同路径内的各组用户相对于用户A的重要程度。接着,基于不同重要程度对不同路径对应的组聚合向量进行融合,得到针对用户A的表示向量。根据表示向量,可以输出用户的金融借贷风险等级。根据风险等级,可以进行后续业务,如限制该用户的借贷金额、禁止用户进行借贷业务等。Use the user financial relationship data graph to predict the user's financial risk level as an application scenario for description. In this application scenario, the entity object is the user AH, and the path type can include the interaction between users (indicated by a two-way arrow), loan method (indicated by a solid line), and income level (indicated by a dotted line). The predicted content is that the user is in The level of risk in financial lending (such as the probability of default). Exemplarily, to predict the financial risk level of user A, the characteristic information of user C and user F in the network can be processed according to the network of the interactive relationship path of user A to obtain the group aggregation vector for user A. In the same way, the characteristic information of user G and users B and H can be processed separately according to the network of the loan mode path and the income level path to obtain two other group aggregation vectors for user A. Then, the importance of each group of users in different paths relative to user A is determined according to the aggregation vector representation of each group. Then, the group aggregation vectors corresponding to different paths are merged based on different degrees of importance to obtain a representation vector for user A. According to the representation vector, the user's financial loan risk level can be output. According to the risk level, follow-up services can be carried out, such as limiting the user's loan amount, prohibiting the user from lending business, etc.
以用户观影数据关系图为另一实施例,说明包含不同类型实体对象的异构图应用场景。实体对象可以包括用户A、C、D、F,电影B、H,以及导演G;路径类型可以包括用户之间的交互路径(以双向箭头表示)、用户对电影的满意度路径(以虚线表示)、用户对导演的满意度路径(以实线表示)。预测内容可以是用户是否偏好某部电影,也可以是电影或导演的评分。示例性的,预测用户A是否偏好某部电影,可以根据用户A的电影满意度路径的网络,对网络中电影B和电影H的特征信息进行处理,得到针对用户A的组聚合向量。同理,可以根据交互路径和导演满意度路径的网络,分别对其中的用户G和导演B、H的特征信息进行处理,得到另外两个针对用户A的组聚合向量。然后,根据各个组聚合向量表征确定不同路径内的各组用户相对于用户A的重要程度。接着,基于不同重要程度对不同路径对应的组聚合向量进行融合,得到针对用户A的表示向量。根据表示向量和某部电影的特征信息,预测用户A对于某部电影的偏好。又一示例性的,预测导演G的评分,可以根据用户对电影的满意度路径的网络,基于网络中的用户A、D、F来进行预测。Taking the user viewing data relationship diagram as another embodiment, an application scenario of heterogeneous diagrams containing different types of entity objects is described. Entity objects can include users A, C, D, F, movies B, H, and director G; path types can include interaction paths between users (indicated by two-way arrows), user satisfaction paths with movies (indicated by dashed lines) ), the path of the user's satisfaction with the director (indicated by a solid line). The prediction content can be whether the user prefers a certain movie, or it can be the rating of the movie or the director. Exemplarily, to predict whether user A prefers a certain movie, the characteristic information of movie B and movie H in the network can be processed according to the network of user A's movie satisfaction path to obtain a group aggregation vector for user A. In the same way, the characteristic information of user G and directors B and H can be processed according to the network of interaction path and director satisfaction path, and two other group aggregation vectors for user A can be obtained. Then, the importance of each group of users in different paths relative to user A is determined according to the aggregation vector representation of each group. Then, the group aggregation vectors corresponding to different paths are merged based on different degrees of importance to obtain a representation vector for user A. According to the representation vector and the feature information of a certain movie, the user A's preference for a certain movie is predicted. As another example, predicting the score of director G may be based on the network of users' satisfaction with the movie, and based on users A, D, and F in the network.
关于基于异构图神经网络模型进行预测的技术方案,参见下文。For the technical solution of prediction based on the heterogeneous graph neural network model, see below.
图2是根据本说明书一些实施例所示的基于异构图神经网络模型进行预测的系统的模块图。Fig. 2 is a block diagram of a system for predicting based on a heterogeneous graph neural network model according to some embodiments of this specification.
如图2所示,该通过实体对象数据判断实体对象类别的系统200可以包括异构图获取模块210、分组模块220和节点预测模块230。As shown in FIG. 2, the system 200 for judging the entity object category based on entity object data may include a heterogeneous graph acquisition module 210, a grouping module 220, and a node prediction module 230.
异构图获取模块210可以获取与预测内容相关的异构图数据,所述异构图数据包 括待预测节点、所述待预测节点的邻居节点、以及连接所述待预测节点与所述邻居节点之间的路径,所述路径包括至少一种类型。异构图获取模块210的详细描述可以参见图3的步骤302,在此不再赘述。The heterogeneous graph acquisition module 210 can acquire heterogeneous graph data related to the predicted content, the heterogeneous graph data including the node to be predicted, the neighbor nodes of the node to be predicted, and the connection between the node to be predicted and the neighbor node The path between, the path includes at least one type. For a detailed description of the heterogeneous graph acquisition module 210, refer to step 302 in FIG. 3, which will not be repeated here.
分组模块220可以用于基于所述路径的类型,对所述邻居节点进行分组,以使得同一组的所述邻居节点的路径的类型相同。分组模块220的详细描述可以参见图3的步骤304,在此不再赘述。The grouping module 220 may be configured to group the neighbor nodes based on the types of the paths, so that the types of paths of the neighbor nodes in the same group are the same. For a detailed description of the grouping module 220, refer to step 304 in FIG. 3, which will not be repeated here.
节点预测模块230可以用于将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量后输入训练好的预测模型进行预测。在一些实施例中,所述节点预测模块还用于融合节点注意力权重和/或路径注意力权重后得到所述待预测节点的表示向量。在一些实施例中,节点预测模块230可以根据同一组的所述邻居节点的特征信息,逐一得到对应于该组的组聚合向量;针对每一所述组聚合向量,将该待预测节点特征信息融合,得到分组融合的待预测节点信息。在一些实施例中,所述节点预测模块还用于基于路径的重要程度确定路径的注意力权重;基于所述分组融合的待预测节点信息和所述路径的注意力权重确定所述待预测节点的表示向量。在一些实施例中,节点预测模块230可以基于邻居节点的重要程度确定节点的注意力权重向量,基于路径的重要程度确定路径的注意力权重向量。在一些实施例中,节点预测模块230可以基于所述分组融合的待预测节点信息、所述节点的注意力权重向量和所述路径的注意力权重向量确定所述待预测节点的表示向量。节点预测模块230的详细描述可以参见图3的步骤306,在此不再赘述。The node prediction module 230 may be used to input the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model, obtain the representation vector of the node to be predicted, and then input the trained Predictive models make predictions. In some embodiments, the node prediction module is further used to obtain the representation vector of the node to be predicted after fusing the node attention weight and/or the path attention weight. In some embodiments, the node prediction module 230 may obtain the group aggregation vector corresponding to the group one by one according to the characteristic information of the neighbor nodes of the same group; for each group aggregation vector, the characteristic information of the node to be predicted Fusion, get the node information to be predicted that is grouped and fused. In some embodiments, the node prediction module is further configured to determine the attention weight of the path based on the importance of the path; determine the node to be predicted based on the information of the node to be predicted fused by the group and the attention weight of the path The representation vector. In some embodiments, the node prediction module 230 may determine the attention weight vector of the node based on the importance of neighbor nodes, and determine the attention weight vector of the path based on the importance of the path. In some embodiments, the node prediction module 230 may determine the representation vector of the node to be predicted based on the grouped fusion of node information to be predicted, the attention weight vector of the node, and the attention weight vector of the path. For a detailed description of the node prediction module 230, refer to step 306 in FIG. 3, which is not repeated here.
在一些实施例中,所述系统100还包括:训练模块,用于对系统的异构图网络神经模型和预测模型进行端对端的训练,具体为基于预测模型的损失函数迭代更新预测模型以及异构图神经网络模型的模型参数,直到满足迭代截止条件。在一些实施例中,可以将若干个异构图数据作为训练数据,将对应于该异构图数据的节点正确结果作为该训练数据的标签数据,所述预测模型的参数和异构图神经网络模型的参数利用所述训练数据和所述标签数据通过训练迭代更新。具体可以参加本文其他地方描述的异构图网络神经模型和预测模型的训练过程,在此不再赘述。In some embodiments, the system 100 further includes: a training module for end-to-end training of the heterogeneous graph network neural model and prediction model of the system, specifically the iterative update of the prediction model based on the loss function of the prediction model and the different Compose the model parameters of the neural network model until the iteration cut-off condition is met. In some embodiments, several heterogeneous graph data can be used as training data, and the correct result of the node corresponding to the heterogeneous graph data can be used as the label data of the training data. The parameters of the prediction model and the heterogeneous graph neural network The parameters of the model are updated through training iterations using the training data and the label data. Specifically, you can participate in the training process of the heterogeneous graph network neural model and prediction model described elsewhere in this article, and I will not repeat it here.
应当理解,图2所示的系统及其模块可以利用各种方式来实现。例如,在一些实施例中,系统及其模块可以通过硬件、软件或者软件和硬件的结合来实现。其中,硬件部分可以利用专用逻辑来实现;软件部分则可以存储在存储器中,由适当的指令执行系统,例如微处理器或者专用设计硬件来执行。本领域技术人员可以理解上述的方法和系统可以使用计算机可执行指令和/或包含在处理器控制代码中来实现,例如在诸如磁盘、CD或DVD-ROM的载体介质、诸如只读存储器(固件)的可编程的存储器或者诸如光学或电子信号载体的数据载体上提供了这样的代码。本说明书的系统及其模块不仅可以有诸如超大规模集成电路或门阵列、诸如逻辑芯片、晶体管等的半导体、或者诸如现场可编程门阵列、可编程逻辑设备等的可编程硬件设备的硬件电路实现,也可以用例如由各种类型的处理器所执行的软件实现,还可以由上述硬件电路和软件的结合(例如,固 件)来实现。It should be understood that the system and its modules shown in FIG. 2 can be implemented in various ways. For example, in some embodiments, the system and its modules may be implemented by hardware, software, or a combination of software and hardware. Among them, the hardware part can be implemented using dedicated logic; the software part can be stored in a memory and executed by an appropriate instruction execution system, such as a microprocessor or dedicated design hardware. Those skilled in the art can understand that the above-mentioned methods and systems can be implemented using computer-executable instructions and/or included in processor control codes, for example on a carrier medium such as a disk, CD or DVD-ROM, such as a read-only memory (firmware Such codes are provided on a programmable memory or a data carrier such as an optical or electronic signal carrier. The system and its modules in this specification can not only be implemented by hardware circuits such as very large-scale integrated circuits or gate arrays, semiconductors such as logic chips, transistors, etc., or programmable hardware devices such as field programmable gate arrays, programmable logic devices, etc. It may also be implemented by software executed by various types of processors, or may be implemented by a combination of the foregoing hardware circuit and software (for example, firmware).
需要注意的是,以上对于候选项显示、确定系统及其模块的描述,仅为描述方便,并不能把本说明书限制在所举实施例范围之内。可以理解,对于本领域的技术人员来说,在了解该系统的原理后,可能在不背离这一原理的情况下,对各个模块进行任意组合,或者构成子系统与其他模块连接。例如,在一些实施例中,例如,图2中披露的异构图获取模块210、分组模块220、节点预测模块230和训练模块可以是一个系统中的不同模块,也可以是一个模块实现上述的两个或两个以上模块的功能。例如,异构图获取模块210、分组模块220可以是两个模块,也可以是一个模块同时具有异构图获取和分组功能。例如,各个模块可以共用一个存储模块,各个模块也可以分别具有各自的存储模块。诸如此类的变形,均在本说明书的保护范围之内。It should be noted that the above description of the candidate item display, determination system and its modules is only for convenience of description, and does not limit this specification to the scope of the examples mentioned. It can be understood that for those skilled in the art, after understanding the principle of the system, it is possible to arbitrarily combine various modules, or form a subsystem to connect with other modules without departing from this principle. For example, in some embodiments, for example, the heterogeneous graph acquisition module 210, the grouping module 220, the node prediction module 230, and the training module disclosed in FIG. The function of two or more modules. For example, the heterogeneous graph acquisition module 210 and the grouping module 220 may be two modules, or one module may have both heterogeneous graph acquisition and grouping functions. For example, each module may share a storage module, and each module may also have its own storage module. Such deformations are all within the protection scope of this specification.
图3是根据本说明书一些实施例所示的基于异构图神经网络模型进行预测的方法的示例性流程图。如图3所示,该通过实体对象数据判断实体对象类别的方法300包括:Fig. 3 is an exemplary flowchart of a method for prediction based on a heterogeneous graph neural network model according to some embodiments of the present specification. As shown in FIG. 3, the method 300 for judging the category of an entity object based on entity object data includes:
步骤302,获取与预测内容相关的异构图数据。所述异构图数据包括待预测节点、所述待预测节点的邻居节点、以及连接所述待预测节点与所述邻居节点之间的路径,所述路径包括至少一种类型。Step 302: Obtain heterogeneous graph data related to the predicted content. The heterogeneous graph data includes a node to be predicted, neighbor nodes of the node to be predicted, and a path connecting the node to be predicted and the neighbor node, and the path includes at least one type.
具体的,步骤302可以由异构图获取模块210执行。Specifically, step 302 may be performed by the heterogeneous graph acquisition module 210.
异构图数据是包括不同类型节点和不同类型节点之间路径的数据关系图。例如,用户观影数据关系图。又例如,用户金融关系数据图。Heterogeneous graph data is a data relationship graph including different types of nodes and paths between different types of nodes. For example, a graph of user viewing data. Another example is a graph of user financial relationship data.
节点是指实体对象。在一些实施例中,节点的类型可以相同。例如,节点可以都是用户。在一些实施例中,节点的类型可以不同。例如,节点可以包括用户、导演和演员等。Nodes refer to physical objects. In some embodiments, the types of nodes may be the same. For example, nodes can all be users. In some embodiments, the types of nodes can be different. For example, nodes can include users, directors, actors, and so on.
路径表示节点之间的连接关系。在一些实施例中,不同类型的路径可以表示节点之间的不同的连接关系。The path represents the connection relationship between nodes. In some embodiments, different types of paths may represent different connection relationships between nodes.
在一些实施例中,连接关系可以是相同类型的实体对象之间的连接关系。在一些实施例中,连接关系可以是相同类型的实体对象之间的交互关系(如用户之间的关注、点赞、评论)。示例性的,路径可以表示为:用户A—点赞—用户B。在一些实施例中,连接关系也可以是相同类型的实体对象之间的共同偏好。例如,用户之间共同偏好的观影方式(如优酷在线观看、电影院观看)。示例性的,路径可以表示为:用户A—优酷在线观看—用户B。In some embodiments, the connection relationship may be a connection relationship between entity objects of the same type. In some embodiments, the connection relationship may be an interaction relationship between entity objects of the same type (such as attention, likes, and comments between users). Exemplarily, the path can be expressed as: User A—Like—User B. In some embodiments, the connection relationship may also be a common preference between entity objects of the same type. For example, the viewing methods that users commonly prefer (such as online viewing on Youku, viewing in movie theaters). Exemplarily, the path may be expressed as: user A—youku online viewing—user B.
在一些实施例中,连接关系还可以是不同类型的实体对象之间的关系。例如,实体对象用户A对实体对象导演K的满意度(如喜欢、无感、讨厌)。示例性的,路径可以表示为:用户A—喜欢—导演K。In some embodiments, the connection relationship may also be a relationship between different types of entity objects. For example, the satisfaction of the entity object user A with the entity object director K (such as like, insensitivity, dislike). Exemplarily, the path can be expressed as: user A-likes-director K.
在一些实施例中,异构图数据可以通过用户授权的社交网络、点评网站、征信网站等存储的数据获取,还可以通过调用相关接口或其他方式获取。本实施例对获取方式不做限制。例如,通过社交网络授权的注册信息可以获取实体对象—用户,通过社交网络用户之间的交互行为获取路径。In some embodiments, the heterogeneous graph data may be obtained through data stored in a social network, a review website, a credit information website, etc. authorized by the user, and may also be obtained by calling related interfaces or other methods. This embodiment does not limit the acquisition method. For example, the entity object-user can be obtained through the registration information authorized by the social network, and the path can be obtained through the interactive behavior between the users of the social network.
待预测内容可以是实体对象的未知数据。在一些实施例中,在一些实施例中,预测模型的预测内容可以包括待预测节点对应的实体对象的类别、风险等级或者偏好习惯。The content to be predicted may be unknown data of the physical object. In some embodiments, in some embodiments, the prediction content of the prediction model may include the category, risk level, or preference habits of the entity object corresponding to the node to be predicted.
示例性的,实体对象的类别可以包括用户的收入水平类别(贫穷、小康和富裕)、导演评价类别(差、一般、好、很好)等。Exemplarily, the category of the entity object may include the user's income level category (poor, well-off, and rich), director evaluation category (poor, fair, good, very good), and so on.
示例性的,实体对象的风险等级可以包括用户的金融借贷风险等级(例如,违约可能性低、违约可能性中、违约可能性高)、投保人的健康风险等级(例如,赔付可能性低、赔付可能性中、赔付可能性高)等。Exemplarily, the risk level of the entity object may include the user's financial lending risk level (for example, the probability of default is low, the probability of default is medium, the probability of default is high), the health risk level of the applicant (for example, the probability of compensation is low, The possibility of compensation is medium, the probability of compensation is high), etc.
示例性的,实体对象的偏好习惯可以包括用户偏好的电影类型(如,喜剧片、动作片、科幻片)、用户偏好的还贷方式(如等额本息、等额本金)。Exemplarily, the preference habits of the entity object may include the user's preferred movie type (eg, comedy, action, science fiction film), and the user's preferred loan repayment method (eg, equal principal and interest, equal principal).
待预测节点是输出待预测内容的节点。可以理解的是,可以根据待预测内容选择待预测节点。以用户偏好的电影类型为例,实体对象包括用户、导演和演员的异构图中,待预测内容是各个用户喜欢的电影类型,则待预测节点可以是用户;待预测内容是各个导演的评分,则待预测节点可以是导演。为了进一步说明,以用户的借贷风险为预测内容的实施例参加下文。The node to be predicted is a node that outputs the content to be predicted. It is understandable that the node to be predicted can be selected according to the content to be predicted. Take the user’s preferred movie type as an example. The entity object includes a heterogeneous graph of users, directors, and actors. If the content to be predicted is the type of movie that each user likes, the node to be predicted may be the user; the content to be predicted is the rating of each director , The node to be predicted can be the director. For further explanation, the following is an example of the prediction content of the user's loan risk.
邻居节点通过路径直接连接待预测节点。在一些实施例中,一个预测节点可以通过多条路径连接到多个邻居节点。在一些实施例中,多条路径的类型可以相同也可以不同。例如,待预测节点用户A“点赞”了1个用户,“评论”了2个用户,则交互类型的路径的邻居节点数量为3;待预测节点用户A和另外2个用户都使用“优酷在线”观看电影,同时和另外3个用户都“电影院观看”电影,则观影方式类型的路径的邻居节点数量为5。The neighbor nodes are directly connected to the node to be predicted through the path. In some embodiments, a prediction node may be connected to multiple neighbor nodes through multiple paths. In some embodiments, the types of multiple paths may be the same or different. For example, if user A of the node to be predicted "likes" 1 user and "comments" 2 users, the number of neighbor nodes of the interactive path is 3; user A of the node to be predicted and the other 2 users both use "Youku" To watch movies online, and to watch movies in cinemas with 3 other users at the same time, the number of neighbor nodes of the path of the movie viewing mode is 5.
为了便于进一步理解,以用户借贷风险应用场景为另一实施例进行说明。例如,实体对象为用户的异构图中,对应的待预测节点为用户,待预测内容是各个用户的借贷风险等级。路径类型可以包括用户之间交互关系、用户的借贷方式、收入等级等,对应的路径可以示例性地对应为:用户A-评论-用户D、用户A-小额信贷公司-用户B、用户D-中产-用户C。In order to facilitate further understanding, an application scenario of user loan risk is taken as another embodiment for description. For example, in a heterogeneous graph where the entity object is a user, the corresponding node to be predicted is the user, and the content to be predicted is the loan risk level of each user. Path types can include interactions between users, user’s lending methods, income levels, etc. The corresponding paths can be exemplarily corresponding to: user A-comment-user D, user A-microfinance company-user B, user D -Middle class-User C.
步骤304,基于所述路径的类型,对所述邻居节点进行分组,以使得同一组的所述邻居节点的路径的类型相同。Step 304: Group the neighboring nodes based on the types of the paths, so that the types of paths of the neighboring nodes in the same group are the same.
具体的,步骤304可以由分组模块220执行。Specifically, step 304 may be performed by the grouping module 220.
在一些实施例中,根据连接待预测节点与其邻居节点之间的路径的类型对邻居节点进行分类。沿用上述例子,交互类型的路径连接的邻居节点分为一组,观影方式类型的路径连接的邻居节点分为一组。In some embodiments, neighbor nodes are classified according to the type of path connecting the node to be predicted and its neighbor nodes. Following the above example, the neighbor nodes connected by the interactive path are grouped into a group, and the neighbor nodes connected by the path of the viewing mode are grouped into a group.
在一些实施例中,每一组邻居节点基于待预测节点和其中的邻居节点的路径连接结构对应于一个图神经网络。In some embodiments, each group of neighbor nodes corresponds to a graph neural network based on the path connection structure of the node to be predicted and the neighbor nodes therein.
步骤306,将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量后输入训练好的预测模型进行预测。Step 306: Input the node to be predicted, the grouped neighbor nodes, and the path between the nodes into the trained heterogeneous graph neural network model to obtain the representation vector of the node to be predicted and then input the trained prediction model to make predictions .
具体的,步骤306可以由节点预测模块230执行。Specifically, step 306 may be performed by the node prediction module 230.
将待预测节点、分组后的邻居节点以及节点之间的路径的特征信息输入训练好的异构图神经网络模型,输出为待预测节点的表示向量。节点的特征信息是可以表征实体对象特征的数据。相应地,路径的特征信息是可以表征路径特征的数据。异构图神经网络模型的详细描述可以参见图4,在此不再赘述。The characteristic information of the node to be predicted, the grouped neighbor nodes, and the path between the nodes are input into the trained heterogeneous graph neural network model, and the output is the representation vector of the node to be predicted. The characteristic information of a node is data that can characterize the characteristics of an entity object. Correspondingly, the characteristic information of the path is data that can characterize the characteristic of the path. The detailed description of the heterogeneous graph neural network model can be seen in Fig. 4, which will not be repeated here.
将待预测节点的表示向量输入训练好的预测模型,输出预测结果。Input the representation vector of the node to be predicted into the trained prediction model, and output the prediction result.
在一些实施例中,测结果是待预测内容的值。如前所述,预测模型的预测内容可以包括待预测节点对应的实体对象的类别、风险等级或者偏好习惯。In some embodiments, the measurement result is the value of the content to be predicted. As mentioned above, the prediction content of the prediction model may include the category, risk level, or preference habits of the entity object corresponding to the node to be predicted.
在一些实施例中,预测模型可以是二分类模型、逻辑回归模型或神经网络等具有分类能力的机器学习模型,本说明书对其不做具体限制。In some embodiments, the prediction model may be a machine learning model with classification capabilities, such as a binary classification model, a logistic regression model, or a neural network, which is not specifically limited in this specification.
在一种可能的实施方式中,将信息表示学习向量输入预测模型后,可能得到多个输出值,每个输出值用于表征该输出值对应的类别的置信度,也即模型对于该类别的预测概率。例如输出值有3个,分别为0.6,0.2,0.2,分别对应用户为喜剧片爱好者、动作片爱好者以及科幻片爱好者的概率,由于喜剧片爱好者对应的概率0.6最大,可以认为该用户属于喜剧片爱好者。In a possible implementation, after the information representation learning vector is input to the prediction model, multiple output values may be obtained, and each output value is used to characterize the confidence of the category corresponding to the output value, that is, the model's Forecast probability. For example, there are three output values, 0.6, 0.2, 0.2, respectively, corresponding to the probability that the user is a comedy fan, an action fan, and a science fiction fan. Since the probability of a comedy fan is 0.6, it can be considered that Users belong to comedy fans.
图4是根据本说明书一些实施例所示的基于异构图神经网络模型进行预测的方法的示例性子流程图。Fig. 4 is an exemplary sub-flow chart of a method for prediction based on a heterogeneous graph neural network model according to some embodiments of the present specification.
在一些实施例中,异构图神经网络模型可以包括组聚合向量层、待预测节点信息层和表示向量层。In some embodiments, the heterogeneous graph neural network model may include a group aggregation vector layer, a node information layer to be predicted, and a representation vector layer.
如图4所述,将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量的方法400包括步骤402~步骤404。As shown in FIG. 4, the method 400 for inputting the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into a trained heterogeneous graph neural network model to obtain a representation vector of the node to be predicted includes step 402 ~Step 404.
步骤402,根据同一组的所述邻居节点的特征信息,逐一得到对应于该组的组聚合向量。Step 402: According to the feature information of the neighbor nodes in the same group, obtain the group aggregation vector corresponding to the group one by one.
在一些实施例中,组聚合向量层可以用于对输入的同一组的邻居节点的特征信息进行聚合,生成组聚合向量。In some embodiments, the group aggregation vector layer may be used to aggregate the input feature information of neighbor nodes of the same group to generate a group aggregation vector.
在一些实施例中,组聚合向量层可以先对邻居节点的特征信息进行向量化,得到特征向量X;同一组邻居节点的特征向量再通过图神经网络进行特征聚合,得到特征聚合向量h;然后基于特征聚合向量得到组聚合向量H。In some embodiments, the group aggregation vector layer can first vectorize the feature information of the neighbor nodes to obtain the feature vector X; the feature vectors of the same group of neighbor nodes are then aggregated through the graph neural network to obtain the feature aggregation vector h; The group aggregation vector H is obtained based on the feature aggregation vector.
如前所述,节点的特征信息是可以表征实体对象特征的数据。例如,实体对象为用户,进而实体对象的特征信息可以包括用户的身份信息数据(如年龄、职业),还可以包括用户的历史偏好数据(如喜欢过的电影、电视剧类型等)。在一些实施例中,实体对象的特征信息可以通过用户注册时输入的授权数据获取,也可以通过用户的实际操作(如点赞、评论等)获取,还可以通过读取存储的数据、调用相关接口或其他方式获取。本实施例对获取方式不做限制。As mentioned earlier, the characteristic information of a node is data that can characterize the characteristics of an entity object. For example, the entity object is a user, and the characteristic information of the entity object may include the user's identity information data (such as age, occupation), and may also include the user's historical preference data (such as favorite movies, TV series, etc.). In some embodiments, the feature information of the entity object can be obtained through the authorization data entered by the user during registration, or through the actual operation of the user (such as likes, comments, etc.), and it can also be obtained by reading stored data and calling related Obtained via interface or other methods. This embodiment does not limit the acquisition method.
在一些实施例中,可以预先通过向量化表示模型对特征信息进行向量化获取对应 的特征向量。在一些实施例中,向量化表示模型可以是词嵌入模型,词嵌入模型可以包括但不限于:Word2vec模型、词频-逆向文件频率模型(Term Frequency–Inverse Document Frequency,TF-IDF)或SSWE-C(skip-gram based combined-sentiment word embedding)模型等。In some embodiments, the feature information may be vectorized through the vectorized representation model in advance to obtain the corresponding feature vector. In some embodiments, the vectorized representation model may be a word embedding model, and the word embedding model may include but is not limited to: Word2vec model, term frequency-inverse document frequency model (Term Frequency-Inverse Document Frequency, TF-IDF) or SSWE-C (skip-gram based combined-sentiment word embedding) model, etc.
在一些实施例中,可以通过对各邻居节点的特征向量X进行特征聚合,得到特征聚合向量h。可以理解的是,一组邻居节点对应一个特征聚合向量h。In some embodiments, the feature aggregation vector h can be obtained by performing feature aggregation on the feature vector X of each neighbor node. It can be understood that a group of neighbor nodes corresponds to a feature aggregation vector h.
在一些实施例中,可以对待预测节点的预定阶数(如2阶)内的邻居节点作为特征聚合的节点,也可以对预定阶数内的邻居进行采样,将采样得到的邻居节点做特征聚合。In some embodiments, neighbor nodes within a predetermined order (such as order 2) of the node to be predicted can be used as feature aggregation nodes, or neighbors within a predetermined order can be sampled, and the neighbor nodes obtained by sampling can be used for feature aggregation. .
在一些实施例中,特征聚合可以通过预先设定的聚合方式对特征向量X进行聚合;也可以基于每一组邻居节点对应的图神经网络,对采用注意力机制对特征向量X进行聚合;还可以通过所述图神经网络对各邻居节点的特征向量X进行卷积。下文将分别对这三种方式进行介绍。In some embodiments, the feature aggregation may aggregate the feature vector X through a preset aggregation method; it may also aggregate the feature vector X using the attention mechanism based on the graph neural network corresponding to each group of neighbor nodes; The feature vector X of each neighbor node can be convolved through the graph neural network. These three methods will be introduced separately below.
在一些实施例中,预先设定的聚合方式可以是利用一个预先训练好的参数矩阵对邻居节点的特征向量X进行加和、求平均、取最大值、求加权和,等等,在此不作限定。In some embodiments, the pre-set aggregation method may be to use a pre-trained parameter matrix to add, average, take the maximum value, calculate the weighted sum, etc., on the feature vector X of the neighbor nodes, etc., which will not be used here. limited.
例如,在得到待预测节点特征向量X u以及邻居节点特征向量
Figure PCTCN2021074479-appb-000001
后,若待预测节点特征向量X u的维度为1*8,第一参数矩阵的维度为8*5,则将待预测节点特征向量X u与第一参数矩阵相乘得到维度为1*5的待预测特征聚合向量h u,第一参数矩阵可以视为对待预测节点特征向量X u的每个元素进行了加权求和操作,进而对待预测节点的特征信息起到了聚合作用,使得待预测节点特征聚合向量h u的元素可以表示更丰富的信息。而对于邻居节点特征向量,以图5所示的邻居节点特征向量
Figure PCTCN2021074479-appb-000002
为例,若其维度为1*10,则将其对应的多个邻居节点特征向量相加后除以3得到维度为1*10的平均向量,然后将该平均向量乘以维度为10*5的第二参数矩阵得到维度为1*5的邻居节点特征聚合向量
Figure PCTCN2021074479-appb-000003
同理可以得到邻居节点特征聚合向量
Figure PCTCN2021074479-appb-000004
For example, after obtaining the feature vector X u of the node to be predicted and the feature vector of the neighbor node
Figure PCTCN2021074479-appb-000001
Then, if the dimension of the feature vector X u of the node to be predicted is 1*8 and the dimension of the first parameter matrix is 8*5, the feature vector X u of the node to be predicted is multiplied by the first parameter matrix to obtain a dimension of 1*5 To be predicted feature aggregation vector h u , the first parameter matrix can be regarded as a weighted sum operation for each element of the feature vector X u of the node to be predicted, and then the feature information of the node to be predicted is aggregated, so that the node to be predicted The elements of the feature aggregation vector h u can represent richer information. For the neighbor node feature vector, the neighbor node feature vector shown in Figure 5
Figure PCTCN2021074479-appb-000002
For example, if its dimension is 1*10, add the feature vectors of multiple neighbor nodes and divide by 3 to obtain an average vector with a dimension of 1*10, and then multiply the average vector by a dimension of 10*5 The second parameter matrix of, obtains the feature aggregation vector of neighbor nodes with dimension 1*5
Figure PCTCN2021074479-appb-000003
In the same way, the feature aggregation vector of neighbor nodes can be obtained
Figure PCTCN2021074479-appb-000004
当然,对于不同类型的路径对应的邻居节点,第二参数矩阵可以不同。为了得到第一参数矩阵和第二参数矩阵,可以预先对异构图神经网络模型进行训练。训练过程可以参见本文其他地方的描述。Of course, for neighboring nodes corresponding to different types of paths, the second parameter matrix may be different. In order to obtain the first parameter matrix and the second parameter matrix, the heterogeneous graph neural network model can be trained in advance. The training process can be referred to the description elsewhere in this article.
如前所述,在一些实施例中,也可以基于每一组邻居节点对应的图神经网络,对采用注意力机制对特征向量X进行聚合。As mentioned above, in some embodiments, it is also possible to aggregate the feature vector X using the attention mechanism based on the graph neural network corresponding to each group of neighbor nodes.
图神经网络可以有一层或多层。在一些实施例中,每一层图神经网络对应的模型参数也可以有多组。The graph neural network can have one or more layers. In some embodiments, there may also be multiple sets of model parameters corresponding to each layer of graph neural network.
注意力机制(Attention Mechanism)是一种深度学习技术,可以使得图神经网络具备专注于其输入(或特征)子集的能力。在一些实施例中,可以通过注意力机制获取神经网络的邻居注意力权重(即邻居权重向量);对于多层的图神经网络,可以通过注意力机制获取隐藏层的特征注意力权重(即特征权重向量),进而使得邻居节点中较为重 要的特征向量被利用的程度更高,减小噪声信息的干扰。Attention Mechanism (Attention Mechanism) is a deep learning technology that enables graph neural networks to have the ability to focus on a subset of their inputs (or features). In some embodiments, the neighbor attention weight (ie, neighbor weight vector) of the neural network can be obtained through the attention mechanism; for multi-layer graph neural networks, the feature attention weight of the hidden layer (ie feature Weight vector), which in turn makes the more important feature vectors in neighbor nodes to be used to a higher degree, and reduces the interference of noise information.
在一些实施例中,每组模型参数为一个权重矩阵,一层图神经网络可以对应多个权重矩阵。对于训练好的图神经网络而言,模型参数可以是经过训练过程中的参数调整确定下来的。In some embodiments, each group of model parameters is a weight matrix, and a layer of graph neural network may correspond to multiple weight matrices. For the trained graph neural network, the model parameters can be determined through parameter adjustment during the training process.
在一些实施例中,权重矩阵可以基于各邻居节点的重要度获取。各个邻居节点的重要度可以通过待预测节点与邻居节点的特征向量确定。In some embodiments, the weight matrix may be obtained based on the importance of each neighbor node. The importance of each neighbor node can be determined by the feature vector of the node to be predicted and the neighbor node.
在一些实施例中,可以通过公式(1)计算邻居权重向量。例如,待预测节点u的第j组邻居节点中,有N个节点。对于待预测节点u,邻居节点k的邻居权重可以为:In some embodiments, the neighbor weight vector can be calculated by formula (1). For example, there are N nodes in the j-th group of neighbor nodes of the node u to be predicted. For the node u to be predicted, the neighbor weight of the neighbor node k can be:
α(u,k)=softmax(V·tanh(W 1[X u||X k])+b 1) α(u,k)=softmax(V·tanh(W 1 [X u ||X k ])+b 1 )
其中,矩阵V(例如称为第一权重矩阵)和W 1(例如称为第二权重矩阵)是图神经网络训练过程中确定的模型参数,b 1是图神经网络训练过程中确定的常数参数,X u、X k分别是节点u、节点k对应的当前的特征表达向量,图神经网络的第一层聚合时,各个节点的当前特征表达向量由相应节点的特征向量确定,[X u||X k]表示两个向量的拼接向量。可以理解的是,激活函数softmax、tanh也可以用其他激活函数(如Relu等)代替,在此不作限定。 Among them, the matrices V (for example, called the first weight matrix) and W 1 (for example, called the second weight matrix) are the model parameters determined during the training process of the graph neural network, and b 1 is the constant parameter determined during the training process of the graph neural network , X u and X k are the current feature expression vector corresponding to node u and node k respectively. When the first layer of graph neural network is aggregated, the current feature expression vector of each node is determined by the feature vector of the corresponding node, [X u | |X k ] represents the stitching vector of two vectors. It is understandable that the activation functions softmax and tanh can also be replaced by other activation functions (such as Relu, etc.), which are not limited here.
如此,可以针对各个邻居节点分别确定相应的邻居权重。在各个邻居节点的当前特征向量各不相同的情况下,针对相应邻居节点的邻居权重也各不相同。值得说明的是,也可以将待预测节点u看作自身的邻居节点,例如称为零阶邻居节点。In this way, the corresponding neighbor weight can be determined for each neighbor node. When the current feature vector of each neighbor node is different, the neighbor weight for the corresponding neighbor node is also different. It is worth noting that the node u to be predicted can also be regarded as its own neighbor node, for example, it is called a zero-order neighbor node.
根据邻居权重对各个邻居节点进行特征聚合,可以采用诸如求加权和等方式进行。例如,通过N u表示待预测节点u的第j组邻居节点集合,待预测节点u的邻居节点经过图神经网络的邻居聚合结果为: According to the neighbor weight, the feature aggregation of each neighbor node can be performed by methods such as weighted sum. For example, by N u represents the j th group of neighbor nodes of node u set to be predicted, to be predicted neighbor nodes after node u neighbor neural network of FIG polymerization results are:
Figure PCTCN2021074479-appb-000005
Figure PCTCN2021074479-appb-000005
可以理解,对于每个节点,经过一层图神经网络之后,都可以得到一个当前层的聚合结果,如节点k的聚合结果(也可以称为特征表达向量)为
Figure PCTCN2021074479-appb-000006
It can be understood that for each node, after a layer of graph neural network, an aggregation result of the current layer can be obtained, for example, the aggregation result of node k (also called a feature expression vector) is
Figure PCTCN2021074479-appb-000006
在一些实施例中,可以将以上邻居聚合结果进一步与待预测节点的特征表达向量聚合,得到待预测节点在图神经网络的当前层的聚合结果。例如,假设图神经网络为多层网络,节点1、节点2、节点3…节点k…为待预测节点u的邻居节点,将它们在第i-1层(i≥2)对应的特征聚合结果分别记为
Figure PCTCN2021074479-appb-000007
待预测节点u在第i-1层对应的特征聚合结果记为
Figure PCTCN2021074479-appb-000008
则在第i层,相应节点的当前特征表达向量为第i-1层的特征聚合结果(即第i-1层输出的特征表达向量)。
In some embodiments, the above neighbor aggregation result may be further aggregated with the feature expression vector of the node to be predicted to obtain the aggregation result of the node to be predicted in the current layer of the graph neural network. For example, suppose that the graph neural network is a multi-layer network, node 1, node 2, node 3...Node k... are neighbor nodes of the node u to be predicted, and aggregate their corresponding features in the i-1th layer (i≥2). Respectively denoted as
Figure PCTCN2021074479-appb-000007
The feature aggregation result corresponding to the node u to be predicted in the i-1th layer is recorded as
Figure PCTCN2021074479-appb-000008
Then in the i-th layer, the current feature expression vector of the corresponding node is the feature aggregation result of the i-1th layer (that is, the feature expression vector output by the i-1th layer).
在一些实施例中,将节点u的各个邻居节点进行聚合,得到邻居聚合结果
Figure PCTCN2021074479-appb-000009
然后,将上一层网络节点u的特征聚合结果
Figure PCTCN2021074479-appb-000010
作为特征向量
Figure PCTCN2021074479-appb-000011
Figure PCTCN2021074479-appb-000012
进行聚合,可以得到待预测节点u在第i层的特征表达向量
Figure PCTCN2021074479-appb-000013
从而,在第j组邻居节点的图神 经网络中,经过预先训练的图神经网络的层层迭代处理,最终可以得到对应的一个特征聚合向量。
In some embodiments, the neighbor nodes of node u are aggregated to obtain the neighbor aggregation result
Figure PCTCN2021074479-appb-000009
Then, the feature aggregation results of the upper network node u
Figure PCTCN2021074479-appb-000010
As feature vector
Figure PCTCN2021074479-appb-000011
with
Figure PCTCN2021074479-appb-000012
Perform aggregation to get the feature expression vector of the node u to be predicted in the i-th layer
Figure PCTCN2021074479-appb-000013
Therefore, in the graph neural network of the j-th group of neighbor nodes, through the iterative processing of the pre-trained graph neural network, a corresponding feature aggregation vector can finally be obtained.
这里,将
Figure PCTCN2021074479-appb-000014
Figure PCTCN2021074479-appb-000015
进行聚合的过程例如可以是求和、求平均或加权求和等。然而,在特征表达向量中,每个特征对节点的表达向量的贡献度也可能不同,因此,在进一步可选的实现方式中,各个特征表达向量还可以具有特征重要度(特征权重)。
Here, will
Figure PCTCN2021074479-appb-000014
with
Figure PCTCN2021074479-appb-000015
The process of aggregation can be, for example, summation, averaging, or weighted summation. However, in the feature expression vector, the contribution of each feature to the expression vector of the node may also be different. Therefore, in a further optional implementation manner, each feature expression vector may also have a feature importance (feature weight).
在一些实施例中,图神经网络的隐藏层,可以通过以下方式确定各个特征表达向量分别对应的特征权重构成的特征权重向量:In some embodiments, the hidden layer of the graph neural network can determine the feature weight vector formed by the feature weights corresponding to each feature expression vector in the following manner:
Figure PCTCN2021074479-appb-000016
Figure PCTCN2021074479-appb-000016
其中,W 2(例如称为第三权重矩阵)、W 3(例如称为第四权重矩阵)均为图神经网络中第i层的权重矩阵,b 2、b 3均为常数参数,这些模型参数均可以在图神经网络训练过程中根据损失函数进行调整确定。在神经网络的某一隐藏层,W 2、W 3、b 2、b 3可以作为通用参数。
Figure PCTCN2021074479-appb-000017
表示两个向量的拼接,其中,
Figure PCTCN2021074479-appb-000018
表示待预测节点u在第i层的特征向量,
Figure PCTCN2021074479-appb-000019
表示待预测节点u的邻居节点在第i层的邻居聚合结果,
Figure PCTCN2021074479-appb-000020
可以通过待预测邻居节点u上一层神经网络中最终的特征聚合结果
Figure PCTCN2021074479-appb-000021
确定。激励函数Relu也可以通过其他合适的激励函数代替,在此不再赘述。
Among them, W 2 (for example, called the third weight matrix) and W 3 (for example, called the fourth weight matrix) are the weight matrices of the i-th layer in the graph neural network, and b 2 and b 3 are constant parameters. These models The parameters can be adjusted and determined according to the loss function during the training process of the graph neural network. In a hidden layer of the neural network, W 2 , W 3 , b 2 , and b 3 can be used as general parameters.
Figure PCTCN2021074479-appb-000017
Represents the splicing of two vectors, where,
Figure PCTCN2021074479-appb-000018
Represents the feature vector of the node u to be predicted in the i-th layer,
Figure PCTCN2021074479-appb-000019
Represents the neighbor aggregation result of the neighbor node of the node u to be predicted in the i-th layer,
Figure PCTCN2021074479-appb-000020
The final feature aggregation result in the upper layer of the neural network of the neighbor node u to be predicted can be used
Figure PCTCN2021074479-appb-000021
Sure. The excitation function Relu can also be replaced by other suitable excitation functions, which will not be repeated here.
特征权重向量β i中的各个元素分别对应各个特征的特征权重。将相应特征权重与邻居聚合结果中的相应元素一一对应相乘,可以得到当前节点u在当前层的特征聚合结果。根据特征权重确定最终的聚合结果的方式可以表示为: Each element in the feature weight vector β i corresponds to the feature weight of each feature. By multiplying the corresponding feature weights with the corresponding elements in the neighbor aggregation result in a one-to-one correspondence, the feature aggregation result of the current node u in the current layer can be obtained. The way to determine the final aggregation result according to the feature weight can be expressed as:
Figure PCTCN2021074479-appb-000022
Figure PCTCN2021074479-appb-000022
其中,⊙表示将两个矩阵的对应元素相乘(如哈达玛积)。对于向量而言,β i中的第k个元素与
Figure PCTCN2021074479-appb-000023
作为聚合结果
Figure PCTCN2021074479-appb-000024
中的第k个元素。例如,向量(A,B,C)⊙(a,b,c)的结果为(Aa,Bb,Cc)。
Among them, ⊙ means multiplying the corresponding elements of two matrices (such as Hadamard product). For a vector, the k-th element in β i is the same as
Figure PCTCN2021074479-appb-000023
As a result of aggregation
Figure PCTCN2021074479-appb-000024
The kth element in. For example, the result of the vector (A, B, C) ⊙ (a, b, c) is (Aa, Bb, Cc).
如此,可以同时考虑节点贡献度和特征贡献度,得到更准确的邻居节点的特征聚合结果。当特征聚合模型为图神经网络时,最后一层得到的聚合结果就是第j组邻居节点对用的特征聚合向量。In this way, the node contribution degree and feature contribution degree can be considered at the same time, and a more accurate feature aggregation result of neighbor nodes can be obtained. When the feature aggregation model is a graph neural network, the aggregation result obtained in the last layer is the feature aggregation vector used by the j-th group of neighbor node pairs.
在一些实施例中,特征聚合向量还可以基于图神经网络进行卷积获取。具体的,可以先对每一个邻居节点的特征向量进行一层或多层卷积,再对输出的各邻居节点的卷积结果进行加和、求平均、取最大值、求加权和,等等,在此不作限定。In some embodiments, the feature aggregation vector may also be obtained by convolution based on the graph neural network. Specifically, one or more layers of convolution can be performed on the feature vector of each neighbor node, and then the output convolution results of each neighbor node can be summed, averaged, maximum value, weighted sum, etc. , It is not limited here.
例如,第j组邻居节点的图神经网络中(图卷积神经网络中),邻居节点v在图神经网络的第l层中间向量可以为:For example, in the graph neural network of the j-th group of neighbor nodes (in the graph convolutional neural network), the intermediate vector of the neighbor node v in the l-th layer of the graph neural network can be:
Figure PCTCN2021074479-appb-000025
Figure PCTCN2021074479-appb-000025
其中:
Figure PCTCN2021074479-appb-000026
是节点v在第j组邻居节点的图神经网络中,第l层的中间向量;N(v)是节点v的邻居节点;d k、d v是归一化因子,比如是相应节点的度,即,与相应节点连接 的连接边数量,或者一阶邻居节点的数量;
Figure PCTCN2021074479-appb-000027
是节点v在图卷积神经网络的第l层的中间向量;
Figure PCTCN2021074479-appb-000028
是节点k在图卷积神经网络的第l层的中间向量;W l是相应节点图卷积神经网络第l层的模型参数。邻居节点有多个时,W l可以是矩阵形式的模型参数,可以称为权重矩阵。公式还可以考虑当前节点的更高阶邻居节点的特征聚合,在此用省略号表示,其原理与一阶邻居节点的特征聚合类似,在此不再赘述。其中,不同的邻居节点的归一化因子不同,特征表达向量不同,从而与权重矩阵相乘的积也不同,因此具有不同的邻居权重。
in:
Figure PCTCN2021074479-appb-000026
Is the intermediate vector of the l-th layer in the graph neural network where node v is in the j-th group of neighbor nodes; N(v) is the neighbor node of node v; d k and d v are normalization factors, such as the degree of the corresponding node , That is, the number of connected edges connected to the corresponding node, or the number of first-order neighbor nodes;
Figure PCTCN2021074479-appb-000027
Is the intermediate vector of node v in the lth layer of the graph convolutional neural network;
Figure PCTCN2021074479-appb-000028
Is the intermediate vector of node k in the lth layer of the graph convolutional neural network; W l is the model parameter of the lth layer of the corresponding node graph convolutional neural network. When there are multiple neighbor nodes, W l can be a model parameter in the form of a matrix, which can be called a weight matrix. The formula can also consider the feature aggregation of higher-order neighbor nodes of the current node, which is represented by an ellipsis here. The principle is similar to the feature aggregation of first-order neighbor nodes, and will not be repeated here. Among them, different neighbor nodes have different normalization factors and different feature expression vectors, so the product multiplied by the weight matrix is also different, so they have different neighbor weights.
可以理解的是,图卷积神经网络的最后一层输出的中间向量即为节点v的卷积结果。例如,
Figure PCTCN2021074479-appb-000029
It is understandable that the intermediate vector output by the last layer of the graph convolutional neural network is the convolution result of node v. E.g,
Figure PCTCN2021074479-appb-000029
由于不同类型的路径对应的邻居节点的特征信息差异较大,同一类型的路径对应的邻居节点的特征信息较为接近,将邻居节点基于路径的类型分类后进行信息聚合,使最终得到的表示学习向量包含的信息更加丰富。Since the feature information of neighbor nodes corresponding to different types of paths are quite different, the feature information of neighbor nodes corresponding to the same type of path is relatively close. The neighbor nodes are classified based on the type of path and the information is aggregated to make the final representation learning vector The information contained is richer.
值得说明的是,特征聚合不限于以上三种方式,还可以采用其他方式进行。It is worth noting that feature aggregation is not limited to the above three methods, and other methods can also be used.
在一些实施例中,可以直接将特征聚合向量作为组聚合向量进行下一步处理。在一些实施例中,也可以对特征聚合向量进行降维。降维可以进一步对特征聚合向量进行特征提取,同时使得在进行后续步骤时效率更高。In some embodiments, the feature aggregation vector can be directly used as the group aggregation vector for further processing. In some embodiments, the feature aggregation vector can also be reduced in dimension. Dimensionality reduction can further perform feature extraction on the feature aggregation vector, and at the same time make the subsequent steps more efficient.
例如,维度为1×3N维的特征聚合向量
Figure PCTCN2021074479-appb-000030
降维为1×5维的组聚合向量
Figure PCTCN2021074479-appb-000031
相应地,m组邻居节点可以得到m个组聚合向量
Figure PCTCN2021074479-appb-000032
For example, a feature aggregation vector with a dimension of 1×3N
Figure PCTCN2021074479-appb-000030
Reduce dimensionality to 1×5 dimensional group aggregation vector
Figure PCTCN2021074479-appb-000031
Correspondingly, m groups of neighbor nodes can get m group aggregation vectors
Figure PCTCN2021074479-appb-000032
在一些实施例中,上述降维的方法可以包括但不限于:主成分分析(Principal Component Analysis,PCA)、线性判别分析(Linear Discriminant Analysis,LDA)、多维尺度变换(Multidimensional scaling,MDS)、局部线性嵌入(Locally Linear Embedding,LLE)、邻接图(ISOMAP,Isometric feature mapping)以及核主成分分析(Kernel Principle Component Analysis,KPCA)等。In some embodiments, the above-mentioned dimensionality reduction methods may include, but are not limited to: principal component analysis (PCA), linear discriminant analysis (LDA), multidimensional scaling (MDS), partial Locally Linear Embedding (LLE), Adjacency Map (ISOMAP, Isometric Feature Mapping), and Kernel Principle Component Analysis (KPCA), etc.
步骤404,针对每一所述组聚合向量,将该待预测节点特征信息融合,得到分组融合的待预测节点信息。Step 404: For each group of aggregation vectors, the feature information of the node to be predicted is fused to obtain grouped and fused information of the node to be predicted.
在一些实施例中,待预测节点信息层可以用于对输入的待预测节点特征信息和组聚合向量进行融合,输出分组融合的待预测节点信息。In some embodiments, the to-be-predicted node information layer may be used to fuse the input feature information of the to-be-predicted node and the group aggregation vector, and output the grouped and fused information of the to-be-predicted node.
分组融合的待预测节点信息是融合了待预测节点特征信息和邻居节点特征信息的向量(即相邻节点融合向量)。The grouped and fused information of the node to be predicted is a vector that combines the feature information of the node to be predicted and the feature information of the neighbor node (ie, the adjacent node fusion vector).
在一些实施例中,待预测节点信息层可以对待预测节点的特征信息进行向量化。在一些实施例中,可以对向量化后的待预测节点特征信息进行降维,得到待预测节点的特征向量。例如,H uIn some embodiments, the information layer of the node to be predicted may vectorize the feature information of the node to be predicted. In some embodiments, the dimensionality of the vectorized feature information of the node to be predicted may be reduced to obtain the feature vector of the node to be predicted. For example, H u .
在一些实施例中,待预测节点特征信息的向量化和降维方法可以参见步骤402,在此不再赘述。In some embodiments, the vectorization and dimensionality reduction methods of the feature information of the node to be predicted may refer to step 402, which will not be repeated here.
在一些实施例中,可以将待预测节点的特征向量分别与步骤402得到的每一个组 聚合向量进行拼接,得到对应的组聚合拼接向量。例如,将H u分别和
Figure PCTCN2021074479-appb-000033
进行拼接,得到m个组聚合拼接向量。在一些实施例中,所述拼接方法可以是直接拼接,也可以是步骤402的聚合,本实施例不做限制。
In some embodiments, the feature vector of the node to be predicted may be spliced with each group aggregation vector obtained in step 402 to obtain the corresponding group aggregation splicing vector. For example, H u and
Figure PCTCN2021074479-appb-000033
Perform splicing to obtain m group aggregation splicing vectors. In some embodiments, the splicing method may be direct splicing or polymerization in step 402, which is not limited in this embodiment.
在一些实施例中,可以基于每一个组聚合拼接向量,得到对应的待预测节点信息(即相邻节点融合向量)。例如,
Figure PCTCN2021074479-appb-000034
具体方法可以参见步骤402,在此不再赘述。
In some embodiments, the splicing vector may be aggregated based on each group to obtain the corresponding node information to be predicted (ie, the adjacent node fusion vector). E.g,
Figure PCTCN2021074479-appb-000034
For the specific method, refer to step 402, which will not be repeated here.
在一些实施例中,将所述分组融合的待预测节点信息输入到所述异构图神经网络模型的下一层处理。在一些实施例中,异构图神经网络模型的下一层即表示向量层。在表示向量层,可以将所述待预测节点信息融合节点注意力权重或路径注意力权重中的至少一个来得到最终的表示学习向量,例如图5所示的实施例中将节点注意力权重和路径注意力权重都进行了融合,又例如,图6中只融合了路径注意力权重。在其他实施例中,在表示向量层,也可以不融合节点注意力权重和路径注意力权重,直接基于所述待预测节点信息得到最终的表示向量。In some embodiments, the grouped and fused node information to be predicted is input to the next layer of processing of the heterogeneous graph neural network model. In some embodiments, the next layer of the heterogeneous graph neural network model represents the vector layer. In the representation vector layer, the information of the node to be predicted can be fused with at least one of the node attention weight or the path attention weight to obtain the final representation learning vector. For example, in the embodiment shown in FIG. 5, the node attention weight and The path attention weights are all fused, and for example, in Figure 6, only the path attention weights are fused. In other embodiments, in the representation vector layer, the node attention weight and the path attention weight may not be merged, and the final representation vector is obtained directly based on the information of the node to be predicted.
图5是根据申请一些实施例所示的基于异构图神经网络模型进行预测方法的示例性子流程图。Fig. 5 is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model according to some embodiments of the application.
在一些实施例中,表示向量层可以基于输入的待预测节点信息、节点注意力权重和路径注意力权重,输出待预测节点的表示向量。In some embodiments, the representation vector layer may output the representation vector of the node to be predicted based on the input node information to be predicted, the node attention weight, and the path attention weight.
在一些实施例中,表示向量层可以先通过待预测节点的组聚合向量和相邻节点融合向量计算节点的注意力权重向量;再基于节点的注意力权重向量和相邻节点融合向量获取对应的信息加权融合向量;然后再通过信息加权融合向量和路径的特征向量计算路径的注意力权重向量;最终由信息加权融合向量和路径的注意力权重向量得到待预测节点的表示向量。In some embodiments, the presentation vector layer can first calculate the attention weight vector of the node through the group aggregation vector of the node to be predicted and the adjacent node fusion vector; then obtain the corresponding attention weight vector based on the node's attention weight vector and the adjacent node fusion vector Information weighted fusion vector; then the attention weight vector of the path is calculated by the information weighted fusion vector and the feature vector of the path; finally, the representation vector of the node to be predicted is obtained from the information weighted fusion vector and the attention weight vector of the path.
步骤502,基于邻居节点的重要程度确定节点的注意力权重向量。Step 502: Determine the attention weight vector of the node based on the importance of neighbor nodes.
邻居节点的注意力权重向量是以各组邻居节点的重要程度为元素的向量。邻居节点的重要程度表征了对于不同的待预测节点,各组邻居节点的特征信息在计算待预测节点的表示向量过程中被利用的程度。The attention weight vector of neighbor nodes is a vector whose elements are the importance of each group of neighbor nodes. The importance of neighbor nodes represents the degree to which the feature information of each group of neighbor nodes is used in the process of calculating the representation vector of the node to be predicted for different nodes to be predicted.
例如,在用户借贷风险预测时,用户交互关系类型的路径重要程度小,用户收入等级类型的路径重要程度大。For example, in the prediction of user loan risk, the path importance of the user interaction relationship type is low, and the path importance of the user income level type is high.
在一些实施例中,各组邻居节点的注意力权重向量可以通过待预测节点的特征向量和每组邻居节点的相邻节点融合向量确定,包括:针对每一组邻居节点的融合向量,将该相邻节点融合向量与所述待预测节点特征向量拼接,并将基于拼接得到的向量乘以预设的第三权重矩阵,得到对应于该相邻节点融合向量的所述节点的注意力权重向量。In some embodiments, the attention weight vector of each group of neighbor nodes can be determined by the feature vector of the node to be predicted and the neighbor node fusion vector of each group of neighbor nodes, including: The adjacent node fusion vector is spliced with the feature vector of the node to be predicted, and the vector obtained based on the splicing is multiplied by the preset third weight matrix to obtain the attention weight vector of the node corresponding to the adjacent node fusion vector .
在一些实施例中,节点的注意力权重向量可以参考步骤402中邻居权重的获取方式,在此不再赘述。In some embodiments, the attention weight vector of the node can refer to the way of obtaining the neighbor weight in step 402, which will not be repeated here.
步骤504,基于节点的注意力权重向量和相邻节点融合向量获取对应的信息加权融合向量。Step 504: Obtain a corresponding information weighted fusion vector based on the attention weight vector of the node and the adjacent node fusion vector.
在一些实施例中,所述相邻节点融合向量与所述节点注意力权重向量的维度相同,且所述节点注意力权重向量的每一元素用于表征所述相邻节点融合向量相同位置的元素的重要程度。所述针对每一所述相邻节点融合向量,基于对应于该相邻节点融合向量的所述节点注意力权重向量与该相邻节点融合向量,得到对应于该相邻节点融合向量的信息加权融合向量,包括:针对每一所述相邻节点融合向量,将该相邻节点融合向量的每一元素与对应于该相邻节点融合向量的节点注意力权重向量的相同位置的元素相乘,并将得到的乘积作为所述信息加权融合向量的相同位置的元素,以得到所述信息加权融合向量。In some embodiments, the adjacent node fusion vector and the node attention weight vector have the same dimension, and each element of the node attention weight vector is used to represent the same position of the adjacent node fusion vector. The importance of the element. For each of the adjacent node fusion vectors, the information weight corresponding to the adjacent node fusion vector is obtained based on the node attention weight vector corresponding to the adjacent node fusion vector and the adjacent node fusion vector The fusion vector includes: for each adjacent node fusion vector, multiplying each element of the adjacent node fusion vector by the element at the same position of the node attention weight vector corresponding to the adjacent node fusion vector, And use the obtained product as the element at the same position of the information weighted fusion vector to obtain the information weighted fusion vector.
例如,维度为1*6的相邻节点融合向量
Figure PCTCN2021074479-appb-000035
为[A1,B1,C1,D1,E1,F1],其对应的维度为1*6的节点注意力权重向量
Figure PCTCN2021074479-appb-000036
为[A2,B2,C2,D2,E2,F2],在一种可能的实施方式中,A2用于表征A1的重要程度,B2用于表征B1的重要程度,其余元素依次类推。则基于上述元素间的乘法得到的对应于该相邻节点融合向量
Figure PCTCN2021074479-appb-000037
的信息加权融合向量
Figure PCTCN2021074479-appb-000038
为[A1*A2,B1*B2,C1*C2,D1*D2,E1*E2,F1*F2]。同理,信息加权融合向量
Figure PCTCN2021074479-appb-000039
可以基于相同的方法得到。
For example, the fusion vector of adjacent nodes with dimension 1*6
Figure PCTCN2021074479-appb-000035
Is [A1, B1, C1, D1, E1, F1], and its corresponding dimension is 1*6 node attention weight vector
Figure PCTCN2021074479-appb-000036
It is [A2, B2, C2, D2, E2, F2], in a possible embodiment, A2 is used to characterize the importance of A1, B2 is used to characterize the importance of B1, and the rest of the elements are analogously. Then based on the multiplication between the above elements, the fusion vector corresponding to the adjacent node is obtained
Figure PCTCN2021074479-appb-000037
Information weighted fusion vector
Figure PCTCN2021074479-appb-000038
It is [A1*A2, B1*B2, C1*C2, D1*D2, E1*E2, F1*F2]. Similarly, the information weighted fusion vector
Figure PCTCN2021074479-appb-000039
It can be obtained based on the same method.
步骤506,基于路径的重要程度确定路径的注意力权重向量。Step 506: Determine the attention weight vector of the path based on the importance of the path.
在一些实施例中,路径的注意力权重向量是基于信息加权融合向量得到。In some embodiments, the attention weight vector of the path is obtained based on the information weighted fusion vector.
在一些实施例中,路径的注意力权重向量的获取方法可以参考步骤402中特征权重向量的获取。In some embodiments, the method for obtaining the attention weight vector of the path may refer to the obtaining of the feature weight vector in step 402.
例如,基于维度为1*6的信息加权融合向量
Figure PCTCN2021074479-appb-000040
以及维度为6*1的第四权重矩阵,得到对应于该信息加权融合向量的路径注意力权重β j。同理对于其他m个信息加权融合向量,基于上述相同的方式得到其他的路径注意力权重:β 1,β 2,…β m。将所有的路径注意力组合可以得到如图5所示的路径注意力权重向量β u。不同的信息加权融合向量的参数向量可以相同,参数向量可以通过对该通过实体对象数据判断实体对象类别的方法对应的模型进行训练得到。
For example, based on the information weighted fusion vector with a dimension of 1*6
Figure PCTCN2021074479-appb-000040
And a fourth weight matrix with a dimension of 6*1 to obtain the path attention weight β j corresponding to the information weighted fusion vector. In the same way, for other m information weighted fusion vectors, other path attention weights are obtained based on the same method as described above: β 1 , β 2 ,...β m . Combining all the path attentions can obtain the path attention weight vector β u as shown in Fig. 5. The parameter vectors of different information weighted fusion vectors can be the same, and the parameter vectors can be obtained by training the model corresponding to the method for judging the entity object category through the entity object data.
步骤508,基于所述分组融合的待预测节点信息、所述节点的注意力权重向量和所述路径的注意力权重向量确定所述待预测节点的表示向量。Step 508: Determine the representation vector of the node to be predicted based on the grouped fusion of the node information to be predicted, the attention weight vector of the node, and the attention weight vector of the path.
在一些实施例中,基于所述信息加权融合向量与对应于该信息加权融合向量的路径注意力权重,得到所述待预测节点的信息表示学习向量,包括:将每一所述信息加权融合向量
Figure PCTCN2021074479-appb-000041
与对应于该信息加权融合向量β u中的路径注意力权重β j相乘,并将所述相乘得到的所有向量相加得到所述信息表示学习向量e u,如:
In some embodiments, obtaining the information representation learning vector of the node to be predicted based on the information weighted fusion vector and the path attention weight corresponding to the information weighted fusion vector includes: weighting each information fusion vector
Figure PCTCN2021074479-appb-000041
Multiply it with the path attention weight β j in the information-weighted fusion vector β u , and add all the vectors obtained by the multiplication to obtain the information representation learning vector e u , such as:
Figure PCTCN2021074479-appb-000042
Figure PCTCN2021074479-appb-000042
举例来讲,经过上述步骤的计算,得到共2个信息加权融合向量
Figure PCTCN2021074479-appb-000043
其对应的路径的注意力权重向量依次为β 1,β 2,则得到的信息表示学习向量e u
Figure PCTCN2021074479-appb-000044
Figure PCTCN2021074479-appb-000045
可直接将该信息表示学习向量e u用于相关计算例如直接输入预测模型进行计算。
For example, after the calculation of the above steps, a total of 2 information weighted fusion vectors are obtained
Figure PCTCN2021074479-appb-000043
The attention weight vector of the corresponding path is β 1 , β 2 , and the obtained information indicates that the learning vector e u is
Figure PCTCN2021074479-appb-000044
Figure PCTCN2021074479-appb-000045
The information can be directly represented by the learning vector e u for related calculations, for example, directly input into the prediction model for calculation.
由于融合了路径注意力权重,节点注意力权重,以及邻居节点与待预测节点的特 征信息,因此该信息表示学习向量能表达更丰富的信息,使得后续的预测结果更准确。Because of the integration of path attention weight, node attention weight, and characteristic information of neighbor nodes and nodes to be predicted, this information indicates that the learning vector can express richer information and make subsequent prediction results more accurate.
如前文所述,在一些是实施例中,还可以不融合节点注意力权重,基于路径注意力权重以及待预测节点的特征信息来得到对应的表示学习向量e u,即参见图6,图6是根据本说明书另一些实施例所示的基于异构图神经网络模型进行预测方法的示例性子流程图。 As mentioned above, in some embodiments, the attention weight of the node may not be merged, and the corresponding representation learning vector e u is obtained based on the path attention weight and the feature information of the node to be predicted, ie, see Fig. 6 and Fig. 6 It is an exemplary sub-flow chart of a prediction method based on a heterogeneous graph neural network model shown in other embodiments of this specification.
步骤602和步骤604基于分组后的邻居节点得到待预测节点信息,具体描述可分别参照步骤402和404,在此不再赘述。步骤606和608分别是基于路径的重要度程度确定路径的注意力权重,以及基于上述待预测节点信息和路径的注意力权重确定待预测节点的表示学习向量e u,具体方法可参见图5中步骤506和508,在此不再赘述。 Step 602 and step 604 obtain the node information to be predicted based on the grouped neighbor nodes. For detailed description, please refer to steps 402 and 404 respectively, which will not be repeated here. Steps 606 and 608 respectively determine the attention weight of the path based on the importance of the path, and determine the representation learning vector e u of the node to be predicted based on the above-mentioned node information to be predicted and the attention weight of the path. The specific method can be seen in Figure 5 Steps 506 and 508 are not repeated here.
在一些实施例中,可以基于大量带有标识的训练样本对异构图网络模型和预测模型进行端到端(End-to-end)训练。所述端到端训练是指多个模型在训练过程中,按照模型对数据的处理步骤,从第一个模型的输入端输入数据,从最后一个模型的输出端得到结果,基于结果与真实值的误差对所有模型的参数进行迭代调整,直到模型满足截止条件。所述端到端的训练可以省去训练每一个独立模型所需要的训练数据,同时模型之间的训练结果不会相互影响。In some embodiments, end-to-end training may be performed on the heterogeneous graph network model and the prediction model based on a large number of training samples with identifications. The end-to-end training refers to multiple models in the training process, according to the model's data processing steps, input data from the input end of the first model, and obtain the result from the output end of the last model, based on the result and the true value The error of all models are adjusted iteratively until the model meets the cut-off condition. The end-to-end training can save the training data required for training each independent model, and at the same time, the training results between the models will not affect each other.
具体的,在一些实施例中,可以将带有标识的训练样本输入异构图网络模型,再将异构网络模型输出的表示向量输入预测模型,基于预测模型输出值构建损失函数,通过训练同时迭代更新异构图网络模型和预测模型的参数。Specifically, in some embodiments, the training samples with the logo may be input into the heterogeneous graph network model, and then the representation vector output by the heterogeneous network model is input into the prediction model, and the loss function is constructed based on the output value of the prediction model. Iteratively update the parameters of the heterogeneous graph network model and the prediction model.
在一些实施例中,训练样本可以是若干份与目标对象相关(例如,该用户的信用度;或者该用户的电影喜好类型)的异构图数据,每份异构图数据包括相关各个节点数据、相邻节点数据以及各节点间的路径信息。所述异构图数据可以基于获取的历史数据信息进行构建,例如根据获取的用户个人数据、用户对电影类型的偏好数据、对电影演员的偏好数据以及对电影导演的偏好数据等进行构建。In some embodiments, the training sample may be several pieces of heterogeneous graph data related to the target object (for example, the user's credit rating; or the user's movie preference type), and each piece of heterogeneous graph data includes related node data, Adjacent node data and path information between each node. The heterogeneous graph data can be constructed based on acquired historical data information, for example, based on acquired user personal data, user preference data for movie types, preference data for movie actors, preference data for movie directors, and the like.
然后,对所述训练样本中的每一份异构图数据进行标记,所述标记数据包括每一份异构图数据中目标内容的评价结果(例如,该用户的信用度良好;或者该用户喜欢的电影类型为喜剧电影)。其中,所述评价结果也可以通过获取所述目标对象(例如,该用户)的历史评价信息来确定。Then, each piece of heterogeneous graph data in the training sample is labeled, and the labeled data includes the evaluation result of the target content in each piece of heterogeneous graph data (for example, the user’s credit rating is good; or the user likes The movie genre is comedy movie). Wherein, the evaluation result may also be determined by obtaining historical evaluation information of the target object (for example, the user).
下面将基于上述带有标记的异构图数据输入异构图网络模型和预测模型进行同时训练,具体的,将所述带有样本标识的异构图数据输入异构图网络模型,将异构图网络模型输出的表示向量作为预测模型的输入数据,所述异构图数据对应的标记数据作为预测模型的输出数据,并将输入数据和输出数据输入预测模型进行训练。在训练过程中,可以基于预测模型的实际输出值构建损失函数,并通过该损失函数同时迭代更新异构图网络模型混合预测模型的参数。The following will be based on the above labeled heterogeneous graph data input into the heterogeneous graph network model and the prediction model for simultaneous training. Specifically, the heterogeneous graph data with sample identification will be input into the heterogeneous graph network model, and the heterogeneous graph data will be input into the heterogeneous graph network model. The representation vector output by the graph network model is used as the input data of the prediction model, the labeled data corresponding to the heterogeneous graph data is used as the output data of the prediction model, and the input data and output data are input to the prediction model for training. In the training process, a loss function can be constructed based on the actual output value of the prediction model, and the parameters of the heterogeneous graph network model hybrid prediction model can be iteratively updated through the loss function.
在一些实施例中,异构网络模型的参数可以包括第一参数矩阵、第二参数矩阵、第一权重矩阵、第二权重矩阵等,预测模型的参数可以包括二分类模型、逻辑回归模型和神经网络模型的权重、阈值等。In some embodiments, the parameters of the heterogeneous network model may include a first parameter matrix, a second parameter matrix, a first weight matrix, a second weight matrix, etc., and the parameters of a prediction model may include a binary classification model, a logistic regression model, and a neural network. The weights, thresholds, etc. of the network model.
在一些实施例中,当训练的模型满足迭代截止条件时,训练结束。其中,迭代截止条件可以是损失函数结果收敛或小于阈值等。In some embodiments, when the trained model meets the iteration cutoff condition, the training ends. Among them, the iterative cut-off condition can be that the result of the loss function converges or is less than a threshold.
应当注意的是,上述有关流程的描述仅仅是为了示例和说明,而不限定本说明书的适用范围。对于本领域技术人员来说,在本说明书的指导下可以对流程进行各种修正和改变。然而,这些修正和改变仍在本说明书的范围之内。It should be noted that the above description of the related process is only for example and description, and does not limit the scope of application of this specification. For those skilled in the art, various modifications and changes can be made to the process under the guidance of this specification. However, these corrections and changes are still within the scope of this specification.
本说明书实施例可能带来的有益效果包括但不限于:(1)该通过实体对象数据判断实体对象类别的方法结合了待预测节点与邻居节点的特征信息,以及节点注意力权重和路径注意力权重,充分提取了异构图中各方面的信息,使得得到的信息表示学习向量包含的信息更加丰富,进而输入预测模型的预测结果更加准确;(2)对于同一类型的路径,融合了不同的邻居节点的特征信息,有效的抽取和利用了异构图中的结构信息,使得得到的信息学习表示向量包含更丰富的语义信息。需要说明的是,不同实施例可能产生的有益效果不同,在不同的实施例里,可能产生的有益效果可以是以上任意一种或几种的组合,也可以是其他任何可能获得的有益效果。The beneficial effects that the embodiments of this specification may bring include, but are not limited to: (1) The method for judging the entity object category through entity object data combines the feature information of the node to be predicted and the neighbor node, as well as the node attention weight and path attention The weight fully extracts all aspects of the information in the heterogeneous graph, so that the obtained information indicates that the information contained in the learning vector is richer, and the prediction result of the input prediction model is more accurate; (2) For the same type of path, different The feature information of neighbor nodes effectively extracts and utilizes the structural information in the heterogeneous graph, so that the obtained information learning representation vector contains richer semantic information. It should be noted that different embodiments may have different beneficial effects. In different embodiments, the possible beneficial effects may be any one or a combination of the above, or any other beneficial effects that may be obtained.
上文已对基本概念做了描述,显然,对于本领域技术人员来说,上述详细披露仅仅作为示例,而并不构成对本说明书的限定。虽然此处并没有明确说明,本领域技术人员可能会对本说明书进行各种修改、改进和修正。该类修改、改进和修正在本说明书中被建议,所以该类修改、改进、修正仍属于本说明书示范实施例的精神和范围。The basic concepts have been described above. Obviously, for those skilled in the art, the above detailed disclosure is only an example, and does not constitute a limitation to this specification. Although it is not explicitly stated here, those skilled in the art may make various modifications, improvements and amendments to this specification. Such modifications, improvements, and corrections are suggested in this specification, so such modifications, improvements, and corrections still belong to the spirit and scope of the exemplary embodiments of this specification.
同时,本说明书使用了特定词语来描述本说明书的实施例。如“一个实施例”、“一实施例”、和/或“一些实施例”意指与本说明书至少一个实施例相关的某一特征、结构或特点。因此,应强调并注意的是,本说明书中在不同位置两次或多次提及的“一实施例”或“一个实施例”或“一个替代性实施例”并不一定是指同一实施例。此外,本说明书的一个或多个实施例中的某些特征、结构或特点可以进行适当的组合。Meanwhile, this specification uses specific words to describe the embodiments of this specification. For example, "one embodiment", "an embodiment", and/or "some embodiments" mean a certain feature, structure, or characteristic related to at least one embodiment of this specification. Therefore, it should be emphasized and noted that “one embodiment” or “one embodiment” or “an alternative embodiment” mentioned twice or more in different positions in this specification does not necessarily refer to the same embodiment. . In addition, some features, structures, or characteristics in one or more embodiments of this specification can be appropriately combined.
此外,本领域技术人员可以理解,本说明书的各方面可以通过若干具有可专利性的种类或情况进行说明和描述,包括任何新的和有用的工序、机器、产品或物质的组合,或对他们的任何新的和有用的改进。相应地,本说明书的各个方面可以完全由硬件执行、可以完全由软件(包括固件、常驻软件、微码等)执行、也可以由硬件和软件组合执行。以上硬件或软件均可被称为“数据块”、“模块”、“引擎”、“单元”、“组件”或“系统”。此外,本说明书的各方面可能表现为位于一个或多个计算机可读介质中的计算机产品,该产品包括计算机可读程序编码。In addition, those skilled in the art can understand that various aspects of this specification can be explained and described through a number of patentable categories or situations, including any new and useful process, machine, product, or combination of substances, or a combination of them. Any new and useful improvements. Correspondingly, various aspects of this specification can be completely executed by hardware, can be completely executed by software (including firmware, resident software, microcode, etc.), or can be executed by a combination of hardware and software. The above hardware or software can all be referred to as "data block", "module", "engine", "unit", "component" or "system". In addition, various aspects of this specification may be embodied as a computer product located in one or more computer-readable media, and the product includes computer-readable program codes.
计算机存储介质可能包含一个内含有计算机程序编码的传播数据信号,例如在基带上或作为载波的一部分。该传播信号可能有多种表现形式,包括电磁形式、光形式等,或合适的组合形式。计算机存储介质可以是除计算机可读存储介质之外的任何计算机可读介质,该介质可以通过连接至一个指令执行系统、装置或设备以实现通讯、传播或传输供使用的程序。位于计算机存储介质上的程序编码可以通过任何合适的介质进行传播,包括无线电、电缆、光纤电缆、RF、或类似介质,或任何上述介质的组合。The computer storage medium may contain a propagated data signal containing a computer program code, for example on a baseband or as part of a carrier wave. The propagated signal may have multiple manifestations, including electromagnetic forms, optical forms, etc., or a suitable combination. The computer storage medium may be any computer readable medium other than the computer readable storage medium, and the medium may be connected to an instruction execution system, device, or device to realize communication, propagation, or transmission of the program for use. The program code located on the computer storage medium can be transmitted through any suitable medium, including radio, cable, fiber optic cable, RF, or similar medium, or any combination of the above medium.
本说明书各部分操作所需的计算机程序编码可以用任意一种或多种程序语言编写,包括面向对象编程语言如Java、Scala、Smalltalk、Eiffel、JADE、Emerald、C++、C#、VB.NET、Python等,常规程序化编程语言如C语言、Visual Basic、Fortran 2003、Perl、 COBOL 2002、PHP、ABAP,动态编程语言如Python、Ruby和Groovy,或其他编程语言等。该程序编码可以完全在用户计算机上运行、或作为独立的软件包在用户计算机上运行、或部分在用户计算机上运行部分在远程计算机运行、或完全在远程计算机或服务器上运行。在后种情况下,远程计算机可以通过任何网络形式与用户计算机连接,比如局域网(LAN)或广域网(WAN),或连接至外部计算机(例如通过因特网),或在云计算环境中,或作为服务使用如软件即服务(SaaS)。The computer program codes required for the operation of each part of this manual can be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB.NET, Python Etc., conventional programming languages such as C language, Visual Basic, Fortran 2003, Perl, COBOL 2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages. The program code can be run entirely on the user's computer, or run as an independent software package on the user's computer, or partly run on the user's computer and partly run on a remote computer, or run entirely on the remote computer or server. In the latter case, the remote computer can be connected to the user's computer through any network form, such as a local area network (LAN) or a wide area network (WAN), or connected to an external computer (for example, via the Internet), or in a cloud computing environment, or as a service Use software as a service (SaaS).
此外,除非权利要求中明确说明,本说明书所述处理元素和序列的顺序、数字字母的使用、或其他名称的使用,并非用于限定本说明书流程和方法的顺序。尽管上述披露中通过各种示例讨论了一些目前认为有用的发明实施例,但应当理解的是,该类细节仅起到说明的目的,附加的权利要求并不仅限于披露的实施例,相反,权利要求旨在覆盖所有符合本说明书实施例实质和范围的修正和等价组合。例如,虽然以上所描述的系统组件可以通过硬件设备实现,但是也可以只通过软件的解决方案得以实现,如在现有的服务器或移动设备上安装所描述的系统。In addition, unless explicitly stated in the claims, the order of processing elements and sequences, the use of numbers and letters, or the use of other names described in this specification are not used to limit the order of processes and methods in this specification. Although the foregoing disclosure uses various examples to discuss some embodiments of the invention that are currently considered useful, it should be understood that such details are only for illustrative purposes, and the appended claims are not limited to the disclosed embodiments. On the contrary, the rights are The requirements are intended to cover all modifications and equivalent combinations that conform to the essence and scope of the embodiments of this specification. For example, although the system components described above can be implemented by hardware devices, they can also be implemented only by software solutions, such as installing the described system on an existing server or mobile device.
同理,应当注意的是,为了简化本说明书披露的表述,从而帮助对一个或多个发明实施例的理解,前文对本说明书实施例的描述中,有时会将多种特征归并至一个实施例、附图或对其的描述中。但是,这种披露方法并不意味着本说明书对象所需要的特征比权利要求中提及的特征多。实际上,实施例的特征要少于上述披露的单个实施例的全部特征。For the same reason, it should be noted that, in order to simplify the expressions disclosed in this specification and help the understanding of one or more embodiments of the invention, in the foregoing description of the embodiments of this specification, multiple features are sometimes combined into one embodiment. In the drawings or its description. However, this method of disclosure does not mean that the subject of the specification requires more features than those mentioned in the claims. In fact, the features of the embodiment are less than all the features of the single embodiment disclosed above.
一些实施例中使用了描述成分、属性数量的数字,应当理解的是,此类用于实施例描述的数字,在一些示例中使用了修饰词“大约”、“近似”或“大体上”来修饰。除非另外说明,“大约”、“近似”或“大体上”表明所述数字允许有±20%的变化。相应地,在一些实施例中,说明书和权利要求中使用的数值参数均为近似值,该近似值根据个别实施例所需特点可以发生改变。在一些实施例中,数值参数应考虑规定的有效数位并采用一般位数保留的方法。尽管本说明书一些实施例中用于确认其范围广度的数值域和参数为近似值,在具体实施例中,此类数值的设定在可行范围内尽可能精确。In some embodiments, numbers describing the number of ingredients and attributes are used. It should be understood that such numbers used in the description of the embodiments use the modifier "about", "approximately" or "substantially" in some examples. Retouch. Unless otherwise stated, "approximately", "approximately" or "substantially" indicates that the number is allowed to vary by ±20%. Correspondingly, in some embodiments, the numerical parameters used in the specification and claims are approximate values, and the approximate values can be changed according to the required characteristics of individual embodiments. In some embodiments, the numerical parameter should consider the prescribed effective digits and adopt the method of general digit retention. Although the numerical ranges and parameters used to confirm the breadth of the ranges in some embodiments of this specification are approximate values, in specific embodiments, the setting of such numerical values is as accurate as possible within the feasible range.
针对本说明书引用的每个专利、专利申请、专利申请公开物和其他材料,如文章、书籍、说明书、出版物、文档等,特此将其全部内容并入本说明书作为参考。与本说明书内容不一致或产生冲突的申请历史文件除外,对本说明书权利要求最广范围有限制的文件(当前或之后附加于本说明书中的)也除外。需要说明的是,如果本说明书附属材料中的描述、定义、和/或术语的使用与本说明书所述内容有不一致或冲突的地方,以本说明书的描述、定义和/或术语的使用为准。For each patent, patent application, patent application publication and other materials cited in this specification, such as articles, books, specifications, publications, documents, etc., the entire contents are hereby incorporated into this specification as a reference. The application history documents that are inconsistent or conflict with the content of this specification are excluded, and the documents that restrict the broadest scope of the claims of this specification (currently or later appended to this specification) are also excluded. It should be noted that if there is any inconsistency or conflict between the description, definition, and/or use of terms in the accompanying materials of this manual and the content of this manual, the description, definition and/or use of terms in this manual shall prevail. .
最后,应当理解的是,本说明书中所述实施例仅用以说明本说明书实施例的原则。其他的变形也可能属于本说明书的范围。因此,作为示例而非限制,本说明书实施例的替代配置可视为与本说明书的教导一致。相应地,本说明书的实施例不仅限于本说明书明确介绍和描述的实施例。Finally, it should be understood that the embodiments described in this specification are only used to illustrate the principles of the embodiments of this specification. Other variations may also fall within the scope of this specification. Therefore, as an example and not a limitation, the alternative configuration of the embodiment of the present specification can be regarded as consistent with the teaching of the present specification. Accordingly, the embodiments of this specification are not limited to the embodiments explicitly introduced and described in this specification.

Claims (18)

  1. 一种基于异构图神经网络模型进行预测的方法,所述方法包括:A method for prediction based on a heterogeneous graph neural network model, the method comprising:
    获取与预测内容相关的异构图数据,所述异构图数据包括待预测节点、所述待预测节点的邻居节点、以及连接所述待预测节点与所述邻居节点之间的路径,所述路径包括至少一种类型;Acquiring heterogeneous graph data related to the predicted content, the heterogeneous graph data including a node to be predicted, neighbor nodes of the node to be predicted, and a path connecting the node to be predicted and the neighbor node, the The path includes at least one type;
    基于所述路径的类型,对所述邻居节点进行分组,以使得同一组的所述邻居节点的路径的类型相同;Grouping the neighboring nodes based on the type of the path, so that the types of paths of the neighboring nodes in the same group are the same;
    将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量后输入训练好的预测模型进行预测。The node to be predicted, the grouped neighbor nodes, and the path between the nodes are input into the trained heterogeneous graph neural network model, and the representation vector of the node to be predicted is obtained and then input into the trained prediction model for prediction.
  2. 如权利要求1所述的方法,将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量还包括:The method according to claim 1, wherein inputting the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model, and obtaining the representation vector of the node to be predicted further comprises:
    融合节点注意力权重和/或路径注意力权重后得到所述待预测节点的表示向量。After fusing the node attention weight and/or the path attention weight, the representation vector of the node to be predicted is obtained.
  3. 如权利要求1所述的方法,将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量包括:The method according to claim 1, wherein inputting the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model to obtain the representation vector of the node to be predicted comprises:
    基于路径的重要程度确定路径的注意力权重;Determine the attention weight of the path based on the importance of the path;
    基于所述分组融合的待预测节点信息和所述路径的注意力权重确定所述待预测节点的表示向量。The representation vector of the node to be predicted is determined based on the grouped fusion of the node information to be predicted and the attention weight of the path.
  4. 如权利要求3所述的方法,将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量还包括:The method according to claim 3, inputting the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model, and obtaining the representation vector of the node to be predicted further comprises:
    基于邻居节点的重要程度确定节点的注意力权重,基于所述分组融合的待预测节点信息、所述节点的注意力权重和所述路径的注意力权重确定所述待预测节点的表示向量。The attention weight of the node is determined based on the importance of the neighbor nodes, and the representation vector of the node to be predicted is determined based on the information of the node to be predicted that is fused by the group, the attention weight of the node, and the attention weight of the path.
  5. 如权利要求1所述的方法,所述预测内容包括预测目标对象的类别、风险等级或者偏好习惯。The method according to claim 1, wherein the prediction content includes predicting the category, risk level, or preference habits of the target object.
  6. 如权利要求1所述的方法,所述训练好的异构图神经网络模型和训练好的预测模型采用如下端到端训练获得:The method according to claim 1, wherein the trained heterogeneous graph neural network model and the trained prediction model are obtained by the following end-to-end training:
    基于预测模型的损失函数迭代更新预测模型以及异构图神经网络模型的模型参数,直到满足迭代截止条件。Based on the loss function of the prediction model, iteratively update the model parameters of the prediction model and the heterogeneous graph neural network model until the iteration cut-off condition is met.
  7. 如权利要求6所述的方法,所述端到端训练还包括:The method of claim 6, wherein the end-to-end training further comprises:
    将若干个异构图数据作为训练数据,将对应于该异构图数据的节点正确结果作为该训练数据的标签数据,所述预测模型的参数和异构图神经网络模型的参数利用所述训练数据和所述标签数据通过训练迭代更新。Several heterogeneous graph data are used as training data, and the correct result of the node corresponding to the heterogeneous graph data is used as the label data of the training data. The parameters of the prediction model and the parameters of the heterogeneous graph neural network model use the training The data and the label data are updated through training iterations.
  8. 如权利要求1所述的方法,所述将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量包括:The method according to claim 1, wherein the inputting the node to be predicted, the neighbor nodes after grouping, and the path between the nodes into the trained heterogeneous graph neural network model to obtain the representation vector of the node to be predicted comprises :
    根据同一组的所述邻居节点的特征信息,逐一得到对应于该组的组聚合向量;According to the feature information of the neighbor nodes of the same group, obtain the group aggregation vector corresponding to the group one by one;
    针对每一所述组聚合向量,将该待预测节点特征信息融合,得到分组融合的待预测节点信息。For each group of aggregation vectors, the feature information of the node to be predicted is fused to obtain grouped and fused node information to be predicted.
  9. 一种基于异构图神经网络模型进行预测的系统,所述系统包括:A prediction system based on a heterogeneous graph neural network model, the system comprising:
    异构图获取模块,用于获取与预测内容相关的异构图数据,所述异构图数据包括待预测节点、所述待预测节点的邻居节点、以及连接所述待预测节点与所述邻居节点之间的路径,所述路径包括至少一种类型;Heterogeneous graph acquisition module for acquiring heterogeneous graph data related to predicted content, the heterogeneous graph data including a node to be predicted, neighbor nodes of the node to be predicted, and connections between the node to be predicted and the neighbor A path between nodes, the path includes at least one type;
    分组模块,用于基于所述路径的类型,对所述邻居节点进行分组,以使得同一组的所述邻居节点的路径的类型相同;A grouping module, configured to group the neighboring nodes based on the type of the path, so that the types of paths of the neighboring nodes in the same group are the same;
    节点预测模块,用于将所述待预测节点、分组后的所述邻居节点以及节点之间的路径输入训练好的异构图神经网络模型,得到待预测节点的表示向量后输入训练好的预测模型进行预测。The node prediction module is used to input the node to be predicted, the grouped neighbor nodes, and the path between the nodes into the trained heterogeneous graph neural network model to obtain the representation vector of the node to be predicted and then input the trained prediction The model makes predictions.
  10. 如权利要求9所述的系统,所述节点预测模块还用于:融合节点注意力权重和/或路径注意力权重后得到所述待预测节点的表示向量。The system according to claim 9, wherein the node prediction module is further used to obtain the representation vector of the node to be predicted after fusing the node attention weight and/or the path attention weight.
  11. 如权利要求9所述的系统,所述节点预测模块还用于:基于路径的重要程度确定路径的注意力权重;基于所述分组融合的待预测节点信息和所述路径的注意力权重确定所述待预测节点的表示向量。The system according to claim 9, wherein the node prediction module is further configured to: determine the attention weight of the path based on the importance of the path; State the representation vector of the node to be predicted.
  12. 如权利要求11所述的系统,所述节点预测模块还用于:The system according to claim 11, wherein the node prediction module is further configured to:
    基于邻居节点的重要程度确定节点的注意力权重,基于所述分组融合的待预测节点信息、所述节点的注意力权重和所述路径的注意力权重确定所述待预测节点的表示向量。The attention weight of the node is determined based on the importance of the neighbor nodes, and the representation vector of the node to be predicted is determined based on the information of the node to be predicted that is fused by the group, the attention weight of the node, and the attention weight of the path.
  13. 如权利要求9所述的系统,所述预测内容包括预测目标对象的类别、风险等级或者偏好习惯。The system according to claim 9, wherein the prediction content includes the category, risk level, or preference habits of the predicted target object.
  14. 如权利要求9所述的系统,所述系统还包括模型训练模块,用于采用如下端到端训练获得训练好的异构图神经网络模型和训练好的预测模型:9. The system according to claim 9, further comprising a model training module for obtaining a trained heterogeneous graph neural network model and a trained prediction model by adopting the following end-to-end training:
    基于预测模型的损失函数迭代更新预测模型以及异构图神经网络模型的模型参数,直到满足迭代截止条件。Based on the loss function of the prediction model, iteratively update the model parameters of the prediction model and the heterogeneous graph neural network model until the iteration cut-off condition is met.
  15. 如权利要求14所述的系统,所述模型训练模块还用于:The system according to claim 14, wherein the model training module is further used for:
    将若干个异构图数据作为训练数据,将对应于该异构图数据的节点正确结果作为该训练数据的标签数据,所述预测模型的参数和异构图神经网络模型的参数利用所述训练 数据和所述标签数据通过训练迭代更新。Several heterogeneous graph data are used as training data, and the correct result of the node corresponding to the heterogeneous graph data is used as the label data of the training data. The parameters of the prediction model and the parameters of the heterogeneous graph neural network model use the training The data and the label data are updated through training iterations.
  16. 如权利要求9所述的系统,所述节点预测模块还用于:The system according to claim 9, wherein the node prediction module is further configured to:
    根据同一组的所述邻居节点的特征信息,逐一得到对应于该组的组聚合向量;According to the feature information of the neighbor nodes of the same group, obtain the group aggregation vector corresponding to the group one by one;
    针对每一所述组聚合向量,将该待预测节点特征信息融合,得到分组融合的待预测节点信息。For each group of aggregation vectors, the feature information of the node to be predicted is fused to obtain grouped and fused node information to be predicted.
  17. 一种基于异构图神经网络模型进行预测的装置,包括处理器,其中,所述处理器用于执行如权利要求1~8中任一项所述的基于异构图神经网络模型进行预测的方法。A device for prediction based on a heterogeneous graph neural network model, comprising a processor, wherein the processor is configured to execute the method for prediction based on a heterogeneous graph neural network model according to any one of claims 1 to 8 .
  18. 一种计算机可读存储介质,所述存储介质存储计算机指令,当计算机读取存储介质中的计算机指令后,计算机执行如权利要求1~8中任一项所述的基于异构图神经网络模型进行预测的方法。A computer-readable storage medium that stores computer instructions. After the computer reads the computer instructions in the storage medium, the computer executes the heterogeneous graph-based neural network model according to any one of claims 1 to 8. Method of making predictions.
PCT/CN2021/074479 2020-03-10 2021-01-29 Prediction method and system based on heterogeneous graph neural network model WO2021179838A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010162355.6 2020-03-10
CN202010162355.6A CN111400560A (en) 2020-03-10 2020-03-10 Method and system for predicting based on heterogeneous graph neural network model

Publications (1)

Publication Number Publication Date
WO2021179838A1 true WO2021179838A1 (en) 2021-09-16

Family

ID=71434461

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/074479 WO2021179838A1 (en) 2020-03-10 2021-01-29 Prediction method and system based on heterogeneous graph neural network model

Country Status (2)

Country Link
CN (1) CN111400560A (en)
WO (1) WO2021179838A1 (en)

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113656604A (en) * 2021-10-19 2021-11-16 之江实验室 Medical term normalization system and method based on heterogeneous graph neural network
CN113744882A (en) * 2021-09-17 2021-12-03 腾讯科技(深圳)有限公司 Method, device and equipment for determining target area and storage medium
CN114066081A (en) * 2021-11-23 2022-02-18 北京恒通慧源大数据技术有限公司 Enterprise risk prediction method and device based on graph attention network and electronic equipment
CN114168804A (en) * 2021-12-17 2022-03-11 中国科学院自动化研究所 Similar information retrieval method and system based on heterogeneous subgraph neural network
CN114329099A (en) * 2021-11-22 2022-04-12 腾讯科技(深圳)有限公司 Overlapping community identification method, device, equipment, storage medium and program product
CN114399028A (en) * 2022-01-14 2022-04-26 马上消费金融股份有限公司 Information processing method, graph convolution neural network training method and electronic equipment
CN114398462A (en) * 2022-03-24 2022-04-26 之江实验室 Destination recommendation method and system based on multi-source heterogeneous information network
CN114528479A (en) * 2022-01-20 2022-05-24 华南理工大学 Event detection method based on multi-scale different composition embedding algorithm
CN114565053A (en) * 2022-03-10 2022-05-31 天津大学 Deep heterogeneous map embedding model based on feature fusion
CN114660993A (en) * 2022-05-25 2022-06-24 中科航迈数控软件(深圳)有限公司 Numerical control machine tool fault prediction method based on multi-source heterogeneous data feature dimension reduction
CN114780867A (en) * 2022-05-10 2022-07-22 杭州网易云音乐科技有限公司 Recommendation method, medium, device and computing equipment
CN114826963A (en) * 2022-03-31 2022-07-29 鹏城实验室 Internet of things equipment detection method and system based on equipment behaviors
CN114863685A (en) * 2022-07-06 2022-08-05 北京理工大学 Traffic participant trajectory prediction method and system based on risk acceptance degree
CN114900364A (en) * 2022-05-18 2022-08-12 桂林电子科技大学 High-level continuous threat detection method based on tracing graph and heterogeneous graph neural network
CN114895985A (en) * 2022-06-08 2022-08-12 华东师范大学 Data loading system for sampling-based graph neural network training
CN115118451A (en) * 2022-05-17 2022-09-27 北京理工大学 Network intrusion detection method combining graph embedded knowledge modeling
CN115130663A (en) * 2022-08-30 2022-09-30 中国海洋大学 Heterogeneous network attribute completion method based on graph neural network and attention mechanism
CN115221976A (en) * 2022-08-18 2022-10-21 抖音视界有限公司 Model training method and device based on graph neural network
CN115660688A (en) * 2022-10-24 2023-01-31 西南财经大学 Financial transaction abnormity detection method and cross-region sustainable training method thereof
CN115660147A (en) * 2022-09-26 2023-01-31 哈尔滨工业大学 Information propagation prediction method and system based on influence modeling between propagation paths and in propagation paths
CN115713986A (en) * 2022-11-11 2023-02-24 中南大学 Attention mechanism-based material crystal property prediction method
CN115983148A (en) * 2022-12-13 2023-04-18 北京景行锐创软件有限公司 CFD simulation cloud picture prediction method, system, electronic device and medium
CN116109381A (en) * 2023-01-10 2023-05-12 深圳峰涛科技有限公司 E-commerce platform data processing method and system
CN116305995A (en) * 2023-03-27 2023-06-23 清华大学 Nonlinear analysis method, nonlinear analysis device, nonlinear analysis equipment and nonlinear analysis medium of structural system
CN116383446A (en) * 2023-04-06 2023-07-04 哈尔滨工程大学 Author classification method based on heterogeneous quotation network
CN116561688A (en) * 2023-05-09 2023-08-08 浙江大学 Emerging technology identification method based on dynamic graph anomaly detection
CN116578915A (en) * 2023-04-11 2023-08-11 广州极点三维信息科技有限公司 Structured house type analysis method and system based on graphic neural network
CN116578884A (en) * 2023-07-07 2023-08-11 北京邮电大学 Scientific research team identification method and device based on heterogeneous information network representation learning
CN116595157A (en) * 2023-07-17 2023-08-15 江西财经大学 Dynamic interest transfer type session recommendation method and system based on user intention fusion
WO2023207790A1 (en) * 2022-04-28 2023-11-02 华为技术有限公司 Classification model training method and device
CN117038105A (en) * 2023-10-08 2023-11-10 武汉纺织大学 Drug repositioning method and system based on information enhancement graph neural network
CN117151279A (en) * 2023-08-15 2023-12-01 哈尔滨工业大学 Isomorphic network link prediction method and system based on line graph neural network
CN117218459A (en) * 2023-11-08 2023-12-12 支付宝(杭州)信息技术有限公司 Distributed node classification method and device
CN117421671A (en) * 2023-12-18 2024-01-19 南开大学 Frequency self-adaptive static heterogeneous graph node classification method for quote network
CN117493490A (en) * 2023-11-17 2024-02-02 南京信息工程大学 Topic detection method, device, equipment and medium based on heterogeneous multi-relation graph
CN117648197A (en) * 2024-01-30 2024-03-05 西安电子科技大学 Serialized microservice resource prediction method based on countermeasure learning and heterograph learning
CN117828514A (en) * 2024-03-04 2024-04-05 清华大学深圳国际研究生院 User network behavior data anomaly detection method based on graph structure learning
CN117976139A (en) * 2024-03-29 2024-05-03 武汉纺织大学 Drug repositioning method and system based on deviation correcting mechanism and contrast learning
CN116578915B (en) * 2023-04-11 2024-06-11 广州极点三维信息科技有限公司 Structured house type analysis method and system based on graphic neural network

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111400560A (en) * 2020-03-10 2020-07-10 支付宝(杭州)信息技术有限公司 Method and system for predicting based on heterogeneous graph neural network model
CN114063459B (en) * 2020-08-10 2024-03-15 海信集团控股股份有限公司 Terminal and intelligent home control method
CN115867919A (en) * 2020-08-17 2023-03-28 华为技术有限公司 Graph structure aware incremental learning for recommendation systems
CN112037038B (en) * 2020-09-02 2024-05-28 中国银行股份有限公司 Bank credit risk prediction method and device
CN112036418A (en) * 2020-09-04 2020-12-04 京东数字科技控股股份有限公司 Method and device for extracting user features
CN112035669B (en) * 2020-09-09 2021-05-14 中国科学技术大学 Social media multi-modal rumor detection method based on propagation heterogeneous graph modeling
CN112085172B (en) * 2020-09-16 2022-09-16 支付宝(杭州)信息技术有限公司 Method and device for training graph neural network
CN116249987A (en) * 2020-09-22 2023-06-09 维萨国际服务协会 Graph-based learning system with update vectors
CN112070216B (en) * 2020-09-29 2023-06-02 支付宝(杭州)信息技术有限公司 Method and system for training graph neural network model based on graph computing system
CN112017784B (en) * 2020-10-22 2021-02-09 平安科技(深圳)有限公司 Coronary heart disease risk prediction method based on multi-modal data and related equipment
CN112257806B (en) * 2020-10-30 2023-06-20 福建师范大学 Heterogeneous user-oriented migration learning method
CN112257959A (en) * 2020-11-12 2021-01-22 上海优扬新媒信息技术有限公司 User risk prediction method and device, electronic equipment and storage medium
CN112529302A (en) * 2020-12-15 2021-03-19 中国人民大学 Method and system for predicting success rate of patent application authorization and electronic equipment
CN112561688A (en) * 2020-12-21 2021-03-26 第四范式(北京)技术有限公司 Credit card overdue prediction method and device based on graph embedding and electronic equipment
CN112861967B (en) * 2021-02-07 2023-04-07 中国电子科技集团公司电子科学研究院 Social network abnormal user detection method and device based on heterogeneous graph neural network
CN113010687B (en) * 2021-03-03 2023-02-03 广州视源电子科技股份有限公司 Exercise label prediction method and device, storage medium and computer equipment
CN113065950A (en) * 2021-04-22 2021-07-02 中国工商银行股份有限公司 Credit card limit evaluation method and device
CN113298221B (en) * 2021-04-26 2023-08-22 上海淇玥信息技术有限公司 User Risk Prediction Method and Device Based on Logistic Regression and Graph Neural Network
CN113095592A (en) * 2021-04-30 2021-07-09 第四范式(北京)技术有限公司 Method and system for performing predictions based on GNN and training method and system
CN113191565B (en) * 2021-05-18 2023-04-07 同盾科技有限公司 Security prediction method, security prediction device, security prediction medium, and security prediction apparatus
CN113239875B (en) * 2021-06-01 2023-10-17 恒睿(重庆)人工智能技术研究院有限公司 Method, system and device for acquiring face characteristics and computer readable storage medium
CN116150425A (en) * 2021-11-19 2023-05-23 腾讯科技(深圳)有限公司 Recommended content selection method, apparatus, device, storage medium and program product
CN114186069B (en) * 2021-11-29 2023-09-29 江苏大学 Depth video understanding knowledge graph construction method based on multi-mode different-composition attention network
CN115455438B (en) * 2022-11-09 2023-02-07 南昌航空大学 Program slicing vulnerability detection method, system, computer and storage medium
CN116094827A (en) * 2023-01-18 2023-05-09 支付宝(杭州)信息技术有限公司 Safety risk identification method and system based on topology enhancement
CN116305461B (en) * 2023-03-13 2023-10-13 清华大学 Structure response calculation method, device, electronic equipment and storage medium
CN116127204B (en) * 2023-04-17 2023-07-18 中国科学技术大学 Multi-view user portrayal method, multi-view user portrayal system, apparatus, and medium
CN116757262B (en) * 2023-08-16 2024-01-12 苏州浪潮智能科技有限公司 Training method, classifying method, device, equipment and medium of graph neural network

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017147157A1 (en) * 2016-02-23 2017-08-31 Linkedin Corporation Graph framework using heterogeneous social networks
CN107145527A (en) * 2017-04-14 2017-09-08 东南大学 Link prediction method based on first path in alignment isomery social networks
CN107491540A (en) * 2017-08-24 2017-12-19 济南浚达信息技术有限公司 A kind of combination depth Bayesian model and the film of collaboration Heterogeneous Information insertion recommend method
CN107944629A (en) * 2017-11-30 2018-04-20 北京邮电大学 A kind of recommendation method and device based on heterogeneous information network representation
CN110569437A (en) * 2019-09-05 2019-12-13 腾讯科技(深圳)有限公司 click probability prediction and page content recommendation methods and devices
CN110851662A (en) * 2019-11-05 2020-02-28 中国人民解放军国防科技大学 Heterogeneous information network link prediction method based on meta path
CN111400560A (en) * 2020-03-10 2020-07-10 支付宝(杭州)信息技术有限公司 Method and system for predicting based on heterogeneous graph neural network model

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109767008A (en) * 2019-01-07 2019-05-17 武汉大学 A kind of polymorphic feature learning method of high isomerism network based on meta schema
CN110677284B (en) * 2019-09-24 2022-06-17 北京工商大学 Heterogeneous network link prediction method based on meta path
CN110598130B (en) * 2019-09-30 2022-06-24 重庆邮电大学 Movie recommendation method integrating heterogeneous information network and deep learning

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017147157A1 (en) * 2016-02-23 2017-08-31 Linkedin Corporation Graph framework using heterogeneous social networks
CN107145527A (en) * 2017-04-14 2017-09-08 东南大学 Link prediction method based on first path in alignment isomery social networks
CN107491540A (en) * 2017-08-24 2017-12-19 济南浚达信息技术有限公司 A kind of combination depth Bayesian model and the film of collaboration Heterogeneous Information insertion recommend method
CN107944629A (en) * 2017-11-30 2018-04-20 北京邮电大学 A kind of recommendation method and device based on heterogeneous information network representation
CN110569437A (en) * 2019-09-05 2019-12-13 腾讯科技(深圳)有限公司 click probability prediction and page content recommendation methods and devices
CN110851662A (en) * 2019-11-05 2020-02-28 中国人民解放军国防科技大学 Heterogeneous information network link prediction method based on meta path
CN111400560A (en) * 2020-03-10 2020-07-10 支付宝(杭州)信息技术有限公司 Method and system for predicting based on heterogeneous graph neural network model

Cited By (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113744882A (en) * 2021-09-17 2021-12-03 腾讯科技(深圳)有限公司 Method, device and equipment for determining target area and storage medium
CN113744882B (en) * 2021-09-17 2023-09-19 腾讯科技(深圳)有限公司 Method, device, equipment and storage medium for determining target area
CN113656604A (en) * 2021-10-19 2021-11-16 之江实验室 Medical term normalization system and method based on heterogeneous graph neural network
CN114329099B (en) * 2021-11-22 2023-07-07 腾讯科技(深圳)有限公司 Overlapping community identification method, device, equipment, storage medium and program product
CN114329099A (en) * 2021-11-22 2022-04-12 腾讯科技(深圳)有限公司 Overlapping community identification method, device, equipment, storage medium and program product
CN114066081B (en) * 2021-11-23 2022-04-26 北京恒通慧源大数据技术有限公司 Enterprise risk prediction method and device based on graph attention network and electronic equipment
CN114066081A (en) * 2021-11-23 2022-02-18 北京恒通慧源大数据技术有限公司 Enterprise risk prediction method and device based on graph attention network and electronic equipment
CN114168804B (en) * 2021-12-17 2022-06-10 中国科学院自动化研究所 Similar information retrieval method and system based on heterogeneous subgraph neural network
CN114168804A (en) * 2021-12-17 2022-03-11 中国科学院自动化研究所 Similar information retrieval method and system based on heterogeneous subgraph neural network
CN114399028A (en) * 2022-01-14 2022-04-26 马上消费金融股份有限公司 Information processing method, graph convolution neural network training method and electronic equipment
CN114528479B (en) * 2022-01-20 2023-03-21 华南理工大学 Event detection method based on multi-scale heteromorphic image embedding algorithm
CN114528479A (en) * 2022-01-20 2022-05-24 华南理工大学 Event detection method based on multi-scale different composition embedding algorithm
CN114565053A (en) * 2022-03-10 2022-05-31 天津大学 Deep heterogeneous map embedding model based on feature fusion
CN114565053B (en) * 2022-03-10 2023-05-19 天津大学 Deep heterogeneous graph embedded model based on feature fusion
CN114398462A (en) * 2022-03-24 2022-04-26 之江实验室 Destination recommendation method and system based on multi-source heterogeneous information network
CN114826963B (en) * 2022-03-31 2023-07-14 鹏城实验室 Internet of things equipment detection method and system based on equipment behaviors
CN114826963A (en) * 2022-03-31 2022-07-29 鹏城实验室 Internet of things equipment detection method and system based on equipment behaviors
WO2023207790A1 (en) * 2022-04-28 2023-11-02 华为技术有限公司 Classification model training method and device
CN114780867B (en) * 2022-05-10 2023-11-03 杭州网易云音乐科技有限公司 Recommendation method, medium, device and computing equipment
CN114780867A (en) * 2022-05-10 2022-07-22 杭州网易云音乐科技有限公司 Recommendation method, medium, device and computing equipment
CN115118451B (en) * 2022-05-17 2023-09-08 北京理工大学 Network intrusion detection method combined with graph embedded knowledge modeling
CN115118451A (en) * 2022-05-17 2022-09-27 北京理工大学 Network intrusion detection method combining graph embedded knowledge modeling
CN114900364A (en) * 2022-05-18 2022-08-12 桂林电子科技大学 High-level continuous threat detection method based on tracing graph and heterogeneous graph neural network
CN114900364B (en) * 2022-05-18 2024-03-08 桂林电子科技大学 Advanced continuous threat detection method based on traceability graph and heterogeneous graph neural network
CN114660993A (en) * 2022-05-25 2022-06-24 中科航迈数控软件(深圳)有限公司 Numerical control machine tool fault prediction method based on multi-source heterogeneous data feature dimension reduction
CN114895985B (en) * 2022-06-08 2023-06-09 华东师范大学 Data loading system for graph neural network training based on sampling
CN114895985A (en) * 2022-06-08 2022-08-12 华东师范大学 Data loading system for sampling-based graph neural network training
CN114863685A (en) * 2022-07-06 2022-08-05 北京理工大学 Traffic participant trajectory prediction method and system based on risk acceptance degree
CN114863685B (en) * 2022-07-06 2022-09-27 北京理工大学 Traffic participant trajectory prediction method and system based on risk acceptance degree
CN115221976B (en) * 2022-08-18 2024-05-24 抖音视界有限公司 Model training method and device based on graph neural network
CN115221976A (en) * 2022-08-18 2022-10-21 抖音视界有限公司 Model training method and device based on graph neural network
CN115130663B (en) * 2022-08-30 2023-10-13 中国海洋大学 Heterogeneous network attribute completion method based on graph neural network and attention mechanism
CN115130663A (en) * 2022-08-30 2022-09-30 中国海洋大学 Heterogeneous network attribute completion method based on graph neural network and attention mechanism
CN115660147A (en) * 2022-09-26 2023-01-31 哈尔滨工业大学 Information propagation prediction method and system based on influence modeling between propagation paths and in propagation paths
CN115660688A (en) * 2022-10-24 2023-01-31 西南财经大学 Financial transaction abnormity detection method and cross-region sustainable training method thereof
CN115660688B (en) * 2022-10-24 2024-04-30 西南财经大学 Financial transaction anomaly detection method and cross-regional sustainable training method thereof
CN115713986A (en) * 2022-11-11 2023-02-24 中南大学 Attention mechanism-based material crystal property prediction method
CN115713986B (en) * 2022-11-11 2023-07-11 中南大学 Attention mechanism-based material crystal attribute prediction method
CN115983148A (en) * 2022-12-13 2023-04-18 北京景行锐创软件有限公司 CFD simulation cloud picture prediction method, system, electronic device and medium
CN115983148B (en) * 2022-12-13 2024-04-12 北京景行锐创软件有限公司 CFD simulation cloud image prediction method, system, electronic equipment and medium
CN116109381B (en) * 2023-01-10 2023-09-29 深圳峰涛科技有限公司 E-commerce platform data processing method and system
CN116109381A (en) * 2023-01-10 2023-05-12 深圳峰涛科技有限公司 E-commerce platform data processing method and system
CN116305995B (en) * 2023-03-27 2023-11-07 清华大学 Nonlinear analysis method, nonlinear analysis device, nonlinear analysis equipment and nonlinear analysis medium of structural system
CN116305995A (en) * 2023-03-27 2023-06-23 清华大学 Nonlinear analysis method, nonlinear analysis device, nonlinear analysis equipment and nonlinear analysis medium of structural system
CN116383446A (en) * 2023-04-06 2023-07-04 哈尔滨工程大学 Author classification method based on heterogeneous quotation network
CN116578915A (en) * 2023-04-11 2023-08-11 广州极点三维信息科技有限公司 Structured house type analysis method and system based on graphic neural network
CN116578915B (en) * 2023-04-11 2024-06-11 广州极点三维信息科技有限公司 Structured house type analysis method and system based on graphic neural network
CN116561688B (en) * 2023-05-09 2024-03-22 浙江大学 Emerging technology identification method based on dynamic graph anomaly detection
CN116561688A (en) * 2023-05-09 2023-08-08 浙江大学 Emerging technology identification method based on dynamic graph anomaly detection
CN116578884A (en) * 2023-07-07 2023-08-11 北京邮电大学 Scientific research team identification method and device based on heterogeneous information network representation learning
CN116578884B (en) * 2023-07-07 2023-10-31 北京邮电大学 Scientific research team identification method and device based on heterogeneous information network representation learning
CN116595157B (en) * 2023-07-17 2023-09-19 江西财经大学 Dynamic interest transfer type session recommendation method and system based on user intention fusion
CN116595157A (en) * 2023-07-17 2023-08-15 江西财经大学 Dynamic interest transfer type session recommendation method and system based on user intention fusion
CN117151279A (en) * 2023-08-15 2023-12-01 哈尔滨工业大学 Isomorphic network link prediction method and system based on line graph neural network
CN117038105A (en) * 2023-10-08 2023-11-10 武汉纺织大学 Drug repositioning method and system based on information enhancement graph neural network
CN117038105B (en) * 2023-10-08 2023-12-15 武汉纺织大学 Drug repositioning method and system based on information enhancement graph neural network
CN117218459A (en) * 2023-11-08 2023-12-12 支付宝(杭州)信息技术有限公司 Distributed node classification method and device
CN117218459B (en) * 2023-11-08 2024-01-26 支付宝(杭州)信息技术有限公司 Distributed node classification method and device
CN117493490A (en) * 2023-11-17 2024-02-02 南京信息工程大学 Topic detection method, device, equipment and medium based on heterogeneous multi-relation graph
CN117493490B (en) * 2023-11-17 2024-05-14 南京信息工程大学 Topic detection method, device, equipment and medium based on heterogeneous multi-relation graph
CN117421671A (en) * 2023-12-18 2024-01-19 南开大学 Frequency self-adaptive static heterogeneous graph node classification method for quote network
CN117421671B (en) * 2023-12-18 2024-03-05 南开大学 Frequency self-adaptive static heterogeneous graph node classification method for quote network
CN117648197B (en) * 2024-01-30 2024-05-03 西安电子科技大学 Serialized microservice resource prediction method based on countermeasure learning and heterograph learning
CN117648197A (en) * 2024-01-30 2024-03-05 西安电子科技大学 Serialized microservice resource prediction method based on countermeasure learning and heterograph learning
CN117828514B (en) * 2024-03-04 2024-05-03 清华大学深圳国际研究生院 User network behavior data anomaly detection method based on graph structure learning
CN117828514A (en) * 2024-03-04 2024-04-05 清华大学深圳国际研究生院 User network behavior data anomaly detection method based on graph structure learning
CN117976139A (en) * 2024-03-29 2024-05-03 武汉纺织大学 Drug repositioning method and system based on deviation correcting mechanism and contrast learning

Also Published As

Publication number Publication date
CN111400560A (en) 2020-07-10

Similar Documents

Publication Publication Date Title
WO2021179838A1 (en) Prediction method and system based on heterogeneous graph neural network model
US10609433B2 (en) Recommendation information pushing method, server, and storage medium
Mongia et al. Deep latent factor model for collaborative filtering
US10354184B1 (en) Joint modeling of user behavior
Li et al. Deep probabilistic matrix factorization framework for online collaborative filtering
Liu et al. Social recommendation with learning personal and social latent factors
Yang et al. Plateclick: Bootstrapping food preferences through an adaptive visual interface
Eckstein et al. Robust risk aggregation with neural networks
CN111639687A (en) Model training and abnormal account identification method and device
Ghanbari et al. Reconstruction of gene networks using prior knowledge
CN106651427B (en) Data association method based on user behaviors
Pan et al. Collaborative recommendation with multiclass preference context
WO2022011553A1 (en) Feature interaction via edge search
CN111340245B (en) Model training method and system
US20220270155A1 (en) Recommendation with neighbor-aware hyperbolic embedding
Khan et al. A study on relationship between prediction uncertainty and robustness to noisy data
US20240119266A1 (en) Method for Constructing AI Integrated Model, and AI Integrated Model Inference Method and Apparatus
Rahim et al. An efficient recommender system algorithm using trust data
Portier et al. Bootstrap testing of the rank of a matrix via least-squared constrained estimation
Guan et al. Enhanced SVD for collaborative filtering
CN112241920A (en) Investment and financing organization evaluation method, system and equipment based on graph neural network
Kim et al. Selection of the most probable best under input uncertainty
WO2023185125A1 (en) Product resource data processing method and apparatus, electronic device and storage medium
US20230124258A1 (en) Embedding optimization for machine learning models
CN112232360A (en) Image retrieval model optimization method, image retrieval device and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21767806

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21767806

Country of ref document: EP

Kind code of ref document: A1