CN115618124A - Propagation popularity prediction method for bidirectional social influence learning - Google Patents

Propagation popularity prediction method for bidirectional social influence learning Download PDF

Info

Publication number
CN115618124A
CN115618124A CN202211196722.XA CN202211196722A CN115618124A CN 115618124 A CN115618124 A CN 115618124A CN 202211196722 A CN202211196722 A CN 202211196722A CN 115618124 A CN115618124 A CN 115618124A
Authority
CN
China
Prior art keywords
graph
node
time
attention
propagation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211196722.XA
Other languages
Chinese (zh)
Inventor
王振宇
黄振华
吴志祥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN202211196722.XA priority Critical patent/CN115618124A/en
Publication of CN115618124A publication Critical patent/CN115618124A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9536Search customisation based on social or collaborative filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/01Social networking

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Business, Economics & Management (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Economics (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a propagation popularity prediction method for bidirectional social influence learning. The method comprises the following steps: collecting information transmission network data and constructing graph sequence data of information transmission; constructing a model based on a graph sequence attention network; constructing a macroscopic dynamic popularity prediction model based on the combination of a graph sequence attention network and a decoder; inputting the graph sequence data into a macroscopic dynamic popularity prediction model based on the combination of a graph sequence attention network and a decoder for iterative training; and inputting the graph sequence data of information propagation, and predicting the popularity of the information propagation at different moments by using the model obtained by training. The invention applies a multilevel bidirectional attention mechanism to learn the space-time characteristics of the transmission dynamic graph, effectively learns the local and global dependency relationship of the transmission network subgraphs at different moments, considers the bidirectional social influence information of the transmission subgraphs at the previous and next moments on the structure and has better usability.

Description

Propagation popularity prediction method for bidirectional social influence learning
Technical Field
The invention relates to the field of social network information propagation research, in particular to a propagation popularity prediction method for bidirectional social influence learning.
Background
The online social network provides a platform channel for information transmission, meets the requirements of users for expressing self, transmitting innovation ideas and transmitting scientific and technical information, and also provides a channel for enterprise competitive product analysis and marketing planning innovation. The social network is convenient for the masses to know the information issued by local government affair handling organizations and enterprises and public institutions in time, and the information gap existing in the online social network is reduced. The information dissemination popularity prediction can help us to mine and recommend microblogs or articles with high quality and capable of arousing wide user interests. By modeling and predicting the dynamic changes of the propagation popularity, the social driving factors of the propagation behaviors among the users can be understood. The method has the advantages that the macroscopic popularity growth trend of information dissemination is researched, the future dissemination popularity is predicted in the early stage of information dissemination, and the method has important social significance and commercial value.
The information cascade propagation track in the social network forms a dynamic information cascade propagation network, the cascade propagation network records the dynamic growth process of the information, the network structure of the cascade propagation network also changes with time, and the propagation scale (popularity) in each time period also changes with time. The information or microblog publisher induces the user to generate forwarding behavior depending on the attraction of the information content and the social influence of the information or microblog publisher, the forwarding behavior of the user also influences the propagation growth, and the social influence in the information propagation has bidirectional property. The social network propagation has the characteristic of bidirectional social influence, and the bidirectional inspiration helps us to learn forward structural information and consider backward structural information when establishing a propagation model. In the prior art, the DeepHawkes model [ Cao Q, shen H, cen K, et al. The method comprises the steps of Bridging the gap between prediction and interpretation of information broadcasts, learning path feature vectors on a propagation path, predicting the static popularity of information propagation by using the time sequence features of a recurrent neural network learning Hooke process, and modeling the dynamic characteristics of propagation subgraphs at each moment.
The traditional graph neural network GNN is mostly only suitable for static graph structure learning and cannot model a dynamic graph network, while an information propagation network, an academic introduction network, a social relation network and the like are dynamic graphs, and the structure of the graphs changes along with time, so that the existing graph neural network is difficult to directly apply to the dynamic graph network.
Disclosure of Invention
The invention aims to provide a propagation popularity prediction method based on bidirectional social influence learning aiming at the defects of the prior art.
The purpose of the invention is realized by at least one of the following technical solutions.
The propagation popularity prediction method for the bidirectional social influence learning comprises the following steps:
s1, collecting information propagation network data, and constructing graph sequence data of information propagation, wherein the graph sequence data comprises a graph adjacent matrix, a node representation feature and time sequence representation feature data;
s2, constructing a model based on a graph sequence attention network, wherein the model comprises a graph deformation module and a deformation coding module based on bidirectional local graph attention at a node level and a bidirectional graph sequence attention mechanism at a graph level;
s3, constructing a macroscopic dynamic popularity prediction model based on the combination of the graph sequence attention network and a decoder, wherein the macroscopic dynamic popularity prediction model comprises the graph sequence attention network and a time sequence perception attention decoder;
s4, inputting the graph sequence data into a macroscopic dynamic popularity prediction model based on the combination of a graph sequence attention network and a decoder for iterative training;
and S5, inputting graph sequence data of information propagation, and predicting the popularity of the information propagation at different moments by using a macro dynamic popularity prediction model obtained by training and based on the combination of a graph sequence attention network and a decoder.
Further, in step S1, information propagation network data is collected, information is propagated in the social network to form a propagation network, and a post m issued by a user u is forwarded by a user v at time t to form a propagation quadruple data (v, u, t, m); at time i, the publisher and all the forwarders of a post m form a propagation network G i (ii) a Modeling a propagation network as a sequence of graphs, the propagation network G at all times i Construct a graph sequence data, denoted as { G } 0 ,G 1 ,...,G T ,G T+1 ,G T+2 A. }, T =0,1,2.., T, where T is the observation time, the node features and the subgraph structure of the corresponding subgraph snapshot at time i are denoted as
Figure BDA0003870067680000021
And
Figure BDA0003870067680000022
N i is subgraph G of time i i F is the initial characteristic dimension of the node, and the time sequence characteristic of the time i is recorded as
Figure BDA0003870067680000023
The time sequence characteristics, the node characteristics and the sub-graph structure at all the moments respectively form time sequence representation characteristics, node representation characteristics and graph adjacency matrix data of graph sequence data.
Further, in step S2, attention based on the bipartite local graph is specifically as follows:
features of a node in a propagation subgraph
Figure BDA0003870067680000031
Not only has a relation with the current neighbor, but also is influenced by the subgraphs of the forward and backward moments, and the calculation formula is as follows:
Figure BDA0003870067680000032
wherein f is combine Is a method of combining the current node characteristics and the previous and subsequent time node characteristics,
Figure BDA0003870067680000033
representing the characteristics of the propagation subgraph of node v at time t.
The social influence of the neighbor node of the node on the node in the attention of the bidirectional local graph is specifically as follows:
initial feature x of node v v Is converted into conversion characteristics through a neural network layerSign
Figure BDA0003870067680000034
The calculation formula is as follows:
Figure BDA0003870067680000035
wherein, W l Are parameters of the neural network layer and,
Figure BDA0003870067680000036
f represents x v L layers of initialized feature dimension, F l Is a layer transformation characteristic
Figure BDA0003870067680000037
The LeaklyReLU is an activation function;
transformation characteristics of node v after transformation through neural network layer
Figure BDA0003870067680000038
And translation characteristics of neighbor node u of node v
Figure BDA0003870067680000039
And carrying out aggregation, wherein u belongs to N (v), and N (v) is a neighbor node set of the node v, and the method comprises the following steps:
Figure BDA00038700676800000310
wherein, f prop For calculating attention values, e.g.
Figure BDA00038700676800000311
Figure BDA00038700676800000312
Is the attention mechanism sharing neural network parameters, e uv The attention value of the node v for the neighbor node u of the node v,
Figure BDA00038700676800000313
to calculate the attention value of node v to node u.
The weight calculation mode of the neighbor node u of the node v acting on the node v is as follows:
Figure BDA00038700676800000314
wherein alpha is uv Is the attention coefficient of the neighbor node u of the node v to the node v after normalization, the node k is one of the neighbor node set of the node v, e kv The attention value of the node v is the neighbor node k of the node v.
Compute node v aggregates updated transformation features
Figure BDA0003870067680000041
The calculation method is as follows:
Figure BDA0003870067680000042
where σ is an activation function
Figure BDA0003870067680000043
The characteristics of a node in the propagation subgraph are also affected by the subgraphs at the forward and backward moments, the subnetwork G at the previous moment t-1 t-1 Social impact of intermediate node u on node v
Figure BDA0003870067680000044
The calculation method is as follows:
Figure BDA0003870067680000045
Figure BDA0003870067680000046
wherein u ∈ V t-1 ,V t-1 F is a neighbor node set of a node v in a sub-network at the previous time t-1, a calculation method of node aggregation coefficients at the current time and the forward time is set as an Euclidean space of node characteristics,
Figure BDA0003870067680000047
n is the dimension of the feature, sub-network G of the later time t +1 t+1 Social impact of intermediate node u on node v
Figure BDA0003870067680000048
The calculation method is as follows:
Figure BDA0003870067680000049
Figure BDA00038700676800000410
wherein u ∈ V t+1 ,V t+1 Which is the set of neighbor nodes for node v in the sub-network at the later time t + 1.
Further, in step S2, the attention mechanism based on the bipartite graph sequence is as follows:
generating a time sequence-related co-production growth dependence relationship between a subgraph and a non-adjacent subgraph at a certain time of a non-adjacent subgraph, and learning the bidirectional long dependence of the subgraph on the time sequence by adopting a multi-head self-attention mechanism; the calculation formula of the space-time characteristic Q integrating the propagation structure and the time sequence characteristic is as follows:
Figure BDA0003870067680000051
wherein,
Figure BDA0003870067680000052
is the characteristic of the propagation subgraph snapshot node characteristic at the moment t after the average pooling, tp t For the time-series representation of the time t, X t For the sub-graph node feature at time t, sumPool is the sum pooling operation, W t Is a Linear neural network parameter;
the space-time characteristics Q are learned through a multi-head self-attention mechanism, so that the graph sequence attention network model can deeply fuse propagation structures and time sequence characteristics, and the formula is as follows:
Figure BDA0003870067680000053
wherein,
Figure BDA0003870067680000054
is the ith attention head, has the same K, V and Q and comes from the same space-time characteristic input,
Figure BDA0003870067680000055
is a scaling factor; operating a plurality of multi-head self-attention machines to learn the sequence characteristics of the graph respectively, finally connecting the output characteristics of the c multi-head self-attention machines together and outputting a group of characteristic vectors
Figure BDA0003870067680000056
F o Is the dimension of the group feature vector H, c is the number of subgraphs; each sub-graph corresponds to a feature vector H, calculated by the formula, where W attn Is a multi-head attention mechanism weight parameter:
Figure BDA0003870067680000057
the multi-head self-attention mechanism can learn a group of influence weights of the level of the bipartite graph of the subgraph and the subgraphs at other moments, and node features of the subgraphs at different moments are fused.
Further, in step S2, the model based on the graph sequence attention network includes a graph deformation module and a deformation coding module;
the graph deformation module comprises a local attention layer of a bipartite graph, a graph pooling layer and a sequence attention layer of the bipartite graph, and the deformation coding module comprises Add + Norm layer and Feed Forward layer; the Add + Norm layer and the Feed Forward layer are the same as the transform Encoder Block. The Add + Norm is subjected to batch normalization processing after a residual connecting network, so that the generalization capability of the model can be enhanced, and the neural network learning characterization capability is prevented from declining along with the layer number. After the graph deformation coding module completes the calculation, a group of graph sequence characteristics H are output gtb Feature H of sequence of images gtb Performing dynamic Pooling (Adaptive Pooling), which is a common operation in deep learning, and fusing features at different moments to output a feature H representing the sequence of the whole graph gtb To train a regression model.
Further, in step S3, the macro dynamic popularity prediction model based on the combination of the graph sequence attention network and the decoder comprises a graph sequence attention network module and a decoder module;
in the structure of a Decoder, a time sequence perception information is added to a graph sequence Attention network to enhance the time-dependent learning of information propagation by a model, the basic structure of a time sequence perception Attention Decoder TTA (Temporal Aware attachment) Decoder Block is close to a Decoder in a transform, a time sequence perception part is added on the basis, the time sequence characteristics of a current time sub-graph are obtained by representing and learning the information representing the position and the current time, and the learning characteristics are closely related to the time sequence through the self-Attention mechanism of the Decoder.
Further, step S4 includes the steps of:
s4.1, a Graph transformation module (Graph transform Block) receives Graph sequence data comprising a Graph adjacency matrix, a node representation characteristic and a time sequence representation characteristic as input;
s4.2, the local attention layer of the bipartite graph of the graph deformation module pools all node features in a propagation subgraph and outputs a feature vector corresponding to each subgraph
Figure BDA0003870067680000063
Learning node features including structural information of mutual influence in the forward and backward directions;
s4.3, running a plurality of multi-head self-attention mechanisms on the bidirectional graph sequence attention layer of the graph deformation module to respectively learn graph sequence characteristics, and for the characteristic vector of each sub-graph
Figure BDA0003870067680000061
The self-attention mechanism can learn a group of influence weights of bipartite graph levels of the subgraph and the subgraphs at other moments, finally, output features of the c multi-head self-attention mechanism are connected together, and a group of feature vectors are output
Figure BDA0003870067680000062
F o Is the dimension of the output feature of the multi-head self-attention mechanism, and the set feature vector H comprises the feature vector H corresponding to the cascade subgraphs of each moment i i Dynamically pooling the group feature vector H, and outputting a vector representing the features of the whole image series;
s4.4, the time sequence perception attention decoder takes the output vector of the image deformation module as V, K and Q, and the output of the time i-1 (Power embedding) at the time on the time sequence perception attention decoder
Figure BDA0003870067680000071
The Position Encoding (Position Encoding) and the time sequence representation feature are added, and the output of the time sequence perception attention decoder time i is
Figure BDA0003870067680000072
S4.5, in the model training stage, shifting (Shift) is carried out on the propagation popularity, the propagation popularity whole sequence is shifted to the left by one bit, the true value of the propagation popularity at the previous moment i-1 becomes the input of the time sequence perception attention decoder of the model at the moment i, and the TeacherForcing training is carried out.
Further, in step S5, the predicted value of the propagation popularity at the current time is output by using the predicted value of the propagation popularity at the previous time as an input of the time-series perceptual attention decoder of the model at the current time.
Compared with the prior art, the invention has the following advantages and beneficial effects:
the propagation popularity prediction method for bidirectional social influence learning provided by the invention aims at the problem that the existing propagation model can only model unidirectional social influence in propagation, bidirectional social influence information of sub-graphs at the previous and later moments is not considered, a graph sequence attention neural network with bidirectional propagation characteristic learning capability is provided, and a dynamic propagation network is modeled into a graph sequence learning problem. The model integrates a multidimensional attention mechanism and a multidimensional bidirectional graph attention mechanism, and in the attention of a bidirectional graph partial graph at a node level, node characteristics at each moment simultaneously pay attention to sub-graph information at front and rear moments; in a self-attention mechanism of a bipartite graph sequence at a graph level, long dependency of propagation sub-graph features at different moments is learned, the problem of common long dependency in information propagation time sequence features is solved, and mutual dependency features of a propagation network in a local structure and a global time sequence are learned respectively.
The propagation popularity prediction method for bidirectional social influence learning combines the graph sequence attention neural network and the self-encoder model, dynamically models the propagation popularity as a decoding process, deeply fuses the space-time propagation characteristics by applying a decoder with time sequence perception, and fuses the static propagation popularity and the dynamic propagation popularity into the same frame.
Drawings
FIG. 1 is a flow chart of a method for predicting popularity of bi-directional social impact learning propagation in an embodiment of the present invention;
FIG. 2 is a schematic partial attention diagram of a bipartite graph according to an embodiment of the invention;
FIG. 3 is a schematic attention diagram of a sequence of bipartite graphs in an embodiment of the invention;
FIG. 4 is a schematic diagram of a graph sequence attention network in an embodiment of the present invention;
FIG. 5 is a block diagram of a map morphing module in an embodiment of the present invention;
FIG. 6 is a block diagram of a timing aware decoder TAA according to an embodiment of the present invention;
fig. 7 is a diagram of a graph sequence attention network and decoder in combination with a predictive macro dynamic popularity framework in an embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to examples and drawings, but the present invention is not limited thereto.
Example 1:
the method for predicting the propagation popularity of the bidirectional social influence learning, as shown in fig. 1, comprises the following steps:
s1, collecting information propagation network data, and constructing graph sequence data of information propagation, wherein the graph sequence data comprises a graph adjacent matrix, a node representation feature and time sequence representation feature data;
collecting information propagation network data, wherein information is propagated in a social network to form a propagation network, and a post m issued by a user u is forwarded by a user v at a time t to form propagation four-tuple data (v, u, t, m); at time i, the publisher and all the forwarders of a post m form a propagation network G i (ii) a Modeling a propagation network as a sequence of graphs, the propagation network G at all times i Construct a graph sequence data, denoted as { G } 0 ,G 1 ,...,G T ,G T+1 ,G T+2 Said, T =0,1,2, T, where T is the observation time instant, the node features and the subgraph structure of the corresponding subgraph snapshot at time instant i being denoted as observation time instant
Figure BDA0003870067680000081
And
Figure BDA0003870067680000082
ni is the number of nodes of the subgraph Gi of the time i, F is the initial characteristic dimension of the nodes, and the time sequence characteristic of the time i is recorded as
Figure BDA0003870067680000083
The time sequence characteristics, the node characteristics and the sub-graph structure at all the moments respectively form time sequence representation characteristics, node representation characteristics and graph adjacency matrix data of graph sequence data.
In this embodiment, the microblog transmission is collectedBroadcasting network data, and constructing microblog propagation popularity map sequence data; in this embodiment, there are 134279 pieces of microblog data collected, including documents [ Huang Z, wang Z, zhuY, et al.prediction of Cascade Structure and Outbranches Recurrence in Microblogs]And literature [ Cao Q, shen H, cen K, et al. Bridging the gap between prediction and interpretation of information cassettes]A collection of (1). The observation time is set to be 1 hour, the microblog information propagation popularity is predicted every 0.5 hour, and the model learning rate is set to be 1e -5
S2, constructing a model based on a graph sequence attention network, wherein the model comprises a graph deformation module and a deformation coding module based on bidirectional local graph attention at a node level and a bidirectional graph sequence attention mechanism at a graph level;
as shown in fig. 2, attention is specifically given as follows based on the bidirectional partial graph:
features of a node in a propagation subgraph
Figure BDA0003870067680000091
Not only has a relation with the current neighbor, but also is influenced by the subgraphs of the forward and backward moments, and the calculation formula is as follows:
Figure BDA0003870067680000092
wherein f is combine Is a method of combining the current node characteristics and the previous and subsequent time node characteristics,
Figure BDA0003870067680000093
representing the characteristics of the propagation subgraph of node v at time t.
The social influence of the neighbor node of the node on the node in the attention of the bidirectional local graph is specifically as follows:
initial feature x of node v v Is converted into conversion characteristics through a neural network layer
Figure BDA0003870067680000094
The calculation formula is as follows:
Figure BDA0003870067680000095
wherein, W l Are parameters of the neural network layer and,
Figure BDA0003870067680000096
f represents x v L layers of (1) initialize the feature dimension, F l Is a layer transformation characteristic
Figure BDA0003870067680000097
The LeaklyReLU is an activation function; in this example, F =32,f l =64。
Transformation characteristics of node v after transformation through neural network layer
Figure BDA0003870067680000098
And translation characteristics of neighbor node u of node v
Figure BDA0003870067680000099
And carrying out aggregation, wherein u belongs to N (v), and N (v) is a neighbor node set of the node v, and the method comprises the following steps:
Figure BDA00038700676800000910
wherein f is prop In order to calculate the attention value, in the present embodiment,
Figure BDA00038700676800000911
is the attention mechanism sharing neural network parameters, e uv The attention value of the node v for the neighbor node u of the node v,
Figure BDA00038700676800000912
to calculate the attention value of node v to node u.
The weight calculation mode of the neighbor node u of the node v acting on the node v is as follows:
Figure BDA0003870067680000101
wherein alpha is uv Is the attention coefficient of the neighbor node u of the node v to the node v after normalization, the node k is one of the neighbor node set of the node v, e kv The attention value of a neighbor node k of the node v to the node v;
compute node v aggregates updated transformation features
Figure BDA0003870067680000102
The calculation method is as follows:
Figure BDA0003870067680000103
where σ is the activation function, note
Figure BDA0003870067680000104
The characteristics of a node in the propagation subgraph are also affected by the subgraphs at the forward and backward moments, the subnetwork G at the previous moment t-1 t-1 Social impact of intermediate node u on node v
Figure BDA0003870067680000105
The calculation method is as follows:
Figure BDA0003870067680000106
Figure BDA0003870067680000107
wherein u ∈ V t-1 ,V t-1 F is a neighbor node set of a node v in a sub-network at the previous time t-1, a calculation method of node aggregation coefficients at the current time and the forward time is set as an Euclidean space of node characteristics,
Figure BDA0003870067680000108
n is the dimension of the feature, sub-network G at a later time t +1 t+1 Social impact of intermediate node u on node v
Figure BDA0003870067680000109
The calculation method is as follows:
Figure BDA00038700676800001010
Figure BDA00038700676800001011
wherein u ∈ V t+1 ,V t+1 Which is the set of neighbor nodes of node v in the subnetwork t +1 at a later time.
In this embodiment, n =64.
As shown in fig. 3, the attention mechanism based on the bipartite graph sequence is as follows:
generating a time sequence correlation co-production growth dependence relationship between the subgraph and a non-adjacent subgraph at a certain moment of non-neighbor subgraph, and learning the bidirectional long dependence of the subgraph on the time sequence by adopting a multi-head self-attention mechanism; the calculation formula of the space-time characteristic Q fusing the propagation structure and the time sequence characteristic is as follows:
Figure BDA0003870067680000111
wherein,
Figure BDA0003870067680000112
is the characteristic of the propagation subgraph snapshot node characteristic at the moment t after average pooling, tp t For time-series representation of the time t, X t For the sub-graph node feature at time t, sumPool is the sum pooling operation, W t Is a Linear neural network parameter;
the space-time characteristics Q are learned through a multi-head self-attention mechanism, so that the graph sequence attention network model can deeply fuse propagation structures and time sequence characteristics, and the formula is as follows:
Figure BDA0003870067680000113
wherein,
Figure BDA0003870067680000114
is the ith attention head, has the same K, V and Q and comes from the same space-time characteristic input,
Figure BDA0003870067680000115
is a scaling factor; operating a plurality of multi-head self-attention machines to learn the sequence characteristics of the graph respectively, finally connecting the output characteristics of the c multi-head self-attention machines together and outputting a group of characteristic vectors
Figure BDA0003870067680000116
F o Is the dimension of the group feature vector H, c is the number of subgraphs; each sub-graph corresponds to a feature vector H, calculated by the formula, where W attn Is a multi-head attention mechanism weight parameter:
Figure BDA0003870067680000117
the multi-head self-attention mechanism can learn a group of influence weights of the bipartite graph level of the subgraph and subgraphs at other moments, and node characteristics of the subgraphs at different moments are fused.
As shown in fig. 4, the model based on the graph sequence attention network includes a graph deformation module and a deformation coding module; as shown in fig. 5, the graph deformation module includes a local attention layer of a bidirectional graph, a graph pooling layer, and a sequence attention layer of the bidirectional graph, and the deformation encoding module includes an Add + Norm layer and a Feed Forward layer; the Add + Norm layer and the Feed Forward layer are the same as the transform Encoder Block. Add + Norm adds a batch normalization after a residual connecting network, can strengthen the generalization ability of the model, prevent godThe characterization ability via web learning declines with the number of layers. After the graph deformation coding module completes the calculation, a group of graph sequence characteristics H are output gtb Feature H of sequence of images gtb Performing dynamic Pooling (Adaptive Pooling), which is a common operation in deep learning, and fusing features at different moments to output a feature H representing the sequence of the whole graph gtb To train a regression model.
S3, constructing a macroscopic dynamic popularity prediction model based on the combination of the graph sequence attention network and a decoder, wherein the macroscopic dynamic popularity prediction model comprises the graph sequence attention network and a time sequence perception attention decoder;
as shown in fig. 7, the macro dynamic popularity prediction model based on the combination of the graph sequence attention network and the decoder comprises a graph sequence attention network module and a decoder module;
as shown in fig. 6, in the Decoder structure, the time-series perception information is added to the graph sequence Attention network to enhance the model to learn the information propagation timeliness, the basic structure of the time-series perception Attention Decoder TTA (Temporal Aware attachment) Decoder Block is close to the Decoder in the transform, a time-series perception part is added on the basis, the presentation position information and the current time information are learned to obtain the time-series characteristics of the current time subgraph, and the learning characteristics and the time series are closely related through the self-Attention mechanism of the Decoder.
S4, inputting the graph sequence data of 1 hour observation time into a macroscopic dynamic popularity prediction model based on the combination of a graph sequence attention network and a decoder for iterative training, and comprising the following steps of:
s4.1, a Graph deformation module (Graph Transformer Block) receives Graph sequence data comprising a Graph adjacency matrix, a node representation feature and a time sequence representation feature as input;
s4.2, the local attention layer of the bipartite graph of the graph deformation module pools all node features in a propagation subgraph and outputs a feature vector corresponding to each subgraph
Figure BDA0003870067680000121
Learning node features including structural information of mutual influence in the forward and backward directions;
s4.3, running a plurality of multi-head self-attention mechanisms on the bidirectional graph sequence attention layer of the graph deformation module to respectively learn graph sequence characteristics, and for the characteristic vector of each sub-graph
Figure BDA0003870067680000131
The self-attention mechanism can learn a group of influence weights of a bipartite graph level about the subgraph and the subgraph at other moments, finally, output features of the c multi-head self-attention mechanism are connected together, and a group feature vector is output
Figure BDA0003870067680000132
F o Is the dimension of the output feature of the multi-head self-attention mechanism, and the group feature vector H comprises a feature vector H corresponding to the cascade subgraph of each moment i i Dynamically pooling the group characteristic vector H, and outputting a vector representing the characteristics of the whole image series;
s4.4, the time sequence perception attention decoder takes the output vector of the graph deformation module as V, K and Q, and the time sequence perception attention decoder outputs the output at the moment i-1 (Power embedding)
Figure BDA0003870067680000133
The Position Encoding (Position Encoding) and the time sequence representation characteristics are added, and the time sequence perception attention decoder time i is output as
Figure BDA0003870067680000134
S4.5, in the model training stage, shifting (Shift) is carried out on the propagation popularity, the whole sequence of the propagation popularity is moved to the left by one bit, the real value of the propagation popularity at the last moment i-1 is used as the input of the time sequence perception attention decoder of the model at the moment i with the probability of 0.5, and the Teacher Forcing training is carried out.
And S5, inputting the graph sequence data of microblog propagation, and predicting the popularity of the microblog propagation at different moments by using a macroscopic dynamic popularity prediction model obtained by training and based on the combination of a graph sequence attention network and a decoder.
And adopting the predicted value of the propagation popularity at the previous moment as the input of a time sequence perception attention decoder of the model at the current moment, and outputting the predicted value of the propagation popularity at the current moment.
Example 2:
the difference between the embodiment and the embodiment 1 is that a method for predicting popularity of an academic citation network based on bidirectional social influence learning is provided;
step S1, collecting academic quotation network data and constructing academic quotation network spreading popularity graph sequence data; in this embodiment, 68433 pieces of academic citation network data are collected. The observation time is set to 5 years, the propagation popularity of the academic quotation network is predicted every 180 days, and the model learning rate is set to 1e -5
In the step S2, the process is carried out,
Figure BDA0003870067680000135
in step S4, the graph sequence data of 5-year observation time is input into a macroscopic dynamic popularity prediction model based on the combination of a graph sequence attention network and a decoder for iterative training
In the step S5, graph sequence data of the academic quotation is input, and popularity of the academic quotation at different moments is predicted by using a macro dynamic popularity prediction model obtained by training and based on the combination of a graph sequence attention network and a decoder, so that the academic quotation propagation popularity prediction values at different moments are obtained.
Example 3:
the embodiment is different from the embodiment 1 in that a twitter propagation popularity prediction method based on bidirectional social influence learning is provided;
step S1, collecting twitter data and constructing twitter propagation popularity observation graph sequence data; in this embodiment, a total of 20000 pieces of twitter data are collected. The observation time is set to 1 hour, the Tutt propagation popularity is predicted every 0.5 hour, and the model learning rate is set to 1e -5
In the step S2, the process is carried out,
Figure BDA0003870067680000141
in step S4, the graph sequence data of 1 hour observation time is input into a macroscopic dynamic popularity prediction model based on the combination of a graph sequence attention network and a decoder for iterative training.
In step S5, graph sequence data of the twitter propagation is input, and the popularity of the twitter propagation at different moments is predicted by using a macro dynamic popularity prediction model obtained by training and based on the combination of a graph sequence attention network and a decoder, so that the twitter propagation popularity prediction values at different moments are obtained.
The above description is only for the preferred embodiments of the present invention, but the protection scope of the present invention is not limited thereto, and any person skilled in the art can substitute or change the technical solution of the present invention and the inventive concept within the scope of the present invention, which is disclosed by the present invention, and the equivalent or change thereof belongs to the protection scope of the present invention.

Claims (10)

1. A propagation popularity prediction method for bidirectional social influence learning is characterized by comprising the following steps:
s1, collecting information propagation network data, and constructing graph sequence data of information propagation, wherein the graph sequence data comprises a graph adjacency matrix, a node representation feature and time sequence representation feature data;
s2, constructing a model based on a graph sequence attention network, wherein the model comprises a bidirectional local graph attention mechanism based on a node level and a graph deformation coding and decoding module based on a bidirectional graph sequence attention mechanism based on a graph level;
s3, constructing a macroscopic dynamic popularity prediction model based on the combination of the graph sequence attention network and a decoder, wherein the macroscopic dynamic popularity prediction model comprises the graph sequence attention network and a time sequence perception attention decoder;
s4, inputting the graph sequence data into a macroscopic dynamic popularity prediction model based on the combination of a graph sequence attention network and a decoder for iterative training;
and S5, inputting graph sequence data of information propagation, and predicting the popularity of the information propagation at different moments by using a model obtained by training.
2. The method for predicting the propagation popularity of the bidirectional social influence learning according to claim 1, wherein in step S1, information propagation network data is collected, information is propagated in a social network to form a propagation network, and a post m issued by a user u is forwarded by a user v at a time t to form a propagation quadruple of data (v, u, t, m);
at time i, the publisher and all forwarders of a post m form a propagation subgraph G i (ii) a Modeling a propagation network as a sequence of graphs, the propagation network G at all times i Construct a graph sequence data, denoted as { G } 0 ,G 1 ,...,G T ,G T+1 ,G T+2 A construction, where T is the observation time, and the node features and the subgraph structure of the corresponding subgraph snapshot at time i are denoted X, respectively i And A i
Figure FDA0003870067670000011
N i Is subgraph G of time i i F is the initial characteristic dimension of the node, and the time sequence characteristic of the time i is marked as tp i
Figure FDA0003870067670000012
The time sequence characteristics, the node characteristics and the sub-graph structure at all the time respectively form time sequence representation characteristics, node representation characteristics and graph adjacency matrix data of graph sequence data.
3. The method for predicting the popularity of propagation of bidirectional social influence learning according to claim 1, wherein in step S2, the attention based on the bidirectional local graph is specifically as follows:
features of a node in a propagation subgraph
Figure FDA0003870067670000013
Not only has a relation with the current neighbor, but also is influenced by the subgraphs of the forward and backward moments, and the calculation formula is as follows:
Figure FDA0003870067670000021
wherein f is combine Is a method of combining the current node characteristics and the previous and subsequent time node characteristics,
Figure FDA0003870067670000022
Figure FDA0003870067670000023
representing the characteristics of the propagation subgraph of node v at time t.
4. The propagation popularity prediction method for bidirectional social influence learning according to claim 3, wherein social influences of neighbor nodes of a node in the attention of the bidirectional local graph on the node are specifically as follows:
initial feature x of node v v Is converted into conversion characteristics through a neural network layer
Figure FDA0003870067670000024
The calculation formula is as follows:
Figure FDA0003870067670000025
wherein, W l Are parameters of the neural network layer and,
Figure FDA0003870067670000026
f represents x v L layers of (1) initialize the feature dimension, F l Is a characteristic of layer conversion
Figure FDA0003870067670000027
Is the activation function;
transformation characteristics of node v after transformation through neural network layer
Figure FDA0003870067670000028
And translation characteristics of neighbor node u of node v
Figure FDA0003870067670000029
And carrying out aggregation, wherein u belongs to N (v), and N (v) is a neighbor node set of the node v, and the method comprises the following steps:
Figure FDA00038700676700000210
wherein f is prop For calculating attention values, e.g.
Figure FDA00038700676700000211
Figure FDA00038700676700000212
Is the attention mechanism sharing neural network parameters, e uv The attention value of the node v for the neighbor node u of the node v,
Figure FDA00038700676700000213
calculating the attention value of the node v to the node u;
the weight calculation mode of the neighbor node u of the node v acting on the node v is as follows:
Figure FDA00038700676700000214
wherein alpha is uv Is the attention coefficient of the neighbor node u of the node v to the node v after normalization, the node k is one of the neighbor node set of the node v, e kv The attention value of a neighbor node k of the node v to the node v;
compute node v aggregates updated transformation features
Figure FDA00038700676700000215
The calculation method is as follows:
Figure FDA0003870067670000031
where σ is an activation function
Figure FDA0003870067670000032
5. The method of claim 4, wherein the characteristics of a node in the propagation subgraph are also influenced by the subgraphs at forward and backward time, and the sub-network G at the previous time t-1 is the sub-network G t-1 Social impact of intermediate node u on node v
Figure FDA0003870067670000033
The calculation method is as follows:
Figure FDA0003870067670000034
Figure FDA0003870067670000035
wherein u ∈ V t-1 ,V t-1 F is a neighbor node set of a node v in a sub-network at the previous time t-1, a calculation method of node aggregation coefficients at the current time and the forward time is set as an Euclidean space of node characteristics,
Figure FDA0003870067670000036
n is the dimension of the feature, sub-network G of the later time t +1 t+1 Middle jointSocial impact of point u on point v
Figure FDA0003870067670000037
The calculation method is as follows:
Figure FDA0003870067670000038
Figure FDA0003870067670000039
wherein u ∈ V t+1 ,V t+1 A neighbor node set of a node v in the sub-network at the later moment t + 1; considering the forward and backward characteristics of a node, the updated node characteristics are as follows:
Figure FDA00038700676700000310
6. the method for predicting the propagation popularity of the bidirectional social influence learning as claimed in claim 1, wherein in the step S2, the attention mechanism based on the digraph sequence is specifically as follows:
the subgraph and a non-adjacent subgraph at a certain time of a non-neighbor generate a long dependency relationship, a multi-head self-attention mechanism is adopted by the model, and bidirectional long dependency of the subgraph on a time sequence is learned; spatio-temporal features Q incorporating propagation structure and timing features t The calculation formula of (c) is:
Figure FDA0003870067670000041
wherein,
Figure FDA0003870067670000042
is the feature of the sub-graph node feature at the time t after average pooling, tp t Is a time-sequential representation of time tCharacteristic of, X t For the sub-graph node feature at time t, sumPool is the sum pooling operation, W t Is a Linear neural network parameter;
the space-time characteristics are learned through a multi-head self-attention mechanism, so that the graph sequence attention network model can deeply fuse the propagation structure and the time sequence characteristics, and the multi-head self-attention mechanism formula is as follows:
Figure FDA0003870067670000043
wherein,
Figure FDA0003870067670000044
is the ith attention head, has the same K, V and Q and comes from the same space-time characteristic input,
Figure FDA0003870067670000045
is a scaling factor; operating a plurality of multi-head self-attention machines to learn the sequence characteristics of the graph respectively, finally connecting the output characteristics of the c multi-head self-attention machines together and outputting a group of characteristic vectors
Figure FDA0003870067670000046
F o Is the dimension of the group feature vector H, c is the number of subgraphs; each sub-graph corresponds to a feature vector H, calculated by the formula, where W attn Is a multi-head attention mechanism weight parameter:
Figure FDA0003870067670000047
the multi-head self-attention mechanism can learn a group of influence weights of the bipartite graph level of the subgraph and subgraphs at other moments, and the characteristics of the subgraphs at different moments are fused.
7. The method for predicting the popularity of propagation of bidirectional social influence learning according to claim 1, wherein in step S2, the model based on the graph sequence attention network includes a graph deformation encoding module and a decoding module;
the graph deformation module comprises a bidirectional graph local attention layer, a graph pooling layer and a bidirectional graph sequence attention layer, and the deformation coding module comprises an Add + Norm layer and a Feed Forward layer; the Add + Norm layer and the Feed Forward layer are the same as a transform Encoder Block; the Add + Norm layer is subjected to batch normalization processing after a residual connecting network, so that the generalization capability of the model can be enhanced, and the neural network learning characterization capability is prevented from declining along with the number of layers;
after the computation of the graph deformation coding module is completed, a group of graph sequence characteristics H are output gtb Features H of map sequence gtb Performing dynamic Pooling (Adaptive Pooling), which is a common operation in deep learning, and fusing features at different moments to output a feature H representing the sequence of the whole graph gtb Based on which a regression model is trained.
8. The method for predicting the popularity of propagation of bi-directional social influence learning according to claim 1, wherein in step S3, the macro dynamic popularity prediction model based on the combination of the graph sequence attention network and the decoder comprises a graph sequence attention network module and a decoder module;
in the structure of a Decoder, a time sequence perception information is added to a graph sequence Attention network to enhance the time-dependent learning of information propagation by a model, the basic structure of a time sequence perception Attention Decoder TTA (Temporal Aware attachment) Decoder Block is close to a Decoder in a Transformer, a time sequence perception part is added on the basis, the time sequence characteristics of a current time subgraph are obtained by representing and learning the position information and the current time information, and the learning characteristics are closely associated with the time sequence by learning through a self-Attention mechanism of the Decoder.
9. The method for predicting the popularity of propagation of bidirectional social influence learning according to claim 1, wherein the step S4 comprises the steps of:
s4.1, a Graph transformation module (Graph transform Block) receives Graph sequence data comprising a Graph adjacency matrix, a node representation characteristic and a time sequence representation characteristic as input;
s4.2, the local attention layer of the bipartite graph of the graph deformation module pools all node features in a propagation subgraph and outputs a feature vector corresponding to each subgraph
Figure FDA0003870067670000051
Learning node characteristics including structural information of mutual influence in the forward and backward directions;
s4.3, running a plurality of multi-head self-attention mechanisms on the bidirectional graph sequence attention layer of the graph deformation module to learn graph sequence characteristics respectively, and for the characteristic vector of each sub-graph
Figure FDA0003870067670000061
The self-attention mechanism can learn a group of influence weights of bipartite graph levels of the subgraph and the subgraphs at other moments, finally, output features of the c multi-head self-attention mechanism are connected together, and a group of feature vectors are output
Figure FDA0003870067670000062
Fo is the dimension of the output feature of the multi-head self-attention mechanism, and the set feature vector H comprises the feature vector H corresponding to the cascade subgraph at each time i i Dynamically pooling the group feature vector H, and outputting a vector representing the features of the whole image series;
s4.4, the time sequence perception attention decoder takes the output vector of the image deformation module as V, K and Q, and the output of the time i-1 (Power embedding) at the time on the time sequence perception attention decoder
Figure FDA0003870067670000063
The Position Encoding (Position Encoding) and the time sequence representation feature are added, and the output of the time sequence perception attention decoder time i is
Figure FDA0003870067670000064
S4.5, in the model training stage, shifting (Shift) is carried out on the propagation popularity, the propagation popularity whole sequence is shifted to the left by one bit, the true value of the propagation popularity at the previous moment i-1 becomes the input of the time sequence perception attention decoder of the model at the moment i, and the TeacherForcing training is carried out.
10. The method as claimed in claim 1, wherein the predicted value of the propagation popularity at the current time is output in step S5 by using the predicted value of the propagation popularity at the previous time as an input to a time-series perceptual attention decoder of the model at the current time.
CN202211196722.XA 2022-09-28 2022-09-28 Propagation popularity prediction method for bidirectional social influence learning Pending CN115618124A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211196722.XA CN115618124A (en) 2022-09-28 2022-09-28 Propagation popularity prediction method for bidirectional social influence learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211196722.XA CN115618124A (en) 2022-09-28 2022-09-28 Propagation popularity prediction method for bidirectional social influence learning

Publications (1)

Publication Number Publication Date
CN115618124A true CN115618124A (en) 2023-01-17

Family

ID=84859991

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211196722.XA Pending CN115618124A (en) 2022-09-28 2022-09-28 Propagation popularity prediction method for bidirectional social influence learning

Country Status (1)

Country Link
CN (1) CN115618124A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117610717A (en) * 2023-11-13 2024-02-27 重庆大学 Information popularity prediction method based on double-variation cascade self-encoder

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117610717A (en) * 2023-11-13 2024-02-27 重庆大学 Information popularity prediction method based on double-variation cascade self-encoder

Similar Documents

Publication Publication Date Title
Chen et al. TrajVAE: A Variational AutoEncoder model for trajectory generation
CN111310063B (en) Neural network-based article recommendation method for memory perception gated factorization machine
CN112488055B (en) Video question-answering method based on progressive graph attention network
CN109117943B (en) Method for enhancing network representation learning by utilizing multi-attribute information
Sheng et al. Enhanced graph neural network for session-based recommendation
CN114971784B (en) Session recommendation method and system based on graph neural network by fusing self-attention mechanism
CN113590976A (en) Recommendation method of space self-adaptive graph convolution network
CN115470406A (en) Graph neural network session recommendation method based on dual-channel information fusion
CN113704438A (en) Conversation recommendation method of abnormal picture based on layered attention mechanism
CN115618124A (en) Propagation popularity prediction method for bidirectional social influence learning
CN114637911A (en) Next interest point recommendation method of attention fusion perception network
CN115017405A (en) Graph neural network travel package recommendation method based on multitask self-coding
CN114925270A (en) Session recommendation method and model
CN117633371B (en) Recommendation method, device and readable storage medium based on multi-attention mechanism
CN114676332A (en) Network API recommendation method facing developers
CN117994011A (en) E-commerce dynamic perception data recommendation method based on memory updating and neighbor transfer
CN114820139A (en) Multi-user recommendation system based on knowledge graph path reasoning
CN114528490A (en) Self-supervision sequence recommendation method based on long-term and short-term interests of user
CN112559904B (en) Conversational social recommendation method based on door mechanism and multi-modal graph network
Zheng et al. Tourism demand forecasting using complex network theory
CN117390267A (en) Knowledge graph-based personalized multitask enhanced recommendation model
CN117237140A (en) Social network influence maximization method fusing graph convolution neural network and transducer
CN118396685B (en) Advertisement marketing recommendation method based on deep reinforcement learning
CN115660147B (en) Information propagation prediction method and system based on influence modeling between propagation paths and in propagation paths
Shou et al. Research on Knowledge Concept Recommendation Algorithm with Spatial-Temporal Information

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination