CN113688600A - Information propagation prediction method based on topic perception attention network - Google Patents

Information propagation prediction method based on topic perception attention network Download PDF

Info

Publication number
CN113688600A
CN113688600A CN202111049168.8A CN202111049168A CN113688600A CN 113688600 A CN113688600 A CN 113688600A CN 202111049168 A CN202111049168 A CN 202111049168A CN 113688600 A CN113688600 A CN 113688600A
Authority
CN
China
Prior art keywords
user
topic
propagation
information
context
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111049168.8A
Other languages
Chinese (zh)
Other versions
CN113688600B (en
Inventor
杨成
石川
王浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Posts and Telecommunications
Original Assignee
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Posts and Telecommunications filed Critical Beijing University of Posts and Telecommunications
Priority to CN202111049168.8A priority Critical patent/CN113688600B/en
Publication of CN113688600A publication Critical patent/CN113688600A/en
Application granted granted Critical
Publication of CN113688600B publication Critical patent/CN113688600B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/126Character encoding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/194Calculation of difference between files
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses an information propagation prediction method based on a theme perception attention network, which integrates a theme context and a propagation history context into a user representation for prediction. The topic context supports model propagation patterns for a particular topic, while the propagation history context can be further decomposed into user-dependent and location-dependent modeling. We can then use the encoded user context to construct a user representation under multiple topics. We then further integrate the user representation through a time-decay aggregation module to obtain a cascaded representation. Wherein all of these modules are driven by information dissemination features. Therefore, the information propagation prediction method based on the topic perception attention network can better fit the diffusion data of the real world and predict more accurately. In addition, predefined good theme distribution is required to be used in the traditional theme perception model, and the theme can be automatically learned by adopting the method.

Description

Information propagation prediction method based on topic perception attention network
Technical Field
The invention relates to the technical field of networks, in particular to an information propagation prediction method based on a theme perception attention network.
Background
Social networking platforms such as twitter and new wave microblog attract millions of users, and a large amount of information is spread among the users every day. The process of information dissemination, also known as cascading, models the dissemination pattern and the user behavior is widely used in many fields, such as popularity prediction, epidemiology and personalized recommendations. Next user prediction has been extensively studied in recent years as a popular micro-cascading prediction task. This problem is defined as the sequence of user infections, ordered over time for a given information item, predicting the next infected user (by convention, researchers will use "infection", "activation" or "influence" to describe that there is an interaction of a user with an information item).
Conventional micro-cascade prediction methods include independent cascade model (IC) based methods and embedding based methods. The independent cascade model allocates an independent diffusion probability between each user pair, and many cascade diffusion models are established on the basic assumption of the model, and are expanded by additionally considering more information, such as continuous time stamps and user attributes. There are also some studies that explored the impact of topic information on cascade modeling. TIC has for the first time studied the information dissemination prediction task from the perspective of topic perception, the main idea being by setting a specific topic probability for each user pair.
As research advances, researchers have proposed an embedding-based approach to cascade prediction, which embeds users into a continuous latent space by improving the expressive power of the model using representation learning techniques, and calculates the propagation probability between each pair of users by using a user embedding function, rather than directly estimating a real-valued parameter. However, neither the IC-based approach nor the embedding-based approach takes into account the modeling of the concatenation history sequence information. Recent work has shown that these models are not as efficient as deep learning models based on considering cascade sequences.
With the success of deep learning, Recurrent Neural Networks (RNNs) have shown a strong ability to model information dissemination. TopolSTM extends the standard LSTM model, building hidden states from Directed Acyclic Graphs (DAGs) extracted from social graphs. CYAN-RNN and DeepDiffuse combine a recurrent neural network with an attention mechanism to account for the propagation structure. RecCTIC proposes a Bayesian topological RNN model for capture tree dependence. The Diffusion-LSTM uses image information to assist in prediction, and builds a Tree-LSTM model to infer propagation paths. The FOREST expands the GRU model, and designs an additional structural context extraction strategy to utilize the underlying social graph information.
Recently, some attention networks have been proposed to better capture the propagation dependence in the cascade sequence. The HiDAN builds a hierarchical attention network, captures a non-sequence structure in the cascade by adopting an attention mechanism, excavates a real dependency relationship from the cascade, designs a time attenuation module by combining timestamp information of a user, and jointly models user dependency and time attenuation, thereby greatly improving the expression capability and interpretability of the model.
With the development of deep learning technology, some documents model information propagation cascade as infection sequences, and a circulating neural network is adopted to obtain a good effect. Although concatenation is usually represented as a sequence of users, ordered by infection timestamp, the real propagation process is usually not strictly ordered, relying on unobserved user connection graphs. Therefore, other studies have employed an attention mechanism to capture non-sequential long-term propagation dependencies.
However, existing neural network-based approaches assume that the propagation behavior and pattern of all information items are homogenous. This assumption may not hold in the real world. In a real information dissemination scenario, a user may have different behavior patterns for information items of different topics. Intuitively, the interests of the user are often diverse, and the propagation behavior of the user may also be diverse according to the topic of the information item. For example, a user may focus on different people and then forward different information under different topics, respectively, and thus have topic-specific dependencies. The existing method based on the neural network rarely utilizes information texts, does not consider the propagation mode and user behavior perceived by a modeling theme, cannot model the propagation mode and dependency relationship under a specific theme, and limits the expression capability of the model. Whereas traditional non-neural approaches have demonstrated the impact of the subject on the user.
Next user prediction has been extensively studied in recent years as a popular micro-cascading prediction task. Traditional modeling typically ignores the textual content of the propagating information item, resulting in learning mixed dependencies from different topics. In contrast, topic-aware modeling aims to explicitly decouple propagation dependencies under a particular topic, thereby enabling more accurate predictions. In fact, the traditional non-neural network method based on independent cascade model has proved the advantage of topic-aware modeling, which can model the behavior of information items from different topics separately. But these early methods were built on strong independent assumptions, and this strategy limited the generalization performance of the model and has been shown to be suboptimal by recent deep learning-based methods. To our knowledge, no previous research has proposed a topic perception model based on neural networks to mine propagation dependencies under different topics.
Disclosure of Invention
In view of the above, the present invention aims to provide an information propagation prediction method based on a topic-aware attention network, which starts from a formalized propagation prediction problem and introduces our embedding strategy to encode user/location/text information into a vector. We will then come up with a topic-aware attention tier aimed at capturing the historical propagation dependence and time-decay effects of different topics. Finally, our model will take a multi-topic cascade representation through a given topic-aware attention layer and then predict the next infected user.
In order to achieve the above purpose, the invention provides the following technical scheme:
the invention provides an information propagation prediction method based on a topic perception attention network, and S1, a topic context and a propagation history context are integrated into a user representation for prediction;
s2, the topic context supports the model building of the propagation mode aiming at the specific topic, and the propagation history context is further decomposed into user dependence model building and position dependence model building;
s3, constructing a user representation under multiple topics by using the user context obtained by encoding;
s4, further integrating the user representation through a time attenuation aggregation module, thereby obtaining a multi-topic cascade representation and then predicting the next infected user;
wherein each module is driven by information dissemination features.
Further, the specific method of step S1 is:
given a set of users U, a concatenation set V and a propagation information set M, the propagation sequence of the ith information item in M is defined as a concatenation
Figure BDA0003252232820000031
Wherein the tuple
Figure BDA0003252232820000032
Representing a user
Figure BDA0003252232820000033
In that
Figure BDA0003252232820000034
The moments are forwarded and the sequence is ordered by the time of infection, the propagation prediction task being defined as a given cascade ciAnd previous infected user sequence
Figure BDA0003252232820000035
Predict next infected user as
Figure BDA0003252232820000036
Wherein n is 1,2, …, | ci|-1。
Further, in step S2, the propagation pattern modeling is to encode semantic information of the propagation information text by using the pre-trained language model BERT.
Further, the propagation pattern modeling in step S2 is to embed the text encoded by BERT into the text by a full link layer
Figure BDA0003252232820000041
Conversion to propagating text embedding
Figure BDA0003252232820000042
yi=Wxxi+bx (1)
Wherein WxAnd bxRespectively a weight matrix and an offset vector.
Further, the user dependent modeling uses the embedded matrix in step S2
Figure BDA0003252232820000043
And encoding the users, wherein | U | represents the number of users, and K and d represent the number of topics and the embedding dimension respectively.
Further, for the cascade sequence
Figure BDA0003252232820000044
Each user in
Figure BDA0003252232820000045
User is embedded into
Figure BDA0003252232820000046
Wherein
Figure BDA0003252232820000047
Is the user embedding of the user under the kth topic.
Further, the position-dependent modeling in step S2 is to set a learnable position-embedded pos for each positionjWherein posjShared among all cascades.
Further, the encoding method in step S3 is:
topic context:
computing user embedding for each topic k
Figure BDA0003252232820000048
And passBroadcast text embedding yiCosine similarity between them, and normalized by softmax function:
Figure BDA0003252232820000049
wherein K is 1,2, …, K, and
Figure BDA00032522328200000410
representing a user
Figure BDA00032522328200000411
A weight under the kth topic; user-embedded representation of the aggregated topic context as
Figure BDA00032522328200000412
Propagation history context:
in a cascade sequence
Figure BDA00032522328200000413
With previous users
Figure BDA00032522328200000414
Is calculated by the following formula:
Figure BDA00032522328200000415
wherein
Figure BDA00032522328200000416
Respectively for subject-specific linear mapping of target users and previous users;
user' s
Figure BDA00032522328200000417
And the user
Figure BDA00032522328200000418
BetweenFull attention score
Figure BDA00032522328200000419
And weight
Figure BDA00032522328200000420
Is used to describe the propagation history context and is calculated by the following formula:
Figure BDA0003252232820000051
Figure BDA0003252232820000052
wherein
Figure BDA0003252232820000053
Is a position dependent score from position m to position j;
full context-aware multi-topic user representation:
users in the k-th theme
Figure BDA0003252232820000054
Expressed as a weighted sum of previously infected users:
Figure BDA0003252232820000055
weighting of topic contexts
Figure BDA0003252232820000056
And location dependent score
Figure BDA0003252232820000057
Shared between different layers.
Further, the modeling method of the time decay aggregation module in step S4 is as follows:
converting the continuous-time decay into discrete time intervals:
Figure BDA0003252232820000058
wherein t islBy dividing the time range [0, Tmax]Is divided into L sub-intervals { [0, t ]1),..,[tL-1,Tmax) Where T ismaxIs the maximum timestamp in the data set, with a corresponding learnable weight for each time interval for each topic
Figure BDA0003252232820000059
Further, the obtaining method of the multi-topic cascade representation in step S4 is as follows:
the complete aggregation weight is calculated according to equation (8):
Figure BDA00032522328200000510
then the
Figure BDA00032522328200000511
Normalizing j to 1,2, …, n by a softmax function;
for each topic k, calculating
Figure BDA00032522328200000512
Being weighted
Figure BDA00032522328200000513
And a feedforward neural network with a ReLU activation function is adopted to endow the model with nonlinearity, and the output of the theme perception attention layer is represented in cascade as
Figure BDA00032522328200000514
Further, the method for predicting the next infected user in step S4 is:
given a cascading sequence
Figure BDA00032522328200000515
By measuring user embedding
Figure BDA00032522328200000516
And cascade embedding
Figure BDA00032522328200000517
To parameterize the next infected user
Figure BDA00032522328200000518
Probability, cascade and user of
Figure BDA00032522328200000519
The interaction probability of (a) is expressed as:
Figure BDA0003252232820000061
wherein Θ represents all parameters that need to be learned;
the training objective for predicting an infected user is defined by equation (10):
Figure BDA0003252232820000062
setting K topic prototype embedding
Figure BDA0003252232820000063
And encourages user embedding under k themes
Figure BDA0003252232820000064
Corresponding subject prototype mkSimilarly, the goal is to maximize:
Figure BDA0003252232820000065
this term is taken as an additional training target and summed over all users:
Figure BDA0003252232820000066
the complete training objective function is
Figure BDA0003252232820000067
Where η is the equilibrium coefficient.
Compared with the prior art, the invention has the beneficial effects that:
the information propagation prediction method based on the topic perception attention network, provided by the invention, combines the advantages of topic-specific propagation modeling and deep learning technology, designs a novel and effective topic perception attention mechanism, and integrates a topic context and a propagation history context into a user representation for prediction. The topic context supports model propagation patterns for a particular topic, while the propagation history context can be further decomposed into user-dependent and location-dependent modeling. We can then use the encoded user context to construct a user representation under multiple topics. We then further integrate the user representation through a time-decay aggregation module to obtain a cascaded representation. Wherein all of these modules are driven by information dissemination features. Therefore, the information propagation prediction method based on the topic perception attention network can better fit the diffusion data of the real world and predict more accurately. In addition, predefined good theme distribution is required to be used in the traditional theme perception model, and the theme can be automatically learned by adopting the method.
Drawings
In order to more clearly illustrate the embodiments of the present application or technical solutions in the prior art, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the present invention, and other drawings can be obtained by those skilled in the art according to the drawings.
Fig. 1 is an architecture diagram of a subject-aware attention network according to an embodiment of the present invention.
Detailed Description
The invention provides an information propagation prediction method based on a topic perception attention network, and S1, a topic context and a propagation history context are integrated into a user representation for prediction;
s2, the topic context supports the model building of the propagation mode aiming at the specific topic, and the propagation history context is further decomposed into user dependence model building and position dependence model building;
s3, constructing a user representation under multiple topics by using the user context obtained by encoding;
s4, further integrating the user representation through a time attenuation aggregation module, thereby obtaining a multi-topic cascade representation and then predicting the next infected user;
wherein each module is driven by information dissemination features.
For a better understanding of the present solution, the method of the present invention is described in detail below with reference to the accompanying drawings.
In this section we will start with a formalized propagation prediction problem and introduce our embedding strategy to encode user/location/text information into a vector. We will then come up with a topic-aware attention tier aimed at capturing the historical propagation dependence and time-decay effects of different topics. Finally, our model will take a multi-topic cascade representation through a given topic-aware attention layer and then predict the next infected user. The complete structure of our proposed TAN is shown in figure 1.
1. Problem definition
Given a set of users U, a concatenation set V and a propagation information set M, the propagation sequence of the ith information item in M can be defined as a concatenation
Figure BDA0003252232820000071
Wherein the tuple
Figure BDA0003252232820000072
Representing a user
Figure BDA0003252232820000073
In that
Figure BDA0003252232820000074
The time of day is forwarded and the sequence is ordered by time of infection. Following the previous task setup, the propagation prediction task is defined as a given cascade ciAnd previous infected user sequence
Figure BDA0003252232820000075
Predicting next infected user
Figure BDA0003252232820000076
Wherein n is 1,2, …, | ci|-1。
2. Embedding layer
1) User embedding:
to capture user interests and dependencies on different topics, we use an embedded matrix
Figure BDA0003252232820000077
Figure BDA0003252232820000081
And encoding the users, wherein | U | represents the number of users, and K and d represent the number of topics and the embedding dimension respectively. For cascade sequences
Figure BDA0003252232820000082
Each user in
Figure BDA0003252232820000083
His user is embedded as
Figure BDA0003252232820000084
Wherein
Figure BDA0003252232820000085
Is the user embedding of the user under the kth topic.
2) Position embedding:
to exploit the cascade infection order information, we are for eachPosition setting a learnable position embedding posjWherein posjShared among all cascades.
3) Text embedding:
the semantic information of the transmission information text is coded by using a pre-training language model BERT. To measure topic similarity between user embedding and text embedding for a particular topic, we embed text encoded by BERT through a fully connected layer
Figure BDA0003252232820000086
Is converted into
Figure BDA0003252232820000087
yi=Wxxi+bx (1)
Wherein WxAnd bxRespectively a weight matrix and an offset vector.
3. Topic perception attention layer
In this section, we will further encode various context information into user representations, which are then aggregated in conjunction with time-decaying weights to generate a cascaded representation for each topic.
3.1 user representation enhancement
We incorporate the topic context and the propagation history context into the multi-topic user representation, respectively. The propagation history context can be further decomposed into user dependencies and location dependencies. Inspired by the multi-head attention mechanism, we treat a topic as a specific head (head) and perform the attention mechanism separately in each topic to extract user and location dependencies.
1) Topic context
Based on propagating text yiWe propose to strengthen the user's embedding if there is a higher similarity to the text embedding under the kth topic
Figure BDA0003252232820000088
Is embedded. Specifically, we compute for each topic k
Figure BDA0003252232820000089
And yiCosine similarity between them, and normalized by a softmax function:
Figure BDA00032522328200000810
wherein K is 1,2, …, K, and
Figure BDA00032522328200000811
representing a user
Figure BDA00032522328200000812
Weight under the k topic. User embedding of the aggregate topic context can then be represented as
Figure BDA00032522328200000813
We can find out that when the k-th topic corresponds to the user embedding
Figure BDA0003252232820000091
And yiThe larger the cosine similarity of (a), the larger the assigned weight, and the more strengthened the user embedding under the theme.
2) Propagating historical context
Intuitively, a user is infected typically due to the spreading of text, while there are only a few users in the spreading sequence that were previously infected. Thus, the goal of propagating historical context is to extract and characterize and user
Figure BDA0003252232820000092
Infecting the associated user. In particular, we employ an attention mechanism to model user dependencies and give more attention weight to those users who may affect an infection. Formally, in a cascade sequence
Figure BDA0003252232820000093
With previous users
Figure BDA0003252232820000094
The dependent attention weight of (a) can be calculated by the following formula:
Figure BDA0003252232820000095
wherein
Figure BDA0003252232820000096
For subject-specific linear mapping of target users and previous users, respectively.
Intuitively, we should also focus on the source user and the recently infected users. Note that such dependencies are independent of the particular user, so we propose to model location dependencies under each topic. Instead of directly adding the predefined location embedding and the user embedding in the past, we compute the location-dependent score using a similar method to user-dependent modeling. In this way, our approach can better capture user-independent location dependencies for better prediction performance.
User' s
Figure BDA0003252232820000097
And the user
Figure BDA0003252232820000098
Complete attention score between
Figure BDA0003252232820000099
And weight
Figure BDA00032522328200000910
Can be used to describe the propagation history context and is calculated by the following formula:
Figure BDA00032522328200000911
Figure BDA00032522328200000912
wherein
Figure BDA00032522328200000913
Is a position dependent score from position m to position j.
3) Complete context-aware multi-topic user representation
To fully exploit the topic and propagation history context, we will refer to the users in the kth topic
Figure BDA00032522328200000914
Represented as a weighted sum of previously infected users.
Figure BDA00032522328200000915
Note that we can also add multiple tiers of the above operations to get a more accurate representation. In this case, the weight of the subject context
Figure BDA0003252232820000101
And location dependent score
Figure BDA0003252232820000102
Shared between different layers.
3.2 obtaining a cascading representation based on time-decaying aggregation
After extracting the multi-user representation under multiple topics, we need to aggregate them to obtain a cascading representation under multiple topics. We assume that the user's influence will decay with time and jointly consider the time decay and the propagation dependent weights in equation 4.
1) Time-decay impact modeling
Specifically, inspired by Deephawkes, we employed a non-parametric time-decay modeling strategy for each topic. Formally, a cascade of sequences giving historical infection
Figure BDA0003252232820000103
The continuous-time decay is first converted into discrete time intervals:
Figure BDA0003252232820000104
wherein t islBy dividing the time range [0, Tmax]Is divided into L sub-intervals { [0, t ]1),..,[tL-1,Tmax) Where T ismaxIs the maximum timestamp in the data set. For each topic, each time interval has a corresponding learnable weight
Figure BDA0003252232820000105
2) Computing cascading representations under multiple topics
The complete aggregate weight is the addition of an additional term to equation 4:
Figure BDA0003252232820000106
then the
Figure BDA0003252232820000107
J will be normalized by the softmax function to 1,2, …, n. Finally, for each topic k, we calculate
Figure BDA0003252232820000108
Being weighted
Figure BDA0003252232820000109
And a feed-forward neural network with a ReLU activation function is used to impart model non-linearity. The output of the topic perception attention layer is a cascade representation and can be represented as
Figure BDA00032522328200001010
3.3 training goals and model details
Given a cascading sequence
Figure BDA00032522328200001011
By measuring user embedding
Figure BDA00032522328200001012
And cascade embedding
Figure BDA00032522328200001013
To parameterize the next infected user
Figure BDA00032522328200001014
The probability of (c). Cascading with Users as follows
Figure BDA00032522328200001015
The interaction probability of (c) can be expressed as:
Figure BDA00032522328200001016
where Θ represents all parameters that need to be learned.
We then predict the training goals of the infected user as defined by the following equation:
Figure BDA0003252232820000111
in addition to this, we want each topic subspace to reflect different semantics, and the embedding of different users under the same topic should be as similar as possible. Therefore, we set K topic prototype embedding
Figure BDA0003252232820000112
And encourages user embedding under k themes
Figure BDA0003252232820000113
Corresponding subject prototype mkSimilarly. Formalization, we aim at maximumAnd (3) conversion:
Figure BDA0003252232820000114
we therefore take this term as an additional training target and sum all users:
Figure BDA0003252232820000115
the complete training objective function is
Figure BDA0003252232820000116
Where η is the equilibrium coefficient. We optimized the parameters using a gradient descent method and Adam optimizer. To avoid the training process from being unstable, we also apply layer normalization and dropout techniques to user embedding.
In the present invention we propose a model TAN to model the subject-specific propagation dependence. Specifically, we jointly model the textual content information of the propagated item and the user propagation history sequence, and then propose a topic-aware attention mechanism for capturing historical propagation dependencies and time-decay effects under different topics. TAN can automatically learn topics and benefit from deep learning, as compared to traditional topic perception models. Compared with the current neural network-based model, TAN can not only effectively model subject-specific propagation patterns, but also better capture user dependence and location dependence. Meanwhile, the interpretability of the model is improved by extracting the theme information, so that the model makes a prediction and gives a prediction reason, namely, the information item of a specific theme is more likely to be forwarded by a user, and the forwarding operation can be carried out according to the attention weight, wherein the user is influenced by the user.
The above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: it is to be understood that modifications may be made to the technical solutions described in the foregoing embodiments, or equivalents may be substituted for some of the technical features thereof, but such modifications or substitutions do not depart from the spirit and scope of the technical solutions of the embodiments of the present invention.

Claims (10)

1. An information propagation prediction method based on a topic awareness attention network is characterized by comprising the following steps:
s1, integrating the topic context and the propagation history context into the user representation for prediction;
s2, the topic context supports the model building of the propagation mode aiming at the specific topic, and the propagation history context is further decomposed into user dependence model building and position dependence model building;
s3, constructing a user representation under multiple topics by using the user context obtained by encoding;
s4, further integrating the user representation through a time attenuation aggregation module, thereby obtaining a multi-topic cascade representation and then predicting the next infected user;
wherein each module is driven by information dissemination features.
2. The information propagation prediction method based on the topic awareness network according to claim 1, wherein the specific method in step S1 is:
given a set of users U, a concatenation set V and a propagation information set M, the propagation sequence of the ith information item in M is defined as a concatenation
Figure FDA0003252232810000011
Wherein the tuple
Figure FDA0003252232810000012
Representing a user
Figure FDA00032522328100000112
In that
Figure FDA00032522328100000113
Time quiltForwarding, and sequencing by time of infection, the propagation prediction task being defined as a given cascade ciAnd previous infected user sequence
Figure FDA0003252232810000013
Predict next infected user as
Figure FDA0003252232810000014
Wherein n is 1,2i|-1。
3. The information propagation prediction method based on the topic awareness attention network as claimed in claim 2, wherein the propagation mode modeling in step S2 is to encode the semantic information of the propagation information text by using the pre-trained language model BERT, specifically, to embed the text encoded by BERT into the propagation information text through a full connection layer
Figure FDA0003252232810000015
Conversion to propagating text embedding
Figure FDA0003252232810000016
yi=Wxxi+bx (1)
Wherein WxAnd bxRespectively a weight matrix and an offset vector.
4. The information propagation prediction method based on topic awareness attention network according to claim 3 wherein the user dependent modeling uses embedded matrix in step S2
Figure FDA0003252232810000017
And encoding the users, wherein | U | represents the number of users, and K and d represent the number of topics and the embedding dimension respectively.
5. The topic aware attention network based information dissemination module of claim 4The method is characterized in that for the cascade sequence
Figure FDA0003252232810000018
Each user in
Figure FDA0003252232810000019
User is embedded into
Figure FDA00032522328100000110
Wherein
Figure FDA00032522328100000111
Is the user embedding of the user under the kth topic.
6. The information propagation prediction method based on topic awareness network according to claim 5, wherein the position-dependent modeling in step S2 is to set a learnable position-embedded pos for each positionjWherein posjShared among all cascades.
7. The information propagation prediction method based on the topic awareness network according to claim 6, wherein the encoding method in step S3 is:
topic context:
computing user embedding for each topic k
Figure FDA0003252232810000021
And propagating text embedding yiCosine similarity between them, and normalized by softmax function:
Figure FDA0003252232810000022
wherein K is 1,2,.., K, and
Figure FDA0003252232810000023
representing a user
Figure FDA0003252232810000024
A weight under the kth topic; user-embedded representation of the aggregated topic context as
Figure FDA0003252232810000025
Propagation history context:
in a cascade sequence
Figure FDA0003252232810000026
With previous users
Figure FDA0003252232810000027
Is calculated by the following formula:
Figure FDA0003252232810000028
wherein
Figure FDA0003252232810000029
For topic-specific linear mapping of target users and previous users, respectively:
user' s
Figure FDA00032522328100000210
And the user
Figure FDA00032522328100000211
Complete attention score between
Figure FDA00032522328100000212
And weight
Figure FDA00032522328100000213
Used for describing propagation historyAnd calculated by the following formula:
Figure FDA00032522328100000214
Figure FDA00032522328100000215
wherein
Figure FDA00032522328100000216
Is a position dependent score from position m to position j;
full context-aware multi-topic user representation:
users in the k-th theme
Figure FDA00032522328100000217
Expressed as a weighted sum of previously infected users:
Figure FDA0003252232810000031
weighting of topic contexts
Figure FDA0003252232810000032
And location dependent score
Figure FDA0003252232810000033
Shared between different layers.
8. The information propagation prediction method based on the topic awareness attention network according to claim 1, wherein the modeling method of the time attenuation aggregation module in step S4 is:
converting the continuous-time decay into discrete time intervals:
Figure FDA0003252232810000034
wherein t is1By dividing the time range [0, Tmax]Is divided into L sub-intervals { [0, t ]1),..,[tL-1,Tmax) Where T ismaxIs the maximum timestamp in the data set, with a corresponding learnable weight for each time interval for each topic
Figure FDA0003252232810000035
9. The information propagation prediction method based on the topic awareness attention network according to claim 8, wherein the obtaining method of the multi-topic cascade representation in step S4 is:
the complete aggregation weight is calculated according to equation (8):
Figure FDA0003252232810000036
then the
Figure FDA0003252232810000037
Normalizing j 1, 2.. and n by a softmax function;
for each topic k, calculating
Figure FDA0003252232810000038
Being weighted
Figure FDA0003252232810000039
And a feedforward neural network with a ReLU activation function is adopted to endow the model with nonlinearity, and the output of the theme perception attention layer is represented in cascade as
Figure FDA00032522328100000310
10. The information dissemination prediction method based on the topic awareness network as claimed in claim 1, wherein the method for predicting the next infected user in step S4 is:
given a cascading sequence
Figure FDA00032522328100000311
By measuring user embedding
Figure FDA00032522328100000312
And cascade embedding
Figure FDA00032522328100000313
To parameterize the next infected user
Figure FDA00032522328100000314
Probability, cascade and user of
Figure FDA00032522328100000315
The interaction probability of (a) is expressed as:
Figure FDA00032522328100000316
wherein Θ represents all parameters that need to be learned;
the training objective for predicting an infected user is defined by equation (10):
Figure FDA00032522328100000317
setting K topic prototype embedding
Figure FDA00032522328100000318
And encourages user embedding under k themes
Figure FDA00032522328100000319
Corresponding subject prototype mkSimilarly, the goal is to maximize:
Figure FDA0003252232810000041
this term is taken as an additional training target and summed over all users:
Figure FDA0003252232810000042
the complete training objective function is
Figure FDA0003252232810000043
Where η is the equilibrium coefficient.
CN202111049168.8A 2021-09-08 2021-09-08 Information propagation prediction method based on topic perception attention network Active CN113688600B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111049168.8A CN113688600B (en) 2021-09-08 2021-09-08 Information propagation prediction method based on topic perception attention network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111049168.8A CN113688600B (en) 2021-09-08 2021-09-08 Information propagation prediction method based on topic perception attention network

Publications (2)

Publication Number Publication Date
CN113688600A true CN113688600A (en) 2021-11-23
CN113688600B CN113688600B (en) 2023-07-28

Family

ID=78585637

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111049168.8A Active CN113688600B (en) 2021-09-08 2021-09-08 Information propagation prediction method based on topic perception attention network

Country Status (1)

Country Link
CN (1) CN113688600B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114334159A (en) * 2022-03-16 2022-04-12 四川大学华西医院 Postoperative risk prediction natural language data enhancement model and method
CN114519606A (en) * 2022-01-29 2022-05-20 北京京东尚科信息技术有限公司 Information propagation effect prediction method and device
CN116955846A (en) * 2023-07-20 2023-10-27 重庆理工大学 Cascade information propagation prediction method integrating theme characteristics and cross attention

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180329884A1 (en) * 2017-05-12 2018-11-15 Rsvp Technologies Inc. Neural contextual conversation learning
AU2019100854A4 (en) * 2019-08-02 2019-09-05 Xi’an University of Technology Long-term trend prediction method based on network hotspot single-peak topic propagation model
CN112182423A (en) * 2020-10-14 2021-01-05 重庆邮电大学 Information propagation evolution trend prediction method based on attention mechanism
CN112380427A (en) * 2020-10-27 2021-02-19 中国科学院信息工程研究所 User interest prediction method based on iterative graph attention network and electronic device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180329884A1 (en) * 2017-05-12 2018-11-15 Rsvp Technologies Inc. Neural contextual conversation learning
AU2019100854A4 (en) * 2019-08-02 2019-09-05 Xi’an University of Technology Long-term trend prediction method based on network hotspot single-peak topic propagation model
CN112182423A (en) * 2020-10-14 2021-01-05 重庆邮电大学 Information propagation evolution trend prediction method based on attention mechanism
CN112380427A (en) * 2020-10-27 2021-02-19 中国科学院信息工程研究所 User interest prediction method based on iterative graph attention network and electronic device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张志扬;张凤荔;陈学勤;王瑞锦;: "基于分层注意力的信息级联预测模型", 计算机科学 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114519606A (en) * 2022-01-29 2022-05-20 北京京东尚科信息技术有限公司 Information propagation effect prediction method and device
CN114334159A (en) * 2022-03-16 2022-04-12 四川大学华西医院 Postoperative risk prediction natural language data enhancement model and method
CN114334159B (en) * 2022-03-16 2022-06-17 四川大学华西医院 Postoperative risk prediction natural language data enhancement model and method
CN116955846A (en) * 2023-07-20 2023-10-27 重庆理工大学 Cascade information propagation prediction method integrating theme characteristics and cross attention
CN116955846B (en) * 2023-07-20 2024-04-16 重庆理工大学 Cascade information propagation prediction method integrating theme characteristics and cross attention

Also Published As

Publication number Publication date
CN113688600B (en) 2023-07-28

Similar Documents

Publication Publication Date Title
CN113688600A (en) Information propagation prediction method based on topic perception attention network
Ji et al. Learning private neural language modeling with attentive aggregation
Xiong et al. Dcn+: Mixed objective and deep residual coattention for question answering
CN108876044B (en) Online content popularity prediction method based on knowledge-enhanced neural network
Han et al. Maximum information exploitation using broad learning system for large-scale chaotic time-series prediction
CN106294618A (en) Searching method and device
Chen et al. A survey on heterogeneous one-class collaborative filtering
Xiao et al. User behavior prediction of social hotspots based on multimessage interaction and neural network
Chen et al. Tracking dynamics of opinion behaviors with a content-based sequential opinion influence model
Hu et al. Self-attention-based temporary curiosity in reinforcement learning exploration
CN115409155A (en) Information cascade prediction system and method based on Transformer enhanced Hooke process
CN116051175A (en) Click rate prediction model and prediction method based on depth multi-interest network
CN117435715B (en) Question answering method for improving time sequence knowledge graph based on auxiliary supervision signals
CN113330462A (en) Neural network training using soft nearest neighbor loss
Sun et al. Tcsa-net: a temporal-context-based self-attention network for next location prediction
Xu et al. Knowledge graph-based reinforcement federated learning for Chinese question and answering
Deng et al. Risk management of investment projects based on artificial neural network
Liu et al. Model design and parameter optimization of CNN for side-channel cryptanalysis
An et al. Multiuser behavior recognition module based on dc-dmn
Zhao et al. Preference-aware Group Task Assignment in Spatial Crowdsourcing: Effectiveness and Efficiency
Yin et al. Time-Aware Smart City Services based on QoS Prediction: A Contrastive Learning Approach
Xue et al. Deep reinforcement learning based ontology meta-matching technique
Feng et al. A New Method of Microblog Rumor Detection Based on Transformer Model
CN113190733B (en) Network event popularity prediction method and system based on multiple platforms
CN116485501B (en) Graph neural network session recommendation method based on graph embedding and attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant