CN115422945A - Rumor detection method and system integrating emotion mining - Google Patents

Rumor detection method and system integrating emotion mining Download PDF

Info

Publication number
CN115422945A
CN115422945A CN202211139407.3A CN202211139407A CN115422945A CN 115422945 A CN115422945 A CN 115422945A CN 202211139407 A CN202211139407 A CN 202211139407A CN 115422945 A CN115422945 A CN 115422945A
Authority
CN
China
Prior art keywords
vector
content
knowledge
text content
word
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211139407.3A
Other languages
Chinese (zh)
Inventor
陈羽中
朱文龙
饶孟宇
万宇杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fuzhou University
Original Assignee
Fuzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fuzhou University filed Critical Fuzhou University
Priority to CN202211139407.3A priority Critical patent/CN115422945A/en
Publication of CN115422945A publication Critical patent/CN115422945A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention provides a rumor detection method integrating emotion mining, which comprises the following steps of; step A: collecting and extracting text content and comment content of a source post in a social network medium, and manually marking a real label of the source post to form a training data set DT; and B: training a deep learning network model N based on multi-level attention and a knowledge graph by using a training data set DT, wherein training content comprises authenticity labels for analyzing source posts and forecasting the source posts; and C: inputting the text content and the comment content of the source post into a trained deep learning network model N to obtain an authenticity label of the source post; the invention can improve the accuracy of rumor detection on the microblog.

Description

Rumor detection method and system integrating emotion mining
Technical Field
The invention relates to the technical field of natural language processing, in particular to a rumor detection method and system integrating emotion mining.
Background
Rumor Detection (Rumor Detection), also known as false news Detection, is an important task in the field of Natural Language Processing (NLP). Rumor detection can be regarded as a text classification problem with supervised learning, and can be generally classified into two types, rumors and not rumors. With the development of internet technology, social network platforms such as microblogs, twitter and the like are rapidly popular in the mass life. In social networking platforms, people are not just recipients of information but also creators of content. The social network platform greatly accelerates the speed and depth of information exchange between people. Social networking platforms are able to provide timely and comprehensive information about events occurring around the world, and therefore an increasing number of people are interested in participating in discussions and communications of hot topics on social networking platforms. This discussion and communication, on the one hand, facilitates the dissemination and dissemination of news, enabling people to more easily and quickly understand what is happening. However, in such a convenient environment, the social networking platform also reduces the cost of disseminating unrealistic information. The false rumors typically use false or forged images and aggressive languages, mislead the reader and spread quickly. The spread of the false rumors can have large-scale negative effects on society, causing social agitation.
In recent years, with the rise of deep learning technology, the technology is also widely applied by rumor detection tasks. The most common of these are Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). Since CNNs perform well in capturing semantic information from text, some researchers have applied them to content-based rumor detection. However, the network cannot take full advantage of the contextual information in the sentence, which is critical to modeling the semantic relationship between an aspect and its context. Therefore, the performance of CNN-based neural network models is limited in rumor detection tasks. To address this problem, many researchers have employed RNNs, particularly Long Short Term Memory (LSTM) and gated round-robin units (GRU), to extract contextual semantic information for rumors. Unlike CNN, RNN regards a sentence as a sequence of words, takes each word in time order, takes the output of the hidden layer as the input of the next hidden layer, and learns the context information in the sequence data continuously. Ma et al use recurrent neural networks to capture semantic changes between each source post and its forwarded comments and predict based on the semantic changes. The RNN-based neural network model is significantly better than the CNN-based neural network model in rumor detection.
Researchers have indicated that rumor characteristics of a given post are often determined by several keywords, rather than by all words in the context. While RNN cannot accurately estimate the contribution of different context words to the overall semantics. In contrast, the attention mechanism may capture the importance of each contextual word by computing an attention weight for each contextual word to the semantics of a given post and utilizing this attention weight to compute a semantic representation of the post.
However, most of these neural network models ignore emotional information in the post, which represents the emotion of the publisher to the content of the post, and this is especially important for correctly judging the authenticity label of the post. Recent learners have focused on finding unique emotional characteristics between the fake rumors and the real rumors. Ajao et al verified that there is a relationship between the authenticity of the news (true and false) and the use of emotional words, and designed an emotional signature (the ratio of the number of negative words to the number of positive words) to help detect false news. Furthermore, giachanou et al extract emotional features from news content for rumor detection based on an emotion dictionary. However, the related research in the past ignores syntax-dependent information and external knowledge information required in the aspect of emotion, so that emotion information is not sufficiently extracted.
Disclosure of Invention
The invention provides a rumor detection method and system integrating emotion mining, which can improve the accuracy of rumor detection on microblogs.
The invention adopts the following technical scheme.
A rumor detection method with emotion mining fused, the method comprising the following steps;
step A: collecting and extracting text content and comment content of a source post in a social network medium, and manually marking a real label of the source post to form a training data set DT;
and B, step B: training a deep learning network model N based on multi-level attention and a knowledge graph by using a training data set DT, wherein training content comprises authenticity of analysis source posts and authenticity labels for predicting the source posts;
and C: inputting the text content and the comment content of the source post into the trained deep learning network model N to obtain the authenticity label of the source post.
The step B comprises the following steps;
step B1: coding each training sample in the training data set DT to obtain an initial characterization vector T of the text content st Initial characterization vector T of comment content rt And syntactic adjacency matrix A st
And step B2: generating corresponding syntactic knowledge subgraph SK of text content from the knowledge map and the syntactic dependency graph according to the syntactic knowledge subgraph construction algorithm, and obtaining an adjacency matrix A thereof SK Then, the nodes are coded to obtain a node knowledge representation vector H of the syntax knowledge subgraph SK SK
And step B3: the text content initial characterization vector T obtained in the step B1 Tabanus Inputting the text content representation vector H into a bidirectional long-short term memory network Bi-LSTM to obtain a context-enhanced text content representation vector H st Let U st =H st (ii) a Then, the token vector T is sk And initial characterization vector T of comment content rt Inputting the data into a multi-head cross attention mechanism together to obtain a comment characterization vector P based on text content sr While characterizing the vector T st Inputting the text content enhancement representation vector Ps into a multi-head self-attention mechanism;then characterizing the vector P by the comments based on the text content sr And respectively inputting the text content enhanced representation vector Ps into the pooling layer to carry out average pooling operation to obtain an average pooled comment content sentence representation vector
Figure BDA0003852860510000021
And average pooled text content enhanced representation vector
Figure BDA0003852860510000022
And step B4: expressing node knowledge of sub-graph SK as vector H SK And the characterization vector U obtained in the step B3 st The method comprises the steps that the information is respectively input into two graph convolution networks with K layers, and the information is recorded as a text knowledge graph convolution network SKGCN and a text content graph convolution network SCGCN and used for learning external knowledge information and extracting syntax information; meanwhile, each layer of nodes of the text content graph convolution network SCGCN and the text knowledge graph convolution network SKGCN are subjected to knowledge guidance by using a knowledge guidance mechanism to obtain a graph knowledge representation vector V of a source post sks
And step B5: characterizing the vector V of the map knowledge obtained in the step B4 by using a cross attention mechanism sks And sentence characterization vector U st Fusing to obtain a knowledge enhanced sentence-level characterization vector E sd To further improve the ability of the model to extract information; then E is drawn by a multi-head self-attention mechanism sd Further strengthening to obtain sentence representation E of aggregated word-level information mt (ii) a Reducing noise from irregular sentences through a gating mechanism to obtain a source post emotion representation vector E sf
Step B6: representing vector of average pooling comment content sentences corresponding to source posts
Figure BDA0003852860510000034
And average pooled text content enhanced representation vector
Figure BDA0003852860510000035
All input into a multi-head cross attention mechanism and pass through an averaging cellComprehensive semantic representation C for transforming and obtaining comment content sr (ii) a The average pooled text content enhancement characterization vector is then
Figure BDA0003852860510000036
And comprehensive semantic representation C of comment content sr Inputting the semantic representation vector V into a fusion gating mechanism to obtain a fine-grained semantic representation vector V of the source post t
And step B7: representing the emotion vector E obtained in the step B5 sf And the fine-grained semantic representation vector V of the source post obtained in the step B6 t Combining to obtain final characterization vector E f (ii) a Then E is f Inputting a full connection layer and a softmax function to obtain a prediction result; calculating the gradient of each parameter in the deep learning network model by using a back propagation method according to the target loss function loss, and updating each parameter by using a random gradient descent method;
and step B8: and when the iterative change of the loss value generated by the deep learning network model N is smaller than a given threshold value or the maximum iteration number is reached, terminating the training process of the deep learning network model N.
The step B1 comprises the following steps;
step B11: traversing a training set DT, performing word segmentation on text content and comment content of a source post, and removing stop words, wherein each training sample in the DT is represented as DT = (st, rt, l); wherein st is the text content of the source post, rt is the comment content corresponding to the source post, l is the authenticity label corresponding to the source post, l belongs to { general fact, rumor, unverified rumor, rumor opened by public rumors };
the text content st of the source post is represented as:
Figure BDA0003852860510000031
wherein the content of the first and second substances,
Figure BDA0003852860510000032
for the i-th word in the text content st, i =1,2, …, nThe number of words that are the source post text content st;
the comment content rt of the source post is represented as:
Figure BDA0003852860510000033
wherein the content of the first and second substances,
Figure BDA0003852860510000041
for the jth word in the comment content rt, i =1,2, …, m, m is the number of words in the comment content rt;
step B12: for step B11, obtaining text content
Figure BDA0003852860510000042
Coding is carried out to obtain an initial characterization vector T of the text content st st ;T st Expressed as:
Figure BDA0003852860510000043
wherein the word vector matrix is pre-trained
Figure BDA0003852860510000044
Can be found to obtain
Figure BDA0003852860510000045
Is the ith word
Figure BDA0003852860510000046
Corresponding word vectors, d represents the dimensionality of the word vectors, and | V | is the number of words in the dictionary V;
step B13: for the comment content obtained in step B11
Figure BDA0003852860510000047
Coding is carried out to obtain an initial characterization vector T of the comment content rt rt ;T rt Expressed as:
Figure BDA0003852860510000048
wherein the word vector matrix is pre-trained
Figure BDA0003852860510000049
Can be found to obtain
Figure BDA00038528605100000410
Denotes the jth word
Figure BDA00038528605100000411
Corresponding word vectors, d represents the dimensionality of the word vectors, and | V | is the number of words in the dictionary V;
step B14: performing syntactic dependency analysis on the text content st to obtain a corresponding syntactic dependency tree DTD and an n-level syntactic adjacency matrix A st (ii) a The syntax dependency tree DTD is represented as,
Figure BDA00038528605100000412
wherein the content of the first and second substances,
Figure BDA00038528605100000413
words representing text content
Figure BDA00038528605100000414
And text content words
Figure BDA00038528605100000415
There is a syntactic dependency between them.
The step B2 comprises the following steps;
step B21: taking each original word node in a syntactic dependency tree DTD as a root node, expanding hop layers from a knowledge graph to generate child nodes, and selecting u nodes which are connected with the nodes of the previous layer in the knowledge graph with edges as the nodes of the layer on each layer, namely each seed node has
Figure BDA00038528605100000416
Expanding child nodes to finally obtain a syntactic knowledge sub-graph SK with the total number of all nodes being z = n + n × q and a z-order adjacency matrix A SK (ii) a The syntactic knowledge sub-graph SK is represented as,
Figure BDA00038528605100000417
wherein the content of the first and second substances,
Figure BDA0003852860510000051
meaning knowledge node words
Figure BDA0003852860510000052
Is a text content word
Figure BDA0003852860510000053
The number of the extended nodes of (1),
Figure BDA0003852860510000054
meaning knowledge node words
Figure BDA0003852860510000055
Is a knowledge node word
Figure BDA0003852860510000056
The knowledge-extending child node of (a),
Figure BDA0003852860510000057
words representing text content
Figure BDA0003852860510000058
And text content words
Figure BDA0003852860510000059
There is a syntactic dependency relationship between the two, u is the number of nodes selected in the knowledge graph, and hop is the number of layers of the topology;
step B22: the nodes of the sentence-method knowledge subgraph SK are encoded by embedding the knowledge graph to obtain the node knowledge expression vector of
Figure BDA00038528605100000510
Order to
Figure BDA00038528605100000511
As the initial input of a text knowledge graph convolution network SKGCN; knowledge word vector matrix in pre-training
Figure BDA00038528605100000512
Can be found to obtain
Figure BDA00038528605100000513
Is the ith word
Figure BDA00038528605100000514
And the corresponding knowledge word vector, wherein d represents the dimension of the knowledge word vector, and | V | is the word number of the knowledge word embedded in V.
The step B3 comprises the following steps;
step B31: initial characterization vector of text content
Figure BDA00038528605100000515
Sequentially and respectively inputting the forward layer and the reverse layer of the first bidirectional long-short term memory network to obtain the state vector sequence of the forward hidden layer and the state vector sequence of the reverse hidden layer, namely
Figure BDA00038528605100000516
And
Figure BDA00038528605100000517
wherein
Figure BDA00038528605100000518
i =1,2,.. N, f is the activation function; text content characterization vectors with context enhancement obtained through connection
Figure BDA00038528605100000519
Wherein the content of the first and second substances,
Figure BDA00038528605100000520
Figure BDA00038528605100000521
i =1,2, · n, ": "denotes a vector join operation; h st Is namely U st
Step B32: an initial characterization vector T of the text content st st And an initial token vector T of the review content rt rt Inputting the two into a multi-head cross attention mechanism together to obtain a comment characterization vector P based on text content sr The calculation formula is as follows:
P sr =MultiHead(T st ,T rt ,T rt ) A formula seven;
MultiHead(Q′,K′,V′)=Concat(head 1 ,head 2 ,…,head h )W o a formula eight;
head i =Attention(Q′W i Q ,K′W i K ,V′W i v ) A formula of nine;
Figure BDA0003852860510000061
wherein, multihead represents a multi-head attention mechanism, Q ', K ' and V ' represent input vectors of the multi-head attention mechanism, and an initial characterization vector T of text content st As a matrix Q', the initial token vector T of the corresponding review content rt rt As K 'and V'; head i The output vector calculated for the ith sub-vector of Q ', K ', V ' using Attention mechanism Attention (·), h being the number of heads of the multi-head Attention mechanism, W o Training parameters for a multi-headed attention mechanism, W i Q ,W i K
Figure BDA0003852860510000062
Is a weight matrix of the linear projection,
Figure BDA0003852860510000063
is a scale factor;
step B33: text content is initially characterized by a vector T st Inputting the text content enhancement representation vector P into a multi-head self-attention mechanism s The calculation formula is as follows:
P s =MultiHead(T st ,T st ,T st ) A formula eleven;
MultiHead(Q′,K′,V′)=Concat(head 1 ,head 2 ,…,head h )W 1 a formula twelve;
head i =Attention(Q′W i Q ,K′W i K ,V′W i V ) A formula thirteen;
Figure BDA0003852860510000064
wherein, multihead represents a multi-head attention mechanism, Q ', K ' and V ' represent input vectors of the multi-head attention mechanism, and an initial characterization vector T of text content st As matrices Q ', K ' and V '; head i The output vector calculated for the ith sub-vector of Q ', K ', V ' using Attention mechanism Attention (·), h being the number of heads of the multi-head Attention mechanism, W 1 Training parameters for a multi-headed attention mechanism, W i Q ,W i K
Figure BDA0003852860510000065
Is a weight matrix of the linear projection,
Figure BDA0003852860510000066
is a scale factor;
step B34: characterizing vectors P based on comments of text content sr And a text content enhanced representation vector P s Respectively inputting the data into a pooling layer to perform average pooling operation to obtain average pooling comment content sentence representation vectors
Figure BDA00038528605100000610
And averaging pooled text content enhancement characterization vectors
Figure BDA00038528605100000611
The calculation formula is as follows:
Figure BDA0003852860510000067
Figure BDA0003852860510000068
wherein the content of the first and second substances,
Figure BDA0003852860510000069
MeanPool is the average pooling function.
The step B4 comprises the following steps;
step B41: the sub-graph node knowledge characterization vector G obtained in the step B22 is used for SK,0 Input text knowledge graph convolution network SKGCN first layer graph convolution network using adjacency matrix A SK Updating the vector representation of each sub-graph node and outputting G SK ,1 And is used as the input of the next layer of graph convolution network;
wherein G is SK,1 Expressed as:
Figure BDA0003852860510000071
wherein the content of the first and second substances,
Figure BDA0003852860510000072
is the output of node i in the first layer graph convolution network,
Figure BDA0003852860510000073
the calculation formula of (a) is as follows:
Figure BDA0003852860510000074
Figure BDA0003852860510000075
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003852860510000076
is a bias term; w is a group of SK 、b SK Are all parameters which can be learnt, and the parameters,
Figure BDA0003852860510000077
as a weight matrix, relu is an activation function; node i in SKGCN and ith word in comment content
Figure BDA0003852860510000078
Correspondingly, the edges between the nodes represent the knowledge connection relationship between the words, d i D is selected to indicate the degree of the node i and to prevent the degree of the node i from being 0 and causing operation errors i +1 as a divisor;
step B42: for the text content graph convolution network SCGCN, the context enhanced text content representation vector U obtained in the step B31 is used st Inputting SCGCN first layer graph convolution network, using adjacency matrix A SK Updating the vector representation of each word and outputting U st,1
Wherein, U st,1 Expressed as:
Figure BDA0003852860510000079
wherein the content of the first and second substances,
Figure BDA00038528605100000710
is the output of node i in the first layer graph convolution network,
Figure BDA00038528605100000711
the calculation formula of (a) is as follows:
Figure BDA00038528605100000712
wherein, W st
Figure BDA00038528605100000713
Are all parameters which can be learnt, and the parameters,
Figure BDA00038528605100000714
in order to be a weight matrix, the weight matrix,
Figure BDA00038528605100000715
is a bias term; relu is an activation function; node i in graph convolution network and ith word in comment content
Figure BDA00038528605100000716
Correspondingly, the edges between nodes in the graph convolution network represent the syntactic dependency between words in the comment content, d i D is selected to indicate the degree of the node i and to prevent the degree of the node i from being 0 and causing operation errors i +1 as a divisor;
for the knowledge boot mechanism, the first layer output G of SKGCN SK,1 Discarding the content except the words in the current comment content sentence to obtain the first layer knowledge representation about the text content
Figure BDA00038528605100000717
Then using a cross attention mechanism to output the SCGCN with the first layer output U st,1 Combine to obtain knowledgeable review content sentence representation G SD,1 And is used as the input of the next layer of the SCGCN,
wherein G is SD,1 Expressed as:
Figure BDA0003852860510000081
wherein, the output of the node i in the SCGCN first layer graph convolution network through the knowledge guiding mechanism is
Figure BDA0003852860510000082
Is calculated byThe following were used:
Figure BDA0003852860510000083
Figure BDA0003852860510000084
Figure BDA0003852860510000085
wherein, (.) T Denotes a transpose operation, α i Is the attention weight of the knowledge about the ith word in the comment content s;
step B43: the input of the next layer graph convolution network of SKGCN and SCGCN is G SK,1 And G SD,1 Repeating the steps B41 and B42;
wherein, for SKGCN,
Figure BDA0003852860510000086
the output of the k layer graph convolution network is used as the input of the k +1 layer graph convolution network, and graph convolution characterization vectors are obtained after iteration is finished
Figure BDA0003852860510000087
For the case of the SCGCN,
Figure BDA0003852860510000088
for the output of the k-th layer graph convolution network, U is converted through a knowledge interaction mechanism st,k And G SD,k The method is used as the input of the (k + 1) th layer of graph convolution network, and graph convolution characterization vectors are obtained after continuous iteration and final end
Figure BDA0003852860510000089
Wherein K is more than or equal to 1 and less than or equal to K, and K is the layer number of the graph convolution network.
The step B5 comprises the following steps;
step B51: enhancing the context obtained in step B31Text content characterization vector U st And V obtained in step B43 sks Inputting an attention network, and selecting important knowledge information through the attention network to obtain a knowledge enhanced sentence-level characterization vector E sd The calculation formula is as follows:
Figure BDA00038528605100000810
Figure BDA0003852860510000091
Figure BDA0003852860510000092
wherein, (.) T Denotes the transposition operation, ∈ i Is the attention weight of the ith word in the comment content s;
step B52: using the knowledge enhanced sentence-level characterization vector E obtained in step 51 sd Inputting the sentence characterization vector E of the aggregated word-level information into a multi-head self-attention mechanism mt
E mt =MuliHead(E sd ,E sd ,E sd ) Twenty-nine of a formula;
step B53: for the noise brought by the non-standard sentence pair model, the sentence characterization vector E of the word-level information is aggregated mt Inputting a gating function to filter the irrelevant information to obtain a vector E sda (ii) a Then inputting the emotion expression vector into a multi-layer perceptron (MLP) to obtain an emotion representation vector E of the source post sf (ii) a The specific calculation process is as follows:
Figure BDA0003852860510000093
Figure BDA0003852860510000094
wherein the content of the first and second substances,
Figure BDA0003852860510000095
and
Figure BDA0003852860510000096
are all parameters which can be learnt, and the parameters,
Figure BDA0003852860510000097
and
Figure BDA0003852860510000098
in the form of a matrix of weights,
Figure BDA0003852860510000099
and
Figure BDA00038528605100000910
is the bias term.
The step B6 comprises the following steps;
step B61: representing vectors of all average pooled comment content sentences corresponding to source posts
Figure BDA00038528605100000911
And average pooled text content enhanced representation vector
Figure BDA00038528605100000912
Inputting the data into a multi-head cross attention mechanism together, and obtaining a comprehensive semantic representation C of the comment content through average pooling sr The calculation process is as follows:
Figure BDA00038528605100000913
C sr = MeanPool (C') formula thirty-three;
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA00038528605100000914
MeanPool is the average pooling function;
step B62: enhancing the average pooled text content with a characterization vector
Figure BDA00038528605100000915
And comprehensive semantic representation C of comment content sr Jointly input into a fusion gating mechanism to obtain a fine-grained semantic representation vector V of a source post t The calculation process is as follows:
Figure BDA0003852860510000101
Figure BDA0003852860510000102
wherein the content of the first and second substances,
Figure BDA0003852860510000103
is a sigmoid activation function, w 1
Figure BDA0003852860510000104
And
Figure BDA0003852860510000105
a parameter learnable in the fusion gating mechanism, which is a dot product operation.
The step B7 comprises the following steps;
step B71: using the source post emotion characterization vector E obtained in the step B53 sf And V obtained in step B62 t Connecting to obtain a final characterization vector E f The calculation formula is as follows:
E f =Concat(E sf ,V t ) A formula thirty-six;
wherein the content of the first and second substances,
Figure BDA00038528605100001010
concat is a vector join operation.
Step B72: final characterization vector E f Input to the fully-connected layer and normalized using softmax, compute text content pairsThe probability that the data belongs to each category is calculated as follows:
y=W 3 E f + b formula thirty-seven;
p c (y) = softmax (y) formula thirty-eight;
where y is the output vector of the fully connected layer,
Figure BDA0003852860510000106
is a weight matrix of the full connection layer,
Figure BDA0003852860510000107
bias term for fully connected layer, p c (y) is the probability of predicting the corresponding category of the text content as c, and p is more than or equal to 0 c (y) ≦ 1,c ∈ { general fact, rumor, unverified rumor, rumor daggered };
step B73: calculating a loss value by using the cross entropy as a loss function, updating the learning rate by using a gradient optimization algorithm Adam, and updating model parameters by using back propagation iteration so as to train a model by using a minimized loss function; the minimum loss function loss is calculated as follows:
Figure BDA0003852860510000108
wherein the content of the first and second substances,
Figure BDA0003852860510000109
is an L2 regularization term, λ is a learning rate, θ includes all parameters, and c is an authenticity label corresponding to the text content.
A rumor detection system integrating emotion mining adopts the rumor detection method, the social network media is microblog, and the rumor detection system comprises the following modules:
a data collection module: the method comprises the steps of extracting text content and comment content of a source post in a microblog, marking authenticity of the source post and constructing a training set;
a preprocessing module: the system is used for preprocessing training samples in a training set, and comprises word segmentation processing and stop word removal;
the coding module: the method comprises the steps of searching word vectors of words in preprocessed text content and comment content in a pre-trained word vector matrix to obtain initial token vectors of the text content and initial token vectors of the comment content, searching word vectors of nodes in a syntactic knowledge subgraph in a pre-trained knowledge graph word vector matrix to obtain initial token vectors of the syntactic knowledge subgraph related to the comment content;
a network training module: the deep learning network training system is used for inputting an initial characterization vector of text content, an initial characterization vector of comment content and a syntactic knowledge subgraph initial characterization vector into the deep learning network to obtain a final characterization vector and train the deep learning network according to the final characterization vector, and training the whole deep learning network by taking the probability that the characterization vector belongs to a certain class and marks in a training set as losses and taking minimized losses as a target to obtain a deep learning network model based on multi-level attention and a knowledge graph;
rumor detection module: and extracting semantic and emotional information in the input source post text content and comment content by using an NLP tool, analyzing the input source post text content and comment content by using a trained deep learning network model based on multi-level attention and knowledge maps, and outputting a predicted source post authenticity label.
Compared with the prior art, the invention has the following beneficial effects: the method obtains the syntactic knowledge subgraph of the corresponding comment sentence by using a knowledge map and subgraph generation strategy, then codes the comment content and the text content respectively, learns the syntactic dependency and the external knowledge in the comment content through two graph convolution networks and a knowledge guide mechanism, and filters sentence noise by using a gating mechanism to enhance the expression of the comment sentence. The invention also learns the fine-grained semantic information between the text content and the comment content by utilizing a multi-level attention mechanism. Compared with the prior art, the method can enhance the characteristic representation of the rumor by utilizing fine-grained semantic information and rich emotional information, so that the precision of the rumor detection is further improved, and the robustness of the rumor is enhanced.
Drawings
The invention is described in further detail below with reference to the following figures and detailed description:
FIG. 1 is a schematic flow chart of a method implementation of an embodiment of the present invention;
FIG. 2 is a schematic diagram of a model architecture in an embodiment of the invention;
fig. 3 is a schematic system configuration of an embodiment of the present invention.
Detailed Description
The invention is further explained below with reference to the drawings and the embodiments.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
As shown in the figure, a rumor detection method fusing emotion mining, the method comprises the following steps;
step A: collecting and extracting text content and comment content of a source post in a social network medium, and manually marking a real label of the source post to form a training data set DT;
and B: training a deep learning network model N based on multi-level attention and a knowledge graph by using a training data set DT, wherein training content comprises authenticity of analysis source posts and authenticity labels for predicting the source posts;
and C: and inputting the text content and the comment content of the source post into the trained deep learning network model N to obtain the authenticity label of the source post.
The step B comprises the following steps;
step B1: encoding each training sample in the training data set DT to obtain an initial characterization vector T of the text content st Initial characterization vector T of comment content rt And syntactic adjacency matrix A st
And step B2: generating a corresponding syntactic knowledge sub-graph SK of the text content from the knowledge graph and the syntactic dependency graph according to a syntactic knowledge sub-graph construction algorithm, and obtaining an adjacency matrix A of the syntactic knowledge sub-graph SK SK Then, the nodes are coded to obtain a node knowledge representation vector H of the syntax knowledge subgraph SK SK
And step B3: the text content initial characterization vector T obtained in the step B1 st Inputting the text content representation vector H into a bidirectional long-short term memory network Bi-LSTM to obtain a context-enhanced text content representation vector H st Let U st =H st (ii) a Then, the token vector T is st And initial characterization vector T of comment content rt Inputting the data into a multi-head cross attention mechanism together to obtain a comment characterization vector P based on text content sr While characterizing the vector T st Inputting the text content enhancement representation vector P into a multi-head self-attention mechanism s (ii) a Then characterizing the vector P by the comments based on the text content sr And a text content enhanced representation vector P s Respectively inputting the average into a pooling layer to perform average pooling operation to obtain average pooled comment content sentence characterization vectors
Figure BDA0003852860510000123
And average pooled text content enhanced representation vector
Figure BDA0003852860510000124
And step B4: expressing node knowledge of sub-graph SK as vector H SK And the characterization vector U obtained in the step B3 st Respectively inputting into two graph convolution networks with K layers, recording as text knowledge graph convolution network SKGCN and text content graph convolution network SCGCN, and learning external knowledge informationAnd extracting syntax information; meanwhile, each layer of nodes of the text content graph convolution network SCGCN and the text knowledge graph convolution network SKGCN are subjected to knowledge guidance by using a knowledge guidance mechanism to obtain a graph knowledge representation vector V of a source post sks
And step B5: characterizing vector V of the map knowledge obtained in the step B4 by using a cross attention mechanism sks And sentence characterization vector U st Fusing to obtain a knowledge enhanced sentence-level characterization vector E sd To further improve the ability of the model to extract information; then E is drawn by a multi-head self-attention mechanism sd Further strengthening to obtain sentence representation E of aggregated word-level information mt (ii) a Then reducing noise from irregular sentences through a gating mechanism to obtain source post emotion characterization vector E sf
And step B6: representing vector of average pooling comment content sentences corresponding to source posts
Figure BDA0003852860510000121
And average pooled text content enhanced representation vector
Figure BDA0003852860510000122
Inputting the data into a multi-head cross attention mechanism together, and obtaining a comprehensive semantic representation C of the comment content through average pooling sr (ii) a The average pooled text content is then enhanced with a characterization vector
Figure BDA0003852860510000131
And comprehensive semantic representation C of comment content sr Inputting the semantic representation vector V into a fusion gating mechanism to obtain a fine-grained semantic representation vector V of the source post t
Step B7: representing the emotion vector E obtained in the step B5 sf And the fine-grained semantic representation vector V of the source post obtained in the step B6 t Combining to obtain final characterization vector E f (ii) a Then E is mixed f Inputting a full connection layer and a softmax function to obtain a prediction result; calculating the gradient of each parameter in the deep learning network model by using a back propagation method according to the loss function loss of the target, and reducing by using a random gradientThe method updates each parameter;
and step B8: and when the iterative change of the loss value generated by the deep learning network model N is smaller than a given threshold value or the maximum iteration number is reached, terminating the training process of the deep learning network model N.
The step B1 comprises the following steps;
step B11: traversing the training set DT, performing word segmentation on text content and comment content of a source post in the training set DT and removing stop words, wherein each training sample in the DT is represented as DT = (st, rt, l); wherein st is the text content of the source post, rt is the comment content corresponding to the source post, l is the authenticity label corresponding to the source post, l belongs to { general fact, rumor, unverified rumor, rumor opened by public rumors };
the text content st of the source post is represented as:
Figure BDA0003852860510000132
wherein the content of the first and second substances,
Figure BDA0003852860510000133
the ith word in the text content st, i =1,2, …, n, n is the number of words in the text content st of the source post;
the comment content rt of the source post is represented as:
Figure BDA0003852860510000134
wherein the content of the first and second substances,
Figure BDA0003852860510000135
for the jth word in the comment content rt, i =1,2, …, m, m is the number of words in the comment content rt;
step B12: for step B11, obtaining text content
Figure BDA0003852860510000136
For encoding to obtain text content stInitial token vector T st ;T st Expressed as:
Figure BDA0003852860510000137
wherein the word vector matrix is pre-trained
Figure BDA0003852860510000138
Can be found to obtain
Figure BDA0003852860510000139
Is the ith word
Figure BDA00038528605100001310
Corresponding word vectors, d represents the dimensionality of the word vectors, and | V | is the number of words in the dictionary V;
step B13: for the comment content obtained in step B11
Figure BDA00038528605100001311
Coding is carried out to obtain an initial characterization vector T of the comment content rt rt ;T rt Expressed as:
Figure BDA0003852860510000141
wherein the word vector matrix is pre-trained
Figure BDA0003852860510000142
Can be found to obtain
Figure BDA0003852860510000143
Denotes the jth word
Figure BDA0003852860510000144
Corresponding word vectors, d represents the dimensionality of the word vectors, and | V | is the number of words in the dictionary V;
step B14: syntactic dependency parsing is carried out on the text content st to obtain a corresponding syntactic Dependency Tree (DTD) so as to obtain the corresponding syntactic dependency treeAnd an n-level syntactic adjacency matrix A st (ii) a The syntax dependency tree DTD is represented as,
Figure BDA0003852860510000145
wherein the content of the first and second substances,
Figure BDA0003852860510000146
words representing text content
Figure BDA0003852860510000147
And text content words
Figure BDA0003852860510000148
There is a syntactic dependency between them.
The step B2 comprises the following steps;
step B21: each original word node in the syntactic dependency tree DTD is used as a root node, hop layers are expanded from a knowledge graph to generate child nodes, and u nodes which are connected with the nodes of the previous layer in the knowledge graph in an edge mode are selected from each layer to serve as the nodes of the layer, namely each seed node has
Figure BDA0003852860510000149
Expanding child nodes to finally obtain a syntactic knowledge sub-graph SK with the total number of all nodes being z = n + n × q and a z-order adjacency matrix A SK (ii) a The syntactic knowledge sub-graph SK is represented as,
Figure BDA00038528605100001410
wherein the content of the first and second substances,
Figure BDA00038528605100001411
meaning knowledge node words
Figure BDA00038528605100001412
Is a text content word
Figure BDA00038528605100001413
The node of the expansion of (1) is,
Figure BDA00038528605100001414
representing knowledge node words
Figure BDA00038528605100001415
Is a knowledge node word
Figure BDA00038528605100001416
The knowledge-extending child node of (a),
Figure BDA00038528605100001417
words representing text content
Figure BDA00038528605100001418
And text content words
Figure BDA00038528605100001419
There is a syntactic dependency relationship between the two, u is the number of nodes selected in the knowledge graph, and hop is the number of layers of the topology;
step B22: the nodes of the sentence-method knowledge subgraph SK are encoded by embedding the knowledge graph to obtain the node knowledge expression vector of
Figure BDA00038528605100001420
Order to
Figure BDA00038528605100001421
As the initial input of a text knowledge graph convolution network SKGCN; knowledge word vector matrix in pre-training
Figure BDA00038528605100001422
Can be found
Figure BDA00038528605100001423
Is the ith word
Figure BDA00038528605100001424
Corresponding knowledge word vector, whichWhere d represents the dimension of the knowledge word vector, | V | is the number of words in which the knowledge word is embedded in V.
The step B3 comprises the following steps;
step B31: initial characterization vector of text content
Figure BDA0003852860510000151
Sequentially and respectively inputting the forward layer and the reverse layer of the first bidirectional long-short term memory network to obtain the state vector sequence of the forward hidden layer and the state vector sequence of the reverse hidden layer, namely
Figure BDA0003852860510000152
Wherein
Figure BDA0003852860510000153
i =1,2,.. N, f is the activation function; text content characterization vectors with context enhancement obtained through connection
Figure BDA0003852860510000154
Wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003852860510000155
Figure BDA0003852860510000156
i =1,2, · n, ": "denotes a vector join operation; h st Is namely U st
Step B32: an initial characterization vector T of the text content st st And an initial token vector T of the review content rt rt Inputting the two into a multi-head cross attention mechanism together to obtain a comment characterization vector P based on text content sr The calculation formula is as follows:
P sr =MultiHead(T st ,T rt ,T rt ) A formula seven;
MultiHead(Q′,K′,V′)=Concat(head 1 ,head 2 ,…,head h )W o a formula eight;
head i =Attention(Q′W i Q ,K′W i K ,V′W i V ) A formula of nine;
Figure BDA0003852860510000157
wherein, multihead represents a multi-head attention mechanism, Q ', K ' and V ' represent input vectors of the multi-head attention mechanism, and an initial characterization vector T of text content st As a matrix Q', the initial token vector T of the corresponding review content rt rt As K 'and V'; head i The output vector calculated for the ith sub-vector of Q ', K ', V ' using Attention mechanism Attention (·), h being the number of heads of the multi-head Attention mechanism, W o For the training parameters of the multi-head attention mechanism,
Figure BDA0003852860510000158
is a weight matrix of the linear projection and,
Figure BDA0003852860510000159
is a scale factor;
step B33: initially characterizing text content by a vector T st Inputting the text content enhancement representation vector P into a multi-head self-attention mechanism s The calculation formula is as follows:
P s =MultiHead(T st ,T st ,T st ) A formula eleven;
MultiHead(Q 1 ,K′,V′)=Concat(head 1 ,head 2 ,…,head h )W 1 a formula twelve;
head i =Attention(Q′W i Q ,K′W i K ,V′W i V ) A formula thirteen;
Figure BDA0003852860510000161
wherein, multihead represents a multi-head attention mechanism, Q ', K ' and V ' represent input vectors of the multi-head attention mechanism, and an initial characterization vector T of text content st As matrices Q ', K ' and V '; head i The output vector calculated using Attention mechanism Attention (-) for the ith sub-vector of Q ', K ', V ', h is the number of heads of the multi-head Attention mechanism, W 1 For the training parameters of the multi-head attention mechanism,
Figure BDA0003852860510000162
Figure BDA0003852860510000163
is a weight matrix of the linear projection,
Figure BDA0003852860510000164
is a scale factor;
step B34: characterizing vectors P based on comments of text content sr And a text content enhanced representation vector P s Respectively inputting the data into a pooling layer to perform average pooling operation to obtain average pooling comment content sentence representation vectors
Figure BDA0003852860510000165
And average pooled text content enhanced representation vector
Figure BDA0003852860510000166
The calculation formula is as follows:
Figure BDA0003852860510000167
Figure BDA0003852860510000168
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003852860510000169
MeanPool is the average pooling function.
The step B4 comprises the following steps;
step B41: the sub-graph node knowledge characterization vector G obtained in the step B22 is used for SK,0 Input text knowledge graph convolution network SKGCN first layer graph convolution network using adjacency matrix A SK Updating the vector representation of each sub-graph node and outputting G SK ,1 And is used as the input of the next layer of graph convolution network;
wherein G is SK,1 Expressed as:
Figure BDA00038528605100001610
wherein the content of the first and second substances,
Figure BDA00038528605100001611
is the output of node i in the first level graph convolution network,
Figure BDA00038528605100001612
the calculation formula of (a) is as follows:
Figure BDA00038528605100001613
Figure BDA00038528605100001614
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA00038528605100001615
is a bias term; w SK 、b SK Are all parameters which can be learnt, and the parameters,
Figure BDA00038528605100001616
as a weight matrix, relu is an activation function; node i in SKGCN and ith word in comment content
Figure BDA00038528605100001617
Correspondingly, the edges between the nodes represent the knowledge existing between the wordsConnection relation, d i D is selected to indicate the degree of the node i and to prevent the degree of the node i from being 0 and causing operation errors i +1 as a divisor;
step B42: for the text content graph convolution network SCGCN, the context enhanced text content representation vector U obtained in the step B31 is used st Inputting SCGCN first layer graph convolution network, using adjacency matrix A SK Updating the vector representation of each word and outputting U st,1
Wherein, U st,1 Expressed as:
Figure BDA0003852860510000171
wherein the content of the first and second substances,
Figure BDA0003852860510000172
is the output of node i in the first layer graph convolution network,
Figure BDA0003852860510000173
the calculation formula of (a) is as follows:
Figure BDA0003852860510000174
wherein, W st
Figure BDA0003852860510000175
Are all parameters which can be learnt, and the parameters,
Figure BDA0003852860510000176
in order to be a weight matrix, the weight matrix,
Figure BDA0003852860510000177
is a bias term; relu is an activation function; node i in graph convolution network and ith word in comment content
Figure BDA0003852860510000178
Correspondingly, edges between nodes in the graph convolution network represent syntactic dependencies between words in the comment content,d i D is selected to indicate the degree of the node i and to prevent the degree of the node i from being 0 and causing operation errors i +1 as a divisor;
for the knowledge-guided mechanism, the first layer output G to SKGCN SK,1 Discarding the content except the words in the current comment content sentence to obtain the first layer knowledge representation about the text content
Figure BDA0003852860510000179
Then using a cross attention mechanism to output the SCGCN with the first layer output U st,1 Combine to obtain knowledgeable review content sentence representation G SD,1 And is used as the input of the next layer of the SCGCN,
wherein G is SD,1 table Shown as follows:
Figure BDA00038528605100001710
wherein, the output of the node i in the SCGCN first layer graph convolution network through the knowledge guiding mechanism is
Figure BDA00038528605100001711
The calculation formula of (a) is as follows:
Figure BDA00038528605100001712
Figure BDA00038528605100001713
Figure BDA00038528605100001714
wherein, (.) T Denotes a transpose operation, α i Is the attention weight of the knowledge about the ith word in the comment content s;
step B43: the input of the next layer graph convolution network of SKGCN and SCGCN is SK,1 And G SD,1 Repeating the steps B41,B42;
Wherein, for SKGCN,
Figure BDA0003852860510000181
the output of the k layer graph convolution network is used as the input of the k +1 layer graph convolution network, and graph convolution characterization vectors are obtained after iteration is finished
Figure BDA0003852860510000182
For the case of the SCGCN,
Figure BDA0003852860510000183
for the output of the k-th layer graph convolution network, U is converted through a knowledge interaction mechanism st,k And G SD,k The method is used as the input of the (k + 1) th layer of graph convolution network, and graph convolution characterization vectors are obtained after continuous iteration and final end
Figure BDA0003852860510000184
Wherein K is more than or equal to 1 and less than or equal to K, and K is the layer number of the graph convolution network.
The step B5 comprises the following steps;
step B51: the text content characterization vector U with the enhanced context obtained in the step B31 is used for carrying out the context enhancement st And V obtained in step B43 sks Inputting an attention network, and selecting important knowledge information through the attention network to obtain a knowledge enhanced sentence-level characterization vector E sd The calculation formula is as follows:
Figure BDA0003852860510000185
Figure BDA0003852860510000186
Figure BDA0003852860510000187
wherein, (.) T Denotes the transposition operation, ∈ i Is to commentAttention weight of ith word in the theory s;
step B52: using the knowledge enhanced sentence-level characterization vector E obtained in step 51 sd Inputting the sentence characterization vector E of the aggregated word-level information into a multi-head self-attention mechanism mt
E mt =MuliHead(E sd ,E sd ,E sd ) Twenty-nine of a formula;
step B53: for the noise brought by the non-standard sentence pair model, the sentence characterization vector E of the word-level information is aggregated mt Inputting a gating function to filter the irrelevant information to obtain a vector E sda (ii) a Then inputting the emotion expression vector into a multi-layer perceptron (MLP) to obtain an emotion representation vector E of the source post sf (ii) a The specific calculation process is as follows:
Figure BDA0003852860510000188
Figure BDA0003852860510000189
wherein the content of the first and second substances,
Figure BDA0003852860510000191
and
Figure BDA0003852860510000192
are all parameters which can be learnt, and the parameters,
Figure BDA0003852860510000193
and
Figure BDA0003852860510000194
in order to be a weight matrix, the weight matrix,
Figure BDA0003852860510000195
and
Figure BDA0003852860510000196
is a bias term.
The step B6 comprises the following steps;
step B61: representing vectors of all average pooled comment content sentences corresponding to source posts
Figure BDA0003852860510000197
And average pooled text content enhanced representation vector
Figure BDA0003852860510000198
Inputting the data into a multi-head cross attention mechanism together, and obtaining a comprehensive semantic representation C of the comment content through average pooling sr The calculation process is as follows:
Figure BDA0003852860510000199
C sr = MeanPool (C') formula thirty-three;
wherein the content of the first and second substances,
Figure BDA00038528605100001910
MeanPool is the average pooling function;
step B62: enhancing the average pooled text content with a characterization vector
Figure BDA00038528605100001911
And comprehensive semantic representation C of comment content sr Jointly input into a fusion gating mechanism to obtain a fine-grained semantic representation vector V of a source post t The calculation process is as follows:
Figure BDA00038528605100001912
Figure BDA00038528605100001913
wherein the content of the first and second substances,
Figure BDA00038528605100001914
is a sigmoid activation function, w 1
Figure BDA00038528605100001915
And
Figure BDA00038528605100001916
a parameter learnable in the fusion gating mechanism, which is a dot product operation.
The step B7 comprises the following steps;
step B71: using the source post emotion characterization vector E obtained in the step B53 sf And V obtained in step B62 t Connecting to obtain a final characterization vector E f The calculation formula is as follows:
E f =Concat(E sf ,V t ) A formula thirty-six;
wherein the content of the first and second substances,
Figure BDA00038528605100001917
concat is a vector join operation.
Step B72: final characterization vector E f Inputting the text content into a full connection layer, normalizing by using softmax, and calculating the probability that the text content correspondingly belongs to each category, wherein the calculation formula is as follows:
y=W 3 E f + b formula thirty-seven;
p c (y) = softmax (y) formula thirty-eight;
where y is the output vector of the fully connected layer,
Figure BDA0003852860510000201
is a matrix of the weights of the full connection layer,
Figure BDA0003852860510000202
bias term for fully connected layer, p c (y) is the probability of predicting the corresponding category of the text content as c, and p is more than or equal to 0 c (y) ≦ 1,c ∈ { general fact, rumor, unverified rumor, rumor daggered };
step B73: calculating a loss value by using the cross entropy as a loss function, updating the learning rate by using a gradient optimization algorithm Adam, and updating model parameters by using back propagation iteration so as to train a model by using a minimized loss function; the minimum loss function loss is calculated as follows:
Figure BDA0003852860510000203
wherein the content of the first and second substances,
Figure BDA0003852860510000204
is an L2 regularization term, λ is a learning rate, θ includes all parameters, and c is an authenticity label corresponding to the text content.
A rumor detection system integrating emotion mining adopts the rumor detection method, the social network media is microblog, and the rumor detection system comprises the following modules:
a data collection module: the method comprises the steps of extracting text content and comment content of a source post in a microblog, marking authenticity of the source post and constructing a training set;
a preprocessing module: the system is used for preprocessing training samples in a training set, and comprises word segmentation processing and stop word removal;
the coding module: the method comprises the steps of searching word vectors of words in preprocessed text content and comment content in a pre-trained word vector matrix to obtain initial token vectors of the text content and initial token vectors of the comment content, searching word vectors of nodes in a syntactic knowledge subgraph in a pre-trained knowledge graph word vector matrix to obtain initial token vectors of the syntactic knowledge subgraph related to the comment content;
a network training module: the deep learning network training system is used for inputting an initial characterization vector of text content, an initial characterization vector of comment content and a syntactic knowledge subgraph initial characterization vector into the deep learning network to obtain a final characterization vector and train the deep learning network according to the final characterization vector, and training the whole deep learning network by taking the probability that the characterization vector belongs to a certain class and marks in a training set as losses and taking minimized losses as a target to obtain a deep learning network model based on multi-level attention and a knowledge graph;
rumor detection module: and extracting semantic and emotional information in the input source post text content and comment content by using an NLP tool, analyzing the input source post text content and comment content by using a trained deep learning network model based on multi-level attention and knowledge maps, and outputting a predicted source post authenticity label.
The foregoing is directed to preferred embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow. However, any simple modification, equivalent change and modification of the above embodiments according to the technical essence of the present invention are within the protection scope of the technical solution of the present invention.

Claims (10)

1. A rumor detection method integrating emotion mining is characterized in that: the method comprises the following steps;
step A: collecting and extracting text content and comment content of a source post in a social network medium, and manually marking a real label of the source post to form a training data set DT;
and B: training a deep learning network model N based on multi-level attention and a knowledge graph by using a training data set DT, wherein training content comprises authenticity of analysis source posts and authenticity labels for predicting the source posts;
and C: inputting the text content and the comment content of the source post into the trained deep learning network model N to obtain the authenticity label of the source post.
2. The method of claim 1, wherein the method comprises: the step B comprises the following steps;
step B1: encoding each training sample in the training data set DT to obtain an initial characterization vector T of the text content st First of comment contentStarting token vector T rt And syntactic adjacency matrix A st
And step B2: generating corresponding syntactic knowledge subgraph SK of text content from the knowledge map and the syntactic dependency graph according to the syntactic knowledge subgraph construction algorithm, and obtaining an adjacency matrix A thereof SK Then, the nodes are coded to obtain a node knowledge representation vector H of the syntax knowledge subgraph SK SK
And step B3: the text content initial characterization vector T obtained in the step B1 st Inputting the text content representation vector H into a bidirectional long-short term memory network Bi-LSTM to obtain a context-enhanced text content representation vector H st Let U st =H st (ii) a Then, the token vector T is st And initial characterization vector T of comment content rt Inputting the data into a multi-head cross attention mechanism together to obtain a comment characterization vector P based on text content sr While characterizing the vector T st Inputting the text content enhancement representation vector P into a multi-head self-attention mechanism s (ii) a Then characterizing the vector P by the comments based on the text content sr And text content enhancement characterization vector P s Respectively inputting the data into a pooling layer to perform average pooling operation to obtain average pooling comment content sentence representation vectors
Figure FDA0003852860500000011
And average pooled text content enhanced representation vector
Figure FDA0003852860500000012
B4, expressing the node knowledge of the sub-graph SK as a vector H SK And the characterization vector U obtained in the step B3 st The method comprises the steps that the data are respectively input into two graph convolution networks with K layers, and are recorded as a text knowledge graph convolution network SKGCN and a text content graph convolution network SCGCN, and the data are used for learning external knowledge information and extracting syntax information; meanwhile, each layer of nodes of the text content graph convolution network SCGCN and the text knowledge graph convolution network SKGCN are subjected to knowledge guidance by using a knowledge guidance mechanism to obtain a graph knowledge representation vector V of a source post sks
B5, characterizing a vector V of the graph knowledge obtained in the step B4 by using a cross attention mechanism sks And sentence characterization vector U st Fusing to obtain a knowledge enhanced sentence-level characterization vector E sd To further improve the ability of the model to extract information; then E is drawn by a multi-head self-attention mechanism sd Further strengthening, obtaining sentence representation E of aggregated word-level information mt (ii) a Reducing noise from irregular sentences through a gating mechanism to obtain a source post emotion representation vector E sf
Step B6: representing vector of average pooling comment content sentences corresponding to source posts
Figure FDA0003852860500000021
And average pooled text content enhanced representation vector
Figure FDA0003852860500000022
Inputting the data into a multi-head cross attention mechanism together, and obtaining a comprehensive semantic representation C of the comment content through average pooling sr (ii) a The average pooled text content is then enhanced with a characterization vector
Figure FDA0003852860500000023
And comprehensive semantic representation C of comment content sr Inputting the semantic representation vector V into a fusion gating mechanism to obtain a fine-grained semantic representation vector V of the source post t
Step B7: representing the emotion vector E obtained in the step B5 sf And the fine-grained semantic representation vector V of the source post obtained in the step B6 t Combining to obtain final characterization vector E f (ii) a Then E is mixed f Inputting a full connection layer and a softmax function to obtain a prediction result; calculating the gradient of each parameter in the deep learning network model by using a back propagation method according to the target loss function loss, and updating each parameter by using a random gradient descent method;
and step B8: and when the iterative change of the loss value generated by the deep learning network model N is smaller than a given threshold value or the maximum iteration number is reached, terminating the training process of the deep learning network model N.
3. The rumor detection method for fusion emotion mining, according to claim 2, characterized in that: the step B1 comprises the following steps;
step B11: traversing the training set DT, performing word segmentation on text content and comment content of a source post in the training set DT and removing stop words, wherein each training sample in the DT is represented as DT = (st, rt, l); wherein st is the text content of the source post, rt is the comment content corresponding to the source post, l is the authenticity label corresponding to the source post, l belongs to { general fact, rumor, unverified rumor, rumor opened by public rumors };
the text content st of the source post is represented as:
Figure FDA0003852860500000024
wherein the content of the first and second substances,
Figure FDA0003852860500000025
the ith word in the text content st, i =1,2, …, n, n is the number of words in the text content st of the source post;
the comment content rt of the source post is represented as:
Figure FDA0003852860500000026
wherein the content of the first and second substances,
Figure FDA0003852860500000027
for the jth word in the comment content rt, i =1,2, …, m, m is the number of words in the comment content rt;
step B12: for step B11, obtaining text content
Figure FDA0003852860500000031
Coding to obtain text contentst initial token vector T st ;T st Expressed as:
Figure FDA0003852860500000032
wherein the word vector matrix is pre-trained
Figure FDA0003852860500000033
Can be found to obtain
Figure FDA0003852860500000034
Is the ith word
Figure FDA0003852860500000035
Corresponding word vectors, d represents the dimensionality of the word vectors, and | V | is the number of words in the dictionary V;
step B13: for the comment content obtained in step B11
Figure FDA0003852860500000036
Coding is carried out to obtain an initial characterization vector T of the comment content rt rt ;T rt Expressed as:
Figure FDA0003852860500000037
wherein the word vector matrix is pre-trained
Figure FDA0003852860500000038
Can be found to obtain
Figure FDA0003852860500000039
Denotes the jth word
Figure FDA00038528605000000310
Corresponding word vectors, d represents the dimensionality of the word vectors, and | V | is the number of words in the dictionary V;
step B14: performing syntactic dependency analysis on the text content st to obtain a corresponding syntactic dependency tree DTD and an n-level syntactic adjacency matrix A st (ii) a The syntax dependency tree DTD is represented as,
Figure FDA00038528605000000311
wherein the content of the first and second substances,
Figure FDA00038528605000000312
words representing text content
Figure FDA00038528605000000313
And text content words
Figure FDA00038528605000000314
There is a syntactic dependency between them.
4. The method of claim 3, wherein the method comprises the steps of: the step B2 comprises the following steps;
step B21: taking each original word node in a syntactic dependency tree DTD as a root node, expanding hop layers from a knowledge graph to generate child nodes, and selecting u nodes which are connected with the nodes of the previous layer in the knowledge graph with edges as the nodes of the layer on each layer, namely each seed node has
Figure FDA00038528605000000315
Expanding child nodes to finally obtain a syntactic knowledge sub-graph SK with the total number of all nodes being z = n + n × q and a z-order adjacency matrix A SK (ii) a The syntactic knowledge sub-graph SK is represented as,
Figure FDA00038528605000000316
wherein the content of the first and second substances,
Figure FDA0003852860500000041
meaning knowledge node words
Figure FDA0003852860500000042
Is a text content word
Figure FDA0003852860500000043
The number of the extended nodes of (1),
Figure FDA0003852860500000044
meaning knowledge node words
Figure FDA0003852860500000045
Is a knowledge node word
Figure FDA0003852860500000046
The knowledge-extending child node of (a),
Figure FDA0003852860500000047
words representing text content
Figure FDA0003852860500000048
And text content words
Figure FDA0003852860500000049
There is a syntactic dependency relationship between the two, u is the number of nodes selected in the knowledge graph, and hop is the number of layers of the topology;
step B22: the nodes of the sentence-method knowledge subgraph SK are encoded by embedding the knowledge graph to obtain the node knowledge expression vector of
Figure FDA00038528605000000410
Order to
Figure FDA00038528605000000411
As the initial input of a text knowledge graph convolution network SKGCN; in the pre-trainingWord recognition vector matrix
Figure FDA00038528605000000412
Can be found to obtain
Figure FDA00038528605000000413
Is the ith word W i kg And the corresponding knowledge word vector, wherein d represents the dimension of the knowledge word vector, and | V | is the word number of the knowledge word embedded in V.
5. The method of claim 4, wherein the method comprises the steps of: the step B3 comprises the following steps;
step B31, initial characterization vectors of the text content
Figure FDA00038528605000000414
Sequentially and respectively inputting the forward layer and the reverse layer of the first bidirectional long-short term memory network to obtain the state vector sequence of the forward hidden layer and the state vector sequence of the reverse hidden layer, namely
Figure FDA00038528605000000415
And
Figure FDA00038528605000000416
wherein
Figure FDA00038528605000000417
Is an activation function; text content characterization vectors with context enhancement obtained through connection
Figure FDA00038528605000000418
Wherein, the first and the second end of the pipe are connected with each other,
Figure FDA00038528605000000419
Figure FDA00038528605000000420
": means a vector join operation; h st Is namely U st
Step B32: an initial characterization vector T of the text content st st And an initial token vector T of the review content rt rt Inputting the two into a multi-head cross attention mechanism together to obtain a comment characterization vector P based on text content sr The calculation formula is as follows:
P sr =MultiHead(T st ,T rt ,T rt ) A formula seven;
MultiHead(Q′,K′,V′)=Concat(head 1 ,head 2 ,…,head h )W o a formula eight;
Figure FDA00038528605000000421
Figure FDA0003852860500000051
wherein, multihead represents a multi-head attention mechanism, Q ', K ' and V ' represent input vectors of the multi-head attention mechanism, and an initial characterization vector T of text content st As a matrix Q', the initial token vector T of the corresponding review content rt rt As K 'and V'; head i The output vector calculated using Attention mechanism Attention (-) for the ith sub-vector of Q ', K ', V ', h is the number of heads of the multi-head Attention mechanism, W o For the training parameters of the multi-head attention mechanism,
Figure FDA0003852860500000052
is a weight matrix of the linear projection,
Figure FDA0003852860500000053
is a scale factor;
step B33: initially characterizing text content by a vector T st Inputting the text content enhancement representation vector P into a multi-head self-attention mechanism s The calculation formula is as follows:
P s =MultiHead(T st ,T st ,T st ) A formula eleven;
MultiHead(Q′,K′,V′)=Concat(head 1 ,head 2 ,…,head h )W 1 a formula twelve;
Figure FDA0003852860500000054
Figure FDA0003852860500000055
wherein, nultiHead represents a multi-head attention mechanism, Q ', K ' and V ' represent input vectors of the multi-head attention mechanism, and an initial characterization vector T of text content st As matrices Q ', K ' and V '; head i The output vector calculated for the ith sub-vector of Q ', K ', V ' using Attention mechanism Attention (·), h being the number of heads of the multi-head Attention mechanism, W 1 For the training parameters of the multi-head attention mechanism,
Figure FDA0003852860500000056
Figure FDA0003852860500000057
is a weight matrix of the linear projection and,
Figure FDA0003852860500000058
is a scale factor;
step B34: characterizing vectors P based on comments of text content sr And text content enhancement characterization vector P s Respectively inputting the data into a pooling layer to perform average pooling operation to obtain average pooling comment content sentence representation vectors
Figure FDA0003852860500000059
And average pooled text content enhanced representation vector
Figure FDA00038528605000000510
The calculation formula is as follows:
Figure FDA00038528605000000511
Figure FDA00038528605000000512
wherein the content of the first and second substances,
Figure FDA00038528605000000513
MeanPool is the average pooling function.
6. The method of claim 5, wherein the method comprises: the step B4 comprises the following steps;
step B41: the sub-graph node knowledge characterization vector G obtained in the step B22 is used for SK,0 Input text knowledge graph convolution network SKGCN first layer graph convolution network using adjacency matrix A SK Updating the vector representation of each sub-graph node and outputting G SK,1 And is used as the input of the next layer of graph convolution network;
wherein, G SK,1 Expressed as:
Figure FDA0003852860500000061
wherein the content of the first and second substances,
Figure FDA0003852860500000062
is the output of node i in the first level graph convolution network,
Figure FDA0003852860500000063
the calculation formula of (a) is as follows:
Figure FDA0003852860500000064
Figure FDA0003852860500000065
wherein the content of the first and second substances,
Figure FDA0003852860500000066
is a bias term; w SK 、b SK Are all parameters which can be learnt, and the parameters,
Figure FDA0003852860500000067
as a weight matrix, relu is an activation function; node i in SKGCN and ith word in comment content
Figure FDA0003852860500000068
Correspondingly, the edges between the nodes represent the knowledge connection relationship between the words, d i D is selected to indicate the degree of the node i and to prevent the degree of the node i from being 0 and causing operation errors i +1 as a divisor;
step B42: for the text content graph convolution network SCGCN, the text content characterization vector U with the enhanced context obtained in the step B31 is used for representing the text content st Inputting SCGCN first layer graph convolution network, using adjacency matrix A SK Updating the vector representation of each word and outputting U st,1
Wherein, U st,1 Expressed as:
Figure FDA0003852860500000069
wherein the content of the first and second substances,
Figure FDA00038528605000000610
is a middle section of a first layer graph convolution networkThe output of the point i is then taken,
Figure FDA00038528605000000611
the calculation formula of (a) is as follows:
Figure FDA00038528605000000612
wherein, W st
Figure FDA00038528605000000613
Are all parameters which can be learnt, and are,
Figure FDA00038528605000000614
in the form of a matrix of weights,
Figure FDA00038528605000000615
is a bias term; relu is an activation function; node i in graph convolution network and ith word in comment content
Figure FDA00038528605000000616
Correspondingly, the edges between nodes in the graph convolution network represent the syntactic dependency between words in the comment content, d i D is selected to indicate the degree of the node i and to prevent the degree of the node i from being 0 to cause operation error i +1 as a divisor;
for the knowledge-guided mechanism, the first layer output G to SKGCN SK,1 Discarding the content except the words in the current comment content sentence to obtain the first layer knowledge representation about the text content
Figure FDA00038528605000000617
Then using a cross attention mechanism to output the SCGCN with the first layer output U st,1 Combine to obtain knowledgeable review content sentence representation G SD,1 And is used as the input of the next layer of the SCGCN,
wherein G is SD,1 Expressed as:
Figure FDA0003852860500000071
wherein, the output of the node i in the SCGCN first layer graph convolution network through the knowledge guiding mechanism is
Figure FDA0003852860500000072
Figure FDA0003852860500000073
The calculation formula of (a) is as follows:
Figure FDA0003852860500000074
Figure FDA0003852860500000075
Figure FDA0003852860500000076
wherein, (. Cndot.) T Denotes a transpose operation, α i Is the attention weight of the knowledge about the ith word in the comment content s;
step B43: the input of the next layer graph convolution network of SKGCN and SCGCN is G SK,1 And G SD,1 Repeating the steps B41 and B42;
wherein, for SKGCN,
Figure FDA0003852860500000077
the output of the k layer graph convolution network is used as the input of the k +1 layer graph convolution network, and graph convolution characterization vectors are obtained after iteration is finished
Figure FDA0003852860500000078
For the case of the SCGCN,
Figure FDA0003852860500000079
for the output of the k-th layer graph convolution network, U is converted through a knowledge interaction mechanism st,k And G SD,k The method is used as the input of the (k + 1) th layer of graph convolution network, and graph convolution characterization vectors are obtained after continuous iteration and final end
Figure FDA00038528605000000710
Wherein K is more than or equal to 1 and less than or equal to K, and K is the layer number of the graph convolution network.
7. The method of claim 6, wherein the method comprises: the step B5 comprises the following steps;
step B51: the text content characterization vector U with the enhanced context obtained in the step B31 is used for carrying out the context enhancement st And V obtained in step B43 sks Inputting an attention network, and selecting important knowledge information through the attention network to obtain a knowledge enhanced sentence-level characterization vector E sd The calculation formula is as follows:
Figure FDA00038528605000000711
Figure FDA0003852860500000081
Figure FDA0003852860500000082
wherein, (. Cndot.) T Denotes the transposition operation, ε i Is the attention weight of the ith word in the comment content s;
step B52: using the knowledge enhancement type sentence level characterization vector E obtained in step 51 sd Inputting the sentence characterization vector E of the aggregated word-level information into a multi-head self-attention mechanism mt
E mt =MuliHead(E sd ,E sd ,E sd ) Twenty-nine of a formula;
step B53: for the noise brought by the non-standard sentence pair model, the sentence characterization vector E of the word-level information is aggregated mt Inputting a gating function to filter the irrelevant information to obtain a vector E sda (ii) a Then inputting the emotion expression vector into a multi-layer perceptron (MLP) to obtain an emotion representation vector E of the source post sf (ii) a The specific calculation process is as follows:
Figure FDA0003852860500000083
Figure FDA0003852860500000084
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0003852860500000085
and
Figure FDA0003852860500000086
are all parameters which can be learnt, and the parameters,
Figure FDA0003852860500000087
and
Figure FDA0003852860500000088
in the form of a matrix of weights,
Figure FDA0003852860500000089
and
Figure FDA00038528605000000810
is the bias term.
8. The method of claim 7, wherein the method comprises: the step B6 comprises the following steps;
step B61: representing vectors of all average pooled comment content sentences corresponding to source posts
Figure FDA00038528605000000811
And average pooled text content enhanced representation vector
Figure FDA00038528605000000812
Inputting the data into a multi-head cross attention mechanism together, and obtaining a comprehensive semantic representation C of the comment content through average pooling sr The calculation process is as follows:
Figure FDA00038528605000000813
C sr = MeanPool (C') formula thirty-three;
wherein the content of the first and second substances,
Figure FDA00038528605000000814
MeanPool is the average pooling function;
step B62: enhancing the average pooled text content with a characterization vector
Figure FDA00038528605000000815
And comprehensive semantic representation C of comment content sr Jointly input into a fusion gating mechanism to obtain a fine-grained semantic representation vector V of a source post t The calculation process is as follows:
Figure FDA0003852860500000091
Figure FDA0003852860500000092
wherein the content of the first and second substances,
Figure FDA0003852860500000093
is a sigmoid activation function that is,
Figure FDA0003852860500000094
and
Figure FDA0003852860500000095
a parameter learnable in the fusion gating mechanism, which is a dot product operation.
9. The method of claim 8, wherein the method comprises: the step B7 comprises the following steps;
step B71: using the source post emotion characterization vector E obtained in the step B53 sf And V obtained in step B62 t Connecting to obtain a final characterization vector E f The calculation formula is as follows:
E f =Concat(E sf ,V t ) A formula thirty-six;
wherein the content of the first and second substances,
Figure FDA0003852860500000096
concat is a vector join operation.
Step B72: final characterization vector E f Inputting the text content into a full connection layer, normalizing by using softmax, and calculating the probability that the text content correspondingly belongs to each category, wherein the calculation formula is as follows:
y=W 3 E f + b formula thirty-seven;
p c (y) = softmax (y) formula thirty-eight;
where y is the output vector of the fully connected layer,
Figure FDA0003852860500000097
is a matrix of the weights of the full connection layer,
Figure FDA0003852860500000098
bias term for fully connected layer, p c (y) predicting the textProbability of content corresponding to class c, 0 ≦ p c (y) ≦ 1,c ∈ { general fact, rumor, unverified rumor, rumor daggered };
step B73: calculating a loss value by using the cross entropy as a loss function, updating the learning rate by using a gradient optimization algorithm Adam, and updating model parameters by using back propagation iteration so as to train a model by using a minimized loss function; the minimum loss function loss is calculated as follows:
Figure FDA0003852860500000099
wherein the content of the first and second substances,
Figure FDA00038528605000000910
is an L2 regularization term, λ is a learning rate, θ includes all parameters, and c is an authenticity label corresponding to the text content.
10. A rumor detection system with emotion mining, which employs the rumor detection method of any one of claims 1 to 9, and is characterized in that: the social network media is a microblog, and the rumor detection system comprises the following modules:
a data collection module: the method comprises the steps of extracting text content and comment content of a source post in a microblog, marking authenticity of the source post and constructing a training set;
a preprocessing module: the system is used for preprocessing training samples in a training set, and comprises word segmentation processing and stop word removal;
and an encoding module: the method comprises the steps of searching word vectors of words in preprocessed text content and comment content in a pre-trained word vector matrix to obtain initial token vectors of the text content and initial token vectors of the comment content, searching word vectors of nodes in a syntactic knowledge subgraph in a pre-trained knowledge graph word vector matrix to obtain initial token vectors of the syntactic knowledge subgraph related to the comment content;
a network training module: the deep learning network model is used for inputting the initial characterization vector of the text content, the initial characterization vector of the comment content and the initial characterization vector of the syntactic knowledge subgraph into the deep learning network to obtain a final characterization vector and train the deep learning network by using the final characterization vector, the probability that the characterization vector belongs to a certain class and the marks in a training set as losses, and the whole deep learning network is trained by using the minimized losses as a target to obtain a deep learning network model based on multi-level attention and a knowledge graph;
rumor detection module: and extracting semantic and emotional information in the input source post text content and comment content by using an NLP tool, analyzing the input source post text content and comment content by using a trained deep learning network model based on multi-level attention and knowledge maps, and outputting a predicted source post authenticity label.
CN202211139407.3A 2022-09-19 2022-09-19 Rumor detection method and system integrating emotion mining Pending CN115422945A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211139407.3A CN115422945A (en) 2022-09-19 2022-09-19 Rumor detection method and system integrating emotion mining

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211139407.3A CN115422945A (en) 2022-09-19 2022-09-19 Rumor detection method and system integrating emotion mining

Publications (1)

Publication Number Publication Date
CN115422945A true CN115422945A (en) 2022-12-02

Family

ID=84204734

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211139407.3A Pending CN115422945A (en) 2022-09-19 2022-09-19 Rumor detection method and system integrating emotion mining

Country Status (1)

Country Link
CN (1) CN115422945A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117252264A (en) * 2023-11-20 2023-12-19 神思电子技术股份有限公司 Relation extraction method combining language model and graph neural network

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117252264A (en) * 2023-11-20 2023-12-19 神思电子技术股份有限公司 Relation extraction method combining language model and graph neural network
CN117252264B (en) * 2023-11-20 2024-02-02 神思电子技术股份有限公司 Relation extraction method combining language model and graph neural network

Similar Documents

Publication Publication Date Title
CN108984724B (en) Method for improving emotion classification accuracy of specific attributes by using high-dimensional representation
CN109376242B (en) Text classification method based on cyclic neural network variant and convolutional neural network
WO2023024412A1 (en) Visual question answering method and apparatus based on deep learning model, and medium and device
CN111274398B (en) Method and system for analyzing comment emotion of aspect-level user product
CN112613303B (en) Knowledge distillation-based cross-modal image aesthetic quality evaluation method
CN108549658A (en) A kind of deep learning video answering method and system based on the upper attention mechanism of syntactic analysis tree
CN111985205A (en) Aspect level emotion classification model
CN114443827A (en) Local information perception dialogue method and system based on pre-training language model
CN111914553B (en) Financial information negative main body judging method based on machine learning
CN114841151B (en) Medical text entity relation joint extraction method based on decomposition-recombination strategy
CN115659966A (en) Rumor detection method and system based on dynamic heteromorphic graph and multi-level attention
CN116028604A (en) Answer selection method and system based on knowledge enhancement graph convolution network
CN114492459A (en) Comment emotion analysis method and system based on convolution of knowledge graph and interaction graph
CN114742069A (en) Code similarity detection method and device
CN113051904B (en) Link prediction method for small-scale knowledge graph
CN115422945A (en) Rumor detection method and system integrating emotion mining
CN116661805B (en) Code representation generation method and device, storage medium and electronic equipment
CN114881038B (en) Chinese entity and relation extraction method and device based on span and attention mechanism
CN116663539A (en) Chinese entity and relationship joint extraction method and system based on Roberta and pointer network
CN116737897A (en) Intelligent building knowledge extraction model and method based on multiple modes
CN116258147A (en) Multimode comment emotion analysis method and system based on heterogram convolution
CN116662924A (en) Aspect-level multi-mode emotion analysis method based on dual-channel and attention mechanism
CN116644760A (en) Dialogue text emotion analysis method based on Bert model and double-channel model
CN116579347A (en) Comment text emotion analysis method, system, equipment and medium based on dynamic semantic feature fusion
CN113254575B (en) Machine reading understanding method and system based on multi-step evidence reasoning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination