CN113947074A - Deep collaborative interaction emotion reason joint extraction method - Google Patents

Deep collaborative interaction emotion reason joint extraction method Download PDF

Info

Publication number
CN113947074A
CN113947074A CN202111188307.5A CN202111188307A CN113947074A CN 113947074 A CN113947074 A CN 113947074A CN 202111188307 A CN202111188307 A CN 202111188307A CN 113947074 A CN113947074 A CN 113947074A
Authority
CN
China
Prior art keywords
emotion
reason
representation
representing
state
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111188307.5A
Other languages
Chinese (zh)
Inventor
史树敏
邬成浩
黄河燕
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202111188307.5A priority Critical patent/CN113947074A/en
Publication of CN113947074A publication Critical patent/CN113947074A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/237Lexical tools
    • G06F40/247Thesauruses; Synonyms
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9536Search customisation based on social or collaborative filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/01Social networking

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Business, Economics & Management (AREA)
  • Economics (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Machine Translation (AREA)

Abstract

The invention relates to a deep collaborative interaction emotion reason combined extraction method, and belongs to the technical field of natural language processing emotion analysis. The method adopts pre-trained word feature vectors to represent vectorized representation of each word in a text sequence, and uses a bidirectional long-time memory network to perform sentence-level text coding on the word representation fused with external knowledge. The importance of each word in the representation learning process is determined through an attention mechanism, so that a shallow emotion representation and a candidate reason representation are obtained. And (4) modeling association of the emotion representation and the reason representation by adopting multilayer collaborative attention network stacking, and outputting to obtain deep interactive emotion representation and reason representation. And finally, calculating the emotion probability vector and the reason probability vector simultaneously in a joint learning mode. The method can better capture the characteristics of the emotion and the reason of the text, can be simultaneously applied to the emotion reason extraction scene of the explicit emotion text and the implicit emotion text, and realizes the synchronous combined extraction of the emotion and the reason thereof.

Description

Deep collaborative interaction emotion reason joint extraction method
Technical Field
The invention relates to a deep collaborative interaction emotion reason joint extraction method, in particular to a method for deeply modeling potential relation between emotion and reason information by collaborative interaction and efficiently extracting emotion and reason contained in a text by joint learning, and belongs to the technical field of natural language processing emotion analysis.
Background
In recent years, with the development of the internet and social networks, text emotion reason extraction has become one of the most popular research directions in the field of natural language processing. The method can comprehensively and accurately understand the emotion expressed by the text, and mine the reason for the emotion, and can be applied to a plurality of scenes such as social public opinion analysis, client feedback tracking, system supervision, medical health supervision and the like, thereby generating wide social benefits.
Emotion cause extraction is a technique for deep text analysis and understanding that has appeared in recent years, and can identify causes of emotion expressed in a text. The existing emotion reason extraction method mainly aims at identifying and extracting reasons under the condition that the text has obvious emotion characteristics, namely, has explicit emotion words.
However, in daily expression, human emotions reflected in objective experiences of things and behaviors thereof are rich and abstract. Besides expressing emotion by using specific explicit emotion words, the emotion of the user is implicitly expressed by using objective statement or revival. Implicit emotions are defined as "language segments (sentences, clauses or phrases) that do not contain explicit emotional words but express subjective emotions". Because the expression of the implicit emotional text is more obscure and understandable than the explicit emotional text, in the process of extracting the emotional reasons from the implicit emotional text, the existing method cannot well capture the deep link between the implicit emotional information and the reason information and effectively extract the emotional reasons.
Therefore, further research on emotion reason extraction is needed, so that emotion reason extraction can be simultaneously applied to the explicit emotion text and the implicit emotion text, and the text information mining performance is improved more comprehensively and more accurately. The method has the advantages that the method plays a positive promoting role in research on aspects of natural language understanding, text expression learning, joint learning and the like, and further promotes the rapid development of application and industry in the relevant fields based on text emotion analysis.
Disclosure of Invention
The invention aims to solve the technical problems that the prior art cannot fully capture the deep connection between the emotion information and the reason information of a text, especially cannot be applied to the scene of an implicit emotion text, better solves the technical problems that the application scene is single, the connection between emotion and reason is sensed and extraction is carried out simultaneously, particularly enables emotion reason extraction to be simultaneously applied to an explicit emotion text and an implicit emotion text, and creatively provides an emotion reason combined extraction method for deep collaborative interaction.
First, the concept will be explained:
definition 1: emotional cause corpus
The method comprises the steps of providing corresponding specific emotion and candidate reasons for an emotion reason extraction task and a document set to be extracted, wherein documents in a corpus comprise explicit emotion texts and implicit emotion texts. Candidate reasons can be labeled at any level, such as clause level or tuple level.
Definition 2: text sequence s
The expression is as follows: { s ═ w1,w2,...,wNAnd indicating a sentence needing emotion analysis. There are N words w in the sentence1,w2,...,wNWhere the subscript N is the sentence word sequence length and w represents a word.
Definition 3: word feature vector for input text sequence
The pre-training vector used for vectorizing the input text sequence comprises a semantic vector and a position vector, wherein the semantic vector refers to semantic feature information of a current word, and the position vector refers to position feature information of the current word in the text sequence.
Definition 4: attention to
Refers to a phenomenon in which a human needs to select a specific portion in a visual region and then focus on it in order to make reasonable use of limited visual information processing resources.
Artificial intelligence exploits this phenomenon to provide neural networks with the ability to select specific inputs. In this method, attention is drawn to: if a word in a sentence is more relevant to the representation of the current sentence, the word is given a higher weight.
Definition 5: deep collaborative interaction network model
Refers to a network model that captures deep connections between two representations through a stack of multiple collaborative interaction layer models.
The invention is realized by adopting the following technical scheme.
A deep collaborative interaction emotion reason joint extraction method comprises the following steps:
step 1: and acquiring an emotion reason corpus, and acquiring a text document needing to be extracted and an emotion candidate reason text thereof.
Step 2: performing word segmentation processing on the input document needing emotion reason extraction and candidate reason text thereof to obtain a text sequence s of the corresponding document1And text sequences s of candidate reasons2
And step 3: respectively representing text sequences s by using pre-trained word feature vectors1And s2To obtain a word feature representation.
And taking the sum of the semantic representation and the position representation of each word as a word feature vector, thereby obtaining the feature vector corresponding to each word in the text sequence.
And 4, step 4: respectively comparing the text sequences s obtained in the step 3 by using a bidirectional long-time memory network LSTM1And s2The words in (1) represent the text encoding at the sentence level.
Specifically, the long-time memory network LSTM performs node state calculation in the network according to formulas 1 to 5:
i(t)=δ(Uix(t)+Wih(t-1)+bi) (1)
f(t)=δ(Ufx(t)+Wfh(t-1)+bf) (2)
o(t)=δ(Uox(t)+Woh(t-1)+bo) (3)
c(t)=f(t)⊙c(t-1)+i(t)⊙tanh(Ucx(t)+Wch(t-1)+bc) (4)
h(t)=o(t)⊙tanh(c(t)) (5) wherein x(t)、i(t)、f(t)、o(t)、c(t)、h(t)Respectively representing the input vector, the input gate state, the forgetting gate state, the output gate state, the memory unit state and the hidden layer state of the LSTM at the moment t, c(t-1)Represents the state of the memory cell of LSTM at time t-1, h(t-1)Representing the hidden layer state of the LSTM at time t-1; w, U, b respectively representing parameters of the circulation network structure, parameters of the input structure and deviation parameters; the symbol [ ] indicates an element product, and a sequence indicating vector corresponding to each sentence at the moment t is obtained according to the output of the model; δ () represents a sigmoid function; biRepresenting the parameter of the deviation term in calculating the state of the input gate, bfRepresenting the parameter of the deviation term when calculating the forgotten door state, boRepresenting the deviation term parameter in calculating the state of the output gate, bcRepresenting a deviation term parameter when calculating the state of the memory unit; u shapeiWeight matrix representing input vectors in calculating the state of the input gate, UfWeight matrix, U, representing input vectors when calculating a forgetting gate stateoWeight matrix representing input vectors in calculating output gate states, UcRepresenting a weight matrix of input vectors when calculating the state of the memory unit; wiWeight matrix representing the state of the hidden layer when calculating the state of the input gate, WfWeight matrix representing the state of the hidden layer when calculating the forgotten gate state, WoWeight matrix representing the state of the hidden layer when calculating the state of the output gate, WcA weight matrix representing the state of the hidden layer when calculating the state of the memory cell.
To obtain the information of the preceding and following words related to each word, the sequence is encoded using bi-directional LSTM, and the representation vectors of the ith element of the sequence obtained by forward and backward LSTM are spliced to obtain the final representation hi
Figure BDA0003300189670000041
Figure BDA0003300189670000042
Figure BDA0003300189670000043
Wherein the content of the first and second substances,
Figure BDA0003300189670000044
a hidden state representation of the ith element representing the forward LSTM network output;
Figure BDA0003300189670000045
represents a forward LSTM network; diA representation representing the ith element in the input sequence; l represents the number of words contained in the sentence;
Figure BDA0003300189670000046
a hidden state representation of the ith element representing the output to the LSTM network,
Figure BDA0003300189670000047
representing a backward LSTM network.
And 5: determining the importance of each word in the step 4 in the representation learning process through an Attention mechanism, and respectively calculating to obtain a text sequence s1E, s shallow emotion representation2Superficial layer reason tableC is shown.
Specifically, the attention score a of each word is calculatedi
Figure BDA0003300189670000048
Wherein h isiRepresenting the representation of the ith element in the input sequence, hjA representation representing the jth element in the input sequence; l is the number of words contained in the sentence; waEach sentence is represented as a group of word sequences, and further represented as a weighted average of the word representations of the word sequences after connection, wherein the weight is the Attention value calculated by formula 9.
Text sequence s1The shallow emotion representation E of (a) is of the form:
Figure BDA0003300189670000049
wherein L is1For a text sequence s1The number of words contained in (1).
Text sequence s2The reason for the shallow layer of (a) indicates that C is in the form:
Figure BDA00033001896700000410
wherein L is2For a text sequence s2The number of words contained in (1).
Step 6: and 5, taking the shallow emotion expression E and the shallow reason expression C output in the step 5 as the input of the deep collaborative interaction network model, and outputting to obtain the deep emotion expression and the deep reason expression.
Specifically, the M layers of cooperative attention network stacking are adopted to model the association of the emotion representation and the reason representation, and the emotion representation and the reason representation of deep interaction are output:
Em+1=Em+Softmax(Em((Cm)T))Cm (12)
Cm+1=Cm+Softmax(Cm((Em)T))Em (13)
wherein E ism+1Representing emotional representations at level m +1 in a collaborative interactive network, EmRepresenting emotion representation of the mth layer in the collaborative interactive network; cm+1A reason representation, C, representing the m +1 th layer in a collaborative interactive networkmRepresenting a reason representation of an m-th layer in the collaborative interactive network; t denotes a matrix transposition.
And 7: expressing the deep emotion obtained in step 6MAnd deep cause of CMRespectively jointly calculating corresponding emotion probability vectors y through a Softmax layerEAnd a cause probability vector yC
Specifically, the emotion probability vector yERepresents the probability of a particular specific emotion of the corresponding document:
Figure BDA0003300189670000051
reason probability vector yCRepresents the probability of whether the current candidate reason is causing the text emotion:
Figure BDA0003300189670000052
wherein the content of the first and second substances,
Figure BDA0003300189670000053
representing the probability of the document being an i-th implicit emotion,
Figure BDA0003300189670000054
a probability indicating a reason why the candidate reason is a document or not; w is aiDenotes the ith weight parameter, wjRepresents the jth weight parameter; biRepresenting the ith deviation term parameter, bjRepresents the jth deviation term parameter; k represents the dimension of the probability vector(ii) a T denotes a matrix transposition.
And after obtaining the emotion probability vector and the reason probability vector, updating the parameters by using the sum of the cross entropies of the emotion probability vector and the reason probability vector as a loss function and using a gradient descent mode to minimize the joint prediction error of the model.
So far, from step 1 to step 7, the emotion probability of the given document and the judgment result of the candidate reason are obtained, and the emotion reason joint extraction of the deep collaborative interaction is completed.
Advantageous effects
Compared with the prior art, the method of the invention has the following advantages:
1. the method can fully capture the deep connection between the text emotion information and the reason information, carry out deep modeling on the emotion and the reason, better capture the characteristics of the text emotion and the reason,
2. the method can be simultaneously applied to the emotion reason extraction scenes of the explicit emotion texts and the implicit emotion texts, can better solve the technical problems that the application scenes are single, the relation between the emotion and the reason is sensed, and extraction is carried out simultaneously, and achieves synchronous combined extraction of the emotion and the reason.
Drawings
FIG. 1 is a flow chart of the method of the present invention.
Detailed Description
The method of the present invention will be described in further detail below with reference to the accompanying drawings and examples.
As shown in fig. 1, a method for extracting emotion reason combination of deep collaborative interaction includes the following steps:
step A: and obtaining a document text to be extracted and a candidate reason text.
Specifically, in this embodiment, the same as the step 1 of the invention is performed;
and B: obtaining a text sequence representation;
specifically, in the embodiment, a document text sequence and a candidate reason text sequence are obtained, a pre-training word feature representation is adopted to initialize the text sequence, and a corresponding text sequence representation is obtained, which is specifically the same as the steps 2 to 3 of the invention content;
and C: obtaining shallow emotion semantic representation and reason semantic representation;
specifically, in the embodiment, the document text sequence representation and the candidate reason text sequence representation are respectively subjected to a bidirectional LSTM and an attention network coding sequence to obtain a shallow emotion semantic representation and a shallow reason semantic representation, which are specifically the same as the steps 4 to 5 of the invention content;
step D: obtaining deep interactive emotion representation and reason representation;
specifically, in the embodiment, the shallow emotion semantic representation and the reason semantic representation are subjected to a multi-layer stacked cooperative attention network, deep contact between the shallow emotion semantic representation and the reason semantic representation is captured, and a deep interactive emotion representation and a deep interactive reason representation are obtained, which are the same as the step 6 of the invention content;
step E: jointly calculating the emotion and reason probabilities;
specifically, in this embodiment, the emotion and reason probabilities are calculated simultaneously by using a joint learning method, which is the same as the step 7 of the invention.
Examples
Taking the emotion reason extraction corpus "funeral that I are eating fried chicken in a dining room and suddenly invited to the best friend" as an embodiment, the embodiment will explain the specific operation steps of the method in detail by using specific examples.
As shown in fig. 1, a method for extracting emotion reason combination of deep collaborative interaction includes the following steps:
step A: and obtaining a document text to be extracted and a candidate reason text.
Specifically, in the present embodiment, the document text "i am eating a fried chicken in a dining room, funeral suddenly invited to the best friend" and the candidate reason text "funeral suddenly invited to the best friend" are obtained.
And B: a text sequence representation is obtained.
Specifically, in the embodiment, the document text and the candidate reason text are segmented, and Word2Vec Word feature vectors pre-trained on a large-scale text data set for each Word are obtained to obtain a sequence representation of the document text "i am eating a fried chicken in a dining room and suddenly invited to the funeral of the best friend" and a sequence representation of the candidate reason text "suddenly invited to the funeral of the best friend".
And C: and obtaining shallow emotion semantic representation and reason semantic representation.
Specifically, in the embodiment, the document text sequence and the reason text sequence are input into the bidirectional LSTM and attention network for coding learning, and a shallow emotion semantic representation of "i eat a fried chicken in a dining room and suddenly invite a funeral to a best friend" and a shallow reason semantic representation of "suddenly invite a funeral to a best friend" are obtained.
Step D: and obtaining deep interactive emotion representation and reason representation.
Specifically, in the embodiment, the shallow emotion semantic representation and the shallow reason semantic representation are input to the 3-layer stacked collaborative attention network together, so as to obtain the deep interactive emotion representation and reason representation.
Step E: and jointly calculating the emotion and reason probabilities.
Specifically, in the embodiment, the deep interactive emotion expression and the deep reason expression are respectively input into the Softmax layer to calculate the corresponding probability vectors, so as to obtain the emotion to which the document most likely belongs and whether the candidate reason is the reason of the document emotion.

Claims (5)

1. A deep collaborative interaction emotion reason joint extraction method is characterized by comprising the following steps:
step 1: acquiring an emotion reason corpus, and acquiring text documents needing to be extracted and emotion candidate reason texts thereof;
the emotion reason corpus is a corresponding specific emotion and candidate reason provided for the emotion reason extraction task and a document set to be extracted, wherein documents in the corpus comprise explicit emotion texts and implicit emotion texts; the candidate reasons can be marked at any level;
step 2: the method comprises the steps of performing word segmentation on an input document needing emotion reason extraction and a candidate reason text thereof,obtaining a text sequence s of a corresponding document1And text sequences s of candidate reasons2
Wherein the expression of the text sequence s is as follows: { s ═ w1,w2,...,wNRepresents a sentence with N words w to be analyzed1,w2,...,wNWherein, subscript N is the length of sentence word sequence, and w represents a word;
and step 3: respectively representing text sequences s by using pre-trained word feature vectors1And s2The vectorization representation of each word in the word graph is obtained to obtain the word feature representation; the sum of semantic representation and position representation of each word is used as a word feature vector, so that a feature vector corresponding to each word in a text sequence is obtained;
the word feature vector is a pre-training vector used for vectorizing the input text sequence, and comprises a semantic vector and a position vector, wherein the semantic vector refers to semantic feature information of a current word, and the position vector refers to position feature information of the current word in the text sequence;
and 4, step 4: respectively comparing the text sequences s obtained in the step 3 by using a bidirectional long-time memory network LSTM1And s2The words in (1) represent the text coding at sentence level;
and 5: determining the importance of each word in the step 4 in the representation learning process through an Attention mechanism, and respectively calculating to obtain a text sequence s1E, s shallow emotion representation2The shallow cause of (a) represents C;
step 6: taking the shallow emotion expression E and the shallow reason expression C output in the step 5 as the input of the deep collaborative interaction network model, and outputting to obtain a deep emotion expression and a deep reason expression;
and 7: expressing the deep emotion obtained in step 6MAnd deep cause of CMRespectively jointly calculating corresponding emotion probability vectors y through a Softmax layerEAnd a cause probability vector yC
And after obtaining the emotion probability vector and the reason probability vector, updating the parameters by using the sum of the cross entropies of the emotion probability vector and the reason probability vector as a loss function and using a gradient descent mode to minimize the joint prediction error of the model.
2. The method for extracting emotion reason combination of deep collaborative interaction according to claim 1, wherein in step 4, the long-time memory network LSTM performs node state calculation in the network according to equations 1 to 5:
i(t)=δ(Uix(t)+Wih(t-1)+bi) (1)
f(t)=δ(Ufx(t)+Wfh(t-1)+bf) (2)
o(t)=δ(Uox(t)+Woh(t-1)+bo) (3)
c(t)=f(t)⊙c(t-1)+i(t)⊙tanh(Ucx(t)+Wch(t-1)+bc) (4)
h(t)=o(t)⊙tanh(c(t)) (5)
wherein x is(t)、i(t)、f(t)、o(t)、c(t)、h(t)Respectively representing the input vector, the input gate state, the forgetting gate state, the output gate state, the memory unit state and the hidden layer state of the LSTM at the moment t, c(t-1)Represents the state of the memory cell of LSTM at time t-1, h(t-1)Representing the hidden layer state of the LSTM at time t-1; w, U, b respectively representing parameters of the circulation network structure, parameters of the input structure and deviation parameters; the symbol [ ] indicates an element product, and a sequence indicating vector corresponding to each sentence at the moment t is obtained according to the output of the model; δ () represents a sigmoid function; biRepresenting the parameter of the deviation term in calculating the state of the input gate, bfRepresenting the parameter of the deviation term when calculating the forgotten door state, boRepresenting the deviation term parameter in calculating the state of the output gate, bcRepresenting a deviation term parameter when calculating the state of the memory unit;Uiweight matrix representing input vectors in calculating the state of the input gate, UfWeight matrix, U, representing input vectors when calculating a forgetting gate stateoWeight matrix representing input vectors in calculating output gate states, UcRepresenting a weight matrix of input vectors when calculating the state of the memory unit; wiWeight matrix representing the state of the hidden layer when calculating the state of the input gate, WfWeight matrix representing the state of the hidden layer when calculating the forgotten gate state, WoWeight matrix representing the state of the hidden layer when calculating the state of the output gate, WcRepresenting a weight matrix of a hidden layer state when calculating the state of the memory unit;
to obtain the information of the preceding and following words related to each word, the sequence is encoded using bi-directional LSTM, and the representation vectors of the ith element of the sequence obtained by forward and backward LSTM are spliced to obtain the final representation hi
Figure FDA0003300189660000021
Figure FDA0003300189660000022
Figure FDA0003300189660000023
Wherein the content of the first and second substances,
Figure FDA0003300189660000024
a hidden state representation of the ith element representing the forward LSTM network output;
Figure FDA0003300189660000025
represents a forward LSTM network; diA representation representing the ith element in the input sequence; l represents the number of words contained in the sentence;
Figure FDA0003300189660000031
a hidden state representation of the ith element representing the output to the LSTM network,
Figure FDA0003300189660000032
representing a backward LSTM network.
3. The method for joint extraction of emotional causes of deep collaborative interaction according to claim 1, wherein in step 5, an attention score a of each word is calculatedi
Figure FDA0003300189660000033
Wherein h isiRepresenting the representation of the ith element in the input sequence, hjA representation representing the jth element in the input sequence; l is the number of words contained in the sentence; waEach sentence is expressed as a group of word sequences and further expressed as a weighted average of the expression of each connected word in the word sequences, wherein the weight is the Attention value calculated by the formula 9;
text sequence s1The shallow emotion representation E of (a) is of the form:
Figure FDA0003300189660000034
wherein L is1For a text sequence s1The number of words contained in (1);
text sequence s2The reason for the shallow layer of (a) indicates that C is in the form:
Figure FDA0003300189660000035
wherein L is2For a text sequence s2The number of words contained in (1).
4. The method for extracting emotion reason combination of deep collaborative interaction according to claim 1, wherein in step 6, the association between emotion representation and reason representation is modeled by stacking M layers of collaborative attention networks, and the emotion representation and reason representation of deep collaborative interaction are output:
Em+1=Em+Softmax(Em((Cm)T))Cm (12)
Cm+1=Cm+Softmax(Cm((Em)T))Em (13)
wherein E ism+1Representing emotional representations at level m +1 in a collaborative interactive network, EmRepresenting emotion representation of the mth layer in the collaborative interactive network; cm+1A reason representation, C, representing the m +1 th layer in a collaborative interactive networkmRepresenting a reason representation of an m-th layer in the collaborative interactive network; t denotes a matrix transposition.
5. The method for jointly extracting emotion reason for deep collaborative interaction according to claim 1, wherein in step 7, emotion probability vector yERepresents the probability of a particular specific emotion of the corresponding document:
Figure FDA0003300189660000036
reason probability vector yCRepresents the probability of whether the current candidate reason is causing the text emotion:
Figure FDA0003300189660000041
wherein the content of the first and second substances,
Figure FDA0003300189660000042
indicating that the document is the ith implicit conditionThe probability of the sensation is that the user is,
Figure FDA0003300189660000043
a probability indicating a reason why the candidate reason is a document or not; w is aiDenotes the ith weight parameter, wjRepresents the jth weight parameter; biRepresenting the ith deviation term parameter, bjRepresents the jth deviation term parameter; k represents the dimension of the probability vector; t denotes a matrix transposition.
CN202111188307.5A 2021-10-12 2021-10-12 Deep collaborative interaction emotion reason joint extraction method Pending CN113947074A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111188307.5A CN113947074A (en) 2021-10-12 2021-10-12 Deep collaborative interaction emotion reason joint extraction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111188307.5A CN113947074A (en) 2021-10-12 2021-10-12 Deep collaborative interaction emotion reason joint extraction method

Publications (1)

Publication Number Publication Date
CN113947074A true CN113947074A (en) 2022-01-18

Family

ID=79330226

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111188307.5A Pending CN113947074A (en) 2021-10-12 2021-10-12 Deep collaborative interaction emotion reason joint extraction method

Country Status (1)

Country Link
CN (1) CN113947074A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117455497A (en) * 2023-11-12 2024-01-26 北京营加品牌管理有限公司 Transaction risk detection method and device
CN117787267A (en) * 2023-12-29 2024-03-29 广东外语外贸大学 Emotion cause pair extraction method and system based on neural network

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117455497A (en) * 2023-11-12 2024-01-26 北京营加品牌管理有限公司 Transaction risk detection method and device
CN117787267A (en) * 2023-12-29 2024-03-29 广东外语外贸大学 Emotion cause pair extraction method and system based on neural network
CN117787267B (en) * 2023-12-29 2024-06-07 广东外语外贸大学 Emotion cause pair extraction method and system based on neural network

Similar Documents

Publication Publication Date Title
CN111444709B (en) Text classification method, device, storage medium and equipment
CN110083705B (en) Multi-hop attention depth model, method, storage medium and terminal for target emotion classification
CN109753566B (en) Model training method for cross-domain emotion analysis based on convolutional neural network
CN109614471B (en) Open type problem automatic generation method based on generation type countermeasure network
CN110110324B (en) Biomedical entity linking method based on knowledge representation
CN110222163A (en) A kind of intelligent answer method and system merging CNN and two-way LSTM
Wen et al. Dynamic interactive multiview memory network for emotion recognition in conversation
CN111460132B (en) Generation type conference abstract method based on graph convolution neural network
CN113435211B (en) Text implicit emotion analysis method combined with external knowledge
CN107247702A (en) A kind of text emotion analysis and processing method and system
CN111382565A (en) Multi-label-based emotion-reason pair extraction method and system
CN108363695B (en) User comment attribute extraction method based on bidirectional dependency syntax tree representation
CN110287323B (en) Target-oriented emotion classification method
CN110765775A (en) Self-adaptive method for named entity recognition field fusing semantics and label differences
CN110969020A (en) CNN and attention mechanism-based Chinese named entity identification method, system and medium
CN112527966B (en) Network text emotion analysis method based on Bi-GRU neural network and self-attention mechanism
CN113947074A (en) Deep collaborative interaction emotion reason joint extraction method
CN113065344A (en) Cross-corpus emotion recognition method based on transfer learning and attention mechanism
CN110991190A (en) Document theme enhanced self-attention network, text emotion prediction system and method
CN109271636B (en) Training method and device for word embedding model
CN111737467A (en) Object-level emotion classification method based on segmented convolutional neural network
CN111046134A (en) Dialog generation method based on replying person personal feature enhancement
CN115374281A (en) Session emotion analysis method based on multi-granularity fusion and graph convolution network
CN115392232A (en) Topic and multi-mode fused emergency emotion analysis method
Yao et al. New Technologies to Enhance Computer Generated Interactive Virtual Humans

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination