CN112905776A - Emotional dialogue model construction method, emotional dialogue system and method - Google Patents
Emotional dialogue model construction method, emotional dialogue system and method Download PDFInfo
- Publication number
- CN112905776A CN112905776A CN202110283821.0A CN202110283821A CN112905776A CN 112905776 A CN112905776 A CN 112905776A CN 202110283821 A CN202110283821 A CN 202110283821A CN 112905776 A CN112905776 A CN 112905776A
- Authority
- CN
- China
- Prior art keywords
- emotion
- reply
- model
- dialogue
- discriminator
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/332—Query formulation
- G06F16/3329—Natural language query formulation or dialogue systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/951—Indexing; Web crawling techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
- G06F40/35—Discourse or dialogue representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Software Systems (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Machine Translation (AREA)
Abstract
The invention discloses an emotional dialogue model construction method, an emotional dialogue system and an emotional dialogue method. The disclosed method adopts production confrontation network training to obtain a corresponding model, wherein a generator is used for generating a dialogue reply, a content discriminator can distinguish whether an input reply text sequence belongs to a 'universal reply', and participates in the confrontation training process of the model; the emotion discriminator can distinguish whether the emotion type of the reply text generated by the generation model is a certain appointed emotion type, and the emotion type of the dialogue text generated by the generation model can be guided to be closer to the certain appointed emotion type by adding the emotion discrimination model. In the man-machine conversation process, the method and the system can generate the conversation reply meeting specific emotion, so that the reply diversity is realized, the emotion consistency is also met, the quality of the reply generated by the machine can be improved, and the user experience in man-machine interaction can be improved.
Description
Technical Field
The invention belongs to the technical field of man-machine conversation, and particularly relates to an emotion conversation model construction method and a related emotion conversation system and method.
Background
As a novel man-machine interaction mode, the man-machine conversation can save manpower to a great extent, is mainly divided into task type conversation and chatting type conversation in function, and is respectively used for meeting different user requirements.
The task type conversation can be used in scenes such as intelligent customer service, ticket booking, weather inquiry and the like in a certain fixed field, such as Taobao intelligent customer service, Jingdong intelligent customer service and the like, and the intention understanding, conversation management and the like of a user are realized according to field knowledge and conversation history related to the task.
The chatting conversation is mostly used in the chatting robot, such as microsoft ice, the conversation has various topics, the system plays a role closer to a human, and the system can give a proper reply to any text input by a user.
In practical applications, however, people not only want the robot to have intelligence and understand the intention of the conversation, but also want the robot to understand the emotion of people and provide personalized emotional communication, so that the chat-type conversation system can just meet the requirement of emotional conversation communication, can continuously converse with the user, understand the emotion of the user in the conversation, and express the emotion in the conversation.
At present, a dialog generation system mainly uses a Sequence-to-Sequence model (Sequence 2Seq), which determines generated words by using a maximum likelihood estimation method, so that words with high frequency in training data are easily selected, but general sentences lacking information quantity, such as ' good ' and ' not known by me, have high frequency, are easy to generate general answers, and the model answers are relatively single. And the model lacks the encoding and decoding of emotion and does not consider the emotional relationship between questions and answers. Natural communication with the user is difficult to achieve, and emotional ties between the user and the machine are established.
Disclosure of Invention
Aiming at the defects or shortcomings of the prior art, the invention provides a method for constructing an emotional dialogue model.
Therefore, the construction method of the emotion conversation model provided by the invention comprises the following steps:
step1, constructing an emotion dialogue corpus data set and a general reply data set, wherein the emotion dialogue corpus data set comprises a plurality of sentences and emotion category labels of the sentences, and the general reply data set comprises a plurality of general sentences;
step2, training by utilizing an emotion dialogue corpus data set and a general reply data set to generate a confrontation network, wherein the generated confrontation network comprises a generator and a discriminator, the generator comprises an encoder and a decoder, and the discriminator comprises a first discriminator and a second discriminator;
the encoder vectorizes each statement in the emotion dialogue corpus data set to obtain a semantic expression vector; the decoder generates replies of sentences according to semantic expression vectors and randomly assigned emotion category labels, the replies contain emotion information, and the objective function of the generation process is-r-log (P (Y | X, e)), r ═ a rewardedc+b*Rewarde(ii) a Log (Y | X) is the objective function of the underlying Seq2Seq dialogue model, X is the semantic representation vector, Y is the reply to X, e is the randomly assigned emotion class label, and P (Y | X) is the probability of generation based on X maximizing Y.
Pc≥P1While, Rewardc=1-Pc,Pc<P1While, Rewardc=PcWherein: 1>P1≥0.4;
Pe≥P2While, Rewarde=Pe,Pe<P2While, Rewarde=1-PeWherein: 1>P2≥0.4;
a is RewardcWeight of term, 0<a<1;
b is RewardeWeight of term, 0<b<1;
r takes any value of [0,1] initially and randomly;
the first discriminator judges and outputs the probability P that each reply belongs to the general reply according to the general reply data setc;
The second discriminator is used for judging each reply bandAnd outputs the consistency probability P of the emotion category label and the randomly designated emotion category label when the reply is generatede。
Further, the first discriminator adopts a keyword extraction method or template matching.
Optionally, the universal reply dataset includes a universal statement dataset and a non-universal statement dataset; the first discriminant data device is trained by the universal reply data set, and the trained first discriminant is used for judging and outputting the probability P that each reply belongs to the universal replyc(ii) a The first discriminator adopts a CNN network, an LSTM model, a GRU model, a Bi-LSTM model and a Bi-GRU model.
The invention provides another construction method of an emotion conversation model, which comprises the following steps:
step 1: constructing an emotion dialogue corpus data set, wherein the emotion dialogue corpus data set comprises a plurality of sentences, emotion category labels of the sentences, a plurality of general sentences and general reply category labels;
step2, training by utilizing an emotion dialogue corpus data set and a general reply data set to generate a confrontation network, wherein the generated confrontation network comprises a generator and a discriminator, the generator comprises an encoder and a decoder, and the discriminator comprises a first discriminator and a second discriminator; the first discriminator adopts a CNN network, an LSTM model, a GRU model, a Bi-LSTM model and a Bi-GRU model;
the encoder vectorizes each statement in the emotion dialogue corpus data set to obtain a semantic expression vector; the decoder generates replies of sentences according to semantic expression vectors and randomly assigned emotion category labels, the replies contain emotion information, and the objective function of the generation process is-r-log (P (Y | X, e)), r ═ a rewardedc+b*Rewarde(ii) a Log (Y | X) is the objective function of the underlying Seq2Seq dialogue model, X is the semantic representation vector, Y is the reply to X, e is the randomly assigned emotion class label, and P (Y | X) is the probability of generation based on X maximizing Y.
Pc≥P1While, Rewardc=1-Pc,Pc<P1While, Rewardc=PcWherein: 1>P1≥0.4;
Pe≥P2While, Rewarde=Pe,Pe<P2While, Rewarde=1-PeWherein: 1>P2≥0.4;
a is RewardcWeight of term, 0<a<1;
b is RewardeWeight of term, 0<b<1;
r takes any value of [0,1] initially and randomly;
the first discriminant data device is trained by the emotion dialogue corpus data set, and the trained first discriminant is adopted to judge and output the probability P that each reply belongs to a universal replyc;
The second discriminator is used for judging the emotion category label to which the emotion information carried by each reply belongs and outputting the consistency probability P of the emotion category label to which the emotion information belongs and the emotion category label randomly designated when the reply is generatede。
Optionally, the second discriminator adopts a CNN network, an LSTM model, a GRU model, a Bi-LSTM model, and a Bi-GRU model.
The invention also provides an emotional dialogue method. The provided dialogue method adopts the emotion dialogue model trained by the method to carry out emotion dialogue.
The invention also provides an emotional dialogue system, which comprises an input module, a dialogue reply generation module and an output module;
the input module is used for inputting conversation content;
the dialogue reply generation module adopts the emotion dialogue model trained by the method and is used for generating the reply of the dialogue content;
the output module is used for outputting the reply content.
The beneficial technical effects of the invention are as follows:
in the man-machine conversation process, the method can generate conversation replies meeting specific emotions, so that the reply diversity is realized, the emotion consistency is also met, the quality of the machine generated replies can be improved, and the user experience in man-machine interaction can be improved.
And (II) in the process of generating the reply, feedback information in the conversation is emphasized, and the current conversation generation model tends to generate a safe reply and contains a low amount of information.
Drawings
The present invention will be explained in further detail with reference to examples.
FIG. 1 is a schematic diagram of the working principle of the model of the present invention;
FIG. 2 shows the emotional resonance of the input and the reply of the model of the present invention in the test results.
Detailed Description
Unless otherwise indicated, the terminology herein is to be understood in light of the knowledge of one of ordinary skill in the relevant art.
The source Conversation data set of the invention can be obtained from crawlers in the network or adopt public data sets, for example, the existing data sets, such as 'electronic Conversation Generation Task 4', are obtained from customer service conversations of platforms such as Taobao, microblog and American groups.
And preprocessing the dialogue data set to obtain a dialogue corpus data set or an emotion dialogue corpus data set. The preprocessing includes deleting the content, such as: deleting redundant punctuations and symbols (such as @, emoticons, repeated symbols and the like) in the conversation data set, and then deleting sentences with unreasonable length, such as sentences with the number of characters less than 3 or more than 25 in the data through length analysis, and in some cases, deleting dialect expressions in different places, such as cantonese; and obtaining a dialogue corpus data set or an emotion dialogue corpus data set after the preprocessing is finished. Preprocessing the existing data set to obtain an emotion dialogue corpus data set; and preprocessing the original data directly obtained from the network, and then carrying out emotion classification to determine the emotion category of each statement so as to obtain an emotion dialogue corpus data set. The emotion classification can be performed manually or by using emotion classification methods (such as emotion classification methods based on keyword matching) or emotion classification models (such as BilSTM models).
The emotion dialogue corpus data set is composed of a plurality of sentences and emotion category labels of the sentences, wherein the emotion category labels are determined according to emotion types of people, for example, the currently recognized emotion types in the field include: happy, angry, sad, liked, disliked, fear, surprised, and others. Examples of sentences and their emotion category labels are shown in table 1.
TABLE 1 Emotion Categories
Input device | Emotion categories | Label (R) |
A tree is arranged in a courtyard | Others | 0 |
Poor alcohol of alcohol | Sadness and sorrow | 1 |
How to say that | Aversion to | 2 |
I are angry | Anger and anger | 3 |
Haha, i see | Happy music | 4 |
Feeling of love | Xi Huan | 5 |
She frightened like nailing on the ground | Fear of | 6 |
It is too beautiful | Is surprised | 7 |
The universal reply data set construction of the invention can adopt different technical ideas, one is that the universal reply data set is directly formed by a universal statement set, and specifically can be formed by short conversation statements used by users in network data, namely daily common statements, for example, statements with the length less than three words in the data and/or statements which continuously and repeatedly appear for more than three times. For example: "to-gether", "kay", "good, yes", "yes, yes", etc. And secondly, a data set formed by non-universal sentences is constructed or added on the basis of the constructed universal sentence set and is used for training the first generator. And thirdly, the emotion dialogue corpus data set contains a universal sentence, and a universal reply category label is added while the emotion dialogue corpus data set is constructed, wherein the universal reply category label comprises that the sentence is a universal reply (for example, represented by '1') and the sentence does not belong to the universal reply (for example, represented by '0').
The generation countermeasure network of the invention is composed of a generator and two discriminators, wherein the generator can adopt a recurrent neural network, such as an LSTM model or a Seq2Seq model; the first and second discriminators can be trained by the data set of the invention or trained CNN model, LSTM model, GRU model, Bi-LSTM model, Bi-GRU model, the training can be based on the emotion dialogue corpus data set of the invention, or trained models in the prior art.
The following embodiments of the present invention are provided, and it should be noted that the present invention is not limited to the following embodiments, and all equivalent changes based on the technical solutions of the present invention are within the protection scope of the present invention.
Example (b):
the embodiment provides an emotional dialogue model construction method provided by the invention. The emotion Conversation data set used in this embodiment is "Emotional conversion Generation Task 4", as shown in Table 2; in the operation of this embodiment, a general reply category label is added to each statement in the "electronic conversion Generation Task 4" dataset, specifically, the general reply category label of a statement in the dataset with a length of less than three words and a statement that appears three times or more continuously is "1", and the other statements are "0";
TABLE 2 Emotion dialogue data set
The generator used in the Generation countermeasure network of this embodiment is based on the seq2seq framework, wherein both the encoder and the decoder use LSTM networks, and the first discriminator uses CNN networks trained by each statement in the "electronic conversion Generation Task 4" dataset and its generic reply category label; the second discriminator adopts a Bi-LSTM model (model disclosed in Zhou H, Huang M, Zhang T, et al. electronic changing Machine: electronic conversion Generation with Internal and External Memory [ J ]. 2017.); the associated parameter settings were vocabulary set to 40000, word embedding size set to 100, all parameters of the model were initialized randomly, batch size set to 128, and initial learning rate was 0.001 using Adam optimizer.
As shown in fig. 1, the generator embeds emotion information during decoding, the first discriminator distinguishes whether the generated reply belongs to the probability of a general reply, the second discriminator distinguishes the emotion generating the reply and the consistency of the randomly assigned emotion types, and the feedback obtained by the two discriminators is returned to the rewrite model to guide the generation of an emotion text, and the specific process is as follows:
step1, converting any statement into vector form by adopting word2vec model, namely, the statement represents vector X ═ { X ═ X1,x2,…xi,…xmM is the length of the statement, xiA word vector corresponding to the ith element in the X;
step2, the encoder will vector X1={x1,x2,…,xmConverting into a semantic expression vector;
step3, randomly assigning emotion category labels, the decoder generates replies (shown in table 3, for example) of each sentence according to each semantic expression vector and the randomly assigned emotion category labels, specifically, one-hot encoding is adopted for the randomly assigned emotion category labels (for example, "like" can be encoded as [0,1,0,0, 0)]) Obtaining the emotion vector of the sentence, embedding the emotion vector into a decoding process, and generating a reply of a corresponding sentence according to the semantic vector representation and the embedded emotion vector; the objective function of the generation process is-r · log (P (Y | X, e)); in this example P1=0.5,P2=0.5,a=0.5b=0.5;
TABLE 3 Emotion dialog example
The first discriminator firstly carries out convolution operation to obtain a feature matrix of each reply, then carries out maximum pooling through a pooling layer, and then outputs the probability P that each reply belongs to a universal reply through the processing of a full connection layerc(ii) a Example (c):
TABLE 4 first discriminator working example
Input device | Recovery | Probability Pc |
The rainbow is good | Is | 0.91 |
The second discriminator judges the emotion category label of the emotion information carried by each reply and outputs the consistency probability P of the emotion category label and the randomly designated emotion category label when the reply is generatede. Example (c):
TABLE 5 second discriminator working example
And (3) simulation comparison:
simulation input: randomly selecting 1000 data in an "electronic conversion Generation Task 4" dataset;
1000 pieces of data were inputted into the Model trained in the above-mentioned examples, the Seq2Seq Model (Vinyals O, Le Q.A Neural switching Model [ J ]. Computer, 2015.), the ECM (Zhou H, Huang M, Zhang T, et al. electronic changing Machine: electronic switching Generation with Internal and External Memory [ J ].2017.), respectively, the output results of each Model were compared and evaluated,
the effectiveness of the emotion-based dialog generation method provided by the invention is evaluated, and the method mainly comprises the following two aspects: diversity and emotional consistency of the generated replies. Where emotional reasonableness is measured by emotional scores and semantic aspects are reflected by diversity and confusion. Meanwhile, manual evaluation is carried out, so that the annotator scores the responses generated by each sentence in the aspect of emotion and the aspect of semantics, and the evaluation rule is shown in table 6.
TABLE 6 rules for human evaluation
The comparative experiment was set up as Seq2Seq model and ECM model: the Seq2Seq method uses an encoder and decoder to generate a sentence, which has been considered the basic model of dialog generation, which the present invention compares with the diversity of generating reply text. The ECM model introduces emotion information in large-scale session generation for the first time, and generates emotion reply by embedding emotion vectors and an internal and external memory storage mechanism. The invention compares it with the emotional intensity and emotional accuracy of the reply. Experimental comparison results are given in table 7 and table 8, respectively.
Table 7 automatic evaluation of experimental results
Perplexity, namely, the confusion degree, whether the model generation reply is smooth in content or not is evaluated, the lower the confusion degree is, the larger the sentence probability is, the better the model effect is, and the concrete reference can be made to: li J, Galley M, Brockett C, et al.A Persona-Based Neural conversion Model [ J ]. 2016;
dist-1 is obtained by dividing the number of different unary groups in the generated reply by the total number of the unary groups in the generated reply, Dist-2 is obtained by dividing the number of different binary groups in the generated reply by the total number of the binary groups in the generated reply, and the larger Dist-1 and Dist-2 are, the more different information is contained in the generated reply; see in particular: xing C, Wu W, Wu Y, et al.Topic Aware Neural Response Generation [ J ]. 2016;
the Accuracy rate that the emotion type of the generated reply is consistent with the randomly designated emotion type is represented by the Accuracy, and the larger the Accuracy is, the reply of the randomly designated emotion type can be generated by the description model; see in particular: study of the see xiwson controllable chat conversation system 2019.
As shown in Table 7, the Seq2Seq model is easy to generate "general reply" with very low emotional intensity, since it produces the same response for different emotional categories; the ECM adopts the Beam Search algorithm to avoid safe and universal reply to a certain extent, and a simple emotion embedding model is used, so that the emotion factors in the reply are weak; after the invention is added with a reward mechanism, various sentences containing abundant information can be generated more easily, and the emotion is richer.
TABLE 8 results of the Manual evaluation experiment
Since "others" do not contain a specific emotion category, they are not scored. From the analysis in table 8, in the responses of the five emotion categories, compared with the Seq2Seq model, after the emotion constraint is added to the model of the present invention, the content and emotion are greatly improved, and the content score is higher than Seq2Seq by 0.35 points and the emotion score is higher than Seq2Seq by 0.12 points in average score. Compared with the ECM model, the model provided by the invention performs better in the emotion categories of like and happy, and the content score is 0.12 points higher than the ECM and the emotion score is 0.03 points higher than the ECM in average score. It can also be seen that the emotional scores generated by the three models for replies in the "aversion" and "anger" emotion categories are low, and the analysis may be due to the relatively small corpus of dialogue training in both emotion categories, resulting in the selection of the generated sentence ignoring the "aversion" and "anger" constraints and optimizing the deviation of replies. In general, compared with two baseline models, the model provided by the invention has good improvement in the reply of five emotion categories, both in content and emotion.
The emotional change between the input (1000 pieces of data) and the reply is observed, noting that the input and reply typically carry the same or similar emotional tags. For example, an input tagged with a "like" tag rarely returns "angry", while each type of emotion interacts emotionally with the rest, indicating that the degree of emotional resonance is not only related to emotional similarity, but also affected by emotional complementarity. The emotional resonance degree of the pleased emotion and the happy emotion is not equal to the resonance degree of the happy emotion and the pleased emotion, and the fact that the interactive emotional resonance degree has directionality is shown. As shown in fig. 2, is not completely symmetrical about the main diagonal. (in the embodiment, the evaluation method of the emotional resonance degree can be specifically referred to Liuning, human-computer interaction emotional anthropomorphic strategy research [ D ] 2020.).
Claims (7)
1. A construction method of an emotional dialogue model is characterized by comprising the following steps:
step1, constructing an emotion dialogue corpus data set and a general reply data set, wherein the emotion dialogue corpus data set comprises a plurality of sentences and emotion category labels of the sentences, and the general reply data set comprises a plurality of general sentences;
step2, training by utilizing an emotion dialogue corpus data set and a general reply data set to generate a confrontation network, wherein the generated confrontation network comprises a generator and a discriminator, the generator comprises an encoder and a decoder, and the discriminator comprises a first discriminator and a second discriminator;
the encoder vectorizes each statement in the emotion dialogue corpus data set to obtain a semantic expression vector; the decoder generates replies of sentences according to semantic expression vectors and randomly assigned emotion category labels, the replies contain emotion information, and the objective function of the generation process is-r-log (P (Y | X, e)), r ═ a rewardedc+b*Rewarde(ii) a Log (Y | X) is the objective function of the underlying Seq2Seq dialogue model, X is the semantic representation vector, Y is the reply to X, e is the randomly assigned emotion class label, and P (Y | X) is the probability of generation based on X maximizing Y.
Pc≥P1While, Rewardc=1-Pc,Pc<P1While, Rewardc=PcWherein: 1>P1≥0.4;
Pe≥P2While, Rewarde=Pe,Pe<P2While, Rewarde=1-PeWherein: 1>P2≥0.4;
a is RewardcWeight of term, 0<a<1;
b is RewardeWeight of term, 0<b<1;
r takes any value of [0,1] initially and randomly;
the first discriminator judges and outputs the probability P that each reply belongs to the general reply according to the general reply data setc;
The second discriminator is used for judging the emotion category label to which the emotion information carried by each reply belongs and outputting the consistency probability P of the emotion category label to which the emotion information belongs and the emotion category label randomly designated when the reply is generatede。
2. The method of claim 1, wherein the first discriminator uses keyword extraction or template matching.
3. The method of constructing an emotional dialog model of claim 1, wherein the universal reply dataset comprises a universal sentence dataset and a non-universal sentence dataset; the first discriminant data device is trained by the universal reply data set, and the trained first discriminant is used for judging and outputting the probability P that each reply belongs to the universal replyc(ii) a The first discriminator adopts a CNN network, an LSTM model, a GRU model, a Bi-LSTM model and a Bi-GRU model.
4. A construction method of an emotional dialogue model is characterized by comprising the following steps:
step 1: constructing an emotion dialogue corpus data set, wherein the emotion dialogue corpus data set comprises a plurality of sentences, emotion category labels of the sentences, a plurality of general sentences and general reply category labels;
step2, training by utilizing an emotion dialogue corpus data set and a general reply data set to generate a confrontation network, wherein the generated confrontation network comprises a generator and a discriminator, the generator comprises an encoder and a decoder, and the discriminator comprises a first discriminator and a second discriminator; the first discriminator adopts a CNN network, an LSTM model, a GRU model, a Bi-LSTM model and a Bi-GRU model;
the encoder vectorizes each statement in the emotion dialogue corpus data set to obtain a semantic expression vector; the decoder generates replies of sentences according to semantic expression vectors and randomly assigned emotion category labels, the replies contain emotion information, and the objective function of the generation process is-r-log (P (Y | X, e)), r ═ a rewardedc+b*Rewarde(ii) a Log (Y | X) is the objective function of the underlying Seq2Seq dialogue model, X is the semantic representation vector, Y is the reply to X, e is the randomly assigned emotion class label, and P (Y | X) is the probability of generation based on X maximizing Y.
Pc≥P1While, Rewardc=1-Pc,Pc<P1While, Rewardc=PcWherein: 1>P1≥0.4;
Pe≥P2While, Rewarde=Pe,Pe<P2While, Rewarde=1-PeWherein: 1>P2≥0.4;
a is RewardcWeight of term, 0<a<1;
b is RewardeWeight of term, 0<b<1;
r takes any value of [0,1] initially and randomly;
the first discriminant data device is trained by the emotion dialogue corpus data set, and the trained first discriminant is adopted to judge and output the probability P that each reply belongs to a universal replyc;
The second discriminator is used for judging the emotion category label to which the emotion information carried by each reply belongs and outputting the consistency probability P of the emotion category label to which the emotion information belongs and the emotion category label randomly designated when the reply is generatede。
5. The method for constructing an emotional dialogue model according to claim 1 or 4, wherein the second discriminator employs a CNN network, an LSTM model, a GRU model, a Bi-LSTM model, and a Bi-GRU model.
6. An emotional dialogue method, characterized in that an emotional dialogue is performed by using an emotional dialogue model trained by the method of any one of claims 1 to 5.
7. The emotion conversation system is characterized by comprising an input module, a conversation reply generation module and an output module;
the input module is used for inputting conversation content;
the dialogue reply generation module adopts an emotional dialogue model trained by the method of any one of claims 1 to 5 to generate the reply of the dialogue content;
the output module is used for outputting the reply content.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110283821.0A CN112905776B (en) | 2021-03-17 | 2021-03-17 | Emotional dialogue model construction method, emotional dialogue system and method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110283821.0A CN112905776B (en) | 2021-03-17 | 2021-03-17 | Emotional dialogue model construction method, emotional dialogue system and method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112905776A true CN112905776A (en) | 2021-06-04 |
CN112905776B CN112905776B (en) | 2023-03-31 |
Family
ID=76105293
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110283821.0A Active CN112905776B (en) | 2021-03-17 | 2021-03-17 | Emotional dialogue model construction method, emotional dialogue system and method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112905776B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113688309A (en) * | 2021-07-23 | 2021-11-23 | 北京三快在线科技有限公司 | Training method for generating model and generation method and device for recommendation reason |
CN114187997A (en) * | 2021-11-16 | 2022-03-15 | 同济大学 | Psychological consultation chat robot implementation method for depressed people |
CN114385802A (en) * | 2022-01-10 | 2022-04-22 | 重庆邮电大学 | Common-emotion conversation generation method integrating theme prediction and emotion inference |
WO2023159759A1 (en) * | 2022-02-22 | 2023-08-31 | 平安科技(深圳)有限公司 | Model training method and apparatus, emotion message generation method and apparatus, device and medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180174020A1 (en) * | 2016-12-21 | 2018-06-21 | Microsoft Technology Licensing, Llc | Systems and methods for an emotionally intelligent chat bot |
CN108874972A (en) * | 2018-06-08 | 2018-11-23 | 青岛里奥机器人技术有限公司 | A kind of more wheel emotion dialogue methods based on deep learning |
CN111128240A (en) * | 2019-12-19 | 2020-05-08 | 浙江大学 | Speech emotion recognition method based on anti-semantic erasure |
CN111241250A (en) * | 2020-01-22 | 2020-06-05 | 中国人民大学 | Emotional dialogue generation system and method |
CN111522936A (en) * | 2020-04-24 | 2020-08-11 | 上海智臻智能网络科技股份有限公司 | Intelligent customer service dialogue reply generation method and device containing emotion and electronic equipment |
US20210034708A1 (en) * | 2019-08-01 | 2021-02-04 | Oracle International Corporation | Using neural network and score weighing to incorporate contextual data in sentiment analysis |
-
2021
- 2021-03-17 CN CN202110283821.0A patent/CN112905776B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180174020A1 (en) * | 2016-12-21 | 2018-06-21 | Microsoft Technology Licensing, Llc | Systems and methods for an emotionally intelligent chat bot |
CN108874972A (en) * | 2018-06-08 | 2018-11-23 | 青岛里奥机器人技术有限公司 | A kind of more wheel emotion dialogue methods based on deep learning |
US20210034708A1 (en) * | 2019-08-01 | 2021-02-04 | Oracle International Corporation | Using neural network and score weighing to incorporate contextual data in sentiment analysis |
CN111128240A (en) * | 2019-12-19 | 2020-05-08 | 浙江大学 | Speech emotion recognition method based on anti-semantic erasure |
CN111241250A (en) * | 2020-01-22 | 2020-06-05 | 中国人民大学 | Emotional dialogue generation system and method |
CN111522936A (en) * | 2020-04-24 | 2020-08-11 | 上海智臻智能网络科技股份有限公司 | Intelligent customer service dialogue reply generation method and device containing emotion and electronic equipment |
Non-Patent Citations (3)
Title |
---|
LANTAO YU: "《SeqGAN:sequence generative adversarial nets with policy gradient》", 《ARXIV》 * |
周震卿等: "基于TextCNN情感预测器的情感监督聊天机器人", 《微型电脑应用》 * |
王明申: "《基于词级权重与对抗性ECM模型的对话生成方法研究》", 《中国优秀硕士学位论文全文数据库信息科技辑》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113688309A (en) * | 2021-07-23 | 2021-11-23 | 北京三快在线科技有限公司 | Training method for generating model and generation method and device for recommendation reason |
CN114187997A (en) * | 2021-11-16 | 2022-03-15 | 同济大学 | Psychological consultation chat robot implementation method for depressed people |
CN114385802A (en) * | 2022-01-10 | 2022-04-22 | 重庆邮电大学 | Common-emotion conversation generation method integrating theme prediction and emotion inference |
WO2023159759A1 (en) * | 2022-02-22 | 2023-08-31 | 平安科技(深圳)有限公司 | Model training method and apparatus, emotion message generation method and apparatus, device and medium |
Also Published As
Publication number | Publication date |
---|---|
CN112905776B (en) | 2023-03-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109241255B (en) | Intention identification method based on deep learning | |
CN108897857B (en) | Chinese text subject sentence generating method facing field | |
CN112905776B (en) | Emotional dialogue model construction method, emotional dialogue system and method | |
CN110321563B (en) | Text emotion analysis method based on hybrid supervision model | |
CN107798140A (en) | A kind of conversational system construction method, semantic controlled answer method and device | |
CN109063164A (en) | A kind of intelligent answer method based on deep learning | |
CN113505209A (en) | Intelligent question-answering system for automobile field | |
CN112989033B (en) | Microblog emotion classification method based on emotion category description | |
CN111914556A (en) | Emotion guiding method and system based on emotion semantic transfer map | |
CN113297364A (en) | Natural language understanding method and device for dialog system | |
CN114911932A (en) | Heterogeneous graph structure multi-conversation person emotion analysis method based on theme semantic enhancement | |
CN111949762B (en) | Method and system for context-based emotion dialogue and storage medium | |
CN113435211A (en) | Text implicit emotion analysis method combined with external knowledge | |
CN114648016A (en) | Event argument extraction method based on event element interaction and tag semantic enhancement | |
CN112818106A (en) | Evaluation method of generating type question and answer | |
CN116049387A (en) | Short text classification method, device and medium based on graph convolution | |
CN116561251A (en) | Natural language processing method | |
TWI734085B (en) | Dialogue system using intention detection ensemble learning and method thereof | |
CN114416969A (en) | LSTM-CNN online comment sentiment classification method and system based on background enhancement | |
CN112200674B (en) | Stock market emotion index intelligent calculation information system | |
CN113326367A (en) | Task type dialogue method and system based on end-to-end text generation | |
CN113486143A (en) | User portrait generation method based on multi-level text representation and model fusion | |
CN113065324A (en) | Text generation method and device based on structured triples and anchor templates | |
CN113220964A (en) | Opinion mining method based on short text in network communication field | |
Chowanda et al. | Generative Indonesian conversation model using recurrent neural network with attention mechanism |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |