CN107766506A - A kind of more wheel dialog model construction methods based on stratification notice mechanism - Google Patents

A kind of more wheel dialog model construction methods based on stratification notice mechanism Download PDF

Info

Publication number
CN107766506A
CN107766506A CN201710986813.6A CN201710986813A CN107766506A CN 107766506 A CN107766506 A CN 107766506A CN 201710986813 A CN201710986813 A CN 201710986813A CN 107766506 A CN107766506 A CN 107766506A
Authority
CN
China
Prior art keywords
sentence
word
implicit function
stratification
encryption
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710986813.6A
Other languages
Chinese (zh)
Inventor
张伟男
汪意发
朱庆福
刘挺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Institute of Technology
Original Assignee
Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Institute of Technology filed Critical Harbin Institute of Technology
Priority to CN201710986813.6A priority Critical patent/CN107766506A/en
Publication of CN107766506A publication Critical patent/CN107766506A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems

Abstract

The present invention relates to a kind of more wheel dialog model construction methods based on stratification notice mechanism, it is to depend on large-scale corpus to solve existing interactive system, training speed is influenceed by language material scale, and due to the reply nonuniqueness of dialogue generation, Seq2Seq models are invariably prone to generate general, the shortcomings that insignificant reply, and a kind of more wheel dialog model construction methods based on stratification notice mechanism are proposed, including:Sentence inputting is received, for each sentence, encryption implicit function is calculated since first word, calculates the Attention weights of each sentence, and calculates topic linguistic context and represents vector, finally calculates decryption implicit function, while result is exported.The present invention is applied to the chat robots system of open field.

Description

A kind of more wheel dialog model construction methods based on stratification notice mechanism
Technical field
The present invention relates to interactive system, and in particular to a kind of more wheel dialog models based on stratification notice mechanism Construction method.
Background technology
1st, foreign technology present situation
(1) conversational system based on artificial template
Based on the technology of artificial template by manually setting session operational scenarios, and some are write to each scene and targetedly talked with Template, template describe the problem of user is possible and corresponding answer template.
Weizenbaum et al. (1966) develops earliest chat robots ELIZA, ELIZA may go out according in dialogue Existing language situation, goes to be pre-designed corresponding language template, and text generator can be according to the input of user by the weight in input Want information to be embedded into template, finally give reply.
Chat is restricted to special scenes or specific topic by them, and generates sound using one group of pattern rule Should.
(2) conversational system based on retrieval
Chat robots based on retrieval technique then using the method for being similar search engine, are previously stored dialogue storehouse simultaneously Establish index, according to user's question sentence, in storehouse talk with progress fuzzy matching find most suitable response content.
Shaikh et al. (2010) constructs a cyberchat robot (VCA), can enter in chatroom with people The preliminary social activity of row, they utilize ongoing conversation subject to carry out web search using a kind of novel method, and Find the related subject that may be inserted into dialogue and change its flow, can be regarded as based on retrieval and and template method melt Close.
(3) the dialogue generation model based on deep learning
Application of the depth learning technology in dialogue generates is mainly towards open field chat robots, because extensive logical It is relatively easy to the acquisition of language material, the Sequence to Sequence models of the most frequently used reference machine translation, dialogue is given birth to Into the whole process by problem to reply be considered as in machine translation translation process from original language to object language.
Ritter et al. (2011) has used the dialogue language material obtained from Twitter, to imitate using Seq2Seq models Fruit has exceeded the dialog model based on searching system.
Sordoni et al. (2015) proposes a dialog generation system, and the system take into account the context letter in dialogue Breath, so as to obtain lifting on uniformity is replied.
Serban et al. (2016) proposes Hierarchical Nerual Network models, it is intended in dialogue Semanteme and interaction hereafter is modeled, so as to build the conversational system taken turns one more.
Jiwei Li et al. (2016) are directed to solving the problems, such as traditional Seq2Seq models generation common replies, introduce Mutual information improves the diversity that generation is replied as object function.Jiwei Li (2016) use improved Seq2Seq simultaneously Model is modeled to user styles, user embedding is introduced as priori in decoding end, so as to improve dialogue system The uniformity and correlation of system.
Louis Shao et al. (2017) improve training method and the decode ends of Seq2Seq models, and add Beam-search, so as to improve the reply length and uniformity and correlation of model generation.
2nd, domestic technique present situation
The country is because start late, the method that the research in terms of conversational system is also mainly based upon deep learning, Li Hang et al. (2015) proposes Neural Responding Machine, using improved Seq2Seq models, adds Attention mechanism simultaneously carries out fusion so as to achieve good result in short text conversational system using multiple models.
Mou Lili (2016), which are conceived to, solves the problems, such as traditional Seq2Seq models generation common replies, proposes first Seq2Bf models, keyword is first predicted by using mutual information, then based on keyword reply the generation of sentence.
Zongcheng Ji (2014) then use the method based on retrieval simultaneously, using state-of-the-art information retrieval technique, By huge dialogue corpus, a relatively intelligent conversational system is created.
3rd, the brief analysis of domestic and foreign literature summary
The research both at home and abroad for open field dialog generation system mainly includes the method based on template at present, based on retrieval Method and method based on deep learning.The method based on template during early stage does not carry out real Language Processing, raw Into language it is stiff, formalization, the problem of often existing in terms of semantic and fluency, relative to open field conversational system, the party Method is more suitable for Task chat robots.
Method based on retrieval be it is existing everybody talk with corpus by sorting learning art and depth matches skill Art finds the optimal reply for being adapted to currently input.The limitation of this method is to be only capable of being replied with fixed language mode, nothing Method realizes the diversity combination of word.
Current most popular method is the method based on deep learning, uses the Seq2Seq moulds from machine translation task Type, usually Encoder-Decoder structures, coordinate fairly large dialogue language material, trained end to end for realizing, so as to Obtain a conversational system.This method can break through before limitation of the method for clause word, mainly user is inputted The problem of be modeled, the generation of word for word (word) is then carried out according to intermediate result, can creatively generate replys, it is exhausted at present Most of researchs are all based on the expansion or improvement of the model.
But the method based on deep learning depends on large-scale corpus, Seq2Seq model training speed is advised by language material The influence of mould, and due to the reply nonuniqueness of dialogue generation, Seq2Seq models are invariably prone to generate general, it is not intended to justice Reply, such as " hello ", " I does not also know ", " heartily " etc..
Furthermore current most of conversational systems are devoted to optimize the single-wheel dialogue i.e. quality of question-response process, and chat is One continuous interactive process for having specific background, the meaning of a word will combine the background ability of the context of dialogue or correlation sometimes It can determine that.The problem of needing to be studied is still for context modeling.
The content of the invention
The invention aims to solve existing interactive system to depend on large-scale corpus, training speed by The influence of language material scale, and due to the reply nonuniqueness of dialogue generation, Seq2Seq models be invariably prone to generate it is general, The shortcomings that insignificant reply, and a kind of more wheel dialog model construction methods based on stratification notice mechanism are proposed, including:
Step 1: receive n+1 sentence inputting co,c1,…cn
Step 2: it is directed to each sentence ci, encryption implicit function h is calculated since first wordi,t=f (xi,t, hi,t-1), wherein wherein xi,tRepresent ciT-th of word;Wherein hi,0It is designated as parameter preset;And finish last calculating hi,tIt is designated as sentence ciEncryption implicit function hi
Step 3: calculate the Attention weights of i-th of sentenceWherein ei=vTtanh(Whi+ Uhn);V, W, U are the parameter preset in Attention mechanism;Tanh is activation primitive;
Step 4: calculating topic linguistic context represents vector T=∑ αihi
Step 5: calculate decryption implicit function st=f (yt-1,st-1, T), yt-1Represent the iteration input quantity at t-1 moment, y0For Preset value;s0=hn
Step 6: by s1,s2,……st……snValue exported as a result.
The present invention also provides another more wheel dialog model construction methods based on stratification notice mechanism, including:
Step 1: receive n+1 sentence inputting co,c1,…cn
Step 2: it is directed to each sentence ci, encryption implicit function h is calculated since first wordi,t=f (xi,t, hi,t-1), wherein wherein xi,tRepresent ciT-th of word;Wherein hi,0It is designated as parameter preset;And finish last calculating hi,tIt is designated as sentence ciEncryption implicit function hi
Step 3: calculate the Attention weights of t-th of word in i-th of sentenceWherein eit= vTtanh(Whi+Ust-1);V, W, U are the parameter preset in Attention mechanism;st-1For the hidden layer state at t-1 moment;
Step 4: calculate dynamic representation vector Dtithi
Step 5: calculate decryption implicit function st=f (yt-1,st-1, T), yt-1Represent the iteration input quantity at t-1 moment, y0For Preset value;s0=hn
Step 6: by s1,s2,……st……snValue exported as a result.
Beneficial effects of the present invention are:
1st, do not influenceed independent of large-scale corpus, training speed by language material scale, do not tend to generate it is general, Insignificant reply;
2nd, the present invention is tested on Opensubtitles data sets and Ubuntu data sets.
On Opensubtitles data sets, Embedding Average of the invention can reach 0.565647, Apparently higher than the 0.557139 of prior art;The Greedy Matching of the present invention can reach 0.523235, hence it is evident that be higher than The 0.503273 of prior art;The Extrema of the present invention can reach 0.393724, higher than the 0.393189 of prior art.
On Ubuntu data sets, Embedding Average of the invention can reach 0.612089, hence it is evident that be higher than The 0.577022 of prior art, Greedy Matching of the invention can reach 0.429328, hence it is evident that higher than prior art 0.416948, Extrema of the invention can reach 0.397543, higher than the 0.391392 of prior art.
Brief description of the drawings
Fig. 1 is the flow chart of more wheel dialog model construction methods based on stratification notice mechanism of the present invention;
Fig. 2 is the schematic diagram that encryption implicit function process is calculated in the step 2 of the specific embodiment of the invention one;Wherein Context refers to sentence inputting co,c1,…cnThe context of composition;
Fig. 3 is schematic diagram of the step 3 of the specific embodiment of the invention one to step 5;Wherein Topic Net represent to calculate Topic linguistic context represents the computing module of vector T, and topic vector represent that topic linguistic context represents vector;Decoder, which is used to calculate, to be solved Close implicit function;By hnThe arrow for pointing to decoder represents that the calculating of decryption implicit function needs to use hnValue.
Embodiment
Embodiment one:
The present invention provides a kind of more wheel dialog model construction methods based on stratification notice mechanism, as shown in figure 1, bag Include:
Step 1: receive n+1 sentence inputting co,c1,…cn
Step 2: it is directed to each sentence ci, encryption implicit function h is calculated since first wordi,t=f (xi,t, hi,t-1), wherein wherein xi,tRepresent ciT-th of word;Wherein hi,0It is designated as parameter preset;And finish last calculating hi,tIt is designated as sentence ciEncryption implicit function hi
Step 3: calculate the Attention weights of i-th of sentenceWherein ei=vTtanh(Whi+ Uhn);V, W, U are the parameter preset in Attention mechanism.
Step 4: calculating topic linguistic context represents vector T=∑ αihi
Step 5: calculate decryption implicit function st=f (yt-1,st-1, T), yt-1The iteration input quantity at t-1 moment is represented,0To be pre- If value;s0=hn;When the sentence inputting received in step 1 is training data, yt-1For default model answer word, work as step When the sentence inputting received in rapid one is measured data, yt-1Value be equal to st-1
Step 6: by stValue exported as a result.
Specifically, mode of the invention is the Seq2Seq models based on machine translation task, and using Encoder-Decoder structures, it is the characteristics of this model and structure, for a sentence ci, each list that will wherein include Word xi,tModel is sequentially inputted, model is for each xi,tCorresponding encryption implicit function (as shown in Figure 2) is calculated according to formula, its Result of calculation is used to calculate next encryption implicit function, and by that analogy, the value of last encryption implicit function output is as whole The encryption implicit function h of sentencei, due to have input n+1 sentence altogether, therefore can also obtain n+1 encryption implicit function.Then root Attention weights are calculated according to these encryption implicit functions, Attention mechanism is also conventional mechanism, and the present invention is to its formula Form is modified, the improvement to formula is mainly when calculating the weight of each, it is contemplated that the encryption of current sentence Implicit function (the i.e. formula e of last in implicit function and all sentencesi=vTtanh(Whi+Uhn) in hiRepresent current sentence Encryption implicit function, hnRepresent last encryption implicit function), vector T then is represented further according to weight calculation topic linguistic context, The vector contains the necessary information at encryption end, and this vector is decrypted last decrypting end (is used for the function i.e. s decryptedt =f (yt-1,st-1, T)) result is obtained, as shown in Figure 3.During actual treatment data, yt-1Be output result, stIt is exactly yt, i.e. t The word of moment output, and in the training process, in order to ensure the effect of training, stIt is actual output, and yt-1It is to preset Word, i.e. word in model answer.
It can be seen that the method for the present invention is based not only on current problem and replied, current context and up and down also can take into account Literary information, this process can be considered as to Memory Process, people read information from memory, with reference to current problem, so as to provide Reply.This item work is handled whole context using memory network (Memory Network), remembers net end to end Network can be used for QA tasks or read understanding, and different tasks is completed with this to obtain the expression of document by modeling.
Last sentence in context is considered as key (i.e. formula e by this item worki=vTtanh(Whi+Uhn) in hn), its Remaining sentence is considered as memory, and representing for whole context is calculated with this, encryption end of the part as Seq2Seq models, as a result Decrypting end is input to, is replied for decoding.The timing information of sentence in memory is considered that simultaneously, assigns each sentence Different weights, represented with this to obtain more preferable context.
Embodiment two:
The present invention also provides another more wheel dialog model construction methods based on stratification notice mechanism, including:
Step 1: receive n+1 sentence inputting co,c1,…cn
Step 2: it is directed to each sentence ci, encryption implicit function h is calculated since first wordi,t=f (xi,t, hi,t-1), wherein wherein xi,tRepresent ciT-th of word;Wherein hi,0It is designated as parameter preset;And finish last calculating hi,tIt is designated as sentence ciEncryption implicit function hi
Step 3: calculate the Attention weights of t-th of word in i-th of sentenceWherein eit= vTtanh(Whi+Ust-1);V, W, U are the parameter preset in Attention mechanism;st-1For the hidden layer state at t-1 moment.
Step 4: calculate dynamic representation vector Dtithi
Step 5: calculate decryption implicit function st=f (yt-1,st-1,Dt), yt-1Represent the iteration input quantity at t-1 moment, y0For Preset value;s0=hn
Step 6: by stValue exported as a result.
The difference of present embodiment and embodiment is, the side of Attention weights is calculated in step 3 Formula is different, and present embodiment is to calculate weight for each each word, and embodiment one be only for Whole sentence calculates a weight.Another place's difference is that last encryption implicit function h is not used when weight specifically calculatesn, and It is the hidden layer state s using last momentt-1.Embodiment one and embodiment two are all based on " considering on entire chapter Hereafter, and from key values weight is calculated " this invention thinking, the key values only chosen one are to select last to add Close implicit function, another is the hidden layer state for selecting decrypting end.
In the present invention, the method for deep learning has been used to be modeled contextual information, so as to optimize dialogue matter Amount, improve correlation and uniformity that generation is replied.The main models that this item patent uses are Seq2Seq models, and structure is Encoder-Decoder structures (encryption end-decryption end structure), because final goal is that decoding generation one is semantic smooth, one The reply of cause property good relationship, Decoder (decrypting end) must be a preferable language model, so Decoder is Realization based on RNN.
Traditional Seq2Seq models simply consider the problem of single-wheel dialogue and reply, but not only nationwide examination for graduation qualification when people engage in the dialogue Dialogue is current, and what is said or talked about, also can take into account current context and contextual information, but in more wheel dialogues, current sentence is namely Distance currently replys nearest sentence and is considered as most important sentence on time, because it is for the direct of the sentence that generation, which is replied, Reply.In order to obtain the information of whole context, so as to be modeled to linguistic context topic, institute of this item work using RNN to context There is sentence to be modeled, so as to obtain multiple expressions, then referring to the notice mechanism in machine translation, layer is carried out to context The notice of secondaryization calculates, and show that the attented of context is represented, and in this, as the expression of topic linguistic context, be added to decoding Hold and be used for auxiliary decoder, preferably replied so as to generate uniformity correlation.
Beneficial effects of the present invention are specifically described below:
The method main flow evaluated at present the chat robots under open field environment has two kinds of thinkings, objective indicator evaluation with Subjective artificial scoring.The objective indicator part that this item work uses is mainly with Embedding Average, Greedy Matching, Vector Extrema are the Evaluations matrix based on term vector of representative.
The general principle of above-mentioned objective indicator is all that the candidate for calculating generation is replied between the target retro of Key for Reference Similarity, in this, as reply quality scoring, basic skills be by understand each word the meaning come judge reply Correlation, term vector are the bases for realizing this evaluation method.According to semantic distribution, distribute a vector to each word and be used for Represent this word, then by various methods can respectively obtain candidate reply with the sentence of target retro vector, then by cosine away from From being compared, it is possible to obtain the similarity of the two.
Opensubtitles is the well-known captions website in a world, there is provided the captions of various countries' language are downloaded, and can therefrom be obtained Obtain the English dialogue data collection being made up of film dialogue;Ubuntu data sets are that the chat record sticked is discussed in Ubuntu forums The English dialogue data collection of composition.This patent work obtains 100000 Opensubtitles session operational scenarios, is processed into The dialogue session of wheel more than 800000, on this basis using word2vec pre-training term vectors, has obtained 30000+ word Table;The dialogue session of wheel more than about 450000 has then been obtained on Ubuntu data sets, has been obtained using word2vec pre-training term vectors 30000+ vocabulary is arrived.
At present topic model (method of embodiment one) and dynamic model (embodiment two Method) tested on Opensubtitles data sets and Ubuntu data sets, by handling former dialogue language material After context, dialogue session form is processed into, carries out experiment test.
This item patent work and three LSTM, HRED model, VHRED models baseline models are in Embedding Compared in Average, Greedy Matching and Vector Extrema indexs.Test result indicates that this item is special The topic model of profit work performances of three indexs on Opensubtitles data sets and Ubuntu data sets are better than baseline;And dynamic model have also exceeded baseline performance in the first two index.It the results are shown in Table 1 and table 2.
Table 1:Result on Opensubtitles data sets, lstm, hred, vhred are baseline results, topic Model, dynamic model are the method applied for a patent
Table 2:Result on Ubuntu data sets, lstm, hred, vhred are baseline results, topic model, Dynamic model are the method applied for a patent
The effect of the present invention can be more intuitively found out by following 4 groups of test cases.
The method that wherein topic mode represent embodiment one, dyna model represent embodiment two Method, hred model, vhred model and LSTM model are existing models.
Example 1:
Above:
I will attempt the exploitation version using PPA now, look at whether it can collapse again.
You are seeing the cpu temperature of computer
No, the problem of I is not run into also in terms of temperature ... I where can see its temperature
topic model:You can attempt to delete your configuration file, then try again
dyna model:Attempt from terminal operating lspci, there is individual list there
hred model:System->Preference->Power management
vhred model:I does not understand your meaning
LSTM model:I does not understand
Example 2:
Above:
Come in.
It thanks.
It is not so polite, it is in own home just to work as
Miss Di La etc. be able to can go out.
Alright
Have fun
topic model:Had fun I believe that can do by myself
dyna model:Thank you
hred model:Come on
vhred model:Alright, we walk
lstm model:I understands careful
Example 3:
Above:
According to the method in link, nomodeset parameters are added in kernel into startup item
In additionIt thanks.
topic model:You can also use nomodeset options editor's configuration text under/etc/default/grub Part
dyna model:Instruct identical with last entry of last time renewal
hred model:It is unfriendly
vhred model:It is unfriendly
LSTM model:I does not understand
Example 4:
Above:
Who knows another problem.Pidgin is allowed to show messages menu either with or without method
It please check down whether you really use NVIDIA driver.Or reset driving ensure to have no problem
topic model:I has a nvidia video card, and a video card of Radeon Mobility 7000.
dyna model:I has one 75 feet of best display.
hred model:How I this do
vhred model:I knows
lstm model:I does not know
Vhred model and lstm model are can be seen that from above-mentioned 4 examples to be more likely to generate general reply, It is difficult to the context generation for complexity meets the reply of linguistic context.Hred model generate general reply sometimes, sometimes Generation meets the reply of linguistic context, but generally speaking, the situation for not meeting linguistic context is more.And the present invention topic model and Dyna model can generate the reply content for meeting linguistic context.
The embodiment one, two of the present invention may be directly applied in the chat robots system of open field.This is specially The chat robots that the application vector of sharp technology is Harbin Institute of Technology's social computing with Research into information retrieval center is developed are " stupid It is stupid ".
More wheel dialogue functions provided by the invention can be as a module of whole chat robots system:The module connects By the enabling signal from middle control module, independently it is responsible for once the progress of more wheel dialogues, when taking turns end-of-dialogue, by control flow more Give back middle control module.
From deployment way, this technology can independently be used as a calculate node, be deployed in Ali's cloud or U.S. On the cloud computing platforms such as group's cloud, the communication between other modules can be carried out by way of binding IP address and port numbers.
In the specific implementation of the technology, because having used deep learning correlation technique, corresponding depth is needed to use Learning framework:The all alternatively frameworks of Tensorflow and Pytorch.No matter any deep learning framework is used, not The external interface of the technology modules can be influenceed.
The present invention can also have other various embodiments, in the case of without departing substantially from spirit of the invention and its essence, this area Technical staff works as can make various corresponding changes and deformation according to the present invention, but these corresponding changes and deformation should all belong to The protection domain of appended claims of the invention.

Claims (4)

1. a kind of more wheel dialog model construction methods based on stratification notice mechanism, including:
Step 1: receive n+1 sentence inputting co, c1... cn
Step 2: it is directed to each sentence ci, encryption implicit function h is calculated since first wordI, t=f (xI, t, hI, t-1), its In wherein xI, tRepresent ciT-th of word;Wherein hI, 0For parameter preset;And the h for finishing last calculatingI, tAs sentence ci Encryption implicit function hi
Step 3: calculate the Attention weights of i-th of sentenceWherein ei=vTtanh(Whi+Uhn);v、 W, U is the parameter preset in Attention mechanism;Tanh is activation primitive;
Step 4: calculating topic linguistic context represents vector T=∑ αihi
Step 5: calculate decryption implicit function st=f (yt-1, st-1, T), yt-1Represent the iteration input quantity at t-1 moment, y0It is default Value;s0=hn
Step 6: by s1, s2... snValue exported as a result.
2. more wheel dialog model construction methods according to claim 1 based on stratification notice mechanism, its feature exist In, when the sentence inputting received in step 1 is training data, y in step 5t-1For default model answer word, work as step When the sentence inputting received in rapid one is measured data, y in step 5t-1Value be equal to st-1
3. a kind of more wheel dialog model construction methods based on stratification notice mechanism, including:
Step 1: receive n+1 sentence inputting co, c1... cn
Step 2: it is directed to each sentence ci, encryption implicit function h is calculated since first wordI, t=f (xI, t, hI, t-1), its In wherein xI, tRepresent ciT-th of word;Wherein hI, 0For parameter preset;And the h for finishing last calculatingI, tAs sentence ci Encryption implicit function hi
Step 3: calculate the Attention weights of t-th of word in i-th of sentenceWherein eit= vTtanh(Whi+Ust-1);V, W, U are the parameter preset in Attention mechanism;st-1For the hidden layer state at t-1 moment; Tanh is activation primitive;
Step 4: calculate dynamic representation vector Dtithi
Step 5: calculate decryption implicit function st=f (yt-1, st-1, Dt), yt-1Represent the iteration input quantity at t-1 moment, y0It is default Value;s0=hn
Step 6: by s1, s2... snValue exported as a result.
4. more wheel dialog model construction methods according to claim 3 based on stratification notice mechanism, its feature exist In, when the sentence inputting received in step 1 is training data, y in step 5t-1For default model answer word, work as step When the sentence inputting received in rapid one is measured data, y in step 5t-1Value be equal to st-1
CN201710986813.6A 2017-10-20 2017-10-20 A kind of more wheel dialog model construction methods based on stratification notice mechanism Pending CN107766506A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710986813.6A CN107766506A (en) 2017-10-20 2017-10-20 A kind of more wheel dialog model construction methods based on stratification notice mechanism

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710986813.6A CN107766506A (en) 2017-10-20 2017-10-20 A kind of more wheel dialog model construction methods based on stratification notice mechanism

Publications (1)

Publication Number Publication Date
CN107766506A true CN107766506A (en) 2018-03-06

Family

ID=61268560

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710986813.6A Pending CN107766506A (en) 2017-10-20 2017-10-20 A kind of more wheel dialog model construction methods based on stratification notice mechanism

Country Status (1)

Country Link
CN (1) CN107766506A (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108388944A (en) * 2017-11-30 2018-08-10 中国科学院计算技术研究所 LSTM neural network chips and its application method
CN108491514A (en) * 2018-03-26 2018-09-04 清华大学 The method and device putd question in conversational system, electronic equipment, computer-readable medium
CN108563628A (en) * 2018-03-07 2018-09-21 中山大学 Talk with generation method based on the emotion of HRED and inside and outside memory network unit
CN108805088A (en) * 2018-06-14 2018-11-13 南京云思创智信息科技有限公司 Physiological signal analyzing subsystem based on multi-modal Emotion identification system
CN108920510A (en) * 2018-05-30 2018-11-30 出门问问信息科技有限公司 Automatic chatting method, device and electronic equipment
CN108959246A (en) * 2018-06-12 2018-12-07 北京慧闻科技发展有限公司 Answer selection method, device and electronic equipment based on improved attention mechanism
CN109658270A (en) * 2018-12-19 2019-04-19 前海企保科技(深圳)有限公司 It is a kind of to read the core compensation system and method understood based on insurance products
CN109684912A (en) * 2018-11-09 2019-04-26 中国科学院计算技术研究所 A kind of video presentation method and system based on information loss function
CN109948089A (en) * 2019-02-21 2019-06-28 中国海洋大学 A kind of method and device for extracting Web page text
CN110297894A (en) * 2019-05-22 2019-10-01 同济大学 A kind of Intelligent dialogue generation method based on auxiliary network
CN110334190A (en) * 2019-07-12 2019-10-15 电子科技大学 A kind of reply automatic generation method towards open field conversational system
CN110347833A (en) * 2019-07-09 2019-10-18 浙江工业大学 A kind of classification method of more wheel dialogues
CN110413729A (en) * 2019-06-25 2019-11-05 江南大学 Talk with generation method based on the more wheels of tail sentence-dual attention model of context
CN110516213A (en) * 2019-09-03 2019-11-29 哈尔滨工业大学 A kind of structural data generation text method based on table hierarchyization modeling
CN111797218A (en) * 2020-07-07 2020-10-20 海南中智信信息技术有限公司 Open domain dialogue generation method based on Cycle-Seq2Seq
CN111914983A (en) * 2019-05-07 2020-11-10 阿里巴巴集团控股有限公司 Interaction method and device, sound box, electronic equipment and storage medium
CN112417089A (en) * 2019-08-21 2021-02-26 东北大学秦皇岛分校 High-parallelism reading understanding method based on deep learning
CN113488036A (en) * 2020-06-10 2021-10-08 海信集团有限公司 Multi-round voice interaction method, terminal and server
US11245648B1 (en) 2020-07-31 2022-02-08 International Business Machines Corporation Cognitive management of context switching for multiple-round dialogues
CN114357129A (en) * 2021-12-07 2022-04-15 华南理工大学 High-concurrency multi-round chat robot system and data processing method thereof

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106126596A (en) * 2016-06-20 2016-11-16 中国科学院自动化研究所 A kind of answering method based on stratification memory network
CN106354710A (en) * 2016-08-18 2017-01-25 清华大学 Neural network relation extracting method
CN106383815A (en) * 2016-09-20 2017-02-08 清华大学 Neural network sentiment analysis method in combination with user and product information
CN106776578A (en) * 2017-01-03 2017-05-31 竹间智能科技(上海)有限公司 Talk with the method and device of performance for lifting conversational system
CN106844368A (en) * 2015-12-03 2017-06-13 华为技术有限公司 For interactive method, nerve network system and user equipment
CN106847271A (en) * 2016-12-12 2017-06-13 北京光年无限科技有限公司 A kind of data processing method and device for talking with interactive system
CN106933785A (en) * 2017-02-23 2017-07-07 中山大学 A kind of abstraction generating method based on recurrent neural network
CN107133211A (en) * 2017-04-26 2017-09-05 中国人民大学 A kind of composition methods of marking based on notice mechanism
CN107256228A (en) * 2017-05-02 2017-10-17 清华大学 Answer selection system and method based on structuring notice mechanism

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106844368A (en) * 2015-12-03 2017-06-13 华为技术有限公司 For interactive method, nerve network system and user equipment
CN106126596A (en) * 2016-06-20 2016-11-16 中国科学院自动化研究所 A kind of answering method based on stratification memory network
CN106354710A (en) * 2016-08-18 2017-01-25 清华大学 Neural network relation extracting method
CN106383815A (en) * 2016-09-20 2017-02-08 清华大学 Neural network sentiment analysis method in combination with user and product information
CN106847271A (en) * 2016-12-12 2017-06-13 北京光年无限科技有限公司 A kind of data processing method and device for talking with interactive system
CN106776578A (en) * 2017-01-03 2017-05-31 竹间智能科技(上海)有限公司 Talk with the method and device of performance for lifting conversational system
CN106933785A (en) * 2017-02-23 2017-07-07 中山大学 A kind of abstraction generating method based on recurrent neural network
CN107133211A (en) * 2017-04-26 2017-09-05 中国人民大学 A kind of composition methods of marking based on notice mechanism
CN107256228A (en) * 2017-05-02 2017-10-17 清华大学 Answer selection system and method based on structuring notice mechanism

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
DZMITRY BAHDANAU: "Neural Machine Translation by Jointly Learning to Align and Translate", 《COMPUTER SCIENCE》 *
HUAYU LI等: "A Context-aware Attention Network for Interactive Question Answering", 《KDD "17: PROCEEDINGS OF THE 23RD ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING》 *
LIFENG SHANG: "Neural Responding Machine for Short-Text Conversation", 《ACL》 *
ZICHAO YANG等: "Hierarchical Attention Networks for Document Classification", 《PROCEEDINGS OF THE 2016 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES》 *

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108388944A (en) * 2017-11-30 2018-08-10 中国科学院计算技术研究所 LSTM neural network chips and its application method
CN108563628A (en) * 2018-03-07 2018-09-21 中山大学 Talk with generation method based on the emotion of HRED and inside and outside memory network unit
CN108491514A (en) * 2018-03-26 2018-09-04 清华大学 The method and device putd question in conversational system, electronic equipment, computer-readable medium
CN108920510A (en) * 2018-05-30 2018-11-30 出门问问信息科技有限公司 Automatic chatting method, device and electronic equipment
CN108959246A (en) * 2018-06-12 2018-12-07 北京慧闻科技发展有限公司 Answer selection method, device and electronic equipment based on improved attention mechanism
CN108959246B (en) * 2018-06-12 2022-07-12 北京慧闻科技(集团)有限公司 Answer selection method and device based on improved attention mechanism and electronic equipment
CN108805088A (en) * 2018-06-14 2018-11-13 南京云思创智信息科技有限公司 Physiological signal analyzing subsystem based on multi-modal Emotion identification system
CN108805088B (en) * 2018-06-14 2021-05-28 南京云思创智信息科技有限公司 Physiological signal analysis subsystem based on multi-modal emotion recognition system
CN109684912A (en) * 2018-11-09 2019-04-26 中国科学院计算技术研究所 A kind of video presentation method and system based on information loss function
CN109658270A (en) * 2018-12-19 2019-04-19 前海企保科技(深圳)有限公司 It is a kind of to read the core compensation system and method understood based on insurance products
CN109948089A (en) * 2019-02-21 2019-06-28 中国海洋大学 A kind of method and device for extracting Web page text
CN111914983A (en) * 2019-05-07 2020-11-10 阿里巴巴集团控股有限公司 Interaction method and device, sound box, electronic equipment and storage medium
CN111914983B (en) * 2019-05-07 2023-10-24 阿里巴巴集团控股有限公司 Interaction method and device, sound box, electronic equipment and storage medium
CN110297894A (en) * 2019-05-22 2019-10-01 同济大学 A kind of Intelligent dialogue generation method based on auxiliary network
CN110297894B (en) * 2019-05-22 2021-03-26 同济大学 Intelligent dialogue generating method based on auxiliary network
CN110413729B (en) * 2019-06-25 2023-04-07 江南大学 Multi-turn dialogue generation method based on clause-context dual attention model
CN110413729A (en) * 2019-06-25 2019-11-05 江南大学 Talk with generation method based on the more wheels of tail sentence-dual attention model of context
CN110347833A (en) * 2019-07-09 2019-10-18 浙江工业大学 A kind of classification method of more wheel dialogues
CN110334190A (en) * 2019-07-12 2019-10-15 电子科技大学 A kind of reply automatic generation method towards open field conversational system
CN112417089A (en) * 2019-08-21 2021-02-26 东北大学秦皇岛分校 High-parallelism reading understanding method based on deep learning
CN112417089B (en) * 2019-08-21 2022-12-09 东北大学秦皇岛分校 High-parallelism reading understanding method based on deep learning
CN110516213A (en) * 2019-09-03 2019-11-29 哈尔滨工业大学 A kind of structural data generation text method based on table hierarchyization modeling
CN110516213B (en) * 2019-09-03 2022-04-15 哈尔滨工业大学 Structured data text generation method based on table hierarchical modeling
CN113488036A (en) * 2020-06-10 2021-10-08 海信集团有限公司 Multi-round voice interaction method, terminal and server
CN111797218A (en) * 2020-07-07 2020-10-20 海南中智信信息技术有限公司 Open domain dialogue generation method based on Cycle-Seq2Seq
CN111797218B (en) * 2020-07-07 2022-03-29 海南中智信信息技术有限公司 Open domain dialogue generation method based on Cycle-Seq2Seq
US11245648B1 (en) 2020-07-31 2022-02-08 International Business Machines Corporation Cognitive management of context switching for multiple-round dialogues
CN114357129A (en) * 2021-12-07 2022-04-15 华南理工大学 High-concurrency multi-round chat robot system and data processing method thereof
CN114357129B (en) * 2021-12-07 2023-02-14 华南理工大学 High-concurrency multi-round chat robot system and data processing method thereof

Similar Documents

Publication Publication Date Title
CN107766506A (en) A kind of more wheel dialog model construction methods based on stratification notice mechanism
Nan et al. Improving factual consistency of abstractive summarization via question answering
CN109460463B (en) Model training method, device, terminal and storage medium based on data processing
CN109271493A (en) A kind of language text processing method, device and storage medium
CN109547331A (en) One kind is based on more wheel voice-enabled chat model building methods
CN107423274A (en) Commentary content generating method, device and storage medium based on artificial intelligence
CN109791549A (en) Machine customer interaction towards dialogue
CN109964223A (en) Session information processing method and its device, storage medium
CN110234018B (en) Multimedia content description generation method, training method, device, equipment and medium
CN109284502B (en) Text similarity calculation method and device, electronic equipment and storage medium
CN110427461A (en) Intelligent answer information processing method, electronic equipment and computer readable storage medium
CN107679225B (en) Reply generation method based on keywords
CN111672098A (en) Virtual object marking method and device, electronic equipment and storage medium
CN108960574A (en) Quality determination method, device, server and the storage medium of question and answer
CN108345612A (en) A kind of question processing method and device, a kind of device for issue handling
CN110457661A (en) Spatial term method, apparatus, equipment and storage medium
CN113761156A (en) Data processing method, device and medium for man-machine interaction conversation and electronic equipment
CN110427454A (en) Text mood analysis method and device, electronic equipment and non-transient storage media
CN116956116A (en) Text processing method and device, storage medium and electronic equipment
Jhan et al. CheerBots: Chatbots toward empathy and emotionusing reinforcement learning
CN107729983A (en) A kind of method, apparatus and electronic equipment using realizing of Robot Vision man-machine chess
CN116258147A (en) Multimode comment emotion analysis method and system based on heterogram convolution
CN110491372A (en) A kind of feedback information generating method, device, storage medium and smart machine
CN116975214A (en) Text generation method, device, storage medium and computer equipment
CN113392640B (en) Title determination method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination