CN109241255A - A kind of intension recognizing method based on deep learning - Google Patents
A kind of intension recognizing method based on deep learning Download PDFInfo
- Publication number
- CN109241255A CN109241255A CN201810945991.9A CN201810945991A CN109241255A CN 109241255 A CN109241255 A CN 109241255A CN 201810945991 A CN201810945991 A CN 201810945991A CN 109241255 A CN109241255 A CN 109241255A
- Authority
- CN
- China
- Prior art keywords
- vector
- word
- deep learning
- sentence
- blstm
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Landscapes
- Machine Translation (AREA)
Abstract
A kind of conversational system intension recognizing method based on deep learning is matched the pre- dialogue D for carrying out intention assessment to obtain intent classifier result P with rule by rule of the word frequency weight extracting keywords as intention assessment from dialogue corpusA;With dialogue corpus training deep learning MODEL C NN-BLSTM, the deep learning MODEL C NN-BLSTM has merged convolutional neural networks CNN and two-way shot and long term memory network BLSTM, then pre- intent classifier result P is obtained with the deep learning MODEL C NN-BLSTM after training, identification dialogue DB;Finally, by by intent classifier result PAWith intent classifier result PBLinear fusion is carried out, the final intention of dialogue D is obtained.The present invention can effectively improve the accuracy of user's intention assessment.
Description
Technical field
The invention belongs to interactive system technical field, in particular to a kind of intention assessment side based on deep learning
Method.
Background technique
Interactive system is one of core technology of artificial intelligence field, it will become the new human-computer interaction of one kind
Mode has great researching value.People are pursuing and can exchanged with natural language with computer for a long time, because
This has great meaning: people, using computer, can be handed over oneself most familiar with the language being most accustomed to computer
Mutually, it does not need to deflorate a large amount of time study again and adapts to computer language.With the arrival of Internet era, human-computer dialogue system
The use demand of system greatly increases.Such as it is widely used in the intelligent customer service in online shopping, not only greatly improve people and meter
The communication efficiency of calculation machine, also facilitates people's lives and work.Intelligent dialogue system is also added in major science and technology giant one after another
Research ranks simultaneously release relevant product such as: the siri of apple, the Cortana of Microsoft, Baidu small degree.Perhaps soon
In the future, people will not be used the input equipment of current mainstream, and natural language will replace as most widely used people
Machine interactive mode.The key step of the interaction of man-machine natural language include: speech recognition, natural language understanding, dialogue state with
Track, spatial term, speech synthesis.
Natural language understanding is a module crucial in interactive system, and effect is that user says computer
Natural language be converted into computer it will be appreciated that semantic expressiveness, achieve the purpose that understand user's natural language.It is appreciated that user
Word, just must be known by field involved in user's natural language or user thinks the intention of expression, and user's intention assessment is just
It is that above-mentioned purpose is reached using the method for classification.The raising of user's intention assessment accuracy rate can greatly help conversational system raw
It is replied at reasonable.
In interactive system, the correct identification that user is intended to is that conversational system generates the basis rationally replied.If
The intention of user can not correct judgment, conversational system will generate the reply given an irrelevant answer, it is such reply it is also just not any
Meaning.Therefore for the performance boost of conversational system, increase user experience, the intention for accurately identifying user is particularly important.
In addition to this, accurately judge that user is intended to, commercial Intelligent dialogue system can provide a user useful consumption, amusement, product
Deng recommendation, there is very big commercial value.In conclusion user's intention assessment has critically important researching value and research meaning
Justice.
Summary of the invention
Problem to be solved by this invention is, using deep learning technology and to improve the accuracy of user's intention assessment.
Technical solution of the present invention provides a kind of conversational system intension recognizing method based on deep learning, firstly, from right
By rule of the word frequency weight extracting keywords as intention assessment in language material, the pre- dialogue D for carrying out intention assessment is advised
It is then matched to obtain intent classifier result PA;With dialogue corpus training deep learning MODEL C NN-BLSTM, the deep learning
MODEL C NN-BLSTM has merged convolutional neural networks CNN and two-way shot and long term memory network BLSTM, then with the depth after training
Learning model CNN-BLSTM is spent, identification dialogue D obtains intent classifier result PB;Finally, by by intent classifier result PAAnd meaning
Figure classification results PBLinear fusion is carried out, the final intention of dialogue D is obtained.
Moreover, described pass through rule of the word frequency weight extracting keywords as intention assessment from dialogue corpus, including right
Each classification carries out the following processing,
Word segmentation processing is carried out, and counts and obtains the number N of all entries under the category, all entries are formed into a word
Table, and count in the category, the number W that i-th of word occursi, the total number M of sentence and the length L of j-th strip sentencej, i=
1,2,3 ... N, j=1,2,3 ... M;
The average length AveL of sentence under the category is calculated, calculation formula is as follows,
Calculate the word frequency weight F of i-th of wordiIt is as follows,
Wherein, ∑ S indicates the summation that adds up to the length S for the sentence for occurring i-th of word in the category;
After acquiring word frequency weight to each word under the category, by each entry in vocabulary by word frequency weight from greatly to
Small sequence proposes to choose rule of the preceding several entries of ranking as keyword, as the category.
Moreover, the dialogue corpus training deep learning MODEL C NN-BLSTM, implementation is as follows,
After training corpus is carried out word segmentation processing, the vector of each word in every words resulting after participle is indicated into xbGroup
The vector for synthesizing word indicates X=[x1, x2, x3... xl], l is the sequence length of vector, b=1,2,3 ... l;
The convolutional layer of the vector X input convolutional neural networks of sentence is calculated, all characteristic pattern s that will be obtainedaGroup closes
To obtain output result S=[s1,s2,s3…sn], n indicates the total number of characteristic pattern, a=1,2,3 ... n;
The structure of S is rearranged, gained vector V after permutatation is inputted in BLSTM neural network model,
BLSTM is two-way shot and long term Memory Neural Networks, by the shot and long term Memory Neural Networks and a backward length of a forward direction
Phase Memory Neural Networks are constituted;For each time step t, the shot and long term Memory Neural Networks of forward direction export hidden layer state
Backward shot and long term Memory Neural Networks export hidden layer stateThe vector for combining two hidden layer states obtains vector ht;According to
The vector of all time steps indicates, obtains the corresponding vector H of entire sentence, implicit context semantic information in H;
Vector O is obtained after carrying out maximum pondization operation to S, includes most important semantic feature in sentence and classification spy in O
Reference breath;
Vector H and vector O are stitched together as vector T,
It is indicated T as final dialogue sentence characteristics vector, connects all sentence characteristics T and obtain intermediate quantity yc, according to yc
The probability of each classification is obtained, chooses the intention classification of maximum probability as intention assessment result PB。
Moreover, setting intent classifier result PAWith intent classifier result PBFinal intention class is obtained after carrying out linear fusion
Other probability distribution P chooses result of the intention classification as final intention assessment corresponding to maximum probability in P.
The present invention provides a kind of intension recognizing methods based on deep learning, on the one hand, passes through word from dialogue corpus
Rule of the frequency weight extracting keywords as intention assessment matches the pre- dialogue D for carrying out intention assessment with these rules
Obtain pre- intent classifier result PA, on the other hand convolutional neural networks (CNN) and two-way length have been merged with dialogue corpus training
The deep learning model (CNN-BLSTM) of phase memory network (BLSTM) is identified to obtain with the model after training to dialogue D
Pre- intent classifier result PB, finally, by by intent classifier result PAWith intent classifier result PBLinear fusion is carried out, is obtained pair
Talk about the final intention of D.According to the intention assessment side provided by the present invention based on combined deep learning model CNN-BLSTM model
Method can only considered timing information in intention assessment efficiently against depth of foundation learning model and have ignored local important
The deficiency of information.Rule is obtained in addition, sentence length is incorporated in word frequency weight calculation, can effectively measure the important journey of word
Degree selects more representational word as rule.Finally, classification results and rule by combining CNN-BLSTM model
The classification results matched can accurately obtain the intention of user.On the data set of multiple official's intention assessments in the world
The comparative test result of best multiple models shows to combine combined deep learning model and involvement according to provided by the invention
The intension recognizing method of the rule match of sentence length information, realizes on accuracy of identification and is obviously improved.The present invention can have
Effect improves the accuracy of user's intention assessment, and the correct identification that user is intended to is that Intelligent dialogue system generates the base rationally replied
Plinth accurately identifies the intention of user, and the performance boost of energy conversational system increases user experience, has very big value and research
Meaning.
Detailed description of the invention
Fig. 1 is the flow chart of intention assessment in the embodiment of the present invention.
Specific implementation method
The present invention proposes a kind of in the two-way shot and long term Memory Neural Networks (LSTM) of convolutional neural networks (CNN) involvement
Combined deep learning model mentions to realize intent classifier using the sentence length information in corpus as influence factor involvement rule
In taking, by the matched classification results of binding rule, the accuracy of intention assessment is further increased.
Combined deep learning model proposed by the present invention is known as CNN-BLSTM model, is used for intention assessment.Traditional depth
Learning model generallys use Recognition with Recurrent Neural Network (RNN) and its mutation shot and long term Memory Neural Networks in being intended to identification mission
(LSTM) etc., this neural network can hold well the timing information of sentence, but be a lack of local important information.It is basic herein
On, the present invention incorporates convolutional neural networks (CNN) on traditional model to obtain the important semantic information in the part in sentence.It should
Built-up pattern can capture user's intention using more information.
The present invention is proposed for Rules extraction method unreasonable in Classical matching process by the sentence in corpus
Length information is as impact factor in view of in word frequency weight calculation, obtaining more reasonable keyword word as rule.Generally
For, significance level of the same word in short sentence it is bigger than significance level in long sentence (such as " I wants to listen song ",
" song " word is greater than the latter in the former importance in " room-mate's sleep can probably be bothered by singing now "), efficiently use sentence
The length information of son, can preferably capture the true intention of user.
Present invention incorporates rule matching method with combined deep learning model method as a result, conventional method and depth
Learning method complements each other, and further increases the accuracy of user's intention assessment.Firstly, being taken out from dialogue corpus by word frequency weight
Rule of the keyword as intention assessment is taken, the pre- dialogue D for carrying out intention assessment is matched with these rules and is intended to
Classification results PA, then convolutional neural networks (CNN) and two-way shot and long term memory network have been merged with dialogue corpus training
(BLSTM) deep learning model (CNN-BLSTM) is identified to obtain intent classifier knot with the model after training to dialogue D
Fruit PB, finally, by by intent classifier result PAWith intent classifier result PBLinear fusion is carried out, the final meaning of dialogue D is obtained
Figure.
Referring to Fig. 1, embodiment the specific implementation process is as follows:
Step 1, to the dialogue corpus (training corpus) for having marked classification, to each classification, following place is carried out respectively
Reason, obtains the rule of each classification:
Word segmentation processing is carried out with Jieba participle kit, and counts and obtains the number N of all entries under the category, by institute
By entry form a vocabulary.And it counts in the category, the number W that i-th of word occursi(i=1,2,3 ... N), sentence
The length L of total number M and j-th strip sentencej(j=1,2,3 ... M).It is existing software tool, this hair that Jieba, which segments kit,
It is bright that it will not go into details.
Then the average length for calculating sentence under the category carries out cumulative summation again to the length of all sentences of the category
Divided by the total number of sentence, the average length for obtaining classification sentence is AveL, and calculation formula is as follows,
The word frequency weight F of i-th of word can be calculated according to formula (1)i, as shown in formula (2):
In above-mentioned formula, FiIndicate the word frequency weight of i-th of word of the category, WiIndicate the number that category word i occurs, N
Indicate the number of all entries under the category, AveL is the average length of sentence in the category, and S indicates to include i-th in the category
The length of some sentence of a word, ∑ S are the summation that adds up to the length for the sentence for occurring i-th of word in the category;
Its word frequency weight is acquired according to formula (2) to each word under the category, each entry in vocabulary is pressed into word
Frequency weight sorts from large to small, and embodiment proposes the entry of ranking 1% before choosing as keyword, these keywords are just used as should
The rule of classification.When it is implemented, user can preset selection ratio, embodiment uses preferred ratio 1%.
Step 2, the rule based on step 1 each classification obtained, and duplicate rule in different classes of is deleted, finally
Result of the obtained rule as final rule extraction.
Step 3, the pre- dialogue D for carrying out intention assessment is matched one by one with the rule that step 2 obtains (once successfully
It is fitted on a certain rule, then terminates matching), if comprising some rule in D, the corresponding probability mark for being intended to classification of the rule
Be denoted as 1, other probability for being intended to classifications are collectively labeled as 0, obtain D it is corresponding intentional classification probability distribution PA=[p1,p2,
p3,…pd], d indicates intention class number in total, p1,p2,p3,…pdThe probability of respectively the 1st~the d intention classification.In order to
Facilitate understanding, below will for example: if " song " word is the rule that music is intended to classification, when conversation sentence D is that " you like
Listen what kind of song " the words when, comprising " song " this rule in D, then judge that the music of D is intended to class probability as 1,
It is 0 that other, which are intended to class probability,.
Step 4, the embodiment of the present invention is trained on Chinese wikipedia large-scale corpus by word2vec tool
Obtain term vector set.Vector in the term vector set comprising a large amount of word indicates.Word2vec tool is one by word
It is converted into the existing software tool of vector form, it will not go into details by the present invention.
Step 5, after training corpus Jieba being segmented kit participle, by each of every words resulting after participle
The vector of word indicates xbThe vector that (can find in the term vector set that step 4 obtains) is combined into word indicates X=[x1,
x2, x3... xl], l be vector sequence length (embodiment of the present invention according in training corpus sentence length statistics and analysis select
Select l=40), when length is more than l, interception is no more than the length thereof of l, zero padding when insufficient, b=1,2,3 ... l.Specific implementation
When, user can preset the value for choosing l.
The convolutional layer of the vector X input convolutional neural networks of sentence is calculated, in convolution process, sliding window is according to sentence
Sub- sequence dimension direction convolution, sliding window length, that is, convolution kernel size are respectively set to 2,3,5.Every kind of convolution nuclear volume is set
It is set to 128.Each convolution kernel sliding distich subvector does convolution operation, obtains different degrees of characteristic pattern sa(each ruler
Very little convolution kernel can generate 128 characteristic pattern feature map), by all sa(a=1,2,3 ... n) combines to obtain defeated
Result S=[s out1,s2,s3…sn], n indicates the total number of characteristic pattern (for example, n=3 × 128 in the present embodiment).
For the relative ranks for keeping sentence, the structure of S is rearranged to obtain result V, permutatation mode is as follows:
S=[s1,s2,…sn] formula (3)
In above-mentioned formula (3), the convolution results that S is are made of, s several characteristic patternsa(a=1,2,3 ... n)
Indicate a-th of the characteristic pattern obtained after convolution operation, each characteristic pattern is a vector, in formula (4) and formula (5)
Indicate the numerical value of the corresponding element of a-th of characteristic pattern b (b=1,2,3 ... l) dimension in convolution results.vbIt is by saAfter rearranging
Corresponding vector, by vbIt combines to obtain final permutatation result V.
V=[v1,v2,…vl] formula (6)
By in the vector V input BLSTM neural network model after permutatation, BLSTM is two-way shot and long term memory nerve net
Network is made of the shot and long term Memory Neural Networks of forward direction and a backward shot and long term Memory Neural Networks.For each
(t=1,2,3 ... l) (input of a word is as a time step in a word), and the shot and long term of forward direction remembers mind by a time step t
Hidden layer state is exported through networkBackward shot and long term Memory Neural Networks export hidden layer stateCombine two hidden layer states
Vector obtains vector ht;
H=[h1,h2,h3…hl] formula (8)
Wherein, H indicates that the vector of all time steps indicates, i.e., the vector of entire sentence indicates that implicit context is semantic in H
Information.
Step 6, convolution results S obtained in step 5 is subjected to pondization operation, pondization operates the output to convolutional layer and carries out
Feature sampling, can merge the feature of various sizes convolution window extraction.The present invention is using maximum pond (max-pooling) side
Method, this method retain the maximum most significant feature of characteristic vector pickup.The output in maximum pond is defined as,
smax=max sa, a ∈ [1, n] formula (9)
O=h (Wsmax+ b) formula (10)
S in above-mentioned formula (9)aIndicate the characteristic pattern of convolutional layer output, smaxIt indicates to choose saIn maximum feature.In public affairs
In formula (10), operation h () is nonlinear activation function, and the present invention is as activation primitive, W and b using LeakyReLU function
Parameter in convolutional network, initial value random value between 0 to 1.Vector O is obtained after operating by maximum pondization, includes in O
Most important semantic feature and category feature information in sentence;
Step 7, the vector H that step 5 obtains and the vector O that step 6 obtains are stitched together as new vector T, splicing
The method of vector H and vector O are as follows:
T=[O, H] formula (11)
It is indicated T as final dialogue sentence characteristics vector, by full articulamentum, connects all sentence characteristics T and obtain
yc, connection type is as follows,
yc=h (Wc×T+bc) formula (12)
By ycIt being input in softmax function and obtains the probability of each classification, calculation formula is as follows,
In formula (12) and formula (13), c (c=1,2,3 ..., d) indicates that c class is intended to classification, ycIt is intermediate quantity, Wc
It is the parameter of the full articulamentum of convolutional neural networks, bcIt is bias term parameter, d indicates intention class number in total, operation h ()
It is nonlinear activation function, for the embodiment of the present invention using tanh function as activation primitive, e is the truth of a matter of natural logrithm, pcTable
Show that user's sentence belongs to the probability that c class is intended to classification.Calculate intentional classification probability PB=[p1,p2,p3,…
pd], p1,p2,p3,…pdThe probability of respectively the 1st~the d intention classification.
Step 8, step 5-7 is repeated to every a word in training corpus and obtains its probability distribution for being intended to classification, chosen
The intention classification of maximum probability is intended to classification as prediction, by the way that (corpus dialogue provides the true of every words with its true intention
It is intended to) it is compared, the parameter in CNN-BLSTM model and constantly iteration optimization model is trained with this, with having trained
CNN-BLSTM model come calculate it is pre- carry out intention assessment dialogue D (being calculated by the corresponding mode of step 5-7), obtain
To D it is corresponding intentional classification probability distribution PB。
Step 9, for the pre- dialogue D for carrying out intention assessment, the intention class probability that step 3 is obtained is distributed PAWith step
8 obtained intention class probability distribution PBIt carries out linear fusion and obtains final intention class probability distribution P, amalgamation mode is such as
Under,
P=α × PA+β×PBFormula (14)
In formula (14), alpha+beta=1 can take α, and (0.0,0.1,0.2 ... 1.0), preset α as needed, β's takes by β ∈
Value, such as α=0.5, β=0.5.
Finally, choosing result of the intention classification as final intention assessment corresponding to maximum probability in P.
Step 1~3 are to realize rule-based matched classification, and step 4~8 are to realize point based on CNN-BLSTM model
Class, step 9 carry out the combination of the two.
When it is implemented, the automatic running that software technology realizes the above process can be used in those skilled in the art.Correspondingly,
Provided that a kind of intension recognizing method based on deep learning technology, including computer or server, in computer or service
The matched CNN-BLSTM model realization intention assessment of the above process binding rule is executed on device, it should also in guarantor of the invention
It protects in range.
Claims (4)
1. a kind of conversational system intension recognizing method based on deep learning, it is characterised in that: firstly, passing through from dialogue corpus
The pre- dialogue D for carrying out intention assessment match with rule by rule of the word frequency weight extracting keywords as intention assessment
To intent classifier result PA;With dialogue corpus training deep learning MODEL C NN-BLSTM, the deep learning MODEL C NN-
BLSTM has merged convolutional neural networks CNN and two-way shot and long term memory network BLSTM, then with the deep learning mould after training
Type CNN-BLSTM, identification dialogue D, obtains intent classifier result PB;Finally, by by intent classifier result PAWith intent classifier knot
Fruit PBLinear fusion is carried out, the final intention of dialogue D is obtained.
2. the conversational system intension recognizing method based on deep learning according to claim 1, it is characterised in that: described from right
By rule of the word frequency weight extracting keywords as intention assessment in language material, including each classification is carried out the following processing,
Word segmentation processing is carried out, and counts and obtains the number N of all entries under the category, all entries are formed into a vocabulary,
And it counts in the category, the number W that i-th of word occursi, the total number M of sentence and the length L of j-th strip sentencej, i=1,
2,3 ... N, j=1,2,3 ... M;
The average length AveL of sentence under the category is calculated, calculation formula is as follows,
Calculate the word frequency weight F of i-th of wordiIt is as follows,
Wherein, ∑ S indicates the summation that adds up to the length S for the sentence for occurring i-th of word in the category;
After acquiring word frequency weight to each word under the category, each entry in vocabulary is arranged from big to small by word frequency weight
Sequence proposes to choose rule of the preceding several entries of ranking as keyword, as the category.
3. the conversational system intension recognizing method based on deep learning according to claim 1, it is characterised in that: the use pair
Language material trains deep learning MODEL C NN-BLSTM, and implementation is as follows,
After training corpus is carried out word segmentation processing, the vector of each word in every words resulting after participle is indicated into xbIt is combined into
The vector of the word indicates X=[x1, x2, x3... xl], l is the sequence length of vector, b=1,2,3 ... l;
The convolutional layer of the vector X input convolutional neural networks of sentence is calculated, all characteristic pattern s that will be obtainedaIt combines to obtain
Export result S=[s1,s2,s3…sn], n indicates the total number of characteristic pattern, a=1,2,3 ... n;
The structure of S is rearranged, by gained vector V input BLSTM neural network model, BLSTM is after permutatation
Two-way shot and long term Memory Neural Networks, by the shot and long term Memory Neural Networks and a backward shot and long term memory mind of a forward direction
It is constituted through network;For each time step t, the shot and long term Memory Neural Networks of forward direction export hidden layer stateBackward length
Short-term memory neural network exports hidden layer stateThe vector for combining two hidden layer states obtains vector ht;According to all time steps
Vector indicate, obtain the corresponding vector H of entire sentence, implicit context semantic information in H;
Vector O is obtained after carrying out maximum pondization operation to S, includes most important semantic feature in sentence and category feature letter in O
Breath;
Vector H and vector O are stitched together as vector T,
It is indicated T as final dialogue sentence characteristics vector, connects all sentence characteristics T and obtain intermediate quantity yc, according to ycIt obtains
The probability of each classification chooses the intention classification of maximum probability as intention assessment result PB。
4. the according to claim 1 or 2 or 3 conversational system intension recognizing methods based on deep learning, it is characterised in that: set
By intent classifier result PAWith intent classifier result PBFinal intention class probability distribution P is obtained after carrying out linear fusion, in P
Result of the intention classification as final intention assessment corresponding to middle selection maximum probability.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810945991.9A CN109241255B (en) | 2018-08-20 | 2018-08-20 | Intention identification method based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810945991.9A CN109241255B (en) | 2018-08-20 | 2018-08-20 | Intention identification method based on deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109241255A true CN109241255A (en) | 2019-01-18 |
CN109241255B CN109241255B (en) | 2021-05-18 |
Family
ID=65071796
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810945991.9A Active CN109241255B (en) | 2018-08-20 | 2018-08-20 | Intention identification method based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109241255B (en) |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109893095A (en) * | 2019-03-11 | 2019-06-18 | 常州市贝叶斯智能科技有限公司 | A kind of intelligent robot system of body composition detection and analysis |
CN110162775A (en) * | 2019-03-11 | 2019-08-23 | 腾讯科技(深圳)有限公司 | Determine the method, apparatus and computer equipment of intention assessment accuracy |
CN110209791A (en) * | 2019-06-12 | 2019-09-06 | 百融云创科技股份有限公司 | It is a kind of to take turns dialogue intelligent speech interactive system and device more |
CN110232114A (en) * | 2019-05-06 | 2019-09-13 | 平安科技(深圳)有限公司 | Sentence intension recognizing method, device and computer readable storage medium |
CN110245348A (en) * | 2019-05-17 | 2019-09-17 | 北京百度网讯科技有限公司 | A kind of intension recognizing method and system |
CN110321564A (en) * | 2019-07-05 | 2019-10-11 | 浙江工业大学 | A kind of more wheel dialogue intension recognizing methods |
CN110334340A (en) * | 2019-05-06 | 2019-10-15 | 北京泰迪熊移动科技有限公司 | Semantic analysis, device and the readable storage medium storing program for executing of rule-based fusion |
CN110377911A (en) * | 2019-07-23 | 2019-10-25 | 中国工商银行股份有限公司 | Intension recognizing method and device under dialogue frame |
CN110795944A (en) * | 2019-10-11 | 2020-02-14 | 腾讯科技(深圳)有限公司 | Recommended content processing method and device, and emotion attribute determining method and device |
CN110928997A (en) * | 2019-12-04 | 2020-03-27 | 北京文思海辉金信软件有限公司 | Intention recognition method and device, electronic equipment and readable storage medium |
CN111400440A (en) * | 2020-02-28 | 2020-07-10 | 深圳市华海同创科技有限公司 | Intention identification method and device |
CN111462752A (en) * | 2020-04-01 | 2020-07-28 | 北京思特奇信息技术股份有限公司 | Client intention identification method based on attention mechanism, feature embedding and BI-L STM |
WO2020155766A1 (en) * | 2019-01-31 | 2020-08-06 | 平安科技(深圳)有限公司 | Method, device and apparatus for identification rejection in intention identification, and storage medium |
CN111597320A (en) * | 2020-05-26 | 2020-08-28 | 成都晓多科技有限公司 | Intention recognition device, method, equipment and storage medium based on hierarchical classification |
CN111639152A (en) * | 2019-08-29 | 2020-09-08 | 上海卓繁信息技术股份有限公司 | Intention recognition method |
CN111737544A (en) * | 2020-05-13 | 2020-10-02 | 北京三快在线科技有限公司 | Search intention recognition method and device, electronic equipment and storage medium |
US10916242B1 (en) | 2019-08-07 | 2021-02-09 | Nanjing Silicon Intelligence Technology Co., Ltd. | Intent recognition method based on deep learning network |
WO2021022816A1 (en) * | 2019-08-07 | 2021-02-11 | 南京硅基智能科技有限公司 | Intent identification method based on deep learning network |
CN112667816A (en) * | 2020-12-31 | 2021-04-16 | 华中师范大学 | Deep learning-based aspect level emotion analysis method and system |
CN112989003A (en) * | 2021-04-01 | 2021-06-18 | 网易(杭州)网络有限公司 | Intention recognition method, device, processing equipment and medium |
CN113094475A (en) * | 2021-06-08 | 2021-07-09 | 成都晓多科技有限公司 | Dialog intention recognition system and method based on context attention flow |
CN113158062A (en) * | 2021-05-08 | 2021-07-23 | 清华大学深圳国际研究生院 | User intention identification method and device based on heterogeneous graph neural network |
CN113678133A (en) * | 2019-04-05 | 2021-11-19 | 三星电子株式会社 | System and method for context-rich attention memory network with global and local encoding for dialog break detection |
CN114547435A (en) * | 2020-11-24 | 2022-05-27 | 腾讯科技(深圳)有限公司 | Content quality identification method, device, equipment and readable storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090012778A1 (en) * | 2007-07-05 | 2009-01-08 | Nec (China) Co., Ltd. | Apparatus and method for expanding natural language query requirement |
CN107679234A (en) * | 2017-10-24 | 2018-02-09 | 上海携程国际旅行社有限公司 | Customer service information providing method, device, electronic equipment, storage medium |
CN107679199A (en) * | 2017-10-11 | 2018-02-09 | 北京邮电大学 | A kind of external the Chinese text readability analysis method based on depth local feature |
US20180157638A1 (en) * | 2016-12-02 | 2018-06-07 | Microsoft Technology Licensing, Llc | Joint language understanding and dialogue management |
US20180173999A1 (en) * | 2016-12-21 | 2018-06-21 | XBrain, Inc. | Natural Transfer of Knowledge Between Human and Artificial Intelligence |
CN108415923A (en) * | 2017-10-18 | 2018-08-17 | 北京邮电大学 | The intelligent interactive system of closed domain |
-
2018
- 2018-08-20 CN CN201810945991.9A patent/CN109241255B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090012778A1 (en) * | 2007-07-05 | 2009-01-08 | Nec (China) Co., Ltd. | Apparatus and method for expanding natural language query requirement |
US20180157638A1 (en) * | 2016-12-02 | 2018-06-07 | Microsoft Technology Licensing, Llc | Joint language understanding and dialogue management |
US20180173999A1 (en) * | 2016-12-21 | 2018-06-21 | XBrain, Inc. | Natural Transfer of Knowledge Between Human and Artificial Intelligence |
CN107679199A (en) * | 2017-10-11 | 2018-02-09 | 北京邮电大学 | A kind of external the Chinese text readability analysis method based on depth local feature |
CN108415923A (en) * | 2017-10-18 | 2018-08-17 | 北京邮电大学 | The intelligent interactive system of closed domain |
CN107679234A (en) * | 2017-10-24 | 2018-02-09 | 上海携程国际旅行社有限公司 | Customer service information providing method, device, electronic equipment, storage medium |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020155766A1 (en) * | 2019-01-31 | 2020-08-06 | 平安科技(深圳)有限公司 | Method, device and apparatus for identification rejection in intention identification, and storage medium |
CN109893095A (en) * | 2019-03-11 | 2019-06-18 | 常州市贝叶斯智能科技有限公司 | A kind of intelligent robot system of body composition detection and analysis |
CN110162775A (en) * | 2019-03-11 | 2019-08-23 | 腾讯科技(深圳)有限公司 | Determine the method, apparatus and computer equipment of intention assessment accuracy |
CN113678133B (en) * | 2019-04-05 | 2024-10-01 | 三星电子株式会社 | System and method for context-rich attention memory network with global and local coding for dialog break detection |
CN113678133A (en) * | 2019-04-05 | 2021-11-19 | 三星电子株式会社 | System and method for context-rich attention memory network with global and local encoding for dialog break detection |
CN110334340A (en) * | 2019-05-06 | 2019-10-15 | 北京泰迪熊移动科技有限公司 | Semantic analysis, device and the readable storage medium storing program for executing of rule-based fusion |
CN110232114A (en) * | 2019-05-06 | 2019-09-13 | 平安科技(深圳)有限公司 | Sentence intension recognizing method, device and computer readable storage medium |
CN110334340B (en) * | 2019-05-06 | 2021-08-03 | 北京泰迪熊移动科技有限公司 | Semantic analysis method and device based on rule fusion and readable storage medium |
CN110245348A (en) * | 2019-05-17 | 2019-09-17 | 北京百度网讯科技有限公司 | A kind of intension recognizing method and system |
CN110245348B (en) * | 2019-05-17 | 2023-11-24 | 北京百度网讯科技有限公司 | Intention recognition method and system |
CN110209791A (en) * | 2019-06-12 | 2019-09-06 | 百融云创科技股份有限公司 | It is a kind of to take turns dialogue intelligent speech interactive system and device more |
CN110209791B (en) * | 2019-06-12 | 2021-03-26 | 百融云创科技股份有限公司 | Multi-round dialogue intelligent voice interaction system and device |
CN110321564A (en) * | 2019-07-05 | 2019-10-11 | 浙江工业大学 | A kind of more wheel dialogue intension recognizing methods |
CN110321564B (en) * | 2019-07-05 | 2023-07-14 | 浙江工业大学 | Multi-round dialogue intention recognition method |
CN110377911A (en) * | 2019-07-23 | 2019-10-25 | 中国工商银行股份有限公司 | Intension recognizing method and device under dialogue frame |
CN110377911B (en) * | 2019-07-23 | 2023-07-21 | 中国工商银行股份有限公司 | Method and device for identifying intention under dialog framework |
US10916242B1 (en) | 2019-08-07 | 2021-02-09 | Nanjing Silicon Intelligence Technology Co., Ltd. | Intent recognition method based on deep learning network |
WO2021022816A1 (en) * | 2019-08-07 | 2021-02-11 | 南京硅基智能科技有限公司 | Intent identification method based on deep learning network |
CN111639152B (en) * | 2019-08-29 | 2021-04-13 | 上海卓繁信息技术股份有限公司 | Intention recognition method |
CN111639152A (en) * | 2019-08-29 | 2020-09-08 | 上海卓繁信息技术股份有限公司 | Intention recognition method |
CN110795944A (en) * | 2019-10-11 | 2020-02-14 | 腾讯科技(深圳)有限公司 | Recommended content processing method and device, and emotion attribute determining method and device |
CN110928997A (en) * | 2019-12-04 | 2020-03-27 | 北京文思海辉金信软件有限公司 | Intention recognition method and device, electronic equipment and readable storage medium |
CN111400440A (en) * | 2020-02-28 | 2020-07-10 | 深圳市华海同创科技有限公司 | Intention identification method and device |
CN111462752A (en) * | 2020-04-01 | 2020-07-28 | 北京思特奇信息技术股份有限公司 | Client intention identification method based on attention mechanism, feature embedding and BI-L STM |
CN111462752B (en) * | 2020-04-01 | 2023-10-13 | 北京思特奇信息技术股份有限公司 | Attention mechanism, feature embedding and BI-LSTM (business-to-business) based customer intention recognition method |
CN111737544A (en) * | 2020-05-13 | 2020-10-02 | 北京三快在线科技有限公司 | Search intention recognition method and device, electronic equipment and storage medium |
CN111597320A (en) * | 2020-05-26 | 2020-08-28 | 成都晓多科技有限公司 | Intention recognition device, method, equipment and storage medium based on hierarchical classification |
CN114547435A (en) * | 2020-11-24 | 2022-05-27 | 腾讯科技(深圳)有限公司 | Content quality identification method, device, equipment and readable storage medium |
CN112667816B (en) * | 2020-12-31 | 2022-07-05 | 华中师范大学 | Deep learning-based aspect level emotion analysis method and system |
CN112667816A (en) * | 2020-12-31 | 2021-04-16 | 华中师范大学 | Deep learning-based aspect level emotion analysis method and system |
CN112989003B (en) * | 2021-04-01 | 2023-04-18 | 网易(杭州)网络有限公司 | Intention recognition method, device, processing equipment and medium |
CN112989003A (en) * | 2021-04-01 | 2021-06-18 | 网易(杭州)网络有限公司 | Intention recognition method, device, processing equipment and medium |
CN113158062A (en) * | 2021-05-08 | 2021-07-23 | 清华大学深圳国际研究生院 | User intention identification method and device based on heterogeneous graph neural network |
CN113094475A (en) * | 2021-06-08 | 2021-07-09 | 成都晓多科技有限公司 | Dialog intention recognition system and method based on context attention flow |
Also Published As
Publication number | Publication date |
---|---|
CN109241255B (en) | 2021-05-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109241255A (en) | A kind of intension recognizing method based on deep learning | |
CN105677873B (en) | Text Intelligence association cluster based on model of the domain knowledge collects processing method | |
CN106529503B (en) | A kind of integrated convolutional neural networks face emotion identification method | |
CN108363790A (en) | For the method, apparatus, equipment and storage medium to being assessed | |
CN110532379B (en) | Electronic information recommendation method based on LSTM (least Square TM) user comment sentiment analysis | |
CN108763326A (en) | A kind of sentiment analysis model building method of the diversified convolutional neural networks of feature based | |
CN108363690A (en) | Dialog semantics Intention Anticipation method based on neural network and learning training method | |
CN108519890A (en) | A kind of robustness code abstraction generating method based on from attention mechanism | |
CN107918782A (en) | A kind of method and system for the natural language for generating description picture material | |
CN104598611B (en) | The method and system being ranked up to search entry | |
CN108597541A (en) | A kind of speech-emotion recognition method and system for enhancing indignation and happily identifying | |
CN111966917A (en) | Event detection and summarization method based on pre-training language model | |
CN110717843A (en) | Reusable law strip recommendation framework | |
CN111414461A (en) | Intelligent question-answering method and system fusing knowledge base and user modeling | |
CN110992988B (en) | Speech emotion recognition method and device based on domain confrontation | |
CN106682089A (en) | RNNs-based method for automatic safety checking of short message | |
CN116205222A (en) | Aspect-level emotion analysis system and method based on multichannel attention fusion | |
CN107145514A (en) | Chinese sentence pattern sorting technique based on decision tree and SVM mixed models | |
CN110415071A (en) | A kind of competing product control methods of automobile based on opining mining analysis | |
CN109325780A (en) | A kind of exchange method of the intelligent customer service system in E-Governance Oriented field | |
CN117236338B (en) | Named entity recognition model of dense entity text and training method thereof | |
CN112561718A (en) | Case microblog evaluation object emotion tendency analysis method based on BilSTM weight sharing | |
CN108875034A (en) | A kind of Chinese Text Categorization based on stratification shot and long term memory network | |
TW202034207A (en) | Dialogue system using intention detection ensemble learning and method thereof | |
CN113268628A (en) | Music emotion recognition method based on modularized weighted fusion neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |