CN112183062B - Spoken language understanding method based on alternate decoding, electronic equipment and storage medium - Google Patents

Spoken language understanding method based on alternate decoding, electronic equipment and storage medium Download PDF

Info

Publication number
CN112183062B
CN112183062B CN202011045822.3A CN202011045822A CN112183062B CN 112183062 B CN112183062 B CN 112183062B CN 202011045822 A CN202011045822 A CN 202011045822A CN 112183062 B CN112183062 B CN 112183062B
Authority
CN
China
Prior art keywords
word
decoding
sequence
level
slot
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011045822.3A
Other languages
Chinese (zh)
Other versions
CN112183062A (en
Inventor
刘广灿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Unisound Intelligent Technology Co Ltd
Xiamen Yunzhixin Intelligent Technology Co Ltd
Original Assignee
Unisound Intelligent Technology Co Ltd
Xiamen Yunzhixin Intelligent Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Unisound Intelligent Technology Co Ltd, Xiamen Yunzhixin Intelligent Technology Co Ltd filed Critical Unisound Intelligent Technology Co Ltd
Priority to CN202011045822.3A priority Critical patent/CN112183062B/en
Publication of CN112183062A publication Critical patent/CN112183062A/en
Application granted granted Critical
Publication of CN112183062B publication Critical patent/CN112183062B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/126Character encoding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Databases & Information Systems (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a spoken language understanding method based on alternate decoding, electronic equipment and a storage medium.

Description

Spoken language understanding method based on alternate decoding, electronic equipment and storage medium
Technical Field
The invention relates to the technical field of man-machine conversation systems, in particular to a spoken language understanding method based on alternate decoding, electronic equipment and a storage medium.
Background
Spoken language understanding mainly includes two subtasks: intent recognition (Intint Detection) and Slot Filling (Slot Filling). The two tasks of intention recognition and slot filling are not independent of each other, and the slot filling is highly dependent on the result of the intention recognition, and the slot filling can also promote the intention recognition. The prior art jointly models two tasks to make full use of knowledge information common to the two tasks, typically using a Multi-task (multitask) framework, two subtasks sharing the coding layer, and then adding the loss functions of the two parts to model.
The existing method which uses a multi-task framework adopts a shared coding layer and loss function addition is an implicit joint modeling method, and the method does not explicitly model the interaction relation between intention recognition and slot filling two subtasks; the Slot-Gated method and the SF-ID method are also researched and proposed for preliminary exploration, but the existing model cannot fully utilize the co-occurrence relation between the groove and the intention, so that the potential performance of the groove and the intention is limited.
Disclosure of Invention
In order to solve the problems, the invention provides that the word-level intention recognition and word-level slot filling are alternately decoded, and the whole decoding state is initialized by using the sentence-level intention recognition result, so that the mutual contribution of the intention recognition and the slot filling can be realized from the two angles of the global and the local, thereby improving the effect of the spoken language understanding task.
According to an aspect of an embodiment of the present invention, there is provided a spoken language understanding method based on alternate decoding, including:
s100, obtaining a semantic vector sequence of a pre-training language model BERT based on an input sequence, and taking a first mark of the input sequence as a classification mark;
s200, obtaining a corresponding final hidden state according to the classification mark, and carrying out sentence-level intention recognition by using a fully connected neural network and a Softmax function based on the final hidden state to obtain sentence-level global intention information;
S300, word level intention recognition and slot filling alternate decoding are carried out based on the semantic vector sequence and the sentence level global intention information;
s400 obtains the result of intention recognition and slot filling based on the result of the alternate decoding.
Preferably, the alternately decoding includes: and obtaining a corresponding decoder hiding state based on the semantic vector sequence and the sentence-level global intention information, decoding the decoder hiding state to obtain an output sequence, wherein the length of the output sequence is 2n, and obtaining a word-level intention recognition sequence and a word-level slot filling sequence based on the output sequence, wherein the decoder is unidirectional LSTM.
Preferably, the decoding is implemented to map intent tags and slot tags to a high-dimensional embedding space, explicitly distinguishing and semantically representing classification categories to facilitate understanding of classification tags.
Preferably, the method further comprises:
S310, when decoding to the ith step and i is an odd number, analyzing the intention of the previous word and the corresponding slot information of the previous word based on the semantic vector sequence when predicting the intention of the ([ i/2] +1) th word, wherein i is [0,2n ].
Preferably, the decoding method is as follows:
y[i/2]+1=argmax(softmax(Wyhi)),
Where W y and W o are both trainable parameters, the symbol [ ] represents a rounding operation, Embedded vector representing intent tag y [i/2],/>An embedded representation of the representation slot label o [i/2],
The semantic vector sequence is represented as e= (e [cls],e1,e2,e3,e4,...,en),
The output sequence is denoted y= (y 1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is y i=(y1,y2,y3,...,yn),
The word level slot filling sequence is y s=(o1,o2,o3,...,on).
Preferably, the method further comprises:
S320, when decoding to the ith step and i is even, analyzing the intention information of the current word and the slot information of the previous word when predicting the slot of the [ i/2] word, wherein i is [0,2n ].
Preferably, the decoding method is as follows:
o[i/2]=argmax(softmax(Wohi)),
Where W y and W o are both trainable parameters, the symbol [ ] represents a rounding operation, Embedded vector representing intent tag y [i/2],/>An embedded representation of the representation slot label o [i/2],
The semantic vector sequence is represented as e= (e [cls],e1,e2,e3,e4,...,en),
The output sequence is denoted y= (y 1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is y i=(y1,y2,y3,...,yn),
The word level slot filling sequence is y s=(o1,o2,o3,...,on).
Preferably, the method further comprises:
s330, loss function calculation of word level alternating decoding: the negative log likelihood is used as a loss function to optimize the model parameters.
Preferably, the decoder is trained using a planned sampling mechanism.
Preferably, in the prediction stage, the intent result of the whole sentence is determined by adopting voting results of all word intentions, and the greedy search is used for sampling to obtain a slot prediction result.
According to another aspect of an embodiment of the present invention, there is provided an electronic apparatus including: one or more processors; and a storage device having one or more programs stored thereon, which when executed by the one or more processors, cause the one or more processors to implement the foregoing method.
According to yet another aspect of embodiments of the present invention, there is provided a non-transitory computer readable storage medium having stored thereon executable instructions that, when run on a processor, implement the foregoing method.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below. It will be apparent to those of ordinary skill in the art that the drawings in the following description are exemplary only and that other implementations can be obtained from the extensions of the drawings provided without inventive effort.
The structures, proportions, sizes, etc. shown in the present specification are shown only for the purposes of illustration and description, and are not intended to limit the scope of the invention, which is defined by the claims, so that any structural modifications, changes in proportions, or adjustments of sizes, which do not affect the efficacy or the achievement of the present invention, should fall within the ambit of the technical disclosure.
Fig. 1 is a schematic diagram of a spoken language understanding model proposed by the present invention.
Detailed Description
Other advantages and advantages of the present invention will become apparent to those skilled in the art from the following detailed description, which, by way of illustration, is to be read in connection with certain specific embodiments, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The model of spoken language understanding proposed by the invention is shown in fig. 1, the model uses a pre-trained language model BERT (Bidirectional Encoder Representations from Transformers) as an encoding layer, and word-level sequence decoding is performed by using semantic vector sequence representations obtained by BERT. The method and the device provided by the invention are used for alternately decoding the intention recognition at the word level and the slot filling at the word level, initializing the whole decoding state by using the intention recognition result at the sentence level, and realizing the mutual contribution of the intention recognition and the slot filling from the global and local angles in such a way, thereby improving the effect of the spoken language understanding task.
The specific method comprises the following steps:
Step 1, obtaining a semantic vector sequence coding representation of a pre-training language model BERT:
The BERT model structure is a multi-layer bi-directional transform-based encoder whose inputs include word embedding, sentence embedding, and position embedding. The first marker of the input sequence is always a special classification marker [ cls ], and the final hidden state corresponding to the special marker is used for classification tasks; and a special tag [ SEP ] is used as the last tag of the sequence. The input of the invention is expressed as x= ([ cls ], x 1,x2,x3,x4,...,xn), and the semantic vector sequence code obtained by BERT is expressed as e= (e [cls],e1,e2,e3,e4,...,en).
Step 2, obtaining a sentence level intention recognition result:
According to the final hidden state corresponding to the special symbol [ CLS ] obtained by BERT, sentence-level intention recognition is carried out by using a fully connected neural network (FCN, full ConnectedNetwork) and a Softmax function, sentence-level global intention information y [cls]=softmax(FCN(e[cls]) is finally obtained, and then sentence-level global intention information y [cls] is used for guiding an alternate decoding process of word-level intention recognition and slot filling, so that overall consistency is ensured.
Wherein, the FCN analysis of the fully connected neural network: for the n-1 layer and the n layer, any node of the n-1 layer is connected with all nodes of the n layer. I.e. each node of the n-th layer, when performing the calculation, the input of the activation function is the weighting of all nodes of the n-1 layer, so as to obtain the weight matrix of the full connection layer.
The full connection layer multiplies the weight matrix by the input vector and adds bias, and maps n (- -infinity, + -infinity) real numbers to K (- -infinity, + -infinity) real numbers (fractions); softmax will be K (- ≡, ++ infinity) is mapped to K real numbers (probabilities) of (0, 1), while ensuring that their sum is 1. The method comprises the following steps:
y[CLS]=softmax(FCN(e[cls]))=softmax(WTx+b)
Where x is the input of the fully connected layer, W n X K is the weight b is the bias term, and y [CLS] is the probability of Softmax output.
The probability of splitting into each category is as follows:
y[cls]=softmax(FCN(e[cls]))=softmax(W[CLS]·x+b[CLS])
wherein W [CLS] is a vector formed by the weights of one node of the n-th layer and all nodes of the n-1 th layer in the full connection layer.
Step 3, word level intention recognition and slot filling alternate decoding are carried out:
Word level intent recognition is used to reduce the negative impact on the slot predictions for all words when the intent of the entire sentence is mispredicted.
The invention expresses an output sequence obtained by decoding a semantic vector sequence as y= (y 1,o1,y2,o2,y3,o3,...,yn,on), and the sequence length is 2n; the hidden state of the corresponding decoder is denoted as h= (h 1,h2,h3,...,h2n); where word level intent identifies the decoding result sequence as y i=(y1,y2,y3,...,yn), word level slot filling results sequence as y s=(o1,o2,o3,...,on).
The invention adopts a unidirectional LSTM as a decoder, the LSTM refers to a state transition equation of a long-time memory network, and when decoding to the ith step, i epsilon [0,2n ], the corresponding calculation process of the hidden state of the decoder is as follows:
and 3.1, when i is an odd number, analyzing the intention of the previous word and the slot information corresponding to the previous word when predicting the intention of the ([ i/2] +1) th word. The word-level intent is decoded using the following formula:
y[i/2]+1=argmax(softmax(Wyhi))。
and 3.2, when i is even, analyzing the intention information of the current word and the slot information of the previous word when predicting the slot of the [ i/2] word. The decoding process is as follows:
o[i/2]=argmax(softmax(Wohi)),
Where W y and W o are both trainable parameters, the symbol [ ] represents a rounding operation. Embedded vector representing intent tag y [i/2],/>The embedded representation of the representation slot label o [i/2], mapping the label to the high-dimensional embedded space, to some extent, explicitly distinguishes and semantically represents the classification category, thereby facilitating understanding of the classification label by the model.
Step 3.3, calculating a loss function of word level alternate decoding: the present invention uses negative log-likelihood as a loss function to optimize the model parameters, i.e., L (y) = -log (y). And a design sampling (SS, schedule Sampling) mechanism is used during decoder training to solve the problem of bias in training and predicted information. In planned sampling, i.e. the sampling rate P is varied during training. The training is insufficient at the beginning, the sampling rate P is reduced, the real label is used as input as much as possible, the sampling rate P is increased along with the progress of the training, and the output of the sampling rate P is mostly used as the input of the next prediction. As training proceeds, the sampling rate P becomes larger and the training model is finally the same as the prediction model, so that deviation between training and predicted information is eliminated.
Step 4, the result acquisition of the prediction stage intention recognition and the slot filling is carried out:
Because the word-level intention recognition mode is used, the intention result of the final whole sentence is determined by adopting the voting result of all word intents; the greedy search is used for sampling to obtain a slot prediction result, such as Dijkstra algorithm, prim algorithm, kruskal algorithm and the like, when solving, the optimal solution is guaranteed each time, the coverage area of each sampling is maximized, so that the best or optimal (i.e. most favorable) selection is adopted in each step of selection, and the result is expected to be the best or optimal. Furthermore, the use of Conditional Random Fields (CRF) may have negative effects.
According to the method, the intention recognition is modeled as a word level classification task, the interactive relation among sentence level intention recognition, word level intention recognition and word level slot filling is explicitly modeled through a word level alternate decoding framework, and the potential performance of spoken language understanding is improved through mutual assistance of the three. Therefore, the problem that the co-occurrence relation between the groove and the intention cannot be fully utilized in the conventional method which takes the intention recognition as a sentence-level classification task and takes the groove filling as a sequence labeling task is remarkably improved.
The alternative decoding-based spoken language understanding method provided by the embodiment of the invention can be realized in the form of a software functional module and sold or used as an independent product, and can be stored in a computer readable storage medium. Based on this understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art or in a part of the technical solution, in the form of a software product stored in a storage medium, comprising several instructions for causing a computer device (which may be a personal computer, a server, a network device, etc.) to perform all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a usb disk, a removable hard disk, a read-only memory (ROM), a random access memory (RAM, random Access Memory), a magnetic disk, or an optical disk, or other various media capable of storing program codes.
While the invention has been described in detail in the foregoing general description and specific examples, it will be apparent to those skilled in the art that modifications and improvements can be made thereto. Accordingly, such modifications or improvements may be made without departing from the spirit of the invention and are intended to be within the scope of the invention as claimed.

Claims (8)

1. A method of spoken language understanding based on alternate decoding, comprising:
s100, obtaining a semantic vector sequence of a pre-training language model BERT based on an input sequence, and taking a first mark of the input sequence as a classification mark;
S200, obtaining a corresponding final hidden state according to the classification mark, and carrying out sentence-level intention recognition by using a fully connected neural network and a Softmax function based on the final hidden state to obtain sentence-level global intention information;
S300, word level intention recognition and slot filling alternate decoding are carried out based on the semantic vector sequence and the sentence level global intention information; wherein the alternately decoding includes: obtaining a corresponding decoder hiding state based on the semantic vector sequence and the sentence-level global intention information, decoding the decoder hiding state to obtain an output sequence, wherein the length of the output sequence is 2n, and obtaining a word-level intention recognition sequence and a word-level slot filling sequence based on the output sequence, wherein the decoder is unidirectional LSTM;
s310, when decoding to the ith step and i is an odd number, analyzing the intention of the previous word and the slot position information corresponding to the previous word based on the semantic vector sequence when predicting the intention of the [ i/2] +1 th word, wherein i is [0,2n ];
s320, when decoding to the ith step and i is even, analyzing the intention information of the current word and the slot information of the previous word when predicting the slot of the [ i/2] word, wherein i is [0,2n ];
S330, calculating a loss function of word level alternate decoding: optimizing model parameters using negative log-likelihood as a loss function;
S400, obtaining results of intention recognition and slot filling based on the results of the alternate decoding.
2. The alternative decoding-based spoken language understanding method of claim 1, wherein the decoding is implemented to map intent tags and slot tags to a high-dimensional embedding space, explicitly distinguishing and semantically representing classification categories.
3. The method for speech understanding based on alternate decoding according to claim 1,
The decoding method comprises the following steps:
y[i/2]+1=argmax(softmax(Wyhi)),
where W y is a trainable parameter, the symbol [ ] represents a rounding operation, The embedded vector representing the intent tag y [i/2],An embedded representation of the representation slot label o [i/2],
The semantic vector sequence is expressed as e= (e [cls],e1,e2,e3,e4,...,en), [ cls ] is a special class label of the input sequence,
The output sequence is denoted y= (y 1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is y i=(y1,y2,y3,...,yn),
The word level slot filling sequence is y s=(o1,o2,o3,...,on).
4. The alternative decoding-based spoken language understanding method of claim 1, wherein the decoding method is as follows:
o[i/2]=argmax(softmax(Wohi)),
Wherein W o is a trainable parameter,
The semantic vector sequence is represented as e= (e [cls],e1,e2,e3,e4,...,en),
The output sequence is denoted y= (y 1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is y i=(y1,y2,y3,...,yn),
The word level slot filling sequence is y s=(o1,o2,o3,...,on);
the symbol [ ] indicates a rounding operation, An embedded vector representing the intent tag y [i/2], [ cls ] is a special class label of the input sequence.
5. The alternative decoding-based spoken language understanding method of claim 1, wherein the decoder is trained using a planned sampling mechanism.
6. The alternative decoding-based spoken language understanding method of any one of claims 1-5,
In the prediction stage, the intention result of the whole sentence is determined by adopting the voting result of all word intentions, and the greedy search is used for sampling to obtain the slot prediction result.
7. An electronic device, comprising: one or more processors; storage means having stored thereon one or more programs which, when executed by the one or more processors, cause the one or more processors to implement the method of any of claims 1-6.
8. A non-transitory computer readable storage medium having stored thereon executable instructions that, when run on a processor, implement the method according to any of claims 1-6.
CN202011045822.3A 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium Active CN112183062B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011045822.3A CN112183062B (en) 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011045822.3A CN112183062B (en) 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112183062A CN112183062A (en) 2021-01-05
CN112183062B true CN112183062B (en) 2024-04-19

Family

ID=73945683

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011045822.3A Active CN112183062B (en) 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112183062B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113204952B (en) * 2021-03-26 2023-09-15 南京邮电大学 Multi-intention and semantic slot joint identification method based on cluster pre-analysis
CN114021582B (en) * 2021-12-30 2022-04-01 深圳市北科瑞声科技股份有限公司 Spoken language understanding method, device, equipment and storage medium combined with voice information

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109858030A (en) * 2019-02-11 2019-06-07 北京邮电大学 The Task dialogue of two-way intention slot value crosscorrelation understands system and method
CN110188342A (en) * 2019-04-19 2019-08-30 杭州电子科技大学 A kind of speech understanding method of knowledge based map and semantic diagram technology
CN110838288A (en) * 2019-11-26 2020-02-25 杭州博拉哲科技有限公司 Voice interaction method and system and dialogue equipment
CN111177341A (en) * 2019-12-11 2020-05-19 江苏艾佳家居用品有限公司 End-to-end ID + SF model-based user conversation demand extraction method and system

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10431207B2 (en) * 2018-02-06 2019-10-01 Robert Bosch Gmbh Methods and systems for intent detection and slot filling in spoken dialogue systems

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109858030A (en) * 2019-02-11 2019-06-07 北京邮电大学 The Task dialogue of two-way intention slot value crosscorrelation understands system and method
CN110188342A (en) * 2019-04-19 2019-08-30 杭州电子科技大学 A kind of speech understanding method of knowledge based map and semantic diagram technology
CN110838288A (en) * 2019-11-26 2020-02-25 杭州博拉哲科技有限公司 Voice interaction method and system and dialogue equipment
CN111177341A (en) * 2019-12-11 2020-05-19 江苏艾佳家居用品有限公司 End-to-end ID + SF model-based user conversation demand extraction method and system

Also Published As

Publication number Publication date
CN112183062A (en) 2021-01-05

Similar Documents

Publication Publication Date Title
CN111783462B (en) Chinese named entity recognition model and method based on double neural network fusion
Yao et al. An improved LSTM structure for natural language processing
Kalchbrenner et al. Recurrent convolutional neural networks for discourse compositionality
JP7087938B2 (en) Question generator, question generation method and program
CN108962224B (en) Joint modeling method, dialogue method and system for spoken language understanding and language model
CN110516253B (en) Chinese spoken language semantic understanding method and system
CN107870902A (en) Neural machine translation system
CN109074517B (en) Global normalized neural network
CN110866401A (en) Chinese electronic medical record named entity identification method and system based on attention mechanism
CN111062217B (en) Language information processing method and device, storage medium and electronic equipment
CN112183061B (en) Multi-intention spoken language understanding method, electronic equipment and storage medium
US20210248471A1 (en) Method and apparatus for creating dialogue, and storage medium
WO2019235103A1 (en) Question generation device, question generation method, and program
US11694677B2 (en) Decoding method and apparatus in artificial neural network for speech recognition
CN112183062B (en) Spoken language understanding method based on alternate decoding, electronic equipment and storage medium
CN110678882A (en) Selecting answer spans from electronic documents using machine learning
CN116204674B (en) Image description method based on visual concept word association structural modeling
CN111737974A (en) Semantic abstract representation method and device for statement
CN115186147B (en) Dialogue content generation method and device, storage medium and terminal
CN111145914B (en) Method and device for determining text entity of lung cancer clinical disease seed bank
CN114912450B (en) Information generation method and device, training method, electronic device and storage medium
JP7342971B2 (en) Dialogue processing device, learning device, dialogue processing method, learning method and program
Park et al. Natural language generation using dependency tree decoding for spoken dialog systems
CN110633473A (en) Implicit discourse relation identification method and system based on conditional random field
CN112364659B (en) Automatic identification method and device for unsupervised semantic representation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant