CN112183062A - Spoken language understanding method based on alternate decoding, electronic equipment and storage medium - Google Patents

Spoken language understanding method based on alternate decoding, electronic equipment and storage medium Download PDF

Info

Publication number
CN112183062A
CN112183062A CN202011045822.3A CN202011045822A CN112183062A CN 112183062 A CN112183062 A CN 112183062A CN 202011045822 A CN202011045822 A CN 202011045822A CN 112183062 A CN112183062 A CN 112183062A
Authority
CN
China
Prior art keywords
word
decoding
sequence
level
intention
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011045822.3A
Other languages
Chinese (zh)
Other versions
CN112183062B (en
Inventor
刘广灿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Unisound Intelligent Technology Co Ltd
Xiamen Yunzhixin Intelligent Technology Co Ltd
Original Assignee
Unisound Intelligent Technology Co Ltd
Xiamen Yunzhixin Intelligent Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Unisound Intelligent Technology Co Ltd, Xiamen Yunzhixin Intelligent Technology Co Ltd filed Critical Unisound Intelligent Technology Co Ltd
Priority to CN202011045822.3A priority Critical patent/CN112183062B/en
Publication of CN112183062A publication Critical patent/CN112183062A/en
Application granted granted Critical
Publication of CN112183062B publication Critical patent/CN112183062B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/126Character encoding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Databases & Information Systems (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a spoken language understanding method based on alternate decoding, electronic equipment and a storage medium.

Description

Spoken language understanding method based on alternate decoding, electronic equipment and storage medium
Technical Field
The invention relates to the technical field of man-machine conversation systems, in particular to a spoken language understanding method based on alternate decoding, electronic equipment and a storage medium.
Background
Spoken language understanding mainly includes two subtasks: intent recognition (Intent Detection) and Slot Filling (Slot Filling). The two tasks of intent recognition and slot filling are not independent of each other, slot filling is highly dependent on the result of intent recognition, and slot filling may also facilitate intent recognition. In the prior art, two tasks are jointly modeled to fully utilize knowledge information common to the two tasks, a Multi-task (Multi-task) framework is generally adopted, two subtasks share a coding layer, and then loss functions of the two parts are added to model.
The existing method for using a multi-task frame and adopting a shared coding layer and loss function addition is an implicit joint modeling method, and the method does not explicitly model the interaction relationship between intention identification and slot position filling two subtasks; the Slot-Gated method and the SF-ID method are also proposed for preliminary exploration, but the existing models cannot fully utilize the co-occurrence relationship between the slots and the intentions, so that the potential performances of the slots and the intentions are limited.
Disclosure of Invention
In order to solve the problems, the invention proposes that the intention recognition at the word level and the slot filling at the word level are alternately decoded, and the whole decoding state is initialized by using the intention recognition result at the sentence level, so that the intention recognition and the slot filling can mutually contribute from the global and local angles, thereby improving the effect of the spoken language understanding task.
According to an aspect of an embodiment of the present invention, there is provided a spoken language understanding method based on alternating decoding, including:
s100, obtaining a semantic vector sequence of a pre-training language model BERT based on an input sequence, and marking a first label of the input sequence as a classification label;
s200, obtaining a corresponding final hidden state according to the classification marks, and performing sentence level intention recognition by using a fully-connected neural network and a Softmax function based on the final hidden state to obtain sentence level global intention information;
s300, performing word level intention recognition and slot filling alternate decoding based on the semantic vector sequence and the sentence level global intention information;
s400 obtains results of intent recognition and slot filling based on the results of the alternating decoding.
Preferably, the alternating decoding comprises: and obtaining a corresponding decoder hidden state based on the semantic vector sequence and the sentence level global intention information, decoding the decoder hidden state to obtain an output sequence, wherein the length of the output sequence is 2n, and obtaining a word level intention identification sequence and a word level slot filling sequence based on the output sequence, wherein the decoder is a unidirectional LSTM.
Preferably, the decoding is implemented to map intent tags and slot tags into a high-dimensional embedding space, explicitly distinguishing and semantically representing classification categories to facilitate understanding of the classification tags.
Preferably, the method further comprises the following steps:
s310, when the ith step is decoded and i is an odd number, when the intention of the ([ i/2] +1) th word is predicted, the intention of the previous word and the slot position information corresponding to the previous word are analyzed based on the semantic vector sequence, wherein i belongs to [0,2n ].
Preferably, the decoding method is as follows:
Figure BDA0002707924630000021
y[i/2]+1=argmax(softmax(Wy i)),
wherein, WyAnd WoAre trainable parameters and symbols]It is shown that the rounding operation is performed,
Figure BDA0002707924630000031
label y for indicating intention[i/2]The embedded vector of (a) is embedded,
Figure BDA0002707924630000032
indicating slot label o[i/2]Is to be used to represent the embedded representation of,
the semantic vector sequence is represented as e ═ (e)[cls],e1,e2,e3,e4,...,en),
The output sequence is represented as y ═ y (y)1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is yi=(y1,y2,y3,...,yn),
The word-level slot fill sequence is ys=(o1,o2,o3,...,on)。
Preferably, the method further comprises the following steps:
s320, when the ith step is decoded and i is an even number, when the slot position of the [ i/2] th word is predicted, analyzing intention information of the current word and slot position information of the previous word, wherein i belongs to [0,2n ].
Preferably, the decoding method is as follows:
Figure BDA0002707924630000033
o[i/2]=argmax(softmax(Wo i)),
wherein, WyAnd WoAre trainable parameters and symbols]It is shown that the rounding operation is performed,
Figure BDA0002707924630000034
label y for indicating intention[i/2]The embedded vector of (a) is embedded,
Figure BDA0002707924630000035
indicating slot label o[i/2]Is to be used to represent the embedded representation of,
the semantic vector sequence is represented as e ═ (e)[cls],e1,e2,e3,e4,...,en),
The output sequence is represented as y ═ y (y)1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is yi=(y1,y2,y3,...,yn),
The word-level slot fill sequence is ys=(o1,o2,o3,...,on)。
Preferably, the method further comprises the following steps:
s330 loss function calculation for word-level alternating decoding: the model parameters are optimized using negative log-likelihood as a loss function.
Preferably, the decoder is trained using a planned sampling mechanism.
Preferably, in the prediction stage, the voting results of all word intentions are adopted for determining the intention result of the whole word, and sampling is performed by using greedy search to obtain a slot prediction result.
According to another aspect of an embodiment of the present invention, there is provided an electronic apparatus including: one or more processors; a storage device having one or more programs stored thereon which, when executed by the one or more processors, cause the one or more processors to implement the aforementioned methods.
According to yet another aspect of embodiments of the present invention, there is provided a non-transitory computer-readable storage medium having stored thereon executable instructions that, when executed on a processor, implement the foregoing method.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below. It should be apparent that the drawings in the following description are merely exemplary, and that other embodiments can be derived from the drawings provided by those of ordinary skill in the art without inventive effort.
The structures, ratios, sizes, and the like shown in the present specification are only used for matching with the contents disclosed in the specification, so as to be understood and read by those skilled in the art, and are not used to limit the conditions that the present invention can be implemented, so that the present invention has no technical significance, and any structural modifications, changes in the ratio relationship, or adjustments of the sizes, without affecting the effects and the achievable by the present invention, should still fall within the range that the technical contents disclosed in the present invention can cover.
Fig. 1 is a schematic diagram of a spoken language understanding model provided by the present invention.
Detailed Description
The present invention is described in terms of particular embodiments, other advantages and features of the invention will become apparent to those skilled in the art from the following disclosure, and it is to be understood that the described embodiments are merely exemplary of the invention and that it is not intended to limit the invention to the particular embodiments disclosed. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The model for spoken language understanding proposed by the present invention is shown in fig. 1, and the model uses a pre-training language model BERT (bidirectional Encoder expressions from transforms) as a coding layer, and performs word-level sequence decoding using semantic vector sequence representation obtained by BERT. The invention provides that the intention recognition at the word level and the slot filling at the word level are alternately decoded, and the whole decoding state is initialized by using the intention recognition result at the sentence level, so that the intention recognition and the slot filling can mutually contribute from the global and local angles, and the effect of the spoken language understanding task is improved.
The specific method comprises the following steps:
step 1, obtaining semantic vector sequence coding representation of a pre-training language model BERT:
the BERT model structure is a multi-layer bidirectional-based Transformer encoder, and the input of the BERT model structure comprises three parts of word embedding, sentence embedding and position embedding. The first marker of the input sequence is always a special class marker [ CLS]The final hidden state corresponding to the special mark is used for a classification task; and use a special mark SEP]As the last marker of the sequence. The input to the present invention is expressed as x ═ ([ cls ]],x1,x2,x3,x4,...,xn) The BERT derived semantic vector sequence code is expressed as e ═ (e)[cls],e1,e2,e3,e4,...,en)。
And step 2, obtaining a sentence level intention recognition result:
special symbols [ CLS ] obtained from BERT]And corresponding to the final hidden state, performing sentence-level intention recognition by using a fully Connected neural Network (FCN) and a Softmax function, and finally obtaining sentence-level global intention information y[cls]=softmax(FCN(e[cls]) And then use sentence-level global intention information y)[cls]And guiding the alternate decoding process of word-level intention identification and slot filling to ensure the overall consistency.
Wherein, the full-connection neural network FCN analyzes: for the n-1 layer and the n layer, any node of the n-1 layer is connected with all nodes of the n layer. That is, when each node of the nth layer performs calculation, the input of the activation function is the weight of all nodes of the n-1 layer, so as to obtain the weight matrix of the fully-connected layer.
The fully-connected layer multiplies the weight matrix by the input vector, adds an offset, and maps n (— infinity, + infinity) real numbers into K (— infinity, + infinity) real numbers (fractions); softmax maps K (— infinity, + ∞) real numbers to K (0,1) real numbers (probabilities), while ensuring that their sum is 1. The method comprises the following specific steps:
y[CLS]=softmax(FCN(e[cls]))=softmax(WTx+b)
where x is the input of the full link layer, WnxK is the weight b is the bias term, y[CLS]Probability of output of Softmax.
The probability of splitting into each class is as follows:
y[cls]=softmax(FCN(e[cls]))=softmax(W[CLS]·x+b[CLS])
wherein, W[CLS]And a vector is formed by the weight of one node of the n-th layer in the full connection layer and all nodes of the n-1-th layer.
And 3, performing word level intention identification and slot filling alternate decoding:
word-level intent recognition is used to reduce the negative impact of slot prediction on all words when the intent of the entire sentence is mispredicted.
The invention expresses the output sequence obtained by decoding the semantic vector sequence as y ═ y1,o1,y2,o2,y3,o3,...,yn,on) The sequence length is 2 n; the hidden state of the corresponding decoder is denoted as ═ c (c)1, 2, 3,..., 2n) (ii) a Wherein the word-level intent recognition decoding result sequence is yi=(y1,y2,y3,...,yn) The result sequence of word-level slot filling is ys=(o1,o2,o3,...,on)。
The invention adopts a unidirectional LSTM as a decoder, the LSTM refers to a state transfer equation of a long-time memory network, when the decoding reaches the ith step, i belongs to [0,2n ], and the calculation process of the hidden state of the corresponding decoder is as follows:
and 3.1, when i is an odd number, when the intention of the ([ i/2] +1) th word is predicted, analyzing the intention of the previous word and the slot position information corresponding to the previous word. Decoding the word-level intent using the following formula:
Figure BDA0002707924630000071
y[i/2]+1=argmax(softmax(Wy i))。
and 3.2, when i is an even number, analyzing intention information of the current word and slot position information of the previous word when the slot position of the [ i/2] th word is predicted. The decoding process is as follows:
Figure BDA0002707924630000072
o[i/2]=argmax(softmax(Wo i)),
wherein WyAnd WoAre trainable parameters and symbols]Indicating a rounding operation.
Figure BDA0002707924630000073
Label y for indicating intention[i/2]The embedded vector of (a) is embedded,
Figure BDA0002707924630000074
indicating slot label o[i/2]Mapping tags into a high-dimensional embedding space to a certain extent explicitly distinguishes and semantically represents classification categories, thereby facilitating the understanding of classification tags by the model.
And 3.3, calculating a loss function of the word-level alternative decoding: the present invention uses negative log-likelihood as a loss function to optimize the model parameters, i.e., l (y) -log (y). And a plan Sampling (SS) mechanism is used in the decoder training process to solve the problem that the training and prediction information is biased. In planning sampling, i.e. the sampling rate P is varied during training. The training is insufficient at the beginning, the sampling rate P is reduced, real labels are used as input as much as possible, the sampling rate P is increased along with the training, and the output of the sampling rate P is mostly used as the input of the next prediction. As training progresses, the sampling rate P becomes larger and larger, and the training model is finally the same as the prediction model, so that the deviation between the training information and the prediction information is eliminated.
And 4, acquiring results of intention identification and slot filling in the prediction stage:
because the word level intention recognition mode is used, the final intention result of the whole sentence is decided by adopting the voting result of all word intentions; sampling is performed by greedy search to obtain a slot prediction result, such as Dijkstra algorithm, Prim algorithm, Kruskal algorithm and the like, and when solving is performed, an optimal solution is guaranteed each time, so that the coverage area of each time is maximized, and the best or optimal (i.e., most favorable) selection is adopted in each step of selection, so that the result is expected to be the best or optimal. Furthermore, the use of Conditional Random Fields (CRF) may have negative effects.
The method models the intention recognition into a word-level classification task, explicitly models the interactive relation among the sentence-level intention recognition, the word-level intention recognition and the word-level slot position through a word-level alternate decoding frame, and improves the potential performance of spoken language understanding through mutual assistance of the three. Therefore, the problem that the conventional method treats intention recognition as a sentence-level classification task and fills the slot position as a sequence labeling task and cannot fully utilize the co-occurrence relationship between the slot and the intention is solved.
The spoken language understanding method based on the alternative decoding provided by the embodiment of the invention can be realized in the form of a software functional module, sold or used as an independent product, and can be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
Although the invention has been described in detail above with reference to a general description and specific examples, it will be apparent to one skilled in the art that modifications or improvements may be made thereto based on the invention. Accordingly, such modifications and improvements are intended to be within the scope of the invention as claimed.

Claims (12)

1. A method for spoken language understanding based on alternating decoding, comprising:
s100, obtaining a semantic vector sequence of a pre-training language model BERT based on an input sequence, and marking a first label of the input sequence as a classification label;
s200, obtaining a corresponding final hidden state according to the classification marks, and performing sentence level intention recognition by using a fully-connected neural network and a Softmax function based on the final hidden state to obtain sentence level global intention information;
s300, performing word level intention recognition and slot filling alternate decoding based on the semantic vector sequence and the sentence level global intention information;
s400 obtains results of intent recognition and slot filling based on the results of the alternating decoding.
2. The alternate decoding-based spoken language understanding method of claim 1, wherein the alternate decoding comprises:
deriving a corresponding decoder hidden state based on the semantic vector sequence and the sentence-level global intent information,
decoding the hidden state of the decoder to obtain an output sequence, wherein the length of the output sequence is 2n, and obtaining a word-level intention identification sequence and a word-level slot filling sequence based on the output sequence, wherein the decoder is a unidirectional LSTM.
3. The spoken language understanding method based on alternating decoding according to claim 2,
the decoding is implemented to map intent tags and slot tags into a high-dimensional embedding space, explicitly distinguishing and semantically representing classification categories to facilitate understanding of the classification tags.
4. The spoken language understanding method based on alternating decoding according to claim 2, further comprising:
s310, when the ith step is decoded and i is an odd number, when the intention of the ([ i/2] +1) th word is predicted, the intention of the previous word and the slot position information corresponding to the previous word are analyzed based on the semantic vector sequence, wherein i belongs to [0,2n ].
5. The spoken language understanding method based on alternating decoding according to claim 4, characterized in that the decoding method is as follows:
Figure FDA0002707924620000021
y[i/2]+1=argmax(softmax(Wyi)),
wherein, WyAnd WoAre trainable parameters and symbols]It is shown that the rounding operation is performed,
Figure FDA0002707924620000022
label y for indicating intention[i/2]The embedded vector of (a) is embedded,
Figure FDA0002707924620000023
indicating slot label o[i/2]Is to be used to represent the embedded representation of,
the semantic vector sequence is represented as e ═ (e)[cls],e1,e2,e3,e4,...,en),
The output sequence is represented as y ═ y (y)1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is yi=(y1,y2,y3,...,yn),
The word-level slot fill sequence is ys=(o1,o2,o3,...,on)。
6. The spoken language understanding method based on alternating decoding according to claim 4, further comprising:
s320, when the ith step is decoded and i is an even number, when the slot position of the [ i/2] th word is predicted, analyzing intention information of the current word and slot position information of the previous word, wherein i belongs to [0,2n ].
7. The spoken language understanding method based on alternating decoding according to claim 6,
the decoding method comprises the following steps:
Figure FDA0002707924620000024
o[i/2]=argmax(softmax(Woi)),
wherein, WyAnd WoAre trainable parameters and symbols]It is shown that the rounding operation is performed,
Figure FDA0002707924620000025
label y for indicating intention[i/2]The embedded vector of (a) is embedded,
Figure FDA0002707924620000031
indicating slot label o[i/2]Is to be used to represent the embedded representation of,
the semantic vector sequence is represented as e ═ (e)[cls],e1,e2,e3,e4,...,en),
The output sequence is represented as y ═ y (y)1,o1,y2,o2,y3,o3,...,yn,on),
The word-level intent recognition sequence is yi=(y1,y2,y3,...,yn),
The word-level slot fill sequence is ys=(o1,o2,o3,...,on)。
8. The spoken language understanding method based on alternating decoding according to claim 6, further comprising:
s330 loss function calculation for word-level alternating decoding: the model parameters are optimized using negative log-likelihood as a loss function.
9. The spoken language understanding method based on alternating decoding according to claim 8,
the decoder is trained using a planned sampling mechanism.
10. The method for spoken language understanding based on alternating decoding according to one of claims 1 to 9,
in the prediction stage, the whole sentence intention result is determined by adopting the voting results of all word intentions, and sampling is carried out by greedy search to obtain a slot position prediction result.
11. An electronic device, comprising: one or more processors; storage means having one or more programs stored thereon which, when executed by the one or more processors, cause the one or more processors to implement the method of any of claims 1-10.
12. A non-transitory computer readable storage medium having stored thereon executable instructions which, when executed on a processor, implement the method of any one of claims 1-10.
CN202011045822.3A 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium Active CN112183062B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011045822.3A CN112183062B (en) 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011045822.3A CN112183062B (en) 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112183062A true CN112183062A (en) 2021-01-05
CN112183062B CN112183062B (en) 2024-04-19

Family

ID=73945683

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011045822.3A Active CN112183062B (en) 2020-09-28 2020-09-28 Spoken language understanding method based on alternate decoding, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112183062B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114021582A (en) * 2021-12-30 2022-02-08 深圳市北科瑞声科技股份有限公司 Spoken language understanding method, device, equipment and storage medium combined with voice information
WO2022198750A1 (en) * 2021-03-26 2022-09-29 南京邮电大学 Semantic recognition method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109858030A (en) * 2019-02-11 2019-06-07 北京邮电大学 The Task dialogue of two-way intention slot value crosscorrelation understands system and method
US20190244603A1 (en) * 2018-02-06 2019-08-08 Robert Bosch Gmbh Methods and Systems for Intent Detection and Slot Filling in Spoken Dialogue Systems
CN110188342A (en) * 2019-04-19 2019-08-30 杭州电子科技大学 A kind of speech understanding method of knowledge based map and semantic diagram technology
CN110838288A (en) * 2019-11-26 2020-02-25 杭州博拉哲科技有限公司 Voice interaction method and system and dialogue equipment
CN111177341A (en) * 2019-12-11 2020-05-19 江苏艾佳家居用品有限公司 End-to-end ID + SF model-based user conversation demand extraction method and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190244603A1 (en) * 2018-02-06 2019-08-08 Robert Bosch Gmbh Methods and Systems for Intent Detection and Slot Filling in Spoken Dialogue Systems
CN109858030A (en) * 2019-02-11 2019-06-07 北京邮电大学 The Task dialogue of two-way intention slot value crosscorrelation understands system and method
CN110188342A (en) * 2019-04-19 2019-08-30 杭州电子科技大学 A kind of speech understanding method of knowledge based map and semantic diagram technology
CN110838288A (en) * 2019-11-26 2020-02-25 杭州博拉哲科技有限公司 Voice interaction method and system and dialogue equipment
CN111177341A (en) * 2019-12-11 2020-05-19 江苏艾佳家居用品有限公司 End-to-end ID + SF model-based user conversation demand extraction method and system

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022198750A1 (en) * 2021-03-26 2022-09-29 南京邮电大学 Semantic recognition method
JP7370033B2 (en) 2021-03-26 2023-10-27 南京郵電大学 Semantic recognition method
CN114021582A (en) * 2021-12-30 2022-02-08 深圳市北科瑞声科技股份有限公司 Spoken language understanding method, device, equipment and storage medium combined with voice information
CN114021582B (en) * 2021-12-30 2022-04-01 深圳市北科瑞声科技股份有限公司 Spoken language understanding method, device, equipment and storage medium combined with voice information

Also Published As

Publication number Publication date
CN112183062B (en) 2024-04-19

Similar Documents

Publication Publication Date Title
Yao et al. An improved LSTM structure for natural language processing
De Mori Spoken language understanding: A survey
CN108962224B (en) Joint modeling method, dialogue method and system for spoken language understanding and language model
Chen et al. Syntax or semantics? knowledge-guided joint semantic frame parsing
CN110516253B (en) Chinese spoken language semantic understanding method and system
CN111344779A (en) Training and/or determining responsive actions for natural language input using coder models
US20210232948A1 (en) Question responding apparatus, question responding method and program
CN111062217B (en) Language information processing method and device, storage medium and electronic equipment
CN110866401A (en) Chinese electronic medical record named entity identification method and system based on attention mechanism
CN109214006B (en) Natural language reasoning method for image enhanced hierarchical semantic representation
CN112183061B (en) Multi-intention spoken language understanding method, electronic equipment and storage medium
CN112037773B (en) N-optimal spoken language semantic recognition method and device and electronic equipment
CN111144110A (en) Pinyin marking method, device, server and storage medium
CN111738016A (en) Multi-intention recognition method and related equipment
WO2019235103A1 (en) Question generation device, question generation method, and program
CN111581970B (en) Text recognition method, device and storage medium for network context
CN113743099B (en) System, method, medium and terminal for extracting terms based on self-attention mechanism
CN115186147B (en) Dialogue content generation method and device, storage medium and terminal
CN112183062A (en) Spoken language understanding method based on alternate decoding, electronic equipment and storage medium
Hori et al. Statistical dialog management applied to WFST-based dialog systems
CN115964459B (en) Multi-hop reasoning question-answering method and system based on food safety cognition spectrum
CN113326367B (en) Task type dialogue method and system based on end-to-end text generation
CN114528387A (en) Deep learning conversation strategy model construction method and system based on conversation flow bootstrap
CN112364659B (en) Automatic identification method and device for unsupervised semantic representation
CN112307179A (en) Text matching method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant