WO2020140487A1 - 用于智能设备的人机交互语音识别方法及系统 - Google Patents
用于智能设备的人机交互语音识别方法及系统 Download PDFInfo
- Publication number
- WO2020140487A1 WO2020140487A1 PCT/CN2019/106778 CN2019106778W WO2020140487A1 WO 2020140487 A1 WO2020140487 A1 WO 2020140487A1 CN 2019106778 W CN2019106778 W CN 2019106778W WO 2020140487 A1 WO2020140487 A1 WO 2020140487A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- slot
- vector
- context
- intent
- word sequence
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 57
- 230000003993 interaction Effects 0.000 title abstract description 6
- 239000013598 vector Substances 0.000 claims abstract description 200
- 230000011218 segmentation Effects 0.000 claims abstract description 35
- 238000005457 optimization Methods 0.000 claims abstract description 11
- 238000004364 calculation method Methods 0.000 claims description 28
- 230000002452 interceptive effect Effects 0.000 claims description 19
- 239000011159 matrix material Substances 0.000 claims description 17
- 230000004913 activation Effects 0.000 claims description 12
- 230000002457 bidirectional effect Effects 0.000 claims description 11
- 238000006243 chemical reaction Methods 0.000 claims description 10
- 238000010586 diagram Methods 0.000 claims description 7
- 238000013139 quantization Methods 0.000 claims description 3
- 238000001514 detection method Methods 0.000 abstract 2
- 230000009286 beneficial effect Effects 0.000 description 5
- 238000010276 construction Methods 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 3
- 241000238558 Eucarida Species 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
Definitions
- the present invention relates to the field of voice recognition technology, and in particular to a human-machine interactive voice recognition method and system for smart devices.
- the CNN+knowledge representation classifier For intent recognition, you can abstract it as a classification problem, and then use the CNN+knowledge representation classifier to train the intent recognition model. In addition to embedding the user's speech problem into the word in the intent recognition model, it also introduces the semantics of knowledge Representation to increase the generalization ability of the presentation layer, but in practical applications, it is found that the model has a defect of slot information filling deviation, which affects the accuracy of the intention recognition model.
- slot filling the essence is to formalize sentence sequences into labeled sequences. There are many methods for labeling sequences, such as hidden Markov model or conditional random field model, but these slot filling models are used in specific applications.
- An object of the present invention is to provide a method and system for human-machine interactive voice recognition for smart devices, by joint optimization training of intention recognition and slot filling, to improve the accuracy of voice recognition.
- one aspect of the present invention provides a human-machine interactive voice recognition method for a smart device, including:
- the user's speech problem is segmented to obtain the original word sequence, and the original word sequence is vectorized through the embedding process;
- slot gate g to perform splicing processing on the slot context vector c i S and the intention context vector c I , and converting and representing the slot label model y i S through the slot gate g;
- An intent prediction model y I and the converted slot label model y i S are jointly optimized to construct an objective function, and intent recognition is performed on the user's speech problem based on the objective function.
- the user's speech problem is segmented to obtain the original word sequence
- the method of vectorizing the original word sequence through the embedding process includes:
- the original word sequence is subjected to word embedding to realize the vectorized representation of each word segmentation in the original word sequence.
- the vector of each sub-word is calculated implicit state vector h i and c i S slots context of the vector, the implied by the state vector h i and the vector c i S slots context obtain slot weighting
- the methods of the label model y i S include:
- the implicit state vector hT and the intention context vector c I of the original word sequence represented by the vectorization are calculated, and the intention prediction model y I is obtained by weighting the implicit state vector hT and the intention context vector c I Methods include:
- the slot gate g is used to splice the slot context vector c i S and the intention context vector c I , and the method of converting and representing the slot label model y i S through the slot gate g includes:
- v represents the weight vector obtained by training
- W represents the weight matrix obtained by training
- the objective function constructed by jointly optimizing the intent prediction model y I and the converted slot label model y i S is:
- X) represents the conditional probability of slot filling and intent prediction output at a given original word sequence, where X is the original word sequence represented by vectorization.
- the method for intent recognition of the user's voice problem based on the objective function includes:
- the word segmentation with the highest probability value is selected and recognized as the intention of the user's voice problem.
- the human-machine interactive voice recognition method for smart devices provided by the present invention has the following beneficial effects:
- the acquired user voice question is first converted into recognized text, and the original word sequence is generated based on the recognizable text segmentation processing, and then the original word sequence is word embedded After processing, the vectorized representation is implemented. After that, the slot label model y i S and the intent prediction model y I are constructed based on the original word sequence of the vector representation. The construction step of the slot label model y i S is through calculation.
- y I is to calculate the implicit state vector hT and the intention context vector c I of the original word sequence, and then weight the implicit state vector hT and the intention context vector c I to obtain the intention prediction model y I.
- Integrating the intent prediction model y I and the slot label model y i S we add a decoder layer to the existing encoder-decoder architecture to construct the intent prediction model y I , and introduce the slot gate g to the slot context vector c i S and the intent context vector c I are stitched together. Finally, the intent prediction model y I and the converted slot label model y i S are jointly optimized to obtain an objective function, and the objective function is used to sequentially obtain the corresponding words in the original word sequence. Intentional conditional probability, and then select the word segmentation with the largest probability value to recognize the user's voice problem intent, which ensures the accuracy of voice recognition.
- Another aspect of the present invention provides a human-machine interactive voice recognition system for smart devices, which is applied to the human-machine interactive voice recognition method for smart devices described in the above technical solution, the system includes:
- the word segmentation processing unit is used to segment the user's speech problem to obtain the original word sequence, and vectorize the original word sequence through embedding processing;
- First calculating means for calculating a hidden state of the sub-word vector and the vector h i c i S slots context of the vector, the implied by the state vector h i and the context slots weighting vector c i S Obtain the slot label model y i S ;
- the second calculation unit is used to calculate the implicit state vector hT and the intention context vector c I of the original word sequence represented by the vectorization, and obtain the intention prediction by weighting the implicit state vector hT and the intention context vector c I Model y I ;
- a model conversion unit used to perform slotting processing on the slot context vector c i S and the intention context vector c I using a slot gate g, and convert and represent the slot label model y i S through the slot gate g;
- the joint optimization unit is used to jointly optimize the intent prediction model y I and the converted slot label model y i S to construct an objective function, and perform intent recognition on the user's voice problem based on the objective function.
- the word segmentation processing unit includes:
- the word segmentation module is used to convert the user's voice question into recognizable text, and use the word segmenter to segment the recognizable text to obtain the original word sequence;
- the embedded processing module is used to embedding the original word sequence to realize the vectorized representation of each word segmentation in the original word sequence.
- the first calculation unit includes:
- Implicit state calculation module for the bidirectional network for each word LSTM vector coding process, and outputs the sub-word vectors corresponding implicit state vector h i;
- Slot context calculation module used to pass formulas Calculate the slot context vector c i S corresponding to each participle vector; where, Represents the attention weight of the slot, and its calculation formula is ⁇ represents the slot activation function, Represents the slot weight matrix;
- Slot tag model module configured to build, based on the implicit tag slot state vector h i and the vector c i S slots context model
- the beneficial effects of the human-machine interactive voice recognition system for smart devices provided by the present invention are the same as the beneficial effects of the human-machine interactive voice recognition method for smart devices provided by the foregoing technical solutions, and are not described here To repeat.
- FIG. 1 is a schematic flowchart of a human-machine interactive voice recognition method for a smart device according to Embodiment 1 of the present invention
- FIG. 2 is an example diagram of an encoder-decoder fusion model in Embodiment 1 of the present invention
- FIG. 3 is an example diagram of the slot gate g in FIG. 2;
- FIG. 4 is a structural block diagram of a human-machine interactive voice recognition system for smart devices in Embodiment 2 of the present invention.
- FIG. 1 is a schematic flowchart of a human-machine interactive voice recognition method for a smart device according to Embodiment 1 of the present invention.
- this embodiment provides a human-machine interactive voice recognition method for a smart device, including:
- the acquired user voice question is first converted into recognized text, and the original word sequence is generated based on the recognizable text segmentation processing, and then the original word sequence is subjected to words
- the embedding process realizes the vectorized representation.
- the slot label model y i S and the intent prediction model y I are constructed based on the original word sequence represented by the vectorization.
- the construction step of the slot label model y i S is after calculating the vectors of the sub-word implicit state vector h i and a slot context of the vector c i S, then the implicit state vector h i and a slot context of the vector c i S obtain slot weighting tag model y i S, intent prediction
- the construction step of model y I is to calculate the implicit state vector hT and intention context vector c I of the original word sequence, and then weight the implicit state vector hT and intention context vector c I to obtain the intention prediction model y I , as shown in the figure
- the vector c i S and the intent context vector c I are stitched together.
- the intent prediction model y I and the converted slot label model y i S are jointly optimized to obtain the target function, and the target function is used to sequentially obtain each participle in the original word sequence.
- Corresponding intent conditional probabilities and then select the word segmentation with the highest probability value to recognize the intent of the user's voice problem, ensuring the accuracy of voice recognition.
- the user's speech word segmentation is processed to obtain the original word sequence
- the method of vectorizing the original word sequence through the embedding process includes:
- the received user's voice question is converted into recognizable text, and the word segmentation is used to segment the recognizable text to obtain the original word sequence; the original word sequence is subjected to word embedding processing to realize the vectorized representation of each word segmentation in the original word sequence.
- LSTM network using the bidirectional hidden state vector h i for each word vector coding process, and outputs the sub-word vectors corresponding to; by the equation Calculate the slot context vector c i S corresponding to each participle vector; where, Represents the attention weight of the slot, and its calculation formula is ⁇ represents the slot activation function, Represents a weight matrix slot; slot tag model constructed based on implicit state vector h i and a slot context of the vector c i S
- a plurality of word vectors LSTM bidirectional input one-output network may be hidden state vector h i, the formula for the context of the vector slot among them Represents the attention weight of the slot, i represents the i-th word segmentation vector, j represents the j-th element in the i-th word segmentation vector, specifically, the calculation formula of the slot's attention weight is T represents the total number of elements in the word segmentation vector, and K represents the Kth element in T.
- the slot activation function ⁇ and slot weight matrix It can be derived based on the vector matrix training of the original word sequence, and the specific training process is a common technical means in the art, which will not be repeated here in this embodiment.
- the method for calculating the implicit state vector hT and the intention context vector c I of the original word sequence represented by the vectorization in the above embodiment, and the method for obtaining the intention prediction model y I after weighting the implicit state vector hT and the intention context vector c I include:
- the training method of the intent prediction model y I and the slot label model The training method is the same, the difference is that the hidden state vector hT can be obtained only by using the hidden units in the bidirectional LSTM network, by one-dimensional processing of the vector matrix, and then calling the formula Calculate the intent context vector c I of the original word sequence; where,
- the attention weight of the table diagram, the calculation formula is ⁇ ′ table schematic activation function, Table schematic weight matrix, for intent activation function ⁇ ′ and intent weight matrix It can be derived based on the processed one-dimensional vector training.
- the specific training process is a common technical means in the art, and this embodiment will not repeat them here.
- the slot gate g is used to splice the slot context vector c i S and the intent context vector c I , and the method of converting and representing the slot label model y i S through the slot gate g includes: :
- v represents the weight vector obtained by training
- W represents the weight matrix obtained by training
- Fig. 3 shows a structural model of the slot gate g.
- the objective function constructed by jointly optimizing the intent prediction model y I and the converted slot label model y i S in the above embodiment is:
- X) represents the conditional probability of slot filling and intent prediction output at a given original word sequence, where X represents the original word sequence represented by vectorization.
- x i represents the i-th word segmentation vector
- T represents the total number of word segmentation vectors.
- this embodiment provides a human-machine interactive voice recognition system for smart devices, including:
- the word segmentation processing unit 1 is used for word segmentation processing of the user's speech problem to obtain an original word sequence, and vectorizing the original word sequence through embedding processing;
- the second calculation unit 3 is used to calculate the implicit state vector hT and the intention context vector c I of the original word sequence represented by the vectorization, and obtain the intention by weighting the implicit state vector hT and the intention context vector c I Prediction model y I ;
- the joint optimization unit 5 is used to jointly optimize the intent prediction model y I and the converted slot label model y i S to construct an objective function, and perform intent recognition on the user's voice problem based on the objective function.
- the word segmentation processing unit includes:
- the word segmentation module is used to convert the user's voice question into recognizable text, and use the word segmenter to segment the recognizable text to obtain the original word sequence;
- the embedded processing module is used to embedding the original word sequence to realize the vectorized representation of each word segmentation in the original word sequence.
- the first calculation unit includes:
- Implicit state calculation module for the bidirectional network for each word LSTM vector coding process, and outputs the sub-word vectors corresponding implicit state vector h i;
- Slot context calculation module used to pass formulas Calculate the slot context vector c i S corresponding to each participle vector; where, Represents the attention weight of the slot, and its calculation formula is ⁇ represents the slot activation function, Represents the slot weight matrix;
- Slot tag model module configured to build, based on the implicit tag slot state vector h i and the vector c i S slots context model
- the beneficial effects of the human-machine interactive voice recognition system for smart devices provided by the embodiments of the present invention are the same as the beneficial effects of the human-machine interactive voice recognition method for smart devices provided by the first embodiment, I will not repeat them here.
- the above program can be stored in a computer-readable storage medium.
- the program When executed, it includes Each step of the method in the foregoing embodiment, and the storage medium may be: ROM/RAM, magnetic disk, optical disk, memory card, or the like.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Machine Translation (AREA)
- Document Processing Apparatus (AREA)
Abstract
Description
Claims (10)
- 一种用于智能设备的人机交互语音识别方法,其特征在于,包括:将用户的语音问题分词处理得到原始词序列,并通过嵌入处理对所述原始词序列进行向量化表示;计算向量化表示的原始词序列隐含状态向量hT和意图上下文向量c I,通过将所述隐含状态向量hT和所述意图上下文向量c I加权处理后得到意图预测模型y I;
- 根据权利要求1所述的方法,其特征在于,将用户的语音问题分词处理得到原始词序列,并通过嵌入处理对所述原始词序列进行向量化表示的方法包括:接收用户的语音问题转换为可识别文本,利用分词器对所述可识别文本分词处理得到原始词序列;将原始词序列进行word embedding处理,实现对原始词序列中各分词的向量化表示。
- 根据权利要求6所述的方法,其特征在于,基于所述目标函数对用户的语音问题进行意图识别的方法包括:通过目标目标函数依次获取原始词序列中各分词对应的意图条件概率;从中筛选出概率值最大的分词识别为用户语音问题的意图。
- 一种用于智能设备的人机交互语音识别系统,其特征在于,包括:分词处理单元,用于将用户的语音问题分词处理得到原始词序列,并通过嵌入处理对所述原始词序列进行向量化表示;第二计算单元,用于计算向量化表示的原始词序列隐含状态向量hT和意图上下文向量c I,通过将所述隐含状态向量hT和所述意图上下文向量c I加权处理后得到意图预测模型y I;
- 根据权利要求8所述的系统,其特征在于,所述分词处理单元包括:分词模块,用于接收用户的语音问题转换为可识别文本,利用分词器对所述可识别文本分词处理得到原始词序列;嵌入处理模块,用于将原始词序列进行word embedding处理,实现对原始词序列中各分词的向量化表示。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA3166784A CA3166784A1 (en) | 2019-01-02 | 2019-09-19 | Human-machine interactive speech recognizing method and system for intelligent devices |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910002748.8A CN109785833A (zh) | 2019-01-02 | 2019-01-02 | 用于智能设备的人机交互语音识别方法及系统 |
CN201910002748.8 | 2019-01-02 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020140487A1 true WO2020140487A1 (zh) | 2020-07-09 |
Family
ID=66499837
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/106778 WO2020140487A1 (zh) | 2019-01-02 | 2019-09-19 | 用于智能设备的人机交互语音识别方法及系统 |
Country Status (3)
Country | Link |
---|---|
CN (1) | CN109785833A (zh) |
CA (1) | CA3166784A1 (zh) |
WO (1) | WO2020140487A1 (zh) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112765959A (zh) * | 2020-12-31 | 2021-05-07 | 康佳集团股份有限公司 | 意图识别方法、装置、设备及计算机可读存储介质 |
CN117151121A (zh) * | 2023-10-26 | 2023-12-01 | 安徽农业大学 | 一种基于波动阈值与分割化的多意图口语理解方法 |
Families Citing this family (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109785833A (zh) * | 2019-01-02 | 2019-05-21 | 苏宁易购集团股份有限公司 | 用于智能设备的人机交互语音识别方法及系统 |
CN110532355B (zh) * | 2019-08-27 | 2022-07-01 | 华侨大学 | 一种基于多任务学习的意图与槽位联合识别方法 |
CN110750628A (zh) * | 2019-09-09 | 2020-02-04 | 深圳壹账通智能科技有限公司 | 会话信息交互处理方法、装置、计算机设备和存储介质 |
CN110795532A (zh) * | 2019-10-18 | 2020-02-14 | 珠海格力电器股份有限公司 | 一种语音信息的处理方法、装置、智能终端以及存储介质 |
CN110853626B (zh) * | 2019-10-21 | 2021-04-20 | 成都信息工程大学 | 基于双向注意力神经网络的对话理解方法、装置及设备 |
CN110827816A (zh) * | 2019-11-08 | 2020-02-21 | 杭州依图医疗技术有限公司 | 语音指令识别方法、装置、电子设备及存储介质 |
CN111090728B (zh) * | 2019-12-13 | 2023-05-26 | 车智互联(北京)科技有限公司 | 一种对话状态跟踪方法、装置及计算设备 |
CN111062209A (zh) * | 2019-12-16 | 2020-04-24 | 苏州思必驰信息科技有限公司 | 自然语言处理模型训练方法和自然语言处理模型 |
CN111177381A (zh) * | 2019-12-21 | 2020-05-19 | 深圳市傲立科技有限公司 | 基于语境向量反馈的槽填充和意图检测联合建模方法 |
US20230040394A1 (en) * | 2020-01-06 | 2023-02-09 | 7Hugs Labs | System and method for controlling a plurality of devices |
CN111339770B (zh) * | 2020-02-18 | 2023-07-21 | 百度在线网络技术(北京)有限公司 | 用于输出信息的方法和装置 |
CN111833849A (zh) * | 2020-03-10 | 2020-10-27 | 北京嘀嘀无限科技发展有限公司 | 语音识别和语音模型训练的方法及存储介质和电子设备 |
CN113505591A (zh) * | 2020-03-23 | 2021-10-15 | 华为技术有限公司 | 一种槽位识别方法及电子设备 |
CN111597342B (zh) * | 2020-05-22 | 2024-01-26 | 北京慧闻科技(集团)有限公司 | 一种多任务意图分类方法、装置、设备及存储介质 |
CN113779975B (zh) * | 2020-06-10 | 2024-03-01 | 北京猎户星空科技有限公司 | 一种语义识别方法、装置、设备及介质 |
CN112069828B (zh) * | 2020-07-31 | 2023-07-04 | 飞诺门阵(北京)科技有限公司 | 一种文本意图的识别方法及装置 |
CN112800190B (zh) * | 2020-11-11 | 2022-06-10 | 重庆邮电大学 | 基于Bert模型的意图识别与槽值填充联合预测方法 |
CN114969339B (zh) * | 2022-05-30 | 2023-05-12 | 中电金信软件有限公司 | 一种文本匹配方法、装置、电子设备及可读存储介质 |
CN115358186B (zh) * | 2022-08-31 | 2023-11-14 | 南京擎盾信息科技有限公司 | 一种槽位标签的生成方法、装置及存储介质 |
CN115273849B (zh) * | 2022-09-27 | 2022-12-27 | 北京宝兰德软件股份有限公司 | 一种关于音频数据的意图识别方法及装置 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180182380A1 (en) * | 2016-12-28 | 2018-06-28 | Amazon Technologies, Inc. | Audio message extraction |
CN108415923A (zh) * | 2017-10-18 | 2018-08-17 | 北京邮电大学 | 封闭域的智能人机对话系统 |
CN108876527A (zh) * | 2018-06-06 | 2018-11-23 | 北京京东尚科信息技术有限公司 | 服务方法和服务装置、应用开放平台和存储介质 |
CN109065053A (zh) * | 2018-08-20 | 2018-12-21 | 百度在线网络技术(北京)有限公司 | 用于处理信息的方法和装置 |
CN109785833A (zh) * | 2019-01-02 | 2019-05-21 | 苏宁易购集团股份有限公司 | 用于智能设备的人机交互语音识别方法及系统 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107491541B (zh) * | 2017-08-24 | 2021-03-02 | 北京丁牛科技有限公司 | 文本分类方法及装置 |
CN108417205B (zh) * | 2018-01-19 | 2020-12-18 | 苏州思必驰信息科技有限公司 | 语义理解训练方法和系统 |
CN108874782B (zh) * | 2018-06-29 | 2019-04-26 | 北京寻领科技有限公司 | 一种层次注意力lstm和知识图谱的多轮对话管理方法 |
-
2019
- 2019-01-02 CN CN201910002748.8A patent/CN109785833A/zh not_active Withdrawn
- 2019-09-19 WO PCT/CN2019/106778 patent/WO2020140487A1/zh active Application Filing
- 2019-09-19 CA CA3166784A patent/CA3166784A1/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180182380A1 (en) * | 2016-12-28 | 2018-06-28 | Amazon Technologies, Inc. | Audio message extraction |
CN108415923A (zh) * | 2017-10-18 | 2018-08-17 | 北京邮电大学 | 封闭域的智能人机对话系统 |
CN108876527A (zh) * | 2018-06-06 | 2018-11-23 | 北京京东尚科信息技术有限公司 | 服务方法和服务装置、应用开放平台和存储介质 |
CN109065053A (zh) * | 2018-08-20 | 2018-12-21 | 百度在线网络技术(北京)有限公司 | 用于处理信息的方法和装置 |
CN109785833A (zh) * | 2019-01-02 | 2019-05-21 | 苏宁易购集团股份有限公司 | 用于智能设备的人机交互语音识别方法及系统 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112765959A (zh) * | 2020-12-31 | 2021-05-07 | 康佳集团股份有限公司 | 意图识别方法、装置、设备及计算机可读存储介质 |
CN112765959B (zh) * | 2020-12-31 | 2024-05-28 | 康佳集团股份有限公司 | 意图识别方法、装置、设备及计算机可读存储介质 |
CN117151121A (zh) * | 2023-10-26 | 2023-12-01 | 安徽农业大学 | 一种基于波动阈值与分割化的多意图口语理解方法 |
CN117151121B (zh) * | 2023-10-26 | 2024-01-12 | 安徽农业大学 | 一种基于波动阈值与分割化的多意图口语理解方法 |
Also Published As
Publication number | Publication date |
---|---|
CN109785833A (zh) | 2019-05-21 |
CA3166784A1 (en) | 2020-07-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020140487A1 (zh) | 用于智能设备的人机交互语音识别方法及系统 | |
CN108733792B (zh) | 一种实体关系抽取方法 | |
CN109033068B (zh) | 基于注意力机制的用于阅读理解的方法、装置和电子设备 | |
CN113268609B (zh) | 基于知识图谱的对话内容推荐方法、装置、设备及介质 | |
WO2021190259A1 (zh) | 一种槽位识别方法及电子设备 | |
CN113239169B (zh) | 基于人工智能的回答生成方法、装置、设备及存储介质 | |
CN110990555B (zh) | 端到端检索式对话方法与系统及计算机设备 | |
CN110678882B (zh) | 使用机器学习从电子文档选择回答跨距的方法及系统 | |
CN114676234A (zh) | 一种模型训练方法及相关设备 | |
CN111625634A (zh) | 词槽识别方法及装置、计算机可读存储介质、电子设备 | |
CN110399454B (zh) | 一种基于变压器模型和多参照系的文本编码表示方法 | |
CN109933792A (zh) | 基于多层双向lstm和验证模型的观点型问题阅读理解方法 | |
CN111814489A (zh) | 口语语义理解方法及系统 | |
CN115203409A (zh) | 一种基于门控融合和多任务学习的视频情感分类方法 | |
CN116304748A (zh) | 一种文本相似度计算方法、系统、设备及介质 | |
CN113705315A (zh) | 视频处理方法、装置、设备及存储介质 | |
CN116341651A (zh) | 实体识别模型训练方法、装置、电子设备及存储介质 | |
CN111597816A (zh) | 一种自注意力命名实体识别方法、装置、设备及存储介质 | |
CN116955644A (zh) | 基于知识图谱的知识融合方法、系统及存储介质 | |
US20240037335A1 (en) | Methods, systems, and media for bi-modal generation of natural languages and neural architectures | |
CN115659242A (zh) | 一种基于模态增强卷积图的多模态情感分类方法 | |
CN116258147A (zh) | 一种基于异构图卷积的多模态评论情感分析方法及系统 | |
CN115240712A (zh) | 一种基于多模态的情感分类方法、装置、设备及存储介质 | |
CN115130461A (zh) | 一种文本匹配方法、装置、电子设备及存储介质 | |
CN113822018A (zh) | 实体关系联合抽取方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19908004 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19908004 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19908004 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 3166784 Country of ref document: CA |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 07.02.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19908004 Country of ref document: EP Kind code of ref document: A1 |