CN106126596A - A kind of answering method based on stratification memory network - Google Patents

A kind of answering method based on stratification memory network Download PDF

Info

Publication number
CN106126596A
CN106126596A CN201610447676.4A CN201610447676A CN106126596A CN 106126596 A CN106126596 A CN 106126596A CN 201610447676 A CN201610447676 A CN 201610447676A CN 106126596 A CN106126596 A CN 106126596A
Authority
CN
China
Prior art keywords
word
sentence
granularity
encoding
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610447676.4A
Other languages
Chinese (zh)
Other versions
CN106126596B (en
Inventor
许家铭
石晶
姚轶群
徐波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201610447676.4A priority Critical patent/CN106126596B/en
Publication of CN106126596A publication Critical patent/CN106126596A/en
Application granted granted Critical
Publication of CN106126596B publication Critical patent/CN106126596B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Machine Translation (AREA)

Abstract

本发明提供了一种基于层次化记忆网络的问答方法,首先进行句粒度记忆编码,并在问题语义编码的刺激下,通过多轮迭代的注意力机制完成句粒度记忆单元的信息推理,通过k最大采样对句子进行筛选,在句粒度记忆编码的基础上还进行词粒度记忆编码,即在两个层次进行记忆编码,形成层次化的记忆编码,利用句粒度和词粒度记忆单元联合预测输出词概率分布,提高了自动问答的准确性,有效解决了低频词和未登录词的答案选择问题。

The present invention provides a question answering method based on a hierarchical memory network. Firstly, sentence granularity memory coding is performed, and under the stimulation of question semantic coding, the information reasoning of the sentence granularity memory unit is completed through a multi-round iterative attention mechanism. The maximum sampling is used to screen the sentences, and word-granularity memory encoding is also performed on the basis of sentence-granularity memory encoding, that is, memory encoding is performed at two levels to form a hierarchical memory encoding, and the sentence-granularity and word-granularity memory units are used to jointly predict the output word Probability distribution improves the accuracy of automatic question answering and effectively solves the problem of answer selection for low-frequency words and unregistered words.

Description

一种基于层次化记忆网络的问答方法A Question Answering Method Based on Hierarchical Memory Network

技术领域technical field

本发明涉及自动问答系统构建技术领域,更具体地涉及一种基于层次记忆网络的端到端问答方法。The present invention relates to the technical field of building an automatic question answering system, and more particularly relates to an end-to-end question answering method based on a hierarchical memory network.

背景技术Background technique

长期以来,自动问答一直都是自然语言处理问题中最具挑战性的任务之一,该任务需要对文本进行深层次理解并筛选出候选答案作为系统响应。目前已有的传统方法包括:采用流水线模式对文本处理过程中各个模块进行独立训练,然后融合输出的模式;构建大规模结构化知识库,并基于此知识库进行信息推理和答案预测。近些年,基于深度学习方法的端到端系统被广泛用于解决各种任务,这些方法无须手工构造特征,且无需对各个模块进行单独调优。Automatic question answering has long been one of the most challenging tasks in natural language processing, which requires deep understanding of text and screening candidate answers as system responses. The existing traditional methods include: using the pipeline mode to independently train each module in the text processing process, and then merging the output mode; building a large-scale structured knowledge base, and based on this knowledge base for information reasoning and answer prediction. In recent years, end-to-end systems based on deep learning methods have been widely used to solve various tasks, these methods do not need to manually construct features, and do not require individual tuning of each module.

问答系统大致可分为两个步骤:首先定位相关语义信息,该步骤称为“激活阶段”,然后基于相关信息进行响应生成,该步骤称为“生成阶段”。最近,神经记忆网络模型在问答系统任务中取得了较好地效果。但是,这些模型最大的缺点是采用单层次句粒度的记忆单元,无法较好地解决低频词或未登录词问题。而且通常情况下,为了降低模型的时间复杂度,常需要减少词典规模。此时,现有的端到端神经网络模型无法较好地选择出低频或未登录词作为答案输出。即当目标答案词在训练词典之外时,现有方法于在线测试阶段无法较好地选择准确答案作为模型输出。以下述对话文本为例:The question answering system can be roughly divided into two steps: firstly locate relevant semantic information, this step is called "activation phase", and then generate a response based on relevant information, this step is called "generation phase". Recently, neural memory network models have achieved good results in question answering system tasks. However, the biggest disadvantage of these models is that they use single-level sentence granular memory units, which cannot solve the problem of low-frequency words or unregistered words well. And usually, in order to reduce the time complexity of the model, it is often necessary to reduce the size of the dictionary. At this time, the existing end-to-end neural network model cannot well select low-frequency or unregistered words as the answer output. That is, when the target answer word is outside the training dictionary, the existing method cannot select the correct answer as the model output in the online test phase. Take the following dialogue text as an example:

1.先生您好,叫什么名字?1. Hello sir, what's your name?

2.呃,我叫威廉姆森。2. Uh, my name is Williamson.

3.请告诉我您的护照号码。3. Please tell me your passport number.

4.好的,是577838771。4. OK, it's 577838771.

5.还有您的电话号码?5. And your phone number?

6.号码是0016178290851。6. The number is 0016178290851.

假定“威廉姆森”,“577838771”和“0016178290851”是低频词或未登录词,如果传统方法将这些词丢弃或者统一以“unk”符号替换的话,这些方法均无法从对话文本中选择出准确的用户信息。然而,在实际应用中,多数答案信息来自于低频词或长尾词,如何设计一种能够有效解决未登录词的答案选择方法是目前自动问答系统领域迫切亟需的任务。Assuming that "Williamson", "577838771" and "0016178290851" are low-frequency words or unregistered words, if traditional methods discard these words or replace them with "unk" symbols, these methods cannot select accurate words from the dialogue text. user information. However, in practical applications, most of the answer information comes from low-frequency words or long-tail words. How to design an answer selection method that can effectively solve unregistered words is an urgent task in the field of automatic question answering systems.

发明内容Contents of the invention

(一)要解决的技术问题(1) Technical problems to be solved

为了解决现有技术问题,本发明提供了一种基于层次化记忆网络的问答方法。In order to solve the problems in the prior art, the present invention provides a question answering method based on a hierarchical memory network.

(二)技术方案(2) Technical solution

本发明提供了一种基于层次化记忆网络的问答方法,包括:步骤S101:融合词的位置和句子的时间序列信息,对句子集合中的句子进行句粒度记忆编码,得到句粒度记忆单元的双通道记忆编码;步骤S102:在问题语义编码的刺激下,通过多轮迭代的注意力机制完成所述句粒度记忆单元的信息推理,得到所述句粒度记忆单元上在词典维度的输出词概率分布;步骤S103:对所述句粒度记忆单元的信息推理结果进行k最大采样,从所述句子集合中筛选出k最大采样重要句子集合;步骤S104:利用双向循环神经网络模型对所述句子集合进行词粒度记忆编码,得到词粒度记忆单元的记忆编码;步骤S105:基于所述问题语义编码、词粒度记忆单元的记忆编码和k最大采样重要句子集合,通过注意力机制得到词粒度输出词概率分布;以及步骤S106:从句粒度和词粒度记忆单元中联合预测输出词概率分布,并利用交叉熵进行监督训练。The present invention provides a question answering method based on a hierarchical memory network, including: Step S101: Fusion word position and sentence time sequence information, carry out sentence granularity memory encoding on the sentences in the sentence set, and obtain the double sentence granularity memory unit Channel memory encoding; step S102: Under the stimulation of question semantic encoding, complete the information reasoning of the sentence granularity memory unit through the attention mechanism of multiple iterations, and obtain the output word probability distribution of the sentence granularity memory unit in the dictionary dimension ; Step S103: Carry out k-maximum sampling on the information reasoning results of the sentence granularity memory unit, and filter out the k-maximum sampling important sentence set from the sentence set; Step S104: Utilize the bidirectional recurrent neural network model to perform Word granularity memory encoding to obtain the memory encoding of the word granularity memory unit; Step S105: Based on the semantic encoding of the problem, the memory encoding of the word granularity memory unit and the k-maximum sampling important sentence set, the word granularity output word probability distribution is obtained through the attention mechanism ; and step S106 : jointly predict the output word probability distribution from the sentence granularity and word granularity memory units, and use cross-entropy for supervised training.

(三)有益效果(3) Beneficial effects

从上述技术方案可以看出,本发明的基于层次化记忆网络的问答方法具有以下有益效果:It can be seen from the above technical solution that the question answering method based on the hierarchical memory network of the present invention has the following beneficial effects:

(1)本发明首先进行句粒度记忆编码,并在问题语义编码的刺激下,通过多轮迭代的注意力机制完成句粒度记忆单元的信息推理,可以提高自动问答的准确性和及时性,有利于低频词和未登录词的答案选择;(1) The present invention first carries out sentence granular memory coding, and under the stimulation of problem semantic coding, completes the information reasoning of sentence granular memory unit through the attention mechanism of multiple iterations, can improve the accuracy and timeliness of automatic question answering, has Conducive to the answer selection of low-frequency words and unregistered words;

(2)通过k最大采样对句子进行筛选,可以提高自动问答的效率,降低计算复杂度;(2) Screening sentences by k-maximum sampling can improve the efficiency of automatic question answering and reduce computational complexity;

(3)在句粒度记忆编码的基础上还进行词粒度记忆编码,即在两个层次进行记忆编码,形成层次化的记忆编码,可以进一步提高自动问答的准确性;(3) On the basis of sentence granular memory coding, word granular memory coding is also carried out, that is, memory coding is carried out at two levels to form a hierarchical memory coding, which can further improve the accuracy of automatic question answering;

(4)利用循环神经网络进行词粒度记忆编码时,是在全句子集合X上操作的,该方式可以在词粒度记忆编码过程中引入词在句子全集合中的上下文环境语义信息,可以提高自动问答的准确性和及时性;(4) When the recurrent neural network is used for word granular memory coding, it is operated on the whole sentence set X. This method can introduce the context semantic information of words in the full sentence set during the word granular memory coding process, which can improve the automatic the accuracy and timeliness of questions and answers;

(5)词粒度记忆单元上的注意力机制是在k采样后的词粒度记忆单元子集合上运算的,避免了记忆编码中的干扰信息,并减少了词粒度注意力机制的计算量;(5) The attention mechanism on the word granularity memory unit is operated on the subset of word granularity memory units after k sampling, which avoids the interference information in the memory encoding and reduces the calculation amount of the word granularity attention mechanism;

(6)利用句粒度和词粒度记忆单元联合预测输出词概率分布,可以进一步提高自动问答的准确性,有效解决了低频词和未登录词的答案选择问题。(6) Using sentence granularity and word granularity memory units to jointly predict the output word probability distribution can further improve the accuracy of automatic question answering, and effectively solve the problem of answer selection for low-frequency words and unregistered words.

附图说明Description of drawings

图1为本发明实施例的基于层次化记忆网络的问答方法的流程图;FIG. 1 is a flowchart of a question answering method based on a hierarchical memory network according to an embodiment of the present invention;

图2为本发明实施例的基于层次化记忆网络的问答方法的框架示意图;FIG. 2 is a schematic framework diagram of a question answering method based on a hierarchical memory network according to an embodiment of the present invention;

图3为本发明实施例的句粒度记忆编码及基于句粒度记忆编码的信息推理示意图;Fig. 3 is a schematic diagram of sentence granularity memory coding and information reasoning based on sentence granularity memory coding according to an embodiment of the present invention;

图4为本发明实施例的词粒度记忆编码及基于词粒度记忆编码的注意力激活示意图;FIG. 4 is a schematic diagram of word granularity memory coding and attention activation based on word granularity memory coding according to an embodiment of the present invention;

图5为本发明实施例的基于层次化记忆网络的问答方法的性能示意图1;FIG. 5 is a performance diagram 1 of a question answering method based on a hierarchical memory network according to an embodiment of the present invention;

图6为本发明实施例的基于层次化记忆网络的问答方法的又一性能示意图。FIG. 6 is another performance diagram of a question answering method based on a hierarchical memory network according to an embodiment of the present invention.

具体实施方式detailed description

为使本发明的目的、技术方案和优点更加清楚明白,以下结合具体实施例,并参照附图,对本发明作进一步的详细说明。In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with specific embodiments and with reference to the accompanying drawings.

本发明公开了一种基于层次化记忆网络的问答方法,其基于一种全神经网络结构的端到端模型,可以实现句子集合中的信息推理,筛选和词粒度选择,有效解决大数据下问答系统对低频词或未登录词的答案选择问题。本发明的问答方法将具有时间序列信息的句子集合分别进行两个层次化的记忆编码,分别为:句粒度记忆编码和词粒度记忆编码。然后基于层次化记忆网络进行信息推理、筛选和激活,联合预测候选答案词概率分布。The invention discloses a question answering method based on a hierarchical memory network, which is based on an end-to-end model of a full neural network structure, which can realize information reasoning, screening and word granularity selection in a sentence set, and effectively solve question answering under big data The system selects answers to low-frequency words or unregistered words. In the question answering method of the present invention, sentence sets with time series information are respectively subjected to two hierarchical memory encodings, namely: sentence granularity memory encoding and word granularity memory encoding. Then information reasoning, screening and activation are performed based on the hierarchical memory network, and the probability distribution of candidate answer words is jointly predicted.

本发明的问答方法,通过层次化记忆网络先对句子集合进行句向量化记忆编码,考虑词在句子中的位置信息及句子在句子集合中的序列时间信息,然后通过多轮迭代的注意力机制完成句粒度记忆单元的信息推理,并基于该推理结果进行k最大采样,筛选出重要的句子信息。然后利用双向循环网络模型对句子集合进行词粒度的序列编码,并通过注意力机制从被筛选的信息中进行词粒度记忆单元的信息激活,最后分别从句粒度和词粒度记忆单元中预测输出词概率分布并通过Softmax进行联合监督训练,学习端到端的自动问答模型。In the question-answering method of the present invention, sentence vectorized memory encoding is first performed on the sentence set through a hierarchical memory network, considering the position information of words in the sentence and the sequence time information of the sentence in the sentence set, and then through the attention mechanism of multiple rounds of iteration Complete the information reasoning of the sentence granularity memory unit, and perform k-maximum sampling based on the reasoning result to screen out important sentence information. Then use the two-way recurrent network model to encode the word-granularity sequence of the sentence set, and use the attention mechanism to activate the information of the word-granularity memory unit from the screened information, and finally predict the output word probability from the sentence-granularity and word-granularity memory units respectively Distributed and jointly supervised training by Softmax to learn an end-to-end automatic question answering model.

下面结合附图对作为本发明实施例的基于层次化记忆网络的问答方法进行详细描述。A question answering method based on a hierarchical memory network as an embodiment of the present invention will be described in detail below with reference to the accompanying drawings.

图1为本发明实施例的基于层次化记忆网络的问答方法的流程图,参照图1,该问答方法包括:Fig. 1 is the flowchart of the question answering method based on hierarchical memory network of the embodiment of the present invention, with reference to Fig. 1, this question answering method comprises:

步骤S101:融合词的位置和句子的时间序列信息,对句子集合中的句子进行句粒度记忆编码,得到句粒度记忆单元的双通道记忆编码。Step S101: Fusing word positions and sentence time series information, performing sentence-granularity memory encoding on the sentences in the sentence set to obtain a dual-channel memory encoding of the sentence-granularity memory unit.

参照图3,步骤S101包括:Referring to Fig. 3, step S101 includes:

子步骤S101a:对句子集合中具有时间序列信息的句子进行双通道词向量映射,得到句子的双通道词向量化编码。Sub-step S101a: Perform dual-channel word vector mapping on the sentences with time series information in the sentence set to obtain the dual-channel word vectorization encoding of the sentences.

子步骤S101a包括:给定具有时间序列信息的句子集合X={xi}i=(1,2,...,n),其中,i为句子的当前时间序列,n为句子集合的最大时间序列长度;随机初始化两个词向量矩阵其中,|V|为词典维度,d为词向量的维度,A和C分别采用标准差为0.1、均值为0的正态分布作为随机初始化参数,对句子集合X中的句子xi进行双通道词向量映射,则句子xi中的词xij的双通道向量化编码为j为词在句子xi中的位置信息。Sub-step S101a includes: Given a set of sentences with time series information X={ xi } i=(1, 2, ..., n) , where i is the current time series of sentences, and n is the largest sentence set Time series length; randomly initialize two word vector matrices and Among them, | V | is the dimension of the dictionary, d is the dimension of the word vector, A and C respectively use the normal distribution with a standard deviation of 0.1 and a mean of 0 as random initialization parameters, and perform dual-channel Word vector mapping, then the two-channel vectorization encoding of word x ij in sentence x i is and j is the position information of the word in the sentence x i .

子步骤S101b:根据词在句子中的位置信息对双通道词向量化编码进行更新。Sub-step S101b: update the dual-channel word vectorization encoding according to the position information of the word in the sentence.

子步骤S101b包括:根据词在句子中的位置信息j及词向量的维度d生成更新矩阵l,则更新后的双通道词向量化编码为lgj·(Axij)和lgj·(Cxij),其中:Sub-step S101b includes: generating an update matrix l according to the position information j of the word in the sentence and the dimension d of the word vector, then the updated two-channel word vectorization encoding is l gj (Ax ij ) and l gj (Cx ij ),in:

lgj=(1-j/Ji)-(g/d)(1-2j/Ji) (1)l gj =(1-j/J i )-(g/d)(1-2j/J i ) (1)

其中,Ji是句子xi中词的个数,而g为维度为d的词向量中的当前维度值,且1≤j≤J和1≤g≤d。Among them, J i is the number of words in the sentence x i , and g is the current dimension value in the word vector with dimension d, and 1≤j≤J and 1≤g≤d.

子步骤S101c:融合句子的时间序列信息对句子进行句粒度记忆编码,得到句粒度记忆单元的双通道记忆编码。Sub-step S101c: Integrating the time series information of the sentence to perform sentence-granularity memory encoding on the sentence to obtain the dual-channel memory encoding of the sentence-granularity memory unit.

子步骤S101c包括:随机初始化两个句子时间序列的向量化矩阵其中,n为句子集合的最大时间序列长度,d为时间向量维度,与词向量的维度相同,TA和TC分别采用标准差为0.1、均值为0的正态分布作为随机初始化参数,则句粒度记忆单元的双通道记忆编码为M(S)={{ai},{ci}},其中:Substep S101c includes: randomly initializing the vectorized matrix of two sentence time series and Among them, n is the maximum time series length of the sentence set, d is the dimension of the time vector, which is the same as the dimension of the word vector, T A and T C respectively adopt the normal distribution with standard deviation of 0.1 and mean value of 0 as random initialization parameters, then The dual-channel memory encoding of the sentence granularity memory unit is M (S) = {{a i }, { ci }}, where:

aa ii == ΣΣ jj ll jj ·&Center Dot; (( AxAx ii jj )) ++ TT AA (( ii )) -- -- -- (( 22 ))

cc ii == ΣΣ jj ll jj ·&Center Dot; (( CxCx ii jj )) ++ TT CC (( ii )) -- -- -- (( 33 ))

其中,lj为中更新矩阵l在句子xi中第j个词的更新向量,操作符·为向量间元素乘法操作,如公式(2)中lj·(Axij)表示向量lj和向量(Axij)进行元素乘法操作。Among them, l j is the update vector of the jth word in the sentence x i in the update matrix l, and the operator · is the element multiplication operation between vectors, such as l j · (Ax ij ) in the formula (2) represents the vector l j and Vector(Ax ij ) performs element-wise multiplication.

步骤S102:在问题语义编码的刺激下,通过多轮迭代的注意力机制完成句粒度记忆单元的信息推理,得到句粒度记忆单元上在词典维度的输出词概率分布。Step S102: Under the stimulation of the semantic encoding of the question, the information reasoning of the sentence granular memory unit is completed through the multi-round iterative attention mechanism, and the output word probability distribution of the sentence granular memory unit in the dictionary dimension is obtained.

步骤S102包括:Step S102 includes:

子步骤S102a:对问题文本进行向量化表示,得到问题的语义编码。Sub-step S102a: Perform vectorized representation of the question text to obtain the semantic code of the question.

子步骤S102a包括:利用所述词向量矩阵对问题文本q中第j个词qj进行向量化表示并基于词在问题文本中的位置j对该向量化表示进行更新,得到问题的语义编码:Sub-step S102a includes: using the word vector matrix Vectorized representation of the jth word q j in the question text q And based on the position j of the word in the question text, the vectorized representation is updated to obtain the semantic encoding of the question:

uu 11 (( SS )) == ΣΣ jj ll jj ·&Center Dot; (( AqQ jj )) -- -- -- (( 44 ))

同所述公式(2)和(3),lj为更新矩阵l在句子xi中第j个词的更新向量。Same as the formulas (2) and (3), l j is the update vector of the jth word of the update matrix l in the sentence x i .

子步骤S102b:在问题语义编码的刺激下,利用注意力机制在句粒度记忆单元的双通道记忆编码中进行信息激活;Sub-step S102b: Under the stimulation of the semantic encoding of the question, use the attention mechanism to activate information in the dual-channel memory encoding of the sentence granularity memory unit;

子步骤S102b包括:采用点积方式计算问题语义编码在句粒度记忆单元的注意力权重:Sub-step S102b includes: using the dot product method to calculate the attention weight of the question semantic code in the sentence granularity memory unit:

αα ii (( SS )) == sthe s oo ff tt maxmax (( aa ii TT uu 11 (( SS )) )) -- -- -- (( 55 ))

则在问题语义编码刺激下,句粒度记忆单元的双通道记忆编码的激活信息为: Then, under the stimulation of question semantic coding, the activation information of the dual-channel memory coding of the sentence granularity memory unit is:

子步骤S102c:通过多轮迭代的注意力机制完成在句粒度记忆单元的信息推理,得到句粒度记忆单元上在词典维度的输出词概率分布。Sub-step S102c: complete the information reasoning in the sentence granular memory unit through the multi-round iterative attention mechanism, and obtain the output word probability distribution in the dictionary dimension on the sentence granular memory unit.

子步骤S102c包括:在句粒度记忆单元上进行R轮信息激活,找到候选句子集合,得到第R轮的激活信息OR,其中,在第r+1轮信息激活中,Sub-step S102c includes: performing R rounds of information activation on the sentence granularity memory unit, finding a set of candidate sentences, and obtaining the Rth round of activation information OR , wherein, in the r +1th round of information activation,

uu rr ++ 11 (( SS )) == oo rr ++ uu rr (( SS )) -- -- -- (( 66 ))

αα ii (( SS )) == sthe s oo ff tt mm aa xx (( aa ii TT uu rr ++ 11 (( SS )) )) -- -- -- (( 77 ))

aa ii == ΣΣ jj ll jj ·· (( AA rr ++ 11 xx ii jj )) ++ TT AA rr ++ 11 (( ii )) -- -- -- (( 88 ))

oo rr ++ 11 == ΣΣ ii αα ii (( SS )) cc ii -- -- -- (( 99 ))

cc ii == ΣΣ jj ll jj ·&Center Dot; (( CC rr ++ 11 xx ii jj )) ++ TT CC rr ++ 11 (( ii )) -- -- -- (( 1010 ))

其中,1≤r≤(R-1),在第r+1轮信息激活中采用独立的词向量矩阵Ar+1和Cr+1和时间向量矩阵对句子集合进行向量化表示,且Cr分别采用标准差为0.1、均值为0的正态分布作为随机初始化参数。Among them, 1≤r≤(R-1), the independent word vector matrix A r+1 and C r+1 and the time vector matrix are used in the r+1 round of information activation and A vectorized representation of a collection of sentences, and C r and A normal distribution with a standard deviation of 0.1 and a mean of 0 was used as random initialization parameters.

通过R轮迭代的注意力机制完成在句粒度记忆单元的信息推理,得到句粒度记忆单元上在词典维度的输出词概率分布为:The information reasoning in the sentence granularity memory unit is completed through the attention mechanism of R rounds of iteration, and the output word probability distribution in the dictionary dimension on the sentence granularity memory unit is obtained as:

pp (( SS )) (( ww )) == sthe s oo ff tt mm aa xx (( (( CC RR )) TT (( oo RR ++ uu RR (( SS )) )) )) -- -- -- (( 1111 ))

其中,为词典维度词集合,为第R轮信息激活的词向量矩阵,而T为转置操作符。in, is the set of dictionary dimension words, is the matrix of word embeddings activated for the R-th round information, and T is the transposition operator.

本发明首先进行句粒度记忆编码,并在问题语义编码的刺激下,通过多轮迭代的注意力机制完成句粒度记忆单元的信息推理,可以提高自动问答的准确性和及时性,有利于低频词和未登录词的答案选择。The present invention first performs sentence granularity memory coding, and under the stimulation of question semantic coding, completes the information reasoning of sentence granularity memory units through the attention mechanism of multiple rounds of iterations, which can improve the accuracy and timeliness of automatic question and answer, and is beneficial to low-frequency words and unregistered word answer choices.

步骤S103:对句粒度记忆单元的信息推理结果进行k最大采样,从句子集合中筛选出k最大采样重要句子集合。Step S103: Perform k-maximum sampling on the information reasoning results of the sentence-granularity memory unit, and select the k-maximum-sampling important sentence set from the sentence set.

步骤S103包括:Step S103 includes:

子步骤S103a:对句粒度记忆单元上第R轮信息激活的注意力权重向量从中选取其k个最大的注意力权重子集合 Sub-step S103a: the attention weight vector activated on the R-th round of information on the sentence granularity memory unit Select the k largest attention weight sub-sets from it

子步骤S103b:选取k个最大的注意力权重子集合对应的句子集合作为k最大采样重要句子集合重要句子集合中的句子为重要句子。Sub-step S103b: Select the k largest attention weight sub-sets The corresponding sentence set is used as the k-maximum sampling important sentence set Sentences in the set of important sentences for important sentences.

本发明k最大采样对句子进行筛选,可以提高自动问答的效率,降低计算复杂度,更加有利于低频词和未登录词的答案选择。The k-maximum sampling of the present invention screens sentences, which can improve the efficiency of automatic question answering, reduce computational complexity, and be more conducive to the answer selection of low-frequency words and unregistered words.

步骤S104:利用双向循环神经网络模型对句子集合进行词粒度记忆编码,得到词粒度记忆单元的记忆编码。Step S104: use the bidirectional recurrent neural network model to perform word-granularity memory encoding on the sentence set, and obtain the memory encoding of the word-granularity memory unit.

参照图4,步骤S104包括:Referring to FIG. 4, step S104 includes:

子步骤S104a:利用双向循环网络模型对重要句子集合中的词按时间序列进行编码,得到双向循环网络模型的隐状态。双向循环网络模型有多种现有模型,本实施例采用其中的一种:门循环网络模型(GRU)。Sub-step S104a: use the bidirectional recurrent network model to encode the words in the important sentence set in time series, and obtain the hidden state of the bidirectional recurrent network model. There are many existing bidirectional recurrent network models, one of which is adopted in this embodiment: the gated recurrent network model (GRU).

子步骤S104a包括:利用门循环网络模型(GRU)分别对句子集合X中的所有词按时间序列进行正向和反向编码,对于t时刻的词特征,前向GRU编码的隐状态为后向GRU编码的隐状态为其中,|t|为按照时间序列对句子集合X中所有词进行排列后的词最大序列长度,的维度与词向量的维度d相同,CR为句粒度记忆单元中第R轮信息激活过程中的词向量矩阵。Sub-step S104a includes: using the Gated Recurrent Network Model (GRU) to respectively classify all words in the sentence set X Forward and reverse encoding according to time series, for word features at time t, the hidden state of forward GRU encoding is The hidden state of the backward GRU encoding is Among them, |t| is the maximum sequence length of words after arranging all the words in the sentence set X according to the time sequence, and The dimension of is the same as the dimension d of the word vector, and C R is the word vector matrix during the R-th round of information activation in the sentence granularity memory unit.

子步骤S104b:对双向循环网络模型的隐状态进行融合,得到词粒度记忆单元的记忆编码。Sub-step S104b: Fusion the hidden state of the bidirectional recurrent network model to obtain the memory encoding of the word granularity memory unit.

子步骤S104b包括:将双向循环网络模型的隐状态直接相加,得到词粒度记忆单元的记忆编码M(W)={mt}t=1,2,..|t|),其中 Sub-step S104b includes: adding the hidden states of the two-way recurrent network model directly to obtain the memory encoding M (W) ={m t } t=1, 2, ..|t|) of the memory unit at the word granularity, where

本发明利用循环神经网络进行词粒度记忆编码时,是在全句子集合X上操作的,该方式可以在词粒度记忆编码过程中引入词在句子全集合中的上下文环境语义信息,可以提高自动问答的准确性和及时性,有利于低频词和未登录词的答案选择。When the present invention utilizes the recurrent neural network to perform word granularity memory encoding, it operates on the entire sentence set X. This method can introduce the context and semantic information of words in the full set of sentences during the word granularity memory encoding process, which can improve automatic question answering. The accuracy and timeliness of the method are beneficial to the answer selection of low-frequency words and unregistered words.

步骤S105:基于问题语义编码、词粒度记忆单元的记忆编码和k最大采样重要句子集合,通过注意力机制得到词粒度输出词概率分布。Step S105: Based on the semantic encoding of the question, the memory encoding of the memory unit at the word granularity, and the k-maximum sampling important sentence set, the output word probability distribution at the word granularity is obtained through the attention mechanism.

步骤S105包括:Step S105 includes:

子步骤S105a:根据问题语义编码和词粒度记忆单元的记忆编码计算词粒度记忆单元的注意力权重;Sub-step S105a: Calculate the attention weight of the word-granularity memory unit according to the semantic code of the question and the memory code of the word-granularity memory unit;

子步骤S105a包括:基于句粒度记忆单元上第R轮信息激活过程中的问题语义编码词粒度记忆单元的记忆编码M(W)={mt}t=1,2,..,|t|)和k最大采样重要句子集合得到归一化后的词粒度记忆单元的注意力权重向量其中:Sub-step S105a includes: based on the semantic encoding of the question during the R-th round of information activation on the sentence granularity memory unit Memory encoding M (W) = {m t } t = 1, 2, .., |t|) and k largest sampling important sentence set of word granularity memory unit Obtain the attention weight vector of the normalized word granularity memory unit in:

αα tt (( ww )) == sthe s oo ff tt mm aa xx (( vv TT tanhtanh (( WuWu RR (( SS )) ++ Uu mm ^^ tt )) )) -- -- -- (( 1212 ))

其中,是k最大采样重要句子集合中的词集合所对应的词粒度记忆编码M(W)={mt}t=(1,2,...,|t|)中的子集合注意力权重向量α(W)的维度与按照时间序列对重要句子集合中所有词进行排列后的词最大序列长度一致,即为 均为学习参数,v、W和U均采用标准差为0.1、均值为0的正态分布进行随机初始化,在训练阶段进行更新。in, is the k-maximally sampled set of important sentences set of words in The corresponding word-granularity memory code M (W) = {m t } t = (1, 2, ..., |t|) in the subset The dimension of the attention weight vector α (W) and the set of important sentences according to the time series all words in The maximum sequence length of the arranged words is the same, that is, and All are learning parameters, v, W and U are randomly initialized with a normal distribution with a standard deviation of 0.1 and a mean of 0, and are updated during the training phase.

子步骤S105b:根据词粒度记忆单元的注意力权重得到词粒度输出词概率分布。在本发明的实施例中,直接采用归一化后的词粒度记忆单元的注意力权重α(W)作为词粒度输出词概率分布:Sub-step S105b: Obtain the word granularity output word probability distribution according to the attention weight of the word granularity memory unit. In an embodiment of the present invention, the attention weight α (W) of the normalized word granularity memory unit is directly adopted as the word granularity output word probability distribution:

pp (( WW )) (( ww ^^ )) == αα (( WW )) -- -- -- (( 1313 ))

此时,词粒度输出词概率分布与注意力权重的维度一致,即 为重要句子集合中所有词的集合 At this time, the word granularity output word probability distribution is consistent with the dimension of attention weight, namely is the set of all words in the important sentence set

本发明在句粒度记忆编码的基础上还进行词粒度记忆编码,即在两个层次进行记忆编码,形成层次化的记忆编码,可以进一步提高自动问答的准确性,更加有利于低频词和未登录词的答案选择。同时,词粒度记忆单元上的注意力机制是在k采样后的词粒度记忆单元子集合上运算的,避免了记忆编码中的干扰信息,并减少了词粒度注意力机制的计算量。The present invention also performs word granularity memory coding on the basis of sentence granularity memory coding, that is, performs memory coding at two levels to form hierarchical memory coding, which can further improve the accuracy of automatic question and answer, and is more conducive to low-frequency words and unregistered words. word answer choice. At the same time, the attention mechanism on the word granularity memory unit is operated on the subset of word granularity memory units after k sampling, which avoids the interference information in the memory encoding and reduces the calculation amount of the word granularity attention mechanism.

步骤S106:从句粒度和词粒度记忆单元中联合预测输出词概率分布,并利用交叉熵进行监督训练。Step S106: Jointly predict the output word probability distribution from the sentence-granularity and word-granularity memory units, and use cross-entropy for supervised training.

步骤S106包括:Step S106 includes:

子步骤S106a:基于句粒度记忆单元上在词典维度的输出词概率分布和词粒度输出词概率分布进行输出词联合预测,联合预测输出词分布p(w)表达式为:Sub-step S106a: Based on the output word probability distribution in the dictionary dimension and the word granularity output word probability distribution on the sentence granularity memory unit, the output word joint prediction is performed, and the expression of the joint prediction output word distribution p(w) is:

pp (( ww )) == pp (( SS )) (( ww )) ++ pp (( WW )) (( ww )) == pp (( SS )) (( ww )) ++ tt rr aa nno sthe s (( pp (( WW )) (( ww ^^ )) )) -- -- -- (( 1414 ))

其中,trans(·)表示将子集合的词粒度输出词概率分布映射到词典维度全集合的词粒度输出词概率分布该映射操作具体是指输出词概率分布中概率值按照其对应词子集合中的词在词典维度词全集合中的位置进行概率值映射,若全集合中的某些词在子集合中未出现,则将其输出概率置为0,得到映射后的词输出概率分布 Among them, trans( ) indicates that the word granularity of the sub-set is output to the word probability distribution Word granularity output word probability distribution mapped to the full set of dictionary dimensions The mapping operation specifically refers to the output word probability distribution The medium probability value according to its corresponding word sub-set The words in the dictionary dimension word set The position in the probability value mapping, if some words in the full set do not appear in the sub-set, the output probability is set to 0, and the output probability distribution of the word after mapping is obtained

子步骤S106b:利用目标答案词分布对联合预测输出词分布进行交叉熵监督训练。给定训练集的目标答案词分布为y,则基于目标答案词分布y与联合预测输出词分布p(w)的交叉熵函数进行联合优化。Sub-step S106b: Perform cross-entropy supervised training on the joint prediction output word distribution by using the target answer word distribution. Given that the target answer word distribution of the training set is y, the joint optimization is performed based on the cross-entropy function of the target answer word distribution y and the joint prediction output word distribution p(w).

本发明的一个示例性实施例中,采用随机梯度下降方法进行误差反向传播对联合优化中的目标函数进行优化,优化参数包括词粒度记忆单元中的词向量矩阵{Ar}r=(1,2,...,R)和{Cr}t=1,2,...,R)和时间向量矩阵词粒度记忆编码过程中所采用双向GRU模型的所有参数集合{θGRU}及计算词粒度记忆单元的注意力权重(公式(12))中的v,W和U。In an exemplary embodiment of the present invention, the stochastic gradient descent method is used to carry out error backpropagation to optimize the objective function in the joint optimization, and the optimization parameters include the word vector matrix {A r } r=(1 , 2,...,R) and {C r } t=1, 2,...,R) and time vector matrix and All parameter sets {θ GRU } of the two-way GRU model used in the encoding process of word granularity memory and v, W and U in calculating the attention weight (formula (12)) of the word granularity memory unit.

本发明在句粒度和词粒度记忆单元中联合预测输出词概率分布,可以进一步提高自动问答的准确性,更加有利于低频词和未登录词的答案选择。The present invention jointly predicts the output word probability distribution in the sentence granularity and word granularity memory units, can further improve the accuracy of automatic question answering, and is more conducive to the answer selection of low-frequency words and unregistered words.

图2为作为本发明一个实施例的基于层次化记忆网络的问答方法的框架示意图。参照图2,基于层次化记忆网络的问答方法共有两个层次的记忆网络单元,分别为:FIG. 2 is a schematic framework diagram of a question answering method based on a hierarchical memory network as an embodiment of the present invention. Referring to Figure 2, the question answering method based on the hierarchical memory network has two levels of memory network units, which are:

记忆单元一:句子集合以时间序列进行句粒度的编码记忆;Memory unit 1: Sentence sets are coded and memorized at sentence granularity in time series;

记忆单元二:句子集合中所有词按照时间序列进行词粒度的编码记忆。Memory unit 2: All words in the sentence set are coded and memorized according to the time sequence of word granularity.

不同记忆单元层次间采用k最大采用进行重要信息筛选和过滤。Between different memory unit levels, k-maximum adoption is used to screen and filter important information.

模型信息处理阶段有两处信息激活机制,分别为:There are two information activation mechanisms in the model information processing stage, which are:

激活机制一:在句粒度记忆单元上采用推理机制进行信息激活;Activation mechanism 1: Use the reasoning mechanism to activate information on the sentence granularity memory unit;

激活机制二:在词粒度记忆单元上采用注意力机制进行词选择。Activation mechanism 2: Use the attention mechanism on the word granularity memory unit for word selection.

整个模型训练阶段共有两处监督信息进行指导,分别为:During the entire model training phase, there are two supervisory information for guidance, namely:

监督信息一:句粒度记忆单元进行信息推理后的输出向量进行解码和Softmax输出后对目标词的拟合信息;Supervision information 1: the output vector after the information reasoning of the sentence granularity memory unit is decoded and the fitting information of the target word is output by Softmax;

监督信息二:词粒度记忆单元进行注意力机制激活和Softmax输出后对目标词的拟合信息。Supervision information 2: the fitting information of the target word after the attention mechanism activation and Softmax output of the word granular memory unit.

为了准确评估本发明方法的自动问答响应性能,本发明通过对比模型选择输出的答案词和数据真实答案词不一致的错误样本数对比本发明方法的性能。In order to accurately evaluate the automatic question and answer response performance of the method of the present invention, the present invention compares the performance of the method of the present invention by comparing the number of error samples in which the answer words output by the model selection and the real answer words of the data are inconsistent.

表1Table 1

数据领域data field 训练/测试问答对train/test question-answer pairs 词典大小(全部/训练/测试)Dictionary size (all/train/test) 未登录目标词(百分比)Unregistered target words (percentage) 机票预订pre-book fly ticket 7,000/7,0007,000/7,000 10,682/5,612/5,61810,682/5,612/5,618 5,070(72.43%)5,070 (72.43%)

本发明的实验中采用一种中文机票订票领域文本数据集,该数据集共包含2,000个完整对话历史,14,000个问答对,将其5∶5分为训练集和测试集。针对这些文本数据集,本发明不做任何处理(包括去停用词和词干还原等操作)。数据集的具体统计信息如表1所示,可以看到,在测试集中的未登录目标词占有72.43%,对传统模型训练会产生比较大的影响。In the experiment of the present invention, a text data set in the field of Chinese airline ticket booking is adopted, which contains 2,000 complete dialogue histories and 14,000 question-answer pairs, which are divided into training set and test set 5:5. For these text data sets, the present invention does not do any processing (including operations such as removing stop words and word stem restoration). The specific statistical information of the data set is shown in Table 1. It can be seen that the unregistered target words in the test set account for 72.43%, which will have a relatively large impact on traditional model training.

本发明的实验中采用以下对比方法:Adopt following comparative method in the experiment of the present invention:

对比方法一:基于注意力机制的指针网络模型,该方法将句子集合中的所有词按照时间序列看成一个长句进行编码,直接利用问题与词编码的注意力机制生成答案;Comparison method 1: Pointer network model based on attention mechanism, which encodes all words in the sentence set as a long sentence according to time sequence, and directly uses the attention mechanism of question and word encoding to generate answers;

对比方法二:神经记忆网络模型,该方法对句子集合进行句粒度编码,利用问题的编码向量进行语义激活后的信息直接在全词典空间上进行答案匹配。Contrast method 2: Neural memory network model, this method encodes sentence sets at sentence granularity, and uses the encoding vector of the question for semantically activated information to directly match the answer in the full dictionary space.

本发明实验中采用参数设置如表2所示:Adopt parameter setting in the experiment of the present invention as shown in table 2:

表2Table 2

nno dd RR kk lrlr bsbs 1616 100100 33 11 0.010.01 1010

表2中,参数n为实验数据的句子集合的句子最大时间序列,d为词向量维度及隐层编码维度,R是句粒度记忆单元上推理机制的迭代次数,k为不同层次记忆间的最大采样数,lr是采用随机梯度下降方法进行模型参数优化时的学习率,bs是进行模型训练时每批样本的数量。In Table 2, the parameter n is the maximum sentence time series of the sentence set of the experimental data, d is the word vector dimension and the hidden layer coding dimension, R is the number of iterations of the reasoning mechanism on the sentence granularity memory unit, and k is the maximum The number of samples, lr is the learning rate when the stochastic gradient descent method is used to optimize the model parameters, and bs is the number of samples in each batch when the model is trained.

本发明实验中,进行15轮迭代训练,所有方法均收敛如图5所示,最终收敛后的实验结果如表3所示:In the experiment of the present invention, 15 rounds of iterative training are carried out, and all methods converge as shown in Figure 5, and the experimental results after the final convergence are shown in Table 3:

表3table 3

方法method 错误样本数error samples 对比方法一Comparison method one 109109 对比方法二Comparison method two 5656 本发明方法The method of the invention 00

图5和表3为本发明方法、对比方法一和对比方法二在数据集上的错误样本数评测结果。实验结果表明,本发明方法的收敛速度显著优越于其他方法。且根据表3中的最终收敛结果,可以看到,本发明方法明显优于其他方法,可以完全解决未登录词集合上的答案选择问题,达到100%的正确率。Figure 5 and Table 3 are the evaluation results of the number of error samples on the data set for the method of the present invention, the first comparison method and the second comparison method. Experimental results show that the convergence speed of the method of the present invention is significantly superior to other methods. And according to the final convergence results in Table 3, it can be seen that the method of the present invention is obviously superior to other methods, and can completely solve the problem of answer selection on unregistered word sets, reaching a 100% correct rate.

同时,本发明实验验证在层次记忆单元间信息筛选的最大采样数k对答案选择问题中错误样本数的性能影响,实验结果如图6和表4所示。可以看到,当最大采样数为1时,本发明方法性能的收敛速度和最终收敛结果可达到最优,进一步说明层次记忆单元间进行信息选择的重要性。At the same time, the experiment of the present invention verifies the performance impact of the maximum number of samples k for information screening among hierarchical memory units on the number of wrong samples in answer selection questions, and the experimental results are shown in Figure 6 and Table 4. It can be seen that when the maximum number of samples is 1, the convergence speed and final convergence result of the performance of the method of the present invention can be optimal, which further illustrates the importance of information selection among hierarchical memory units.

表4Table 4

最大采样数Maximum number of samples 错误样本数error samples 33 55 22 44 11 00

至此,已经结合附图对本发明实施例进行了详细描述。依据以上描述,本领域技术人员应当对本发明的一种基于层次化记忆网络的问答方法有了清楚的认识。So far, the embodiments of the present invention have been described in detail with reference to the accompanying drawings. Based on the above description, those skilled in the art should have a clear understanding of the question answering method based on the hierarchical memory network of the present invention.

本发明一种基于层次化记忆网络的问答方法,首先进行句粒度记忆编码,并在问题语义编码的刺激下,通过多轮迭代的注意力机制完成句粒度记忆单元的信息推理,可以提高自动问答的准确性和及时性,有利于低频词和未登录词的答案选择;并通过k最大采样对句子进行筛选,可以提高自动问答的效率,降低计算复杂度,句粒度记忆编码的基础上还进行词粒度记忆编码,即在两个层次进行记忆编码,形成层次化的记忆编码,可以进一步提高自动问答的准确性;利用循环神经网络进行词粒度记忆编码时,是在全句子集合X上操作的,该方式可以在词粒度记忆编码过程中引入词在句子全集合中的上下文环境语义信息,可以提高自动问答的准确性和及时性;词粒度记忆单元上的注意力机制是在k采样后的词粒度记忆单元子集合上运算的,避免了记忆编码中的干扰信息,并减少了词粒度注意力机制的计算量;利用句粒度和词粒度记忆单元联合预测输出词概率分布,可以进一步提高自动问答的准确性,有效解决了低频词和未登录词的答案选择问题。A question-and-answer method based on a hierarchical memory network of the present invention first performs sentence-granularity memory encoding, and under the stimulation of question semantic encoding, completes the information reasoning of sentence-granularity memory units through a multi-round iterative attention mechanism, which can improve automatic question-answering The accuracy and timeliness of the method are beneficial to the answer selection of low-frequency words and unregistered words; and the k-maximum sampling is used to screen sentences, which can improve the efficiency of automatic question answering and reduce computational complexity. On the basis of sentence granularity memory coding, it can also perform Word-granular memory encoding, that is, memory encoding at two levels to form a hierarchical memory encoding, which can further improve the accuracy of automatic question and answer; when using a recurrent neural network for word-granular memory encoding, it is operated on the entire sentence set X , this method can introduce the context semantic information of words in the full set of sentences in the process of word granular memory encoding, which can improve the accuracy and timeliness of automatic question answering; the attention mechanism on the word granular memory unit is after k sampling The operation on the sub-set of word granularity memory unit avoids the interference information in the memory encoding and reduces the calculation amount of the word granularity attention mechanism; using the sentence granularity and word granularity memory unit to jointly predict the output word probability distribution can further improve the automatic The accuracy of the question and answer effectively solves the problem of answer selection for low-frequency words and unregistered words.

需要说明的是,在附图或说明书正文中,未绘示或描述的实现方式,均为所属技术领域中普通技术人员所知的形式,并未进行详细说明。此外,上述对各元件的定义并不仅限于实施例中提到的各种方式,本领域普通技术人员可对其进行简单地更改或替换,例如:It should be noted that, in the accompanying drawings or in the text of the specification, implementations that are not shown or described are forms known to those of ordinary skill in the art, and are not described in detail. In addition, the above definitions of each element are not limited to the various methods mentioned in the embodiments, and those of ordinary skill in the art can easily modify or replace them, for example:

(1)实施例中提到的方向用语,例如“上”、“下”、“前”、“后”、“左”、“右”等,仅是参考附图的方向,并非用来限制本发明的保护范围;(1) The directional terms mentioned in the embodiments, such as "up", "down", "front", "back", "left", "right", etc., are only referring to the directions of the drawings, and are not used to limit The protection scope of the present invention;

(2)上述实施例可基于设计及可靠度的考虑,彼此混合搭配使用或与其他实施例混合搭配使用,即不同实施例中的技术特征可以自由组合形成更多的实施例。(2) The above embodiments can be mixed and matched with each other or with other embodiments based on design and reliability considerations, that is, technical features in different embodiments can be freely combined to form more embodiments.

以上所述的具体实施例,对本发明的目的、技术方案和有益效果进行了详细说明,所应理解的是,以上所述仅为本发明的具体实施例而已,并不用于限制本发明,凡在本发明的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The specific embodiments described above have described the purpose, technical solutions and beneficial effects of the present invention in detail. It should be understood that the above descriptions are only specific embodiments of the present invention, and are not intended to limit the present invention. Within the spirit and principles of the present invention, any modifications, equivalent replacements, improvements, etc., shall be included in the protection scope of the present invention.

Claims (10)

1.一种基于层次化记忆网络的问答方法,其特征在于,包括:1. A question answering method based on a hierarchical memory network, characterized in that, comprising: 步骤S101:融合词的位置和句子的时间序列信息,对句子集合中的句子进行句粒度记忆编码,得到句粒度记忆单元的双通道记忆编码;Step S101: Fusing the position of the word and the time series information of the sentence, performing sentence-granularity memory encoding on the sentences in the sentence set, and obtaining the dual-channel memory encoding of the sentence-granularity memory unit; 步骤S102:在问题语义编码的刺激下,通过多轮迭代的注意力机制完成所述句粒度记忆单元的信息推理,得到所述句粒度记忆单元上在词典维度的输出词概率分布;Step S102: Under the stimulation of the semantic encoding of the question, complete the information reasoning of the sentence granular memory unit through the attention mechanism of multiple iterations, and obtain the output word probability distribution of the sentence granular memory unit in the dictionary dimension; 步骤S103:对所述句粒度记忆单元的信息推理结果进行k最大采样,从所述句子集合中筛选出k最大采样重要句子集合;Step S103: performing k-maximum sampling on the information reasoning results of the sentence-granularity memory unit, and selecting the k-maximum-sampling important sentence set from the sentence set; 步骤S104:利用双向循环神经网络模型对所述句子集合进行词粒度记忆编码,得到词粒度记忆单元的记忆编码;Step S104: using a bidirectional recurrent neural network model to perform word-granularity memory encoding on the sentence set to obtain memory encoding of word-granularity memory units; 步骤S105:基于所述问题语义编码、词粒度记忆单元的记忆编码和k最大采样重要句子集合,通过注意力机制得到词粒度输出词概率分布;以及Step S105: Based on the semantic encoding of the problem, the memory encoding of the word granularity memory unit and the k-maximum sampling important sentence set, the word granularity output word probability distribution is obtained through the attention mechanism; and 步骤S106:从句粒度和词粒度记忆单元中联合预测输出词概率分布,并利用交叉熵进行监督训练。Step S106: Jointly predict the output word probability distribution from the sentence-granularity and word-granularity memory units, and use cross-entropy for supervised training. 2.如权利要求1所述的问答方法,其特征在于,所述步骤S101包括:2. The question answering method according to claim 1, characterized in that said step S101 comprises: 子步骤S101a:给定具有时间序列信息的句子集合X={xi}i=(1,2,...,n),随机初始化词向量矩阵句子xi中的词xij的双通道向量化编码为 Sub-step S101a: Given a sentence set X={ xi } i=(1, 2,...,n) with time series information, randomly initialize the word vector matrix and The two-channel vectorization encoding of word x ij in sentence x i is and 其中,i为句子的当前时间序列;n为句子集合的最大时间序列长度;|V|为词典维度;d为词向量的维度;j为词在句子xi中的位置信息;Among them, i is the current time series of the sentence; n is the maximum time series length of the sentence set; |V| is the dictionary dimension; d is the dimension of the word vector; j is the position information of the word in the sentence x i ; 子步骤S101b:根据词在句子中的位置信息对双通道词向量化编码进行更新;以及Sub-step S101b: update the dual-channel word vectorization encoding according to the position information of the word in the sentence; and 子步骤S101c:融合句子的时间序列信息对句子进行句粒度记忆编码,得到句粒度记忆单元的双通道记忆编码。Sub-step S101c: Integrating the time series information of the sentence to perform sentence-granularity memory encoding on the sentence to obtain the dual-channel memory encoding of the sentence-granularity memory unit. 3.如权利要求2所述的问答方法,其特征在于,所述子步骤S101b包括:3. The question answering method according to claim 2, characterized in that, said substep S101b comprises: 更新后的双通道词向量化编码为lgj·(Axij)和lgj·(Cxij),其中,The updated two-channel word vectorization is encoded as l gj (Ax ij ) and l gj (Cx ij ), where, lgj=(1-j/Ji)-(g/d)(1-2j/Ji) (1)l gj =(1-j/J i )-(g/d)(1-2j/J i ) (1) 其中,Ji是句子xi中词的个数,而g为维度为d的词向量中的当前维度值,且1≤j≤J和1≤g≤d。Among them, J i is the number of words in the sentence x i , and g is the current dimension value in the word vector with dimension d, and 1≤j≤J and 1≤g≤d. 4.如权利要求3所述的问答方法,其特征在于,所述子步骤S101c包括:4. The question answering method according to claim 3, characterized in that, said sub-step S101c comprises: 随机初始化句子的时间向量矩阵则句粒度记忆单元的双通道记忆编码为M(S)={{ai},{ci}},其中,Randomly initialize the matrix of time vectors for sentences and Then the dual-channel memory encoding of the sentence granularity memory unit is M (S) = {{a i }, { ci }}, where, ai=∑jlj·(Axij)+TA(i) (2)a i =∑ j l j ·(Ax ij )+T A (i) (2) ci=∑jlj·(Cxij)+TC(i) (3)c i =∑ j l j ·(Cx ij )+T C (i) (3) 其中,lj为中更新矩阵l在句子xi中第j个词的更新向量;操作符·为向量间元素乘法操作;n为句子集合的最大时间序列长度;d为时间向量维度,与词向量的维度相同。Among them, l j is the update vector of the jth word in the sentence x i in the update matrix l; the operator is the element multiplication operation between vectors; n is the maximum time series length of the sentence set; d is the time vector dimension, and the word The vectors have the same dimensions. 5.如权利要求4所述的问答方法,其特征在于,所述步骤S102包括:5. The question answering method according to claim 4, characterized in that said step S102 comprises: 子步骤S102a:利用词向量矩阵对问题文本q中第j个词qj进行向量化表示得到问题语义编码:Sub-step S102a: use word vector matrix Vectorized representation of the jth word q j in the question text q Get the problem semantic encoding: uu 11 (( SS )) == ΣΣ jj ll jj ·&Center Dot; (( AqQ jj )) -- -- -- (( 44 )) 其中,lj为更新矩阵l在句子xi中第j个词的更新向量;Among them, l j is the update vector of the jth word in the sentence x i of the update matrix l; 子步骤S102b:计算问题语义编码在句粒度记忆单元的注意力权重 Sub-step S102b: Calculating the attention weight of the question semantic code in the sentence granularity memory unit αα ii (( SS )) == sthe s oo ff tt mm aa xx (( aa ii TT uu 11 (( SS )) )) -- -- -- (( 55 )) 在问题语义编码刺激下,句粒度记忆单元的双通道记忆编码的激活信息为:以及Under the stimulation of question semantic encoding, the activation information of the dual-channel memory encoding of the sentence granularity memory unit is: as well as 子步骤S102c:通过多轮迭代的注意力机制完成在句粒度记忆单元的信息推理,得到句粒度记忆单元上在词典维度的输出词概率分布。Sub-step S102c: complete the information reasoning in the sentence granular memory unit through the multi-round iterative attention mechanism, and obtain the output word probability distribution in the dictionary dimension on the sentence granular memory unit. 6.如权利要求5所述的问答方法,其特征在于,所述子步骤S102c包括:6. The question answering method according to claim 5, wherein said sub-step S102c comprises: 在句粒度记忆单元上进行R轮信息激活,得到第R轮的激活信息OR,其中,在第r+1轮信息激活中,Perform R rounds of information activation on the sentence granularity memory unit to obtain the Rth round of activation information OR , where, in the r +1th round of information activation, uu rr ++ 11 (( SS )) == oo rr ++ uu rr (( SS )) -- -- -- (( 66 )) αα ii (( SS )) == sthe s oo ff tt maxmax (( aa ii TT uu rr ++ 11 (( SS )) )) -- -- -- (( 77 )) aa ii == ΣΣ jj ll jj ·&Center Dot; (( AA rr ++ 11 xx ii jj )) ++ TT AA rr ++ 11 (( ii )) -- -- -- (( 88 )) oo rr ++ 11 == ΣΣ ii αα ii (( SS )) cc ii -- -- -- (( 99 )) cc ii == ΣΣ jj ll jj ·· (( CC rr ++ 11 xx ii jj )) ++ TT CC rr ++ 11 (( ii )) -- -- -- (( 1010 )) 其中,1≤r≤(R-1);Ar+1=Cr Among them, 1≤r≤(R-1); A r+1 =C r , 句粒度记忆单元上在词典维度的输出词概率分布为:The output word probability distribution in the dictionary dimension on the sentence granularity memory unit is: pp (( SS )) (( ww )) == sthe s oo ff tt mm aa xx (( (( CC RR )) rr (( oo RR ++ uu RR (( SS )) )) )) -- -- -- (( 1111 )) 其中,w={wt}t=(1,2,...,|V|)为词典维度词集合;为第R轮信息激活的词向量矩阵;T为转置操作符。Wherein, w={w t } t=(1, 2, ..., |V|) is a dictionary dimension word set; The word vector matrix activated for the R-th round of information; T is the transpose operator. 7.如权利要求6所述的问答方法,其特征在于,所述步骤S103包括:7. The question answering method according to claim 6, characterized in that said step S103 comprises: 子步骤S103a:对句粒度记忆单元上第R轮信息激活的注意力权重向量从中选取其k个最大的注意力权重子集合以及Sub-step S103a: the attention weight vector activated on the R-th round of information on the sentence granularity memory unit Select the k largest attention weight sub-sets from it as well as 子步骤S103b:选取k个最大的注意力权重子集合对应的句子集合作为k最大采样重要句子集合 Sub-step S103b: Select the k largest attention weight sub-sets The corresponding sentence set is used as the k-maximum sampling important sentence set 8.如权利要求7所述的问答方法,其特征在于,所述步骤S104包括:8. The question answering method according to claim 7, wherein said step S104 comprises: 子步骤S104a:利用门循环网络模型分别对句子集合X中的所有词按时间序列进行正向和反向编码,对于t时刻的词特征,前向GRU编码的隐状态为后向GRU编码的隐状态为 Sub-step S104a: Use the gated recurrent network model to classify all the words in the sentence set X Forward and reverse encoding according to time series, for word features at time t, the hidden state of forward GRU encoding is The hidden state of the backward GRU encoding is 其中,|t|为按照时间序列对句子集合X中所有词进行排列后的词最大序列长度;的维度与词向量的维度d相同;Among them, |t| is the maximum sequence length of words after arranging all the words in the sentence set X according to the time sequence; and The dimension of is the same as the dimension d of the word vector; 子步骤S104b:得到词粒度记忆单元的记忆编码M(W)={mt}t=(1,2,...,|t|),其中 Sub-step S104b: Obtain the memory code M (W) ={m t } t=(1, 2, ..., |t|) of the word granularity memory unit, where 9.如权利要求8所述的问答方法,其特征在于,所述步骤S105包括:9. The question answering method according to claim 8, characterized in that said step S105 comprises: 子步骤S105a:计算归一化后的词粒度记忆单元的注意力权重向量其中:Sub-step S105a: Calculating the attention weight vector of the normalized word granularity memory unit in: αα tt (( WW )) == sthe s oo ff tt mm aa xx (( vv TT tanhtanh (( WuWu RR (( SS )) ++ Uu mm ^^ tt )) )) -- -- -- (( 1212 )) 其中,是k最大采样重要句子集合中的词集合所对应的词粒度记忆编码M(W)={mt}t=(1,2,...,|t|)中的子集合注意力权重向量α(W)的维度为 为学习参数;in, is the k-maximally sampled set of important sentences set of words in The corresponding word-granularity memory code M (W) = {m t } t = (1, 2, ..., |t|) in the subset The dimension of the attention weight vector α (W) is and is the learning parameter; 子步骤S105b:词粒度输出词概率分布为:Sub-step S105b: word granularity output word probability distribution for: pp (( WW )) (( ww ^^ )) == αα (( WW )) -- -- -- (( 1313 )) 其中, 为重要句子集合中所有词的集合 in, is the set of all words in the important sentence set 10.如权利要求9所述的问答方法,其特征在于,所述步骤S106包括:10. The question answering method according to claim 9, wherein said step S106 comprises: 子步骤S106a:基于句粒度记忆单元上在词典维度的输出词概率分布和词粒度输出词概率分布进行输出词联合预测,联合预测输出词分布p(w)表达式为:Sub-step S106a: Based on the output word probability distribution in the dictionary dimension and the word granularity output word probability distribution on the sentence granularity memory unit, the output word joint prediction is performed, and the expression of the joint prediction output word distribution p(w) is: pp (( ww )) == pp (( SS )) (( ww )) ++ pp (( WW )) (( ww )) == pp (( SS )) (( ww )) ++ tt rr aa nno sthe s (( pp (( WW )) (( ww ^^ )) )) -- -- -- (( 1414 )) 其中,trans(·)表示将子集合的词粒度输出词概率分布映射到词典维度全集合的词粒度输出词概率分布 Among them, trans( ) indicates that the word granularity of the sub-set is output to the word probability distribution Word granularity output word probability distribution mapped to the full set of dictionary dimensions 子步骤S106b:利用目标答案词分布对联合预测输出词分布进行交叉熵监督训练。Sub-step S106b: Perform cross-entropy supervised training on the joint prediction output word distribution by using the target answer word distribution.
CN201610447676.4A 2016-06-20 2016-06-20 A kind of answering method based on stratification memory network Active CN106126596B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610447676.4A CN106126596B (en) 2016-06-20 2016-06-20 A kind of answering method based on stratification memory network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610447676.4A CN106126596B (en) 2016-06-20 2016-06-20 A kind of answering method based on stratification memory network

Publications (2)

Publication Number Publication Date
CN106126596A true CN106126596A (en) 2016-11-16
CN106126596B CN106126596B (en) 2019-08-23

Family

ID=57470348

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610447676.4A Active CN106126596B (en) 2016-06-20 2016-06-20 A kind of answering method based on stratification memory network

Country Status (1)

Country Link
CN (1) CN106126596B (en)

Cited By (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106776578A (en) * 2017-01-03 2017-05-31 竹间智能科技(上海)有限公司 Talk with the method and device of performance for lifting conversational system
CN106778014A (en) * 2016-12-29 2017-05-31 浙江大学 A kind of risk Forecasting Methodology based on Recognition with Recurrent Neural Network
CN107273487A (en) * 2017-06-13 2017-10-20 北京百度网讯科技有限公司 Generation method, device and the computer equipment of chat data based on artificial intelligence
CN107491541A (en) * 2017-08-24 2017-12-19 北京丁牛科技有限公司 File classification method and device
CN107766506A (en) * 2017-10-20 2018-03-06 哈尔滨工业大学 A kind of more wheel dialog model construction methods based on stratification notice mechanism
CN107818306A (en) * 2017-10-31 2018-03-20 天津大学 A kind of video answering method based on attention model
CN107844533A (en) * 2017-10-19 2018-03-27 云南大学 A kind of intelligent Answer System and analysis method
CN108108428A (en) * 2017-12-18 2018-06-01 苏州思必驰信息科技有限公司 A kind of method, input method and system for building language model
CN108388561A (en) * 2017-02-03 2018-08-10 百度在线网络技术(北京)有限公司 Neural network machine interpretation method and device
CN108417210A (en) * 2018-01-10 2018-08-17 苏州思必驰信息科技有限公司 A word embedding language model training method, word recognition method and system
CN108549850A (en) * 2018-03-27 2018-09-18 联想(北京)有限公司 A kind of image-recognizing method and electronic equipment
CN108628935A (en) * 2018-03-19 2018-10-09 中国科学院大学 A kind of answering method based on end-to-end memory network
CN108959246A (en) * 2018-06-12 2018-12-07 北京慧闻科技发展有限公司 Answer selection method, device and electronic equipment based on improved attention mechanism
CN109033463A (en) * 2018-08-28 2018-12-18 广东工业大学 A kind of community's question and answer content recommendation method based on end-to-end memory network
CN109388706A (en) * 2017-08-10 2019-02-26 华东师范大学 A kind of problem fine grit classification method, system and device
CN109558487A (en) * 2018-11-06 2019-04-02 华南师范大学 Document Classification Method based on the more attention networks of hierarchy
CN109597884A (en) * 2018-12-28 2019-04-09 北京百度网讯科技有限公司 Talk with method, apparatus, storage medium and the terminal device generated
CN109614473A (en) * 2018-06-05 2019-04-12 安徽省泰岳祥升软件有限公司 Knowledge reasoning method and device applied to intelligent interaction
CN109658270A (en) * 2018-12-19 2019-04-19 前海企保科技(深圳)有限公司 It is a kind of to read the core compensation system and method understood based on insurance products
CN109829631A (en) * 2019-01-14 2019-05-31 北京中兴通网络科技股份有限公司 A kind of business risk early warning analysis method and system based on memory network
CN109840322A (en) * 2018-11-08 2019-06-04 中山大学 It is a kind of based on intensified learning cloze test type reading understand analysis model and method
CN109977428A (en) * 2019-03-29 2019-07-05 北京金山数字娱乐科技有限公司 A kind of method and device that answer obtains
CN109992657A (en) * 2019-04-03 2019-07-09 浙江大学 A Conversational Question Generation Method Based on Enhanced Dynamic Reasoning
CN110019719A (en) * 2017-12-15 2019-07-16 微软技术许可有限责任公司 Based on the question and answer asserted
CN110046244A (en) * 2019-04-24 2019-07-23 中国人民解放军国防科技大学 An Answer Selection Method for Question Answering System
CN110134771A (en) * 2019-04-09 2019-08-16 广东工业大学 An Implementation Method of Fusion Network Question Answering System Based on Multi-Attention Mechanism
CN110147532A (en) * 2019-01-24 2019-08-20 腾讯科技(深圳)有限公司 Coding method, device, equipment and storage medium
CN110334195A (en) * 2019-06-26 2019-10-15 北京科技大学 A question answering method and system based on local attention mechanism memory network
CN110348462A (en) * 2019-07-09 2019-10-18 北京金山数字娱乐科技有限公司 A kind of characteristics of image determination, vision answering method, device, equipment and medium
CN110389996A (en) * 2018-04-16 2019-10-29 国际商业机器公司 Realize the full sentence recurrent neural network language model for being used for natural language processing
CN110555097A (en) * 2018-05-31 2019-12-10 罗伯特·博世有限公司 Slot filling with joint pointer and attention in spoken language understanding
CN110866403A (en) * 2018-08-13 2020-03-06 中国科学院声学研究所 End-to-end conversation state tracking method and system based on convolution cycle entity network
CN111047482A (en) * 2019-11-14 2020-04-21 华中师范大学 Knowledge tracking system and method based on hierarchical memory network
CN111291803A (en) * 2020-01-21 2020-06-16 中国科学技术大学 A kind of image grading granularity migration method, system, device and medium
CN111310848A (en) * 2020-02-28 2020-06-19 支付宝(杭州)信息技术有限公司 Training method and device of multi-task model
CN112732879A (en) * 2020-12-23 2021-04-30 重庆理工大学 Downstream task processing method and model of question-answering task
CN113704437A (en) * 2021-09-03 2021-11-26 重庆邮电大学 Knowledge base question-answering method integrating multi-head attention mechanism and relative position coding

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104834747A (en) * 2015-05-25 2015-08-12 中国科学院自动化研究所 Short text classification method based on convolution neutral network
CN105159890A (en) * 2014-06-06 2015-12-16 谷歌公司 Generating representations of input sequences using neural networks

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105159890A (en) * 2014-06-06 2015-12-16 谷歌公司 Generating representations of input sequences using neural networks
CN104834747A (en) * 2015-05-25 2015-08-12 中国科学院自动化研究所 Short text classification method based on convolution neutral network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SAINBAYAR SUKHBAATAR ET AL.: "End-To-End Memory Networks", 《ARXIV:1503.08895V5》 *
SARATH CHANDAR ET AL.: "Hierarchical Memory Networks", 《ARXIV:1605.07427V1》 *

Cited By (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106778014A (en) * 2016-12-29 2017-05-31 浙江大学 A kind of risk Forecasting Methodology based on Recognition with Recurrent Neural Network
CN106778014B (en) * 2016-12-29 2020-06-16 浙江大学 Disease risk prediction modeling method based on recurrent neural network
CN106776578A (en) * 2017-01-03 2017-05-31 竹间智能科技(上海)有限公司 Talk with the method and device of performance for lifting conversational system
US11403520B2 (en) 2017-02-03 2022-08-02 Baidu Online Network Technology (Beijing) Co., Ltd. Neural network machine translation method and apparatus
CN108388561A (en) * 2017-02-03 2018-08-10 百度在线网络技术(北京)有限公司 Neural network machine interpretation method and device
CN108388561B (en) * 2017-02-03 2022-02-25 百度在线网络技术(北京)有限公司 Neural network machine translation method and device
CN107273487A (en) * 2017-06-13 2017-10-20 北京百度网讯科技有限公司 Generation method, device and the computer equipment of chat data based on artificial intelligence
US10762305B2 (en) 2017-06-13 2020-09-01 Beijing Baidu Netcom Science And Technology Co., Ltd. Method for generating chatting data based on artificial intelligence, computer device and computer-readable storage medium
CN109388706A (en) * 2017-08-10 2019-02-26 华东师范大学 A kind of problem fine grit classification method, system and device
CN107491541A (en) * 2017-08-24 2017-12-19 北京丁牛科技有限公司 File classification method and device
CN107491541B (en) * 2017-08-24 2021-03-02 北京丁牛科技有限公司 Text classification method and device
CN107844533A (en) * 2017-10-19 2018-03-27 云南大学 A kind of intelligent Answer System and analysis method
CN107766506A (en) * 2017-10-20 2018-03-06 哈尔滨工业大学 A kind of more wheel dialog model construction methods based on stratification notice mechanism
CN107818306B (en) * 2017-10-31 2020-08-07 天津大学 Video question-answering method based on attention model
CN107818306A (en) * 2017-10-31 2018-03-20 天津大学 A kind of video answering method based on attention model
CN110019719B (en) * 2017-12-15 2023-04-25 微软技术许可有限责任公司 Assertion-based question and answer
CN110019719A (en) * 2017-12-15 2019-07-16 微软技术许可有限责任公司 Based on the question and answer asserted
CN108108428B (en) * 2017-12-18 2020-05-12 苏州思必驰信息科技有限公司 Method, input method and system for constructing language model
CN108108428A (en) * 2017-12-18 2018-06-01 苏州思必驰信息科技有限公司 A kind of method, input method and system for building language model
CN108417210A (en) * 2018-01-10 2018-08-17 苏州思必驰信息科技有限公司 A word embedding language model training method, word recognition method and system
CN108417210B (en) * 2018-01-10 2020-06-26 苏州思必驰信息科技有限公司 Word embedding language model training method, word recognition method and system
CN108628935B (en) * 2018-03-19 2021-10-15 中国科学院大学 A Question Answering Method Based on End-to-End Memory Network
CN108628935A (en) * 2018-03-19 2018-10-09 中国科学院大学 A kind of answering method based on end-to-end memory network
CN108549850B (en) * 2018-03-27 2021-07-16 联想(北京)有限公司 Image identification method and electronic equipment
CN108549850A (en) * 2018-03-27 2018-09-18 联想(北京)有限公司 A kind of image-recognizing method and electronic equipment
CN110389996A (en) * 2018-04-16 2019-10-29 国际商业机器公司 Realize the full sentence recurrent neural network language model for being used for natural language processing
CN110555097A (en) * 2018-05-31 2019-12-10 罗伯特·博世有限公司 Slot filling with joint pointer and attention in spoken language understanding
CN109614473A (en) * 2018-06-05 2019-04-12 安徽省泰岳祥升软件有限公司 Knowledge reasoning method and device applied to intelligent interaction
CN108959246A (en) * 2018-06-12 2018-12-07 北京慧闻科技发展有限公司 Answer selection method, device and electronic equipment based on improved attention mechanism
CN108959246B (en) * 2018-06-12 2022-07-12 北京慧闻科技(集团)有限公司 Answer selection method and device based on improved attention mechanism and electronic equipment
CN110866403B (en) * 2018-08-13 2021-06-08 中国科学院声学研究所 End-to-end conversation state tracking method and system based on convolution cycle entity network
CN110866403A (en) * 2018-08-13 2020-03-06 中国科学院声学研究所 End-to-end conversation state tracking method and system based on convolution cycle entity network
CN109033463B (en) * 2018-08-28 2021-11-26 广东工业大学 Community question-answer content recommendation method based on end-to-end memory network
CN109033463A (en) * 2018-08-28 2018-12-18 广东工业大学 A kind of community's question and answer content recommendation method based on end-to-end memory network
CN109558487A (en) * 2018-11-06 2019-04-02 华南师范大学 Document Classification Method based on the more attention networks of hierarchy
CN109840322A (en) * 2018-11-08 2019-06-04 中山大学 It is a kind of based on intensified learning cloze test type reading understand analysis model and method
CN109658270A (en) * 2018-12-19 2019-04-19 前海企保科技(深圳)有限公司 It is a kind of to read the core compensation system and method understood based on insurance products
CN109597884A (en) * 2018-12-28 2019-04-09 北京百度网讯科技有限公司 Talk with method, apparatus, storage medium and the terminal device generated
CN109829631A (en) * 2019-01-14 2019-05-31 北京中兴通网络科技股份有限公司 A kind of business risk early warning analysis method and system based on memory network
CN110147532B (en) * 2019-01-24 2023-08-25 腾讯科技(深圳)有限公司 Encoding method, apparatus, device and storage medium
CN110147532A (en) * 2019-01-24 2019-08-20 腾讯科技(深圳)有限公司 Coding method, device, equipment and storage medium
US11995406B2 (en) 2019-01-24 2024-05-28 Tencent Technology (Shenzhen) Company Limited Encoding method, apparatus, and device, and storage medium
CN109977428B (en) * 2019-03-29 2024-04-02 北京金山数字娱乐科技有限公司 Answer obtaining method and device
CN109977428A (en) * 2019-03-29 2019-07-05 北京金山数字娱乐科技有限公司 A kind of method and device that answer obtains
CN109992657A (en) * 2019-04-03 2019-07-09 浙江大学 A Conversational Question Generation Method Based on Enhanced Dynamic Reasoning
CN110134771B (en) * 2019-04-09 2022-03-04 广东工业大学 Implementation method of multi-attention-machine-based fusion network question-answering system
CN110134771A (en) * 2019-04-09 2019-08-16 广东工业大学 An Implementation Method of Fusion Network Question Answering System Based on Multi-Attention Mechanism
CN110046244B (en) * 2019-04-24 2021-06-08 中国人民解放军国防科技大学 An Answer Selection Method for Question Answering System
CN110046244A (en) * 2019-04-24 2019-07-23 中国人民解放军国防科技大学 An Answer Selection Method for Question Answering System
CN110334195A (en) * 2019-06-26 2019-10-15 北京科技大学 A question answering method and system based on local attention mechanism memory network
CN110348462B (en) * 2019-07-09 2022-03-04 北京金山数字娱乐科技有限公司 Image feature determination and visual question and answer method, device, equipment and medium
CN110348462A (en) * 2019-07-09 2019-10-18 北京金山数字娱乐科技有限公司 A kind of characteristics of image determination, vision answering method, device, equipment and medium
CN111047482A (en) * 2019-11-14 2020-04-21 华中师范大学 Knowledge tracking system and method based on hierarchical memory network
CN111291803B (en) * 2020-01-21 2022-07-29 中国科学技术大学 Image grading granularity migration method, system, equipment and medium
CN111291803A (en) * 2020-01-21 2020-06-16 中国科学技术大学 A kind of image grading granularity migration method, system, device and medium
CN111310848B (en) * 2020-02-28 2022-06-28 支付宝(杭州)信息技术有限公司 Training method and device for multi-task model
CN111310848A (en) * 2020-02-28 2020-06-19 支付宝(杭州)信息技术有限公司 Training method and device of multi-task model
CN112732879A (en) * 2020-12-23 2021-04-30 重庆理工大学 Downstream task processing method and model of question-answering task
CN113704437A (en) * 2021-09-03 2021-11-26 重庆邮电大学 Knowledge base question-answering method integrating multi-head attention mechanism and relative position coding
CN113704437B (en) * 2021-09-03 2023-08-11 重庆邮电大学 A Knowledge Base Question Answering Method Fusion of Multi-Head Attention Mechanism and Relative Position Encoding

Also Published As

Publication number Publication date
CN106126596B (en) 2019-08-23

Similar Documents

Publication Publication Date Title
CN106126596B (en) A kind of answering method based on stratification memory network
CN113544703B (en) Efficient off-policy credit allocation
Qin et al. Dynamic fusion network for multi-domain end-to-end task-oriented dialog
Chen et al. Self-play fine-tuning converts weak language models to strong language models
CN108681610B (en) Generative multi-round chat dialogue method, system and computer-readable storage medium
Zhang et al. Deconvolutional paragraph representation learning
US20180329884A1 (en) Neural contextual conversation learning
CN108415923B (en) Closed Domain Intelligent Human-Machine Dialogue System
CN109657041A (en) The problem of based on deep learning automatic generation method
CN111985245A (en) Relation extraction method and system based on attention loop-gated graph convolutional network
US20200134449A1 (en) Training of machine reading and comprehension systems
CN112541060B (en) An end-to-end task-based dialogue learning framework and method based on adversarial training
EP3602411B1 (en) Training neural networks using posterior sharpening
CN107662617A (en) Vehicle-mounted interactive controlling algorithm based on deep learning
CN107992528B (en) Multi-relational question-answering system using interpretable reasoning network
CN111563146B (en) An inference-based method for problem generation with controllable difficulty
CN114912419B (en) Unified machine reading comprehension method based on recombinant adversarial
CN116955576A (en) Question-answer reply method, system and equipment based on human feedback and reinforcement learning
CN115510814A (en) A Method for Generating Complex Questions at the Chapter Level Based on Dual Programming
Ni et al. Enhancing student performance prediction on learnersourced questions with sgnn-llm synergy
Olabiyi et al. Adversarial bootstrapping for dialogue model training
CN117473279A (en) Method and system for unbiasing VQA model during test based on problem type perception
CN115796187A (en) Open domain dialogue method based on dialogue structure diagram constraint
CN111104806A (en) Construction method and device of neural machine translation model, and translation method and device
CN116050523A (en) Attention-directed enhanced common sense reasoning framework based on mixed knowledge graph

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant