CN113988300A - Topic structure reasoning method and system - Google Patents

Topic structure reasoning method and system Download PDF

Info

Publication number
CN113988300A
CN113988300A CN202111281369.0A CN202111281369A CN113988300A CN 113988300 A CN113988300 A CN 113988300A CN 202111281369 A CN202111281369 A CN 202111281369A CN 113988300 A CN113988300 A CN 113988300A
Authority
CN
China
Prior art keywords
reasoning
sentence
level
word
knowledge
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111281369.0A
Other languages
Chinese (zh)
Inventor
余伟江
卢宇彤
郑馥丹
文英鹏
陈志广
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sun Yat Sen University
Original Assignee
Sun Yat Sen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Yat Sen University filed Critical Sun Yat Sen University
Priority to CN202111281369.0A priority Critical patent/CN113988300A/en
Publication of CN113988300A publication Critical patent/CN113988300A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a topic structure reasoning method and a system, wherein the method comprises the following steps: receiving an input text and constructing the mutual relation among contents to obtain a semantic relation; acquiring prior knowledge and fusing the prior knowledge with an input text to obtain implicit knowledge; combining the implicit knowledge and the semantic relation and generating a tree structure target; and analyzing the tree structure target according to a preset traversal sequence to obtain an equation expression. The system comprises: the system comprises a hierarchical reasoning encoder, a knowledge encoder, a tree structure encoder and a tree structure decoder. By using the invention, the construction of the mathematical expression based on the problem description is realized. The question structure reasoning method and the question structure reasoning system can be widely applied to the field of natural language processing.

Description

Topic structure reasoning method and system
Technical Field
The invention relates to the field of natural language processing, in particular to a question structure reasoning method and system.
Background
The math question task is an inference task for answering math inquiry based on question description, and is a cross-disciplinary research topic for connecting math and natural language processing. A question is described by a brief narrative and a question is posed about an unknown number. In recent years, multitasking research based on deep learning methods has received increasing attention. Typically, to solve the reasoning task of such mathematical queries, the user not only needs to parse the question and understand the context, but also needs to exercise external knowledge. However, previous approaches only learn textual descriptions from brief and limited narratives, without using any background knowledge that does not appear in the description, which limits the ability of the model to infer mathematical problems from a global perspective.
Disclosure of Invention
In order to solve the technical problems, the invention aims to provide a topic structure reasoning method and system, and a mathematical expression is constructed based on problem description.
The first technical scheme adopted by the invention is as follows: a topic structure reasoning method comprises the following steps:
receiving an input text and constructing the mutual relation among contents to obtain a semantic relation;
acquiring prior knowledge and fusing the prior knowledge with an input text to obtain implicit knowledge;
combining the implicit knowledge and the semantic relation and generating a tree structure target;
and analyzing the tree structure target according to a preset traversal sequence to obtain an equation expression.
Further, the interrelation between the contents includes a word-level reasoning relation and a sentence-level reasoning relation, and the step of receiving the input text and constructing the interrelation between the contents to obtain the semantic relation specifically includes:
receiving an input text;
constructing a word level reasoning relation based on the word level reasoning layer;
constructing a sentence-level reasoning relation based on the sentence-level reasoning layer;
the word-level reasoning layer and the sentence-level reasoning layer both adopt sequence coding based on GRU.
Further, the step of constructing a word-level inference relationship based on the word-level inference layer specifically includes:
encoding the words based on the bidirectional GRU;
merging the context information into word-level representation to generate word representation;
an attention mechanism is introduced to extract important words, and the importance of the words is measured by using word-level context vectors to obtain word-level reasoning relations;
the important word tokens are aggregated into a sentence vector.
Further, the step of constructing a sentence-level inference relationship based on the sentence-level inference layer specifically includes:
encoding the sentence based on the bidirectional GRU;
merging the information of adjacent sentences into sentence-level representation to generate sentence representation;
and (4) introducing an attention mechanism and measuring the importance of the sentence by using the sentence-level context vector to obtain a sentence-level reasoning relation.
Further, the step of obtaining the prior knowledge and fusing the prior knowledge with the input text to obtain the implicit knowledge specifically includes:
acquiring priori knowledge based on a Chinese pre-training model on a large corpus;
and fusing the prior knowledge and the input text, and extracting to obtain the implicit knowledge of the input text.
Further, the step of combining implicit knowledge and semantic relations and generating a tree structure target specifically includes:
combining implicit knowledge and semantic relations based on dot product self-adaptation, and generating enhanced representation through a linear mapping function;
dividing nodes of the tree structure into mathematical operators, common sense numerical values and numbers;
initializing a root node vector according to the enhanced representation;
iteratively using a root node vector with a trainable vector in conjunction with a target vocabulary pre-prepared with candidate words to predict probabilities belonging to nodes;
and generating tree nodes according to the prediction probability to obtain a tree structure target.
Further, the expression of the enhanced representation is as follows:
Y=F([wpYp,whYh])
in the above formula, Y represents an enhancement expression, Y representspRepresenting a semantic relationship, YhRepresenting implicit knowledge, wpAnd whRepresenting the corresponding importance [. ]]Representing a chaining operation and F a linear mapping function.
Further, the step of analyzing the tree structure target according to a predetermined traversal order to obtain an equation expression specifically includes:
analyzing the tree structure target;
generating an intermediate operator according to the topmost node of the tree structure target;
and recursively completing the analysis of all the nodes from the left child node to the right child node to obtain an equation expression.
The second technical scheme adopted by the invention is as follows: a topic structure reasoning system comprising:
the hierarchical reasoning encoder is used for receiving an input text and constructing the mutual relation among contents to obtain a semantic relation;
the knowledge encoder is used for acquiring the prior knowledge and fusing the prior knowledge with the input text to obtain the implicit knowledge;
the tree structure encoder is used for combining the implicit knowledge and the semantic relation and generating a tree structure target;
a tree structure decoder; and the method is used for analyzing the tree structure target according to a preset traversal sequence to obtain an equation expression.
The method and the system have the beneficial effects that: the method effectively fuses the implicit knowledge into the model based on the knowledge encoder, can help the model to correctly analyze the semantics of words from complex texts, constructs the relation between words and sentences based on the hierarchical reasoning encoder, realizes the connection between an entity domain and a context domain, and further realizes the construction of a mathematical expression based on problem description.
Drawings
FIG. 1 is a flow chart of the steps of a topic structure inference method of the present invention;
FIG. 2 is a block diagram of a topic structure inference system of the present invention;
FIG. 3 is a schematic flow chart of an embodiment of the present invention.
FIG. 4 is a diagram of hierarchical inference according to an embodiment of the present invention.
Detailed Description
The invention is described in further detail below with reference to the figures and the specific embodiments. The step numbers in the following embodiments are provided only for convenience of illustration, the order between the steps is not limited at all, and the execution order of each step in the embodiments can be adapted according to the understanding of those skilled in the art.
The mathematical problem (MWP) can be represented as (P, E), where P is the problem text and E is the solution expression. Suppose that a description of an MWP has L sentences siEach sentence containing TiA word, wit(t∈[1,T]) Representing the word in the ith sentence. Our proposed encoder projects the original problem description into a vector representation, on the basis of which a tree-structured decoder is constructed to predict the mathematical expression.
Referring to fig. 1 and 3, the present invention provides a topic structure inference method, including the steps of:
s1, receiving the input text and constructing the mutual relation among the contents to obtain the semantic relation;
s2, acquiring prior knowledge and fusing the prior knowledge with the input text to obtain implicit knowledge;
s3, combining the implicit knowledge and the semantic relation and generating a tree structure target;
and S4, analyzing the tree structure target according to a preset traversal sequence to obtain an equation expression.
As a further preferred embodiment of the method, the interrelation between the contents includes a word-level inference relationship and a sentence-level inference relationship, and the step of receiving the input text and constructing the interrelation between the contents to obtain the semantic relationship specifically includes:
s11, receiving an input text;
s12, constructing a word level reasoning relation based on the word level reasoning layer;
s13, constructing sentence-level reasoning relation based on the sentence-level reasoning layer;
the word-level reasoning layer and the sentence-level reasoning layer both adopt sequence coding based on GRU.
In particular, this step is implemented by means of a hierarchical inference encoder, a schematic diagram of which, with reference to fig. 4, takes into account that the different parts of the mathematical description do not have similar relevant information. Furthermore, determining relevant parts involves modeling the interactions between words, not just their isolated presence in the text. Thus, in this regard, the model includes two levels of reasoning mechanisms. One is word-level reasoning and the other is sentence-level reasoning, which allows the model to focus more or less on individual words and sentences in building the entire descriptive representation. The hierarchical inference encoder consists of two layers. The first layer is a word-level inference layer and the second layer is a sentence-level inference layer. Both of these two layer generalizations use GRU-based sequence coding.
Based on the coding of the sequence of the GRU, the GRU uses a gating mechanism to track the state of the sequence without using a separate memory unit. There are two types of doors: reset gate rtAnd an update gate zt. They combine to control how information is updated to the state. At time t, the GRU calculates the new state as:
Figure BDA0003331199620000041
this is the last state ht-1And the current new state calculated by using the new sequence information
Figure BDA0003331199620000042
Linear interpolation of (2). Updating the door ztIt is decided how much past information to keep and how much new information to add. z is a radical oftThe updating is as follows:
zt=σ(Wzxt+Uzht-1+bz)
wherein xtIs the sequence vector at time t. Candidate states
Figure BDA0003331199620000043
The calculation is as follows:
Figure BDA0003331199620000044
wherein r istTo reset the gate, the contribution of the previous state to the candidate state is controlled. If r istA value of 0 indicates that the user has forgotten to go. r istThe updating is as follows:
rt=σ(Wrxt+Urht-1+br)
the W and U are learnable matrix weights, and b is a learnable offset vector.
Further, as a preferred embodiment of the method, the step of constructing the word-level inference relationship based on the word-level inference layer specifically includes:
s121, coding the words based on the bidirectional GRU;
s122, combining the context information into word level representation to generate word representation;
s123, extracting important words by introducing an attention mechanism, and measuring the importance of the words by using word-level context vectors to obtain word-level reasoning relations;
and S124, aggregating the important word characteristics into a sentence vector.
In particular, word level reasoning. At this level, the model uses bidirectional GRUs to generate representations of words by integrating information from both directions. Thus, it incorporates context information into the word-level representation. Given a worded word wit,t∈[1,T]Sentences and an embedding matrix WeA bidirectional GRU includes a forward direction
Figure BDA0003331199620000045
For sentence siFrom the word wi1Read the word wiTAnd a reverse direction
Figure BDA0003331199620000051
From the word wiTRead the word wi1
xit=Wewit,t∈[1,T],
Figure BDA0003331199620000052
Figure BDA0003331199620000053
Given word witIs obtained by concatenating the forward hidden state and the backward hidden state, i.e.
Figure BDA0003331199620000054
Figure BDA0003331199620000055
Can combine with witInformation of the entire sentence as a center. Note that not all words have the same effect on the meaning of a representation sentence. Therefore, we introduce an attention mechanism to extract words important to sentences and aggregate the representations of these words with large amounts of information into a sentence vector. In particular, the first and second (c) substrates,
uit=tanh(Wwhit+bw),
Figure BDA0003331199620000056
si=∑αithit.
we first input word-level properties h through a single-layer MLPitTo obtain uitAs hitIs hidden from view. Then using the word-level context vector uwMeasure the importance of the word and obtain a normalized importance weight alpha by a softmax functionit. Then, we compute a weighted sum of word representations as a sentence vector s from the learnable weightsi. In the training process, the context vector u of the wordwPerforming random initialization and joint learning
Further, as a preferred embodiment of the method, the step of constructing the sentence-level inference relationship based on the sentence-level inference layer specifically includes:
s131, coding the sentence based on the bidirectional GRU;
s132, merging the information of the adjacent sentences into sentence level representation to generate sentence representation;
s133, an attention mechanism is introduced, and the importance of the sentence is measured by using the sentence-level context vector, so that a sentence-level reasoning relation is obtained.
Specifically, sentence-level reasoning. Given sentence vector siWe get the problem description vector in a similar way. We encode sentences using bidirectional GRUs:
Figure BDA0003331199620000057
Figure BDA0003331199620000058
wherein the content of the first and second substances,
Figure BDA0003331199620000059
and
Figure BDA00033311996200000510
forward GRU and reverse GRU are indicated, respectively. We will want to
Figure BDA00033311996200000511
And
Figure BDA00033311996200000512
connected to obtain the target representation of sentence i
Figure BDA00033311996200000513
hiThe information of the neighboring sentences of sentence i is integrated, but still focused on sentence i. To reward sentences associated with correctly parsing problem descriptions, we again use the attention mechanism to introduce a sentence-level context vector usTo measure the importance of the sentence, the formula is:
uit=tanh(Wshi+bs),
Figure BDA0003331199620000061
v=∑αihi,
where v is a global text vector that summarizes all the information of the sentence in one description. Also, during the training process, the sentence-level context vector usRandom initialization and joint learning are also possible.
Further, as a preferred embodiment of the method, the step of obtaining the prior knowledge and fusing the prior knowledge with the input text to obtain the implicit knowledge specifically includes:
s21, acquiring priori knowledge based on a Chinese pre-training model on a large corpus;
and S22, fusing the prior knowledge and the input text, and extracting to obtain the implicit knowledge of the input text.
Specifically, the step is realized by a pre-training knowledge encoder, a language model based on a transform is adopted, an encoder is established, and a Roberta model which is pre-trained on BooksCorpus and Wikipedia of a large corpus is used for pre-training so as to capture implicit knowledge. As in BERT, we mark a description with WordPiece
Figure BDA0003331199620000062
To obtain
Figure BDA0003331199620000063
The sequence of tokens is embedded into a pre-trained Roberta embedding, and the position code of the Roberta is modified to obtain a series of d-dimensional token representations
Figure BDA0003331199620000064
We input this information into a transformer-based pre-trained knowledge encoder, fine-tuning the representation during the training process. We average the outputs of all transform steps to get the combined implicit knowledge representation Yp
Further, as a preferred embodiment of the method, the step of combining implicit knowledge and semantic relations and generating a tree structure target specifically includes:
s31, combining implicit knowledge and semantic relations based on dot product self-adaptation, and generating an enhanced representation through a linear mapping function;
specifically, the result Y is obtained by a pre-training knowledge encoder and a hierarchical inference encoder, respectivelypAnd YhLater, we adaptively merge Y using the two encoder-end parsers as shown in FIG. 1pAnd YhAn enhanced representation Y is obtained for final decoding. The enhancement representation Y may be expressed as:
Y=F([wpYp,whYh]),
wherein wpAnd whFrom YpAnd YhDerived, to calculate the importance of the task. [. the]Indicating a linking operation. We use a simple dot product to merge YpAnd YhThese two representations. The enhanced representation Y is then generated for final decoding using a linear mapping function F, such as a fully connected layer. w is apAnd whCan be calculated as:
Figure BDA0003331199620000065
Figure BDA0003331199620000071
wherein, WpAnd WhAre all trainable weighting matrices that are used in a training process,
Figure BDA0003331199620000072
and
Figure BDA0003331199620000073
representing different MLPs.
S32, combining implicit knowledge and semantic relations based on dot product self-adaptation, and generating an enhanced representation through a linear mapping function;
s33, dividing the nodes of the tree structure into mathematical operators, common sense numerical values and numbers;
s34, initializing a root node vector according to the enhancement representation;
s35, combining a target vocabulary table prepared with candidate words, and predicting the probability of the node by iteratively using a root node vector with a trainable vector;
and S36, generating tree nodes according to the prediction probability to obtain a tree structure target.
In particular, our model initializes the root node vector according to the global context representation Y from both encoders. The expression tree in the decoder contains three types of nodes: mathematical operator VopA common sense value V encountered in the target expression but not present in the problem textcon(e.g., a rabbit has 4 legs) and the number n encountered in question PP. For target expression VtarIs defined as:
Figure BDA0003331199620000074
wherein M isopAnd MconAre two trainable word embedding matrices independent of the specific problem.For one nPFrom the encoder, we adopt the corresponding hidden state
Figure BDA0003331199620000075
As its token embedding, where loc (y, P) is the index position of the value y in P. Mathematical operator VopOccupying the non-leaf position. n isPThe representation of (c) relies on certain MWP descriptions. Since y must derive the corresponding hidden state from the encoder output
Figure BDA0003331199620000076
VopAnd VconAre respectively represented by two embedded matrixes MopAnd MconAre independently obtained.
The tree decoder includes a tree encoder and a tree decoder, and as with the tree encoder, we prepare candidate words for operators and numbers in the target vocabulary, and then iteratively use a root vector with a trainable vector to predict the probability of node token y in the target vocabulary. Then, the specific y with the highest probability is replaced with a tree node according to the rule in equation (19).
Further as a preferred embodiment of the method, the expression of the enhanced representation is as follows:
Y=F([wpYp,whYh])
in the above formula, Y represents an enhancement expression, Y representspRepresenting a semantic relationship, YhRepresenting implicit knowledge, wpAnd whRepresenting the corresponding importance [. ]]Representing a chaining operation and F a linear mapping function.
Further, as a preferred embodiment of the method, the step of analyzing the tree structure target according to a predetermined traversal order to obtain an equation expression specifically includes:
s41, analyzing the tree structure target;
s42, generating an intermediate operator according to the topmost node of the tree structure target;
and S43, recursively completing the analysis of all the nodes from the left child node to the right child node to obtain an equation expression.
In particular, mathematical equations are usually composed of operators and variables. First, defining variables as leaf nodes, each operation node needs two child nodes. Then, the tree structure decoder parses the equation expressions in a predetermined traversal order. The middle most operator is generated first, and then the left child node is generated. The generation process recurses until the last leaf node is generated. Next, the right child node is generated in the same manner.
Further as a preferred embodiment of the method, since the MWP task can be expressed as (P, E), we define its loss function as
Figure BDA0003331199620000081
It can be expressed as predicting the token y of the t nodetIs determined by the sum of the negative log-likelihood probabilities of (c). Formally, the objective function of the training optimizer is:
Figure BDA0003331199620000082
wherein m is the size of E, qtAnd YtThe target vector of the t-th node and the context vector thereof. p is calculated using a distributed computation function in the target-driven tree structure.
As shown in fig. 2, a topic structure inference system includes:
the hierarchical reasoning encoder is used for receiving an input text and constructing the mutual relation among contents to obtain a semantic relation;
the knowledge encoder is used for acquiring the prior knowledge and fusing the prior knowledge with the input text to obtain the implicit knowledge;
the tree structure encoder is used for combining the implicit knowledge and the semantic relation and generating a tree structure target;
a tree structure decoder; and the method is used for analyzing the tree structure target according to a preset traversal sequence to obtain an equation expression.
The contents in the above method embodiments are all applicable to the present system embodiment, the functions specifically implemented by the present system embodiment are the same as those in the above method embodiment, and the beneficial effects achieved by the present system embodiment are also the same as those achieved by the above method embodiment.
While the preferred embodiments of the present invention have been illustrated and described, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (9)

1. A topic structure reasoning method is characterized by comprising the following steps:
receiving an input text and constructing the mutual relation among contents to obtain a semantic relation;
acquiring prior knowledge and fusing the prior knowledge with an input text to obtain implicit knowledge;
combining the implicit knowledge and the semantic relation and generating a tree structure target;
and analyzing the tree structure target according to a preset traversal sequence to obtain an equation expression.
2. The topic structure reasoning method according to claim 1, wherein the interrelations between the contents include a word-level reasoning relationship and a sentence-level reasoning relationship, and the step of receiving the input text and constructing the interrelations between the contents to obtain the semantic relationship specifically comprises:
receiving an input text;
constructing a word level reasoning relation based on the word level reasoning layer;
constructing a sentence-level reasoning relation based on the sentence-level reasoning layer;
combining the word level reasoning relation and the sentence level reasoning relation to obtain a semantic relation;
the word-level reasoning layer and the sentence-level reasoning layer both adopt sequence coding based on GRU.
3. The topic structure reasoning method of claim 2, wherein the step of constructing the word-level reasoning relationship based on the word-level reasoning layer specifically comprises:
encoding the words based on the bidirectional GRU;
merging the context information into word-level representation to generate word representation;
an attention mechanism is introduced to extract important words, and the importance of the words is measured by using word-level context vectors to obtain word-level reasoning relations;
the important word tokens are aggregated into a sentence vector.
4. The topic structure reasoning method of claim 3, wherein the step of constructing a sentence-level reasoning relationship based on a sentence-level reasoning layer specifically comprises:
encoding the sentence based on the bidirectional GRU;
merging the information of adjacent sentences into sentence-level representation to generate sentence representation;
and (4) introducing an attention mechanism and measuring the importance of the sentence by using the sentence-level context vector to obtain a sentence-level reasoning relation.
5. The question structure reasoning method of claim 2, wherein the step of obtaining the prior knowledge and fusing the prior knowledge with the input text to obtain the implicit knowledge specifically comprises:
acquiring priori knowledge based on a Chinese pre-training model on a large corpus;
and fusing the prior knowledge and the input text, and extracting to obtain the implicit knowledge of the input text.
6. The topic structure reasoning method according to claim 3, wherein the step of combining implicit knowledge and semantic relations and generating a tree structure target specifically comprises:
combining implicit knowledge and semantic relations based on dot product self-adaptation, and generating enhanced representation through a linear mapping function;
dividing nodes of the tree structure into mathematical operators, common sense numerical values and numbers;
initializing a root node vector according to the enhanced representation;
iteratively using a root node vector with a trainable vector in conjunction with a target vocabulary pre-prepared with candidate words to predict probabilities belonging to nodes;
and generating tree nodes according to the prediction probability to obtain a tree structure target.
7. The topic structure reasoning method of claim 6, wherein the expression of the enhanced representation is as follows:
Y=F([wpYp,whYh])
in the above formula, Y represents an enhancement expression, Y representspRepresenting a semantic relationship, YhRepresenting implicit knowledge, wpAnd whRepresenting the corresponding importance [. ]]Representing a chaining operation and F a linear mapping function.
8. The topic structure reasoning method of claim 7, wherein the step of resolving the tree structure target according to a predetermined traversal order to obtain an equation expression specifically comprises:
analyzing the tree structure target;
generating an intermediate operator according to the topmost node of the tree structure target;
and recursively completing the analysis of all the nodes from the left child node to the right child node to obtain an equation expression.
9. A topic structure reasoning system, comprising:
the hierarchical reasoning encoder is used for receiving an input text and constructing the mutual relation among contents to obtain a semantic relation;
the knowledge encoder is used for acquiring the prior knowledge and fusing the prior knowledge with the input text to obtain the implicit knowledge;
the tree structure encoder is used for combining the implicit knowledge and the semantic relation and generating a tree structure target;
a tree structure decoder; and the method is used for analyzing the tree structure target according to a preset traversal sequence to obtain an equation expression.
CN202111281369.0A 2021-11-01 2021-11-01 Topic structure reasoning method and system Pending CN113988300A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111281369.0A CN113988300A (en) 2021-11-01 2021-11-01 Topic structure reasoning method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111281369.0A CN113988300A (en) 2021-11-01 2021-11-01 Topic structure reasoning method and system

Publications (1)

Publication Number Publication Date
CN113988300A true CN113988300A (en) 2022-01-28

Family

ID=79745229

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111281369.0A Pending CN113988300A (en) 2021-11-01 2021-11-01 Topic structure reasoning method and system

Country Status (1)

Country Link
CN (1) CN113988300A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114943276A (en) * 2022-04-16 2022-08-26 西北工业大学 Depth knowledge tracking method based on tree attention mechanism
CN116680502A (en) * 2023-08-02 2023-09-01 中国科学技术大学 Intelligent solving method, system, equipment and storage medium for mathematics application questions
CN117272992A (en) * 2023-08-21 2023-12-22 华中师范大学 Mathematical application problem machine answering system and method integrating physical property knowledge prompt

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114943276A (en) * 2022-04-16 2022-08-26 西北工业大学 Depth knowledge tracking method based on tree attention mechanism
CN114943276B (en) * 2022-04-16 2024-03-05 西北工业大学 Depth knowledge tracking method based on tree-type attention mechanism
CN116680502A (en) * 2023-08-02 2023-09-01 中国科学技术大学 Intelligent solving method, system, equipment and storage medium for mathematics application questions
CN116680502B (en) * 2023-08-02 2023-11-28 中国科学技术大学 Intelligent solving method, system, equipment and storage medium for mathematics application questions
CN117272992A (en) * 2023-08-21 2023-12-22 华中师范大学 Mathematical application problem machine answering system and method integrating physical property knowledge prompt

Similar Documents

Publication Publication Date Title
Neubig Neural machine translation and sequence-to-sequence models: A tutorial
CN110334354B (en) Chinese relation extraction method
CN110929515B (en) Reading understanding method and system based on cooperative attention and adaptive adjustment
CN110597947B (en) Reading understanding system and method based on global and local attention interaction
CN113988300A (en) Topic structure reasoning method and system
US20180329884A1 (en) Neural contextual conversation learning
CN108062388A (en) Interactive reply generation method and device
CN111738003B (en) Named entity recognition model training method, named entity recognition method and medium
CN110390397B (en) Text inclusion recognition method and device
CN109214006B (en) Natural language reasoning method for image enhanced hierarchical semantic representation
US20180144234A1 (en) Sentence Embedding for Sequence-To-Sequence Matching in a Question-Answer System
CN106875940B (en) Machine self-learning construction knowledge graph training method based on neural network
CN111241807B (en) Machine reading understanding method based on knowledge-guided attention
CN112527966B (en) Network text emotion analysis method based on Bi-GRU neural network and self-attention mechanism
CN113569932A (en) Image description generation method based on text hierarchical structure
CN112115687A (en) Problem generation method combining triples and entity types in knowledge base
CN112784532B (en) Multi-head attention memory system for short text sentiment classification
Tang et al. Modelling student behavior using granular large scale action data from a MOOC
CN111259632A (en) Semantic alignment-based tree structure mathematical application problem solving method and system
CN113254581B (en) Financial text formula extraction method and device based on neural semantic analysis
US20220129450A1 (en) System and method for transferable natural language interface
CN112069827B (en) Data-to-text generation method based on fine-grained subject modeling
CN115186147B (en) Dialogue content generation method and device, storage medium and terminal
CN113591482A (en) Text generation method, device, equipment and computer readable storage medium
CN111145914B (en) Method and device for determining text entity of lung cancer clinical disease seed bank

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination