CN111061851B - Question generation method and system based on given facts - Google Patents

Question generation method and system based on given facts Download PDF

Info

Publication number
CN111061851B
CN111061851B CN201911276552.4A CN201911276552A CN111061851B CN 111061851 B CN111061851 B CN 111061851B CN 201911276552 A CN201911276552 A CN 201911276552A CN 111061851 B CN111061851 B CN 111061851B
Authority
CN
China
Prior art keywords
question
input information
history
sequence
generation model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911276552.4A
Other languages
Chinese (zh)
Other versions
CN111061851A (en
Inventor
刘康
何世柱
赵军
刘操
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201911276552.4A priority Critical patent/CN111061851B/en
Publication of CN111061851A publication Critical patent/CN111061851A/en
Application granted granted Critical
Publication of CN111061851B publication Critical patent/CN111061851B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation
    • G06N5/022Knowledge engineering; Knowledge acquisition
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention relates to a question generation method and a question generation system based on given facts, wherein the question generation method comprises the following steps: acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users; expanding each history input information to obtain a corresponding context representation; establishing a question generation model according to each piece of input information and the corresponding context representation; and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model. According to the invention, a question generation model is established through historical reference data; the question generation model can be based on the question generation model, and the question sequence corresponding to the current input information can be accurately determined according to a small amount of current input information given by the current user.

Description

Question generation method and system based on given facts
Technical Field
The invention relates to the technical field of natural language processing, in particular to a question generation method and a question generation system based on given facts.
Background
With the vigorous development of the internet and the increasing popularity of network communication terminals, people are exposed to massive amounts of information related to various fields every day. Knowledge base questions and answers can help people to quickly acquire knowledge from massive information, so that the learning cost of human beings is reduced. However, knowledge base question-answering is severely dependent on manual annotation data, and the annotation data of question-answering pairs (pairs) becomes bottleneck resources for restricting the development of question-sentence technology and question-answering systems, so that question-sentence generation can effectively solve the problem.
The task of question generation mainly automatically generates questions from given answers and auxiliary information thereof. The given answer and ancillary information may be in plain text form or may be a structured knowledge base. The question generation has the following purposes: 1. automatically constructing data resources of questions and answers, or reducing the workload of manually marking question and answer pairs; 2. the method is used for data enhancement and improves the performance of a question-answering system; 3. as a typical text generation task, development and progress of text generation technology can be facilitated.
However, the conventional question generation method easily generates a question whose predicate is not matched, such as given input < statue of liberty, position, new york city > in table 1, may generate a question whose given predicate cannot be expressed like Q1 (who authored statue of liberty.
TABLE 1
Disclosure of Invention
In order to solve the above problems in the prior art, that is, to solve the problem of accurately determining based on a small number of given facts, the present invention provides a question generation method and system based on the given facts.
In order to solve the technical problems, the invention provides the following scheme:
a question generation method based on given facts, the question generation method comprising:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
Optionally, the history reference data further includes a plurality of pieces of supervision information, and each piece of supervision information includes a manually noted question and a reference answer corresponding to the history input information;
the question generation method further comprises the following steps:
and correcting the question generation model according to the supervision information to obtain a corrected question generation model.
Optionally, the correcting the question generation model according to the supervision information to obtain a corrected question generation model specifically includes:
based on the question generation model, determining a corresponding historical question sequence according to each historical input information;
according to each history question sequence and the corresponding manual labeling question, calculating and generating question loss
Calculating auxiliary answer loss according to each history question sequence and the corresponding reference answer
Wherein each reference answer includes answer type words corresponding to the history input information, the history question sequence includes generated words corresponding to the history input information,is a set of answer type words, and A represents the answer type words in the set of answer type wordsQuantity of->Is a generated word y in a question sequence t And the corresponding answer type word a n Is a loss of (2);
according to the generated question lossAuxiliary answer loss->Determining supervisory information loss
Wherein λ represents a reference coefficient;
according to the supervision information lossAnd correcting the question generation model to obtain a corrected question generation model.
Optionally, the format of the history input information of the different users is a head entity-relation-tail entity;
the expanding of each history input information to obtain a corresponding context representation specifically comprises:
for the head entity and/or the tail entity, using the type information in the knowledge base as a context representation of the head entity and/or the tail entity;
for a relationship, at least one of a sentence corresponding to a domain, a value domain, a topic, and a distance supervision back label in a knowledge base is used as a contextual representation of the relationship.
Optionally, when the knowledge base has multiple types of information, the most frequently used and most differentiated type is selected as the context representation of the head entity and/or the tail entity.
Optionally, the establishing a question generation model according to each input information and the corresponding context representation specifically includes:
for each pair of input information and corresponding context representation,
training the input information to obtain training information;
based on a first sequence model, obtaining a representation sequence according to the context representation;
fusing the training information and the representation sequence to obtain fusion information;
coding the fusion information to obtain a hidden layer sequence;
and decoding each hidden layer sequence, and calculating to obtain a corresponding decoding sequence function, wherein the decoding sequence function is a question generation model.
Optionally, decoding each hidden layer sequence, and calculating to obtain a corresponding decoding sequence function, which specifically includes:
decoding each hidden layer state sequence based on a second sequence model to obtain decoding information;
according to the decoding information, respectively calculating the knowledge base copy mode probability of the names corresponding to the copy history input information in the knowledge base, the context copy mode probability of the copy context representation and the vocabulary generation mode probability of the words generated from the vocabulary;
according to the knowledge base copy mode probability p cpkb Context copy mode probability p cptx Vocabulary generation pattern probability p genv Calculating the predictive probability P (y) t |s t ,y t-1 ,F,C):
Wherein genev, cpkb and cpctx represent vocabulary generation mode, knowledge base replication mode and context replication mode, respectively, p.represents three different speciesProbability of pattern, P (x) represents probability of generating target word in various patterns, F and C represent input information and context, s, respectively t Representing the current decoding status, y t Representing words generated at the current moment;
according to the predictive probability P (y) t |s t ,y t-1 F, C), decoding word by word to obtain a decoded question sequence function.
In order to solve the technical problems, the invention also provides the following scheme:
a question generation system based on a given fact, the question generation system comprising:
the acquisition unit is used for acquiring history reference data, wherein the history reference data comprises a plurality of pieces of history input information of different users;
the expansion unit is used for expanding each history input information to obtain a corresponding context representation;
the modeling unit is used for establishing a question generation model according to the input information and the corresponding context representation;
and the determining unit is used for determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
In order to solve the technical problems, the invention also provides the following scheme:
a question generation system based on given facts, comprising:
a processor; and
a memory arranged to store computer executable instructions that, when executed, cause the processor to:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
In order to solve the technical problems, the invention also provides the following scheme:
a computer-readable storage medium storing one or more programs that, when executed by an electronic device comprising a plurality of application programs, cause the electronic device to:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
According to the embodiment of the invention, the following technical effects are disclosed:
according to the invention, a question generation model is established through historical reference data; the question generation model can be based on the question generation model, and the question sequence corresponding to the current input information can be accurately determined according to a small amount of current input information given by the current user.
Drawings
FIG. 1 is a flow chart of a question generation method based on a given fact of the present invention;
FIG. 2 is a schematic diagram of answer-assisted supervision
Fig. 3 is a schematic block diagram of a question generation system based on given facts of the present invention.
Symbol description:
the system comprises an acquisition unit-1, an expansion unit-2, a modeling unit-3 and a determination unit-4.
Detailed Description
Preferred embodiments of the present invention are described below with reference to the accompanying drawings. It should be understood by those skilled in the art that these embodiments are merely for explaining the technical principles of the present invention, and are not intended to limit the scope of the present invention.
The invention aims to provide a question generation method based on given facts, which establishes a question generation model through historical reference data; the question generation model can be based on the question generation model, and the question sequence corresponding to the current input information can be accurately determined according to a small amount of current input information given by the current user.
In order that the above-recited objects, features and advantages of the present invention will become more readily apparent, a more particular description of the invention will be rendered by reference to the appended drawings and appended detailed description.
As shown in fig. 1, the question generation method based on the given facts of the present invention includes:
step 100: acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users;
step 200: expanding each history input information to obtain a corresponding context representation;
step 300: establishing a question generation model according to each piece of input information and the corresponding context representation;
step 400: and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
Further, the history reference data further comprises a plurality of pieces of supervision information, and each piece of supervision information comprises a manually marked question and a reference answer corresponding to the history input information.
In order to improve the determination accuracy, the question generation method based on the given facts further comprises the following steps:
and correcting the question generation model according to the supervision information to obtain a corrected question generation model.
The traditional method only has supervision information of question ends, namely, the generated question and the manually marked question are compared, the difference between the questions is taken as loss, and the training is carried out by using an optimizer, so that the ambiguous question, namely, one question corresponds to a plurality of correct answers, can be easily obtained. For this purpose, in addition to manually annotating question sentences as supervision information, answer information is also utilized as auxiliary supervision, with the goal that the generated question sentences contain any one answer type word.
Further, the step of correcting the question generation model according to the supervision information to obtain a corrected question generation model specifically includes:
step S1: based on the question generation model, corresponding historical question sequences are determined according to each historical input information.
Step S2: according to each history question sequence and the corresponding manual labeling question, calculating and generating question loss
Step S3: calculating auxiliary answer loss according to each history question sequence and the corresponding reference answer(as shown in fig. 2):
wherein each reference answer includes answer type words corresponding to the history input information, the history question sequence includes generated words corresponding to the history input information,is a set of answer type words (e.g. type words of new york city include cities, administrative areas, etc.), |a| represents the number of answer type words in the set of answer type words, ++>Is a generated word y in a question sequence t And the corresponding answer type word a n Is a loss of (2). Wherein, auxiliary answer is lost->Is the minimum of two calculations.
Step S4: according to the generated question lossAuxiliary answer loss->Determining supervision information loss->
Where λ represents a reference coefficient.
Step S5: according to the supervision information lossAnd correcting the question generation model to obtain a corrected question generation model.
In this embodiment, the format of the history input information of the different users is a head entity (subject) -relationship (predicate) -tail entity (object).
Further, in step 200, the expanding each history input information to obtain a corresponding context representation specifically includes:
for the head entity and/or the tail entity, using the type information in the knowledge base as a context representation of the head entity and/or the tail entity;
for a relationship, at least one of a sentence corresponding to a domain, a value domain, a topic, and a distance supervision back label in a knowledge base is used as a contextual representation of the relationship.
Wherein when there are a plurality of types of information of the knowledge base, the most frequently used and most differentiated type is selected as the context representation of the head entity and/or the tail entity.
In step 300, a question generation model is built according to each input information and the corresponding context representation, which specifically includes:
for each pair of input information and corresponding context representation,
training the input information to obtain training information;
based on a first sequence model, obtaining a representation sequence according to the context representation;
fusing the training information and the representation sequence to obtain fusion information;
coding the fusion information to obtain a hidden layer sequence;
and decoding each hidden layer sequence, and calculating to obtain a corresponding decoding sequence function, wherein the decoding sequence function is a question generation model.
For the fact of input symbolization, a knowledge base such as TransE (transformation) can be used for representing a learning method to pretrain on a large-scale corpus, and the training can be carried out along with the first sequence model by random initialization. For context information, the modeling may be performed with a first sequence model, which may be a recurrent neural network (RNN, recurrent Neural Networks), a gated loop unit (GRU, gated Recurrent Unit), a long short term memory network (LSTM, long Short Term Memory), a transducer model, and the like. Thus, each element in the reference information has both a symbolic representation and a contextual representation, and the fusion can be performed by Gate, and finally, the fusion information can be encoded into a hidden layer sequence (e.g., H f =[h s ;h p ;h o ])。
Further, decoding each hidden layer sequence, and calculating to obtain a corresponding decoding sequence function, which specifically includes:
decoding each hidden layer state sequence based on a second sequence model to obtain decoding information;
according to the decoding information, respectively calculating the knowledge base copy mode probability of the name corresponding to the copy history input information in the knowledge base, the context copy mode probability of the copy context representation and the vocabulary generation mode probability of generating words from the vocabulary;
according to the knowledge base copy mode probability p cpkb Context copy mode probability p cpctx Vocabulary generation pattern probability p genv Calculating the predictive probability P (y) t |s t ,y t-1 ,F,C):
Wherein genev, cpkb and cpctx represent vocabulary generation mode, knowledge base replication mode and context replication mode, respectively, p. represent probabilities of three different modes, P (x) represents probabilities of generating words in various modes, F and C represent input information and context, s, respectively t Representing the current decoding status, y t Representing words generated at the current moment;
according to the predictive probability P (y) t |s t ,y t-1 F, C), decoding the words to obtain a question sequence function.
Likewise, the decoder may decode with a second sequence model, which may be a recurrent neural network (RNN, recurrent Neural Networks), a gated loop unit (GRU, gated Recurrent Unit), a long and short term memory network (LSTM, long Short Term Memory), a transducer model, and the like. In the decoding process, in order to better capture the input information, several copy forms can be selectively adopted: 1. considering that the head entity often appears in the generated question sentence, so that the name corresponding to the head entity symbol of the copy knowledge base; context of copy extension it is noted that the input context may have many repeated words, so the mechanism of using maxout pointers, i.e. when multiple identical token (words) appear, takes the token with the highest copy score as the score of the token in copy mode instead of adding. Finally, there are two different copy forms and three modes of token generation from the word list, the weighted sum of the three modes is used as the final token selection probability, then word-by-word decoding is carried out according to the prediction probability of the target word to obtain a question generation model, the question generation model is further corrected to obtain a corrected question generation model, and the question sequence information can be accurately determined based on a small number of given facts according to the corrected question generation model.
The effectiveness of the invention is verified by the following experiments:
test corpus
SimpleQuestions: the current rule models the largest knowledge base question-answer dataset.
The comparison method comprises the following steps:
template: template generation question
Serpan et al (2016): sequence-to-Sequence model generating questions
Elsahar et al (2018): generating questions by introducing a single context
The experimental results (as in Table 2)
TABLE 2
Overall performance comparison: the effectiveness of the present invention is illustrated by comparing the effects of the existing method and the present invention. The performance of the invention is obviously better than that of the standard method, and the performance is continuously improved by adding answer auxiliary supervision (the last line).
TABLE 3 Table 3
Predicate override comparison (as shown in table 3): the predicate of a given input is evaluated manually to determine whether the given predicate is correctly expressed in the generated question, and the best performance of the invention is found by calculating the ratio of correctly expressing the given predicate (predicate coverage Predicate Identification).
TABLE 4 Table 4
Answer coverage comparison: the invention further defines an evaluation index Ans of the proportion of answer type words contained in the generated question cov For evaluating the degree of certainty of generating answers corresponding to questions. The weight lambda of the answer-assisted supervision information is adjusted. When answer auxiliary supervision is found, BLEU score is higher, ans cov The lifting is more obvious.
In addition, the invention also provides a question generation system based on the given facts, which can accurately determine the problems based on a small number of given facts.
As shown in fig. 3, the question generation system based on the given facts of the present invention includes: an acquisition unit 1, an expansion unit 2, a modeling unit 3 and a determination unit 4.
Specifically, the acquiring unit 1 is configured to acquire history reference data, where the history reference data includes history input information of a plurality of different users;
the expansion unit 2 is used for expanding each history input information to obtain a corresponding context representation;
the modeling unit 3 is configured to establish a question generation model according to each piece of input information and the corresponding context representation;
the determining unit 4 is configured to determine, based on the question generation model, a question sequence corresponding to current input information of a current user according to the current input information.
Further, the invention also provides a question generation system based on given facts, which comprises the following steps:
a processor; and
a memory arranged to store computer executable instructions that, when executed, cause the processor to:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
The present invention also provides a computer-readable storage medium storing one or more programs that, when executed by an electronic device comprising a plurality of application programs, cause the electronic device to:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
Compared with the prior art, the question generation system and the computer readable storage medium based on the given facts have the same beneficial effects as the question generation method based on the given facts, and are not repeated here.
Thus far, the technical solution of the present invention has been described in connection with the preferred embodiments shown in the drawings, but it is easily understood by those skilled in the art that the scope of protection of the present invention is not limited to these specific embodiments. Equivalent modifications and substitutions for related technical features may be made by those skilled in the art without departing from the principles of the present invention, and such modifications and substitutions will fall within the scope of the present invention.

Claims (8)

1. A question generation method based on given facts, characterized in that the question generation method comprises:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users; the history reference data also comprises a plurality of pieces of supervision information, and each piece of supervision information comprises a manually marked question corresponding to the history input information and a reference answer;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
the correction method of the question generation model further comprises the following steps:
correcting the question generation model according to the supervision information to obtain a corrected question generation model:
based on the question generation model, determining a corresponding historical question sequence according to each historical input information;
according to each history question sequence and the corresponding manual labeling question, calculating and generating question loss
Calculating auxiliary answer loss according to each history question sequence and the corresponding reference answer
Wherein each reference answer includes answer type words corresponding to the history input information, the history question sequence includes generated words corresponding to the history input information,is a set of answer type words, |a| represents the number of answer type words in the set of answer type words, |a| is +.>Is a generated word y in a question sequence t And corresponding answer classWord a n Is a loss of (2);
according to the generated question lossAuxiliary answer loss->Determining supervision information loss->
Wherein λ represents a reference coefficient;
according to the supervision information lossCorrecting the question generation model to obtain a corrected question generation model;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
2. The question generation method based on given facts according to claim 1, wherein the format of the history input information of the different users is a head entity-relation-tail entity;
the expanding of each history input information to obtain a corresponding context representation specifically comprises:
for the head entity and/or the tail entity, using the type information in the knowledge base as a context representation of the head entity and/or the tail entity;
for a relationship, at least one of a sentence corresponding to a domain, a value domain, a topic, and a distance supervision back label in a knowledge base is used as a contextual representation of the relationship.
3. A question generation method based on given facts according to claim 2, characterized in that when there are a plurality of types of information of said knowledge base, the most frequently used and most differentiated type is selected as a context representation of the head entity and/or the tail entity.
4. The question generation method based on given facts according to claim 1, wherein the creating a question generation model according to each of the input information and the corresponding context representation specifically comprises:
for each pair of input information and corresponding context representation,
training the input information to obtain training information;
based on a first sequence model, obtaining a representation sequence according to the context representation;
fusing the training information and the representation sequence to obtain fusion information;
coding the fusion information to obtain a hidden layer sequence;
and decoding each hidden layer sequence, and calculating to obtain a corresponding decoding sequence function, wherein the decoding sequence function is a question generation model.
5. The question generation method based on given facts according to claim 4, wherein said decoding each hidden layer sequence, calculating a corresponding decoding sequence function, specifically includes:
decoding each hidden layer state sequence based on a second sequence model to obtain decoding information;
according to the decoding information, respectively calculating the knowledge base copy mode probability of the names corresponding to the copy history input information in the knowledge base, the context copy mode probability of the copy context representation and the vocabulary generation mode probability of the words generated from the vocabulary;
according to the knowledge base copy mode probability p cpkb Context copy modeProbability p cpctx Vocabulary generation pattern probability p genυ Calculating the predictive probability P (y) t |s t ,y t-1 ,F,C):
P(y t |s t ,y t-1 ,F,C)=p genυ P genυ (y t |s t ,V)+p cpkb P cpkb (y t |s t ,F)+p cpctx P cpctx (y t |s t ,C);
Wherein genev, cpkb and cpctx represent vocabulary generation mode, knowledge base replication mode and context replication mode, respectively, p. represent probabilities of three different modes, P (x) represents probabilities of generating words in various modes, F and C represent input information and context, s, respectively t Representing the current decoding status, y t Representing words generated at the current moment;
according to the predictive probability P (y) t |s t ,y t-1 F, C), decoding word by word to obtain a question sequence function.
6. A question generation system based on a given fact, the question generation system comprising:
the acquisition unit is used for acquiring history reference data, wherein the history reference data comprises a plurality of pieces of history input information of different users; the history reference data also comprises a plurality of pieces of supervision information, and each piece of supervision information comprises a manually marked question corresponding to the history input information and a reference answer;
the expansion unit is used for expanding each history input information to obtain a corresponding context representation;
the modeling unit is used for establishing a question generation model according to the input information and the corresponding context representation;
the correction method of the question generation model further comprises the following steps:
correcting the question generation model according to the supervision information to obtain a corrected question generation model:
based on the question generation model, determining a corresponding historical question sequence according to each historical input information;
according to each history question sequence and the corresponding manual labeling question, calculating and generating question loss
Calculating auxiliary answer loss according to each history question sequence and the corresponding reference answer
Wherein each reference answer includes answer type words corresponding to the history input information, the history question sequence includes generated words corresponding to the history input information,is a set of answer type words, |a| represents the number of answer type words in the set of answer type words, |a| is +.>Is a generated word y in a question sequence t And the corresponding answer type word a n Is a loss of (2);
according to the generated question lossAuxiliary answer loss->Determining supervision information loss->
Wherein λ represents a reference coefficient;
according to the supervision information lossCorrecting the question generation model to obtain a corrected question generation model;
and the determining unit is used for determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
7. A question generation system based on given facts, comprising:
a processor; and
a memory arranged to store computer executable instructions that, when executed, cause the processor to:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users; the history reference data also comprises a plurality of pieces of supervision information, and each piece of supervision information comprises a manually marked question corresponding to the history input information and a reference answer;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
the correction method of the question generation model further comprises the following steps:
correcting the question generation model according to the supervision information to obtain a corrected question generation model:
based on the question generation model, determining a corresponding historical question sequence according to each historical input information;
according to each history question sequence and the corresponding manual labeling question, calculating and generating a questionSentence loss
Calculating auxiliary answer loss according to each history question sequence and the corresponding reference answer
Wherein each reference answer includes answer type words corresponding to the history input information, the history question sequence includes generated words corresponding to the history input information,is a set of answer type words, |a| represents the number of answer type words in the set of answer type words, |a| is +.>Is a generated word y in a question sequence t And the corresponding answer type word a n Is a loss of (2);
according to the generated question lossAuxiliary answer loss->Determining supervision information loss->
Wherein λ represents a reference coefficient;
according to the supervision information lossCorrecting the question generation model to obtain a corrected question generation model;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
8. A computer-readable storage medium storing one or more programs that, when executed by an electronic device comprising a plurality of application programs, cause the electronic device to:
acquiring historical reference data, wherein the historical reference data comprises historical input information of a plurality of different users; the history reference data also comprises a plurality of pieces of supervision information, and each piece of supervision information comprises a manually marked question corresponding to the history input information and a reference answer;
expanding each history input information to obtain a corresponding context representation;
establishing a question generation model according to each piece of input information and the corresponding context representation;
the correction method of the question generation model further comprises the following steps:
correcting the question generation model according to the supervision information to obtain a corrected question generation model:
based on the question generation model, determining a corresponding historical question sequence according to each historical input information;
according to each history question sequence and the corresponding manual labeling question, calculating and generating question loss
According to each history question sequence and correspondingTo calculate the auxiliary answer loss
Wherein each reference answer includes answer type words corresponding to the history input information, the history question sequence includes generated words corresponding to the history input information,is a set of answer type words, |a| represents the number of answer type words in the set of answer type words, |a| is +.>Is a generated word y in a question sequence t And the corresponding answer type word a n Is a loss of (2);
according to the generated question lossAuxiliary answer loss->Determining supervision information loss->
Wherein λ represents a reference coefficient;
according to the supervision information lossCorrecting the question generation model to obtain a corrected question generation model;
and determining a question sequence corresponding to the current input information according to the current input information of the current user based on the question generation model.
CN201911276552.4A 2019-12-12 2019-12-12 Question generation method and system based on given facts Active CN111061851B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911276552.4A CN111061851B (en) 2019-12-12 2019-12-12 Question generation method and system based on given facts

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911276552.4A CN111061851B (en) 2019-12-12 2019-12-12 Question generation method and system based on given facts

Publications (2)

Publication Number Publication Date
CN111061851A CN111061851A (en) 2020-04-24
CN111061851B true CN111061851B (en) 2023-08-08

Family

ID=70300718

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911276552.4A Active CN111061851B (en) 2019-12-12 2019-12-12 Question generation method and system based on given facts

Country Status (1)

Country Link
CN (1) CN111061851B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111813907A (en) * 2020-06-18 2020-10-23 浙江工业大学 Question and sentence intention identification method in natural language question-answering technology

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102737042A (en) * 2011-04-08 2012-10-17 北京百度网讯科技有限公司 Method and device for establishing question generation model, and question generation method and device
GB201419051D0 (en) * 2014-10-27 2014-12-10 Ibm Automatic question generation from natural text
CN105095444A (en) * 2015-07-24 2015-11-25 百度在线网络技术(北京)有限公司 Information acquisition method and device
CN106485370A (en) * 2016-11-03 2017-03-08 上海智臻智能网络科技股份有限公司 A kind of method and apparatus of information prediction
CN106599215A (en) * 2016-12-16 2017-04-26 广州索答信息科技有限公司 Question generation method and question generation system based on deep learning
CN107329967A (en) * 2017-05-12 2017-11-07 北京邮电大学 Question answering system and method based on deep learning
CN107463699A (en) * 2017-08-15 2017-12-12 济南浪潮高新科技投资发展有限公司 A kind of method for realizing question and answer robot based on seq2seq models
CN108363743A (en) * 2018-01-24 2018-08-03 清华大学深圳研究生院 A kind of intelligence questions generation method, device and computer readable storage medium
CN109522393A (en) * 2018-10-11 2019-03-26 平安科技(深圳)有限公司 Intelligent answer method, apparatus, computer equipment and storage medium
KR20190056184A (en) * 2017-11-16 2019-05-24 주식회사 마인즈랩 System for generating question-answer data for maching learning based on maching reading comprehension
CN109992657A (en) * 2019-04-03 2019-07-09 浙江大学 A kind of interactive problem generation method based on reinforcing Dynamic Inference
CN110162613A (en) * 2019-05-27 2019-08-23 腾讯科技(深圳)有限公司 A kind of problem generation method, device, equipment and storage medium
CN110188182A (en) * 2019-05-31 2019-08-30 中国科学院深圳先进技术研究院 Model training method, dialogue generation method, device, equipment and medium
CN110196896A (en) * 2019-05-23 2019-09-03 华侨大学 A kind of intelligence questions generation method towards the study of external Chinese characters spoken language
CN110209790A (en) * 2019-06-06 2019-09-06 阿里巴巴集团控股有限公司 Question and answer matching process and device
CN110543553A (en) * 2019-07-31 2019-12-06 平安科技(深圳)有限公司 question generation method and device, computer equipment and storage medium

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10572595B2 (en) * 2017-04-13 2020-02-25 Baidu Usa Llc Global normalized reader systems and methods
US10902738B2 (en) * 2017-08-03 2021-01-26 Microsoft Technology Licensing, Llc Neural models for key phrase detection and question generation
US11250038B2 (en) * 2018-01-21 2022-02-15 Microsoft Technology Licensing, Llc. Question and answer pair generation using machine learning

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102737042A (en) * 2011-04-08 2012-10-17 北京百度网讯科技有限公司 Method and device for establishing question generation model, and question generation method and device
GB201419051D0 (en) * 2014-10-27 2014-12-10 Ibm Automatic question generation from natural text
CN105095444A (en) * 2015-07-24 2015-11-25 百度在线网络技术(北京)有限公司 Information acquisition method and device
CN106485370A (en) * 2016-11-03 2017-03-08 上海智臻智能网络科技股份有限公司 A kind of method and apparatus of information prediction
CN106599215A (en) * 2016-12-16 2017-04-26 广州索答信息科技有限公司 Question generation method and question generation system based on deep learning
CN107329967A (en) * 2017-05-12 2017-11-07 北京邮电大学 Question answering system and method based on deep learning
CN107463699A (en) * 2017-08-15 2017-12-12 济南浪潮高新科技投资发展有限公司 A kind of method for realizing question and answer robot based on seq2seq models
KR20190056184A (en) * 2017-11-16 2019-05-24 주식회사 마인즈랩 System for generating question-answer data for maching learning based on maching reading comprehension
CN108363743A (en) * 2018-01-24 2018-08-03 清华大学深圳研究生院 A kind of intelligence questions generation method, device and computer readable storage medium
CN109522393A (en) * 2018-10-11 2019-03-26 平安科技(深圳)有限公司 Intelligent answer method, apparatus, computer equipment and storage medium
CN109992657A (en) * 2019-04-03 2019-07-09 浙江大学 A kind of interactive problem generation method based on reinforcing Dynamic Inference
CN110196896A (en) * 2019-05-23 2019-09-03 华侨大学 A kind of intelligence questions generation method towards the study of external Chinese characters spoken language
CN110162613A (en) * 2019-05-27 2019-08-23 腾讯科技(深圳)有限公司 A kind of problem generation method, device, equipment and storage medium
CN110188182A (en) * 2019-05-31 2019-08-30 中国科学院深圳先进技术研究院 Model training method, dialogue generation method, device, equipment and medium
CN110209790A (en) * 2019-06-06 2019-09-06 阿里巴巴集团控股有限公司 Question and answer matching process and device
CN110543553A (en) * 2019-07-31 2019-12-06 平安科技(深圳)有限公司 question generation method and device, computer equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Yankai Lin等.Neural Relation Extraction with Selective Attention over Instances.Association for Computational Linguistics.2016,全文. *

Also Published As

Publication number Publication date
CN111061851A (en) 2020-04-24

Similar Documents

Publication Publication Date Title
CN109359294B (en) Ancient Chinese translation method based on neural machine translation
US20180329884A1 (en) Neural contextual conversation learning
CN110390397B (en) Text inclusion recognition method and device
CN109214006B (en) Natural language reasoning method for image enhanced hierarchical semantic representation
CN107862087A (en) Sentiment analysis method, apparatus and storage medium based on big data and deep learning
CN110688489B (en) Knowledge graph deduction method and device based on interactive attention and storage medium
CN111401084A (en) Method and device for machine translation and computer readable storage medium
CN108845990A (en) Answer selection method, device and electronic equipment based on two-way attention mechanism
CN112417092B (en) Intelligent text automatic generation system based on deep learning and implementation method thereof
Ren The use of machine translation algorithm based on residual and LSTM neural network in translation teaching
CN111563146B (en) Difficulty controllable problem generation method based on reasoning
CN115618045B (en) Visual question answering method, device and storage medium
US20220261546A1 (en) Method and apparatus for selecting answers to idiom fill-in-the-blank questions, and computer device
Guo et al. Learning to query, reason, and answer questions on ambiguous texts
CN109933792A (en) Viewpoint type problem based on multi-layer biaxially oriented LSTM and verifying model reads understanding method
CN111538838B (en) Problem generating method based on article
CN115759042A (en) Sentence-level problem generation method based on syntax perception prompt learning
CN115034208A (en) Chinese ASR output text repair method and system based on BERT
CN115345169A (en) Knowledge enhancement-based text generation model and training method thereof
CN111061851B (en) Question generation method and system based on given facts
CN113486174B (en) Model training, reading understanding method and device, electronic equipment and storage medium
CN110489761B (en) Chapter-level text translation method and device
CN112052663B (en) Customer service statement quality inspection method and related equipment
CN113705207A (en) Grammar error recognition method and device
CN116362242A (en) Small sample slot value extraction method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant