CN112100351A - Method and equipment for constructing intelligent question-answering system through question generation data set - Google Patents
Method and equipment for constructing intelligent question-answering system through question generation data set Download PDFInfo
- Publication number
- CN112100351A CN112100351A CN202010956043.2A CN202010956043A CN112100351A CN 112100351 A CN112100351 A CN 112100351A CN 202010956043 A CN202010956043 A CN 202010956043A CN 112100351 A CN112100351 A CN 112100351A
- Authority
- CN
- China
- Prior art keywords
- question
- entity
- knowledge graph
- entities
- knowledge
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 54
- 239000013598 vector Substances 0.000 claims abstract description 36
- 238000012549 training Methods 0.000 claims abstract description 24
- 230000011218 segmentation Effects 0.000 claims abstract description 19
- 238000000605 extraction Methods 0.000 claims abstract description 16
- 238000004364 calculation method Methods 0.000 claims abstract description 12
- 230000008569 process Effects 0.000 claims abstract description 11
- 238000004590 computer program Methods 0.000 claims description 13
- 238000004422 calculation algorithm Methods 0.000 claims description 12
- 230000002457 bidirectional effect Effects 0.000 claims description 10
- 238000010276 construction Methods 0.000 claims description 7
- 239000000284 extract Substances 0.000 claims description 4
- PCHJSUWPFVWCPO-UHFFFAOYSA-N gold Chemical compound [Au] PCHJSUWPFVWCPO-UHFFFAOYSA-N 0.000 claims description 2
- 239000010931 gold Substances 0.000 claims description 2
- 229910052737 gold Inorganic materials 0.000 claims description 2
- 238000003058 natural language processing Methods 0.000 description 8
- 239000011159 matrix material Substances 0.000 description 7
- 230000000694 effects Effects 0.000 description 6
- 230000006870 function Effects 0.000 description 6
- 238000013528 artificial neural network Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 238000013527 convolutional neural network Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 238000006243 chemical reaction Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 241000282414 Homo sapiens Species 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000000306 recurrent effect Effects 0.000 description 2
- 239000013589 supplement Substances 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 238000000429 assembly Methods 0.000 description 1
- 230000000712 assembly Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013145 classification model Methods 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/332—Query formulation
- G06F16/3329—Natural language query formulation or dialogue systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/211—Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/289—Phrasal analysis, e.g. finite state techniques or chunking
- G06F40/295—Named entity recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/049—Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Human Computer Interaction (AREA)
- Databases & Information Systems (AREA)
- Machine Translation (AREA)
Abstract
A method and apparatus for constructing an intelligent question-answering system through a question generation data set, by constructing a travel domain knowledge map; analyzing a question of a natural language proposed by a user, performing word segmentation and word vector training on the question, and adding a preset tourism field dictionary by using a jieba tool in the word segmentation process; performing entity extraction on the natural language question by using a Bert-BilSTM + CRF model; matching the extracted entity with an entity in a knowledge graph; if the knowledge graph has a matched entity, selecting the entity; if the knowledge graph has no matched entities, performing semantic similarity calculation, and selecting the closest entities; matching the selected entities and attributes with triples in a knowledge graph; and returning the corresponding attribute value as an answer of the question to the user. The invention also provides a device, a terminal and a readable storage medium for realizing the method. The invention can conveniently and accurately return the information required by the user.
Description
Technical Field
The invention belongs to the field of information query and retrieval, and particularly relates to a method and equipment for constructing an intelligent question-answering system through a question generation data set, which are particularly suitable for information search of a knowledge map in the field of tourism.
Background
Since the concept of knowledge graph was proposed in 2012, all internet companies construct next-generation intelligent search engines based on the concept, such as google knowledge graph, dog knowledge cube, etc., and hope to create a brand-new information retrieval mode and provide a new idea for solving the information retrieval problem. Meanwhile, the domain knowledge graph is developed rapidly in recent years, and compared with the traditional knowledge graph, the domain knowledge graph has the advantages that data are more concentrated, processing is more convenient, ambiguity between entities is greatly reduced, and searching accuracy is improved. Meanwhile, as knowledge is gradually converted from static state to dynamic state, the information updating speed of the domain knowledge graph is superior to that of the general knowledge graph.
Due to the reasons of high construction difficulty, time consumption, labor consumption and the like of the existing data set, most of natural language processing tasks use universal data sets, such as NLPCC QA data sets, WikiQA data sets, TREC QA data sets and the like. However, with the proposition and development of domain knowledge maps, more and more methods aiming at a certain domain are proposed, and the methods commonly used at present are improved on the basis of manually collected data, including a heuristic marking method and a machine-assisted marking method, and some adopt heuristic marking based on a general disambiguation rule. For example, when marking an entity, all adjectives are deleted, just "entity header" is marked (e.g., "a girl in red booth" is changed to "girl"). It can be seen that the effort to tag large-scale datasets across a knowledge base is very large. In the KBQA task, because relationship detection is more difficult than entity linking, the QA model often fails to answer questions because of invisible predicates or phrases. On the one hand QA models generally tend to give invisible predicates a lower score. On the other hand, even if the training set contains true predicates, the QA model is difficult to answer if the corresponding paraphrases are not visible. At present, the question-answering system based on the knowledge graph also has different classification modes in structure. In the early days of template-based knowledge-graph question-answering, the method needs to prepare an intention template in advance, can match through similarity and can also solve through a classification problem in machine learning. Although this method is fast and can answer complex questions, it requires a huge template library and thus is very labor intensive. The main current research direction is that templates can be automatically updated to reduce the workload of human beings. The knowledge-graph question-answer based on semantic analysis is converted into semantic representation which can be understood by a knowledge base through semantic analysis of natural language, and then reasoning or inquiry is carried out through knowledge in the knowledge base to obtain a final answer. In short, what is needed for semantic parsing is to convert the natural language question into a semantic representation that can be understood by the knowledge base, and the semantic representation is a logical form. In addition, the existing popular intelligent question-answering system based on deep learning also appears, and some people construct a problem analysis model by using a neural network based on word vectors and long-short term memory, and for example, introduce an LSTM model and a semantic enhancement method to improve a domain feature knowledge graph to obtain good effects.
A large number of experiments prove that better results can be obtained by adopting a deep learning algorithm under the condition of large data volume compared with the traditional natural language algorithm. And the deep learning algorithm does not need to manually extract features, so the implementation is relatively simple and convenient. The CNN algorithm is the earliest algorithm used, mainly because the convolution model of the CNN algorithm is simple to implement and can capture characteristic information of data positions. But for time-series data, the LSTM algorithm is more suitable than the CNN algorithm. The LSTM algorithm is a special type of RNN, comprehensively considers the characteristics of the problem time sequence, calculates the state characteristics of the data through 3 gate functions, and has better effect than the CNN model. The LSTM model can be generally matched with other models to be used together, so that the experimental effect is better, the bidirectional LSTM model and the CRF model can be combined to establish a knowledge base question-answering system, the advantages of a conditional random field and a long-short term memory network are tried to be comprehensively utilized, and the performance of the question-answering system can also be improved. With the release of the Bert model of google, researchers began to apply the Bert model to various Natural Language Processing (NLP) questions, including domain knowledge graph-based question-answering systems. The Bert model can be simply understood as a two-segment NLP model, i.e. a pre-training and fine-tuning step. The Bert model is represented by a transform's bi-directional encoder, which, unlike most recent other language representation models, aims to pre-train the deep bi-directional representation by jointly adjusting the context in all layers. Therefore, the pre-trained Bert representation can be fine-tuned through an additional output layer, and is suitable for building the most advanced models of a wide range of tasks, such as question-answering tasks and language reasoning.
Disclosure of Invention
The invention aims to solve the problems that the data set construction in the prior art is time-consuming and labor-consuming and the tourism question-answering system is incomplete, provides a method and equipment for constructing an intelligent question-answering system by using a question generation data set, supplements the question-answering data set by using a question generation model, makes up the condition of tourism data loss, and can conveniently and accurately return information required by a user.
In order to achieve the purpose, the invention has the following technical scheme:
a method of constructing an intelligent question-answering system from question-generating datasets, comprising the steps of:
step one, establishing a tourist field knowledge map;
storing entity related information in a knowledge graph in a triple form of 'entity-attribute value', when a problem is raised, firstly, extracting entities in a question through named entity identification, linking the entities with the knowledge graph to find corresponding entities, then, utilizing question information and a preset rule template to perform entity attribute matching, and returning corresponding attribute values;
step two, performing question analysis on a natural language question provided by a user, performing word segmentation and word vector training on the question, and adding a preset tourism field dictionary by using a jieba tool in the word segmentation process;
step three, utilizing a Bert-BilSTM + CRF model to perform entity extraction on the natural language question;
step four, matching the extracted entity with an entity in the knowledge graph; if the knowledge graph has a matched entity, selecting the entity; if the knowledge graph has no matched entities, performing semantic similarity calculation, and selecting the closest entities;
matching the selected entities and attributes with triples in the knowledge graph;
and step six, returning the corresponding attribute value as an answer of the question to the user.
Preferably, the tourist domain knowledge map constructed in the step one uses question generation to construct a question-and-answer data set, the task of question generation is to generate corresponding natural language questions given the input answers, and the question-and-answer is the reverse task of finding suitable answers in the given questions.
Preferably, in the third step, firstly, a Bert pre-training language model is used for coding a single character to obtain a word vector corresponding to the single character, then, a BilSTM layer is used for carrying out bidirectional coding on an input text, and finally, a semantic vector containing context information is input into a CRF layer for decoding, and the CRF layer can output a label sequence with the maximum probability so as to obtain the category of each character.
Preferably, in the fourth step, a cosine similarity algorithm is used for semantic similarity calculation, and the cosine similarity calculation is expressed as follows:
wherein X represents an entity extracted from the question, Y represents an entity in the knowledge graph, and n represents the dimension of a word vector; and selecting the entity in the knowledge graph with the highest similarity as matching by calculating the similarity.
Preferably, the QG and QA models are combined, and by means of a knowledge base and a text corpus, the QA and QG models are firstly trained on the gold data set in a combined mode, and then the QA model is finely adjusted by means of text by means of a supplementary data set constructed by the QG model.
Preferably, the entity related information in the first step is crawled by a Python tool, the entity related information comprises tourist attractions, names, introduction, construction time and scenic spot levels, and the crawled data is cleaned and sorted to form a tourist knowledge map.
The invention also provides a device for constructing the intelligent question-answering system by the question generation data set, which comprises the following components:
the tourism domain knowledge map building module stores entity related information in a triple form of entity-attribute value, extracts entities in question sentences through named entity identification when a problem is raised, links the entities with a knowledge map to find corresponding entities, matches entity attributes by using question sentence information and combining a rule template set in advance, and returns corresponding attribute values;
the natural language question analyzing module is used for performing word segmentation and word vector training on the natural language question, and a jieba tool is used and a preset tourism field dictionary is added in the word segmentation process;
the entity extraction module is used for firstly utilizing a Bert pre-training language model to encode a single character to obtain a word vector corresponding to the single character, then utilizing a BilSTM layer to carry out bidirectional encoding on an input text, and finally inputting a semantic vector containing context information into a CRF layer to be decoded, wherein the CRF layer outputs a label sequence with the maximum probability to obtain the category of each character;
the entity matching module is used for matching the extracted entity with the entity in the knowledge graph; if the knowledge graph has a matched entity, selecting the entity; if the knowledge graph does not have the matched entity, selecting the closest entity through semantic similarity;
and the answer feedback module is used for matching the selected entity and attribute with the triples in the knowledge graph, returning corresponding attribute values, and providing the attributes serving as answers of the question to the user.
The invention also provides a terminal, which comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, wherein the processor realizes the steps of the method for constructing the intelligent question-answering system through the question generation data set when executing the computer program.
The present invention also provides a computer readable storage medium storing a computer program which, when executed by a processor, performs the steps of the method for constructing an intelligent question-and-answer system from question generated data sets.
Compared with the prior art, the invention has the following beneficial effects:
the invention searches the information based on the knowledge graph in the tourism field, can greatly reduce ambiguity of some nouns, concentrates on the inquiry in the aspect of tourism and improves the accuracy and efficiency. According to the intelligent question-answering system, the question-answering data set is supplemented by the question generation model based on the intelligent question-answering strategy constructed by the question generation, the condition of travel data loss is made up, and the intelligent question-answering system constructed by the Bert model can return information required by the user more conveniently and accurately. The invention combines the advantages of the Bert model, and constructs the Bert-BilSTM-CRF model on the basis of the Bert model, aiming at better identifying the entity. Meanwhile, the Bert model is applied to the relation extraction, and the purpose of the relation extraction is to identify the relation in the problem and obtain the attribute in the triple. The accuracy rate of text similarity calculation by using the Bert model is superior to that of word2vec models and the like, and the effect is obvious.
Drawings
FIG. 1 is a flow chart of the operation of an embodiment of the present invention to construct a question-answering system.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples.
A method of constructing an intelligent question-answering system from question-generating datasets, comprising the steps of:
step one, establishing a tourist field knowledge map;
storing entity related information in a knowledge graph in a triple form of 'entity-attribute value', when a problem is raised, firstly, extracting entities in a question through named entity identification, linking the entities with the knowledge graph to find corresponding entities, then, utilizing question information and a preset rule template to perform entity attribute matching, and returning corresponding attribute values;
using a Python tool to crawl relevant information of required tourism, wherein the relevant information comprises tourist attractions, names, introduction, construction time, scenic spot levels and the like; and cleaning and sorting the crawled data to form a tourism knowledge map.
The whole construction framework of the knowledge graph consists of a logic framework and a technical framework adopted by construction, and the logic framework can be divided into 2 levels of a data layer and a mode layer. Wherein the basic unit of knowledge of the data layer is "fact" (namely a triple of "head entity-relation-tail entity" or "entity-attribute value"); and the mode layer above the data layer manages the knowledge graph in a mode of ontology library, so that the knowledge graph is stored by the mode layer and is highly condensed. The data sources in the knowledge acquisition stage mainly include three types of structured data (such as link data and a relational database), semi-structured data (such as encyclopedia and vertical domain data) and unstructured data (namely text data). The main work of the knowledge integration phase is: due to multiple sources of knowledge acquisition channels, knowledge acquired through knowledge acquisition usually contains a large amount of redundant abnormal information, and therefore the result needs to be further processed and integrated, including operations such as entity linking, knowledge merging, entity alignment and the like.
Step two, performing question analysis on a natural language question provided by a user, performing word segmentation and word vector training on the question, and adding a preset tourism field dictionary by using a jieba tool in the word segmentation process;
step three, utilizing a Bert-BilSTM + CRF model to perform entity extraction on the natural language question;
step four, matching the extracted entity with an entity in the knowledge graph; if the knowledge graph has a matched entity, selecting the entity; if the knowledge graph has no matched entities, performing semantic similarity calculation, and selecting the closest entities;
matching the selected entities and attributes with triples in the knowledge graph;
and step six, returning the corresponding attribute value as an answer of the question to the user.
Specifically, in the first step, a question and answer data set is generated and constructed by using questions, and entity related information is stored in a knowledge graph; identifying entities and relations by natural language processing of questions, and finding out corresponding attribute values in the question and answer data set;
question generation can be viewed as the reverse task of QA, generating a corresponding question given a question, in different QA/QG tasks the answer a can be different, e.g. a sentence in a document or a fact in a knowledge base.
In this task, the KBQA task of the present invention is a triple, so that different methods need to be designed to calculate the corresponding term in the probability formula. To address the challenges of unseen predicates and phrases, the current approach is to train a sequence-to-sequence model that can generate unseen predicates based on triples extracted from the knowledge base. In addition, the QA model can be fine-tuned by inputting the generated questions and the triples extracted from the KB.
The answer in KBQA is a fact from the knowledge base, while the answer in text QA is a sentence in a given document. In this work, the focus should be on KBQA, which is considered a scoring and ranking problem, using a scalar to estimate the correlation between the problem and the triples. The QA task is simplified here to a relationship detection task that takes as input the question and candidate relationships and outputs a relationship that is the maximum probability of the correct relationship, assuming that the subject entity has been detected, so that the answer fact can be queried in the knowledge base once the correct relationship is determined. The task of the QG is to take a sentence or a triple as input and output a question that can be answered by the triple.
Typically such a frame needs to be composed of two part assemblies. The first is a double learning component. It uses the probabilistic correlation between QA and QG to direct the parameters of the QA/QG model to a more appropriate direction. The second is a fine tuning component that improves the ability of the QA model to handle invisible predicates and phrases by combining the QG model with the text corpus and knowledge base triples. The framework is very flexible and does not depend on a particular QA or QG model.
The training of the KBQA system relies on a high quality annotation data set that is not only large-scale, but also unbiased. It is very difficult to build such a data set that covers a large number of triples in the knowledge base.
Therefore, a fine-tuning framework is proposed to supplement the QA dataset, increasing the capacity of the QA model.
QA model
Generally, a Question Answering (QA) model can be simplified into a relationship classification model, and compared with other subtasks such as entity links in KBQA, relationship extraction has a more significant influence on a final result. The entity link accuracy in the existing KBQA method is high, and the performance of relation extraction is not ideal due to invisibility of predicates or paraphrases.
The invention provides a relation extraction model based on a recurrent neural network. To better support invisible relationships, relationship names are decomposed into word sequences, and relationship extraction is described as a sequence matching and ordering task. For example, the relationship location, country, language may be divided into { location, country, language }. Each of the above tokens is converted into a pre-trained word embedding, and then a two-way long-short term memory model (BilSTM) is used to obtain a hidden representation. Finally, the local features are extracted by utilizing the maximum local feature set. The invention uses the same neural network to obtain problem representation, and then uses cosine distance function to calculate similarity. To learn a more general representation at the syntactic level, the present invention replaces the entity with a generic symbol < e >, such as "where is < e > from". However, this mechanism can lose all entity information and in some cases can confuse the model. Thus, the present invention detects the type of entity and concatenates the type representation with the problem representation, which can significantly improve performance.
The model described above is trained using a scoring ranking training method, and the loss function is as follows:
lrel=max{0,λ-Ss(q,r+)+Ss(q,r-)}
wherein, question and correct predicate (q, r)+) Must be at least the margin λ, Ss(q,r)=Cosine(hq,hr). The candidate relationship R consists of all predicates in q that are relevant to the subject entity e.
QG model
The QG model is a task that generates natural language, inputting triples in a knowledge base. Generating a question involves the subject-to-predicate relationship of a fact whose object represents a valid answer to the generated question. Assuming that the input fact a is { s, p, o }, the QG model approximation generates a problem q is { w }1,w2,…,wnThe conditional probabilities are as follows:
wherein w<tRepresenting all previous generated words until the time step reaches t. Under the successful initiation of the sequence learning in the recent machine translation, the QG problem is regarded as a translation task and is solved by adopting the structure of the encoder and decoder. Specifically, the encoder encodes a given fact a ═ { s, p, o } into three fixed-size vectors hs=Efes,hp=EfepAnd ho=EfeoIn which EfIs a KB embedding matrix using TransE learning, es,ep,eoIs a hotspot vector for s, p, o. Concatenating these three vectors to obtain the coded fact hf=[hs;hp;ho]. Then the decoder takes hfThe questions are generated in order.
Complementary question-answer pairs are generated in a fine-tuning component using the QG model to fine-tune the trained QA model. The tag data that is desired to be supplemented contains predicates and phrases that the QA model does not encounter during the training process. Thereby improving the performance of the QA model.Therefore, the QG model should be able to generate problems given triples and invisible predicates. For each fact a, n pieces of textual evidence D ═ D are collected from the wiki document1,d2…,dn-encoding each text evidence using a set of n-gated recurrent neural networks (GRUs) with shared parameters. The hidden state of the ith word in the j text evidences is calculated as:
Connecting each hidden state of the text evidence to obtain the final coded textFor the decoder, a GRU is used which has an attention mechanism acting on the evidence of the input text. Given a set of encoded input vectors I ═ h1,h2,…,hkH hidden state before decodert-1Attention mechanism calculation of at={ai,t,…,ak,tAs a weight vector, each ai,tTo determine its corresponding encoded input vector hiThe weight of (c).
ei,t=va Ttanh(Wast-1+Uahi)
Wherein v isa,Wa,UaIs a trainable weight matrix of the attention module.
Then, for all the marks in all the text evidences:
whereinIs a scalar value that determines the jth textual evidence d at time tiThe weight of the ith word.
Preferably, the process of identifying entities and relationships by natural language processing of questions and finding corresponding attribute values in the question-answer dataset comprises: the knowledge graph-based question answering method can be roughly divided into two steps, namely a named entity identification step and an attribute mapping step, most of knowledge graphs in most of the prior art adopt a template-based method due to data quantity, and a Bert model is used in the invention in order to improve the question answering efficiency and accuracy. The Bert model is an NLP model proposed by google in 2018, i.e., Encoder of bidirectional Transformer, because decoder cannot obtain information to be predicted. The main innovation points of the model are all on a pre-train method, namely two methods, namely Masked LM and Next sequence Prediction, are used for capturing the representation at the word level and the Sentence level respectively. The invention divides the whole intelligent question-answering process into two parts. The method uses a Bert-BilSTM-CRF model, and the accuracy of the model in named entity recognition is greatly improved compared with the prior model. And then, judging the attributes of the problems, wherein the attribute judgment and conversion is to solve the problem of text similarity, and a Bert model is still used for performing binary classification. After the Bert model is used, the effect is remarkable.
Further, the main problems to be solved by the question-answering system are named entity identification and relationship extraction. The object of the entity identification step is to find the name of the entity queried in the question, and the object of the relationship extraction step is to find the relevant attribute queried in the question. The current methods for obtaining pre-trained representation models are mainly feature-based methods or fine-tuning methods. A typical feature-based approach is the ELMo model. And based on a fine tuning method, the openAI GPT model only considers the information of the previous word, the vector of the previous word is input into a Transformer model, and L layers are stacked to obtain pre-training representation characteristics. GPT has a strong feature extraction capability, and the effect of ELMo can be achieved without the need for a special design structure for the processing task. For the problem of sentence level, it is necessary to combine the context to accurately grasp its meaning, that is, it is not possible to take into account only the information of the previous word, and it is also necessary to take into account the meaning of the following word. The Bert model is then derived.
The Bert model is a large-scale pre-training language model based on a bidirectional transducer issued by Google, shows striking performance in machine reading understanding top level test SQuAD1.1, comprehensively surpasses human on all two measurement indexes, creates optimal performance in a plurality of NLP tests, and is a great trend when the Bert model is applied to various models.
The Bert model has the following two characteristics: first, this model is very deep, having 12 layers and not wide, with only 1024 intermediate layers, whereas the previous Transformer model intermediate layer has 2048. This also corroborates the view of computer image processing that deep, narrow models are superior to shallow, wide models. Second, the Masked Language Model uses words on both the left and right sides, and the Model considering this point only combines the left-to-right and right-to-left training separately. The author randomized 15% tokens during training, rather than predicting each word over as cbow does. The final loss function only computes which token was dropped by the mask. Structurally, compared with the GPT model, the information of backward words is added, so that the model fully considers the information of the context. In contrast, for the ELMo model, although two bidirectional RNN models are trained, one RNN can only see one direction, and therefore, information in two directions, front and rear, cannot be simultaneously utilized. To solve the problem of being able to use only one-way information, Bert uses Mask language model instead of the ordinary language model. The Mask language model simply gives a sentence, blocks a certain word in the sentence, and predicts the blocked word. Here, 15% of words are randomly masked, and then Bert is used to predict the words of Mask, and the probability of correct model prediction is made as large as possible by adjusting the parameters of the model, which is equivalent to the loss function of cross entropy. Such a Transformer will refer to contextual information when encoding a word. If a token is in the selected 15% tokens, then it is randomly executed as follows: replacement to MASK with 80% probability; a 10% probability of replacement with a random word; 10% probability replaced itself; this has the advantage that Bert does not know which word the MASK replaces, and any word may be replaced, thus forcing the model to be too dependent on the current word when encoding the current time, taking into account its context.
BilSTM-CRF
The present invention utilizes a bi-directional LSTM network that enables efficient use of past and future features within a specified time frame. And (3) taking the word embedding vector of each word in the sentence obtained by the Bert model as the input of each time step of the bidirectional LSTM, and outputting a prediction label corresponding to each word. If the input sentence is composed of 120 words, each word is represented by a 100-dimensional word vector, the input corresponding to the model is (120,100), the hidden vector changes to T1(120,128) after BiLSTM, where 128 is the output dimension of BiLSTM in the model. If the CRF layer is not used, a fully connected layer can be added to the model for classification. And setting target labels of the word segmentation task as B (begin), M (middle), E (end) and S (Single), and finally outputting the vector with the latitude (120,4) by the model. And respectively representing the probability of corresponding BMES for 4 floating point values corresponding to each word, and finally taking the label with high probability as the predicted label. Through a large amount of labeled data and continuous iterative optimization of the model, the method can learn a good word segmentation model. Although relying on the powerful nonlinear fitting capability of neural networks, good models can be theoretically learned. The above-described BiLSTM model only takes into account context information on the tags. For the sequence labeling task, the label L _ t of the current position has a potential relation with the previous position L _ t-1 and the next position L _ t + 1. Therefore, most of the current cases follow the BilTM model with a CRF layer for the entire BiLSTM modelThe optimal tag sequence is learned on the sequence. For an input sentence X ═ X1,x2…,xn) Let P be the fractional matrix of the output of the BilSTM network, the size of P being n x k, where k is the number of different labels, Pi,jThe score of the jth token corresponding to the ith word in the sentence. Followed by the predicted sequence y ═ y (y)1,y2…,yn) Then the score is defined as:
where A is a matrix of conversion scores, thus Ai,jRepresenting the conversion score, y, from token i to token j0And ynAre the beginning and ending tokens of a sentence, which the present invention adds to the set of possible tokens. Thus a is a square matrix of size k + 2.
It can be seen that the score for the entire sequence is equal to the sum of the scores for the individual positions, each of which is obtained in two parts, one part being the p output by the LSTMiAnd the other part is determined by a transfer matrix A of the CRF model.
Then the normalized probability can be obtained using Softmax:
during training, the log probability of the correct tag sequence is maximized:
wherein Y isXRepresenting all the marker sequences of sentence X. From the above description, it is clear that the present invention contemplates that the network architecture of the present invention generates efficient output tag sequences. In decoding, the present invention predicts the output sequence by (2):
since only binary interactions between outputs are modeled, dynamic programming can be used to compute the maximum a posteriori sequence y in sum (1) and in sum (2)*。
Bert-BiLSTM-CRF
The model is composed of three parts. Firstly, a Bert pre-training language model is utilized to encode a single character to obtain a word vector corresponding to the single character, then a BilSTM layer is utilized to carry out bidirectional encoding on an input text, finally, a semantic vector containing context information is input into a CRF layer to be decoded, and the CRF layer outputs a label sequence with the maximum probability to obtain the category of each character.
The method for generating and constructing the question and answer data set through the questions comprises the following specific steps of:
firstly, the QA and QG models are jointly trained by utilizing the probability correlation between the QA and the QG, and different methods are used for calculating corresponding items in the probability formula.
And step two, in order to solve the problem of invisible predicates, a fine adjustment component is used, a sequence-to-sequence model is trained by using copy operation and text evidences from Wikipedia, and the model can generate the invisible predicates according to triples extracted from a knowledge base.
And step three, optimizing the QA model by providing the generated problems and the triples extracted from the knowledge base, wherein the QA model is finely adjusted by inputting the generated problems and the triples extracted from the knowledge base.
Referring to fig. 1, the specific steps of constructing the question-answering system of the present invention are:
step one, analyzing a natural language question provided by a user; and performing word segmentation and word vector training on the question sentence. In the word segmentation process, a jieba tool is used, and a preset tourism field special dictionary is added, so that the word segmentation accuracy is improved.
And step two, carrying out named entity identification and relationship identification by using a Bert + BiLSTM + CRF model.
Thirdly, inquiring corresponding entities in the knowledge graph by using the identified entities and relations; if the knowledge graph has a matched entity, selecting the entity; and if the knowledge graph does not have the matched entity, selecting the closest entity by utilizing semantic similarity calculation. The invention utilizes a cosine similarity algorithm to calculate the semantic similarity, and the cosine similarity calculation is expressed as follows:
wherein X represents an entity extracted from the question, Y represents an entity in the knowledge graph, and n represents the dimension of the word vector. And selecting the entity in the knowledge graph with the highest similarity as matching by calculating the similarity.
Step four, matching the selected entities and attributes with triples in the knowledge graph;
and step five, returning the corresponding attribute value as an answer of the question to the user.
The invention also provides a device for constructing the intelligent question-answering system by the question generation data set, which comprises the following components:
the tourism domain knowledge map building module stores entity related information in a triple form of entity-attribute value, extracts entities in question sentences through named entity identification when a problem is raised, links the entities with a knowledge map to find corresponding entities, matches entity attributes by using question sentence information and combining a rule template set in advance, and returns corresponding attribute values;
the natural language question analyzing module is used for performing word segmentation and word vector training on the natural language question, and a jieba tool is used and a preset tourism field dictionary is added in the word segmentation process;
the entity extraction module is used for firstly utilizing a Bert pre-training language model to encode a single character to obtain a word vector corresponding to the single character, then utilizing a BilSTM layer to carry out bidirectional encoding on an input text, and finally inputting a semantic vector containing context information into a CRF layer to be decoded, wherein the CRF layer outputs a label sequence with the maximum probability to obtain the category of each character;
the entity matching module is used for matching the extracted entity with the entity in the knowledge graph; if the knowledge graph has a matched entity, selecting the entity; if the knowledge graph does not have the matched entity, selecting the closest entity through semantic similarity;
and the answer feedback module is used for matching the selected entity and attribute with the triples in the knowledge graph, returning corresponding attribute values, and providing the attributes serving as answers of the question to the user.
The invention also provides a terminal, which comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, wherein the processor realizes the steps of the method for constructing the intelligent question-answering system through the question generation data set when executing the computer program.
The present invention also provides a computer readable storage medium storing a computer program which, when executed by a processor, performs the steps of the method for constructing an intelligent question-and-answer system from question generated data sets.
The computer program may be divided into one or more modules/units, which are stored in the memory and executed by the processor to perform the method of the invention. The terminal can be a desktop computer, a notebook, a palm computer, a cloud server and other computing equipment, and can also be a processor and a memory.
The processor may be a Central Processing Unit (CPU), other general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, etc. The memory can be used for storing the computer program and/or the module, and the processor can realize various functions of the device for screening the wiring relation of the characteristic values of the backplane signals by running or executing the computer program and/or the module stored in the memory and calling the data stored in the memory.
The above description is only a preferred embodiment of the present invention, and is not intended to limit the technical solution of the present invention, and it should be understood by those skilled in the art that the technical solution can be modified and replaced by a plurality of simple modifications and replacements without departing from the spirit and principle of the present invention, and the modifications and replacements also fall within the protection scope covered by the claims.
Claims (9)
1. A method for constructing an intelligent question-answering system through question generation data sets is characterized by comprising the following steps:
step one, establishing a tourist field knowledge map;
storing entity related information in a knowledge graph in a triple form of 'entity-attribute value', when a problem is raised, firstly, extracting entities in a question through named entity identification, linking the entities with the knowledge graph to find corresponding entities, then, utilizing question information and a preset rule template to perform entity attribute matching, and returning corresponding attribute values;
step two, performing question analysis on a natural language question provided by a user, performing word segmentation and word vector training on the question, and adding a preset tourism field dictionary by using a jieba tool in the word segmentation process;
step three, utilizing a Bert-BilSTM + CRF model to perform entity extraction on the natural language question;
step four, matching the extracted entity with an entity in the knowledge graph; if the knowledge graph has a matched entity, selecting the entity; if the knowledge graph has no matched entities, performing semantic similarity calculation, and selecting the closest entities;
matching the selected entities and attributes with triples in the knowledge graph;
and step six, returning the corresponding attribute value as an answer of the question to the user.
2. The method for constructing an intelligent question-and-answer system through question generation data sets according to claim 1, wherein the constructed travel domain knowledge graph constructs question-and-answer data sets using question generation, wherein the task of question generation is to generate corresponding natural language questions given input answers, and wherein question answering is a reverse task of finding suitable answers in given questions.
3. The method for constructing an intelligent question-answering system through question generation data sets according to claim 1, wherein in the third step, a Bert pre-training language model is used to encode a single character to obtain a word vector corresponding to the single character, a BilSTM layer is used to encode an input text bidirectionally, and finally a semantic vector containing context information is input into a CRF layer to be decoded, wherein the CRF layer can output a label sequence with the maximum probability to obtain the category of each character.
4. The method for constructing an intelligent question-answering system through question generation data sets according to claim 1, wherein the fourth step utilizes a cosine similarity algorithm to perform semantic similarity calculation, the cosine similarity calculation being expressed as follows:
wherein X represents an entity extracted from the question, Y represents an entity in the knowledge graph, and n represents the dimension of a word vector; and selecting the entity in the knowledge graph with the highest similarity as matching by calculating the similarity.
5. The method for constructing an intelligent question-answering system through question generation data sets according to claim 1, characterized in that QG and QA models are combined, and by means of knowledge base and text corpus, the QA and QG models are first trained jointly on gold data sets, and then the QA models are fine-tuned by means of text by means of supplementary data sets constructed by the QG models.
6. The method according to claim 1, wherein the information related to the entity in step one is crawled by using Python tool, the information related to the entity comprises tourist attractions, names, profiles, construction time and scenic spot levels, and the crawled data is cleaned and collated to form a tourist knowledge map.
7. An apparatus for constructing an intelligent question-answering system from question-generating data sets, comprising:
the tourism domain knowledge map building module stores entity related information in a triple form of entity-attribute value, extracts entities in question sentences through named entity identification when a problem is raised, links the entities with a knowledge map to find corresponding entities, matches entity attributes by using question sentence information and combining a rule template set in advance, and returns corresponding attribute values;
the natural language question analyzing module is used for performing word segmentation and word vector training on the natural language question, and a jieba tool is used and a preset tourism field dictionary is added in the word segmentation process;
the entity extraction module is used for firstly utilizing a Bert pre-training language model to encode a single character to obtain a word vector corresponding to the single character, then utilizing a BilSTM layer to carry out bidirectional encoding on an input text, and finally inputting a semantic vector containing context information into a CRF layer to be decoded, wherein the CRF layer outputs a label sequence with the maximum probability to obtain the category of each character;
the entity matching module is used for matching the extracted entity with the entity in the knowledge graph; if the knowledge graph has a matched entity, selecting the entity; if the knowledge graph does not have the matched entity, selecting the closest entity through semantic similarity;
and the answer feedback module is used for matching the selected entity and attribute with the triples in the knowledge graph, returning corresponding attribute values, and providing the attributes serving as answers of the question to the user.
8. A terminal comprising a memory, a processor, and a computer program stored in the memory and executable on the processor, characterized in that: the processor, when executing the computer program, carries out the steps of the method of building an intelligent question-answering system from question-generating data sets according to any one of claims 1 to 6.
9. A computer-readable storage medium storing a computer program, characterized in that: the computer program realizing the steps of a method of building an intelligent question-answering system from question-generating data sets as claimed in any one of claims 1 to 6 when executed by a processor.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010956043.2A CN112100351A (en) | 2020-09-11 | 2020-09-11 | Method and equipment for constructing intelligent question-answering system through question generation data set |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010956043.2A CN112100351A (en) | 2020-09-11 | 2020-09-11 | Method and equipment for constructing intelligent question-answering system through question generation data set |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112100351A true CN112100351A (en) | 2020-12-18 |
Family
ID=73750928
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010956043.2A Pending CN112100351A (en) | 2020-09-11 | 2020-09-11 | Method and equipment for constructing intelligent question-answering system through question generation data set |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112100351A (en) |
Cited By (55)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112463944A (en) * | 2020-12-22 | 2021-03-09 | 安徽商信政通信息技术股份有限公司 | Retrieval type intelligent question-answering method and device based on multi-model fusion |
CN112507139A (en) * | 2020-12-28 | 2021-03-16 | 深圳力维智联技术有限公司 | Knowledge graph-based question-answering method, system, equipment and storage medium |
CN112528003A (en) * | 2020-12-24 | 2021-03-19 | 北京理工大学 | Multi-item selection question-answering method based on semantic sorting and knowledge correction |
CN112527999A (en) * | 2020-12-22 | 2021-03-19 | 江苏省农业科学院 | Extraction type intelligent question and answer method and system introducing agricultural field knowledge |
CN112650845A (en) * | 2020-12-30 | 2021-04-13 | 西安交通大学 | Question-answering system and method based on BERT and knowledge representation learning |
CN112667794A (en) * | 2020-12-31 | 2021-04-16 | 民生科技有限责任公司 | Intelligent question-answer matching method and system based on twin network BERT model |
CN112685550A (en) * | 2021-01-12 | 2021-04-20 | 腾讯科技(深圳)有限公司 | Intelligent question answering method, device, server and computer readable storage medium |
CN112733526A (en) * | 2021-01-28 | 2021-04-30 | 成都不问科技有限公司 | Extraction method for automatically identifying taxation objects in finance and tax file |
CN112749567A (en) * | 2021-03-01 | 2021-05-04 | 哈尔滨理工大学 | Question-answering system based on reality information environment knowledge graph |
CN112883175A (en) * | 2021-02-10 | 2021-06-01 | 武汉大学 | Meteorological service interaction method and system combining pre-training model and template generation |
CN113011196A (en) * | 2021-04-28 | 2021-06-22 | 广西师范大学 | Concept-enhanced representation and one-way attention-containing subjective question automatic scoring neural network model |
CN113076718A (en) * | 2021-04-09 | 2021-07-06 | 苏州爱语认知智能科技有限公司 | Commodity attribute extraction method and system |
CN113127626A (en) * | 2021-04-22 | 2021-07-16 | 广联达科技股份有限公司 | Knowledge graph-based recommendation method, device and equipment and readable storage medium |
CN113297368A (en) * | 2021-06-29 | 2021-08-24 | 中国人民解放军国防科技大学 | Intelligent auxiliary method for data link network guarantee demand analysis and application thereof |
CN113297369A (en) * | 2021-07-26 | 2021-08-24 | 中国科学院自动化研究所 | Intelligent question-answering system based on knowledge graph subgraph retrieval |
CN113312490A (en) * | 2021-04-28 | 2021-08-27 | 乐山师范学院 | Event knowledge graph construction method for emergency |
CN113449528A (en) * | 2021-08-30 | 2021-09-28 | 企查查科技有限公司 | Address element extraction method and device, computer equipment and storage medium |
CN113505586A (en) * | 2021-06-07 | 2021-10-15 | 中电鸿信信息科技有限公司 | Seat-assisted question-answering method and system integrating semantic classification and knowledge graph |
CN113535917A (en) * | 2021-06-30 | 2021-10-22 | 山东师范大学 | Intelligent question-answering method and system based on travel knowledge map |
CN113569554A (en) * | 2021-09-24 | 2021-10-29 | 北京明略软件系统有限公司 | Entity pair matching method and device in database, electronic equipment and storage medium |
CN113590844A (en) * | 2021-08-09 | 2021-11-02 | 北京智源人工智能研究院 | Knowledge graph-based question-answer library generation method and device, electronic equipment and storage medium |
CN113626553A (en) * | 2021-07-15 | 2021-11-09 | 人民网股份有限公司 | Cascade binary Chinese entity relation extraction method based on pre-training model |
CN113657125A (en) * | 2021-07-14 | 2021-11-16 | 内蒙古工业大学 | Knowledge graph-based Mongolian non-autoregressive machine translation method |
CN113672720A (en) * | 2021-09-14 | 2021-11-19 | 国网天津市电力公司 | Power audit question and answer method based on knowledge graph and semantic similarity |
CN113724819A (en) * | 2021-08-31 | 2021-11-30 | 平安国际智慧城市科技股份有限公司 | Training method, device, equipment and medium for medical named entity recognition model |
CN113763135A (en) * | 2021-09-18 | 2021-12-07 | 京东科技信息技术有限公司 | Commodity data comparison method and device, electronic equipment and storage medium |
CN113961667A (en) * | 2021-09-23 | 2022-01-21 | 哈尔滨工业大学(深圳) | Intelligent question-answering system based on Bert's dynamic threshold adjustment |
CN113987145A (en) * | 2021-10-22 | 2022-01-28 | 智联(无锡)信息技术有限公司 | Method, system, equipment and storage medium for accurately reasoning user attribute entity |
CN114003709A (en) * | 2021-12-30 | 2022-02-01 | 中国电子科技集团公司第二十八研究所 | Intelligent question-answering system and method based on question matching |
CN114036956A (en) * | 2021-11-18 | 2022-02-11 | 清华大学 | Tourism knowledge semantic analysis method and device |
CN114048298A (en) * | 2021-11-23 | 2022-02-15 | 中国民用航空总局第二研究所 | Intention understanding method, device, equipment and medium |
CN114048321A (en) * | 2021-08-12 | 2022-02-15 | 湖南达德曼宁信息技术有限公司 | Multi-granularity text error correction data set generation method, device and equipment |
CN114065741A (en) * | 2021-11-16 | 2022-02-18 | 北京有竹居网络技术有限公司 | Method, device, apparatus and medium for verifying the authenticity of a representation |
CN114153993A (en) * | 2022-02-07 | 2022-03-08 | 杭州远传新业科技有限公司 | Automatic knowledge graph construction method and system for intelligent question answering |
CN114218372A (en) * | 2021-12-17 | 2022-03-22 | 北京北大软件工程股份有限公司 | Knowledge graph retrieval method and system based on knowledge base representation |
CN114238653A (en) * | 2021-12-08 | 2022-03-25 | 华东师范大学 | Method for establishing, complementing and intelligently asking and answering knowledge graph of programming education |
CN114416947A (en) * | 2022-01-17 | 2022-04-29 | 中国科学技术大学 | Relation perception similar problem identification and evaluation method, system, equipment and storage medium |
CN114691845A (en) * | 2022-02-22 | 2022-07-01 | 北京市农林科学院 | Semantic search method and device, electronic equipment, storage medium and product |
CN115129828A (en) * | 2021-03-25 | 2022-09-30 | 科沃斯商用机器人有限公司 | Human-computer interaction method and device, intelligent robot and storage medium |
CN115168565A (en) * | 2022-07-07 | 2022-10-11 | 北京数美时代科技有限公司 | Cold start method, device, equipment and storage medium for vertical domain language model |
WO2022222716A1 (en) * | 2021-04-21 | 2022-10-27 | 华东理工大学 | Construction method and apparatus for chemical industry knowledge graph, and intelligent question and answer method and apparatus |
CN115248854A (en) * | 2022-09-22 | 2022-10-28 | 中邮消费金融有限公司 | Knowledge graph-based automatic question answering method, system and storage medium |
CN115599902A (en) * | 2022-12-15 | 2023-01-13 | 西南石油大学(Cn) | Oil-gas encyclopedia question-answering method and system based on knowledge graph |
CN115827844A (en) * | 2022-12-12 | 2023-03-21 | 之江实验室 | Knowledge graph question-answering method and system based on spark ql statement generation |
CN115859989A (en) * | 2023-02-13 | 2023-03-28 | 神州医疗科技股份有限公司 | Entity identification method and system based on remote supervision |
CN116010583A (en) * | 2023-03-17 | 2023-04-25 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | Cascade coupling knowledge enhancement dialogue generation method |
CN116662523A (en) * | 2023-08-01 | 2023-08-29 | 宁波甬恒瑶瑶智能科技有限公司 | Biochemical knowledge question-answering method, system and storage medium based on GPT model |
CN116737967A (en) * | 2023-08-15 | 2023-09-12 | 中国标准化研究院 | Knowledge graph construction and perfecting system and method based on natural language |
WO2023246093A1 (en) * | 2022-06-24 | 2023-12-28 | 重庆长安汽车股份有限公司 | Common question answering method and system, device and medium |
CN117520522A (en) * | 2023-12-29 | 2024-02-06 | 华云天下(南京)科技有限公司 | Intelligent dialogue method and device based on combination of RPA and AI and electronic equipment |
CN117539996A (en) * | 2023-11-21 | 2024-02-09 | 北京拓医医疗科技服务有限公司 | Consultation question-answering method and system based on user portrait |
CN117669718A (en) * | 2023-12-05 | 2024-03-08 | 广州鸿蒙信息科技有限公司 | Fire control knowledge training model and training method based on artificial intelligence |
CN117688189A (en) * | 2023-12-27 | 2024-03-12 | 珠江水利委员会珠江水利科学研究院 | Knowledge graph, knowledge base and large language model fused question-answering system construction method |
CN118132735A (en) * | 2024-05-07 | 2024-06-04 | 支付宝(杭州)信息技术有限公司 | Medical rule base generation method and device |
CN118210907A (en) * | 2024-05-20 | 2024-06-18 | 成都信息工程大学 | Knowledge graph vectorization representation-based intelligent question-answer model construction method |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111353030A (en) * | 2020-02-26 | 2020-06-30 | 陕西师范大学 | Knowledge question and answer retrieval method and device based on travel field knowledge graph |
CN111639163A (en) * | 2020-04-29 | 2020-09-08 | 深圳壹账通智能科技有限公司 | Problem generation model training method, problem generation method and related equipment |
-
2020
- 2020-09-11 CN CN202010956043.2A patent/CN112100351A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111353030A (en) * | 2020-02-26 | 2020-06-30 | 陕西师范大学 | Knowledge question and answer retrieval method and device based on travel field knowledge graph |
CN111639163A (en) * | 2020-04-29 | 2020-09-08 | 深圳壹账通智能科技有限公司 | Problem generation model training method, problem generation method and related equipment |
Non-Patent Citations (3)
Title |
---|
SEN HU ETC.: "How Question Generation Can Help Question Answering over Knowledge Base", 《NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING》 * |
王文华: "一种多模型融合的问题生成算法设计与实现", 《中国优秀硕士学位论文全文数据库(信息科技辑)》 * |
赵平 等: "基于BERT+BiLSTM+CRF的中文景点命名实体识别", 《计算机系统应用》 * |
Cited By (85)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112527999A (en) * | 2020-12-22 | 2021-03-19 | 江苏省农业科学院 | Extraction type intelligent question and answer method and system introducing agricultural field knowledge |
CN112527999B (en) * | 2020-12-22 | 2024-05-14 | 江苏省农业科学院 | Extraction type intelligent question-answering method and system for introducing knowledge in agricultural field |
CN112463944B (en) * | 2020-12-22 | 2023-10-24 | 安徽商信政通信息技术股份有限公司 | Search type intelligent question-answering method and device based on multi-model fusion |
CN112463944A (en) * | 2020-12-22 | 2021-03-09 | 安徽商信政通信息技术股份有限公司 | Retrieval type intelligent question-answering method and device based on multi-model fusion |
CN112528003B (en) * | 2020-12-24 | 2022-10-04 | 北京理工大学 | Multi-item selection question-answering method based on semantic sorting and knowledge correction |
CN112528003A (en) * | 2020-12-24 | 2021-03-19 | 北京理工大学 | Multi-item selection question-answering method based on semantic sorting and knowledge correction |
CN112507139A (en) * | 2020-12-28 | 2021-03-16 | 深圳力维智联技术有限公司 | Knowledge graph-based question-answering method, system, equipment and storage medium |
CN112507139B (en) * | 2020-12-28 | 2024-03-12 | 深圳力维智联技术有限公司 | Knowledge graph-based question and answer method, system, equipment and storage medium |
CN112650845B (en) * | 2020-12-30 | 2023-01-03 | 西安交通大学 | Question-answering system and method based on BERT and knowledge representation learning |
CN112650845A (en) * | 2020-12-30 | 2021-04-13 | 西安交通大学 | Question-answering system and method based on BERT and knowledge representation learning |
CN112667794A (en) * | 2020-12-31 | 2021-04-16 | 民生科技有限责任公司 | Intelligent question-answer matching method and system based on twin network BERT model |
CN112685550A (en) * | 2021-01-12 | 2021-04-20 | 腾讯科技(深圳)有限公司 | Intelligent question answering method, device, server and computer readable storage medium |
CN112685550B (en) * | 2021-01-12 | 2023-08-04 | 腾讯科技(深圳)有限公司 | Intelligent question-answering method, intelligent question-answering device, intelligent question-answering server and computer readable storage medium |
CN112733526B (en) * | 2021-01-28 | 2023-11-17 | 成都不问科技有限公司 | Extraction method for automatically identifying tax collection object in financial file |
CN112733526A (en) * | 2021-01-28 | 2021-04-30 | 成都不问科技有限公司 | Extraction method for automatically identifying taxation objects in finance and tax file |
CN112883175B (en) * | 2021-02-10 | 2022-06-14 | 武汉大学 | Meteorological service interaction method and system combining pre-training model and template generation |
CN112883175A (en) * | 2021-02-10 | 2021-06-01 | 武汉大学 | Meteorological service interaction method and system combining pre-training model and template generation |
CN112749567A (en) * | 2021-03-01 | 2021-05-04 | 哈尔滨理工大学 | Question-answering system based on reality information environment knowledge graph |
CN115129828A (en) * | 2021-03-25 | 2022-09-30 | 科沃斯商用机器人有限公司 | Human-computer interaction method and device, intelligent robot and storage medium |
CN113076718A (en) * | 2021-04-09 | 2021-07-06 | 苏州爱语认知智能科技有限公司 | Commodity attribute extraction method and system |
WO2022222716A1 (en) * | 2021-04-21 | 2022-10-27 | 华东理工大学 | Construction method and apparatus for chemical industry knowledge graph, and intelligent question and answer method and apparatus |
CN113127626B (en) * | 2021-04-22 | 2024-04-30 | 广联达科技股份有限公司 | Recommendation method, device, equipment and readable storage medium based on knowledge graph |
CN113127626A (en) * | 2021-04-22 | 2021-07-16 | 广联达科技股份有限公司 | Knowledge graph-based recommendation method, device and equipment and readable storage medium |
CN113312490A (en) * | 2021-04-28 | 2021-08-27 | 乐山师范学院 | Event knowledge graph construction method for emergency |
CN113011196A (en) * | 2021-04-28 | 2021-06-22 | 广西师范大学 | Concept-enhanced representation and one-way attention-containing subjective question automatic scoring neural network model |
CN113011196B (en) * | 2021-04-28 | 2023-01-10 | 陕西文都教育科技有限公司 | Concept-enhanced representation and one-way attention-containing subjective question automatic scoring neural network model |
CN113505586A (en) * | 2021-06-07 | 2021-10-15 | 中电鸿信信息科技有限公司 | Seat-assisted question-answering method and system integrating semantic classification and knowledge graph |
CN113297368B (en) * | 2021-06-29 | 2023-09-19 | 中国人民解放军国防科技大学 | Intelligent auxiliary method for data link network guarantee demand analysis and application thereof |
CN113297368A (en) * | 2021-06-29 | 2021-08-24 | 中国人民解放军国防科技大学 | Intelligent auxiliary method for data link network guarantee demand analysis and application thereof |
CN113535917A (en) * | 2021-06-30 | 2021-10-22 | 山东师范大学 | Intelligent question-answering method and system based on travel knowledge map |
CN113657125B (en) * | 2021-07-14 | 2023-05-26 | 内蒙古工业大学 | Mongolian non-autoregressive machine translation method based on knowledge graph |
CN113657125A (en) * | 2021-07-14 | 2021-11-16 | 内蒙古工业大学 | Knowledge graph-based Mongolian non-autoregressive machine translation method |
CN113626553B (en) * | 2021-07-15 | 2024-02-20 | 人民网股份有限公司 | Cascade binary Chinese entity relation extraction method based on pre-training model |
CN113626553A (en) * | 2021-07-15 | 2021-11-09 | 人民网股份有限公司 | Cascade binary Chinese entity relation extraction method based on pre-training model |
CN113297369B (en) * | 2021-07-26 | 2022-04-01 | 中国科学院自动化研究所 | Intelligent question-answering system based on knowledge graph subgraph retrieval |
CN113297369A (en) * | 2021-07-26 | 2021-08-24 | 中国科学院自动化研究所 | Intelligent question-answering system based on knowledge graph subgraph retrieval |
CN113590844A (en) * | 2021-08-09 | 2021-11-02 | 北京智源人工智能研究院 | Knowledge graph-based question-answer library generation method and device, electronic equipment and storage medium |
CN114048321A (en) * | 2021-08-12 | 2022-02-15 | 湖南达德曼宁信息技术有限公司 | Multi-granularity text error correction data set generation method, device and equipment |
CN114048321B (en) * | 2021-08-12 | 2024-08-13 | 湖南达德曼宁信息技术有限公司 | Multi-granularity text error correction data set generation method, device and equipment |
CN113449528B (en) * | 2021-08-30 | 2021-11-30 | 企查查科技有限公司 | Address element extraction method and device, computer equipment and storage medium |
CN113449528A (en) * | 2021-08-30 | 2021-09-28 | 企查查科技有限公司 | Address element extraction method and device, computer equipment and storage medium |
CN113724819B (en) * | 2021-08-31 | 2024-04-26 | 平安国际智慧城市科技股份有限公司 | Training method, device, equipment and medium for medical named entity recognition model |
CN113724819A (en) * | 2021-08-31 | 2021-11-30 | 平安国际智慧城市科技股份有限公司 | Training method, device, equipment and medium for medical named entity recognition model |
CN113672720A (en) * | 2021-09-14 | 2021-11-19 | 国网天津市电力公司 | Power audit question and answer method based on knowledge graph and semantic similarity |
CN113763135A (en) * | 2021-09-18 | 2021-12-07 | 京东科技信息技术有限公司 | Commodity data comparison method and device, electronic equipment and storage medium |
CN113961667A (en) * | 2021-09-23 | 2022-01-21 | 哈尔滨工业大学(深圳) | Intelligent question-answering system based on Bert's dynamic threshold adjustment |
CN113569554A (en) * | 2021-09-24 | 2021-10-29 | 北京明略软件系统有限公司 | Entity pair matching method and device in database, electronic equipment and storage medium |
CN113987145B (en) * | 2021-10-22 | 2024-02-02 | 智联网聘信息技术有限公司 | Method, system, equipment and storage medium for accurately reasoning user attribute entity |
CN113987145A (en) * | 2021-10-22 | 2022-01-28 | 智联(无锡)信息技术有限公司 | Method, system, equipment and storage medium for accurately reasoning user attribute entity |
CN114065741A (en) * | 2021-11-16 | 2022-02-18 | 北京有竹居网络技术有限公司 | Method, device, apparatus and medium for verifying the authenticity of a representation |
CN114036956A (en) * | 2021-11-18 | 2022-02-11 | 清华大学 | Tourism knowledge semantic analysis method and device |
CN114048298B (en) * | 2021-11-23 | 2024-05-31 | 中国民用航空总局第二研究所 | Intent understanding method, device, apparatus and medium |
CN114048298A (en) * | 2021-11-23 | 2022-02-15 | 中国民用航空总局第二研究所 | Intention understanding method, device, equipment and medium |
CN114238653A (en) * | 2021-12-08 | 2022-03-25 | 华东师范大学 | Method for establishing, complementing and intelligently asking and answering knowledge graph of programming education |
CN114238653B (en) * | 2021-12-08 | 2024-05-24 | 华东师范大学 | Method for constructing programming education knowledge graph, completing and intelligently asking and answering |
CN114218372A (en) * | 2021-12-17 | 2022-03-22 | 北京北大软件工程股份有限公司 | Knowledge graph retrieval method and system based on knowledge base representation |
CN114003709A (en) * | 2021-12-30 | 2022-02-01 | 中国电子科技集团公司第二十八研究所 | Intelligent question-answering system and method based on question matching |
CN114416947A (en) * | 2022-01-17 | 2022-04-29 | 中国科学技术大学 | Relation perception similar problem identification and evaluation method, system, equipment and storage medium |
CN114153993A (en) * | 2022-02-07 | 2022-03-08 | 杭州远传新业科技有限公司 | Automatic knowledge graph construction method and system for intelligent question answering |
CN114153993B (en) * | 2022-02-07 | 2022-05-06 | 杭州远传新业科技有限公司 | Automatic knowledge graph construction method and system for intelligent question answering |
CN114691845A (en) * | 2022-02-22 | 2022-07-01 | 北京市农林科学院 | Semantic search method and device, electronic equipment, storage medium and product |
CN114691845B (en) * | 2022-02-22 | 2024-10-01 | 北京市农林科学院 | Semantic search method, semantic search device, electronic equipment, storage medium and product |
WO2023246093A1 (en) * | 2022-06-24 | 2023-12-28 | 重庆长安汽车股份有限公司 | Common question answering method and system, device and medium |
CN115168565A (en) * | 2022-07-07 | 2022-10-11 | 北京数美时代科技有限公司 | Cold start method, device, equipment and storage medium for vertical domain language model |
CN115248854B (en) * | 2022-09-22 | 2023-03-03 | 中邮消费金融有限公司 | Automatic question-answering method, system and storage medium based on knowledge graph |
CN115248854A (en) * | 2022-09-22 | 2022-10-28 | 中邮消费金融有限公司 | Knowledge graph-based automatic question answering method, system and storage medium |
CN115827844A (en) * | 2022-12-12 | 2023-03-21 | 之江实验室 | Knowledge graph question-answering method and system based on spark ql statement generation |
CN115827844B (en) * | 2022-12-12 | 2023-08-08 | 之江实验室 | Knowledge graph question-answering method and system based on Sparql statement generation |
CN115599902A (en) * | 2022-12-15 | 2023-01-13 | 西南石油大学(Cn) | Oil-gas encyclopedia question-answering method and system based on knowledge graph |
CN115859989A (en) * | 2023-02-13 | 2023-03-28 | 神州医疗科技股份有限公司 | Entity identification method and system based on remote supervision |
CN116010583B (en) * | 2023-03-17 | 2023-07-18 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | Cascade coupling knowledge enhancement dialogue generation method |
CN116010583A (en) * | 2023-03-17 | 2023-04-25 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | Cascade coupling knowledge enhancement dialogue generation method |
CN116662523B (en) * | 2023-08-01 | 2023-10-20 | 宁波甬恒瑶瑶智能科技有限公司 | Biochemical knowledge question-answering method, system and storage medium based on GPT model |
CN116662523A (en) * | 2023-08-01 | 2023-08-29 | 宁波甬恒瑶瑶智能科技有限公司 | Biochemical knowledge question-answering method, system and storage medium based on GPT model |
CN116737967B (en) * | 2023-08-15 | 2023-11-21 | 中国标准化研究院 | Knowledge graph construction and perfecting system and method based on natural language |
CN116737967A (en) * | 2023-08-15 | 2023-09-12 | 中国标准化研究院 | Knowledge graph construction and perfecting system and method based on natural language |
CN117539996A (en) * | 2023-11-21 | 2024-02-09 | 北京拓医医疗科技服务有限公司 | Consultation question-answering method and system based on user portrait |
CN117669718B (en) * | 2023-12-05 | 2024-07-12 | 广州鸿蒙信息科技有限公司 | Fire control knowledge training model and training method based on artificial intelligence |
CN117669718A (en) * | 2023-12-05 | 2024-03-08 | 广州鸿蒙信息科技有限公司 | Fire control knowledge training model and training method based on artificial intelligence |
CN117688189A (en) * | 2023-12-27 | 2024-03-12 | 珠江水利委员会珠江水利科学研究院 | Knowledge graph, knowledge base and large language model fused question-answering system construction method |
CN117520522B (en) * | 2023-12-29 | 2024-03-22 | 华云天下(南京)科技有限公司 | Intelligent dialogue method and device based on combination of RPA and AI and electronic equipment |
CN117520522A (en) * | 2023-12-29 | 2024-02-06 | 华云天下(南京)科技有限公司 | Intelligent dialogue method and device based on combination of RPA and AI and electronic equipment |
CN118132735A (en) * | 2024-05-07 | 2024-06-04 | 支付宝(杭州)信息技术有限公司 | Medical rule base generation method and device |
CN118210907A (en) * | 2024-05-20 | 2024-06-18 | 成都信息工程大学 | Knowledge graph vectorization representation-based intelligent question-answer model construction method |
CN118210907B (en) * | 2024-05-20 | 2024-07-09 | 成都信息工程大学 | Knowledge graph vectorization representation-based intelligent question-answer model construction method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112100351A (en) | Method and equipment for constructing intelligent question-answering system through question generation data set | |
Li et al. | Oscar: Object-semantics aligned pre-training for vision-language tasks | |
Zhu et al. | Simple is not easy: A simple strong baseline for textvqa and textcaps | |
CN108829719B (en) | Non-fact question-answer selection method and system | |
CN110059217B (en) | Image text cross-media retrieval method for two-stage network | |
Zhu et al. | Knowledge-based question answering by tree-to-sequence learning | |
CN108628935B (en) | Question-answering method based on end-to-end memory network | |
CN111611361A (en) | Intelligent reading, understanding, question answering system of extraction type machine | |
CN112100346B (en) | Visual question-answering method based on fusion of fine-grained image features and external knowledge | |
CN112232053B (en) | Text similarity computing system, method and storage medium based on multi-keyword pair matching | |
Wu et al. | Switchable novel object captioner | |
CN111598041A (en) | Image generation text method for article searching | |
CN111291188A (en) | Intelligent information extraction method and system | |
CN113626589A (en) | Multi-label text classification method based on mixed attention mechanism | |
CN114169312A (en) | Two-stage hybrid automatic summarization method for judicial official documents | |
CN110516145B (en) | Information searching method based on sentence vector coding | |
CN114417851B (en) | Emotion analysis method based on keyword weighted information | |
CN113392265A (en) | Multimedia processing method, device and equipment | |
Cheng et al. | A semi-supervised deep learning image caption model based on Pseudo Label and N-gram | |
CN112183083A (en) | Abstract automatic generation method and device, electronic equipment and storage medium | |
CN114429132A (en) | Named entity identification method and device based on mixed lattice self-attention network | |
Gomez-Perez et al. | ISAAQ--Mastering Textbook Questions with Pre-trained Transformers and Bottom-Up and Top-Down Attention | |
CN112115253A (en) | Depth text ordering method based on multi-view attention mechanism | |
CN116662502A (en) | Method, equipment and storage medium for generating financial question-answer text based on retrieval enhancement | |
Song et al. | LSTM-in-LSTM for generating long descriptions of images |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201218 |