CN108052588B - Method for constructing automatic document question-answering system based on convolutional neural network - Google Patents

Method for constructing automatic document question-answering system based on convolutional neural network Download PDF

Info

Publication number
CN108052588B
CN108052588B CN201711309921.6A CN201711309921A CN108052588B CN 108052588 B CN108052588 B CN 108052588B CN 201711309921 A CN201711309921 A CN 201711309921A CN 108052588 B CN108052588 B CN 108052588B
Authority
CN
China
Prior art keywords
answer
question
vector
neural network
matrix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711309921.6A
Other languages
Chinese (zh)
Other versions
CN108052588A (en
Inventor
吴明晖
范旭民
金苍宏
朱凡微
赵品通
方格格
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University City College ZUCC
Original Assignee
Zhejiang University City College ZUCC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University City College ZUCC filed Critical Zhejiang University City College ZUCC
Priority to CN201711309921.6A priority Critical patent/CN108052588B/en
Publication of CN108052588A publication Critical patent/CN108052588A/en
Application granted granted Critical
Publication of CN108052588B publication Critical patent/CN108052588B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3344Query execution using natural language analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3347Query execution using vector based model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/335Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Abstract

A method for constructing a document automatic question-answering system based on a convolutional neural network comprises the following steps: step 1, constructing a theme document library; step 2, constructing a word vector model; step 3, matching the subjects; step 4, constructing a word vector matrix; step 5, semantic matching is carried out based on a semantic model of the convolutional neural network; the semantic model of the convolutional neural network is divided into three layers; the first layer is a convolutional neural network layer; the second layer is an attention layer; the third layer is a full connection layer; and 6, an answer selection process, namely selecting a matched answer. The method of the invention does not need to manually construct a synonym dictionary, saves a large amount of labor and time cost, can purposefully sample the semantics of the word context in the training process of the model, adds an attention mechanism in the network, and improves the contribution degree of some representative words to the semantics of the whole sentence.

Description

Method for constructing automatic document question-answering system based on convolutional neural network
Technical Field
The invention relates to the field of natural language processing and artificial intelligence, and provides a scheme for performing semantic modeling and semantic matching on a question and an answer under the background that a deep learning algorithm is applied to natural language processing in a large scale by applying a convolutional neural network algorithm.
Background
The most important technology in automatic question answering is sentence semantic matching, and most of the traditional methods construct rule combinations conforming to scenes on the basis of HowNet (HowNet), large-scale dictionaries and Harmony large synonym word forest tools to achieve the purpose of calculation. The method has the advantages that the calculation model is quickly established, the self semantics of the words can be effectively utilized, the model is quickly adjusted, and the method has the defect that the semantics of the word context, even the semantics of the whole sentence or paragraph, cannot be effectively utilized. Therefore, the traditional method easily causes the semantic loss of the word context, and the obtained result cannot be accurately used for calculating the matching degree between sentences.
Some traditional methods for semantic matching of sentences cannot effectively utilize the semantics of word context and have high requirements on labor and time cost, are not as good as the current popular deep learning methods in matching effect, and are difficult to adapt to the requirements of enterprises on automatic question and answer technologies in the background of the internet economic era with explosive increase of data volume.
Disclosure of Invention
The invention aims to provide a method for constructing a document automatic question-answering system based on a convolutional neural network. Therefore, the present invention adopts the following technical solutions.
A method for constructing a document automatic question-answering system based on a convolutional neural network comprises the following steps:
step 1, constructing a theme document library; establishing a theme document library according to different application scenes, wherein the theme document library comprises k theme documents and aims at k types of problems; each topic document corresponds to a question type and is a candidate answer set of the question type;
step 2, constructing a word vector model; obtaining a corpus, training the corpus by using a word2vec tool to obtain a word vector model, wherein one word corresponds to one word vector in the word vector model, L is the dimension of the word vector, the word vector can represent the distance between the words in a multi-dimensional space, and the word vector model can accurately represent the semantic similarity between the words;
step 3, matching the subjects; receiving a first question provided by a user, classifying the first question according to the topic document library constructed in the step 1, and finding out a first topic document corresponding to the first question, wherein n are contained in the first topic document1A plurality of alternative answers;
step 4, constructing a word vector matrix; dividing the first question into m words, and constructing a first question matrix A according to the word vector model in the step 2; according to the stepsThe first theme document obtained by theme matching in step 3 divides the alternative answers in the first theme document into m words, and constructs n words according to the word vector model in step 21Each first answer matrix Q corresponds to one alternative answer; construction n1A word vector matrix M, wherein M ═ M<A,Q>;
Step 5, semantic matching is carried out based on a semantic model of the convolutional neural network; the semantic model of the convolutional neural network is divided into three layers;
the first layer is a convolutional neural network layer, the input of the convolutional neural network layer is a word vector matrix M, the width of a convolutional kernel of the convolutional neural network layer is the dimension of the word vector, and the number of the convolutional kernels is n2(ii) a Inputting the word vector matrix M into the convolutional neural network layer to obtain n2Problem feature vector sum n of dimensions2A dimensional answer feature vector; take n3Convolution kernel of different heights to obtain n3A question feature vector and an answer feature vector; n is to be3Combining the problem feature vectors into a problem feature matrix, and combining n3Combining the answer feature vectors into an answer feature matrix;
the second layer is an attention layer, and the attention layer is used for weighting the question feature vector and the answer feature vector; the input of the attention layer is a question feature matrix and an answer feature matrix, and the output is a question sentence vector and an answer sentence vector;
the third layer is a full connection layer, the full connection layer is used for calculating the semantic matching degree between the question sentence vector and the answer sentence vector, and the semantic matching degree is expressed by a semantic matching degree value;
step 6, answer selection process; according to step 5, according to n1A word vector matrix M, calculating n1And selecting a matching answer according to the semantic matching degree value.
Preferably, the weighting formula of the attention layer is as follows: g ═ softmax (tanh (IW + b) e), i ═ GI; wherein, I is an input feature matrix, W is a weighting matrix, b and e are weighting vectors, softmax is a normalization function, G is a weighting weight vector, and I is an output sentence vector; substituting the question feature matrix and the answer feature matrix into the formula, and outputting a question sentence vector and an answer sentence vector.
Preferably, the word segmentation process of the first question and the first answer adopts a word segmentation method based on an N-gram model.
Preferably, said step 6 comprises adding n1And arranging the semantic matching degree values, and selecting the alternative answer corresponding to the maximum semantic matching degree value as the matching answer.
Preferably, said step 6 comprises adding n1And arranging the semantic matching degree values, setting a first threshold value, and selecting an answer set of alternative answers corresponding to the semantic matching degree values larger than the first threshold value as matching answers.
The invention has the beneficial effects that: (1) the method of the present invention aims to learn "features" generated from a large set of input data, which play an important role in semantic modeling. (2) The semantics are represented by the vector model obtained by training, and a synonym dictionary does not need to be manually constructed, so that a large amount of labor and time cost is saved. (3) The semantics of the word context can be purposefully sampled during the training of the model. (4) And an attention mechanism is added into the network, so that the contribution degree of some representative words to the whole sentence semantics is improved.
Drawings
Fig. 1 is a schematic overall flow chart of an embodiment of the present invention.
FIG. 2 is a schematic diagram of the structure of the document library of the present invention.
FIG. 3 is a flow chart illustrating the matching process of the subject invention.
FIG. 4 is a flowchart illustrating an application environment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
As shown in fig. 1 to 4, a method for constructing a document automatic question-answering system based on a convolutional neural network includes the following steps:
step 1, constructing a theme document library; establishing a theme document library according to different application scenes, wherein the theme document library comprises k theme documents and aims at k problems; each topic document corresponds to a question type and is a candidate answer set of the question type. Each question has its corresponding answer set, and a document can be regarded as an alternative answer set for a certain topic. A complete document library is built based on the types of questions serviced by the system. For example, "ask for your store for all goods supporting national joint insurance? "," ask for return of goods? "and the like with respect to online shopping returns. The corresponding answer set can enjoy national joint guarantee service by 'certificate of warranty and local store invoice', and the customer can return goods without reason within 7 days of purchasing goods. "etc., each answer being a sentence. Based on the above description, the subject documents in the scenario are established according to different application scenarios, and each document corresponds to a question type.
Step 2, constructing a word vector model; obtaining a corpus, training the corpus by using a word2vec tool to obtain a word vector model, wherein one word corresponds to one word vector w in the word vector model, L is the dimension of the word vector, the word vector can represent the distance between words in a multidimensional space, the word vector model can accurately represent the semantic similarity between words, and L can be 100. The Chinese corpus provided by Chinese Wikipedia and dog searching laboratories and comment data crawled from microblogs are collected, and the size of the Chinese corpus is about 5 GB. The method comprises the steps of training the materials by using a word2vec tool open by Google to obtain a word vector model, wherein each word in the model corresponds to a vector and is used for representing the distance between words in a multidimensional space, and the model can accurately represent the semantic similarity between Chinese words.
Step 3, matching the subjects; receiving a first question provided by a user, classifying the first question according to the topic document library constructed in the step 1, and finding out a first topic document corresponding to the first question, wherein n is in the first topic document1The number of alternative answers, where n should not exceed 50 in general per set1And taking 50. On the basis that step 1 has been completed, it is assumed that k documents have already been created, for the k-class problem. After receiving a question posed by a user, the question needs to be classified to find out a corresponding subject. On this result, the answer that best matches the question is found from the document for that topic. In order to accurately match the theme, the invention provides two matching modes: (1) the user selects the theme such as 'return goods', 'warranty' and the like autonomously. (2) And the system is matched automatically.
Step 4, constructing a word vector matrix; dividing the first question into m words, and constructing a first question matrix A according to the word vector model in the step 2; dividing the alternative answers in the first subject document into m words according to the first subject document obtained by subject matching in the step 3, and constructing n words according to the word vector model in the step 21A first answer matrix Q, each alternative answer corresponding to a first answer matrix, i.e. forming an m × L first question matrix A, n1M × L first answer matrix
Figure BDA0001502772030000041
Where n is1And taking 50.
Figure BDA0001502772030000042
w1,w2,…wmAfter the first problem matrix A is divided into m words, a word vector corresponding to each word.
Figure BDA0001502772030000043
w1,w2,…wmAfter the first question matrix Q is divided into m words, a word vector corresponding to each word. The word vector may represent the distance of the word in L space. Where m may be 50, when a problem arisesOr the answer is too short to be divided into 50 words, the corresponding positions of the first question matrix a and the first question matrix Q may be complemented by 0 or by random numbers. Construction n1I.e. 50 word vector matrices M, where M ═ M<A,Q>. And (4) according to the result of the theme matching in the step (2), obtaining a first document D corresponding to the Q, wherein 50 alternative answers exist in the D. The step is to divide the words of the question and the answer, obtain word vectors corresponding to the question and the answer according to the word vector model in the step 2, then construct 50 question-answer pairs, and use a word vector matrix M as the input<A,Q>Represents, a<A,Q>A word vector matrix corresponding to the question and the answer. Each one of which is<A,Q>Will be used as input for the next step. In order to disambiguate the characters of the questions and the answers as much as possible, a word segmentation method based on an N-gram model is adopted in the word segmentation process of the questions and the answers.
Step 5, semantic matching is carried out based on a semantic model of the convolutional neural network; the semantic model of the convolutional neural network is divided into three layers;
the first layer is a convolutional neural network layer, the input of the convolutional neural network layer is a word vector matrix M, the width of a convolutional kernel of the convolutional neural network layer is the dimension of a word vector, and the number of the convolutional kernels is n2Where n is2Taking 100; inputting the word vector matrix M into the convolutional neural network layer to obtain n2Dimensions, namely a 100-dimensional question feature vector and a 100-dimensional answer feature vector. The height of the convolution kernel can be regarded as the height of a window for sampling word context, and because the sampled features of different window heights have a certain difference, 4 kinds of height convolution kernels are taken here, the height is between 2 and 5, and the number of each convolution kernel is 100, so as to take enough features as possible. The number of layers of the neural network here is generally 1. . After sampling, maximum pooling and other operations are performed on each convolution kernel, a 100-dimensional feature vector is obtained, for a sentence, 4 kinds of convolution kernels obtain 4 100-dimensional feature vectors, and the 4 feature vectors are spliced to form a 4 x 100 matrix. Taking 4 kinds of high convolution kernel to obtain 4 100-dimensional question feature vectors and answer feature vectors, combining 4 question feature vectors into 4 x 100 questionTopic feature matrix AcCombining 4 answer feature vectors into a 4 × 100 answer feature matrix Qc
The second layer is an attention layer for weighting the question feature vector and the answer feature vector. And adding an attention mechanism, weighting word vectors corresponding to different words in the first question and the first answer by different weight values, and improving the contribution degree of some representative words to the whole sentence semantics. The input of the attention layer is a problem feature matrix AcAnd answer feature matrix QcThe outputs are question sentence vector a and answer sentence vector b. The weighting formula for the attention layer is: g ═ softmax (tanh (IW + b) e), i ═ GI. Wherein, I is an input feature matrix of 4 × 100, W is a weighting matrix of 100 × 100, b and e are weighting vectors of 100 dimensions, softmax is a normalization function, G is a weighting weight vector, and I is an output sentence vector. W, b and e are all random values initially, and are optimized by an Adam random optimization method in the network training process. The values in the 4-dimensional vector G obtained after normalization by the softmax function are added to 1. Problem feature matrix AcAnd answer feature matrix QcSubstituting the above formula, and outputting 100-dimensional question sentence vector a and answer sentence vector b.
The third layer is a full-connection layer, the full-connection layer is used for calculating semantic matching degree between the question sentence vector a and the answer sentence vector b, the semantic matching degree is expressed by semantic matching degree values, the value of each score value is [0,1], and represents the cosine similarity after normalization. The semantic matching degree can be calculated based on the distance between the question sentence vector a and the answer sentence vector b.
And 6, answer selection process. According to the step 5, 50 semantic matching degree values are calculated according to the 50 word vector matrixes M, and matching answers are selected according to the semantic matching degree values. The 50 semantic matching degree values can be arranged, and the alternative answer corresponding to the largest semantic matching degree value is selected as the matching answer. Or arranging the 50 semantic matching degree values, setting a first threshold value, and selecting an answer set of alternative answers corresponding to the semantic matching degree values larger than the first threshold value as matching answers.
Missing problems and documents can be remedied. If the system successfully matches the user question with the answer, the round of question answering is considered to be completed successfully. Otherwise, the background needs to record and is used as a basis for perfecting the document library later. The data to be recorded are: all questions, documents to which the system matches the question, questions to which the system fails to successfully match the document, questions to which the system does not successfully match the answer in the document. On the basis of the data, the system user can carry out corresponding document and answer compensation work.
The foregoing embodiments are merely illustrative of the principles and utilities of the present invention and are not intended to limit the invention. Any person skilled in the art can modify or change the above-mentioned embodiments without departing from the spirit and scope of the present invention. Accordingly, it is intended that all equivalent modifications or changes which can be made by those skilled in the art without departing from the spirit and technical spirit of the present invention be covered by the claims of the present invention.

Claims (4)

1. A method for constructing a document automatic question-answering system based on a convolutional neural network is characterized by comprising the following steps:
step 1, constructing a theme document library; establishing a theme document library according to different application scenes, wherein the theme document library comprises k theme documents and aims at k types of problems; each topic document corresponds to a question type and is a candidate answer set of the question type;
step 2, constructing a word vector model; obtaining a corpus, training the corpus by using a word2vec tool to obtain a word vector model, wherein one word corresponds to one word vector in the word vector model, L is the dimension of the word vector, the word vector can represent the distance between the words in a multi-dimensional space, and the word vector model can accurately represent the semantic similarity between the words;
step 3, matching the subjects; receiving a first question provided by a user, classifying the first question according to the topic document library constructed in the step 1, and finding out a first topic document corresponding to the first questionN in the first subject document1A plurality of alternative answers;
step 4, constructing a word vector matrix; dividing the first question into m words, and constructing a first question matrix A according to the word vector model in the step 2; dividing the alternative answers in the first subject document into m words according to the first subject document obtained by subject matching in the step 3, and constructing n words according to the word vector model in the step 21Each first answer matrix Q corresponds to one alternative answer; construction n1A word vector matrix M, wherein M ═ M<A,Q>;
Step 5, semantic matching is carried out based on a semantic model of the convolutional neural network; the semantic model of the convolutional neural network is divided into three layers;
the first layer is a convolutional neural network layer, the input of the convolutional neural network layer is a word vector matrix M, the width of a convolutional kernel of the convolutional neural network layer is the dimension of the word vector, and the number of the convolutional kernels is n2(ii) a Inputting the word vector matrix M into the convolutional neural network layer to obtain n2Problem feature vector sum n of dimensions2A dimensional answer feature vector; take n3Convolution kernel of different heights to obtain n3A question feature vector and an answer feature vector; n is to be3Combining the problem feature vectors into a problem feature matrix, and combining n3Combining the answer feature vectors into an answer feature matrix;
the second layer is an attention layer, and the attention layer is used for weighting the question feature vector and the answer feature vector; the input of the attention layer is a question feature matrix and an answer feature matrix, and the output is a question sentence vector and an answer sentence vector;
the third layer is a full connection layer, the full connection layer is used for calculating the semantic matching degree between the question sentence vector and the answer sentence vector, and the semantic matching degree is expressed by a semantic matching degree value;
step 6, answer selection process; according to step 5, according to n1A word vector matrix M, calculating n1Selecting a matching answer according to the semantic matching degree value;
the input of the attention layer is a problem feature matrix AcAnd answer feature matrix QcOutputting a question sentence vector a and an answer sentence vector b; the weighting formula for the attention layer is: g ═ softmax (tanh (IW + b) e), i ═ GI; wherein, I is an input feature matrix of 4 multiplied by 100, W is a weighting matrix of 100 multiplied by 100, b and e are weighting vectors of 100 dimensions, softmax is a normalization function, G is a weighting weight vector, and I is an output sentence vector; w, b and e are all random values initially, and are optimized by an Adam random optimization method in the network training process; adding the values in the 4-dimensional vector G obtained by the normalization of the softmax function to be 1; problem feature matrix AcAnd answer feature matrix QcSubstituting the above formula, and outputting 100-dimensional question sentence vector a and answer sentence vector b.
2. The method for constructing the convolutional neural network-based document automatic question-answering system according to claim 1, wherein a word segmentation method based on an N-gram model is adopted in a word segmentation process performed on the first question and the first answer.
3. The method for constructing a convolutional neural network-based document automatic question-answering system according to claim 1 or 2, wherein the step 6 comprises the step of connecting n1And arranging the semantic matching degree values, and selecting the alternative answer corresponding to the maximum semantic matching degree value as the matching answer.
4. The method for constructing a convolutional neural network-based document automatic question-answering system according to claim 1 or 2, wherein the step 6 comprises the step of connecting n1And arranging the semantic matching degree values, setting a first threshold value, and selecting an answer set of alternative answers corresponding to the semantic matching degree values larger than the first threshold value as matching answers.
CN201711309921.6A 2017-12-11 2017-12-11 Method for constructing automatic document question-answering system based on convolutional neural network Active CN108052588B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711309921.6A CN108052588B (en) 2017-12-11 2017-12-11 Method for constructing automatic document question-answering system based on convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711309921.6A CN108052588B (en) 2017-12-11 2017-12-11 Method for constructing automatic document question-answering system based on convolutional neural network

Publications (2)

Publication Number Publication Date
CN108052588A CN108052588A (en) 2018-05-18
CN108052588B true CN108052588B (en) 2021-03-26

Family

ID=62123968

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711309921.6A Active CN108052588B (en) 2017-12-11 2017-12-11 Method for constructing automatic document question-answering system based on convolutional neural network

Country Status (1)

Country Link
CN (1) CN108052588B (en)

Families Citing this family (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110597966A (en) * 2018-05-23 2019-12-20 北京国双科技有限公司 Automatic question answering method and device
CN108763535B (en) * 2018-05-31 2020-02-07 科大讯飞股份有限公司 Information acquisition method and device
CN108875074B (en) * 2018-07-09 2021-08-10 北京慧闻科技发展有限公司 Answer selection method and device based on cross attention neural network and electronic equipment
CN109145290B (en) * 2018-07-25 2020-07-07 东北大学 Semantic similarity calculation method based on word vector and self-attention mechanism
CN109086386B (en) * 2018-07-26 2023-04-28 腾讯科技(深圳)有限公司 Data processing method, device, computer equipment and storage medium
CN109522920B (en) * 2018-09-18 2020-10-13 义语智能科技(上海)有限公司 Training method and device of synonymy discriminant model based on combination of semantic features
CN111177328B (en) * 2018-11-12 2023-04-28 阿里巴巴集团控股有限公司 Question-answer matching system and method, question-answer processing device and medium
CN109582773B (en) * 2018-11-29 2020-11-27 深圳爱问科技股份有限公司 Intelligent question-answer matching method and device
CN111382573A (en) * 2018-12-12 2020-07-07 百度在线网络技术(北京)有限公司 Method, apparatus, device and storage medium for answer quality assessment
CN109815318A (en) * 2018-12-24 2019-05-28 平安科技(深圳)有限公司 The problems in question answering system answer querying method, system and computer equipment
CN111382264B (en) * 2018-12-27 2023-06-09 阿里巴巴集团控股有限公司 Session quality evaluation method and device and electronic equipment
CN109766423A (en) * 2018-12-29 2019-05-17 上海智臻智能网络科技股份有限公司 Answering method and device neural network based, storage medium, terminal
CN109766424B (en) * 2018-12-29 2021-11-19 安徽省泰岳祥升软件有限公司 Filtering method and device for reading understanding model training data
CN109783817B (en) * 2019-01-15 2022-12-06 浙江大学城市学院 Text semantic similarity calculation model based on deep reinforcement learning
CN114298310A (en) * 2019-01-29 2022-04-08 北京金山数字娱乐科技有限公司 Length loss determination method and device
CN111611355A (en) * 2019-02-25 2020-09-01 北京嘀嘀无限科技发展有限公司 Dialog reply method, device, server and storage medium
CN110046244B (en) * 2019-04-24 2021-06-08 中国人民解放军国防科技大学 Answer selection method for question-answering system
CN110298037B (en) * 2019-06-13 2023-08-04 同济大学 Convolutional neural network matching text recognition method based on enhanced attention mechanism
CN110956039A (en) * 2019-12-04 2020-04-03 中国太平洋保险(集团)股份有限公司 Text similarity calculation method and device based on multi-dimensional vectorization coding
CN111192682B (en) * 2019-12-25 2024-04-09 上海联影智能医疗科技有限公司 Image exercise data processing method, system and storage medium
CN111488438B (en) * 2020-02-21 2022-07-29 天津大学 Question-answer matching attention processing method, computer equipment and storage medium
CN111831789B (en) * 2020-06-17 2023-10-24 广东工业大学 Question-answering text matching method based on multi-layer semantic feature extraction structure
CN112597311B (en) * 2020-12-28 2023-07-11 东方红卫星移动通信有限公司 Terminal information classification method and system based on low-orbit satellite communication

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9659248B1 (en) * 2016-01-19 2017-05-23 International Business Machines Corporation Machine learning and training a computer-implemented neural network to retrieve semantically equivalent questions using hybrid in-memory representations
CN106776545A (en) * 2016-11-29 2017-05-31 西安交通大学 A kind of method that Similarity Measure between short text is carried out by depth convolutional neural networks
CN106897268A (en) * 2017-02-28 2017-06-27 科大讯飞股份有限公司 Text semantic understanding method, device and system
CN107368547A (en) * 2017-06-28 2017-11-21 西安交通大学 A kind of intelligent medical automatic question-answering method based on deep learning
CN107391623A (en) * 2017-07-07 2017-11-24 中国人民大学 A kind of knowledge mapping embedding grammar for merging more background knowledges

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9659248B1 (en) * 2016-01-19 2017-05-23 International Business Machines Corporation Machine learning and training a computer-implemented neural network to retrieve semantically equivalent questions using hybrid in-memory representations
CN106776545A (en) * 2016-11-29 2017-05-31 西安交通大学 A kind of method that Similarity Measure between short text is carried out by depth convolutional neural networks
CN106897268A (en) * 2017-02-28 2017-06-27 科大讯飞股份有限公司 Text semantic understanding method, device and system
CN107368547A (en) * 2017-06-28 2017-11-21 西安交通大学 A kind of intelligent medical automatic question-answering method based on deep learning
CN107391623A (en) * 2017-07-07 2017-11-24 中国人民大学 A kind of knowledge mapping embedding grammar for merging more background knowledges

Also Published As

Publication number Publication date
CN108052588A (en) 2018-05-18

Similar Documents

Publication Publication Date Title
CN108052588B (en) Method for constructing automatic document question-answering system based on convolutional neural network
US11176328B2 (en) Non-factoid question-answering device
CN108363790B (en) Method, device, equipment and storage medium for evaluating comments
CN109783817B (en) Text semantic similarity calculation model based on deep reinforcement learning
CN110110062B (en) Machine intelligent question and answer method and device and electronic equipment
CN109670191B (en) Calibration optimization method and device for machine translation and electronic equipment
CN111597314B (en) Reasoning question-answering method, device and equipment
CN108549658B (en) Deep learning video question-answering method and system based on attention mechanism on syntax analysis tree
TW202009749A (en) Human-machine dialog method, device, electronic apparatus and computer readable medium
WO2020062770A1 (en) Method and apparatus for constructing domain dictionary, and device and storage medium
CN111259127B (en) Long text answer selection method based on transfer learning sentence vector
CN110083693B (en) Robot dialogue reply method and device
CN107870964B (en) Statement ordering method and system applied to answer fusion system
CN108763535B (en) Information acquisition method and device
CN110516245A (en) Fine granularity sentiment analysis method, apparatus, computer equipment and storage medium
CN110008327B (en) Legal answer generation method and device
CN107480196B (en) Multi-modal vocabulary representation method based on dynamic fusion mechanism
CN110516070B (en) Chinese question classification method based on text error correction and neural network
CN111914067A (en) Chinese text matching method and system
CN110162596B (en) Training method and device for natural language processing, automatic question answering method and device
CN107832297B (en) Feature word granularity-oriented domain emotion dictionary construction method
CN111460145A (en) Learning resource recommendation method, device and storage medium
CN111507093A (en) Text attack method and device based on similar dictionary and storage medium
CN112632253A (en) Answer extraction method and device based on graph convolution network and related components
CN109977203A (en) Statement similarity determines method, apparatus, electronic equipment and readable storage medium storing program for executing

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20220712

Address after: 310015 No. 51, Huzhou street, Hangzhou, Zhejiang

Patentee after: Zhejiang University City College

Address before: 310015 No. 51 Huzhou street, Hangzhou, Zhejiang, Gongshu District

Patentee before: Zhejiang University City College

TR01 Transfer of patent right