CN110750616A - Retrieval type chatting method and device and computer equipment - Google Patents

Retrieval type chatting method and device and computer equipment Download PDF

Info

Publication number
CN110750616A
CN110750616A CN201910985485.7A CN201910985485A CN110750616A CN 110750616 A CN110750616 A CN 110750616A CN 201910985485 A CN201910985485 A CN 201910985485A CN 110750616 A CN110750616 A CN 110750616A
Authority
CN
China
Prior art keywords
question
semantic
answer
text
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910985485.7A
Other languages
Chinese (zh)
Other versions
CN110750616B (en
Inventor
邵建智
毛晓曦
范长杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Netease Hangzhou Network Co Ltd
Original Assignee
Netease Hangzhou Network Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Netease Hangzhou Network Co Ltd filed Critical Netease Hangzhou Network Co Ltd
Priority to CN201910985485.7A priority Critical patent/CN110750616B/en
Publication of CN110750616A publication Critical patent/CN110750616A/en
Application granted granted Critical
Publication of CN110750616B publication Critical patent/CN110750616B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/31Indexing; Data structures therefor; Storage structures
    • G06F16/316Indexing structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3344Query execution using natural language analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a retrieval type chatting method, a retrieval type chatting device and computer equipment, relates to the technical field of natural language processing, and solves the technical problem that sentences with many co-occurring characters and large difference in actual meanings influence retrieval results to obtain replies far away from expectation. The method is applied to a chat robot, a text index library and a semantic index library are prestored in the chat robot, problem indexes in the text index library are text feature indexes, and problem indexes in the semantic index library are semantic vector indexes; the method comprises the following steps: determining a first alternative answer set through text features of a target question to be answered based on a text index library; determining a second alternative answer set through a semantic vector of the target question based on a semantic index library; and determining at least one target answer with the highest matching degree with the target question as a reply of the target question according to the first candidate answer set and the second candidate answer set based on a pre-trained matching model.

Description

Retrieval type chatting method and device and computer equipment
Technical Field
The present application relates to the field of natural language processing technologies, and in particular, to a retrieval type chat method, apparatus, and computer device.
Background
With the rapid development of networks and information technologies, the internet has become a main means for people to acquire information, and there are two general means for acquiring information from the internet: one is to search keywords through a search engine to obtain related information; the other is to ask questions and obtain corresponding answers by providing customer service to a website or social software. The former is based on keyword search, users are difficult to express own retrieval intention clearly, related web pages returned by a search engine are numerous, and users are difficult to quickly locate required information; the latter is based on natural language to ask questions, can clearly express the user's intention, but manual asking and answering needs to provide a large amount of customer service to answer, causes the sharp increase of human cost. To solve the human cost problem, various large websites or product manufacturers adopt an "auto-answer" chat robot to answer questions for users, such as smismimi in korea, Siri in apple, kyoto customer service, mini i robot, microsoft mini ice, microsoft mini na (Cortana), and the like.
The chat robot is an important field to which a Natural Language Processing (NLP) technique is applied, and replies based mainly on a search method. Namely, reply is made by relying on the existing question-answering library. For knowledge question answering in the general/open field, a best matching answer can be found out by maintaining a large amount of general/open field knowledge and using a retrieval mode; the current approach is primarily to determine whether two problems are close by co-occurrence of words where the two problems occur. However, due to the complexity of natural language, two sentences with many co-occurring words have great difference in actual meanings, and thus, the reply is far from the expected reply.
Disclosure of Invention
The invention aims to provide a retrieval type chatting method, a retrieval type chatting device and computer equipment, which are used for solving the technical problem that sentences with many co-occurring characters and large difference in actual meanings influence retrieval results to obtain replies far away from expectation.
In a first aspect, an embodiment of the present application provides an indexing type chat method, which is applied to a chat robot, where a text index library and a semantic index library are prestored in the chat robot, a problem index in the text index library is a text feature index, and a problem index in the semantic index library is a semantic vector index; the method comprises the following steps:
determining a first alternative answer set through text features of a target question to be answered based on the text index library;
determining a second alternative answer set through a semantic vector of the target question based on the semantic index library;
and determining at least one target answer with the highest matching degree with the target question as a reply of the target question according to the first candidate answer set and the second candidate answer set based on a pre-trained matching model.
In one possible implementation, the text index library includes a first question-and-answer corpus, and a question index in the text index library is used for indexing a question in the first question-and-answer corpus; the step of determining a first alternative answer set by the text features of the target question to be answered based on the text index library comprises the following steps:
determining a first preset number of first questions with the maximum text similarity with the target question according to the text characteristics of the target question to be answered based on the text index library;
and determining a first alternative answer set according to the corresponding answer of each first question in the first question-answer pair corpus.
In one possible implementation, the step of determining, based on the text index library and according to text features of the target question to be answered, a first preset number of first questions having a maximum text similarity with the target question includes:
based on the text index library, determining the text similarity of the current question and the target question according to one or more items of the number of co-occurring words appearing between the retrieved current question and the target question, the proportion of the co-occurring words in the current question and the target question respectively, and the importance degree of the co-occurring words in the first question-answer pair corpus;
determining a first preset number of first questions with the maximum text similarity with the target question in the questions of the first question-and-answer corpus;
the current question is any one of the questions retrieved from the text index library according to the text features of the target question, and the lower the frequency of the co-occurrence word appearing in the first question-answer pair corpus is, the higher the importance degree is.
In one possible implementation, the step of determining, based on the text index library and according to text features of the target question to be answered, a first preset number of first questions having a maximum text similarity with the target question includes:
based on the text index library, determining the text similarity of the current question and the target question according to the retrieved Jacard distance between the two word sets corresponding to the current question and the target question;
determining a first preset number of first questions with the maximum text similarity with the target question in the questions of the first question-and-answer corpus;
and the current question is any one of the questions retrieved from the text index library according to the text characteristics of the target question.
In one possible implementation, the semantic index library includes a first question-and-answer corpus, and the question index in the semantic index library is used for indexing a question in the first question-and-answer corpus; a step of determining a second candidate answer set by a semantic vector of the target question based on the semantic index library, comprising:
determining a second preset number of second questions with the maximum semantic similarity with the target question according to the semantic vector of the target question to be answered based on the semantic index library;
and determining a second alternative answer set according to the corresponding answer of each second question in the first question-answer pair corpus.
In one possible implementation, the step of determining, based on the semantic index library, a second preset number of second questions having a maximum semantic similarity with the target question according to the semantic vector of the target question to be answered includes:
determining a semantic vector of the target problem;
based on the semantic index library, sorting the problems in the semantic index library according to semantic similarity between the current semantic vector and the semantic vector of the target problem;
selecting a second preset number of second problems with the largest semantic similarity according to the sequence of the problems in the semantic index library;
and the current semantic vector is any one of the semantic vectors in the semantic index library.
In one possible implementation, the semantic vector is extracted by a pre-trained natural language processing NLP model.
In one possible implementation, the step of ordering the questions in the semantic index library according to semantic similarity between the current semantic vector and the semantic vector of the target question based on the semantic index library includes:
based on the semantic index library, determining cosine similarity between a front semantic vector and a semantic vector of the target problem;
and sequencing the problems in the semantic index library according to the cosine similarity.
In one possible implementation, the NLP model includes any one of a Bidirectional Encoding Representation (BERT) model, a word vector Language (ELMo) model, and a Generative Pre-Training Language (GPT) model.
In one possible implementation, the pre-trained matching model is obtained by training an initial matching model according to a training sample, where the training sample includes a positive sample and a negative sample; the positive sample is an original question in a second question-answer pair material and an original answer corresponding to the original question; the negative sample is the least similar answer in a third specified number of answers most similar to the original answer in the original question and the second question-and-answer corpus.
In one possible implementation, the method further includes: and sequentially taking a third preset number of answers which are most similar to the text similarity and/or the semantic similarity of the original answer in the second question-answer corpus as current answers to execute the following steps: if the text similarity and/or the semantic similarity between the current answer and the original answer is smaller than a preset value, determining that the current answer is the most dissimilar answer to the original answer;
and if the text similarity and/or the semantic similarity between the third preset number of most similar answers and the original answer are larger than preset values, randomly selecting one answer from the third preset number of most similar answers as the least similar answer to the original answer.
In one possible implementation, the first question-answer pair corpus is the same as the second question-answer pair corpus.
In a second aspect, an indexing chat device is provided, which is applied to a chat robot, wherein a text index library and a semantic index library are prestored in the chat robot, problem indexes in the text index library are text feature indexes, and problem indexes in the semantic index library are semantic vector indexes; the device comprises:
the first determination module is used for determining a first alternative answer set through text characteristics of a target question to be answered based on the text index library;
a second determining module, configured to determine, based on the semantic index library, a second candidate answer set through a semantic vector of the target question;
and the third determining module is used for determining at least one target answer with the highest matching degree with the target question as the reply of the target question according to the first candidate answer set and the second candidate answer set based on a pre-trained matching model.
In a third aspect, an embodiment of the present application further provides a computer device, including a memory and a processor, where the memory stores a computer program executable on the processor, and the processor implements the method of the first aspect when executing the computer program.
In a fourth aspect, this embodiment of the present application further provides a computer-readable storage medium storing machine executable instructions, which, when invoked and executed by a processor, cause the processor to perform the method of the first aspect.
The embodiment of the application brings the following beneficial effects: the embodiment of the application provides a retrieval type chatting method, a retrieval type chatting device and computer equipment. Firstly, a plurality of alternative answer sets are determined from a text index library and a semantic index library respectively through text features and semantic vectors of a target question, then, based on a pre-trained matching model, final answers are selected from the alternative answer sets, similarity is measured from the perspective of text and semantics, high-quality alternative answers are finally selected, the final answers determined based on the high-quality alternative answers can better accord with expected results, and user experience is more friendly.
In order to make the aforementioned objects, features and advantages of the present application more comprehensible, preferred embodiments accompanied with figures are described in detail below.
Drawings
In order to more clearly illustrate the detailed description of the present application or the technical solutions in the prior art, the drawings needed to be used in the detailed description of the present application or the prior art description will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present application, and other drawings can be obtained by those skilled in the art without creative efforts.
Fig. 1 is a schematic flowchart of a retrieval chat method according to an embodiment of the present application;
FIG. 2 is a schematic flow chart illustrating a training process of a matching model according to an embodiment of the present disclosure;
fig. 3 is another schematic flow chart of a retrieval chat method according to an embodiment of the present application;
fig. 4 is a schematic structural diagram of an indexing chat apparatus according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of a computer device according to an embodiment of the present application.
Detailed Description
To make the objects, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions of the present application will be clearly and completely described below with reference to the accompanying drawings, and it is obvious that the described embodiments are some, but not all embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms "comprises" and "comprising," and any variations thereof, as used in the embodiments of this application, are intended to cover non-exclusive inclusions. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those steps or elements but may alternatively include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
Embodiments of the present invention are further described below with reference to the accompanying drawings.
Fig. 1 is a schematic flowchart of a retrieval chat method according to an embodiment of the present application. The method is applied to a chat robot, a text index library and a semantic index library are pre-stored in the chat robot, problem indexes in the text index library are text feature indexes, and problem indexes in the semantic index library are semantic vector indexes.
As an example, the chat robot may correspond to a terminal provided with a graphical user interface for interacting with a user, and the terminal may transmit a target question to be answered to the chat robot.
As another example, the chat robot may also be provided locally with a graphical user interface for interacting with the user to retrieve the targeted questions to be answered.
For example, a graphical user interface provided by the chat robot or terminal may be used to display a question input area and an answer display area. The user may enter a target question to be answered for the question entry area. The terminal determines an input target question to be answered in response to an input operation for the question input area. The input operation may include a variety of ways, such as a physical input operation (e.g., a keyboard input) triggered in response to pressing a physical key; triggering touch input operation by responding to the touch virtual key; or by responding to a slide operation on a touch screen, a triggered handwriting input operation, or the like.
Of course, the terminal or the chat robot may also provide an audio device locally for interacting with the user to retrieve the target question to be answered. For example, a microphone provided by the chat robot or the terminal may acquire audio uttered by the user, and then, by recognizing the audio, obtain a target question to be answered, and the terminal determines a reply to the target question in response to the target question. The chat robot or terminal may also provide a speaker to output answers corresponding to the target questions.
As shown in fig. 1, the method may specifically include the following steps:
and S110, determining a first alternative answer set through text characteristics of the target question to be answered based on the text index library.
The text index library may be built for the material based on the first question and answer. An inverted index (inverted index) table may be established according to text features corresponding to questions in the first question-answer pair corpus, and each entry in the inverted index table may include one text feature and each question having the text feature. Each question-answer pair in the question-answer pair corpus can be indexed by the inverted index table. The text feature may refer to a word or phrase.
For example, the question-answer pair is "question: you are beautiful; and (3) answer: thank you for your likes, "the text index library generated based on this question-answer pair may be as shown in table 1.
TABLE 1
Figure BDA0002235662800000081
Figure BDA0002235662800000091
Through the text features included in the target question, a plurality of questions with the same text features can be retrieved from the text index library, and the first alternative answer set can be determined according to answers corresponding to the questions with the same text features.
In addition, in the chat robot, a first question-answer corpus may be predetermined; the first question-answer pair material can be collected in advance and obtained through pretreatment. Preprocessing the collected corpus pairs may include data cleansing, which may include filtering out yellow-related, political, and storm-related sentences via sensitive thesaurus and rules; and filtering sentences with too long or too short length, then putting the data into a labeling platform for manual labeling, and further manually filtering all improper question-answer pairs to obtain a high-quality first question-answer pair corpus.
And S120, determining a second alternative answer set through the semantic vector of the target question based on the semantic index library.
The semantic index library may also be built for the material based on the first question and answer. The semantic index table may be established according to a semantic vector corresponding to a question in the first question-answer pair corpus, and each entry in the semantic index table may include a semantic vector and a question-answer pair. Each question-answer pair in the question-answer pair corpus can be indexed by the inverted index table. The text feature may refer to a word or phrase. Additionally, in some embodiments, the semantic index library may also be built based on question-answer corpus that is different from the first question-answer corpus.
For example, the question-answer pair is "question: you are beautiful; and (3) answer: thank you for your likes, "the text index library generated based on this question-answer pair may be as shown in table 1.
And S130, determining at least one target answer with the highest matching degree with the target question as a reply of the target question according to the first candidate answer set and the second candidate answer set based on a pre-trained matching model.
According to the retrieval type chatting method provided by the embodiment of the application, a plurality of alternative answer sets are determined from a text index library and a semantic index library respectively through text features and semantic vectors of target questions, final answers are selected from the alternative answer sets based on a pre-trained matching model, the similarity is measured from two angles of texts and semantics, high-quality alternative answers are selected finally, the final answers determined based on the high-quality alternative answers can better accord with expected results, and therefore user experience is improved.
The pre-trained matching model may be obtained by training an initial matching model according to a training sample. As shown in fig. 2, the training samples may include positive samples and negative samples.
In some embodiments, the positive sample may be a second question-answer pair of an original question in the material and an original answer corresponding to the original question; the negative examples are the least similar answers of a third specified number of answers in the original question and the second challenge-response corpus that are most similar to the original answers.
For example, in the second question-answer corpus, there are 5 answers most similar to the original answer, and similarity comparisons are performed within these 5 answers, and the answer most dissimilar to the original answer is taken as a negative sample.
The negative sample obtained by the method can ensure that the positive sample and the negative sample have certain similarity in return and are not too similar. Therefore, the situation that the positive sample and the negative sample have overlarge difference, the positive sample and the original problem contain co-occurrence words, the negative sample basically cannot have the situation, the matching model excessively relates to the characteristics related to the co-occurrence words, and the matching scores of all replies to the problem and the co-occurrence words are high during prediction, so that the matching effect is reduced. On the other hand, the situation that the difference between the positive sample and the negative sample is too small is avoided, and the negative sample can also be used as the reply of the original problem, so that a large amount of noise appears in training data and the normal training of the matching model is interfered.
In some embodiments, further comprising: and sequentially taking a third preset number of answers which are most similar to the text similarity and/or the semantic similarity of the original answer in the second question-answer pair corpus as current answers to execute the following steps:
if the text similarity and/or the semantic similarity between the current answer and the original answer is smaller than a preset value, determining that the current answer is the most dissimilar answer to the original answer;
and if the text similarity and/or the semantic similarity between the third preset number of most similar answers and the original answer are larger than a preset value, randomly selecting one answer from the third preset number of most similar answers as the least similar answer to the original answer.
As shown in fig. 2, a plurality of replies most similar to the original reply (original answer) are retrieved from the question-answer corpus by using the inverted index table, after the nth most similar reply is obtained, whether the similarity between the reply and the original reply is lower than a threshold value or not can be calculated, if so, the nth most similar reply is taken as a negative sample, and if not, the negative sample is randomly sampled from the most similar replies so as to train the matching model by using the positive sample and the negative sample.
For example, in the second question-answer corpus, there are 5 answers with the similarity greater than the first threshold with the original answer, and in the 5 answers, a negative sample is selected from the answers with the similarity lower than the second threshold with the original answer, of course, the similarity of the second threshold is greater than the first threshold; if there is no answer with similarity lower than the second threshold from among the 5 answers, one answer is randomly selected as a negative sample from among the 5 answers.
By the negative sample determining method, a relatively more appropriate negative sample can be determined under various conditions, so that the responses of the positive sample and the negative sample are ensured to have certain similarity and are not too similar.
As shown in fig. 3, the second question-answer corpus may be the same as the first question-answer corpus in step S110 (question-answer corpus in fig. 3). For example, a third preset number of answers with the most similar text similarity to the original answer may be retrieved by using the text index library in the above steps, and a plurality of answers with the same text features are retrieved in the text index library according to the text features included in the original answer; or, the third preset number of answers with the most similar semantic similarity to the original answer can be retrieved by separately or simultaneously utilizing the semantic index library in the above steps, and a plurality of answers with the same semantic vector are retrieved in the semantic index library through the semantic vector included in the original answer.
For example, the original answer in the first question-answer corpus is input into the text index library to obtain a plurality of answers similar to the original answer text. Of course, the original answer may also be input into the semantic index library to obtain a plurality of answers similar to the original answer text.
Therefore, a plurality of answers similar to the original answer text and/or semantics can be quickly and accurately obtained by directly utilizing the text index library and/or the semantic index library in the steps, and the efficiency of data processing is improved.
In some embodiments, the step S110 may specifically include:
step 1.1), determining a first preset number of first questions with the maximum text similarity to the target question according to the text characteristics of the target question to be answered based on a text index library;
step 1.2), determining a first alternative answer set according to the corresponding answer of each first question in the first question-answer corpus.
For example, as shown in fig. 3, a number of first questions having the most similar text features to the target question to be answered may be retrieved from the text index library by using the inverted index table, and then answers corresponding to the first questions are queried in the first question-answer corpus, and the corresponding answers are used as the first candidate answer set.
The first alternative answer set obtained by the method can fully utilize the text characteristics of the target question to be answered, and retrieve the answer corresponding to the question with the maximum text similarity to the target question.
In some embodiments, the text similarity of the two questions is determined according to one or more of the number of co-occurring words appearing between the two questions, the proportion of the co-occurring words in each question of the two questions, and the importance degree of the co-occurring words in the question-answer corpus; wherein the lower the frequency of occurrence of a word in question-answering expectations, the higher the degree of importance. Exemplarily, the step 1.1) may specifically include:
based on a text index library, determining the text similarity of the current question and the target question according to one or more items of the number of co-occurring words between the retrieved current question and the target question, the ratio of the co-occurring words in the current question and the target question respectively, and the importance degree of the co-occurring words in the first question-answer pair corpus;
determining a first preset number of first questions with the maximum similarity with the target question text in the questions of the first question-and-answer corpus;
the current question is any one of the questions retrieved from the text index library according to the text features of the target question, and the lower the frequency of the co-occurrence words appearing in the first question-answer pair corpus is, the higher the importance degree is.
Of course, the BM25 algorithm may be used to retrieve the problem with higher similarity to the target problem from the text index database. It should be noted that the BM25 algorithm can calculate the matching degree between the target question and all the questions in the text index library, and obtain a plurality of questions with the highest matching degree.
For example, the BM25 algorithm is used to analyze the words and phrases in the target question, retrieve the question that the words and phrases also appear from the text index database, and consider the frequency of the common words and phrases appearing between the target question and all the questions in the text index database, and the importance of the words and phrases (e.g., the less the number of times the words and phrases appear in all the questions, the greater the importance).
By the retrieval method, various factors can be considered based on the text to retrieve the problem most similar to the target problem text in the text index database.
In some embodiments, the two question text similarity is determined according to the Jacard distance between the two sets of words to which the two questions correspond. Exemplarily, the step 1.1) may specifically include:
determining the text similarity of the current problem and the target problem according to the retrieved Jacard distance between two word sets corresponding to the current problem and the target problem based on a text index library;
determining a first preset number of first questions with the maximum similarity with the target question text in the questions of the first question-and-answer corpus;
the current question is any one of the questions retrieved from the text index database according to the text features of the target question.
It should be noted that the Jaccard distance (Jaccard) can be used to calculate the similarity index between the texts of two questions to measure the similarity between the two questions. In this embodiment, the similarity index between the two questions is: the number of the morphemes of the intersection between the two word sets corresponding to the texts of the two questions is divided by the number of the morphemes of the union. The formula is as follows:
Figure BDA0002235662800000141
for example, there are 3 words in the current question and 5 words in the target question, where there are 2 words that co-occur between the current question and the target question. Then the jackard distance between the current question and the target question is: 2/(3+5) ═ 1/4.
Of course, the similarity threshold may also be preset. For example, when calculating the similarity between the current question and the target question, assuming that the number of words appearing in both questions is less than 4 and the jlcard distance is less than 0.3, the answer may be considered to be dissimilar from the original answer, and is taken as a negative sample. It can be seen that the number of words (2 words) that the two questions appear together is less than 4, and moreover, the jaccard distance 1/4 between the current question and the target question is less than the preset 0.3, so that it can be determined that the text similarity between the current question and the target question is low.
By the method for determining the text similarity between the two texts by using the Jacard distance, the similarity between the two problems can be more accurately and reasonably calculated, so that the problem with the maximum text similarity with the target problem can be more accurately determined.
In some embodiments, the semantic index library comprises a first query-response corpus, and the question index in the semantic index library is used for indexing a question in the first query-response corpus; the step S120 specifically includes:
step 2.1), determining a second preset number of second questions with the maximum semantic similarity with the target question according to the semantic vector of the target question to be answered based on the semantic index library;
and 2.2) determining a second alternative answer set according to the corresponding answer of each second question in the first question-answer pair corpus.
By the method for determining the semantic similarity according to the semantic vector, two questions which do not contain the same words but have similar semantics can be searched, so that the second alternative answer set can better meet the actual conditions of various questions.
For example, the target question is "you are beautiful", compared with the question of "you are really looking good" in the first question-answer corpus, the two questions do not have the same words except for the "you", but the semantics are very close, and the existing text retrieval mode cannot retrieve similar questions. The method provided by the embodiment of the application can be used for calculating that the beautiful semantic vector is very similar to the long-life and good-looking semantic vector, so that the problem that the question is long-life and good-looking is similar to the target problem is determined, the corresponding answer of the question that the question is long-life and good-looking in the first question-answer corpus is also included into the alternative answer set, and the problem that the answer of the more suitable target problem is omitted due to the fact that no common words exist is avoided.
In some embodiments, the pre-establishing a corresponding relationship between the semantic vector of each question in the first question-answer pair corpus and the question, for example, step 2.1) may specifically include:
step 2.11), determining a semantic vector of the target problem;
step 2.12), based on the semantic index library, sorting the problems in the semantic index library according to the semantic similarity between the current semantic vector and the semantic vector of the target problem;
step 2.13), selecting a second preset number of second problems with the largest semantic similarity according to the sequence of the problems in the semantic index library;
and the current semantic vector is any one of the semantic vectors in the semantic index library.
For step 2.11), the target question may be encoded into a semantic vector using an encoder, as shown in fig. 3. Specifically, the target question is input to an encoder, encoded into a fixed-length vector by the encoder, and the fixed-length vector is used as a semantic vector.
For step 2.12), because the semantic vector of each question in the first question-answer corpus has a corresponding relationship with the question, the similarity between each question in the first question-answer corpus and the target question can be obtained according to the corresponding relationship and the semantic similarity by calculating the semantic similarity between any one semantic vector in the semantic index library and the semantic vector of the target question, and then all the questions in the first question-answer corpus can be subjected to similarity ranking according to the similarity of each question.
Step 2.13), selecting a plurality of second questions with the largest semantic similarity from the first question-answer pair material according to the sequence obtained in the step 2.12).
By the method for determining the semantic similarity according to the semantic vector, two problems which do not contain the same word and have similar semantics can be searched, and the semantic similarity between the two problems is calculated through the semantic vector, so that the search result is more accurate and precise.
In some embodiments, the semantic vectors are extracted from a pre-trained natural language processing NLP model.
Illustratively, a natural language processing NLP model may be trained in advance, and each question in the target question and the first question-answer corpus may be converted into a semantic word vector of a fixed length by using the trained NLP model, where each question may correspond to several semantic word vectors.
Through the pre-trained NLP model, the problem can be converted into the corresponding semantic word vector more quickly and efficiently, and the semantic word vector output by the NLP model is more accurate.
In some embodiments, the NLP model includes any one of a BERT model, an ELMo model, and a GPT model.
For the specific conversion process of the word vector, the BERT model is taken as an example for explanation.
Firstly, inputting a target problem into a BERT model, wherein the BERT model converts each word in the target problem into a one-dimensional original word vector through codes in a searched word vector table (dictionary), and the original word vector is used for expressing the self word meaning of the word; then, aiming at each word in the target problem, the BERT model calculates the mapping relation between the word and other words in the target problem by using the original word vector, and the mapping relation is blended into the original word vector of the word to obtain a multi-dimensional (the BERT model is 512-dimensional) semantic word vector, wherein the semantic word vector represents the comprehensive semantics of the word in the target problem where the word is located and can also be understood as vector information after full text (all texts in the target problem) semantic information is blended; and then, outputting the vector fused with the full-text semantic information corresponding to each word in the target problem by the BERT model. It can also be understood that the output of the BERT model is a vector representation that enhances the semantics of each word in the target problem separately.
It should be noted that the meaning of a word (a word or a phrase, which is used to refer to the word in this embodiment) expressed in a text is generally related to its context. For example, it is difficult to analyze the semantic expression of "target" by analyzing only one semantic expression of "target", but it is easy to obtain a more precise and richer semantic expression of "target" by analyzing in connection with the context "target" of "target". Therefore, the context information of a word helps to enhance the semantic representation of the word, and the exact and rich semantic expression of the word is convenient to obtain. Furthermore, different words in the context tend to play different roles in enhancing the semantic representation. For example, "hong" has the greatest effect on analyzing the "target" semantic, while "it" has a relatively small effect.
In this embodiment, the BERT model can determine the action relationship between each word in the input target problem and other words in the target problem, so as to enhance the semantic representation of the word by using the context word information, so that the semantic word vector corresponding to each word output by the BERT model can comprehensively and accurately represent the semantics of the word in the whole target problem.
The concrete implementation method of the BERT model is described below.
The BERT model uses a Transformer coder as a language model, the Transformer language model utilizes a self-attention (self-attention) mechanism, and upper and lower layers in the BERT model are directly and completely connected with each other due to the self-attention mechanism.
For the Attention mechanism, it mainly involves three concepts: query, Key, and Value. In the semantic expression process of the enhanced word, the target word and the word of the context thereof have respective original Value (original word vector), the Attention mechanism takes the target word as Query and each word of the context thereof as Key, takes the calculated similarity between Query and each Key as weight, and fuses the weight of each word of the context into the original Value of the target word to obtain the Value (semantic word vector) of the enhanced semantic of the target word, and takes the Value as the output of Attention.
For a text (question) input into a BERT model, the Attention mechanism needs to enhance the semantic expression vector of each word, and therefore, each word needs to be used as Query, and semantic information of all other words in the text is weighted and fused to obtain the enhanced semantic vector of each word. In this case, the vectors of Query, Key and Value are all from the same input text (same target problem), and therefore, the Attention mechanism is also called Self-Attention.
The Self-orientation may have a plurality of headers (Multi-head), for which the Self-orientation mechanism may be understood as a Multi-way vector fusion for a plurality of different semantic scenes. For example, "a city changhuangqiao" may have different semantic expressions for this word under different semantic scenes: "the bridge of Yangtze river/the city of A", or "the bridge of Yangtze river/the city of A". For the length in the sentence, the correct semantic unit can be formed only by combining the length with the length in the former semantic scene; in the latter semantic scenario, it needs to be combined with "city" to form a correct semantic unit. The purpose of the Self-orientation mechanism is to enhance the semantic representation of the target word with other words in the text, and under different semantic scenarios, the words focused on by Self-orientation are different. Therefore, the Multi-head Self-orientation can be understood as considering different fusion modes of semantic vectors of a target word and other words in a text under various semantic scenes. It can be seen that the output form of each head in the Multi-head Self-orientation is completely the same, and is a semantic word vector formed by fusing full-text semantic information with each word. Finally, the Self-extension can linearly combine the semantic word vectors of a plurality of heads corresponding to each word, thereby obtaining a final semantic word vector with the same length as the original word vector after semantic enhancement.
For the whole BERT model, besides outputting the semantic word vector fused with the full text (target problem) semantic information, the semantic sentence vector of the target problem can also be output. For example, after inputting a target question, the BERT model inserts a [ CLS ] symbol before the target question, and takes the semantic sentence vector output corresponding to the symbol as the overall text semantic representation of the target question, which can be used for text classification. It can also be understood that the semantic sentence vector can more comprehensively and objectively fuse semantic information of all words in the target question, such as overall positive emotion, negative emotion of a sentence, and the like, compared with each semantic word vector in the target question. In practical application, a vector corresponding to the first character "[ CLS ]" output by the BERT model is taken as a semantic sentence vector.
Furthermore, for a 512-dimensional semantic word vector in the BERT model, it can be understood that 512 dimensions are used to represent the semantic word vector, and one dimension may correspond to a feature value of a feature.
For example, to determine the relationship of some words to Gender characteristics, assuming that the Gender of a male is-1 and the Gender of a female is +1, then the characteristic value for the Gender characteristics of man may be-1 and the characteristic value for the Gender characteristics of woman may be + 1. According to this rule, it can be found that the gender feature value of king is-0.95, that of queen is +0.97, and that neither apple nor orange is sexed. As another example, another feature may be the Royal degree of these words, in which man, wman and nobility are hardly related, and apple and orange are also hardly related, so that the feature value of their nobility feature may be close to 0. And king and queen are expensive, and the eigenvalues of their expensive characteristics can be determined to be large.
Of course, there may be many other features such as Size, Cost, alive, Action, Noun, etc.
In the BERT model, there are 512 different features, so that a 512-dimensional vector can be formed to represent the man word, and a 512-dimensional vector can also be used to represent the wman word. For example, by using this representation method to represent the two words "apple" and "orange", the two words "apple" and "orange" belong to the similar fruit, and the two words may be very similar in their characteristic representations, but some characteristics may be different, such as color and taste of orange and color and taste of apple, or some other characteristics may be different. In general, most of the features of the applet and the orange are similar in nature, have many similar feature values, and also have a small part of different feature values, so that the features of the orange and the juice can be effectively quantized, and the vector relationship between the orange and the juice can be conveniently calculated.
For a 512-dimensional feature vector, it can be assumed that the 512-dimensional data is embedded in a two-dimensional space or a three-dimensional space, so that the 512-dimensional data can be visualized. In the two-dimensional or three-dimensional space, the words man and wman are gathered together, the words king and queen are gathered together, the words are human and also are gathered together, the words are gathered together, the animals are gathered together, and the fruits are gathered together, so that the characteristic relation between each word is mapped in the visual space. For a 512-dimensional feature vector, a 512-dimensional space, similar to a two-dimensional space or a three-dimensional space, may be assumed, and for the BERT model, the 512-dimensional space is used to represent the semantic word vector of each word.
In some embodiments, the similarity of the semantic vectors of the two questions is determined according to the cosine similarity of the two semantic vectors. Exemplarily, the step 2.12) may specifically include:
based on a semantic index library, determining cosine similarity between the front semantic vector and the semantic vector of the target problem;
and sorting the problems in the semantic index library according to the cosine similarity.
It should be noted that, since similar words have similar semantic word vectors, the distance of the similar semantic word vectors in the feature space is closer. In a high-dimensional space, the similarity in the vector space can be used to represent the similarity in the word meaning in the text, and the distance between words with similar meanings in the feature vector space is relatively close, and the directions pointed by the two vectors are also very close.
In this embodiment, the cosine similarity between the target word vector and the semantic word vector in the semantic index library is used to determine the semantic similarity therebetween. That is to say, the value of the included angle between two word vectors can be used as a measure of the similarity between the two words, and the more similar the two words are, the more similar the corresponding semantic word vectors are, the smaller the included angle between the semantic word vectors is, and the higher the cosine similarity is. By means of a cosine function, the degree of closeness between two word vectors can be calculated, as follows:
Figure BDA0002235662800000201
a and B are two word vectors respectively, cos (theta) represents the cosine of an included angle between the two word vectors A and B, and similarity represents the similarity between words corresponding to the two word vectors A and B.
Therefore, the distance between two semantic word vectors can be calculated more accurately by utilizing the cosine similarity, so that the similarity between two words can be determined more accurately, and the problem similarity sequencing result in the semantic index library is more accurate.
Fig. 4 provides a schematic diagram of a structure of a retrievable chat device. The device is applied to a chat robot, a text index library and a semantic index library are prestored in the chat robot, problem indexes in the text index library are text feature indexes, and problem indexes in the semantic index library are semantic vector indexes; as shown in fig. 4, the retrieval type chatting apparatus 400 includes:
a first determining module 401, configured to determine, based on the text index library, a first candidate answer set according to text features of a target question to be answered;
a second determining module 402, configured to determine a second candidate answer set through a semantic vector of the target question based on the semantic index library;
and a third determining module 403, configured to determine, based on a pre-trained matching model, at least one target answer with the highest matching degree with the target question according to the first candidate answer set and the second candidate answer set, as a reply to the target question.
In some embodiments, the text index library comprises a first question-and-answer corpus, and the question index in the text index library is used for indexing the question in the first question-and-answer corpus;
the first determination module 401 includes:
the first determining submodule is used for determining a first preset number of first questions with the maximum text similarity with the target question according to the text characteristics of the target question to be answered based on the text index library;
and the second determining submodule is used for determining a first alternative answer set according to the corresponding answer of each first question in the first question-answer pair corpus.
In some embodiments, the first determination submodule is specifically configured to:
based on a text index library, determining the text similarity of the current question and the target question according to one or more items of the number of co-occurring words between the retrieved current question and the target question, the ratio of the co-occurring words in the current question and the target question respectively, and the importance degree of the co-occurring words in the first question-answer pair corpus;
determining a first preset number of first questions with the maximum similarity with the target question text in the questions of the first question-and-answer corpus;
the current question is any one of the questions retrieved from the text index library according to the text features of the target question, and the lower the frequency of the co-occurrence words appearing in the first question-answer pair corpus is, the higher the importance degree is.
In some embodiments, the first determination submodule is specifically configured to:
determining the text similarity of the current problem and the target problem according to the retrieved Jacard distance between two word sets corresponding to the current problem and the target problem based on a text index library;
determining a first preset number of first questions with the maximum similarity with the target question text in the questions of the first question-and-answer corpus;
the current question is any one of the questions retrieved from the text index database according to the text features of the target question.
In some embodiments, the semantic index library comprises a first query-response corpus, and the question index in the semantic index library is used for indexing a question in the first query-response corpus; the second determining module 402 includes:
the third determining submodule is used for determining a second preset number of second questions with the maximum semantic similarity with the target question according to the semantic vector of the target question to be answered based on the semantic index library;
and the fourth determining submodule is used for determining a second alternative answer set according to the corresponding answer of each second question in the first question-answer pair corpus.
In some embodiments, the third determination submodule is specifically configured to:
determining a semantic vector of the target problem;
based on a semantic index library, sorting the problems in the semantic index library according to semantic similarity between the current semantic vector and the semantic vector of the target problem;
selecting a second preset number of second problems with the largest semantic similarity according to the sequence of the problems in the semantic index library;
and the current semantic vector is any one of the semantic vectors in the semantic index library.
In some embodiments, the semantic vectors are extracted from a pre-trained natural language processing NLP model.
In some embodiments, the third determination submodule is specifically configured to:
based on a semantic index library, determining cosine similarity between the front semantic vector and the semantic vector of the target problem;
and sorting the problems in the semantic index library according to the cosine similarity.
In some embodiments, the NLP model includes any one of a BERT model, an ELMo model, and a GPT model.
In some embodiments, the pre-trained matching model is obtained by training an initial matching model according to training samples, the training samples including positive samples and negative samples; the positive sample is an original question in the second question-answer pair material and an original answer corresponding to the original question; the negative examples are the least similar answers among the third specified number of answers in the original question and the second question-and-answer corpus that are most similar to the original answers.
In some embodiments, the above apparatus further comprises:
the execution module is used for sequentially taking a third preset number of answers which are most similar to the text similarity and/or the semantic similarity of the original answer in the second question-answer pair corpus as current answers to execute the following steps: if the text similarity and/or the semantic similarity between the current answer and the original answer is smaller than a preset value, determining that the current answer is the most dissimilar answer to the original answer;
and the selecting module randomly selects one answer from the third preset number of most similar answers as the least similar answer to the original answer if the text similarity and/or the semantic similarity between the third preset number of most similar answers and the original answer are larger than the preset value.
In some embodiments, the first question-answer corpus is the same as the second question-answer corpus.
The retrieval type chat device provided by the embodiment of the application has the same technical characteristics as the retrieval type chat method provided by the embodiment, so that the same technical problems can be solved, and the same technical effects can be achieved.
As shown in fig. 5, an embodiment of the present application provides a computer device 500, including: a processor 501, a memory 502 and a bus, wherein the memory 502 stores machine-readable instructions executable by the processor 501, when the electronic device is operated, the processor 501 communicates with the memory 502 through the bus, and the processor 501 executes the machine-readable instructions to perform the steps of the retrieval chat method as described above.
Specifically, the memory 502 and the processor 501 can be general-purpose memory and processor, which are not limited in particular, and the retrieval chat method can be executed when the processor 501 runs a computer program stored in the memory 502.
Corresponding to the retrieval type chat method, the embodiment of the application also provides a computer readable storage medium, wherein a computer program is stored on the computer readable storage medium, and the computer program is executed by a processor to execute the steps of the retrieval type chat method.
The retrieval type chat device provided by the embodiment of the application can be specific hardware on the equipment or software or firmware installed on the equipment. The device provided by the embodiment of the present application has the same implementation principle and technical effect as the foregoing method embodiments, and for the sake of brief description, reference may be made to the corresponding contents in the foregoing method embodiments where no part of the device embodiments is mentioned. It is clear to those skilled in the art that, for convenience and brevity of description, the specific working processes of the foregoing systems, apparatuses and units may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. The above-described embodiments of the apparatus are merely illustrative, and for example, the division of the units is only one logical division, and there may be other divisions when actually implemented, and for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some communication interfaces, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments provided in the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit.
The functions, if implemented in the form of software functional units and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application or portions thereof that substantially contribute to the prior art may be embodied in the form of a software product stored in a storage medium and including instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the mobile control method according to the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus once an item is defined in one figure, it need not be further defined and explained in subsequent figures, and moreover, the terms "first", "second", "third", etc. are used merely to distinguish one description from another and are not to be construed as indicating or implying relative importance.
Finally, it should be noted that: the above-mentioned embodiments are only specific embodiments of the present application, and are used for illustrating the technical solutions of the present application, but not limiting the same, and the scope of the present application is not limited thereto, and although the present application is described in detail with reference to the foregoing embodiments, those skilled in the art should understand that: any person skilled in the art can modify or easily conceive the technical solutions described in the foregoing embodiments or equivalent substitutes for some technical features within the technical scope disclosed in the present application; such modifications, changes or substitutions do not depart from the scope of the embodiments of the present application. Are intended to be covered by the scope of the present application.

Claims (15)

1. An indexing type chatting method is characterized in that the method is applied to a chatting robot, a text index base and a semantic index base are prestored in the chatting robot, problem indexes in the text index base are text feature indexes, and problem indexes in the semantic index base are semantic vector indexes; the method comprises the following steps:
determining a first alternative answer set through text features of a target question to be answered based on the text index library;
determining a second alternative answer set through a semantic vector of the target question based on the semantic index library;
and determining at least one target answer with the highest matching degree with the target question as a reply of the target question according to the first candidate answer set and the second candidate answer set based on a pre-trained matching model.
2. The method of claim 1, wherein the text index library comprises a first query-response corpus, and wherein a question index in the text index library is used for indexing a question in the first query-response corpus; the step of determining a first alternative answer set by the text features of the target question to be answered based on the text index library comprises the following steps:
determining a first preset number of first questions with the maximum text similarity with the target question according to the text characteristics of the target question to be answered based on the text index library;
and determining a first alternative answer set according to the corresponding answer of each first question in the first question-answer pair corpus.
3. The method according to claim 2, wherein the step of determining a first preset number of first questions with the maximum text similarity to the target question according to the text features of the target question to be answered based on the text index library comprises:
based on the text index library, determining the text similarity of the current question and the target question according to one or more items of the number of co-occurring words appearing between the retrieved current question and the target question, the proportion of the co-occurring words in the current question and the target question respectively, and the importance degree of the co-occurring words in the first question-answer pair corpus;
determining a first preset number of first questions with the maximum text similarity with the target question in the questions of the first question-and-answer corpus;
the current question is any one of the questions retrieved from the text index library according to the text features of the target question, and the lower the frequency of the co-occurrence word appearing in the first question-answer pair corpus is, the higher the importance degree is.
4. The method according to claim 2, wherein the step of determining a first preset number of first questions with the maximum text similarity to the target question according to the text features of the target question to be answered based on the text index library comprises:
based on the text index library, determining the text similarity of the current question and the target question according to the retrieved Jacard distance between the two word sets corresponding to the current question and the target question;
determining a first preset number of first questions with the maximum text similarity with the target question in the questions of the first question-and-answer corpus;
and the current question is any one of the questions retrieved from the text index library according to the text characteristics of the target question.
5. The method according to claim 1, wherein the semantic index library comprises a first query-response corpus, and the question index in the semantic index library is used for indexing a question in the first query-response corpus; a step of determining a second candidate answer set by a semantic vector of the target question based on the semantic index library, comprising:
determining a second preset number of second questions with the maximum semantic similarity with the target question according to the semantic vector of the target question to be answered based on the semantic index library;
and determining a second alternative answer set according to the corresponding answer of each second question in the first question-answer pair corpus.
6. The method according to claim 5, wherein the step of determining a second preset number of second questions with the largest semantic similarity with the target question according to the semantic vector of the target question to be answered based on the semantic index library comprises:
determining a semantic vector of the target problem;
based on the semantic index library, sorting the problems in the semantic index library according to semantic similarity between the current semantic vector and the semantic vector of the target problem;
selecting a second preset number of second problems with the largest semantic similarity according to the sequence of the problems in the semantic index library;
and the current semantic vector is any one of the semantic vectors in the semantic index library.
7. The method of claim 6, wherein the semantic vector is extracted by a pre-trained Natural Language Processing (NLP) model.
8. The method according to claim 7, wherein the step of ordering the questions in the semantic index library according to semantic similarity between the current semantic vector and the semantic vector of the target question based on the semantic index library comprises:
based on the semantic index library, determining cosine similarity between a front semantic vector and a semantic vector of the target problem;
and sequencing the problems in the semantic index library according to the cosine similarity.
9. The method of claim 7, wherein the NLP model comprises any one of a BERT model, an ELMo model, and a GPT model.
10. The method of claim 7, wherein the pre-trained matching model is obtained by training an initial matching model according to training samples, wherein the training samples comprise positive samples and negative samples; the positive sample is an original question in a second question-answer pair material and an original answer corresponding to the original question; the negative sample is the least similar answer in a third specified number of answers most similar to the original answer in the original question and the second question-and-answer corpus.
11. The method of claim 10, further comprising: and sequentially taking a third preset number of answers which are most similar to the text similarity and/or the semantic similarity of the original answer in the second question-answer corpus as current answers to execute the following steps: if the text similarity and/or the semantic similarity between the current answer and the original answer is smaller than a preset value, determining that the current answer is the most dissimilar answer to the original answer;
and if the text similarity and/or the semantic similarity between the third preset number of most similar answers and the original answer are larger than preset values, randomly selecting one answer from the third preset number of most similar answers as the least similar answer to the original answer.
12. The method of claim 10, wherein said first question-answer corpus is identical to said second question-answer corpus.
13. An indexing type chatting device is applied to a chatting robot, a text index base and a semantic index base are prestored in the chatting robot, problem indexes in the text index base are text feature indexes, and problem indexes in the semantic index base are semantic vector indexes; the device comprises:
the first determination module is used for determining a first alternative answer set through text characteristics of a target question to be answered based on the text index library;
a second determining module, configured to determine, based on the semantic index library, a second candidate answer set through a semantic vector of the target question;
and the third determining module is used for determining at least one target answer with the highest matching degree with the target question as the reply of the target question according to the first candidate answer set and the second candidate answer set based on a pre-trained matching model.
14. A computer device comprising a memory, a processor, and a computer program stored in the memory and executable on the processor, wherein the processor implements the steps of the method of any of claims 1 to 12 when executing the computer program.
15. A computer readable storage medium having stored thereon machine executable instructions which, when invoked and executed by a processor, cause the processor to perform the method of any of claims 1 to 12.
CN201910985485.7A 2019-10-16 2019-10-16 Retrieval type chatting method and device and computer equipment Active CN110750616B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910985485.7A CN110750616B (en) 2019-10-16 2019-10-16 Retrieval type chatting method and device and computer equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910985485.7A CN110750616B (en) 2019-10-16 2019-10-16 Retrieval type chatting method and device and computer equipment

Publications (2)

Publication Number Publication Date
CN110750616A true CN110750616A (en) 2020-02-04
CN110750616B CN110750616B (en) 2023-02-03

Family

ID=69278631

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910985485.7A Active CN110750616B (en) 2019-10-16 2019-10-16 Retrieval type chatting method and device and computer equipment

Country Status (1)

Country Link
CN (1) CN110750616B (en)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111414746A (en) * 2020-04-10 2020-07-14 中国建设银行股份有限公司 Matching statement determination method, device, equipment and storage medium
CN111523304A (en) * 2020-04-27 2020-08-11 华东师范大学 Automatic generation method of product description text based on pre-training model
CN111581354A (en) * 2020-05-12 2020-08-25 金蝶软件(中国)有限公司 FAQ question similarity calculation method and system
CN111680158A (en) * 2020-06-10 2020-09-18 创新奇智(青岛)科技有限公司 Short text classification method, device, equipment and storage medium in open field
CN111797216A (en) * 2020-06-28 2020-10-20 北京百度网讯科技有限公司 Retrieval item rewriting method, device, equipment and storage medium
CN111859987A (en) * 2020-07-28 2020-10-30 网易(杭州)网络有限公司 Text processing method, and training method and device of target task model
CN111881695A (en) * 2020-06-12 2020-11-03 国家电网有限公司 Audit knowledge retrieval method and device
CN111897943A (en) * 2020-08-17 2020-11-06 腾讯科技(深圳)有限公司 Session record searching method and device, electronic equipment and storage medium
CN111930880A (en) * 2020-08-14 2020-11-13 易联众信息技术股份有限公司 Text code retrieval method, device and medium
CN112182159A (en) * 2020-09-30 2021-01-05 中国人民大学 Personalized retrieval type conversation method and system based on semantic representation
CN112241626A (en) * 2020-10-14 2021-01-19 网易(杭州)网络有限公司 Semantic matching and semantic similarity model training method and device
CN113139043A (en) * 2021-04-29 2021-07-20 北京百度网讯科技有限公司 Question and answer sample generation method and device, electronic equipment and storage medium
CN113159187A (en) * 2021-04-23 2021-07-23 北京金山数字娱乐科技有限公司 Classification model training method and device, and target text determining method and device
CN113326420A (en) * 2021-06-15 2021-08-31 北京百度网讯科技有限公司 Question retrieval method, device, electronic equipment and medium
CN114090747A (en) * 2021-10-14 2022-02-25 特斯联科技集团有限公司 Automatic question answering method, device, equipment and medium based on multiple semantic matching
CN116414958A (en) * 2023-02-06 2023-07-11 飞算数智科技(深圳)有限公司 Text corpus generation method and device, storage medium and electronic equipment
US12079215B2 (en) 2020-06-28 2024-09-03 Beijing Baidu Netcom Science And Technology Co., Ltd. Query rewriting method and apparatus, device and storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108536708A (en) * 2017-03-03 2018-09-14 腾讯科技(深圳)有限公司 A kind of automatic question answering processing method and automatically request-answering system
CN109271505A (en) * 2018-11-12 2019-01-25 深圳智能思创科技有限公司 A kind of question answering system implementation method based on problem answers pair
CN109684442A (en) * 2018-12-21 2019-04-26 东软集团股份有限公司 A kind of text searching method, device, equipment and program product
CN109740077A (en) * 2018-12-29 2019-05-10 北京百度网讯科技有限公司 Answer searching method, device and its relevant device based on semantic indexing
CN110019695A (en) * 2017-08-07 2019-07-16 芋头科技(杭州)有限公司 A kind of automatic chatting response method
CN110019644A (en) * 2017-09-06 2019-07-16 腾讯科技(深圳)有限公司 Searching method, device and computer readable storage medium in dialogue realization
US20190260694A1 (en) * 2018-02-16 2019-08-22 Mz Ip Holdings, Llc System and method for chat community question answering
CN110287494A (en) * 2019-07-01 2019-09-27 济南浪潮高新科技投资发展有限公司 A method of the short text Similarity matching based on deep learning BERT algorithm
CN110321925A (en) * 2019-05-24 2019-10-11 中国工程物理研究院计算机应用研究所 A kind of more granularity similarity comparison methods of text based on semantics fusion fingerprint

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108536708A (en) * 2017-03-03 2018-09-14 腾讯科技(深圳)有限公司 A kind of automatic question answering processing method and automatically request-answering system
CN110019695A (en) * 2017-08-07 2019-07-16 芋头科技(杭州)有限公司 A kind of automatic chatting response method
CN110019644A (en) * 2017-09-06 2019-07-16 腾讯科技(深圳)有限公司 Searching method, device and computer readable storage medium in dialogue realization
US20190260694A1 (en) * 2018-02-16 2019-08-22 Mz Ip Holdings, Llc System and method for chat community question answering
CN109271505A (en) * 2018-11-12 2019-01-25 深圳智能思创科技有限公司 A kind of question answering system implementation method based on problem answers pair
CN109684442A (en) * 2018-12-21 2019-04-26 东软集团股份有限公司 A kind of text searching method, device, equipment and program product
CN109740077A (en) * 2018-12-29 2019-05-10 北京百度网讯科技有限公司 Answer searching method, device and its relevant device based on semantic indexing
CN110321925A (en) * 2019-05-24 2019-10-11 中国工程物理研究院计算机应用研究所 A kind of more granularity similarity comparison methods of text based on semantics fusion fingerprint
CN110287494A (en) * 2019-07-01 2019-09-27 济南浪潮高新科技投资发展有限公司 A method of the short text Similarity matching based on deep learning BERT algorithm

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
谢靖,等: "科技文献检索系统语义丰富化框架的设计与实践", 《中国科学院文献情报中心》 *

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111414746A (en) * 2020-04-10 2020-07-14 中国建设银行股份有限公司 Matching statement determination method, device, equipment and storage medium
CN111414746B (en) * 2020-04-10 2023-11-07 建信金融科技有限责任公司 Method, device, equipment and storage medium for determining matching statement
CN111523304A (en) * 2020-04-27 2020-08-11 华东师范大学 Automatic generation method of product description text based on pre-training model
CN111523304B (en) * 2020-04-27 2022-08-02 华东师范大学 Automatic generation method of product description text based on pre-training model
CN111581354A (en) * 2020-05-12 2020-08-25 金蝶软件(中国)有限公司 FAQ question similarity calculation method and system
CN111680158A (en) * 2020-06-10 2020-09-18 创新奇智(青岛)科技有限公司 Short text classification method, device, equipment and storage medium in open field
CN111881695A (en) * 2020-06-12 2020-11-03 国家电网有限公司 Audit knowledge retrieval method and device
CN111797216A (en) * 2020-06-28 2020-10-20 北京百度网讯科技有限公司 Retrieval item rewriting method, device, equipment and storage medium
US12079215B2 (en) 2020-06-28 2024-09-03 Beijing Baidu Netcom Science And Technology Co., Ltd. Query rewriting method and apparatus, device and storage medium
CN111797216B (en) * 2020-06-28 2024-04-05 北京百度网讯科技有限公司 Search term rewriting method, apparatus, device and storage medium
CN111859987B (en) * 2020-07-28 2024-05-17 网易(杭州)网络有限公司 Text processing method, training method and device for target task model
CN111859987A (en) * 2020-07-28 2020-10-30 网易(杭州)网络有限公司 Text processing method, and training method and device of target task model
CN111930880A (en) * 2020-08-14 2020-11-13 易联众信息技术股份有限公司 Text code retrieval method, device and medium
CN111897943A (en) * 2020-08-17 2020-11-06 腾讯科技(深圳)有限公司 Session record searching method and device, electronic equipment and storage medium
CN112182159A (en) * 2020-09-30 2021-01-05 中国人民大学 Personalized retrieval type conversation method and system based on semantic representation
CN112182159B (en) * 2020-09-30 2023-07-07 中国人民大学 Personalized search type dialogue method and system based on semantic representation
CN112241626A (en) * 2020-10-14 2021-01-19 网易(杭州)网络有限公司 Semantic matching and semantic similarity model training method and device
CN112241626B (en) * 2020-10-14 2023-07-07 网易(杭州)网络有限公司 Semantic matching and semantic similarity model training method and device
CN113159187A (en) * 2021-04-23 2021-07-23 北京金山数字娱乐科技有限公司 Classification model training method and device, and target text determining method and device
CN113139043B (en) * 2021-04-29 2023-08-04 北京百度网讯科技有限公司 Question-answer sample generation method and device, electronic equipment and storage medium
CN113139043A (en) * 2021-04-29 2021-07-20 北京百度网讯科技有限公司 Question and answer sample generation method and device, electronic equipment and storage medium
CN113326420B (en) * 2021-06-15 2023-10-27 北京百度网讯科技有限公司 Question retrieval method, device, electronic equipment and medium
US11977567B2 (en) 2021-06-15 2024-05-07 Beijing Baidu Netcom Science Technology Co., Ltd. Method of retrieving query, electronic device and medium
CN113326420A (en) * 2021-06-15 2021-08-31 北京百度网讯科技有限公司 Question retrieval method, device, electronic equipment and medium
CN114090747A (en) * 2021-10-14 2022-02-25 特斯联科技集团有限公司 Automatic question answering method, device, equipment and medium based on multiple semantic matching
CN116414958A (en) * 2023-02-06 2023-07-11 飞算数智科技(深圳)有限公司 Text corpus generation method and device, storage medium and electronic equipment

Also Published As

Publication number Publication date
CN110750616B (en) 2023-02-03

Similar Documents

Publication Publication Date Title
CN110750616B (en) Retrieval type chatting method and device and computer equipment
Zhao et al. Automatic detection of cyberbullying on social networks based on bullying features
CN108536852B (en) Question-answer interaction method and device, computer equipment and computer readable storage medium
CN106776532B (en) Knowledge question-answering method and device
WO2013125286A1 (en) Non-factoid question answering system and computer program
Kiyoumarsi Evaluation of automatic text summarizations based on human summaries
US10671619B2 (en) Information processing system and information processing method
WO2021212801A1 (en) Evaluation object identification method and apparatus for e-commerce product, and storage medium
KR20160026892A (en) Non-factoid question-and-answer system and method
CN111859940B (en) Keyword extraction method and device, electronic equipment and storage medium
CN111209372B (en) Keyword determination method and device, electronic equipment and storage medium
CN111694941B (en) Reply information determining method and device, storage medium and electronic equipment
CN113569011A (en) Training method, device and equipment of text matching model and storage medium
KR101319413B1 (en) Summary Information Generating System and Method for Review of Product and Service
Houjeij et al. A novel approach for emotion classification based on fusion of text and speech
CN116541493A (en) Interactive response method, device, equipment and storage medium based on intention recognition
JPWO2016178337A1 (en) Information processing apparatus, information processing method, and computer program
JP5718405B2 (en) Utterance selection apparatus, method and program, dialogue apparatus and method
CN115238039A (en) Text generation method, electronic device and computer-readable storage medium
CN111274366A (en) Search recommendation method and device, equipment and storage medium
US20210406291A1 (en) Dialog driven search system and method
CN111401070B (en) Word meaning similarity determining method and device, electronic equipment and storage medium
CN113761125A (en) Dynamic summary determination method and device, computing equipment and computer storage medium
Magnolini A Survey on Paraphrase Recognition.
CN117725153B (en) Text matching method, device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant