CN113987155B - Conversational retrieval method integrating knowledge graph and large-scale user log - Google Patents

Conversational retrieval method integrating knowledge graph and large-scale user log Download PDF

Info

Publication number
CN113987155B
CN113987155B CN202111412012.1A CN202111412012A CN113987155B CN 113987155 B CN113987155 B CN 113987155B CN 202111412012 A CN202111412012 A CN 202111412012A CN 113987155 B CN113987155 B CN 113987155B
Authority
CN
China
Prior art keywords
entity
representation
query
document
representations
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111412012.1A
Other languages
Chinese (zh)
Other versions
CN113987155A (en
Inventor
窦志成
文继荣
左笑晨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Renmin University of China
Original Assignee
Renmin University of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Renmin University of China filed Critical Renmin University of China
Priority to CN202111412012.1A priority Critical patent/CN113987155B/en
Publication of CN113987155A publication Critical patent/CN113987155A/en
Application granted granted Critical
Publication of CN113987155B publication Critical patent/CN113987155B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Abstract

The invention realizes a conversational retrieval method integrating a knowledge graph and a large-scale user log by a method in the artificial intelligence field. First, by learning the representation of the entity in the current session on a fused entity graph, the context information corresponding to the entity is taken into account in the learning of the representation. The method then uses the learned entity representation and the original entity representation and word representation to effect a correlation calculation of the session context with the candidate document. And finally, sorting the candidate documents through the relevance scores. The method provided by the invention fuses the knowledge graph with the entity graph extracted from the query log, and calculates the fused entity representation through the graph attention neural network to fuse additional knowledge. The computed fusion entity representation is used for the matching process of the session context with the candidate document. Compared with the prior conversational search model, the method and the device try to introduce knowledge beyond the current conversation for the first time to improve the document search effect, and further improve the use experience of search engine users.

Description

Conversational retrieval method integrating knowledge graph and large-scale user log
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to a conversational retrieval method integrating a knowledge graph and a large-scale user log.
Background
There are many session-based retrieval models. Some models learn about user intent by modeling remodelling relationships between multiple queries within a session. For example, the user has deleted certain terms compared to previous queries, indicating that these terms do not match the user's current query intent. Or the user has newly added words that are likely to more satisfy the user's current query intent. With the widespread use of depth models, many models began to attempt to obtain a representation of user intent through semantic modeling, and use this representation to measure the relevance of candidate documents to user intent. The latest session-based retrieval model orders two tasks through comprehensive query remodeling and documents, so that the two tasks are promoted to each other, and the effect of document retrieval is improved. In addition, the model [7] with the best effect at present improves the retrieval effect by modeling different behaviors of the user in the query process. These user actions include clicking and skipping certain documents.
The related effort of previous conversational searches uses only the information inside the conversation, i.e., the user's historical queries and historical click documents. However, such information is often limited, and it is sometimes insufficient to understand the intent of the user. For example, when a user mentions some ambiguous entities, the search engine is likely to misunderstand the user's query intent without the assistance of a knowledge-graph. In addition, co-occurrence relationships among many entities are implied in the large-scale user log, and are built by other users through their query behavior during the search process, which also plays an important role in understanding the user's intent of the current session, which was not considered in the past.
Disclosure of Invention
Therefore, the invention firstly provides a conversational retrieval method integrating a knowledge graph and a large-scale user log, firstly, the representation of the entity in the current conversation is learned on a fused entity graph, the context information corresponding to the entity is considered in the process of learning the representation of the entity in the current conversation,
specifically: for a given session context S, it contains a historical query and corresponding historical click documentWherein q is i Representing the i-th query in the session, +.>User click document set corresponding to the ith query>Wherein->Indicating the 1 st clicked document corresponding to the i-th query, n indicating the number of clicked documents, and the current query being q t The candidate document is d t First, word representations in each document and query are obtained, original entity representations and fusion entity representations, for all words [ w ] in a certain query 1 ,…,w |q| ]Obtaining the embedding of each word through Fasttext: /> Using these embeddings as input, contextual semantic information of words is fused into the representation of words by bi-directional LSTM, specifically let +.>And +.>Hidden representations of the i-th word obtained for forward and backward LSTM respectively, and concatenating them together results in a representation of each word:
the learned entity representation is then used with the original entity representation and the word representation to effect a correlation calculation of the session context with the candidate document,
specifically: for the original entity representation, acquiring an initial embedded representation of each entity through a TransE, and embedding the entity into the entity through a fully connected layer:
acquiring new entity representation by means of a fused entity representation moduleThe fusion entity representation module obtains the representation of the entity, and calculates a ranking feature for the s-th historical query and the click document corresponding to the s-th historical query through a single-round interaction module>The ranking features of each historical query are then integrated together into an overall contextual ranking feature>Wherein->Representing the importance of the s-th query, and calculating by designing an attention mechanism capable of perceiving the semantics of the query:
wherein the method comprises the steps ofAnd->Respectively represent the current query q t And history query q s And this representation is spliced by three representations: />Wherein->And->The word representation, the original entity representation and the fused entity representation are summed up, respectively:
the resulting overall context ranking feature f context Obtaining a context matching score S through a full connectivity layer c =MLP(f context ) In addition, for the ranking features corresponding to the current queryLikewise, a current query matching score is obtained via a full connectivity layer>Taking the average tf-idf values of words and entities into consideration, the number of public words in candidate documents and all queries, embedding the calculated similarity values between the candidate documents and the words and entities of historical queries as statistical features, and obtaining another matching score S through a full connection layer by using the statistical features f The three scores are spliced together to obtain the final matching score R (d) of the candidate document through a full connection layer t |q t ,S);
Finally, ranking the candidate documents by taking the matching score output by the model as a basis;
the model training process adopts a model way, and the design of the loss function is as follows:
wherein R (d|q) is shorthand for R (d|q, S). S is S T Including all of the sessions in the training set,containing all clicked and un-clicked documents corresponding to query q, where d + Representing clicked document d - Representing an untclicked document. All model parameters are updated by a back propagation algorithm.
The realization mode of the fusion entity representation module is as follows: firstly, constructing a fusion entity diagram comprising a knowledge graph, entities in a query log and relations among the entities, and specifically, nodes in the fusion entity diagram are formed by two parts: querying entities in the log and entities in the knowledge graph;
the entities in the query log refer to the entities contained in the query and click documents in each session in the query log, and on the basis of the entities in the query log, triples are screened in the knowledge graph, wherein the screening condition is that the head entity or the tail entity of the triples appears in the entity set of the query log, and the edges between the entities are divided into three categories: two entities appear in the same triplet in the knowledge graph, two entities appear in two queries before and after the query log, and two entities respectively appear in one query in the query day and the click document corresponding to the two entities;
the representation of the node of a given entity, in particular of a given entity e, is then calculated by means of a graph attention neural network i The representation of the corresponding node is calculated by a multi-layer graph attention neural network and the output of the last layer of network is used as the final entity representation, each layer of representation being calculated in the following manner:
wherein the method comprises the steps ofRepresenting a layer k neural network calculationOutput of N (e) i ) Representing entity e i Neighbor node set,/->Representing entity e i The weight of the j-th neighbor of (2), which weight needs to be based on this neighbor and e i The context association degree is calculated by the following calculation modes:
wherein R is e Is entity e i A representation of the query or document in which it resides, which is obtained by summing the original representations of all entities within the query or document.
The single-round interactive module designs three kinds of representations: word representation, original entity representation, new fusion entity representation, and nine interaction matrices obtained by calculation:
[M ww ,M we ,M wi ,M ew ,M ee ,M ei ,M iw ,M ie ,M ii ]
the elements of these matrices correspond to interaction scores of two representations, where w represents a word, e represents an entity, and i is a new fused entity representation. The nine interaction matrices are calculated as follows:
wherein W is M Is used to map the fused entity representation with the word or original entity representation into the same semantic space, and then, at each matchObtaining k-dimensional matching features on the matching matrix by using a kernel pooling method in a KNRM model
Wherein mu k Sum sigma k Is a parameter of kernel pooling for measuring different similarity distribution, and then obtaining an interaction feature by weighting and summing the matching features obtained by 9 matrix calculation
The single-round interaction module is designed with a matching method based on representation, namely, for single documents in a click document set corresponding to each query, word representations are used, three document representations are calculated by an original entity representation and a fusion entity representation, the calculation mode is that the representations of each word or entity are added together, and then the representations of each document in the click document set are averaged to obtain the representations of the three click document sets:three representations corresponding to the candidate document are obtained by adding the representations of each word or entity in the candidate document: />For each representation, matching features of the set of click documents and the candidate documents are calculated by:
can be obtained in the same wayAnd->The three matching features and the interaction feature obtained before +.>Spliced together, a single round interactive ordering feature is obtained>
The invention has the technical effects that:
additional knowledge is fused by fusing the knowledge graph with the entity graph extracted from the query log and computing a fused entity representation through the graph attention neural network. The computed fusion entity representation is used for the matching process of the session context with the candidate document. Compared with the prior conversational search model, the method and the device try to introduce knowledge beyond the current conversation for the first time to improve the document search effect, and further improve the use experience of search engine users.
Drawings
FIG. 1 is a diagram of the overall structure of the model;
FIG. 2 is a schematic diagram of a single-wheel interactive module structure
Detailed Description
The following is a preferred embodiment of the present invention and a technical solution of the present invention is further described with reference to the accompanying drawings, but the present invention is not limited to this embodiment.
The method for improving the conversational search effect by introducing the external knowledge is tried for the first time, and the applied external knowledge comprises a knowledge graph based on the real relationship and a large-scale user log recorded by a search engine. . Firstly, constructing an entity diagram by utilizing a knowledge graph and a user log, and learning the representation of the entity in the current session on the entity diagram, wherein the context information corresponding to the entity is considered in the process of learning the representation. The method then uses the learned entity representation and the original entity representation and word representation to effect a correlation calculation of the session context with the candidate document. And finally, sorting the candidate documents through the relevance scores.
In the conversational search problem, given the context S of a conversation, a historical query and a corresponding historical click document are includedWherein q is i Representing the i-th query in the session, +.>User click document set corresponding to the ith query>Wherein->The 1 st click document corresponding to the i-th query is represented, and n represents the number of the clicked documents. In addition, given the current query q t And candidate document d t We calculate the relevance score R (d t |q t ,S)。
The overall structure of the model is shown in fig. 1. First we obtain word representations, original entity representations, and fused entity representations in each document and query. For all words in a query [ w ] 1 ,…,w |q| ]First, the embedding of each word is obtained through Fasttext:using these embeddings as input, the contextual semantic information of the word is fused into the representation of the word through the bi-directional LSTM. Specifically, let->And +.>Hidden representations of the i-th word obtained by forward and backward LSTM, respectively. Stitching them together results in a representation of each word: />
For the original entity representation, an initial embedded representation of each entity is first obtained by a TransE. Since the entities in each query or document do not form a continuous sequence, the use of LS TM to obtain semantic representations of the entities is not considered. However, in order to obtain an entity representation in the same semantic space as the word representation, it is necessary to embed the entity additionally through a fully connected layer:
in addition, in order to introduce additional information in the knowledge graph and the query log, the invention obtains a new entity representation through a fusion entity representation module (IER)The IER module obtains representations of entities by using the graph ideographic neural network on a map of fused entities, as will be described in detail in the following section.
After the word representation, the original entity representation and the fused entity representation are obtained, the candidate documents may be computed as relevance scores for each of the historical queries and their corresponding click documents, respectively. In particular, for the s-th historical query and its corresponding click document, a ranking feature may be computed by a Single-turn interaction module (Single-turn interaction)The ranking features of each historical query are then integrated together into an overall contextual ranking featureWherein->Importance of representing the s-th queryThe degree is calculated by a query-level attention mechanism:
wherein the method comprises the steps ofAnd->Respectively represent the current query q t And history query q s And this representation is spliced by three representations: />Wherein->And->The word representation, the original entity representation and the fused entity representation are summed up, respectively:
the resulting overall context ranking feature f context Obtaining a context matching score S through a full connectivity layer c =MLP(f context ). In addition, for the ranking features corresponding to the current queryLikewise, a current query matching score is obtained via a full connectivity layer>In addition, consider some statistical features, including the average tf-idf values of words and entities, candidate documentsThe number of public words in the files and all queries, and the calculated similarity value is embedded in the words and entities of the candidate documents and the historical queries. These features are passed through a full connection layer to obtain another matching score S f The three scores are spliced together to obtain the final matching score R (d) of the candidate document through a full connection layer t |q t ,S)。
Fusion entity representation module (IER):
in order to introduce additional information contained in the knowledge graph and the query log into the model, the invention designs a fusion entity representation module. The module first needs to build a fusion entity graph. The fused entity graph includes relationships between knowledge maps and entities in the query log. Specifically, the nodes in the fusion entity graph are composed of two parts: the entities in the query log and the entities in the knowledge graph. First, the entities in the query log refer to the queries within each session in the query log and the entities contained within the click document. And screening the triples in the knowledge graph on the basis of the entities in the query log. The condition for filtering is that the head entity or the tail entity of the triplet appears in the entity set of the query log. After all entities in the graph have been determined, edges between the entities also need to be determined. Edges between entities fall into three categories: 1) Two entities appear in the same triplet in the knowledge graph; 2) Two entities appear in both the front and back queries of the query log; 3) The two entities appear in one query and its corresponding click document on the query day, respectively. These three classes of edges describe objective associations between entities and two different co-occurrence associations, respectively. It should be noted that there are no direct edges between entities that appear in the same query or document, but they all have common neighbors, thus equating to indirectly existing associations. This design is to avoid redundancy of edges in the figure.
After the fusion graph is constructed, the present invention calculates a representation of a given entity node through the graph attention neural network. Specifically, given an entity e i Computing a representation of a corresponding node through a multi-layer graph attention neural network and using the output of the last layer network as the finalIs an entity representation of (c). Each layer represents the way in which the computation is as follows:
wherein the method comprises the steps ofRepresenting the output of the calculation of the layer k neural network, N (e i ) Representing entity e i Is described herein). />Representing entity e i The weight of the j-th neighbor of (2), which weight needs to be based on this neighbor and e i The context association degree is calculated by the following calculation modes:
wherein R is e Is entity e i A representation of the query or document in which it resides, which is obtained by summing the original representations of all entities within the query or document.
Single-wheel interaction module (Single-turn interaction)
The block diagram of the single-wheel interactive module is shown in fig. 2. The structure is inspired by the EDRM model, which uses four interaction matrices: query word-document word, query entity-document entity, query entity-document word, and query word-document entity. Three representations are designed in the present invention: word representation, original entity representation, new fusion entity representation, thus nine interaction matrices can be calculated:
[M ww ,M we ,M wi ,M ew ,M ee ,M ei ,M iw ,M ie ,M ii ]
the elements of these matrices correspond to interaction scores of the two representations, M we And M wi The following are examples:
wherein W is M Is used to map the fused entity representation into the same semantic space as the word or original entity representation. Subsequently, a kernel pooling method as mentioned in the KNRM model is used on each matching matrix to obtain k-dimensional matching features
Wherein mu k Sum sigma k Is a parameter of kernel pooling for measuring different similarity distributions. The matching features obtained by the calculation of the 9 matrixes are weighted and summed to obtain an interaction feature
In addition, the module designs a representation-based matching method in order to consider the role of clicking on documents corresponding to each historical query. For a single document in the set of click documents corresponding to each query, three document representations are computed using the word representations, the original entity representation and the fused entity representation, in such a way that the representations of each word or entity are added together. The representation of each document within the set of click documents is then averaged to obtain three representations of the set of click documents:similarly, three representations corresponding to the candidate document are obtained by summing the representations of each word or entity in the candidate document: />For each representation, matching features of the set of click documents and the candidate documents are calculated by:
can be obtained in the same wayAnd->The three matching features and the interaction feature obtained before +.>Spliced together, a single round interactive ordering feature is obtained>
Model training
The model training process adopts a model way, and the design of the loss function is as follows:
wherein R (d|q) is shorthand for R (d|q, S). S is S T Including all of the sessions in the training set,containing all clicked and un-clicked documents corresponding to query q, where d + Representing clicked document d - Representing an untclicked document. All model parameters are updated by a back propagation algorithm.

Claims (3)

1. A conversational search method integrating knowledge graph and large-scale user log, which improves conversational search effect by introducing external knowledge, wherein the external knowledge comprises knowledge graph based on reality relationship and large-scale user log recorded by search engine, and is characterized in that: firstly, constructing an entity graph by utilizing a knowledge graph and a user log, and learning the representation of the entity in the current session on the entity graph, so that the context information corresponding to the entity can be considered in the process of learning the representation of the entity,
specifically: for a given session context S, it contains a historical query and corresponding historical click documentWherein q is i Representing the i-th query in the session, +.>User click document set corresponding to the ith query>Wherein->Indicating the 1 st clicked document corresponding to the i-th query, n indicating the number of clicked documents, and the current query being q t The candidate document is d t First, word representations in each document and query are obtained, original entity representations and fusion entity representations, for all words [ w ] in a certain query 1 ,…,w |q| ]Obtaining the embedding of each word through Fasttext: /> Using these embeddings as input, contextual semantic information of words is fused into the representation of words by bi-directional LSTM, specifically let +.>And +.>Hidden representations of the i-th word obtained for forward and backward LSTM respectively, and concatenating them together results in a representation of each word:
the learned entity representation is then used with the original entity representation and the word representation to effect a correlation calculation of the session context with the candidate document,
specifically: for the original entity representation, acquiring an initial embedded representation of each entity through a TransE, and embedding the entity into the entity through a fully connected layer:
acquiring new entity representation by means of a fused entity representation moduleThe fusion entity representation module obtains the representation of the entity, and calculates a ranking feature for the s-th historical query and the click document corresponding to the s-th historical query through a single-round interaction module>The ranking features of each historical query are then integrated together into an overall contextual ranking feature>Wherein->Representing the importance of the s-th query, and calculating by designing an attention mechanism capable of perceiving the semantics of the query:
wherein the method comprises the steps ofAnd->Respectively represent the current query q t And history query q s And this representation is spliced by three representations: />Wherein->And->The word representation, the original entity representation and the fused entity representation are summed up, respectively:
the resulting overall context ranking feature f context Obtaining a context matching score S through a full connectivity layer c =MLP(f context ) In addition, for the ranking features corresponding to the current queryLikewise, a current query matching score is obtained via a full connectivity layer>Considering the average tf-idf values of words and entities, candidate documents and the number of common words in all queries, and candidate documents and historical queriesEmbedding the counted similarity value between the words and the entities as statistical features, and obtaining another matching score S by using the statistical features through a full connection layer f The three scores are spliced together to obtain the final matching score R (d) of the candidate document through a full connection layer t |q t ,S);
Finally, ranking the candidate documents by taking the matching score output by the model as a basis;
the model training process adopts a model way, and the design of the loss function is as follows:
wherein R (d|q) is shorthand for R (d|q, S) T Including all of the sessions in the training set,containing all clicked and un-clicked documents corresponding to query q, where d + Representing clicked document d - Representing an uncapped document, all model parameters are updated by a back propagation algorithm.
2. The conversational search method of claim 1, wherein the knowledge graph and the large-scale user logs are integrated, wherein the conversational search method is characterized by: the realization mode of the fusion entity representation module is as follows: firstly, constructing a fusion entity diagram comprising a knowledge graph, entities in a query log and relations among the entities, and specifically, nodes in the fusion entity diagram are formed by two parts: querying entities in the log and entities in the knowledge graph;
the entities in the query log refer to the entities contained in the query and click documents in each session in the query log, and on the basis of the entities in the query log, triples are screened in the knowledge graph, wherein the screening condition is that the head entity or the tail entity of the triples appears in the entity set of the query log, and the edges between the entities are divided into three categories: two entities appear in the same triplet in the knowledge graph, two entities appear in two queries before and after the query log, and two entities respectively appear in one query in the query day and the click document corresponding to the two entities;
the representation of the node of a given entity, in particular of a given entity e, is then calculated by means of a graph attention neural network i The representation of the corresponding node is calculated by a multi-layer graph attention neural network and the output of the last layer of network is used as the final entity representation, each layer of representation being calculated in the following manner:
wherein the method comprises the steps ofRepresenting the output of the calculation of the layer k neural network, N (e i ) Representing entity e i Neighbor node set,/->Representing entity e i The weight of the j-th neighbor of (2), which weight needs to be based on this neighbor and e i The context association degree is calculated by the following calculation modes:
wherein R is e Is entity e i A representation of the query or document in which it resides, which is obtained by summing the original representations of all entities within the query or document.
3. The conversational search method of claim 2, wherein the knowledge graph and the large-scale user logs are integrated, wherein the conversational search method is characterized by: the single-round interactive module designs three kinds of representations: word representation, original entity representation, new fusion entity representation, and nine interaction matrices obtained by calculation:
[M ww ,M we ,M wi ,M ew ,M ee ,M ei ,M iw ,M ie ,M ii ]
the elements of these matrices correspond to interaction scores of two representations, where w represents a word, e represents an entity, i is a new fused entity representation, and the nine interaction matrices are calculated as follows:
wherein W is M Is used for mapping the fusion entity representation and the word or the original entity representation into the same semantic space, and then k-dimensional matching features are obtained on each matching matrix by using a kernel pooling method in a KNRM model
Wherein mu k Sum sigma k Is a parameter of kernel pooling for measuring different similarity distribution, and then obtaining an interaction feature by weighting and summing the matching features obtained by 9 matrix calculation
The single-wheel interactive module is designed with a baseIn the matching method of the representations, namely, for single documents in the click document set corresponding to each query, three document representations are calculated by using word representations, original entity representations and fusion entity representations, wherein the calculation mode is to add the representations of each word or entity together, and then the representations of each document in the click document set are averaged to obtain the representations of the three click document sets:three representations corresponding to the candidate document are obtained by adding the representations of each word or entity in the candidate document: />For each representation, matching features of the set of click documents and the candidate documents are calculated by:
can be obtained in the same wayAnd->The three matching features and the interaction feature obtained before +.>Spliced together, a single round interactive ordering feature is obtained>
CN202111412012.1A 2021-11-25 2021-11-25 Conversational retrieval method integrating knowledge graph and large-scale user log Active CN113987155B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111412012.1A CN113987155B (en) 2021-11-25 2021-11-25 Conversational retrieval method integrating knowledge graph and large-scale user log

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111412012.1A CN113987155B (en) 2021-11-25 2021-11-25 Conversational retrieval method integrating knowledge graph and large-scale user log

Publications (2)

Publication Number Publication Date
CN113987155A CN113987155A (en) 2022-01-28
CN113987155B true CN113987155B (en) 2024-03-26

Family

ID=79731813

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111412012.1A Active CN113987155B (en) 2021-11-25 2021-11-25 Conversational retrieval method integrating knowledge graph and large-scale user log

Country Status (1)

Country Link
CN (1) CN113987155B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114741627B (en) * 2022-04-12 2023-03-24 中国人民解放军32802部队 Internet-oriented auxiliary information searching method
CN116304111B (en) * 2023-04-10 2024-02-20 深圳市兴海物联科技有限公司 AI call optimization processing method and server based on visual service data
CN116775848B (en) * 2023-08-23 2023-11-07 宁波吉利汽车研究开发有限公司 Control method, device, computing equipment and storage medium for generating dialogue information

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017041372A1 (en) * 2015-09-07 2017-03-16 百度在线网络技术(北京)有限公司 Man-machine interaction method and system based on artificial intelligence
CN111177357A (en) * 2019-12-31 2020-05-19 中国人民大学 Memory neural network-based conversational information retrieval method
CN112069399A (en) * 2020-08-25 2020-12-11 中国人民大学 Personalized search system based on interactive matching
CN112732944A (en) * 2021-01-30 2021-04-30 吉林大学 New method for text retrieval

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10803055B2 (en) * 2017-12-15 2020-10-13 Accenture Global Solutions Limited Cognitive searches based on deep-learning neural networks

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017041372A1 (en) * 2015-09-07 2017-03-16 百度在线网络技术(北京)有限公司 Man-machine interaction method and system based on artificial intelligence
CN111177357A (en) * 2019-12-31 2020-05-19 中国人民大学 Memory neural network-based conversational information retrieval method
CN112069399A (en) * 2020-08-25 2020-12-11 中国人民大学 Personalized search system based on interactive matching
CN112732944A (en) * 2021-01-30 2021-04-30 吉林大学 New method for text retrieval

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
一种面向多源知识图谱的样例查询方法;汤楠;申德荣;寇月;聂铁铮;;计算机研究与发展;20151015(S1);全文 *
马飞翔 ; 廖祥文 ; 於志勇 ; 吴运兵 ; 陈国龙 ; .基于知识图谱的文本观点检索方法.山东大学学报(理学版).(11),全文. *

Also Published As

Publication number Publication date
CN113987155A (en) 2022-01-28

Similar Documents

Publication Publication Date Title
CN113987155B (en) Conversational retrieval method integrating knowledge graph and large-scale user log
CN112784092B (en) Cross-modal image text retrieval method of hybrid fusion model
CN110826336B (en) Emotion classification method, system, storage medium and equipment
CN111597314B (en) Reasoning question-answering method, device and equipment
CN110674850A (en) Image description generation method based on attention mechanism
CN110347932B (en) Cross-network user alignment method based on deep learning
CN106021364A (en) Method and device for establishing picture search correlation prediction model, and picture search method and device
CN112084331A (en) Text processing method, text processing device, model training method, model training device, computer equipment and storage medium
US11768869B2 (en) Knowledge-derived search suggestion
CN112966091B (en) Knowledge map recommendation system fusing entity information and heat
CN114358657B (en) Post recommendation method and device based on model fusion
CN112988917A (en) Entity alignment method based on multiple entity contexts
CN116601626A (en) Personal knowledge graph construction method and device and related equipment
CN114780748A (en) Priori weight enhancement-based completion method of knowledge graph
CN114528898A (en) Scene graph modification based on natural language commands
CN113779996A (en) Standard entity text determination method and device based on BilSTM model and storage medium
CN115203550A (en) Social recommendation method and system for enhancing neighbor relation
CN113380360B (en) Similar medical record retrieval method and system based on multi-mode medical record map
CN114840747A (en) News recommendation method based on comparative learning
CN114706989A (en) Intelligent recommendation method based on technical innovation assets as knowledge base
CN114332519A (en) Image description generation method based on external triple and abstract relation
CN114996490A (en) Movie recommendation method, system, storage medium and device
CN115344794A (en) Scenic spot recommendation method based on knowledge map semantic embedding
CN113821610A (en) Information matching method, device, equipment and storage medium
CN113076490A (en) Case-related microblog object-level emotion classification method based on mixed node graph

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant