CN110334219B - Knowledge graph representation learning method based on attention mechanism integrated with text semantic features - Google Patents

Knowledge graph representation learning method based on attention mechanism integrated with text semantic features Download PDF

Info

Publication number
CN110334219B
CN110334219B CN201910629813.XA CN201910629813A CN110334219B CN 110334219 B CN110334219 B CN 110334219B CN 201910629813 A CN201910629813 A CN 201910629813A CN 110334219 B CN110334219 B CN 110334219B
Authority
CN
China
Prior art keywords
entity
attention
vector
word
relation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910629813.XA
Other languages
Chinese (zh)
Other versions
CN110334219A (en
Inventor
惠孛
罗光春
张栗粽
卢国明
李攀成
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201910629813.XA priority Critical patent/CN110334219B/en
Publication of CN110334219A publication Critical patent/CN110334219A/en
Application granted granted Critical
Publication of CN110334219B publication Critical patent/CN110334219B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Machine Translation (AREA)

Abstract

The invention relates to a knowledge graph, discloses a knowledge graph representation learning method based on a attention mechanism and integrating text semantic features, and solves the problems that semantic features are insufficient due to the fact that a translation model fails to utilize descriptive texts of entities and relations, the multi-source information embedding method fails to integrate the semantic features for the entities and the relations at the same time, and the text extraction effect is poor. The method can be summarized as follows: firstly, acquiring and processing description texts of entities and relations to obtain text semantic features, then constructing a projection matrix of the entities by utilizing the semantic features of the entities and the relations, projecting entity vectors into a relation space, modeling and carrying out representation learning in the relation space by utilizing the translation thought, and thus modeling the many-to-many complex relation. The method is suitable for the representation learning of the knowledge graph.

Description

Knowledge graph representation learning method based on attention mechanism integrated with text semantic features
Technical Field
The invention relates to a knowledge graph, in particular to a knowledge graph representation learning method based on a attention mechanism integrated with text semantic features.
Background
With the development of internet technology, data has shown explosive growth. However, because of heterogeneous content on the internet, the organization structure is loose, and it is difficult to efficiently use the information therein, google proposes the concept of a Knowledge Graph (knowledgegraph) in 2012, and aims to convert massive unstructured or semi-structured data into structured Knowledge which is unified and reliable in specification, thereby forming a semantic web with high internet, and providing support for data mining and intelligent services.
Knowledge graph can be regarded as a network of directed graph structures, wherein graph nodes represent entities or concepts, and edges in the graph represent relationships between entities and entities or between entities and concepts. Knowledge, i.e., (subject, predicate, object) or (entity, relationship, entity), is typically described in the form of triples. Knowledge-graph representation learning (Knowledge Graph Representation Learning) aims to learn a vectorized representation of entity relationships, converting knowledge in symbolic form into computable real-valued vectors.
In the conventional technology, a plurality of schemes for learning knowledge graph representation based on translation models exist:
mikolov et al found that there was a shift-invariant phenomenon in the word vector space using word2vec, such as v (king) -v (queen) ≡v (man) -v (woman), where v (king) represents the vector of word king obtained using word2 vec. Inspired by this phenomenon, bordes et al propose translation (translation) operations in which the relationships in the knowledge-graph are treated as head-entity to tail-entity translations in the embedding space by the TransE model: if a triplet (h, r, t) exists or holds, then in the embedding space the head entity vector plus the relation vector should be as close as possible to the tail entity vector, i.e. h+r≡t. Which defines a scoring function of
Figure GDA0004134145070000013
The TransE model is simple and effective, has expansibility on a large-scale knowledge graph, but has serious defects. The relationship in the knowledge graph can be divided into four types of 1-1, 1-N, N-1 and N-N according to the number of entities connected at two ends of the knowledge graph, and the TransE model determines that the knowledge graph is effective only for the relationship of 1-1, and has great problems for other relationship types, such as under the N-1 relationship,
Figure GDA0004134145070000012
Figure GDA0004134145070000014
(h i r, T) ∈T, meaning h 0 =h 1 =…=h m This is obviously unreasonable.
Aiming at the defect of the TransE on the complex relation, the head and tail entity vectors are respectively projected to the relation plane by the TransH model and then the translation operation is carried out, so that the entities can have different representations under different relations. W for TransH r 、d r The two vectors represent the relationship r, where w r Is the normal vector of the relation hyperplane, d r For translation operation corresponding to the relation, first, the head entity vector and the tail entity vector are respectively projected to a relation plane, then the translation operation is carried out, and the corresponding scoring function is as follows:
Figure GDA0004134145070000011
both TransE and TransH assume that entities and relationships are in the same semantic space, while relationships and entities are different objects, and TransR models entities and relationships in different spaces. For a triplet (h, r, t), entity embedding
Figure GDA0004134145070000021
Relational embedding->
Figure GDA0004134145070000022
For each relation r a projection matrix is set +.>
Figure GDA0004134145070000023
For projecting entities from the entity space to the relationship space. Similarly, its scoring function becomes:
Figure GDA0004134145070000024
/>
TransD proposes a method of dynamically changing matrices to address multiple semantic representations of relationships. It defines two representations for each entity or relationship, one (h, r, t) representing its own semantics and the other (h p ,r p ,t p ) Representing the way in which the entity vector is projected into the relationship vector space, and the second representation will be used to construct the mapping matrix:
Figure GDA0004134145070000025
Figure GDA0004134145070000026
the projected entity relation vector and scoring function can be obtained after the mapping matrix is provided:
h =M rh h,t =M rt t (6)
Figure GDA0004134145070000027
it can be seen that the models of TransD and TransE, transH are essentially modeling only the structural features within a triplet, i.e., a "translation", while ignoring other semantic features of entities and relationships.
Some multi-source information embedding methods in the conventional technology introduce more semantic features for entity and relation representation by embedding text corpus:
the DKRL takes the text of the entity as the entity description, and proposes a knowledge representation learning method integrated with the entity description information. Each entity has two representations: structure-based representation e s And description-based representation e d The triplet score consists of two parts: e=e S +E D . The representation of the structure uses the TransE model: e (E) S =||h s +r-t s To relate the learning process of the description-based representation to E S Adaptation, E D And is further divided into E DD =||h d +r-t d ||、E DS =||h d +r-t s ||、E SD =||h s +r-t d Three parts. Based on the representation of the description obtained by processing the text of the entity description, the author designed two ways of CBOW (Continuous Bag of Words) encoders and convolutional neural network encoders to extract semantic features of the entity description. It can be seen that DKRL uses the transition model when combining entity description information, however, the transition cannot model many-to-many relationships. In addition, the DKRL method only introduces descriptive information for the entity, and does not consider semantic features of the relationship.
TEKE is also a representation learning method that uses text to enhance entity relationship semantics: given a knowledge graph KG and a text corpus expressed as a word sequence, TEKE firstly marks words in a language library by using an entity linking tool to obtain a marking sequence D= (x) of the entities in the corresponding knowledge graph 1 ,x 2 ,…x n )。In order to combine the knowledge graph KG with the text information D, the author constructs a co-occurrence network g= (X, Y) consisting of entities and words, where X i Representing nodes of the network, corresponding to a word or an entity, y ij Represents x i and xj Co-occurrence frequency between them. Based on the co-occurrence network, a set of tagged words whose co-occurrence frequency exceeds a given threshold is selected as the semantic context of the corresponding entity and a vector representation thereof is constructed. The text processing mode of the TEKE construction co-occurrence network is traditional, the operation is complex, and semantic information among words in the sequence is not fully utilized.
In summary, the translation-based model essentially models only the structural features inside the triples, and does not use the descriptive text of the entities and relationships, so that other semantic features of the entity relationships in the knowledge graph are ignored. Under the circumstance, the sparsity of the knowledge graph leads to insufficient learning of the entity relation vector, often only roughly meets the translation characteristic, has low quality, leads to difficult distinction of some entities with the same relation but different meanings, and brings negative influence to the accuracy of the tasks such as subsequent knowledge fusion, knowledge graph completion and the like.
The multi-source information embedding methods such as DKRL, TEKE and the like expand the semantics of the entity by embedding entity description text corpus, but the methods have the following defects: firstly, DKRL is subjected to structure embedding by using a TransE method, so that the complex relation of many-to-many in a knowledge graph cannot be satisfied; secondly, DKRL only embeds the entity text description, so that semantic features are integrated for the entity, but the semantic features of the relation are not considered; thirdly, when the DKRL and the TEKE process entity text descriptions, a convolutional neural network and a co-occurrence network of words are respectively used, and interaction among the words in the sequence is not considered.
Disclosure of Invention
The technical problems to be solved by the invention are as follows: the knowledge graph representation learning method based on the attention mechanism and integrating the text semantic features is provided, and the problems that the semantic features are insufficient due to the fact that a translation model fails to utilize the descriptive text of the entity and the relation, the semantic features cannot be integrated for the entity and the relation by the multi-source information embedding method at the same time, and the text extraction effect is poor are solved.
The technical scheme adopted for solving the technical problems is as follows:
the knowledge graph representation learning method based on the attention mechanism integrated with the text semantic features comprises the following steps:
step 1, defining two representations for each entity and relation in a knowledge graph, wherein the two representations comprise semantic feature vector representations of the entity and the relation in the knowledge graph;
step 2, aiming at each entity in the knowledge graph, acquiring sentences containing the entity from a corpus, preprocessing, and extracting semantic features of the sentences by adopting a self-attention mechanism to obtain text semantic feature vectors of the entity;
step 3, aiming at each relation in the knowledge graph, word segmentation is carried out on the name description of the relation to obtain a tag word set, and semantic features of the tag word set are extracted by adopting a self-attention mechanism to obtain text semantic feature vectors of the relation;
step 4, constructing a mapping matrix based on text semantic feature vectors of the entities and the relations, and constructing a triplet scoring function based on the thought of the translation model;
and 5, constructing a loss function based on the interval according to the triplet scoring function, taking the knowledge-graph triplet as a training set, training a model by adopting a gradient descent optimization algorithm, and finally obtaining vector representation of the entity and the relation.
As a further optimization, step 2 specifically includes:
step 2.1, acquiring entity description text and preprocessing:
for each entity e in the knowledge graph, acquiring at least one sentence containing the entity from a corpus as a description text of the entity, using a word segmentation tool to segment each sentence, and then removing stop words to obtain a preprocessed word sequence;
step 2.2, constructing a text feature extraction model:
constructing a network model consisting of multi-layer multi-unit self-attention modules, wherein the model is formed by stacking 3 identical layers in total, namely 3 identical layers exist in the longitudinal direction, each layer transversely comprises RH self-attention units for processing input so as to learn the characteristics of word sequences from different aspects, and each self-attention unit has different parameter matrixes; RH can be set in a self-defined manner;
step 2.3, obtaining vector representation x of each word as input of a model:
the vector representation of each word is represented by its word vector
Figure GDA0004134145070000041
And position vector->
Figure GDA0004134145070000042
And composition of (a) in a calculation manner of
x=l word +l pos (8)
Initializing word vectors by using a word embedding tool word2 vec; each word is calculated to have a position code determined by its position order pos in the entity description text sequence in such a way that the value of the ith dimension of the position vector is calculated as
Figure GDA0004134145070000043
Figure GDA0004134145070000044
Step 2.4, calculating the influence degree of each word and all other words in the sequence by using a self-attention mechanism so as to obtain the attention distribution of the self-attention to the other words, namely the weight value:
the influence degree between words is calculated by using multiplicative attention, and then multiplied by the original word vector to obtain the vector after attention
Figure GDA0004134145070000045
The calculation formula is that
Figure GDA0004134145070000046
Where n is the number of words in the sequence,
Figure GDA0004134145070000047
is a matrix of vectors of all words in the sequence,
Figure GDA0004134145070000048
as a parameter matrix, W 1 Initializing the values of (2) by adopting normal distribution at the beginning of training;
step 2.5, after feature extraction of 3 attention layers, adding all output vectors of the sequence, and taking the added output vectors as entity semantic features through a ReLU activation function, wherein a calculation formula is as follows:
Figure GDA0004134145070000049
step 2.6, processing RH different attention units, mapping the RH different attention units into a final entity semantic feature vector, and calculating the final entity semantic feature vector according to the following formula:
e p =ReLU(W 2 E+b) (13)
wherein
Figure GDA00041341450700000410
Is a mapping matrix, E is a matrix of RH outputs of different attention units, b is a bias vector, W 2 And the value of b is initialized with a normal distribution at the beginning of training.
As a further optimization, step 3 specifically includes:
step 3.1, preprocessing the relation name:
for each relation r in the knowledge graph, word segmentation is carried out on the names of the relations r by using a word segmentation tool to obtain a tag word sequence;
step 3.2, constructing a text feature extraction model:
constructing a network model consisting of a single-layer multi-unit self-attention module, wherein the model is provided with 1 self-attention layer in the longitudinal direction, and the layer transversely comprises RH self-attention units for processing input to learn the characteristics of word sequences from different aspects, and each self-attention unit is provided with a different parameter matrix; RH can be set in a self-defined manner;
step 3.3, obtaining vector representation of each tag word in the tag word sequence as input of a model;
step 3.4, calculating the matching degree between the tag words by using a self-attention mechanism, and multiplying the matching degree with the original word vector to obtain an attention vector;
step 3.5, after feature extraction of the attention layer, adding all output vectors of the sequence and using a ReLU activation function as a relation semantic feature;
step 3.6, processing RH different attention units, mapping the RH different attention units into a relation semantic feature vector r p
As a further optimization, step 4 specifically includes:
step 4.1, for a triplet (h, r, t), setting a projection matrix M for the head entity and the tail entity, respectively rh and Mrt For projecting an entity from an entity space to a relationship space; the projection matrix is constructed by semantic feature vectors of the entities and the relations obtained in the step 2 and the step 3, and the calculation mode is that
Figure GDA0004134145070000051
Figure GDA0004134145070000052
wherein Bk×d Is a parameter matrix to be learned;
step 4.2, multiplying the head entity, the tail entity and the respective projection matrix to calculate the projection of the entity in the relation space, namely: h is a =M rh h、t =M rt t;
And 4.3, in a relation space, following the thought of a translation model, regarding the relation as translation operation from a head entity to a tail entity, and constructing a triplet scoring function as follows:
Figure GDA0004134145070000053
as a further optimization, step 5 specifically includes:
step 5.1, taking all original triples T in the knowledge graph as training sets, defining an interval-based hinge loss function to train a model, wherein the goal is to enable a triplet scoring function to obtain a lower score for a positive case triplet and a higher score for a negative case triplet, and the loss function is as follows:
Figure GDA0004134145070000054
wherein ,T′(h,r,t) = { (h ', r, t) |h' ∈e, h 'noteqh }) u { (h, r, t')|t '∈e, t' noteqt } is a negative example set constructed on the basis of triples (h, r, t), and interval value γ > 0 is a super parameter;
step 5.2, for any entity, forcing its vector L2 norm to be 1, i.e
Figure GDA0004134145070000061
Thereby regularizing the entity embedded vector into a unit sphere;
in the training process, the fact triples of the knowledge graph are randomly traversed for a plurality of times, and when each fact triplet is accessed, a negative example triplet is constructed for the fact triplet, and the selection mode of a negative example entity is as follows: adopting a K-neighbor method, firstly calculating the similarity of an entity to be replaced and other entities by using a cosine similarity algorithm, sequencing from high to low, and then taking top-K entities as negative example candidate sets of the entity to be replaced;
step 5.4 optimizing the objective function Using a small batch gradient descent algorithm
Figure GDA0004134145070000062
Gradients are then calculated and model parameters are updated.
The beneficial effects of the invention are as follows:
(1) The invention integrates structural features and text semantic features for the entities and the relations at the same time:
the invention respectively embeds the text corpus of the entity description and the relation description, and uses the text corpus to construct a projection matrix of the entity to the relation space, and finally performs representation learning based on translation ideas in the relation space, thereby not only simultaneously considering semantic features of the entity and the relation, but also skillfully combining structure embedding and text embedding.
(2) Compared with other multisource information embedding methods, the method can extract richer semantic features:
the multi-layer self-attention method adopted by the invention processes entity description and relation description, and can efficiently extract semantic features with higher quality by virtue of the advantage of attention mechanism on natural language processing.
(3) The negative sampling method of the K-neighbor can enable the model to show better distinguishing capability:
the K-neighbor negative sampling method can improve the quality of the negative case triples, thereby enhancing the learning of the model and enabling the final model to better distinguish correct triples from error triples.
Drawings
FIG. 1 is a schematic diagram of a knowledge graph representation learning method based on attention mechanism incorporated text semantic features;
FIG. 2 is a flow chart of a learning method of knowledge graph representation based on attention mechanism merging text semantic features;
fig. 3 is a schematic diagram of text feature extraction based on an attention mechanism.
Detailed Description
The invention aims to provide a knowledge graph representation learning method based on the integration of a attention mechanism into text semantic features, which solves the problems that the semantic features are insufficient due to the fact that a translation model fails to utilize descriptive texts of entities and relations, the multi-source information embedding method fails to integrate the semantic features into the entities and the relations at the same time, and the text extraction effect is poor.
The knowledge graph representation learning method combines text embedding and translation ideas. Firstly, acquiring and processing description texts of entities and relations to obtain text semantic features, then constructing a projection matrix of the entities by utilizing the semantic features of the entities and the relations, projecting entity vectors into a relation space, modeling and carrying out representation learning in the relation space by utilizing the translation thought, and thus modeling the many-to-many complex relation. The implementation principle is shown in fig. 1, and in the relation space, the relation is regarded as a translation operation from a head entity to a tail entity according to the idea of a translation model.
The knowledge graph representation learning method of the invention is shown in fig. 2, and comprises the following implementation steps:
step 1, defining two types of representation for each entity e in the knowledge graph, wherein one type of representation is semantic features of the entity itself and is represented as e. The other is the text semantic feature of the entity, denoted as e p . The two representations are also defined for each relationship r in the knowledge-graph.
Step 2, for each entity e in the knowledge graph, acquiring sentences containing the entity from a corpus, preprocessing, and extracting semantic features of the sentences by adopting a self-attention mechanism to obtain text semantic feature vectors e of the entities p
Step 3, for each relation r in the knowledge graph, word segmentation is carried out on the name description of each relation r to obtain a tag word set, and semantic features of the tag word set are extracted by adopting a self-attention mechanism to obtain semantic feature vectors r of the relations p
And 4, constructing a mapping matrix by utilizing semantic feature vectors of the entities and the relations, and constructing a triplet scoring function, namely an energy equation, based on the translation thought.
And 5, constructing a loss function based on the interval according to the triplet scoring function, taking the knowledge-graph triplet as a training set, training a model by adopting a gradient descent optimization algorithm, and finally obtaining vector representation of the entity and the relation.
In the implementation, the required original data are a triplet set of the knowledge graph and a corpus-text set of the same language as the knowledge graph. The specific implementation means of each step are further described below:
in step 1, all entities and relations of the knowledge graph are firstly obtained, two vectors of the entities and the relations are initialized by using tensorflow, and the dimension values of the entities and the relations are respectively super-parameters d and k and can be selected from {50, 70, 80, 100 }. The semantic feature vectors of entities and relationships themselves use boundaries as
Figure GDA0004134145070000071
Is initialized for the uniform distribution of the memory. Text semantic feature e of entities and relationships p and rp Not randomly initialized, but calculated by the steps 2 and 3.
Step 2, using the description text of the entity as input, extracting the semantic features of sentences by adopting a self-attention mechanism, and outputting a vector e p . The method comprises the following specific steps:
step 2.1, preprocessing entity description text:
for each entity e in the knowledge graph, at least one sentence containing the entity is obtained from the corpus as a description text of the entity, a word segmentation tool is used for segmenting each sentence, and then the stop words are removed to obtain a preprocessed word sequence.
Step 2.2, constructing a text feature extraction model:
the basic processing unit of feature extraction is to apply a self-attention mechanism to the sequence, the model consists of a multi-layer, multi-unit self-attention module, each layer having RH self-attention units, processing the input to learn the features of the sequence from different aspects. The model is stacked with a total of ch=3 identical layers, i.e. 3 identical layers in the longitudinal direction, each layer containing RH self-attention units in the lateral direction, as shown in fig. 3. Wherein each self-attention unit has a different parameter matrix. In the network model building, RH can be set in a self-defined manner, and is generally selected from {1,2,3,4 }.
Step 2.3, input of the model isThe vector for each word represents x. The vector representation of each word is represented by its word vector
Figure GDA0004134145070000081
Figure GDA00041341450700000812
And position vector->
Figure GDA0004134145070000082
And composition of (a) in a calculation manner of
x=l word +l pos (8)
The word vector is initialized with the word embedding tool word2 vec. Each word is calculated to have a position code determined by its position order pos in the entity description text sequence in such a way that the value of the ith dimension of the position vector is calculated as
Figure GDA0004134145070000083
Figure GDA0004134145070000084
And 2.4, calculating the influence degree of each word and all other words in the sequence by using a self-attention mechanism to obtain the attention allocation (namely weight value) of the self-attention to the other words, wherein the weight value determines how much each word is expressed in the belonged position. The influence degree between words is calculated by using multiplicative attention, and then multiplied by the original word vector to obtain the vector after attention
Figure GDA0004134145070000085
The calculation formula is that
Figure GDA0004134145070000086
Where n is the number of words in the sequence.
Figure GDA0004134145070000087
Is a matrix of vectors of all words in the sequence.
Figure GDA0004134145070000088
As a parameter matrix, W 1 The values of (2) may be initialized with a normal distribution at the beginning of training. Divided by->
Figure GDA0004134145070000089
The purpose of (a) is to scale the weight value to prevent its value from becoming too large.
Step 2.5, after feature extraction of CH attention layers, adding all output vectors of the sequence, and taking the result as entity semantic features through a ReLU activation function, wherein a calculation formula is as follows
Figure GDA00041341450700000810
And 2.6, processing RH different attention units and mapping the RH different attention units into a final entity semantic feature vector in order to integrate semantic features learned in different aspects. The calculation mode is that
e p =ReLU(W 2 E+b) (13)
wherein
Figure GDA00041341450700000811
Is a mapping matrix, E is a matrix of outputs of RH different attention cells, and b is a bias vector. W (W) 2 And the value of b may be initialized with a normal distribution at the beginning of training.
Step 3, using the name tag words of the relation as input, extracting semantic features of the tag word set by adopting a self-attention mechanism, and outputting a vector r p . The method specifically comprises the following steps:
step 3.1, preprocessing the relation names: and for each relation r in the knowledge graph, word segmentation is carried out on the names of the relations r by using a word segmentation tool, so as to obtain a tag word sequence. The tag word set of { accident, traffic, accident, response, part } is obtained after the relationship name "/accident/traffic_accident/res-active_part" is processed.
Step 3.2, constructing a text feature extraction model:
similar to the model of entity semantic feature extraction, the model is composed of a single-layer multi-unit self-attention module, each layer has RH self-attention units, and as the relation description contains fewer words and fewer word ranges, the extraction model of the relation semantic features only contains CH=1 self-attention layers, namely 1 self-attention layer in the longitudinal direction, and each layer transversely contains RH self-attention units. Wherein each self-attention unit has a different parameter matrix. In the network model building, RH can be set in a self-defined manner, and is generally selected from {1,2,3,4 }.
Step 3.3, the input of the model is the vector representation of each tag word in a sequence, and the calculation mode is consistent with that of step 2.3, firstly, the word vector of each word is initialized by using word2vec to obtain the word vector of each word
Figure GDA0004134145070000091
The embedding dimension k is selected from {50, 70, 80, 100 }. Calculating a position vector of each tag word using equation (8) and equation (9)
Figure GDA0004134145070000092
And 3.4, calculating the matching degree between the tag words by using a self-attention mechanism, and multiplying the matching degree with the original word vector to obtain an attention vector, wherein the calculation mode is consistent with that of the step 2.4.
And 3.5, after feature extraction of CH attention layers, adding all output vectors of the sequence and using a ReLU activation function as a relation semantic feature. The calculation mode is consistent with the step 2.5.
Step 3.6, in order to integrate the semantic features learned in different aspects, at last, the RH different attention units are processed and mapped into a final relational semantic feature vector r p . The calculation mode is consistent with the step 2.6
And 4, constructing a mapping matrix by utilizing semantic feature vectors of the entities and the relations, and constructing a triplet scoring function, namely an energy equation, based on the translation thought. The method specifically comprises the following steps:
step 4.1, for a triplet (h, r, t), setting a projection matrix M for the head entity and the tail entity, respectively rh and Mrt For projecting an entity from an entity space to a relationship space. The projection matrix is constructed by semantic feature vectors of the entities and the relations obtained in the step 2 and the step 3, and the calculation mode is that
Figure GDA0004134145070000093
Figure GDA0004134145070000094
wherein Bk×d Is a parameter matrix to be learned.
Step 4.2, multiplying the head and tail entities and respective projection matrixes to calculate projections of the entities in a relation space, namely: h is a =M rh h、t =M rt t。
Step 4.3, constructing a triplet scoring function (i.e. an energy equation) as follows the idea of a translation model in a relation space, regarding the relation as a translation operation from a head entity to a tail entity
Figure GDA0004134145070000095
And 5, constructing a loss function based on the interval according to the triplet scoring function, taking the knowledge-graph triplet as a training set, training a model by adopting a gradient descent optimization algorithm, and finally obtaining vector representation of the entity and the relation. The method comprises the following detailed steps:
and 5.1, defining an interval-based hinge loss function by taking all original triples T in the knowledge graph as a training set so as to train a model. The goal is to have the triplet scoring function get a lower score (energy) for positive case triples and a higher score for negative case triples. The loss function is
Figure GDA0004134145070000101
wherein ,T′(h,r,t) = { (h ', r, t) |h' ∈e, h '+.h }) u { (h, r, t')|t '∈e, t' +.t } is a negative example set constructed on the basis of triples (h, r, t). The interval value gamma > 0 is a super parameter and can be selected from {1,2,3,4 }.
Step 5.2, for any entity, forcing its vector L2 norm to be 1, i.e
Figure GDA0004134145070000102
Therefore, the entity embedding vector is regularized into a unit sphere, and invalid convergence of the objective function can be prevented by artificially increasing the entity embedding specification. />
In the training process, the fact triples (training sets) of the knowledge graph are randomly traversed a plurality of times, and when each fact triplet is accessed, a negative example triplet is constructed for the fact triplet. The negative example entity is not selected in the entity set, but a K-neighbor method is adopted, the similarity between the entity to be replaced and other entities is calculated by utilizing a cosine similarity algorithm, the entities are ordered from high to low, and top-K entities are taken as negative example candidate sets of the entity to be replaced.
Step 5.4 optimization of the objective function Using a small batch gradient descent (Mini-batch Gradient Descent)
Figure GDA0004134145070000103
The learning rate μ is selected from {0.1,0.01,0.001}, and the batch size value B is selected from {200, 500, 1400, 4800 }. After a small batch, gradients are calculated and model parameters are updated.
Based on the above scheme of the invention, compared with the traditional technology, the invention has at least the following advantages:
(1) The invention integrates structural features and text semantic features for the entities and the relations at the same time:
the translation models TransE, transH, transR, transD and the like are all structural features only modeling the interior of the triples, and have the disadvantage of ignoring other semantic features of entity relations in the knowledge graph. The text description of the entity is embedded by other multisource information embedding methods such as TEKE, DKRL and the like on the basis of the TransE, so that semantic features for describing the text are introduced for the entity, and the following disadvantages still exist: firstly, the used TransE cannot meet the complex relation of many-to-many in the knowledge graph, and secondly, semantic features are only integrated for the entity. The invention respectively embeds the text corpus of the entity description and the relation description, and uses the text corpus to construct a projection matrix of the entity to the relation space, and finally performs representation learning based on translation ideas in the relation space. The semantic features of the entity and the relation are considered at the same time, and the structure embedding and the text embedding are combined skillfully.
(2) Compared with other multisource information embedding methods, the method can extract richer semantic features:
the entity description text is processed by the TEKE based on the co-occurrence network of the word and the entity, the entity description text is processed by the DKRL by using a continuous word bag model or a convolutional neural network, the modes belong to more traditional modes on natural language processing, and the advantage of an attention mechanism on natural language processing is benefited.
(3) The negative sampling method of the K-neighbor can enable the model to show better distinguishing capability:
choosing one arbitrarily from the whole entity set for replacement may result in the generation of very easily distinguishable negative triples, such as for triples (beijin, caps of, china), by replacing head entities (Water, caps of, china) or replacing tail entities (beijin, caps of, airland), which are all obviously erroneous or even not logical. The Beijing and Water correspondence vectors are inherently far apart in the same space, so that the negative case triples have little learning contribution to the model. In contrast, (Hong Kong, caps of, china) is a close but erroneous triplet. The K-neighbor negative sampling method can improve the quality of the negative case triples, thereby enhancing the learning of the model and enabling the final model to better distinguish correct triples from error triples.

Claims (4)

1. The knowledge graph representation learning method based on the attention mechanism integrated with the text semantic features is characterized by comprising the following steps of:
step 1, defining two representations for each entity and relation in a knowledge graph, wherein the two representations comprise semantic feature vector representations of the entity and the relation in the knowledge graph;
step 2, aiming at each entity in the knowledge graph, acquiring sentences containing the entity from a corpus, preprocessing, and extracting semantic features of the sentences by adopting a self-attention mechanism to obtain text semantic feature vectors of the entity;
step 3, aiming at each relation in the knowledge graph, word segmentation is carried out on the name description of the relation to obtain a tag word set, and semantic features of the tag word set are extracted by adopting a self-attention mechanism to obtain text semantic feature vectors of the relation;
step 4, constructing a mapping matrix based on text semantic feature vectors of the entities and the relations, and constructing a triplet scoring function based on the thought of the translation model; the step 4 specifically comprises the following steps:
step 4.1, for a triplet (h, r, t), setting a projection matrix M for the head entity and the tail entity, respectively rh and Mrt For projecting an entity from an entity space to a relationship space; the projection matrix is constructed by semantic feature vectors of the entities and the relations obtained in the step 2 and the step 3, and the calculation mode is as follows:
Figure FDA0004148970210000011
Figure FDA0004148970210000012
wherein ,Bk×d Is a parameter matrix to be learned; h is a p A semantic feature vector representing the head entity h,
Figure FDA0004148970210000013
is h p Is a transpose of (2); t is t p Semantic feature vector representing tail entity t, +.>
Figure FDA0004148970210000014
At t p Is a transpose of (2); r is (r) p A semantic feature vector representing a relationship r;
step 4.2, multiplying the head entity, the tail entity and the respective projection matrix to calculate the projection of the entity in the relation space, namely: h is a =M rh h、t =M rt t;
And 4.3, in a relation space, following the thought of a translation model, regarding the relation as translation operation from a head entity to a tail entity, and constructing a triplet scoring function as follows:
Figure FDA0004148970210000015
and 5, constructing a loss function based on the interval according to the triplet scoring function, taking the knowledge-graph triplet as a training set, training a model by adopting a gradient descent optimization algorithm, and finally obtaining vector representation of the entity and the relation.
2. The knowledge graph representation learning method based on the attention mechanism integrated with text semantic features according to claim 1,
the method is characterized in that the step 2 specifically comprises the following steps:
step 2.1, acquiring entity description text and preprocessing:
for each entity e in the knowledge graph, acquiring at least one sentence containing the entity from a corpus as a description text of the entity, using a word segmentation tool to segment each sentence, and then removing stop words to obtain a preprocessed word sequence;
step 2.2, constructing a text feature extraction model:
constructing a network model consisting of multi-layer multi-unit self-attention modules, wherein the model is formed by stacking 3 identical layers in total, namely 3 identical layers exist in the longitudinal direction, each layer transversely comprises RII self-attention units for processing input so as to learn the characteristics of word sequences from different aspects, and each self-attention unit has different parameter matrixes; RII can be set in a self-defined manner;
step 2.3, obtaining vector representation x of each word as input of a model:
the vector representation of each word is represented by its word vector
Figure FDA0004148970210000021
And position vector->
Figure FDA0004148970210000022
And composition of (a) in a calculation manner of
x=l word +l pos (8)
Where d represents the dimension of the feature vector,
Figure FDA0004148970210000023
representing real space +.>
Figure FDA0004148970210000024
D-dimensional vector of (a);
initializing word vectors by using a word embedding tool word2 vec; each word is calculated to have a position code determined by its position order pos in the entity description text sequence in such a way that the value of the ith dimension of the position vector is calculated as
Figure FDA0004148970210000025
Figure FDA0004148970210000026
Step 2.4, calculating the influence degree of each word and all other words in the sequence by using a self-attention mechanism so as to obtain the attention distribution of the self-attention to the other words, namely the weight value:
the influence degree between words is calculated by using multiplicative attention, and then multiplied by the original word vector to obtain the vector after attention
Figure FDA0004148970210000027
The calculation formula is that
Figure FDA0004148970210000028
Where n is the number of words in the sequence,
Figure FDA0004148970210000029
is a matrix of vectors of all words in the sequence,
Figure FDA00041489702100000210
as a parameter matrix, W 1 Initializing the values of (2) by adopting normal distribution at the beginning of training; />
Figure FDA00041489702100000211
Representing real space +.>
Figure FDA00041489702100000212
Vector with medium dimension d×d, +.>
Figure FDA00041489702100000213
Representing real space +.>
Figure FDA00041489702100000214
A vector of dimension n x d; softmax () is a normalized exponential function;
step 2.5, after feature extraction of 3 attention layers, adding all output vectors of the sequence, and taking the added output vectors as entity semantic features through a ReLU activation function, wherein a calculation formula is as follows:
Figure FDA0004148970210000031
wherein ,|LSr The i indicates the number of attention layers; m is m i Representing the features extracted by entity e at the ith attention layer;
step 2.6, processing RH different attention units, mapping the RH different attention units into a final entity semantic feature vector, and calculating the final entity semantic feature vector according to the following formula:
e p =ReLU(W 2 E+b) (13)
wherein ,
Figure FDA0004148970210000032
is a mapping matrix, E is a matrix of RH outputs of different attention units, b is a bias vector, W 2 And the value of b is initialized with a normal distribution at the beginning of training.
3. The knowledge graph representation learning method based on the attention mechanism integrated with text semantic features according to claim 1,
the method is characterized in that the step 3 specifically comprises the following steps:
step 3.1, preprocessing the relation name:
for each relation r in the knowledge graph, word segmentation is carried out on the names of the relations r by using a word segmentation tool to obtain a tag word sequence;
step 3.2, constructing a text feature extraction model:
constructing a network model consisting of a single-layer multi-unit self-attention module, wherein the model is provided with 1 self-attention layer in the longitudinal direction, and the layer transversely comprises RH self-attention units for processing input to learn the characteristics of word sequences from different aspects, and each self-attention unit is provided with a different parameter matrix; RH can be set in a self-defined manner;
step 3.3, obtaining vector representation of each tag word in the tag word sequence as input of a model;
step 3.4, calculating the matching degree between the tag words by using a self-attention mechanism, and multiplying the matching degree with the original word vector to obtain an attention vector;
step 3.5, after feature extraction of the attention layer, adding all output vectors of the sequence and using a ReLU activation function as a relation semantic feature;
step 3.6, processing RH different attention units, mapping the RH different attention units into a relation semantic feature vector r p
4. The knowledge graph representation learning method based on the attention mechanism integrated with text semantic features according to claim 1,
the method is characterized in that the step 5 specifically comprises the following steps:
step 5.1, taking all original triples T in the knowledge graph as training sets, defining an interval-based hinge loss function to train a model, wherein the goal is to enable a triplet scoring function to obtain a lower score for a positive case triplet and a higher score for a negative case triplet, and the loss function is as follows:
Figure FDA0004148970210000041
wherein ,T′(h,r,t) = { (h ', r, t) |h' ∈e, h 'noteqh }) u { (h, r, t')|t '∈e, t' noteqt } is a negative example set constructed on the basis of triples (h, r, t), and interval value γ > 0 is a super parameter;
step 5.2, for any entity, forcing its vector L2 norm to be 1, i.e
Figure FDA0004148970210000042
Thereby regularizing the entity embedded vector into a unit sphere;
in the training process, the fact triples of the knowledge graph are randomly traversed for a plurality of times, and when each fact triplet is accessed, a negative example triplet is constructed for the fact triplet, and the selection mode of a negative example entity is as follows: adopting a K-neighbor method, firstly calculating the similarity of an entity to be replaced and other entities by using a cosine similarity algorithm, sequencing from high to low, and then taking top-K entities as negative example candidate sets of the entity to be replaced;
step 5.4 optimizing the objective function Using a small batch gradient descent algorithm
Figure FDA0004148970210000043
Gradients are then calculated and model parameters are updated. />
CN201910629813.XA 2019-07-12 2019-07-12 Knowledge graph representation learning method based on attention mechanism integrated with text semantic features Active CN110334219B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910629813.XA CN110334219B (en) 2019-07-12 2019-07-12 Knowledge graph representation learning method based on attention mechanism integrated with text semantic features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910629813.XA CN110334219B (en) 2019-07-12 2019-07-12 Knowledge graph representation learning method based on attention mechanism integrated with text semantic features

Publications (2)

Publication Number Publication Date
CN110334219A CN110334219A (en) 2019-10-15
CN110334219B true CN110334219B (en) 2023-05-09

Family

ID=68146717

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910629813.XA Active CN110334219B (en) 2019-07-12 2019-07-12 Knowledge graph representation learning method based on attention mechanism integrated with text semantic features

Country Status (1)

Country Link
CN (1) CN110334219B (en)

Families Citing this family (66)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110851620B (en) * 2019-10-29 2023-07-04 天津大学 Knowledge representation method based on text embedding and structure embedding combination
CN111046187B (en) * 2019-11-13 2023-04-18 山东财经大学 Sample knowledge graph relation learning method and system based on confrontation type attention mechanism
CN110866119B (en) * 2019-11-14 2021-06-15 腾讯科技(深圳)有限公司 Article quality determination method and device, electronic equipment and storage medium
CN111160564B (en) * 2019-12-17 2023-05-19 电子科技大学 Chinese knowledge graph representation learning method based on feature tensor
CN111061843B (en) * 2019-12-26 2023-08-25 武汉大学 Knowledge-graph-guided false news detection method
CN111209410B (en) * 2019-12-27 2023-04-18 中国地质大学(武汉) Anchor point-based dynamic knowledge graph representation learning method and system
CN111191004B (en) * 2019-12-27 2023-09-22 咪咕文化科技有限公司 Text label extraction method, text label extraction device and computer readable storage medium
CN111159485B (en) * 2019-12-30 2020-11-13 科大讯飞(苏州)科技有限公司 Tail entity linking method, device, server and storage medium
CN111159431A (en) * 2019-12-30 2020-05-15 深圳Tcl新技术有限公司 Knowledge graph-based information visualization method, device, equipment and storage medium
CN111462914B (en) * 2020-03-13 2023-07-25 云知声智能科技股份有限公司 Entity linking method and device
CN111428047B (en) * 2020-03-19 2023-04-21 东南大学 Knowledge graph construction method and device based on UCL semantic indexing
CN111444343B (en) * 2020-03-24 2021-04-06 昆明理工大学 Cross-border national culture text classification method based on knowledge representation
CN111496784B (en) * 2020-03-27 2021-05-07 山东大学 Space environment identification method and system for robot intelligent service
CN111552817A (en) * 2020-04-14 2020-08-18 国网内蒙古东部电力有限公司 Electric power scientific and technological achievement knowledge map completion method
CN111539197B (en) * 2020-04-15 2023-08-15 北京百度网讯科技有限公司 Text matching method and device, computer system and readable storage medium
CN113536742A (en) * 2020-04-20 2021-10-22 阿里巴巴集团控股有限公司 Method and device for generating description text based on knowledge graph and electronic equipment
CN111680163A (en) * 2020-04-21 2020-09-18 国网内蒙古东部电力有限公司 Knowledge graph visualization method for electric power scientific and technological achievements
CN111581392B (en) * 2020-04-28 2022-07-05 电子科技大学 Automatic composition scoring calculation method based on statement communication degree
CN111538848B (en) * 2020-04-29 2023-09-01 华中科技大学 Knowledge representation learning method integrating multi-source information
CN111581395B (en) * 2020-05-06 2023-09-19 西安交通大学 Model fusion triplet representation learning system and method based on deep learning
CN111737591B (en) * 2020-06-01 2024-03-15 山西大学 Product recommendation method based on heterogeneous heavy side information network translation model
CN112035672B (en) * 2020-07-23 2023-05-09 深圳技术大学 Knowledge graph completion method, device, equipment and storage medium
CN112100393B (en) * 2020-08-07 2022-03-15 浙江大学 Knowledge triple extraction method under low-resource scene
CN112036189A (en) * 2020-08-10 2020-12-04 中国人民大学 Method and system for recognizing gold semantic
CN111897975A (en) * 2020-08-12 2020-11-06 哈尔滨工业大学 Local training method for learning training facing knowledge graph representation
CN111897974B (en) * 2020-08-12 2024-04-16 吉林大学 Heterogeneous knowledge graph learning method based on multilayer attention mechanism
CN112380325B (en) * 2020-08-15 2022-05-31 电子科技大学 Knowledge graph question-answering system based on joint knowledge embedded model and fact memory network
CN112000689B (en) * 2020-08-17 2022-10-18 吉林大学 Multi-knowledge graph fusion method based on text analysis
CN111932026B (en) * 2020-08-27 2022-03-04 西南交通大学 Urban traffic pattern mining method based on data fusion and knowledge graph embedding
CN112052685B (en) * 2020-09-11 2024-06-04 河南合众伟奇云智科技有限公司 End-to-end text entity relation identification method based on two-dimensional time sequence network
CN112084428B (en) * 2020-09-17 2024-02-02 辽宁工程技术大学 Collaborative filtering recommendation method based on coupling network embedding and knowledge graph
CN112131404B (en) * 2020-09-19 2022-09-27 哈尔滨工程大学 Entity alignment method in four-risk one-gold domain knowledge graph
CN112307777B (en) * 2020-09-27 2022-03-11 和美(深圳)信息技术股份有限公司 Knowledge graph representation learning method and system
CN111950303B (en) * 2020-10-19 2021-01-08 平安科技(深圳)有限公司 Medical text translation method, device and storage medium
CN112364174A (en) * 2020-10-21 2021-02-12 山东大学 Patient medical record similarity evaluation method and system based on knowledge graph
CN112668719A (en) * 2020-11-06 2021-04-16 北京工业大学 Knowledge graph construction method based on engineering capacity improvement
CN112347268B (en) * 2020-11-06 2024-03-19 华中科技大学 Text-enhanced knowledge-graph combined representation learning method and device
CN112507039A (en) * 2020-12-15 2021-03-16 苏州元启创人工智能科技有限公司 Text understanding method based on external knowledge embedding
CN113312487A (en) * 2021-01-16 2021-08-27 江苏网进科技股份有限公司 Knowledge representation learning method facing legal text based on TransE model
CN112667824B (en) * 2021-01-17 2024-03-15 北京工业大学 Knowledge graph completion method based on multi-semantic learning
CN112800239B (en) * 2021-01-22 2024-04-12 中信银行股份有限公司 Training method of intention recognition model, and intention recognition method and device
CN112784049B (en) * 2021-01-28 2023-05-12 电子科技大学 Text data-oriented online social platform multi-element knowledge acquisition method
CN112732944A (en) * 2021-01-30 2021-04-30 吉林大学 New method for text retrieval
CN112925953B (en) * 2021-03-09 2024-02-20 南京航空航天大学 Dynamic network representation method and system
CN112950325B (en) * 2021-03-16 2023-10-03 山西大学 Self-attention sequence recommendation method for social behavior fusion
CN113254663B (en) * 2021-04-21 2022-06-17 浙江工业大学 Knowledge graph joint representation learning method integrating graph convolution and translation model
CN113204647B (en) * 2021-04-29 2023-01-03 哈尔滨工程大学 Joint weight-based encoding and decoding framework knowledge graph embedding method
CN113312498B (en) * 2021-06-09 2022-06-17 上海交通大学 Text information extraction method for embedding knowledge graph by undirected graph
CN113360678B (en) * 2021-07-08 2022-07-15 电子科技大学 Elementary mathematic knowledge graph construction method based on Neo4j and big data
CN113488165B (en) * 2021-07-26 2023-08-22 平安科技(深圳)有限公司 Text matching method, device, equipment and storage medium based on knowledge graph
CN113590837B (en) * 2021-07-29 2024-08-16 华中农业大学 Deep learning-based food and health knowledge graph construction method
CN113569773B (en) * 2021-08-02 2023-09-15 南京信息工程大学 Interference signal identification method based on knowledge graph and Softmax regression
CN113626610A (en) * 2021-08-10 2021-11-09 南方电网数字电网研究院有限公司 Knowledge graph embedding method and device, computer equipment and storage medium
CN113535984B (en) * 2021-08-11 2023-05-26 华侨大学 Knowledge graph relation prediction method and device based on attention mechanism
CN113590799B (en) * 2021-08-16 2022-11-18 东南大学 Weak supervision knowledge graph question-answering method based on multi-view reasoning
CN113761224A (en) * 2021-09-01 2021-12-07 东北大学 Long text friendly knowledge graph representation learning method
CN114491070B (en) * 2022-01-24 2024-08-27 广东技术师范大学 Graph embedding method and system based on semantic attributes of knowledge graph nodes
CN114582443B (en) * 2022-02-23 2023-08-18 西北大学 Knowledge graph-based drug relation extraction method
CN114979705A (en) * 2022-04-12 2022-08-30 杭州电子科技大学 Automatic editing method based on deep learning, self-attention mechanism and symbolic reasoning
CN114860877A (en) * 2022-04-29 2022-08-05 华侨大学 Problem chain generation method and system based on knowledge graph relation prediction
CN115438674B (en) * 2022-11-08 2023-03-24 腾讯科技(深圳)有限公司 Entity data processing method, entity linking method, entity data processing device, entity linking device and computer equipment
CN115936737B (en) * 2023-03-10 2023-06-23 云筑信息科技(成都)有限公司 Method and system for determining authenticity of building material
CN116187446B (en) * 2023-05-04 2023-07-04 中国人民解放军国防科技大学 Knowledge graph completion method, device and equipment based on self-adaptive attention mechanism
CN116702898B (en) * 2023-08-04 2023-11-03 北京语言大学 Knowledge representation learning-based cultural relics and literary knowledge migration method and system
CN117251583B (en) * 2023-11-20 2024-01-26 湖北大学 Text enhanced knowledge graph representation learning method and system based on local graph structure
CN118349671B (en) * 2024-02-04 2024-08-27 杭州微未计算机技术有限公司 Multi-platform new media information source comprehensive label learning system

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107590237A (en) * 2017-09-11 2018-01-16 桂林电子科技大学 A kind of knowledge mapping based on dynamic translation principle represents learning method

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102016223193A1 (en) * 2016-11-23 2018-05-24 Fujitsu Limited Method and apparatus for completing a knowledge graph
CN107885760B (en) * 2016-12-21 2021-06-08 桂林电子科技大学 Knowledge graph representation learning method based on multiple semantics
CN107291687B (en) * 2017-04-27 2021-03-26 同济大学 Chinese unsupervised open type entity relation extraction method based on dependency semantics
CN109408812A (en) * 2018-09-30 2019-03-01 北京工业大学 A method of the sequence labelling joint based on attention mechanism extracts entity relationship

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107590237A (en) * 2017-09-11 2018-01-16 桂林电子科技大学 A kind of knowledge mapping based on dynamic translation principle represents learning method

Also Published As

Publication number Publication date
CN110334219A (en) 2019-10-15

Similar Documents

Publication Publication Date Title
CN110334219B (en) Knowledge graph representation learning method based on attention mechanism integrated with text semantic features
CN111382272B (en) Electronic medical record ICD automatic coding method based on knowledge graph
CN112347268B (en) Text-enhanced knowledge-graph combined representation learning method and device
CN112214995B (en) Hierarchical multitasking term embedded learning for synonym prediction
CN112131404B (en) Entity alignment method in four-risk one-gold domain knowledge graph
CN109635109B (en) Sentence classification method based on LSTM and combined with part-of-speech and multi-attention mechanism
CN109299216A (en) A kind of cross-module state Hash search method and system merging supervision message
CN110110122A (en) Image based on multilayer semanteme depth hash algorithm-text cross-module state retrieval
CN109800437A (en) A kind of name entity recognition method based on Fusion Features
CN109062897A (en) Sentence alignment method based on deep neural network
CN109062910A (en) Sentence alignment method based on deep neural network
CN112733866A (en) Network construction method for improving text description correctness of controllable image
CN113537384B (en) Hash remote sensing image retrieval method, device and medium based on channel attention
CN110598022B (en) Image retrieval system and method based on robust deep hash network
CN111460824A (en) Unmarked named entity identification method based on anti-migration learning
CN107480194B (en) Method and system for constructing multi-mode knowledge representation automatic learning model
CN113515632A (en) Text classification method based on graph path knowledge extraction
CN114896434B (en) Hash code generation method and device based on center similarity learning
CN116932722A (en) Cross-modal data fusion-based medical visual question-answering method and system
CN112732872A (en) Biomedical text-oriented multi-label classification method based on subject attention mechanism
CN111582506A (en) Multi-label learning method based on global and local label relation
CN114781382A (en) Medical named entity recognition system and method based on RWLSTM model fusion
CN116822534A (en) Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium
CN115329120A (en) Weak label Hash image retrieval framework with knowledge graph embedded attention mechanism
CN109033304B (en) Multi-modal retrieval method based on online deep topic model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant