CN117251583B - Text enhanced knowledge graph representation learning method and system based on local graph structure - Google Patents

Text enhanced knowledge graph representation learning method and system based on local graph structure Download PDF

Info

Publication number
CN117251583B
CN117251583B CN202311544608.6A CN202311544608A CN117251583B CN 117251583 B CN117251583 B CN 117251583B CN 202311544608 A CN202311544608 A CN 202311544608A CN 117251583 B CN117251583 B CN 117251583B
Authority
CN
China
Prior art keywords
vector
input
entity
representing
obtaining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202311544608.6A
Other languages
Chinese (zh)
Other versions
CN117251583A (en
Inventor
李志飞
蹇越
张龑
王时绘
张淼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hubei University
Original Assignee
Hubei University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hubei University filed Critical Hubei University
Priority to CN202311544608.6A priority Critical patent/CN117251583B/en
Publication of CN117251583A publication Critical patent/CN117251583A/en
Application granted granted Critical
Publication of CN117251583B publication Critical patent/CN117251583B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation
    • G06N5/022Knowledge engineering; Knowledge acquisition
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Abstract

The invention provides a text enhanced knowledge graph representation learning method based on a local graph structure, which comprises the following steps: s1: acquiring a knowledge graph of a local graph structure, and calculating to acquire a first input vector and a second input vector through the knowledge graph and an input triplet; s2: constructing a learning model, and training a first input vector and a second input vector through the learning model to obtain a first prediction vector and a second prediction vector; s3: constructing a scoring function through the first prediction vector and the second prediction vector, and calculating through the scoring function to obtain a loss function; s4: repeating the steps S1-S3 until the loss function is smaller than a preset value, and obtaining a trained learning model. The learning model constructed by the invention fuses the text coding and the local diagram structure information of the knowledge graph, thereby being beneficial to simultaneously acquiring the contextualization and the structured knowledge; by integrating the u-hop neighbor entity into the input vector, the problem of insufficient structural information in the text-based method is solved.

Description

Text enhanced knowledge graph representation learning method and system based on local graph structure
Technical Field
The invention relates to the technical field of knowledge maps, in particular to a text enhanced knowledge map representation learning method and system based on a local map structure.
Background
In recent years, knowledge patterns (KGs) have been widely used in various natural language processing fields. Significant efficacy has been demonstrated in search engines, knowledge questions and answers, recommendation systems, and other tasks. Various types of knowledge bases such as Freebase, wordNet, NELL have been created for storing structural information of human knowledge. Knowledge maps are mainly composed of two elements, which are described as (subject entities, relationships, guest entities) for storing knowledge. While representation learning of the knowledge-graph aims at mapping entities and relationships to a continuous, low-dimensional vector space, which facilitates the computation of the inference process between entities.
In order to obtain a high-level representation of a knowledge graph, the concept of projecting entities and relationships inherent in the knowledge graph onto different points in the vector space has been proposed in recent years. The former knowledge graph representation learning method is mainly a graph embedding method, and aims to acquire representations of elements in the graph in the form of low-dimensional vectors. But this approach ignores the context and text information that may be helpful in enhancing entity characterization learning. This deficiency makes them particularly vulnerable to the challenges presented by the imperfection of the drawing. These limitations greatly impair their ability to efficiently generalize and represent high quality.
For this reason, many studies have employed text encoding methods to obtain a representation. The text-based strategy utilizes supplemental input signals and contextual representations of natural language text associated with knowledge triples to enable missing component prediction in knowledge-graph representation learning. The text may contain text attributes of the entities and relationships, including their names or descriptions. By pre-training the language model, the text encoder can be seamlessly generalized to new graph elements and is not affected by the challenges presented by graph imperfection. However, the existing text-based knowledge graph representation learning method still has the problems of lack of structural knowledge and excessive reliance on semantic information due to insufficient efficacy of a text encoder in acquiring structural insights. This will affect the accuracy and performance of the representation learning.
Disclosure of Invention
In order to solve the technical problems, the invention provides a text enhanced knowledge graph representation learning method based on a local graph structure, which comprises the following steps:
s1: obtaining a knowledge graph of a local graph structure, and obtaining a first input vector V through calculation of the knowledge graph and an input triplet sr And a second input vector V o
S2: constructing a learning model, and obtaining a first input vector V through the learning model sr And a second input vector V o Training to obtain a first predictive vector e sr And a second predictive vector e o
S3: by a first predictive vector e sr And a second predictive vector e o Constructing a scoring function, and calculating to obtain a loss function through the scoring function;
s4: and repeating the steps S1-S3 until the loss function is smaller than a preset value, obtaining a trained learning model, and carrying out a link prediction task through the trained learning model.
Preferably, step S1 specifically includes:
s11: obtaining a knowledge graph of a local graph structure,/>Representing a set of all entities->Representing a set of all relationships->Representing the correct triplet set; by->Obtaining an input triplet (s, r, o), wherein s represents a subject entity embedding, r represents a relationship embedding, and o represents a guest entity embedding;
s12: obtaining a hop neighbor list L of a subject entity embedded in s s And text description T of subject entity s Obtaining a jump neighbor list L of an object entity with the object entity embedded in o o And text of guest entityDescription T o
S13: acquisition of L s U-hop neighbor entity L of middle body entity s [u]Will s, r, L s [u]And T s Combining to obtain a first input vector V sr
S14: acquisition of L o U-hop neighbor entity L of middle guest entity o [u]O, L o [u]And T o Combining to obtain a second input vector V o
Preferably:
first input vector V sr The calculation formula of (2) is as follows:
V sr =s+L s [u]+r+T s
second input vector V o The calculation formula of (2) is as follows:
V o =o+L o [u]+T o
preferably, step S2 specifically includes:
s21: to the first input vector V sr And a second input vector V o Input learning model for V sr And V o Word segmentation processing is carried out to obtain word vectors, and a first hidden layer vector is obtained through calculation of attention, relative position distance and semantic information among the word vectorsAnd a second hidden layer vector->
S22: by text encoder pairsAnd->Performing feature extraction forward propagation to obtain a first language model vector H sr And a second language model vector H o
S23: will H sr And H o Vector processing is carried out through pooling and regularization, and a first pre-preparation with unified dimension is obtainedMeasuring a vector e sr And a second predictive vector e o
Preferably, step S21 specifically includes:
s211: for V sr Word segmentation processing is carried out to obtain a first word vector set:
wherein token represents word segmentation processing, n represents the total number of first word vectors,representing an nth first word vector;
for V o Performing word segmentation processing to obtain a second word vector set:
where m represents the total number of second word vectors,representing an mth second word vector;
s212: the attention among the word vectors is obtained through calculation, and the calculation formula is as follows:
wherein H represents a semantic information hiding state associated with the word vector; q (Q) s 、K s And V s Respectively through projection matrix W qs 、W ks And W is vs Generated semantic vector matrix, K p And Q p Respectively through projection matrix W kp And W is qp A generated relative position vector matrix;
s213: the relative position distance is calculated and obtained, and the calculation formula is as follows:
wherein a and b are labels of a word vector, D (a, b) represents a relative position distance from label a to label b, and k represents a maximum relative position distance;
s214: calculating to obtain attention vector A from tag a to tag b a,b The calculation formula is as follows:
wherein,attention information representing semantics to semantics, +.>Represents Q s Line a, < >>Represent K s Line b, < >>Attention information representing semantics to relative position, < ->Represent K p Line D (a, b), +.>Attention information representing relative position to semantics, < ->Represents Q p Line D (a, b);
s215: constructing attention operation attention through attention vectors among all labels, and carrying out attention operation on all word vectors to obtain a first hidden layer vectorAnd a second hidden layer vector->The calculation formula is as follows:
preferably, the scoring function in step S3The expression of (2) is:
wherein d edu Representing the euclidean distance of the first predictive vector from the second predictive vector,representing cosine similarity of the first predictive vector and the second predictive vector,/and the like>Representing a reorder fraction obtained by reordering the predicted entities,/->Is the Euclidean distance weight.
A text-enhanced knowledge-graph representation learning system based on a local graph structure, comprising:
the input vector acquisition module is used for acquiring a knowledge graph of the local graph structure and acquiring a first input vector V through calculation of the knowledge graph and an input triplet sr And a second input vector V o
A predictive vector acquisition module for constructing a learning model by which the first input vector V is subjected to sr And a second input vector V o Training to obtain a first predictive vector e sr And a second predictive vector e o
A loss function calculation module for passing through the first prediction vector e sr And a second predictive vector e o Constructing a scoring function, and calculating to obtain a loss function through the scoring function;
and the link prediction module is used for repeatedly training the learning model until the loss function is smaller than a preset value, obtaining a trained learning model, and carrying out a link prediction task through the trained learning model.
The invention has the following beneficial effects:
the learning model constructed by the invention fuses the text coding and the local diagram structure information of the knowledge graph, thereby being beneficial to simultaneously acquiring the contextualization and the structured knowledge; the problem of insufficient structural information in a text-based method is solved by integrating the u-hop neighbor entity into the input vector; a structure enhancement mechanism combining the relative position distance is provided to offset prediction deviation caused by semantic difference, so that the finally obtained trained learning model can perform more accurate link prediction task.
Drawings
FIG. 1 is a flow chart of a method according to an embodiment of the present invention;
the achievement of the objects, functional features and advantages of the present invention will be further described with reference to the accompanying drawings, in conjunction with the embodiments.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
Referring to fig. 1, the invention provides a text enhanced knowledge graph representation learning method based on a local graph structure, which comprises the following steps:
s1: obtaining a knowledge graph of a local graph structure, and obtaining a first input vector V through calculation of the knowledge graph and an input triplet sr And a second input vector V o
S2: constructing a learning model, and obtaining a first input vector V through the learning model sr And a second input vector V o Training to obtain a first predictive vector e sr And a second predictive vector e o
S3: by a first predictive vector e sr And a second predictive vector e o Constructing a scoring function byCalculating to obtain a loss function;
s4: and repeating the steps S1-S3 until the loss function is smaller than a preset value, obtaining a trained learning model, and carrying out a link prediction task through the trained learning model.
Further, the step S1 specifically includes:
s11: obtaining a knowledge graph of a local graph structure,/>Representing a set of all entities->Representing a set of all relationships->Representing the correct triplet set; by->Obtaining an input triplet (s, r, o), wherein s represents a subject entity embedding, r represents a relationship embedding, and o represents a guest entity embedding;
s12: obtaining a hop neighbor list L of a subject entity embedded in s s And text description T of subject entity s Obtaining a jump neighbor list L of an object entity with the object entity embedded in o o Textual description T of a guest entity o
S13: acquisition of L s U-hop neighbor entity L of middle body entity s [u]Will s, r, L s [u]And T s Combining to obtain a first input vector V sr
S14: acquisition of L o U-hop neighbor entity L of middle guest entity o [u]O, L o [u]And T o Combining to obtain a second input vector V o
Specifically, the hop neighbor list L is obtained s And L o When the subject entity s and the object entity o in the input triplet (s, r, o) are excludedIn the neighbor list; this is because in one triplet (s, r, o), for the subject entity s and the guest entity o, they are 1-hop neighbors of each other; thus, if the subject entity s is merged to the u-hop neighbor entity L of the guest entity o o [u]Potentially causing a risk of tag leakage in the predicted outcome and vice versa.
Further:
first input vector V sr The calculation formula of (2) is as follows:
V sr =s+L s [u]+r+T s
second input vector V o The calculation formula of (2) is as follows:
V o =o+L o [u]+T o
specifically, when the hop count u of the u-hop neighbor entity is selected, u=1 is generally selected as the hop count, so that excessive neighbor nodes caused by excessive hop count selection can be avoided, and the acquired neighbor entity information dilutes the information of the original entity and causes the performance degradation of the model.
Further, a text encoder is adopted in the process of the step S2, and a specific decoupling attention mechanism is combined, namely, the attention calculation in the original self-attention mechanism is decoupled and divided into two parts of semantic information and relative position confidence, and the attention bias is calculated respectively, so that the relative position information among all the labels is fused into the text encoder;
the step S2 specifically comprises the following steps:
s21: to the first input vector V sr And a second input vector V o Input learning model for V sr And V o Word segmentation processing is carried out to obtain word vectors, and a first hidden layer vector is obtained through calculation of attention, relative position distance and semantic information among the word vectorsAnd a second hidden layer vector->
S22: by text editingEncoder pairAnd->Performing feature extraction forward propagation to obtain a first language model vector H sr And a second language model vector H o
S23: will H sr And H o Vector processing is carried out through pooling and regularization, and a first prediction vector e with unified dimension is obtained sr And a second predictive vector e o
Furthermore, in the embodiment of the invention, the text encoder end-to-end architecture integrating the relative position information, wherein in the input stage of the encoder, semantic information among original labels is utilized, the relative position information obtained by combining the relative position coding is utilized, and 5 attention offset vector matrixes are obtained through linear change and expressed as follows: k (K) p ,K s ,Q p ,Q s ,V s . And then the five attention bias information are added through a multi-head attention mechanism to calculate, so that the relative position information and the semantic information are successfully fused. And then performing text coding through forward propagation and aggregation and regularization operations of the encoder, finally, fusing absolute position information among tags, and transmitting the information to a next layer of encoder to perform the same coding operation, wherein the whole framework consists of N layers of encoder structures. Through the calculation of the encoder, the invention strengthens the extraction of the knowledge of the graph structure by fusing the relative position information, and lightens the prediction deviation caused by semantic ambiguity which is caused by only containing semantic information through a decoupled attention mechanism in the encoder, and the structure has obvious improvement on the performance of connection prediction;
the step S21 specifically includes:
s211: for V sr Word segmentation processing is carried out to obtain a first word vector set:
wherein token represents word segmentation processing, n represents the total number of first word vectors,representing an nth first word vector;
for V o Performing word segmentation processing to obtain a second word vector set:
where m represents the total number of second word vectors,representing an mth second word vector;
s212: the attention among the word vectors is obtained through calculation, and the calculation formula is as follows:
wherein H represents a semantic information hiding state associated with the word vector; q (Q) s 、K s And V s Respectively through projection matrix W qs 、W ks And W is vs Generated semantic vector matrix, K p And Q p Respectively through projection matrix W kp And W is qp A generated relative position vector matrix;
s213: the relative position distance is calculated and obtained, and the calculation formula is as follows:
wherein a and b are labels of a word vector, D (a, b) represents a relative position distance from label a to label b, and k represents a maximum relative position distance;
s214: calculating to obtain attention vector A from tag a to tag b a,b The calculation formula is as follows:
wherein,attention information representing semantics to semantics, +.>Represents Q s Line a, < >>Represent K s Line b, < >>Attention information representing semantics to relative position, < ->Represent K p Line D (a, b), +.>Attention information representing relative position to semantics, < ->Represents Q p Line D (a, b);
s215: constructing attention operation attention through attention vectors among all labels, and carrying out attention operation on all word vectors to obtain a first hidden layer vectorAnd a second hidden layer vector->The calculation formula is as follows:
furthermore, the invention adopts a double encoder structure, and uses the forward propagation of a text encoder and aggregation and regularization processes to carry outInformation extraction, step S22, of the first language model vector H sr And a second language model vector H o The generation process of (2) is as follows:
where the encoder represents the forward propagation and aggregation operations of the text encoder.
Furthermore, when vector processing is performed on the language model vectors, we firstly adopt an average pooling technology to pool the two vectors, reduce the dimension while retaining the main features, and then regularize the vectors through an L2 regularization technology to ensure that the dimensions of the two predicted vectors are uniform, and the first predicted vector e in step S23 sr And a second predictive vector e o The generation process of (2) is as follows:
wherein,representing pooling and regularization operations.
Furthermore, in order to obtain a more accurate triplet prediction result, the invention combines Euclidean distance on the basis of cosine function and introduces a reordering score, so that the strategy can ensure that the similarity score between two prediction vectors can be accurately calculated;
scoring function in step S3The expression of (2) is:
wherein d edu Representing the euclidean distance of the first predictive vector from the second predictive vector,representing cosine similarity of the first predictive vector and the second predictive vector,/and the like>Representing a reorder fraction obtained by reordering the predicted entities,/->Is the Euclidean distance weight.
Specifically, in order to obtain the spatial similarity of two prediction vectors more accurately, euclidean distance score is introduced, and is calculated as follows:
wherein, given an n-dimensional predictive vector e sr And e oAnd->Representing the projection of each of these two vectors along the respective dimensions, the Euclidean distance is:
in order to control the weights of Euclidean distances in the overall similarity score, we are in the scoring functionThe parameter controlling the Euclidean distance weight is also introduced>In addition, in order to enhance the local graph structure information and optimize the prediction performance, a direct reordering strategy is employed. This is because entities that are closer in distance are expected to have stronger relevance than entities that are farther apart;
the calculation formula of (2) is as follows:
the calculation formula of (2) is as follows:
wherein o is i Representing the calculated predicted guest entity if predicted guest entity o i U-hop neighbor list L of its corresponding subject entity s [u]In (c), then increasing a score assigned to the candidate entity. The scoring function can comprehensively calculate the scores of the two prediction vectors, and plays an important role in improving the prediction performance.
Further, in step S3, the loss function selects an InfoNCE loss function, which is expressed as follows:
wherein,representing additional intervals added during the calculation, +.>Temperature factor indicative of the negative influence of control difficulties during calculation +.>A score representing the negative sampling triples, D being the total number of negative sampling triples;
a text-enhanced knowledge-graph representation learning system based on a local graph structure, comprising:
the input vector acquisition module is used for acquiring a knowledge graph of the local graph structure and acquiring a first input vector V through calculation of the knowledge graph and an input triplet sr And a second input vector V o
A predictive vector acquisition module for constructing a learning model by which the first input vector V is subjected to sr And a second input vector V o Training to obtain a first predictive vector e sr And a second predictive vector e o
A loss function calculation module for passing through the first prediction vector e sr And a second predictive vector e o Constructing a scoring function, and calculating to obtain a loss function through the scoring function;
and the link prediction module is used for repeatedly training the learning model until the loss function is smaller than a preset value, obtaining a trained learning model, and carrying out a link prediction task through the trained learning model.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or system that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or system. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or system that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments. In the unit claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The use of the terms first, second, third, etc. do not denote any order, but rather the terms first, second, third, etc. are used to interpret the terms as labels.
The foregoing description is only of the preferred embodiments of the present invention, and is not intended to limit the scope of the invention, but rather is intended to cover any equivalents of the structures or equivalent processes disclosed herein or in the alternative, which may be employed directly or indirectly in other related arts.

Claims (4)

1. A text enhanced knowledge graph representation learning method based on a local graph structure is characterized by comprising the following steps:
s1: obtaining a knowledge graph of a local graph structure, and obtaining a first input vector V through calculation of the knowledge graph and an input triplet sr And a second input vector V o
S2: constructing a learning model, and obtaining a first input vector V through the learning model sr And a second input vector V o Training to obtain a first predictive vector e sr And a second predictive vector e o
S3: by a first predictive vector e sr And a second predictive vector e o Constructing a scoring function, and calculating to obtain a loss function through the scoring function;
s4: repeating the steps S1-S3 until the loss function is smaller than a preset value, obtaining a trained learning model, and carrying out a link prediction task through the trained learning model;
the step S1 specifically comprises the following steps:
s11: obtaining a knowledge graph of a local graph structure,/>Representing a set of all entities->Representing a set of all relationships->Representing the correct triplet set; by->Obtaining an input triplet (s, r, o), wherein s represents a subject entity embedding, r represents a relationship embedding, and o represents a guest entity embedding;
s12: obtaining a hop neighbor list L of a subject entity embedded in s s And text description T of subject entity s Obtaining a jump neighbor list L of an object entity with the object entity embedded in o o Textual description T of a guest entity o
S13: acquisition of L s U-hop neighbor entity L of middle body entity s [u]Will s, r, L s [u]And T s Combining to obtain a first input vector V sr
S14: acquisition of L o U-hop neighbor entity L of middle guest entity o [u]O, L o [u]And T o Combining to obtain a second input vector V o
The step S2 specifically comprises the following steps:
s21: to the first input vector V sr And a second input vector V o Input learning model for V sr And V o Word segmentation processing is carried out to obtain word vectors, and a first hidden layer vector is obtained through calculation of attention, relative position distance and semantic information among the word vectorsAnd a second hidden layer vector->
S22: by text encoder pairsAnd->Performing feature extraction forward propagation to obtain a first language model vector H sr And a second language model vector H o
S23: will H sr And H o Vector processing is carried out through pooling and regularization, and a first prediction vector e with unified dimension is obtained sr And a second predictive vector e o
Scoring function in step S3The expression of (2) is:
wherein d edu Representing the euclidean distance of the first predictive vector from the second predictive vector,representing cosine similarity of the first predictive vector and the second predictive vector,/and the like>Representing a reorder fraction obtained by reordering the predicted entities,/->Is the Euclidean distance weight.
2. The text-enhanced knowledge graph representation learning method based on the partial graph structure of claim 1, wherein:
first input vector V sr The calculation formula of (2) is as follows:
V sr =s+L s [u]+r+T s
second input vector V o The calculation formula of (2) is as follows:
V o =o+L o [u]+T o
3. the text-enhanced knowledge graph representation learning method based on the partial graph structure according to claim 1, wherein the step S21 is specifically:
s211: for V sr Word segmentation processing is carried out to obtain a first word vector set:
wherein token represents word segmentation processing, n represents the total number of first word vectors,representing an nth first word vector;
for V o Performing word segmentation processing to obtain a second word vector set:
where m represents the total number of second word vectors,representing an mth second word vector;
s212: the attention among the word vectors is obtained through calculation, and the calculation formula is as follows:
wherein H represents a semantic information hiding state associated with the word vector; q (Q) s 、K s And V s Respectively through projection matrix W qs 、W ks And W is vs Generated semantic vector matrix, K p And Q p Respectively through projection matrix W kp And W is qp A generated relative position vector matrix;
s213: the relative position distance is calculated and obtained, and the calculation formula is as follows:
wherein a and b are labels of a word vector, D (a, b) represents a relative position distance from label a to label b, and k represents a maximum relative position distance;
s214: calculating to obtain attention vector A from tag a to tag b a,b The calculation formula is as follows:
wherein,attention information representing semantics to semantics, +.>Represents Q s Line a, < >>Represent K s In the row b of (c),attention information representing semantics to relative position, < ->Represent K p Line D (a, b), +.>Attention information representing relative position to semantics, < ->Represents Q p Line D (a, b);
s215: constructing attention operation attention through attention vectors among all labels, and carrying out attention operation on all word vectors to obtain a first hidden layer vectorAnd a second hiddenTibet vector->The calculation formula is as follows:
4. a text-enhanced knowledge graph representation learning system based on a local graph structure, comprising:
the input vector acquisition module is used for acquiring a knowledge graph of the local graph structure and acquiring a first input vector V through calculation of the knowledge graph and an input triplet sr And a second input vector V o
A predictive vector acquisition module for constructing a learning model by which the first input vector V is subjected to sr And a second input vector V o Training to obtain a first predictive vector e sr And a second predictive vector e o
A loss function calculation module for passing through the first prediction vector e sr And a second predictive vector e o Constructing a scoring function, and calculating to obtain a loss function through the scoring function;
the link prediction module is used for repeatedly training the learning model until the loss function is smaller than a preset value, obtaining a trained learning model, and carrying out a link prediction task through the trained learning model;
the workflow of the input vector acquisition module is as follows:
s11: obtaining a knowledge graph of a local graph structure,/>Representing a set of all entities->Representing allSet of relations->Representing the correct triplet set; by->Obtaining an input triplet (s, r, o), wherein s represents a subject entity embedding, r represents a relationship embedding, and o represents a guest entity embedding;
s12: obtaining a hop neighbor list L of a subject entity embedded in s s And text description T of subject entity s Obtaining a jump neighbor list L of an object entity with the object entity embedded in o o Textual description T of a guest entity o
S13: acquisition of L s U-hop neighbor entity L of middle body entity s [u]Will s, r, L s [u]And T s Combining to obtain a first input vector V sr
S14: acquisition of L o U-hop neighbor entity L of middle guest entity o [u]O, L o [u]And T o Combining to obtain a second input vector V o
The working flow of the prediction vector acquisition module is as follows:
s21: to the first input vector V sr And a second input vector V o Input learning model for V sr And V o Word segmentation processing is carried out to obtain word vectors, and a first hidden layer vector is obtained through calculation of attention, relative position distance and semantic information among the word vectorsAnd a second hidden layer vector->
S22: by text encoder pairsAnd->Performing feature extraction forward propagation to obtain a first language model vector H sr And a second language model vector H o
S23: will H sr And H o Vector processing is carried out through pooling and regularization, and a first prediction vector e with unified dimension is obtained sr And a second predictive vector e o
Scoring function in loss function calculation moduleThe expression of (2) is:
wherein d edu Representing the euclidean distance of the first predictive vector from the second predictive vector,representing cosine similarity of the first predictive vector and the second predictive vector,/and the like>Representing a reorder fraction obtained by reordering the predicted entities,/->Is the Euclidean distance weight.
CN202311544608.6A 2023-11-20 2023-11-20 Text enhanced knowledge graph representation learning method and system based on local graph structure Active CN117251583B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311544608.6A CN117251583B (en) 2023-11-20 2023-11-20 Text enhanced knowledge graph representation learning method and system based on local graph structure

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311544608.6A CN117251583B (en) 2023-11-20 2023-11-20 Text enhanced knowledge graph representation learning method and system based on local graph structure

Publications (2)

Publication Number Publication Date
CN117251583A CN117251583A (en) 2023-12-19
CN117251583B true CN117251583B (en) 2024-01-26

Family

ID=89135505

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311544608.6A Active CN117251583B (en) 2023-11-20 2023-11-20 Text enhanced knowledge graph representation learning method and system based on local graph structure

Country Status (1)

Country Link
CN (1) CN117251583B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110334219A (en) * 2019-07-12 2019-10-15 电子科技大学 The knowledge mapping for incorporating text semantic feature based on attention mechanism indicates learning method
CN111191044A (en) * 2019-12-25 2020-05-22 湖北大学 Knowledge extraction and fusion method based on big data
CN112035672A (en) * 2020-07-23 2020-12-04 深圳技术大学 Knowledge graph complementing method, device, equipment and storage medium
CN113051408A (en) * 2021-03-30 2021-06-29 电子科技大学 Sparse knowledge graph reasoning method based on information enhancement
WO2022011681A1 (en) * 2020-07-17 2022-01-20 国防科技大学 Method for fusing knowledge graph based on iterative completion
CN116186216A (en) * 2022-12-30 2023-05-30 福州大学 Question generation method and system based on knowledge enhancement and double-graph interaction
CN116415661A (en) * 2023-03-20 2023-07-11 上海大学 Generalized relation prediction method integrating subgraph enhancement and relation semantics and application

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220164683A1 (en) * 2020-11-25 2022-05-26 Fmr Llc Generating a domain-specific knowledge graph from unstructured computer text

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110334219A (en) * 2019-07-12 2019-10-15 电子科技大学 The knowledge mapping for incorporating text semantic feature based on attention mechanism indicates learning method
CN111191044A (en) * 2019-12-25 2020-05-22 湖北大学 Knowledge extraction and fusion method based on big data
WO2022011681A1 (en) * 2020-07-17 2022-01-20 国防科技大学 Method for fusing knowledge graph based on iterative completion
CN112035672A (en) * 2020-07-23 2020-12-04 深圳技术大学 Knowledge graph complementing method, device, equipment and storage medium
CN113051408A (en) * 2021-03-30 2021-06-29 电子科技大学 Sparse knowledge graph reasoning method based on information enhancement
CN116186216A (en) * 2022-12-30 2023-05-30 福州大学 Question generation method and system based on knowledge enhancement and double-graph interaction
CN116415661A (en) * 2023-03-20 2023-07-11 上海大学 Generalized relation prediction method integrating subgraph enhancement and relation semantics and application

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
:learning knowledge graph embedding with heterogeneous relation attention networks:bing;LI Z;IEEE;全文 *
基于表示学习的知识图谱推理研究综述;李志飞 等;计算机科学;全文 *
融合对抗训练的端到端知识三元组联合抽取;黄培馨;赵翔;方阳;朱慧明;肖卫东;;计算机研究与发展(12);全文 *

Also Published As

Publication number Publication date
CN117251583A (en) 2023-12-19

Similar Documents

Publication Publication Date Title
CN110781683B (en) Entity relation joint extraction method
CN111985369B (en) Course field multi-modal document classification method based on cross-modal attention convolution neural network
CN110414432B (en) Training method of object recognition model, object recognition method and corresponding device
CN112015868B (en) Question-answering method based on knowledge graph completion
CN111666406B (en) Short text classification prediction method based on word and label combination of self-attention
CN113313232B (en) Functional brain network classification method based on pre-training and graph neural network
CN112733866A (en) Network construction method for improving text description correctness of controllable image
Ling et al. Solving optimization problems through fully convolutional networks: An application to the traveling salesman problem
CN109447096B (en) Glance path prediction method and device based on machine learning
CN112766507B (en) Complex problem knowledge base question-answering method based on embedded and candidate sub-graph pruning
CN116932722A (en) Cross-modal data fusion-based medical visual question-answering method and system
CN113065409A (en) Unsupervised pedestrian re-identification method based on camera distribution difference alignment constraint
CN114780766A (en) Fine-grained image-text retrieval method and system based on Transformer model
CN114612767A (en) Scene graph-based image understanding and expressing method, system and storage medium
CN113297369A (en) Intelligent question-answering system based on knowledge graph subgraph retrieval
CN113642602B (en) Multi-label image classification method based on global and local label relation
CN111259938B (en) Manifold learning and gradient lifting model-based image multi-label classification method
CN117251583B (en) Text enhanced knowledge graph representation learning method and system based on local graph structure
CN116681128A (en) Neural network model training method and device with noisy multi-label data
CN116310545A (en) Cross-domain tongue image classification method based on depth layering optimal transmission
CN116187349A (en) Visual question-answering method based on scene graph relation information enhancement
CN115934966A (en) Automatic labeling method based on remote sensing image recommendation information
CN116403608A (en) Speech emotion recognition method based on multi-label correction and space-time collaborative fusion
CN116704272A (en) Text embedded representation method and system for medical vision-language multi-modal task
Li et al. Structured neural motifs: Scene graph parsing via enhanced context

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant