CN105630901A - Knowledge graph representation learning method - Google Patents

Knowledge graph representation learning method Download PDF

Info

Publication number
CN105630901A
CN105630901A CN201510961791.9A CN201510961791A CN105630901A CN 105630901 A CN105630901 A CN 105630901A CN 201510961791 A CN201510961791 A CN 201510961791A CN 105630901 A CN105630901 A CN 105630901A
Authority
CN
China
Prior art keywords
vector
relation
entity
characteristic
tlv triple
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510961791.9A
Other languages
Chinese (zh)
Inventor
孙茂松
林衍凯
刘知远
栾焕博
刘奕群
马少平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201510961791.9A priority Critical patent/CN105630901A/en
Publication of CN105630901A publication Critical patent/CN105630901A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/288Entity relationship models

Abstract

The invention discloses a knowledge graph representation learning method. The method comprises the following steps: defining the correlation between entity vectors and relation vectors in a relation triple (head, relation, tail) by utilizing a translation-based model between the entity vector and the relation vector; defining the correlation between entity vectors and feature vectors in a feature triple (entity, attribute, value) by utilizing a neural network classification model; and correlating the entity vectors, the relation vectors and the feature vectors with one another through an evaluation function, and minimizing the evaluation function to learn the entity vectors, the relation vectors and the feature vectors so as to achieve the optimization aim. By adopting the method disclosed in the invention, the relation among the entity, the relation and the feature can be accurately represented.

Description

A kind of knowledge mapping represents learning method
Technical field
The present invention relates to natural language processing and knowledge mapping field, represent learning method particularly to a kind of knowledge mapping.
Background technology
Along with society develops rapidly, we come into the information explosion epoch, and the entity that magnanimity is new all can be had every day to produce with information. The Internet is as current information acquisition platform the most easily, and user is day by day urgent with the demand concluded to effective information screening, and how obtaining valuable information from mass data becomes a difficult problem. Knowledge mapping arises at the historic moment in this.
Knowledge mapping the proper noun such as all personages, place name, title, team's name and things will be expressed as entity in the world, internal relation between entity is expressed as relation, it is intended to be expressed as between entity using the mass knowledge in data base utilizing relation as the ternary relation group of bridge. Such as, Beijing is this knowledge of capital of China, then utilizes tlv triple relation (Beijing is ... capital, China) to be indicated in knowledge mapping. The different semantemes of one key word can be set up different entities by knowledge mapping, eliminate the interference of language polysemy, deeper wider array of internal relation between target entity and other related entities can be represented simultaneously, be widely used in multiple fields such as data mining, information retrieval, question answering systems. Owing to existing entity is huge, and entity and relation every day all change with increase, it is difficult to manual maintenance with excavate new information, be study hotspot important now to expression and the auto-complete of knowledge mapping.
Knowledge mapping physical quantities is huge, and network structure is openness seriously. And in the research of knowledge mapping, achieve significant progress in recent years, represent that study applies in knowledge mapping, by in all entities and relationship map to low-dimensional vector row space, the openness and efficiency produced during knowledge mapping study before solving. But, current existing knowledge mapping represents that the characteristic of the relation between entity and entity is lumped together by learning method when study, uses same model to be modeled, it is impossible to accurate presentation-entity, connecting each other between relation and characteristic.
Summary of the invention
In view of this, the goal of the invention of the present invention is: solve prior art exists cannot Precise Representation entity, the problem of contact between relation and characteristic, to improve the quality that knowledge mapping represents.
For reaching above-mentioned purpose, technical scheme is specifically achieved in that
The invention provides a kind of knowledge mapping and represent learning method, the method includes: utilize the model based on translation between entity vector and relation vector, in definition relation tlv triple (head, relation, tail) entity vector and relation vector between interrelated; Utilize neural network classification model, in defined property tlv triple (entity, attribute, value) entity vector and eigen vector between interrelated; By evaluation function, entity vector, relation vector and eigen vector are associated, and minimize evaluation function, with learning object vector, relation vector and eigen vector, reach optimization aim.
As seen from the above technical solutions, relation and characteristic are distinguished and are treated by the present invention, so the form that knowledge mapping adopts relation tlv triple and characteristic tlv triple represents knowledge. Therefore, the present invention adopts the model between entity vector and relation vector based on translation, and what represent between the entity vector sum relation vector in relation tlv triple is interrelated; Adopting neural network classification model, that comes between the entity vector sum eigen vector in characterization tlv triple is interrelated; Then pass through evaluation function entity vector, relation vector and eigen vector to be associated, and optimize this evaluation function, when reaching optimization aim, each entity vector, relation vector and the eigen vector that just can simultaneously learn in knowledge mapping, thus accurate presentation-entity, connecting each other between relation and characteristic.
Accompanying drawing explanation
Fig. 1 is the exemplary plot of inclusion relation tlv triple and characteristic tlv triple in knowledge mapping.
Fig. 2 is the schematic flow sheet that knowledge mapping of the present invention represents learning method.
According to prior art knowledge mapping, Fig. 3 a represents that the triple table that learning method obtains advises the exemplary plot of knowledge.
According to knowledge mapping of the present invention, Fig. 3 b represents that the triple table that learning method obtains advises the exemplary plot of knowledge.
Detailed description of the invention
For making the purpose of the present invention, technical scheme and advantage clearly understand, developing simultaneously embodiment referring to accompanying drawing, the present invention is described in more detail.
In prior art, and do not differentiate between relation and characteristic, characteristic is also served as the one of relation, knowledge mapping mainly adopts the form of (entity 1, relation, entity 2) tlv triple to represent knowledge, namely relation tlv triple (head, relation, tail) is adopted to represent. What therefore represent between the entity vector sum relation vector in relation tlv triple in prior art only with a kind of model is interrelated, knowledge mapping represents that the characteristic of the relation between entity and entity be cannot be distinguished by out by learning method when study, it is impossible to accurate presentation-entity, connecting each other between relation and characteristic.
Relation and characteristic are distinguished and are treated by the present invention, so the form that knowledge mapping adopts relation tlv triple and characteristic tlv triple represents knowledge. Relation tlv triple represents with (head, relation, tail), and relation is used for connecting two entities, portrays the association between two entities. Characteristic tlv triple represents with (entity, attribute, value), and (a v) is used for portraying the intrinsic characteristic of correspondent entity to each characteristic-value. In knowledge mapping, relation tlv triple interior joint presentation-entity, even limit represents relation; Connecting limit characterization in characteristic tlv triple, even one end node presentation-entity on limit, even the other end node on limit represents the characteristic value of this entity. Fig. 1 is the exemplary plot of inclusion relation tlv triple and characteristic tlv triple in knowledge mapping. Wherein, the node " Clinton " that circle represents and " Hillary " are entity, Lian Bianwei " wife " relation between them. It addition, it will be seen that two entities are each with one's own characteristic, such as " occupation ", " sex ", " birthplace " etc., characteristic value for entity " Clinton " occupation is US President, and the characteristic value for entity " Hillary " occupation is U.S. Secretary of State.
Embodiment one
The invention discloses a kind of knowledge mapping and represent learning method, its schematic flow sheet is as in figure 2 it is shown, the method includes:
Step 21, utilize the model based on translation between entity vector and relation vector, interrelated between entity vector and relation vector in definition relation tlv triple (head, relation, tail).
Wherein, utilizing the model based on translation between entity vector and relation vector, in definition relation tlv triple, the method for being mutually related between entity vector and relation vector includes:
S211, definition relation tlv triple probability are p ( h | r , t , X ) = exp ( g ( h , r , t ) ) Σ h ‾ exp ( g ( h ‾ , r , t ) ) ;
Represent any entity in knowledge mapping;It is the normalization factor of relation tlv triple probability function, refers to all entity h in traversal knowledge mapping so that normalization factor is 1. p ( h | r , t , X ) = exp ( g ( h , r , t ) ) Σ h ‾ exp ( g ( h ‾ , r , t ) ) It it is softmax function.
S212, utilize entity vector and relation vector between based on translation model, definition measurement relation r and entity to (h, the function g connected each other between t).
Between entity vector and relation vector, the model based on translation can have multiple, for instance, TransE and TransR etc., if what adopt is the energy function of TransE, then g can be defined as:
G (h, r, t)=-| | h+r-t | |L1/L2+b1
If what adopt is the energy function of TransR, then g can be defined as:
G (h, r, t)=-| | hMr+r-tMr||L1/L2+b1
Wherein, L1 is L1 normal form, and L2 is L2 normal form, MrFor projection matrix relevant to relation in TransR model, b1It is an offset constant, for making the average of g function return value be maintained at about 0.
It should be noted that relation tlv triple probability tables is shown as p (h | r, t, X) by the embodiment of the present invention, it is also possible to replace with p (t | r, h, X) or p (r | h, t, X). X is the vector representation of r, h, t.
Step 22, utilize neural network classification model, interrelated between entity vector and eigen vector in defined property tlv triple (entity, attribute, value).
Wherein, utilizing neural network classification model, in defined property tlv triple, the method for being mutually related between entity vector and eigen vector includes:
S221, define the first characteristic tlv triple probability and be p ( v | e , a , X ) = exp ( k ( e , a , v ) ) Σ e ‾ exp ( k ( e ‾ , a , v ) ) ;
Represent any entity in knowledge mapping;It is the normalization factor of the first characteristic tlv triple probability function, refers to all entity e in traversal knowledge mapping so that normalization factor is 1. p ( v | e , a , X ) = exp ( k ( e , a , v ) ) Σ e ‾ exp ( k ( e ‾ , a , v ) ) It it is softmax function.
S222, utilize neural network classification model, definition weigh characteristic-value (a, v) and the function k connected each other between entity e.
When neural network classification model is monolayer neural networks model, k (e, a, v)=-| | f (eWa+ba)-Vav||L1/L2+b2; WaAnd baFor the model parameter in monolayer neural networks model; | | f (eWa+ba)-Vav||L1/L2Represent representing entity e in the vectorial subspace being projected to corresponding characteristic a by the neural network model of a monolayer, then calculate the similarity represented between vector of the vector sum correspondence characteristic value v after projection; b2It is an offset constant, for making the average of k function return value be maintained at about 0.
Step 23, by evaluation function, entity vector, relation vector and eigen vector are associated, and minimize evaluation function, with learning object vector, relation vector and eigen vector, reach optimization aim.
Specifically include:
S231, definition evaluation function are O (X)=log (P (S, Y | X))+�� C (X);
S232, minimizing described evaluation function, study obtains each entity vector, relation vector and eigen vector in knowledge mapping. The method minimizing evaluation function can have multiple, it is possible to the method adopting stochastic gradient descent, etc.
S represents the set of all relation tlv triple in knowledge mapping, and Y represents the set of all characteristic tlv triple in knowledge mapping, and P (S, Y | X) represent the product of all relation tlv triple probability and all characteristic tlv triple probability in knowledge mapping; Characteristic tlv triple probability is the first characteristic tlv triple probability;
�� is hyper parameter, for controlling the weight of penalty term; C (X) is penalty, is used for preventing parameter learning over-fitting, and penalty C (X) definition is as follows:
C ( X ) = Σ e ∈ E [ | | e | | - 1 ] + + Σ r ∈ R [ | | r | | - 1 ] + + Σ e ∈ E Σ i [ | | eW i + b i | | - 1 ] + + Σ i [ | | V i | | - 1 ] + , Wherein, [x]+=max (0, x) represent that an input is for x, return value is the function of number bigger between 0 and x.
It should be noted that the process minimizing evaluation function is exactly the process reaching optimization aim. If the g in relation tlv triple probability function, what adopt is TransE model, then minimize in the process of evaluation function, by constantly adjusting h, r and the t vector of relation and tail (head), make each (h+r) in P (S | X) equal with t as far as possible, i.e. h+r=t. If the k in the first characteristic tlv triple probability function, what adopt is monolayer neural networks model, then minimize in the process of evaluation function, by constantly adjusting e, a and v (entity, the vector of attribute and value) so that in P (Y | X), the value v probability of each entity e correspondence characteristic a is 100%.
Thus, study obtains each entity vector, relation vector and eigen vector in knowledge mapping. According to prior art knowledge mapping, Fig. 3 a represents that the triple table that learning method obtains advises the exemplary plot of knowledge. According to knowledge mapping of the present invention, Fig. 3 b represents that the triple table that learning method obtains advises the exemplary plot of knowledge. In Fig. 3 a, not differentiating between relation and the characteristic of entity, still using the characteristic one as relation, knowledge mapping adopts the form of (entity 1, relation, entity 2) tlv triple to represent knowledge. And in Fig. 3 b, distinguish relation and the characteristic of entity, knowledge mapping adopts the form of relation tlv triple and characteristic tlv triple to represent knowledge. It can be seen that e6, e7, e8 and e9 are characteristic value from Fig. 3 b, e6, e7 belong to the value of a kind of characteristic A1, and e8 and e9 belongs to the value of another kind of characteristic A2, and specifically, in characteristic tlv triple, the value of entity e1 correspondence characteristic A1 is e6; The value of entity e2 correspondence characteristic A1 is e6; The value of entity e3 correspondence characteristic A1 is e7; The value of entity e3 correspondence characteristic A2 is e8; The value of entity e5 correspondence characteristic A2 is e8; The value of entity e4 correspondence characteristic A2 is e9. In simul relation tlv triple, the relation of entity e1 and entity e3 is r1; The relation of entity e1 and entity e2 is r5; The relation of entity e2 and entity e4 is r4; The relation of entity e3 and entity e4 is r3; The relation of entity e3 and entity e5 is r2; The relation of entity e4 and entity e5 is r4. It can thus be seen that the prior art of Fig. 3 a is compared with the present invention of Fig. 3 b, the knowledge mapping of the present invention represents learning method, it is possible to accurately represent connecting each other between entity, relation and characteristic.
Embodiment two
Treat owing to the knowledge mapping of the present invention represents that relation and characteristic are distinguished by learning method, it is possible to further consider connecting each other between characteristic.
The knowledge mapping of the embodiment of the present invention two represents that learning method comprises the following steps:
Step 31, utilize the model based on translation between entity vector and relation vector, interrelated between entity vector and relation vector in definition relation tlv triple (head, relation, tail).
Wherein, utilizing the model based on translation between entity vector and relation vector, in definition relation tlv triple, the method for being mutually related between entity vector and relation vector includes:
S311, definition relation tlv triple probability are p ( h | r , t , X ) = exp ( g ( h , r , t ) ) Σ h ‾ exp ( g ( h ‾ , r , t ) ) ; Represent any entity in knowledge mapping;
S312, utilize entity vector and relation vector between based on translation model, definition measurement relation r and entity to (h, the function g connected each other between t).
Between entity vector and relation vector, the model based on translation can have multiple, for instance, TransE and TransR etc., if what adopt is the energy function of TransE, then g can be defined as:
G (h, r, t)=-| | h+r-t | |L1/L2+b1
If what adopt is the energy function of TransR, then g can be defined as:
G (h, r, t)=-| | hMr+r-tMr||L1/L2+b1
Wherein, L1 is L1 normal form, and L2 is L2 normal form, MrFor projection matrix relevant to relation in TransR model, b1It is an offset constant, for making the average of g function return value be maintained at about 0.
It should be noted that relation tlv triple probability tables is shown as p (h | r, t, X) by the embodiment of the present invention, it is also possible to replace with p (t | r, h, X) or p (r | h, t, X). X is the vector representation of r, h, t.
Step 32, utilize neural network classification model, interrelated between entity vector and eigen vector in defined property tlv triple (entity, attribute, value).
Wherein, utilizing neural network classification model, in defined property tlv triple, the method for being mutually related between entity vector and eigen vector includes:
S321, define the first characteristic tlv triple probability and be p ( v | e , a , X ) = exp ( k ( e , a , v ) ) Σ e ‾ exp ( k ( e ‾ , a , v ) ) ; Represent any entity in knowledge mapping.
S322, utilize neural network classification model, definition weigh characteristic-value (a, v) and the function k connected each other between entity e.
When neural network classification model is monolayer neural networks model, k (e, a, v)=-| | f (eWa+ba)-Vav||L1/L2+b2; WaAnd baFor the model parameter in monolayer neural networks model; | | f (eWa+ba)-Vav||L1/L2Represent representing entity e in the vectorial subspace being projected to corresponding characteristic a by the neural network model of a monolayer, then calculate the similarity represented between vector of the vector sum correspondence characteristic value v after projection; b2It is an offset constant, for making the average of k function return value be maintained at about 0.
Step 323,
Define the second characteristic tlv triple Probability p ((e, a, v) | X) �� p (v | e, a, X) p (v | e, a, Y (e)); Wherein, p ( v | e , a , Y ( e ) ) = exp ( z ( e , a , v , Y ( e ) ) ) Σ v ‾ ∈ V a exp ( z ( e , a , v ‾ , Y ( e ) ) ) , Y (e) for entity e in knowledge mapping except characteristic-value (a, all known characteristics outside v); p ( v | e , a , Y ( e ) ) = exp ( z ( e , a , v , Y ( e ) ) ) Σ v ‾ ∈ V a exp ( z ( e , a , v ‾ , Y ( e ) ) ) It it is softmax function.
Step 324, definition weigh characteristic-value (a, v) and the function z connected each other between other characteristic-values. Assume characteristic-value (the combination positive correlation of every other characteristic-value in a, conditional probability v) and knowledge mapping, it is defined as: z ( e , a , v , Y ( e ) ) ∝ Σ ( e , a ‾ , v ‾ ) ∈ Y ( e ) P ( ( a , v ) | ( a ‾ , v ‾ ) ) ( A a · A a ‾ ) ; Wherein,For AaWithInner product, for weighing degree of contact between two characteristics;Characteristic is had for known entitiesTime (a, conditional probability v).
Step 33, by evaluation function, entity vector, relation vector and eigen vector are associated, and minimize evaluation function, with learning object vector, relation vector and eigen vector, reach optimization aim.
Specifically include:
S331, definition evaluation function are O (X)=log (P (S, Y | X))+�� C (X);
S232, minimizing described evaluation function, study obtains each entity vector, relation vector and eigen vector in knowledge mapping.
S represents the set of all relation tlv triple in knowledge mapping, and Y represents the set of all characteristic tlv triple in knowledge mapping, and P (S, Y | X) represent the product of all relation tlv triple probability and all characteristic tlv triple probability in knowledge mapping; Characteristic tlv triple probability is the second characteristic tlv triple probability;
�� is hyper parameter, for controlling the weight of penalty term; C (X) is penalty, is used for preventing parameter learning over-fitting, and penalty C (X) definition is as follows:
C ( X ) = Σ e ∈ E [ | | e | | - 1 ] + + Σ r ∈ R [ | | r | | - 1 ] + + Σ e ∈ E Σ i [ | | eW i + b i | | - 1 ] + + Σ i [ | | V i | | - 1 ] + , Wherein, [x]+=max (0, x) represent that an input is for x, return value is the function of number bigger between 0 and x.
Further, calculating softmax function owing to there being many places to need in former evaluation function, wherein the amount of calculation of normalization item is very big, greatly reduces the speed of algorithm, the embodiment of the present invention is before minimizing evaluation function, it is preferred to use softmax function is similar to by negative sampling algorithm.
When including relation tlv triple Probability p (h | r, t, X) and the first characteristic tlv triple probability when evaluation function, softmax function therein is converted into formula calculated below by negative sampling by us:
p ( h | r , t , X ) = Π ( h , r , t ) ∈ S [ σ ( g ( h , r , t ) ) Π i = 1 C E ( h i , r , t ) ~ P ( S - ) σ ( g ( h i , r , t ) ) ] ;
p ( v | e , a , X ) = Π ( h , r , t ) ∈ Y [ σ ( k ( e , a , v ) ) Π i = 1 C E ( e , a , v i ) ~ P ( Y - ) σ ( k ( e , a , v i ) ) ] .
When evaluation function include p (h | r, t, X) and the second characteristic tlv triple Probability p ((e, a, v) | X) �� p (v | e, a, X) p (v | e, a, Y (e)) time, softmax function therein is converted into formula calculated below by negative sampling by us:
p ( h | r , t , X ) = Π ( h , r , t ) ∈ S [ σ ( g ( h , r , t ) ) Π i = 1 C E ( h i , r , t ) ~ P ( S - ) σ ( g ( h i , r , t ) ) ] ;
p ( v | e , a , X ) = Π ( h , r , t ) ∈ Y [ σ ( k ( e , a , v ) ) Π i = 1 C E ( e , a , v i ) ~ P ( Y - ) σ ( k ( e , a , v i ) ) ] ;
p ( v | e , a , Y ( e ) ) = Π ( h , r , t ) ∈ Y [ σ ( z ( e , a , v , Y ( e ) ) Π i = 1 C E ( e , a , v i ) ~ P ( Y - ) σ ( z ( e , a , v i , Y ( e ) ) ] .
Obviously, above-mentioned formula p ( h | r , t , X ) = Π ( h , r , t ) ∈ S [ σ ( g ( h , r , t ) ) Π i = 1 C E ( h i , r , t ) ~ P ( S - ) σ ( g ( h i , r , t ) ) ] , Can also be replaced by:
p ( r | h , t , X ) = Π ( h , r , t ) ∈ S [ σ ( g ( h , r , t ) ) Π i = 1 C E ( h i , r , t ) ~ P ( S - ) σ ( g ( h , r i , t ) ) ] ; Or,
p ( h | r , t , X ) = Π ( h , r , t ) ∈ S [ σ ( g ( h , r , t ) ) Π i = 1 C E ( h i , r , t ) ~ P ( S - ) σ ( g ( h i , r , t ) ) ] .
��=1/ (1+exp (-x)) is sogmoid function, S-For the negative example set of relation tlv triple, P (S-) bear the probability function of all elements, Y in example set for relation tlv triple-For the negative example set of characteristic tlv triple, P (Y-) bear the probability function of all elements in example set for characteristic tlv triple.
Relation tlv triple bears example set S-Producing method as follows: we are by each tlv triple (head in positive example set S, relation, tail) random with arbitrarily replacing head entity head for his entity or random with arbitrarily replacing tail entity tail for his entity or random with arbitrarily replacing head entity relation for his relation, so we will produce a new tlv triple not appeared in positive example set, we are treated as a negative example, and the combination of all this negative examples just constitutes S-��
Same, characteristic tlv triple bears example set Y-Producing method as follows: we are by each tlv triple (entity in positive example set Y, attribute, it is value) random that with arbitrarily replacing existing characteristic value value for his possible characteristic value, so we will produce a new tlv triple not appeared in positive example set, we are treated as a negative example, and the combination of all this negative examples just constitutes Y-��
To sum up, the invention has the beneficial effects as follows: compared with prior art, the while that the present invention proposing, in learning knowledge collection of illustrative plates, the knowledge mapping of entity, relation and personality presentation represents learning method, what solve to exist in prior art cannot Precise Representation entity, the problem of contact between relation and characteristic, to improve the quality that knowledge graph represents, there is good practicality.
The foregoing is only presently preferred embodiments of the present invention, be not intended to limit protection scope of the present invention. All any amendment of making, equivalent replace and improvement etc. within the spirit and principles in the present invention, should be included within protection scope of the present invention.

Claims (7)

1. knowledge mapping represents a learning method, and the method includes:
Utilize the model based on translation between entity vector and relation vector, interrelated between entity vector and relation vector in definition relation tlv triple (head, relation, tail);
Utilize neural network classification model, in defined property tlv triple (entity, attribute, value) entity vector and eigen vector between interrelated;
By evaluation function, entity vector, relation vector and eigen vector are associated, and minimize evaluation function, with learning object vector, relation vector and eigen vector, reach optimization aim.
2. the method for claim 1, it is characterised in that
Utilizing the model based on translation between entity vector and relation vector, in definition relation tlv triple, the method for being mutually related between entity vector and relation vector includes:
Definition relation tlv triple probability is Represent any entity in knowledge mapping;
Utilizing the model based on translation between entity vector and relation vector, definition measurement relation r and entity are to (h, the function g connected each other between t).
3. method as claimed in claim 2, it is characterised in that
Utilizing neural network classification model, in defined property tlv triple, the method for being mutually related between entity vector and eigen vector includes:
Defining the first characteristic tlv triple probability is Represent any entity in knowledge mapping;
Utilize neural network classification model, definition weigh characteristic-value (a, v) and the function k connected each other between entity e.
4. method as claimed in claim 3, it is characterised in that when neural network classification model is monolayer neural networks model, k (e, a, v)=-| | f (eWa+ba)-Vav||L1/L2+b2;
WaAnd baFor the model parameter in monolayer neural networks model; | | f (eWa+ba)-Vav||L1/L2Represent representing entity e in the vectorial subspace being projected to corresponding characteristic a by the neural network model of a monolayer, then calculate the similarity represented between vector of the vector sum correspondence characteristic value v after projection; b2It is an offset constant, for making the average of k function return value be maintained at about 0.
5. method as claimed in claim 4, it is characterised in that after defined function k, the method farther includes:
Define the second characteristic tlv triple Probability p ((e, a, v) | X) �� p (v | e, a, X) p (v | e, a, Y (e)); Wherein,Y (e) for entity e in knowledge mapping except characteristic-value (a, all known characteristics outside v);
Definition weigh characteristic-value (a, v) and the function z connected each other between other characteristic-values; Assume characteristic-value (the combination positive correlation of every other characteristic-value in a, conditional probability v) and knowledge mapping, it is defined as:Wherein,For AaWithInner product, for weighing degree of contact between two characteristics;Characteristic is had for known entitiesTime (a, conditional probability v).
6. the method described in claim 5, it is characterised in that
Entity vector, relation vector and eigen vector being associated by evaluation function, and minimize evaluation function, with each entity vector, relation vector and eigen vector in learning knowledge collection of illustrative plates, the method reaching optimization aim includes:
Definition evaluation function is O (X)=log (P (S, Y | X))+�� C (X);
Minimizing described evaluation function, study obtains each entity vector, relation vector and eigen vector in knowledge mapping;
Wherein, S represents the set of all relation tlv triple in knowledge mapping, and Y represents the set of all characteristic tlv triple in knowledge mapping, and P (S, Y | X) represent the product of all relation tlv triple probability and all characteristic tlv triple probability in knowledge mapping; Characteristic tlv triple probability is the first characteristic tlv triple probability or the second characteristic tlv triple probability;
�� is hyper parameter, for controlling the weight of penalty term; C (X) is penalty, is used for preventing parameter learning over-fitting, and penalty C (X) definition is as follows:
Wherein, [x]+=max (0, x), it is a function returning number bigger between 0 and x.
7. method as claimed in claim 6, it is characterised in that before minimizing, adopts negative sampling algorithm, the softmax function in evaluation function is similar to, to accelerate pace of learning;
Softmax function
Softmax function
Softmax function
��=1/ (1+exp (-x)) is sogmoid function, S-For the negative example set of relation tlv triple, P (S-) bear the probability function of all elements, Y in example set for relation tlv triple-For the negative example set of characteristic tlv triple, P (Y-) bear the probability function of all elements in example set for characteristic tlv triple.
CN201510961791.9A 2015-12-21 2015-12-21 Knowledge graph representation learning method Pending CN105630901A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510961791.9A CN105630901A (en) 2015-12-21 2015-12-21 Knowledge graph representation learning method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510961791.9A CN105630901A (en) 2015-12-21 2015-12-21 Knowledge graph representation learning method

Publications (1)

Publication Number Publication Date
CN105630901A true CN105630901A (en) 2016-06-01

Family

ID=56045834

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510961791.9A Pending CN105630901A (en) 2015-12-21 2015-12-21 Knowledge graph representation learning method

Country Status (1)

Country Link
CN (1) CN105630901A (en)

Cited By (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106528609A (en) * 2016-09-28 2017-03-22 厦门理工学院 Vector constraint embedded transformation knowledge graph inference method
CN106528610A (en) * 2016-09-28 2017-03-22 厦门理工学院 Knowledge graph representation learning method based on path tensor decomposition
CN106844658A (en) * 2017-01-23 2017-06-13 中山大学 A kind of Chinese text knowledge mapping method for auto constructing and system
CN106909622A (en) * 2017-01-20 2017-06-30 中国科学院计算技术研究所 Knowledge mapping vector representation method, knowledge mapping relation inference method and system
CN106934042A (en) * 2017-03-16 2017-07-07 中国人民解放军国防科学技术大学 A kind of knowledge mapping represents model and its method
CN107330125A (en) * 2017-07-20 2017-11-07 云南电网有限责任公司电力科学研究院 The unstructured distribution data integrated approach of magnanimity of knowledge based graphical spectrum technology
CN107480191A (en) * 2017-07-12 2017-12-15 清华大学 A kind of entity alignment model of iteration
CN107590237A (en) * 2017-09-11 2018-01-16 桂林电子科技大学 A kind of knowledge mapping based on dynamic translation principle represents learning method
CN107783973A (en) * 2016-08-24 2018-03-09 慧科讯业有限公司 The methods, devices and systems being monitored based on domain knowledge spectrum data storehouse to the Internet media event
CN107871158A (en) * 2016-09-26 2018-04-03 清华大学 A kind of knowledge mapping of binding sequence text message represents learning method and device
CN107885760A (en) * 2016-12-21 2018-04-06 桂林电子科技大学 It is a kind of to represent learning method based on a variety of semantic knowledge mappings
CN107943874A (en) * 2017-11-13 2018-04-20 平安科技(深圳)有限公司 Knowledge mapping processing method, device, computer equipment and storage medium
CN108280061A (en) * 2018-01-17 2018-07-13 北京百度网讯科技有限公司 Text handling method based on ambiguity entity word and device
CN108388580A (en) * 2018-01-24 2018-08-10 平安医疗健康管理股份有限公司 Merge the dynamic knowledge collection of illustrative plates update method of medical knowledge and application case
CN108491491A (en) * 2018-03-14 2018-09-04 广州杰赛科技股份有限公司 The construction method and system of intelligent customer service knowledge mapping
CN108510110A (en) * 2018-03-13 2018-09-07 浙江禹控科技有限公司 A kind of water table trend analysis method of knowledge based collection of illustrative plates
CN108647258A (en) * 2018-01-24 2018-10-12 北京理工大学 A kind of expression learning method based on entity associated constraint
CN108664615A (en) * 2017-05-12 2018-10-16 华中师范大学 A kind of knowledge mapping construction method of discipline-oriented educational resource
CN108694201A (en) * 2017-04-10 2018-10-23 华为软件技术有限公司 A kind of entity alignment schemes and device
CN108876849A (en) * 2018-04-24 2018-11-23 哈尔滨工程大学 Deep learning target identification and localization method based on accessory ID
CN108885626A (en) * 2017-02-22 2018-11-23 谷歌有限责任公司 Optimize graph traversal
CN108959328A (en) * 2017-05-27 2018-12-07 株式会社理光 Processing method, device and the electronic equipment of knowledge mapping
CN109063032A (en) * 2018-07-16 2018-12-21 清华大学 A kind of noise-reduction method of remote supervisory retrieval data
CN109241290A (en) * 2017-07-10 2019-01-18 华东师范大学 A kind of knowledge mapping complementing method, device and storage medium
CN109376864A (en) * 2018-09-06 2019-02-22 电子科技大学 A kind of knowledge mapping relation inference algorithm based on stacking neural network
CN109508419A (en) * 2018-11-23 2019-03-22 成都品果科技有限公司 A kind of recommended method and system of knowledge based study
CN109634939A (en) * 2018-12-28 2019-04-16 中国农业银行股份有限公司 A kind of the determination method, apparatus and electronic equipment of missing values
CN109829057A (en) * 2019-01-11 2019-05-31 中山大学 A kind of knowledge mapping Entity Semantics spatial embedding method based on figure second order similitude
CN110147450A (en) * 2019-05-06 2019-08-20 北京科技大学 A kind of the knowledge complementing method and device of knowledge mapping
CN110188147A (en) * 2019-05-22 2019-08-30 厦门无常师教育科技有限公司 The document entity relationship of knowledge based map finds method and system
CN110413793A (en) * 2019-06-11 2019-11-05 福建奇点时空数字科技有限公司 A kind of knowledge mapping substance feature method for digging based on translation model
CN110569364A (en) * 2019-08-21 2019-12-13 北京大米科技有限公司 online teaching method, device, server and storage medium
CN110647620A (en) * 2019-09-23 2020-01-03 中国农业大学 Knowledge graph representation learning method based on confidence hyperplane and dictionary information
CN110688489A (en) * 2019-09-09 2020-01-14 中国电子科技集团公司电子科学研究院 Knowledge graph deduction method and device based on interactive attention and storage medium
WO2020020085A1 (en) * 2018-07-24 2020-01-30 华为技术有限公司 Representation learning method and device
CN110766435A (en) * 2018-12-19 2020-02-07 北京嘀嘀无限科技发展有限公司 Vector training method and device, electronic equipment and computer readable storage medium
CN111046193A (en) * 2019-12-27 2020-04-21 南京邮电大学 Domain knowledge graph completion method based on meta-learning
CN111159356A (en) * 2019-12-31 2020-05-15 重庆和贯科技有限公司 Knowledge graph construction method based on teaching content
CN111339320A (en) * 2020-03-02 2020-06-26 北京航空航天大学 Knowledge graph embedding and reasoning method introducing entity type automatic representation
CN111392538A (en) * 2020-03-17 2020-07-10 浙江新再灵科技股份有限公司 Elevator comprehensive fault early warning method based on multi-dimensional Internet of things atlas big data
WO2020147594A1 (en) * 2019-01-16 2020-07-23 阿里巴巴集团控股有限公司 Method, system, and device for obtaining expression of relationship between entities, and advertisement retrieval system
CN111753101A (en) * 2020-06-30 2020-10-09 华侨大学 Knowledge graph representation learning method integrating entity description and type
CN114969385A (en) * 2022-08-03 2022-08-30 北京长河数智科技有限责任公司 Knowledge graph optimization method and device based on document attribute assignment entity weight
CN117688121A (en) * 2024-02-04 2024-03-12 南京师范大学 Geographic knowledge map representation learning method for subtgnn injection spatial features
CN117688121B (en) * 2024-02-04 2024-04-26 南京师范大学 SubGNN geographic knowledge graph representation learning method for injecting spatial features

Cited By (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107783973A (en) * 2016-08-24 2018-03-09 慧科讯业有限公司 The methods, devices and systems being monitored based on domain knowledge spectrum data storehouse to the Internet media event
CN107871158A (en) * 2016-09-26 2018-04-03 清华大学 A kind of knowledge mapping of binding sequence text message represents learning method and device
CN106528609A (en) * 2016-09-28 2017-03-22 厦门理工学院 Vector constraint embedded transformation knowledge graph inference method
CN106528610A (en) * 2016-09-28 2017-03-22 厦门理工学院 Knowledge graph representation learning method based on path tensor decomposition
CN107885760B (en) * 2016-12-21 2021-06-08 桂林电子科技大学 Knowledge graph representation learning method based on multiple semantics
CN107885760A (en) * 2016-12-21 2018-04-06 桂林电子科技大学 It is a kind of to represent learning method based on a variety of semantic knowledge mappings
CN106909622A (en) * 2017-01-20 2017-06-30 中国科学院计算技术研究所 Knowledge mapping vector representation method, knowledge mapping relation inference method and system
CN106844658B (en) * 2017-01-23 2019-12-13 中山大学 Automatic construction method and system of Chinese text knowledge graph
CN106844658A (en) * 2017-01-23 2017-06-13 中山大学 A kind of Chinese text knowledge mapping method for auto constructing and system
CN108885626A (en) * 2017-02-22 2018-11-23 谷歌有限责任公司 Optimize graph traversal
US11551003B2 (en) 2017-02-22 2023-01-10 Google Llc Optimized graph traversal
CN106934042A (en) * 2017-03-16 2017-07-07 中国人民解放军国防科学技术大学 A kind of knowledge mapping represents model and its method
CN106934042B (en) * 2017-03-16 2020-05-29 中国人民解放军国防科学技术大学 Knowledge graph representation system and implementation method thereof
CN108694201A (en) * 2017-04-10 2018-10-23 华为软件技术有限公司 A kind of entity alignment schemes and device
CN108664615A (en) * 2017-05-12 2018-10-16 华中师范大学 A kind of knowledge mapping construction method of discipline-oriented educational resource
CN108959328B (en) * 2017-05-27 2021-12-21 株式会社理光 Knowledge graph processing method and device and electronic equipment
CN108959328A (en) * 2017-05-27 2018-12-07 株式会社理光 Processing method, device and the electronic equipment of knowledge mapping
US11270212B2 (en) 2017-05-27 2022-03-08 Ricoh Company, Ltd. Knowledge graph processing method and device
CN109241290A (en) * 2017-07-10 2019-01-18 华东师范大学 A kind of knowledge mapping complementing method, device and storage medium
CN107480191B (en) * 2017-07-12 2020-08-21 清华大学 Iterative entity alignment model
CN107480191A (en) * 2017-07-12 2017-12-15 清华大学 A kind of entity alignment model of iteration
CN107330125B (en) * 2017-07-20 2020-06-30 云南电网有限责任公司电力科学研究院 Mass unstructured distribution network data integration method based on knowledge graph technology
CN107330125A (en) * 2017-07-20 2017-11-07 云南电网有限责任公司电力科学研究院 The unstructured distribution data integrated approach of magnanimity of knowledge based graphical spectrum technology
CN107590237B (en) * 2017-09-11 2021-04-30 桂林电子科技大学 Knowledge graph representation learning method based on dynamic translation principle
CN107590237A (en) * 2017-09-11 2018-01-16 桂林电子科技大学 A kind of knowledge mapping based on dynamic translation principle represents learning method
CN107943874B (en) * 2017-11-13 2019-08-23 平安科技(深圳)有限公司 Knowledge mapping processing method, device, computer equipment and storage medium
CN107943874A (en) * 2017-11-13 2018-04-20 平安科技(深圳)有限公司 Knowledge mapping processing method, device, computer equipment and storage medium
CN108280061A (en) * 2018-01-17 2018-07-13 北京百度网讯科技有限公司 Text handling method based on ambiguity entity word and device
CN108280061B (en) * 2018-01-17 2021-10-26 北京百度网讯科技有限公司 Text processing method and device based on ambiguous entity words
US11455542B2 (en) 2018-01-17 2022-09-27 Beijing Baidu Netcom Science And Technology Co., Ltd. Text processing method and device based on ambiguous entity words
WO2019144587A1 (en) * 2018-01-24 2019-08-01 平安医疗健康管理股份有限公司 Dynamic knowledge graph updating method fusing medical knowledge and application cases
CN108647258A (en) * 2018-01-24 2018-10-12 北京理工大学 A kind of expression learning method based on entity associated constraint
CN108647258B (en) * 2018-01-24 2020-12-22 北京理工大学 Representation learning method based on entity relevance constraint
CN108388580A (en) * 2018-01-24 2018-08-10 平安医疗健康管理股份有限公司 Merge the dynamic knowledge collection of illustrative plates update method of medical knowledge and application case
CN108510110A (en) * 2018-03-13 2018-09-07 浙江禹控科技有限公司 A kind of water table trend analysis method of knowledge based collection of illustrative plates
CN108491491B (en) * 2018-03-14 2021-04-09 广州杰赛科技股份有限公司 Method and system for constructing intelligent customer service knowledge graph
CN108491491A (en) * 2018-03-14 2018-09-04 广州杰赛科技股份有限公司 The construction method and system of intelligent customer service knowledge mapping
CN108876849A (en) * 2018-04-24 2018-11-23 哈尔滨工程大学 Deep learning target identification and localization method based on accessory ID
CN108876849B (en) * 2018-04-24 2021-11-23 哈尔滨工程大学 Deep learning target identification and positioning method based on auxiliary identification
CN109063032A (en) * 2018-07-16 2018-12-21 清华大学 A kind of noise-reduction method of remote supervisory retrieval data
CN109063032B (en) * 2018-07-16 2020-09-11 清华大学 Noise reduction method for remote supervision and retrieval data
WO2020020085A1 (en) * 2018-07-24 2020-01-30 华为技术有限公司 Representation learning method and device
CN109376864A (en) * 2018-09-06 2019-02-22 电子科技大学 A kind of knowledge mapping relation inference algorithm based on stacking neural network
CN109508419A (en) * 2018-11-23 2019-03-22 成都品果科技有限公司 A kind of recommended method and system of knowledge based study
CN110766435A (en) * 2018-12-19 2020-02-07 北京嘀嘀无限科技发展有限公司 Vector training method and device, electronic equipment and computer readable storage medium
CN109634939A (en) * 2018-12-28 2019-04-16 中国农业银行股份有限公司 A kind of the determination method, apparatus and electronic equipment of missing values
CN109829057A (en) * 2019-01-11 2019-05-31 中山大学 A kind of knowledge mapping Entity Semantics spatial embedding method based on figure second order similitude
CN109829057B (en) * 2019-01-11 2023-02-21 中山大学 Knowledge graph entity semantic space embedding method based on graph second-order similarity
WO2020147594A1 (en) * 2019-01-16 2020-07-23 阿里巴巴集团控股有限公司 Method, system, and device for obtaining expression of relationship between entities, and advertisement retrieval system
CN110147450B (en) * 2019-05-06 2021-08-03 北京科技大学 Knowledge complementing method and device for knowledge graph
CN110147450A (en) * 2019-05-06 2019-08-20 北京科技大学 A kind of the knowledge complementing method and device of knowledge mapping
CN110188147A (en) * 2019-05-22 2019-08-30 厦门无常师教育科技有限公司 The document entity relationship of knowledge based map finds method and system
CN110413793A (en) * 2019-06-11 2019-11-05 福建奇点时空数字科技有限公司 A kind of knowledge mapping substance feature method for digging based on translation model
CN110569364A (en) * 2019-08-21 2019-12-13 北京大米科技有限公司 online teaching method, device, server and storage medium
CN110688489A (en) * 2019-09-09 2020-01-14 中国电子科技集团公司电子科学研究院 Knowledge graph deduction method and device based on interactive attention and storage medium
CN110647620B (en) * 2019-09-23 2022-07-01 中国农业大学 Knowledge graph representation learning method based on confidence hyperplane and dictionary information
CN110647620A (en) * 2019-09-23 2020-01-03 中国农业大学 Knowledge graph representation learning method based on confidence hyperplane and dictionary information
CN111046193B (en) * 2019-12-27 2022-05-13 南京邮电大学 Domain knowledge graph completion method based on meta-learning
CN111046193A (en) * 2019-12-27 2020-04-21 南京邮电大学 Domain knowledge graph completion method based on meta-learning
CN111159356B (en) * 2019-12-31 2023-06-09 重庆和贯科技有限公司 Knowledge graph construction method based on teaching content
CN111159356A (en) * 2019-12-31 2020-05-15 重庆和贯科技有限公司 Knowledge graph construction method based on teaching content
CN111339320B (en) * 2020-03-02 2021-03-26 北京航空航天大学 Knowledge graph embedding and reasoning method introducing entity type automatic representation
CN111339320A (en) * 2020-03-02 2020-06-26 北京航空航天大学 Knowledge graph embedding and reasoning method introducing entity type automatic representation
CN111392538A (en) * 2020-03-17 2020-07-10 浙江新再灵科技股份有限公司 Elevator comprehensive fault early warning method based on multi-dimensional Internet of things atlas big data
CN111753101B (en) * 2020-06-30 2022-06-07 华侨大学 Knowledge graph representation learning method integrating entity description and type
CN111753101A (en) * 2020-06-30 2020-10-09 华侨大学 Knowledge graph representation learning method integrating entity description and type
CN114969385A (en) * 2022-08-03 2022-08-30 北京长河数智科技有限责任公司 Knowledge graph optimization method and device based on document attribute assignment entity weight
CN117688121A (en) * 2024-02-04 2024-03-12 南京师范大学 Geographic knowledge map representation learning method for subtgnn injection spatial features
CN117688121B (en) * 2024-02-04 2024-04-26 南京师范大学 SubGNN geographic knowledge graph representation learning method for injecting spatial features

Similar Documents

Publication Publication Date Title
CN105630901A (en) Knowledge graph representation learning method
CN107273490B (en) Combined wrong question recommendation method based on knowledge graph
CN108492200B (en) User attribute inference method and device based on convolutional neural network
He et al. Mining transition rules of cellular automata for simulating urban expansion by using the deep learning techniques
CN104268197B (en) A kind of industry comment data fine granularity sentiment analysis method
CN101409634B (en) Quantitative analysis tools and method for internet news influence based on information retrieval
CN107885760A (en) It is a kind of to represent learning method based on a variety of semantic knowledge mappings
CN105843897A (en) Vertical domain-oriented intelligent question and answer system
CN105975596A (en) Query expansion method and system of search engine
CN105307200A (en) Method for detecting abnormal value of multidimensional data of wireless sensor network based on trajectory
CN104346379A (en) Method for identifying data elements on basis of logic and statistic technologies
CN110135459A (en) A kind of zero sample classification method based on double triple depth measure learning networks
CN105808689A (en) Drainage system entity semantic similarity measurement method based on artificial neural network
CN108984775A (en) A kind of public sentiment monitoring method and system based on comment on commodity
Sadr et al. Unified topic-based semantic models: A study in computing the semantic relatedness of geographic terms
CN109920476A (en) The disease associated prediction technique of miRNA- based on chaos game playing algorithm
Ma et al. Research on machine learning prediction of air quality index based on SPSS
CN107862037A (en) A kind of event masterplate building method based on entity connected graph
CN116843074A (en) Typhoon disaster damage prediction method based on CNN-LSTM model
CN112163069A (en) Text classification method based on graph neural network node feature propagation optimization
Yu et al. Vulnerability assessment and spatiotemporal differentiation of provinces tourism economic system based on the projection pursuit clustering model
CN106327404A (en) Multi-source data collision detection and resolution method for comprehensive disaster assessment
CN106126606A (en) A kind of short text new word discovery method
Meng et al. Application of Data Mining in the Guidance of Sports Training
Raj On the method of overlapping maps in sample surveys

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20160601

RJ01 Rejection of invention patent application after publication