CN108647258B - Representation learning method based on entity relevance constraint - Google Patents
Representation learning method based on entity relevance constraint Download PDFInfo
- Publication number
- CN108647258B CN108647258B CN201810377516.6A CN201810377516A CN108647258B CN 108647258 B CN108647258 B CN 108647258B CN 201810377516 A CN201810377516 A CN 201810377516A CN 108647258 B CN108647258 B CN 108647258B
- Authority
- CN
- China
- Prior art keywords
- entity
- entities
- relevance
- batch
- vector
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 50
- 230000014509 gene expression Effects 0.000 claims abstract description 16
- 238000012549 training Methods 0.000 claims abstract description 13
- 238000013519 translation Methods 0.000 claims abstract description 10
- 238000005070 sampling Methods 0.000 claims abstract description 7
- 230000006870 function Effects 0.000 claims description 10
- 238000010276 construction Methods 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 4
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 claims description 3
- 238000005065 mining Methods 0.000 claims description 3
- 230000004069 differentiation Effects 0.000 claims description 2
- 230000009191 jumping Effects 0.000 claims description 2
- 230000000694 effects Effects 0.000 abstract description 5
- 238000003058 natural language processing Methods 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 238000013441 quality evaluation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Machine Translation (AREA)
Abstract
The invention relates to an expression learning method based on entity relevance constraint, and belongs to the technical field of natural language processing and knowledge maps. According to the method, a strong association entity set and a weak association entity set of an entity are obtained by performing annotation and association division on a description text of the entity, the association is used as an auxiliary loss item and is fused into a translation-based representation learning method, and specifically, embedded representation of the entity and the relationship is obtained through sample negative sampling and model training, namely a head entity h and a tail entity t in a knowledge graph and a relationship r between the head entity and the tail entity are respectively embedded into a vector h, the vector t and the vector r. The method is superior to representation learning methods based on translation and text models in reasoning effect.
Description
Technical Field
The invention relates to an expression learning method based on entity relevance constraint, and belongs to the technical field of natural language processing and knowledge maps.
Background
A Knowledge Graph (Knowledge Graph) is a Knowledge representation method based on a semantic network, provides an efficient and concise structural representation mode, and plays a key role in the fields of Web search and intelligent question answering. The knowledge graph represents real world data as entities and relations, knowledge is stored in (entities, relations, entities) triples, and the entities are connected through relations to form a network knowledge structure. Although a knowledge graph contains a large number of entities and relationships, the knowledge graph is not complete due to the fact that all of the knowledge graph cannot be extracted during the construction process of the knowledge graph. The knowledge graph reasoning technology can be applied to automatic completion of the graph, for example, the possible relation between two entities is predicted, and the quality evaluation can be carried out on the information extraction result by combining with the information extraction in the open field. The expression learning method is to embed the entities and relations in the knowledge graph into a low-dimensional space and finish the inference of the knowledge graph in the low-dimensional space.
Currently, the mainstream expression learning method regards the relationship as a translation process between entities, which is called a translation-based model, such as a trans model (border a, user N, Weston J, et al. translation expressions for modeling multi-translation data [ C ] in International Conference on Neural Information Processing systems, current associations inc.2013: 2787) which proposes that the relationship is a translation operation between a head entity and a tail entity. Therefore, some researchers propose a Representation Learning method combining an Entity description text and a graph structure, which belongs to a text Representation model, wherein a Representation DKRL model (Xie, r., Liu, z., Jia, j., Luan, h., & Sun, M. (2016, library). reproduction Learning of Knowledge Graphs with Entity descriptions.
Although the existing representation learning method has good effect on knowledge graph reasoning, semantic association between entities in a text is lack of mining, and a great space exists for improving reasoning performance. The invention aims to overcome the technical defect that semantic association information among entities is lost in the training process of the traditional expression learning method, and provides an expression learning method based on entity association constraint.
Disclosure of Invention
The invention aims to provide a representation learning method based on entity relevance constraint aiming at the problems that a translation-based model does not utilize rich semantic information in a text and the text representation model loses semantic relevance information among entities in the training process.
The core idea of the invention is as follows: mining out a relevance entity based on the entity description text, grading the relevance, and fusing the relevance serving as an auxiliary constraint into a translation-based representation learning method; the annotated entity description text is mainly used for obtaining the co-occurrence information between the entities, and the information is used as a standard for measuring the semantic association degree between the two entities, and the association degree is directional. The specific implementation is that a head entity h, a tail entity t and a relation r between the head entity and the tail entity in the knowledge graph are respectively embedded into a vector h, a vector t and a vector r.
The invention is realized by the following steps:
firstly, annotating and performing relevance division on a description text of an entity to obtain a strong relevance entity set and a weak relevance entity set of the entity; the method specifically comprises the following substeps:
step 1.1, annotating description texts of entities to obtain entity annotation results;
wherein the entity refers to an entity in the knowledge graph and is represented by e; e description text by DeseThe expression is an ordered word set, and is expressed by formula (1):
Dese=<w1,...,wm> (1)
wherein, w1,...,wmThe words are used, m is the number of words in the description text, the entity extracted from the description text is composed of more than or equal to 1 word, and when the entity is composed of more than or equal to two words, the extracted words need to be spliced;
the process of extracting entities from descriptive text is called descriptive text annotation; and (3) forming a set by the entities extracted from the description text to obtain an entity annotation result:
Dese'=<w1,...,wm'> (2)
wherein m' is less than or equal to m, wiRepresents an entity, DeseIs' DeseThe entity annotation result of (1);
step 1.2, relevance division;
obtaining the association degree value of the entity j to the entity i by using the ith entity and the jth entity in the entity annotation result output in the step 1.1 through a formula (3), and using WijRepresents:
if W isij2, a strongly associated Entity (Strong Relevant Entity) denoted j as i; if W isij1, recording j as a Weak associated Entity (Weak relevance Entity) of i, if two entities appear in the description of each other, the association becomes strong, and then obtaining a strong associated Entity set and a Weak associated Entity set of an Entity e;
degree of association value WijTraversing all entities in the entity annotation result to obtain an entity incidence matrix composed of incidence degree values, and recording the entity incidence matrix asE is the set of entities in the knowledge-graph, | E | represents the total number of entities in the knowledge-graph:
wherein, the strongly associated entity set of the entity e is denoted as s (e):
wherein e isiWhich represents the (i) th entity, the (ii) th entity,representing entity e and entity eiThe relationship between the two is strong association entity;
the weakly associated entity set of entity e is denoted as w (e):
secondly, carrying out sample negative sampling and model training to obtain an embedded expression of an entity and a relation;representing entity e and entity eiThe relationship between the two is weak association entity;
the model training is based on batch random gradient descent algorithm;
step two, the following substeps are specifically included:
step 2.1, initializing a cycle count value to be 1 and a cycle count maximum value;
wherein, the cycle count value is marked as k; the maximum cycle count, denoted iter;
step 2.2, enabling S to represent a triple set in the knowledge graph, wherein one triple in the knowledge graph is a positive sample, namely S is a positive sample set; randomly extracting B positive samples from S to obtain a subset SbatchLet us orderTbatchThe construction of (1) comprises the following substeps:
step 2.2.1, traverse SbatchEach positive sample (h, r, t) was sampled negatively as described in document 1(Feng J. knowledge Graph Embedding by transformation on Hyperplanes [ C ]]in aaai.2014): given the relation r, the average number tph of tail entities corresponding to each head entity in the sample negative sampling method in document 1 corresponds to tph of the patentrThe average number hpt of head entities corresponding to each tail entity corresponds to hpt of the patentr;
Produce a [0,1 ]]Uniformly distributed random number p of interval, if p is less than or equal to tphr/(tphr+hptr) Then equal probability is given from the set of entities E of the knowledge-graphAn entity is extracted to replace a head entity in the positive sample, and the replaced triple is ensured not to belong to S; if p is greater than tphr/(tphr+hptr) Extracting an entity from the entity set E of the knowledge graph with equal probability to replace the tail entity in the positive sample, and ensuring that the replaced triple does not belong to S;
step 2.2.2, after the replacement is completed, S can be obtainedbatchA negative sample (h ', r, T') corresponding to each positive sample (h, r, T), adding each positive sample and negative sample to TbatchIn the set:
Tbatch←Tbatch∪{(h,r,t),(h',r,t')} (6)
after the steps 2.2.1 and 2.2.2, T is obtainedbatchSet, extract TbatchThe entity set in (1), denoted as Ebatch;
2.3, training the model based on a batch random gradient descent algorithm;
step 2.3.1, calculating the scores of the triplets (h, r, t) through the scoring function of the triplets (h, r, t) in the formula (7), and marking the scores as fr(h,t);
step 2.3.2, calculating a loss term L based on entity relevance through a formula (8)r:
Wherein, alpha and beta are a strong correlation weight and a weak correlation weight, alpha determines the strength of the strong correlation constraint, and beta determines the strength of the weak correlation constraint; e represents EbatchThe entity of (1); on the left hand side of equation (8), e' represents the strongly associated set of entities for e, in equation(8) The right entry of (e), the weakly associated entity set of table e;represents the square of the 2 norm of the vector e-e'; SC and WC represent strong correlation hyper-parameters and weak correlation hyper-parameters specified by a user, respectively represent distance limits between two associated entities, and when an entity pair is in a corresponding range, the loss is 0, and L isrSuch that the distance of the associated entity pair in vector space does not exceed a certain range and does not monotonically minimize the associated entity pair distance;
and 2.3.3, calculating a loss function value of the model according to a formula (9):
wherein, Loss represents the Loss function value of the model; f. ofr(h, t) represents the score of a positive sample (h, r, t), fr(h ', t') represents the scores of the negative samples (h ', r, t'), and the positive samples tend to be low in score and the negative samples tend to be high in score when training; gamma is the loss interval, gamma is used to control fr(h, t) and fr(h ', t');
step 2.3.4, calculating the derivative of the independent variable in the formula (9) and updating according to the formula (10);
wherein theta is an independent variable, all h, r and t are included, rate is a learning rate,representing the differentiation of the Loss function value Loss of the model with respect to the argument θ;
step 2.3.5, judging whether the cycle count value k reaches the maximum value iter, and if k is iter, finishing the method; otherwise k is k +1, jumping to step 2.2;
so far, from step one to step two, an embedded representation of entities and relationships is obtained: and completing a representation learning method based on entity relevance constraint by using the vector h, the vector t and the vector r.
Advantageous effects
Compared with the prior art, the representation learning method based on entity relevance constraint has the following beneficial effects:
1. the traditional representation learning method is based on the structural information of the knowledge graph, and the description text information of the entity is not fully utilized, the invention provides a method for measuring the semantic association of the entity from the entity description text, a constraint item based on the entity association is constructed by using the method, the constraint item is fused into the traditional representation learning method, and the experimental result shows that the invention has better reasoning effect on the link prediction task and the triple classification task on the open data set compared with the traditional method, and the speed is consistent with that of the traditional method;
2. the expression learning method based on the text mostly vectorizes the text, omits semantic association among entities in the text, excavates entity association constraint items from the entities in the text, more carefully models the semantic association among the entities, and the experimental result shows that the expression learning method based on the text has better reasoning effect compared with a text expression model DKRL.
Drawings
Fig. 1 is a schematic flow diagram of a representation learning method based on entity association constraint and an embodiment 1 of the invention.
Detailed Description
The invention is further illustrated and described in detail below with reference to the figures and examples.
Example 1
This embodiment describes a specific implementation process of the representation learning method based on entity association constraint according to the present invention, and fig. 1 is an implementation flow diagram of this embodiment.
As can be seen from fig. 1, the specific implementation steps of the present invention and the embodiment are as follows:
a, annotating and performing relevance division on a description text of an entity to obtain a strong relevance entity set and a weak relevance entity set of the entity; the method specifically comprises the following substeps:
a.1, annotating description texts of entities to obtain entity annotation results;
wherein the entity refers to an entity in the knowledge graph and is represented by e; e description text by DeseThe expression is an ordered word set, and is expressed by formula (1):
Dese=<w1,...,wm> (11)
wherein, w1,...,wmThe words are used, m is the number of words in the description text, the entity extracted from the description text is composed of more than or equal to 1 word, and when the entity is composed of more than or equal to two words, the extracted words need to be spliced;
the process of extracting entities from descriptive text is called descriptive text annotation; and (3) forming a set by the entities extracted from the description text to obtain an entity annotation result:
Dese'=<w1,...,wm'> (12)
wherein m' is less than or equal to m, wiRepresents an entity, DeseIs' DeseThe entity annotation result of (1);
step A.2, relevance division;
obtaining the association degree value of the entity j to the entity i by the ith entity and the jth entity in the entity annotation result output in the step A.1 through a formula (3), and using WijRepresents:
if W isij2, a strongly associated Entity (Strong Relevant Entity) denoted j as i; if W isij1, recording j as a Weak associated Entity (Weak relevance Entity) of i, if two entities appear in the description of each other, the association becomes strong, and then obtaining a strong associated Entity set and a Weak associated Entity set of an Entity e;
degree of association value WijIs directed, traverses entity annotationsAll entities in the result obtain an entity incidence matrix, which is marked asE is the set of entities in the knowledge-graph, | E | represents the total number of entities in the knowledge-graph:
wherein, the strongly associated entity set of the entity e is denoted as s (e):
the weakly associated entity set of entity e is denoted as w (e):
b, carrying out sample negative sampling and model training to obtain an embedded expression of an entity and a relation;
the model training is based on batch random gradient descent algorithm; step two, the following substeps are specifically included:
step B.1, initializing a cycle count value, wherein the cycle count value is recorded as k, and the initialization k is 1;
b.2, enabling S to represent a triple set in the knowledge graph, wherein one triple in the knowledge graph is a positive sample, namely S is a positive sample set; randomly extracting B positive samples from S to obtain a subset SbatchWherein B is 100, 1TbatchThe construction of (1) comprises the following substeps:
step B.2.1, traverse SbatchEach positive sample (h, r, t) was sampled negatively as described in document 1(Feng J. knowledge Graph Embedding by transformation on Hyperplanes [ C ]]in aaai.2014): given the relation r, the average number tph of tail entities corresponding to each head entity in the sample negative sampling method in document 1 corresponds to tph of the patentrAverage of each pair of tail entitiesThe number hpt of head entities corresponds to hpt of the patentr;
Produce a [0,1 ]]Uniformly distributed random number p of interval, if p is less than or equal to tphr/(tphr+hptr) Extracting an entity from the entity set E of the knowledge graph with equal probability to replace the head entity in the positive sample, and ensuring that the replaced triple does not belong to S; if p is greater than tphr/(tphr+hptr) Extracting an entity from the entity set E of the knowledge graph with equal probability to replace the tail entity in the positive sample, and ensuring that the replaced triple does not belong to S;
step B.2.2, after the replacement is finished, S can be obtainedbatchA negative sample (h ', r, T') corresponding to each positive sample (h, r, T), adding each positive sample and negative sample to TbatchIn the set:
Tbatch←Tbatch∪{(h,r,t),(h',r,t')} (16)
obtaining T after the step B.2.1 and the step B.2.2batchSet, extract TbatchThe entity set in (1), denoted as Ebatch;
B.3, training the model based on a batch random gradient descent algorithm;
and B.3.1, calculating the scores of the triples (h, r, t) through the scoring function of the triples (h, r, t) in the formula (7), and marking the scores as fr(h,t);
step B.3.2, calculating a loss term L based on entity relevance through a formula (8)r:
Wherein α and β are a strong association weight and a weak association weight, SC and WC are a strong association range and a weak association range, respectively, where α ═ 1, β ═ 0.3, SC ═ 1, and WC ═ 1;
and 2.3.3, calculating a loss function value of the model according to a formula (9):
wherein gamma is a loss interval and is 1;
step 2.3.4, calculating the derivative of the independent variable in the formula (9) and updating according to the formula (10);
wherein, θ is an independent variable including all h, r and t, rate is a learning rate, and rate is 0.1;
step 2.3.5, judging whether the cycle count value k reaches the maximum value iter, wherein iter is 500, and if k is iter, completing the method; otherwise k is k +1, go to step 2.1.
So far, from step one to step two, an embedded representation of entities and relationships is obtained: and completing a representation learning method based on entity relevance constraint by using the vector h, the vector t and the vector r.
Claims (1)
1. A representation learning method based on entity relevance constraint is characterized in that: the core idea is as follows: mining out a relevance entity based on the entity description text, grading the relevance, and fusing the relevance serving as an auxiliary constraint into a translation-based representation learning method; obtaining co-occurrence information between entities by mainly utilizing the annotated entity description text, wherein the information is used as a standard for measuring semantic association degree between two entities, and the association degree is directional; the specific implementation is that a head entity h, a tail entity t and a relation r between the head entity and the tail entity in a knowledge graph are respectively embedded into a vector h, a vector t and a vector r; the method is realized by the following steps:
firstly, annotating and performing relevance division on a description text of an entity to obtain a strong relevance entity set and a weak relevance entity set of the entity; the method specifically comprises the following substeps:
step 1.1, annotating description texts of entities to obtain entity annotation results;
wherein the entity refers to an entity in the knowledge graph and is represented by e; e description text by DeseThe expression is an ordered word set, and is expressed by formula (1):
Dese=<w1,...,wm> (1)
wherein, w1,...,wmThe words are used, m is the number of words in the description text, the entity extracted from the description text is composed of more than or equal to 1 word, and when the entity is composed of more than or equal to two words, the extracted words need to be spliced;
the process of extracting entities from descriptive text is called descriptive text annotation; and (3) forming a set by the entities extracted from the description text to obtain an entity annotation result:
Dese'=<w1,...,wm'> (2)
wherein m' is less than or equal to m, wiRepresents an entity, DeseIs' DeseThe entity annotation result of (1);
step 1.2, relevance division;
obtaining the association degree value of the entity j to the entity i by using the ith entity and the jth entity in the entity annotation result output in the step 1.1 through a formula (3), and using WijRepresents:
if W isij2, a strongly associated Entity (Strong Relevant Entity) denoted j as i; if W isijA weakly associated Entity (Weak relevance) denoted j by i, which becomes strongly associated if two entities appear to each other in the description of each otherThen, a strong associated entity set and a weak associated entity set of the entity e are obtained;
traversing all entities in the entity annotation result to obtain an entity association matrix consisting of association degree values, and marking as W e ∈|E|×|E|E is the set of entities in the knowledge-graph, | E | represents the total number of entities in the knowledge-graph:
wherein, the strongly associated entity set of the entity e is denoted as s (e):
wherein e isiWhich represents the (i) th entity, the (ii) th entity,representing entity e and entity eiThe relationship between the two is strong association entity;
the weakly associated entity set of entity e is denoted as w (e):
secondly, carrying out sample negative sampling and model training to obtain an embedded expression of an entity and a relation;representing entity e and entity eiThe method specifically comprises the following substeps:
step 2.1, initializing a cycle count value to be 1 and a cycle count maximum value;
wherein, the cycle count value is marked as k; the maximum cycle count, denoted iter;
step 2.2, enabling S to represent a triple set in the knowledge graph, wherein one triple in the knowledge graph is a positive sample, namely S is a positive sample set; randomly extracting B positive samples from S to obtain a subset SbatchLet us orderThe construction of (1) comprises the following substeps:
step 2.2.1, traverse Sbatch-negative sampling each positive sample (h, r, t);
produce a [0,1 ]]Uniformly distributed random number p of interval, if p is less than or equal to tphr/(tphr+hptr) Extracting an entity from the entity set E of the knowledge graph with equal probability to replace the head entity in the positive sample, and ensuring that the replaced triple does not belong to S; if p is greater than tphr/(tphr+hptr) Extracting an entity from the entity set E of the knowledge graph with equal probability to replace the tail entity in the positive sample, and ensuring that the replaced triple does not belong to S;
step 2.2.2, after the replacement is completed, S can be obtainedbatchA negative sample (h ', r, T') corresponding to each positive sample (h, r, T), adding each positive sample and negative sample to TbatchIn the set:
after the steps 2.2.1 and 2.2.2, T is obtainedbatchSet, extract TbatchThe entity set in (1), denoted as Ebatch;
2.3, training the model based on a batch random gradient descent algorithm;
step 2.3.1, calculating the scores of the triplets (h, r, t) through the scoring function of the triplets (h, r, t) in the formula (7), and marking the scores as fr(h,t);
step 2.3.2, calculating a loss term L based on entity relevance through a formula (8)r:
Wherein, alpha and beta are a strong correlation weight and a weak correlation weight, alpha determines the strength of the strong correlation constraint, and beta determines the strength of the weak correlation constraint; e represents EbatchThe entity of (1); in the left term of formula (8), e 'represents the strongly associated entity set of e, and in the right term of formula (8), e' represents the weakly associated entity set of e;represents the square of the 2 norm of the vector e-e'; SC and WC represent strong correlation hyper-parameters and weak correlation hyper-parameters specified by a user, respectively represent distance limits between two associated entities, and when an entity pair is in a corresponding range, the loss is 0, and L isrSuch that the distance of the associated entity pair in vector space does not exceed a certain range and does not monotonically minimize the associated entity pair distance;
and 2.3.3, calculating a loss function value of the model according to a formula (9):
wherein, Loss represents the Loss function value of the model; f. ofr(h, t) represents the score of a positive sample (h, r, t), fr(h ', t') represents the scores of the negative samples (h ', r, t'), and the positive samples tend to be low in score and the negative samples tend to be high in score when training; gamma is the loss interval, gamma is used to control fr(h, t) and fr(h ', t');
step 2.3.4, calculating the derivative of the independent variable in the formula (9) and updating according to the formula (10);
wherein theta is an independent variable, all h, r and t are included, rate is a learning rate,representing the differentiation of the Loss function value Loss of the model with respect to the argument θ;
step 2.3.5, judging whether the cycle count value k reaches the maximum value iter, and if k is iter, finishing the method; otherwise k is k +1, jumping to step 2.2;
so far, from step one to step two, an embedded representation of entities and relationships is obtained: vector h, vector t, and vector r.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2018100675556 | 2018-01-24 | ||
CN201810067555 | 2018-01-24 |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108647258A CN108647258A (en) | 2018-10-12 |
CN108647258B true CN108647258B (en) | 2020-12-22 |
Family
ID=63747612
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810377516.6A Active CN108647258B (en) | 2018-01-24 | 2018-04-25 | Representation learning method based on entity relevance constraint |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108647258B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110674637B (en) * | 2019-09-06 | 2023-07-11 | 腾讯科技(深圳)有限公司 | Character relationship recognition model training method, device, equipment and medium |
CN110909881B (en) * | 2019-11-01 | 2022-11-04 | 中电科大数据研究院有限公司 | Knowledge representation method for cross-media knowledge reasoning task |
CN111428047B (en) * | 2020-03-19 | 2023-04-21 | 东南大学 | Knowledge graph construction method and device based on UCL semantic indexing |
CN113220833A (en) * | 2021-05-07 | 2021-08-06 | 支付宝(杭州)信息技术有限公司 | Entity association degree identification method and device |
CN114330323B (en) * | 2022-03-08 | 2022-06-28 | 成都数联云算科技有限公司 | Entity relationship joint extraction method and device, computer terminal and storage medium |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105630901A (en) * | 2015-12-21 | 2016-06-01 | 清华大学 | Knowledge graph representation learning method |
CN107122399A (en) * | 2017-03-16 | 2017-09-01 | 中国科学院自动化研究所 | Combined recommendation system based on Public Culture knowledge mapping platform |
CN107273349A (en) * | 2017-05-09 | 2017-10-20 | 清华大学 | A kind of entity relation extraction method and server based on multilingual |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9251248B2 (en) * | 2010-06-07 | 2016-02-02 | Microsoft Licensing Technology, LLC | Using context to extract entities from a document collection |
-
2018
- 2018-04-25 CN CN201810377516.6A patent/CN108647258B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105630901A (en) * | 2015-12-21 | 2016-06-01 | 清华大学 | Knowledge graph representation learning method |
CN107122399A (en) * | 2017-03-16 | 2017-09-01 | 中国科学院自动化研究所 | Combined recommendation system based on Public Culture knowledge mapping platform |
CN107273349A (en) * | 2017-05-09 | 2017-10-20 | 清华大学 | A kind of entity relation extraction method and server based on multilingual |
Non-Patent Citations (3)
Title |
---|
Representation learning of knowledge graphs with entity descriptions;Ruobing Xie等;《Thirtieth AAAI Conference on Artificial Intelligence》;20161231;2659-2665 * |
SSP:semantic space projection for knowledge graph embedding with text descriptions;Han Xiao等;《Thirty-First AAAI conference on artificial intelligence》;20171231;3014-3110 * |
知识图谱构建技术综述;刘峤等;《计算机研究与发展》;20161231;第53卷(第03期);582-600 * |
Also Published As
Publication number | Publication date |
---|---|
CN108647258A (en) | 2018-10-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108647258B (en) | Representation learning method based on entity relevance constraint | |
CN113254599B (en) | Multi-label microblog text classification method based on semi-supervised learning | |
CN108536870B (en) | Text emotion classification method fusing emotional features and semantic features | |
CN110569508A (en) | Method and system for classifying emotional tendencies by fusing part-of-speech and self-attention mechanism | |
CN105975573B (en) | A kind of file classification method based on KNN | |
CN111144448A (en) | Video barrage emotion analysis method based on multi-scale attention convolutional coding network | |
CN109635109A (en) | Sentence classification method based on LSTM and combination part of speech and more attention mechanism | |
CN112395393B (en) | Remote supervision relation extraction method based on multitask and multiple examples | |
CN110619121B (en) | Entity relation extraction method based on improved depth residual error network and attention mechanism | |
CN110298403A (en) | The sentiment analysis method and system of enterprise dominant in a kind of financial and economic news | |
CN111581385A (en) | Chinese text type identification system and method for unbalanced data sampling | |
CN111274814B (en) | Novel semi-supervised text entity information extraction method | |
CN112395417A (en) | Network public opinion evolution simulation method and system based on deep learning | |
CN111506728B (en) | Hierarchical structure text automatic classification method based on HD-MSCNN | |
CN110334187A (en) | Burmese sentiment analysis method and device based on transfer learning | |
CN111914555B (en) | Automatic relation extraction system based on Transformer structure | |
CN113723083A (en) | Weighted negative supervision text emotion analysis method based on BERT model | |
CN113886562A (en) | AI resume screening method, system, equipment and storage medium | |
Chu et al. | Co-training based on semi-supervised ensemble classification approach for multi-label data stream | |
CN115630156A (en) | Mongolian emotion analysis method and system fusing Prompt and SRU | |
CN114548117A (en) | Cause-and-effect relation extraction method based on BERT semantic enhancement | |
CN113011192B (en) | Text emotion feature extraction method based on attention causal interpretation | |
CN116050419B (en) | Unsupervised identification method and system oriented to scientific literature knowledge entity | |
CN114943216B (en) | Case microblog attribute level view mining method based on graph attention network | |
CN111708896B (en) | Entity relationship extraction method applied to biomedical literature |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |