CN111831783A - Chapter-level relation extraction method - Google Patents

Chapter-level relation extraction method Download PDF

Info

Publication number
CN111831783A
CN111831783A CN202010644404.XA CN202010644404A CN111831783A CN 111831783 A CN111831783 A CN 111831783A CN 202010644404 A CN202010644404 A CN 202010644404A CN 111831783 A CN111831783 A CN 111831783A
Authority
CN
China
Prior art keywords
entity
representation
document
sentence
abstract semantic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010644404.XA
Other languages
Chinese (zh)
Other versions
CN111831783B (en
Inventor
张世琨
叶蔚
李博
胡文蕙
张君福
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Peking University Software Engineering Co ltd
Original Assignee
Beijing Peking University Software Engineering Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Peking University Software Engineering Co ltd filed Critical Beijing Peking University Software Engineering Co ltd
Priority to CN202010644404.XA priority Critical patent/CN111831783B/en
Publication of CN111831783A publication Critical patent/CN111831783A/en
Application granted granted Critical
Publication of CN111831783B publication Critical patent/CN111831783B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/31Indexing; Data structures therefor; Storage structures
    • G06F16/313Selection or weighting of terms for indexing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/353Clustering; Classification into predefined classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention provides a chapter-level relation extraction method, relates to the technical field of natural language processing, and mainly solves the technical problems of calculation resource consumption and logical reasoning between a target entity and a non-target entity aiming at chapter-level documents. The invention comprises the following steps: inputting a document to be processed, wherein the document is a chapter-level document; processing the document based on bidirectional attention constraints to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information; and judging the relation type of the target entity pair in the document based on the abstract semantic representation. The developer can efficiently and accurately extract the discourse-level relationship by using the method of the invention, and simultaneously solve two main problems of discourse-level relationship extraction, namely the problem of the calculation cost caused by traversing all entities to generate alternative samples and the problem of logical reasoning between the target entity and the non-target entity.

Description

Chapter-level relation extraction method
Technical Field
The invention relates to the technical field of natural language processing, in particular to a chapter-level relation extraction method in the technical field of bidirectional attention constraint and graph-convolution neural network intersection. The invention obtains abstract representation of an entity and a document through bidirectional attention mechanism constraint, and uses a chapter-level relation extraction technology for logical reasoning by using a graph convolution neural network.
Background
Relationship extraction is a judgment based on a given context and two target entitiesTwo target entitiesWhat is the relationship between them. Relational extraction is one of the very important techniques for constructing large-scale knowledge graphs, and can also be used to assist some downstream tasks, such as question-answering systems and the like. The research on relation extraction is mainly divided into two directions, wherein one direction is based on a traditional machine learning idea for constructing a large number of artificial feature projects, and the other direction is a deep learning method for building a neural network.
In traditional machine learning solutions, some researchers have constructed different features for the entire sentence and for a given target entity pair. The traditional machine learning solution method is very dependent on feature engineering and professional knowledge, so that a large amount of manpower is needed for constructing features on a model, meanwhile, the generalization capability of the model greatly fluctuates along with the change of a data set, and in recent years, a large number of researchers use a deep learning method to solve the problem of relationship extraction. In deep learning, researchers have focused their research on applying Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), and attention mechanisms, as well as improvements to the above models, thus also yielding a number of relevant solutions based on deep learning methods.
The existing task definition and method for extracting the relation are based on the task of sentence level, namely, one sentence is taken as a context, and the relation of target entities is judged according to two target entities appearing in the current sentence. In practical application scenarios, the processed text is not just as simple as sentence-level text in most cases, and relatively speaking, a large number of entities appear in each chapter, and each target entity appears multiple times in a document, in which case some classical methods in relation extraction for sentence-level are not available, such as entity position vector, PCNN, etc. Meanwhile, the length of the text at the article level is often hundreds or even thousands, which is far greater than that of the text at the sentence level, the extraction of the whole document information is difficult due to the overlong text, the entity is difficult to be well represented and interacted due to the long-distance loss of the dependent information, and therefore the effect of the chapter-level relation extraction model is greatly reduced. Meanwhile, because a large number of entities exist, the combination between each entity should become a potential target entity pair, and through traversing all entity combinations in each article, a sentence and a target entity sample are constructed, and a sentence-level relation extraction model is applied, so that the consumption of calculation data is greatly increased. More importantly, due to the increase of the text length, a large number of entities exist in the input document, and the relationship between two target entities is obtained not only by direct interaction, but also by one or more times of reasoning between the entities possibly through logical reasoning between non-target entities, so that the accurate classification result cannot be obtained for the relationship needing reasoning only by interacting the target entities.
Disclosure of Invention
One of the purposes of the present invention is to provide a chapter-level relationship extraction method, which solves the technical problems of the prior art, such as the chapter-level problem, the calculation resource consumption problem and the logical reasoning between the target entity and the non-target entity. Advantageous effects can be achieved in preferred embodiments of the present invention, as described in detail below.
In order to achieve the purpose, the invention provides the following technical scheme:
the invention relates to a chapter-level relation extraction method, which comprises the following steps:
inputting a document to be processed, wherein the document is a chapter-level document;
processing the document based on bidirectional attention constraints to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information;
and judging the relation type of the target entity pair in the document based on the abstract semantic representation.
Further, the processing the document based on the bidirectional attention constraint to obtain an abstract semantic representation of an entity and a sentence includes:
obtaining a representation of each entity and a representation of each sentence in the document;
obtaining a global entity interaction matrix based on the representation of each entity;
calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention;
calculating the representation of each sentence based on the entity attention based on a graph convolution network to obtain the representation of each sentence with logical reasoning information;
and calculating the global entity interaction matrix and the representation of each sentence with the logical reasoning information based on a second attention mechanism to obtain a new global entity interaction matrix with the global information and the logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence.
Further, the obtaining a representation of each entity and a representation of each sentence in the document includes:
acquiring abstract semantic representation of the document;
a representation of each entity and a representation of each sentence in the document are obtained based on the abstract semantic representation of the document.
Further, the obtaining a representation of each sentence in the document based on the abstract semantic representation of the document includes:
and extracting the composition of each sentence from the abstract semantic representation of the document according to the initial position and the end position of each sentence, performing maximum pooling on the composition of each sentence, and taking a sentence vector obtained after the maximum pooling as the representation of each sentence.
Further, the obtaining a representation of each entity in the document based on the abstract semantic representation of the document includes:
determining words forming the entity according to the position of each entity, calculating an average word vector of the words, taking the average word vector as an entity vector, and taking the average vector of the entity vectors of the same entity as the representation of each entity.
Further, the obtaining the abstract semantic representation of the document includes:
performing word vector conversion on the document to obtain a word vector matrix of the document;
and carrying out bidirectional LSTM operation on the word vector matrix to obtain abstract semantic representation of the document.
Further, the obtaining a global entity interaction matrix based on the representation of each entity includes:
carrying out information interaction on every two entities, and converting two entity vectors into an entity interaction vector;
and forming a global entity interaction matrix by all entity interaction vectors.
Further, the matrix weights used by the first attention mechanism and the second attention mechanism are the same.
Further, the determining a relationship type of a target entity pair in the document based on the abstract semantic representation includes:
carrying out column-based weighting operation on the new global entity interaction matrix to obtain a global entity interaction vector;
and inputting the global entity interaction vector into a preset classification function for distinguishing entity pair relationship types to obtain a classification result, and taking the classification result as the relationship type of a target entity pair in the document.
Further, the classification function adopts a softmax function.
The chapter-level relation extraction method provided by the invention at least has the following beneficial technical effects:
the invention predicts the relationship result of all entity combinations in the document at the same time through the representation of each entity and the coding of the relative position between the entities, namely, no matter how many different entities are in the current document, the invention can efficiently and simultaneously obtain the entity representation, and further input the relationship result of all potential entity pairs in the document at one time. In addition, aiming at the problem of logical reasoning between non-target entities, the invention adopts a bidirectional attention mechanism, namely, firstly acquiring each sentence and the coded representation of each entity, and then, aiming at each entity, traversing to acquire whether the combination of the entity pairs has a relationship or not, and if so, which relationship in a given relationship set.
The developer can efficiently and accurately extract the discourse-level relationship by using the method of the invention, and simultaneously solve two main problems of discourse-level relationship extraction, namely the problem of the calculation cost caused by traversing all entities to generate alternative samples and the problem of logical reasoning between the target entity and the non-target entity.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a flow chart of a chapter-level relationship extraction method according to the present invention;
FIG. 2 is a schematic structural diagram of a chapter-level relationship extraction method according to the present invention;
FIG. 3 is a flow diagram illustrating the processing of the document based on a bi-directional attention constraint according to the present invention;
fig. 4 is a schematic diagram of an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the technical solutions of the present invention will be described in detail below. It is to be understood that the described embodiments are merely exemplary of the invention, and not restrictive of the full scope of the invention. All other embodiments, which can be derived by a person skilled in the art from the examples given herein without any inventive step, are within the scope of the present invention.
Referring to fig. 1 and 2, the present invention is a chapter-level relationship extraction method, which includes the following steps:
s1: inputting a document to be processed, wherein the document is a chapter-level document;
s2: processing the document based on bidirectional attention constraints to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information;
s3: and judging the relation type of the target entity pair in the document based on the abstract semantic representation.
It should be noted that the bidirectional attention mechanism is based on the following intuitive observation: if a given sentence contains a large amount of information that determines what the target entity pair is, then the target entity pair, in turn, may find it more closely associated with the given sentence when computing the association with each sentence in the document.
The invention mainly aims at documents at chapter level, namely, texts with the length often hundreds or even thousands which are far larger than that of sentences, and a large number of target entities exist in the documents at chapter level. By the method for extracting the relationship of the target entity in the discourse-level document based on the bidirectional attention constraint, the input discourse-level document is subjected to a series of transformations and information extraction, and whether the combination of the target entity pair contains the relationship or not is finally obtained, and if the relationship exists, which relationship is in a given relationship set. The developer can efficiently and accurately extract the discourse-level relationship and simultaneously solve two main problems of discourse-level relationship extraction, namely the problem of calculation cost caused by traversing all entities to generate alternative samples and the problem of logical reasoning between a target entity and a non-target entity.
Referring to fig. 3, the processing the document based on the bidirectional attention constraint to obtain an abstract semantic representation of an entity and a sentence includes:
s21: obtaining a representation of each entity and a representation of each sentence in the document;
s22: obtaining a global entity interaction matrix based on the representation of each entity;
s23: calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention;
s24: calculating the representation of each sentence based on the entity attention based on a graph convolution network to obtain the representation of each sentence with logical reasoning information;
s25: and calculating the global entity interaction matrix and the representation of each sentence with the logical reasoning information based on a second attention mechanism to obtain a new global entity interaction matrix with the global information and the logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence.
In step S21, the obtaining of the representation of each entity and the representation of each sentence in the document includes:
acquiring abstract semantic representation of the document;
a representation of each entity and a representation of each sentence in the document are obtained based on the abstract semantic representation of the document.
Wherein the obtaining of the abstract semantic representation of the document comprises:
performing word vector conversion on the document to obtain a word vector matrix of the document;
and carrying out bidirectional LSTM operation on the word vector matrix to obtain abstract semantic representation of the document.
It should be noted that, the pre-training word vector matrix is used to encode the input document to be processed, and convert it into word vectors, so as to obtain word vector representation of the document, and further obtain abstract semantic representation of the document through bidirectional LSTM.
Suppose that there are n words, m sentences, and k different entities in the document. Suppose that the ith word is denoted wiThere is a pre-trained word vector matrix of size N x d, where N is the number of words in the word vector matrix and d is the dimension of each word in the word vector matrix. Converting each word into a word vector using a pre-trained pair of word vector matrices, i.e., pairs wiAnd finding out the corresponding word vectors, and splicing the word vectors of all words in the document to obtain the word vector representation of the whole document, wherein the size of the word vector representation is n multiplied by d.
Assuming that the number of hidden units in the bidirectional LSTM is H, the abstract semantic meaning of the document after the bidirectional LSTM is represented as H, and the size of the abstract semantic meaning is n multiplied by 2H.
The obtaining a representation of each sentence in the document based on the abstract semantic representation of the document comprises:
and extracting the composition of each sentence from the abstract semantic representation of the document according to the initial position and the end position of each sentence, performing maximum pooling on the composition of each sentence, and taking a sentence vector obtained after the maximum pooling as the representation of each sentence.
It should be noted that, when the article to be processed sets the start position and the end position of each sentence in advance, it is easy to extract each sentence in the document and obtain the vector representation of each sentence by using the maximum pooling operation.
Assuming that the beginning and the end of a sentence in a document are the a-th word and the b-th word, respectively, the expression of the sentence is:
Figure BDA0002572607970000071
wherein the content of the first and second substances,
Figure BDA0002572607970000072
is 1 × 2 h.
The obtaining a representation of each entity in the document based on the abstract semantic representation of the document comprises:
determining words forming the entity according to the position of each entity, calculating an average word vector of the words, taking the average word vector as an entity vector, and taking the average vector of the entity vectors of the same entity as the representation of each entity.
It should be noted that, assuming that a given target entity appears 3 times, the entities are respectively denoted as e1 1,e2 1And e3 1Then the representation of the current entity
Figure BDA0002572607970000073
Figure BDA0002572607970000074
Is 1 × 2 h.
In step S22, the obtaining a global entity interaction matrix based on the representation of each entity includes:
carrying out information interaction on every two entities, and converting two entity vectors into an entity interaction vector;
and forming a global entity interaction matrix by all entity interaction vectors.
It should be noted that, traversing all entity pairs in the current document, considering the sequence of two entities (i.e. the direction of the entities), and then performing information interaction on every two entities, the interaction method is to transform two entity vectors into a high-level entity interaction vector v through a bilinear layerij
Suppose that two entity vectors are
Figure BDA0002572607970000075
Operation of the bilinear layer is vij=σ(eiWvej+bv) Wherein W isv,bvAre trainable parameters.
All entity interaction vectors are constructed as a global entity interaction matrix. Assuming that there are k entities in the current document, there are k (k-1) interaction vectors in the constructed global entity interaction matrix.
Step S23: calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention; the method comprises the following specific steps:
each sentence representation is subjected to attention mechanism operation with a global entity interaction matrix, so that the method obtains certain logic reasoning capability, and the representation of each sentence based on entity attention is obtained.
Specifically, it can be seen from the foregoing that a sentence is represented as
Figure BDA0002572607970000081
If the entity interaction vector is denoted by v, the attention weight of the current sentence and the entity interaction vector is calculated as
Figure BDA0002572607970000082
Calculating the current sentence and each entity interaction vector to obtain k (k-1) attention weights, then weighting all entity interaction vectors to obtain the representation of each sentence based on the attention of the entity interaction vectors
Figure BDA0002572607970000083
Namely, it is
Figure BDA0002572607970000084
All sentences are combined into a feature matrix X by column.
Step S24: calculating the representation of each sentence based on the entity attention based on a graph convolution network to obtain the representation of each sentence with logical reasoning information; the method comprises the following specific steps:
and (3) carrying out graph convolution operation on the feature matrix X of the sentence through the following transformation: l ═ ρ (AXW)2) Where p is a non-linear activation function, e.g. sigmoid function, W2Are trainable parameters. Thus obtainingL of (a) is the same size as X, where each column is a representation of a sentence, L.
It should be noted that, through a graph convolution neural network layer, the sequential contact information and the logical inference information between each sentence are obtained. After passing through the layer, each sentence representation with the sequential contact information and the logical reasoning information is obtained.
Step S25: and calculating the global entity interaction matrix and the representation of each sentence with the logical reasoning information based on a second attention mechanism to obtain a new global entity interaction matrix with the global information and the logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence. The method specifically comprises the following steps:
specifically, it can be seen from step S24 that a sentence is represented as
Figure BDA0002572607970000085
If the interaction vector of an entity is represented as v, the attention weight of the current interaction vector of the entity and the sentence is calculated as
Figure BDA0002572607970000086
Calculating the current entity interaction vector and each sentence to obtain m attention weights, then weighting all sentence vectors to obtain the representation of each entity interaction vector based on the sentence attention
Figure BDA0002572607970000091
Namely, it is
Figure BDA0002572607970000092
It should be noted that the matrix weights used in the first attention mechanism and the second attention mechanism are the same. This is because for a given two target entities and context, the magnitude of the degree of association between the target entity and the context should be consistent with the magnitude of the degree of association between the context and the target entity, so a constraint is placed on the matrices resulting from the two attention mechanisms to be as close as possible. And through bidirectional attention constraint, the representation of each entity for each sentence is obtained, then a new global entity interaction matrix with global information and logical reasoning information is obtained, and the matrix is used for carrying out classification of relation extraction.
In step S3, determining the relationship type of the target entity pair in the document based on the abstract semantic representation includes:
carrying out column-based weighting operation on the new global entity interaction matrix to obtain a global entity interaction vector;
and inputting the global entity interaction vector into a preset classification function for distinguishing entity pair relationship types to obtain a classification result, and taking the classification result as the relationship type of a target entity pair in the document.
Preferably, the classification function employs a softmax function.
For each entity interaction vector generated from step S25
Figure BDA0002572607970000094
Judging which relationship it belongs to, and using a softmax function:
Figure BDA0002572607970000093
taking the relation type corresponding to the y with the maximum probability value as an entity interaction vector
Figure BDA0002572607970000095
The relationship (2) of (c).
It should be noted that, assuming that there are 3 predefined relationships, a global entity interaction vector will obtain 3 probability values after passing softmax, and which probability is the largest is predicted as which relationship corresponding to the probability values of the 3 relationships. For example, the target entities of a document are: the method comprises the steps of obtaining a relation of belonging, capital and road after a classification function operation is carried out on a Beijing capital airport, a Beijing capital airport and a Beijing subway airport field line by a target entity pair, wherein the corresponding probabilities of the relation are 0.7, 0.2 and 0.1 respectively, and the relation of the target entity pair belongs to the relation.
The implementation is illustrated with respect to fig. 4 as follows:
1. firstly, acquiring representations of all entities (such as Beijing capital airport, Beijing subway airport lines and the like) and sentences in an input document;
2. obtaining an attention weight matrix based on a given target entity and sentence, and a representation of a new entity and sentence;
3. after the new sentence representation passes through the graph convolution neural network, obtaining the sentence representation with the logic reasoning information, and further performing the second attention mechanism calculation with the entity representation;
4. constraining the obtained calculation results of the two attention mechanisms, and calculating the interaction information of each entity pair;
5. and judging whether a given target entity pair has a relationship or not by using the interaction information of each entity pair, and if so, determining which relationship is.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the appended claims.

Claims (10)

1. A chapter-level relation extraction method is characterized by comprising the following steps:
inputting a document to be processed, wherein the document is a chapter-level document;
processing the document based on bidirectional attention constraints to obtain abstract semantic representations of entities and sentences, wherein the abstract semantic representations have global information and logical reasoning information;
and judging the relation type of the target entity pair in the document based on the abstract semantic representation.
2. The extraction method according to claim 1, wherein the processing the document based on the bidirectional attention constraint to obtain an abstract semantic representation of an entity and a sentence comprises:
obtaining a representation of each entity and a representation of each sentence in the document;
obtaining a global entity interaction matrix based on the representation of each entity;
calculating the global entity interaction matrix and the representation of each sentence based on a first attention mechanism to obtain the representation of each sentence based on entity attention;
calculating the representation of each sentence based on the entity attention based on a graph convolution network to obtain the representation of each sentence with logical reasoning information;
and calculating the global entity interaction matrix and the representation of each sentence with the logical reasoning information based on a second attention mechanism to obtain a new global entity interaction matrix with the global information and the logical reasoning information, and taking the new global entity interaction matrix as the abstract semantic representation of the entity and the sentence.
3. The extraction method according to claim 2, wherein the obtaining of the representation of each entity and the representation of each sentence in the document comprises:
acquiring abstract semantic representation of the document;
a representation of each entity and a representation of each sentence in the document are obtained based on the abstract semantic representation of the document.
4. The extraction method according to claim 3, wherein the obtaining a representation of each sentence in the document based on the abstract semantic representation of the document comprises:
and extracting the composition of each sentence from the abstract semantic representation of the document according to the initial position and the end position of each sentence, performing maximum pooling on the composition of each sentence, and taking a sentence vector obtained after the maximum pooling as the representation of each sentence.
5. The extraction method according to claim 3, wherein the obtaining a representation of each entity in the document based on the abstract semantic representation of the document comprises:
determining words forming the entity according to the position of each entity, calculating an average word vector of the words, taking the average word vector as an entity vector, and taking the average vector of the entity vectors of the same entity as the representation of each entity.
6. The extraction method according to claim 3, wherein the obtaining of the abstract semantic representation of the document comprises:
performing word vector conversion on the document to obtain a word vector matrix of the document;
and carrying out bidirectional LSTM operation on the word vector matrix to obtain abstract semantic representation of the document.
7. The extraction method according to claim 5, wherein the obtaining a global entity interaction matrix based on the representation of each entity comprises:
carrying out information interaction on every two entities, and converting two entity vectors into an entity interaction vector;
and forming a global entity interaction matrix by all entity interaction vectors.
8. The extraction method according to claim 2, wherein the first attention mechanism and the second attention mechanism use the same matrix weight.
9. The extraction method according to claim 2, wherein the determining the relationship type of the target entity pair in the document based on the abstract semantic representation comprises:
carrying out column-based weighting operation on the new global entity interaction matrix to obtain a global entity interaction vector;
and inputting the global entity interaction vector into a preset classification function for distinguishing entity pair relationship types to obtain a classification result, and taking the classification result as the relationship type of a target entity pair in the document.
10. The extraction method according to claim 9, characterized in that the classification function employs a softmax function.
CN202010644404.XA 2020-07-07 2020-07-07 Method for extracting chapter-level relation Active CN111831783B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010644404.XA CN111831783B (en) 2020-07-07 2020-07-07 Method for extracting chapter-level relation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010644404.XA CN111831783B (en) 2020-07-07 2020-07-07 Method for extracting chapter-level relation

Publications (2)

Publication Number Publication Date
CN111831783A true CN111831783A (en) 2020-10-27
CN111831783B CN111831783B (en) 2023-12-08

Family

ID=72900251

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010644404.XA Active CN111831783B (en) 2020-07-07 2020-07-07 Method for extracting chapter-level relation

Country Status (1)

Country Link
CN (1) CN111831783B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113297838A (en) * 2021-05-21 2021-08-24 华中科技大学鄂州工业技术研究院 Relationship extraction method based on graph neural network
CN113435190A (en) * 2021-05-18 2021-09-24 北京理工大学 Chapter relation extraction method integrating multilevel information extraction and noise reduction
CN113468325A (en) * 2021-06-09 2021-10-01 广西电网有限责任公司 Document level relation extraction method based on associated sentence selection and relation graph reasoning

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106354710A (en) * 2016-08-18 2017-01-25 清华大学 Neural network relation extracting method
CN108280062A (en) * 2018-01-19 2018-07-13 北京邮电大学 Entity based on deep learning and entity-relationship recognition method and device
US20190370338A1 (en) * 2017-06-22 2019-12-05 Tencent Technology (Shenzhen) Company Limited Summary generation method, apparatus, computer device, and storage medium
CN111008529A (en) * 2019-07-24 2020-04-14 贵州大学 Chinese relation extraction method based on neural network
CN111274800A (en) * 2020-01-19 2020-06-12 浙江大学 Inference type reading understanding method based on relational graph convolution network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106354710A (en) * 2016-08-18 2017-01-25 清华大学 Neural network relation extracting method
US20190370338A1 (en) * 2017-06-22 2019-12-05 Tencent Technology (Shenzhen) Company Limited Summary generation method, apparatus, computer device, and storage medium
CN108280062A (en) * 2018-01-19 2018-07-13 北京邮电大学 Entity based on deep learning and entity-relationship recognition method and device
CN111008529A (en) * 2019-07-24 2020-04-14 贵州大学 Chinese relation extraction method based on neural network
CN111274800A (en) * 2020-01-19 2020-06-12 浙江大学 Inference type reading understanding method based on relational graph convolution network

Non-Patent Citations (10)

* Cited by examiner, † Cited by third party
Title
LEOWOOD: "基于Attention机制的Bi-LSTM关系抽取(分类)读书报告", pages 1, Retrieved from the Internet <URL:https://blog.csdn.net/u014108004/article/details/84141870> *
TIANWEIDADADA: "实体关系抽取:BiLSTM + Attention (含代码+详细注释)", pages 1, Retrieved from the Internet <URL:https://blog.csdn.net/Tianweidadada/article/details/102755094> *
YE, WEI* 等: "Exploiting Entity BIO Tag Embeddings and Multi-task Learning for Relation Extraction with Imbalanced Data", COMPUTER SCIENCE, pages 1 - 10 *
ZHIXIN LI 等: "Improve relation extraction with dual attention-guided graph convolutional networks", NEURAL COMPUTING AND APPLICATIONS, pages 1773 *
凡子威 等: "基于BiLSTM并结合自注意力机制和句法信息的隐式篇章关系分类", 计算机科学, no. 05, pages 214 - 220 *
刘鉴 等: "基于双向LSTM和自注意力机制的中文关系抽取研究", 山西大学学报(自然科学版), no. 01, pages 8 - 13 *
李京谕 等: "基于联合注意力机制的篇章级机器翻译", 中文信息学报, no. 12, pages 45 - 53 *
李智超: "图文知识图谱中的关系抽取算法研究", 中国优秀硕士学位论文全文数据库 信息科技辑, no. 11, pages 138 - 572 *
郭凤羽 等: "基于语境交互感知和模式筛选的隐式篇章关系识别", 计算机学报, no. 05, pages 901 - 915 *
马语丹 等: "结合实体共现信息与句子语义特征的关系抽取方法", 中国科学:信息科学, no. 11, pages 1533 - 1545 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113435190A (en) * 2021-05-18 2021-09-24 北京理工大学 Chapter relation extraction method integrating multilevel information extraction and noise reduction
CN113435190B (en) * 2021-05-18 2022-08-05 北京理工大学 Chapter relation extraction method integrating multilevel information extraction and noise reduction
CN113297838A (en) * 2021-05-21 2021-08-24 华中科技大学鄂州工业技术研究院 Relationship extraction method based on graph neural network
CN113468325A (en) * 2021-06-09 2021-10-01 广西电网有限责任公司 Document level relation extraction method based on associated sentence selection and relation graph reasoning

Also Published As

Publication number Publication date
CN111831783B (en) 2023-12-08

Similar Documents

Publication Publication Date Title
CN106502985B (en) neural network modeling method and device for generating titles
CN110765966B (en) One-stage automatic recognition and translation method for handwritten characters
CN108717574B (en) Natural language reasoning method based on word connection marking and reinforcement learning
WO2018218705A1 (en) Method for recognizing network text named entity based on neural network probability disambiguation
CN111831783B (en) Method for extracting chapter-level relation
CN110287323B (en) Target-oriented emotion classification method
CN111783394B (en) Training method of event extraction model, event extraction method, system and equipment
CN109947931B (en) Method, system, device and medium for automatically abstracting text based on unsupervised learning
CN113312501A (en) Construction method and device of safety knowledge self-service query system based on knowledge graph
CN108009148A (en) Text emotion classification method for expressing based on deep learning
EP4131076A1 (en) Serialized data processing method and device, and text processing method and device
CN110362797B (en) Research report generation method and related equipment
CN114881042B (en) Chinese emotion analysis method based on graph-convolution network fusion of syntactic dependency and part of speech
WO2023159767A1 (en) Target word detection method and apparatus, electronic device and storage medium
Wang et al. A deep reinforcement learning based multi-step coarse to fine question answering (mscqa) system
Shijia et al. Aspect-based Financial Sentiment Analysis with Deep Neural Networks.
CN113946684A (en) Electric power capital construction knowledge graph construction method
CN114239574A (en) Miner violation knowledge extraction method based on entity and relationship joint learning
CN112287106A (en) Online comment emotion classification method based on dual-channel hybrid neural network
CN111597816A (en) Self-attention named entity recognition method, device, equipment and storage medium
CN113987187A (en) Multi-label embedding-based public opinion text classification method, system, terminal and medium
CN111145914B (en) Method and device for determining text entity of lung cancer clinical disease seed bank
CN111597815A (en) Multi-embedded named entity identification method, device, equipment and storage medium
CN114064852A (en) Method and device for extracting relation of natural language, electronic equipment and storage medium
Touati-Hamad et al. Arabic quran verses authentication using deep learning and word embeddings

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant