CN113222119A - Argument extraction method for multi-view encoder by using topological dependency relationship - Google Patents

Argument extraction method for multi-view encoder by using topological dependency relationship Download PDF

Info

Publication number
CN113222119A
CN113222119A CN202110594279.0A CN202110594279A CN113222119A CN 113222119 A CN113222119 A CN 113222119A CN 202110594279 A CN202110594279 A CN 202110594279A CN 113222119 A CN113222119 A CN 113222119A
Authority
CN
China
Prior art keywords
candidate
argument
node
arguments
entity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110594279.0A
Other languages
Chinese (zh)
Other versions
CN113222119B (en
Inventor
罗森林
祁佳俊
吴舟婷
周妍汝
董勃
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202110594279.0A priority Critical patent/CN113222119B/en
Publication of CN113222119A publication Critical patent/CN113222119A/en
Application granted granted Critical
Publication of CN113222119B publication Critical patent/CN113222119B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/285Clustering or classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/288Entity relationship models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Machine Translation (AREA)

Abstract

The invention relates to an argument extraction method of a multi-view graph encoder by using topological dependency relationship, belonging to the field of natural language processing and machine learning. The method mainly aims to solve the problem that when single-type feature modeling is adopted to extract arguments, argument feature representation of multiple roles is easily interfered by semantic-free association information of candidate arguments, and feature representation of the multiple-role arguments is inaccurate. Firstly, realizing text embedding on a data set by using a BERT pre-training model to obtain a text embedding vector, and triggering a word category embedding vector and an entity category embedding vector; then modeling topological relation entries among the candidate arguments, the entity categories and the trigger words, and constructing a multi-view map information network; and finally, respectively coding the multi-view graphs by using a graph convolution network, aggregating to obtain candidate argument multi-view graph embedded vectors, and classifying and extracting event arguments from the candidate arguments through a SoftMax full-link layer. Experiments are carried out on the ACE2005 English corpus, and results show that the method can achieve a good argument extraction effect.

Description

Argument extraction method for multi-view encoder by using topological dependency relationship
Technical Field
The invention relates to an argument extraction method of a multi-view graph encoder by using topological dependency relationship, belonging to the field of natural language processing and machine learning.
Background
The argument extraction aims at extracting corresponding argument entities in the event sentences and marking argument roles such as time, place, people and the like for the argument entities, so that structured output of unstructured texts containing event information is realized.
In the argument extraction task, one trigger word class corresponds to arguments of several specific roles, a certain argument is represented by an entity of a specific class, and the arguments are connected with the arguments through a syntactic structure, so that a certain topological relation exists among candidate arguments, the entity class and the trigger word.
The triggering word category, the entity category, the candidate argument and other multi-type features in the candidate sentence can provide effective guidance for accurate extraction of the argument, and the existing argument extraction methods are mainly divided into a vector splicing-based method, a sequence modeling-based method and a topological structure construction-based method according to different feature construction modes.
1. Vector splicing-based method
The argument extraction method based on vector splicing realizes the construction of multiple types of features by utilizing the splicing mode of different types of feature vectors. However, the method often utilizes a mode of directly introducing or directly calculating a single-type feature vector to construct features, and the feature construction mode does not consider the guiding effect of a syntactic structure on argument distribution, so that the argument entities in events are difficult to directly and accurately position, and the problem of inaccurate role labeling in candidate entities is caused.
2. Method based on sequence modeling
The argument extraction method based on sequence modeling realizes multi-type feature fusion in a sequence model construction mode. The method utilizes the same characteristic construction mode as the vector splicing method and directly calculates the single type of characteristic vector to construct the characteristic, so that the argument entity of the event is difficult to be directly and accurately positioned, and finally the problem of inaccurate entity role labeling is caused.
3. Method based on topological structure
The argument extraction method based on the topological structure construction mainly researches the construction of the topological structure among different types of features. The method considers the effectiveness of topological structure information on the guidance of argument extraction, but only constructs the syntactic relation among candidate arguments, and does not consider the construction of the topological relation among candidate arguments, a trigger word category and a candidate argument, an entity category, and influences the accuracy of candidate argument feature representation, so that the problems of insufficient utilization of guidance information and low accuracy of argument identification and classification when the candidate arguments corresponding to various roles are subjected to feature modeling are caused.
In summary, in the existing method, only the construction of the dependency relationship between the candidate argument and the candidate argument is usually considered, the trigger word class and the entity class information are often introduced in a vector splicing or sequence modeling manner, and the topological relationship construction between the candidate argument, the trigger word class and the candidate argument-entity class characteristics is not considered, so that when the argument extraction method of the single-type feature modeling is adopted to extract the arguments in the candidate sentences with the same co-occurrence words, the argument feature representation corresponding to multiple roles is easily interfered by the candidate argument without semantic association information, the feature representation of the multi-role argument is inaccurate, and the argument extraction effect is affected.
Disclosure of Invention
The invention aims to provide an argument extraction method of a multi-view graph encoder by using a topological dependency relationship, aiming at the problem that argument feature representation of multiple roles is easily interfered by semantic association information-free candidate arguments when single-type feature modeling is adopted to extract arguments, so that the feature representation of the multi-role arguments is inaccurate.
The design principle of the invention is as follows: firstly, embedding a text into a data set by using a BERT pre-training model; secondly, modeling the correlation among the three types of characteristics of the candidate argument, the trigger word category and the entity category in a multi-view diagram constructing mode; then, encoding three graphs constructed from different angles by using a Graph Convolution Network (GCN), and obtaining a multi-view graph embedding vector of a candidate argument; and finally, classifying and extracting the event arguments through a Softmax full-connection layer.
The technical scheme of the invention is realized by the following steps:
step 1, text embedding is realized on a data set ACE2005 by using a BERT pre-training model.
Step 1.1, regarding the sentence as a sequence formed by words, dividing the words into a group of limited public subword units to obtain word block embedded vectors, and setting [ CLS ] and [ SEP ] labels at the beginning and the end of the sentence respectively.
And step 1.2, coding the position information of the words into a feature vector, wherein the position coding mode is the same as that used in a transform model, and a position embedding vector is obtained by calculating a sine and cosine function.
And 1.3, setting different characteristic values for two different sentences, and distinguishing the two different sentences by setting the different characteristic values to obtain the segmentation embedded vector.
And step 1.4, inputting the word block embedding vector, the position embedding vector and the segmentation embedding vector obtained in the steps 1.1, 1.2 and 1.3 into a BERT model to obtain a text embedding vector.
Step 1.5, embedding 34 trigger word categories defined by the ACE2005 data set by searching a randomly initialized trigger word category vector table to obtain trigger word category embedded vectors.
Step 1.6, the entity scope is labeled by using a BIO labeling strategy, and at the same time, 45 entity categories defined in the ACE2005 are embedded by searching a randomly initialized entity category vector table to obtain entity category embedded vectors.
And 2, modeling the topological relation among the candidate argument, the entity category and the trigger word, and constructing a multi-view diagram.
And 2.1, constructing a candidate argument node-candidate argument node view angle information network graph according to the dependency syntax relationship among the candidate arguments to obtain a corresponding adjacency matrix.
And 2.2, constructing a candidate argument node-trigger word class node view angle information network graph according to the topological relation between the candidate arguments and the trigger word class, and obtaining a corresponding adjacency matrix.
And 2.3, constructing a candidate argument node-entity class node view angle information network graph according to the topological relation between the candidate arguments and the entity class, and obtaining a corresponding adjacency matrix.
And 3, respectively encoding the multi-view graphs by using a Graph Convolution Network (GCN), and aggregating to obtain multi-view graph embedded vectors of candidate arguments.
And 3.1, respectively encoding the candidate argument node-candidate argument node graph, the candidate argument node-trigger word class node graph and the selected argument node-entity class node graph by using GCN to obtain three corresponding network embedded vectors.
And 3.2, aggregating the network embedded vectors of the three views by using a bidirectional gating circulation unit (BiGRU) to obtain the multi-view map embedded vector of the candidate argument.
And 4, classifying and extracting the event arguments from the candidate arguments through a Softmax full-connection layer.
Advantageous effects
Compared with the argument extraction method constructed based on the topological structure, the construction method of the multi-view graph not only utilizes the dependency syntax relationship between the candidate arguments, but also utilizes the topological dependency relationship between the candidate arguments, the trigger word category and the candidate arguments, the entity category, solves the problem of insufficient utilization of guidance information when the candidate arguments of various roles are subjected to feature modeling, and improves the accuracy of argument identification and classification.
Compared with a argument extraction method based on vector splicing or sequence modeling, the method provided by the invention takes the guiding effect of the syntactic structure on argument distribution into consideration, so that the problem of inaccurate role labeling in candidate entities is avoided.
Drawings
FIG. 1 is a schematic diagram of argument extraction method of a multi-view encoder using topological dependency relationship according to the present invention.
Detailed Description
In order to better illustrate the objects and advantages of the present invention, embodiments of the method of the present invention are described in further detail below with reference to examples.
The experimental data is ACE2005 English open corpus, is the mainstream data set of the event extraction task, contains 599 files in total, and covers 6 different fields: broadcast conversion (bc), broadcast news (bn), telestone conversion (cts), newwire (nw), usenet (un) and weblogs (wl). And (3) carrying out test set, verification set and training set division on the original data set, wherein the experimental data division condition is shown in table 1.
TABLE 1 data set partitioning
Figure BDA0003090593950000041
In the experimental process, the entity category embedding dimension and the trigger word category embedding dimension are both set to be 50, and the multi-view map embedding dimension is set to be 200. The number of layers of the Graph Convolution Network (GCN) is 2, and dropout is 0.2. The dropout of BERT is 0.5. The learning rate of the model is 2 e-5.
The test uses Precision (Precision), Recall (Recall) and F1 values to evaluate the effect of argument extraction. Precision represents the proportion of the entity with the correct role label in all the entities predicted as arguments on the premise of correctly classifying the trigger words; recall represents the proportion of the entity with the correct role label in all argument entities on the premise of correct classification of the trigger words; f1 is the harmonic mean of Precision and Recall. Each calculation formula is shown as formula 1, formula 2, and formula 3.
Figure BDA0003090593950000042
Figure BDA0003090593950000043
Figure BDA0003090593950000044
Wherein, TP represents that the real category is a certain argument role, and is also marked as the sample number of the argument role in the prediction result; FP represents that the real category is non-argument and the sample number marked as a certain argument role in the prediction result; FN represents the number of samples of which the real category is a certain argument role and which are marked as non-argument in the prediction result; TN indicates the number of samples whose true category is non-argument and which is also non-argument in the prediction result.
The experiment is carried out on a computer and a server, and the computer is specifically configured as follows: inter i7-6700, CPU 2.40GHz, memory 4G, operating system windows 7, 64 bit; the specific configuration of the server is as follows: e7-4820v4, RAM 256G, operating system is Linux Ubuntu 64 bit.
The specific process of the experiment is as follows:
step 1, text embedding is realized on a data set ACE2005 by using a BERT pre-training model.
Step 1.1, sentence
Figure BDA0003090593950000051
Is regarded as being composed of NwDividing a word into a limited group of common sub-word units to obtain word block embedding vectors, and setting [ CLS ] at the beginning and the end of a sentence respectively]And [ SEP ]]And (4) a label.
And step 1.2, coding the position information of the words into a feature vector, wherein the position coding mode is the same as that used in a transform model, and a position embedding vector is obtained by calculating a sine and cosine function.
And 1.3, setting different characteristic values for two different sentences, and distinguishing the two different sentences by setting the different characteristic values to obtain the segmentation embedded vector.
Step 1.4, inputting the word block embedding vector, the position embedding vector and the segmentation embedding vector obtained in the steps 1.1, 1.2 and 1.3 into a BERT model to obtain a text embedding vector
Figure BDA0003090593950000052
Step 1.5, embedding 34 trigger word categories defined by the ACE2005 data set by searching a randomly initialized trigger word category vector table to obtain trigger word category embedding directionsMeasurement of
Figure BDA0003090593950000053
Figure BDA0003090593950000054
Step 1.6, labeling entity scope by using BIO labeling strategy, and embedding 45 entity categories defined in ACE2005 by searching randomly initialized entity category vector table to obtain entity category embedded vector
Figure BDA0003090593950000055
And 2, modeling the topological relation among the candidate argument, the entity category and the trigger word, and constructing a multi-view diagram.
Step 2.1, constructing a candidate argument node-candidate argument node view angle information network graph according to the dependency syntax relation among the candidate arguments
Figure BDA0003090593950000056
Wherein upsilon iswwIs a node, epsilonwwIs an edge. First, a dependency syntax tree of a candidate sentence is generated by using a Stanford Parser dependency syntax analysis tool, and a dependency syntax relation R (w) of a candidate primitive layer is determinedi,wj) Building edge (w)i,wj) While adding a reversal edge (w)j,wi) And self-ringing side (w)i,wi) The calculation mode of the candidate argument-candidate argument boundary is as follows:
Figure BDA0003090593950000057
Figure BDA0003090593950000058
finally, an adjacency matrix of the dependency relationship between the candidate argument nodes is obtained
Figure BDA0003090593950000061
Wherein n iswThe number of the candidate word layer nodes is obtained.
In the step 2.2, the step of the method,constructing a candidate argument node-trigger word class node view angle information network graph according to the topological relation between the candidate arguments and the trigger word classes
Figure BDA0003090593950000062
Wherein upsilon iswtIs a node, epsilonwtIs an edge. The edge construction rule is to judge whether the current candidate argument is a trigger word, and if the current candidate argument is the trigger word, an edge is established between the current candidate argument and the trigger word class node to which the current candidate argument belongs. According to the figure
Figure BDA0003090593950000063
Obtaining an adjacency matrix of dependency relationships between candidate argument nodes and trigger word class nodes
Figure BDA0003090593950000064
Wherein n iswIs the number of candidate word level nodes, ntThe number of the trigger word category layer nodes.
Step 2.3, constructing a candidate argument node-entity category node view angle information network graph according to the topological relation between the candidate arguments and the entity categories
Figure BDA0003090593950000065
Wherein upsilon isweIs a node, epsilonweIs an edge. The edge construction rule is used for judging whether the current candidate argument belongs to a certain entity class, and if the dependency relationship exists, an edge is established between a candidate argument node and an entity class node to which the argument node belongs according to the BI label. According to the figure
Figure BDA0003090593950000066
The adjacency matrix for obtaining the dependency relationship between the candidate argument nodes and the entity class nodes is
Figure BDA0003090593950000067
Wherein n iswIs the number of candidate word level nodes, neThe number of the nodes of the entity category layer.
And 3, respectively encoding the multi-view graphs by using a Graph Convolution Network (GCN), and aggregating to obtain multi-view graph embedded vectors of candidate arguments.
Step 3.1, using GCN to respectively encode candidate argument nodes-candidate argument node graph, candidate argument nodes-trigger word class node graph and selected argument nodes-entity class node graph, wherein the encoding process is as follows: h(l+1)=σ(M-1/2A′M-1/2·H(l)·W(l)) Where a' is a + I, a being an adjacency matrix derived from a multi-view map
Figure BDA0003090593950000068
And
Figure BDA0003090593950000069
i is an identity matrix representing self-connection, W(l)Is the weight matrix of the l-th layer, σ (-) represents the activation function, H(l)Initializing H for hidden layer representation of layer I node(0)=X,
Figure BDA00030905939500000610
For a regularized laplacian matrix, the formula is calculated as:
Figure BDA00030905939500000611
respectively carrying out GCN coding on adjacent matrixes of the multi-view map to obtain three corresponding network embedded vectors
Figure BDA00030905939500000612
And
Figure BDA00030905939500000613
step 3.2, aggregating the network embedded vectors of three views by using a bidirectional gating circulation unit (BiGRU), wherein the formula is as follows:
Figure BDA00030905939500000614
wherein
Figure BDA00030905939500000615
Finally, obtaining a multi-view map embedding vector H of candidate argumentmpge
And 4, classifying and extracting the event arguments from the candidate arguments through a Softmax full-connection layer.
And (3) testing results: according to the method, the event argument extraction is carried out on the ACE2005 English corpus by using the argument extraction method of the topological dependency relationship multi-view graph encoder, the accuracy of argument extraction on the ACE2005 English corpus is 61.4%, the recall rate is 62.6%, the F1 value is 62%, the recall rate and the F1 are respectively improved by 4.5% and 0.4% compared with a PLMEE-model, and the event argument extraction effect is improved.
The above detailed description is intended to illustrate the objects, aspects and advantages of the present invention, and it should be understood that the above detailed description is only exemplary of the present invention and is not intended to limit the scope of the present invention, and any modifications, equivalents, improvements and the like made within the spirit and principle of the present invention should be included in the scope of the present invention.

Claims (4)

1. An argument extraction method for a multi-view encoder using topological dependency, the method comprising the steps of:
step 1, embedding texts into a data set ACE2005 by using a BERT pre-training model, firstly, sentences
Figure FDA0003090593940000011
Is regarded as being composed of NwA sequence of words is obtained by dividing a word into a limited set of common sub-word units to obtain word block embedding vectors, and [ CLS ] is set at the beginning and end of a sentence respectively]And [ SEP ]]The label is used for coding the position information of the word into a characteristic vector in the same position coding mode as that used in the Transformer model, obtaining a position embedding vector through sine and cosine function calculation, setting different characteristic values for two different sentences, distinguishing the two different sentences through setting different characteristic values to obtain a segmentation embedding vector, and then inputting the obtained word block embedding vector, the position embedding vector and the segmentation embedding vector into the BERT model to obtain a text embedding vector
Figure FDA0003090593940000012
Figure FDA0003090593940000013
Finally, embedding 34 trigger word categories defined by the ACE2005 data set by searching a randomly initialized trigger word category vector table to obtain trigger word category embedded vectors
Figure FDA0003090593940000014
Simultaneously, the BIO marking strategy is used for marking the entity range, 45 entity categories defined in the ACE2005 are embedded by searching a randomly initialized entity category vector table, and entity category embedded vectors are obtained
Figure FDA0003090593940000015
Step 2, modeling topological relations among candidate arguments, entity categories and trigger words, and constructing a multi-view map, wherein a candidate argument node-candidate argument node view information network map is constructed according to a dependency syntax relation among the candidate arguments, then a candidate argument node-trigger word category node view information network map is constructed according to a topological relation between the candidate arguments and the trigger word categories, and finally a candidate argument node-trigger word category view information network map is constructed according to a topological relation between the candidate arguments and the trigger word categories;
step 3, respectively encoding the multi-view map by using a Graph Convolution Network (GCN) to obtain multi-view map embedded vectors of candidate arguments, firstly, respectively encoding candidate argument nodes-candidate argument node maps, candidate argument nodes-trigger word class node maps and selected argument nodes-entity class node maps by using the GCN, wherein the encoding process is as follows: h(l+1)=σ(M-1/2A′M-1/2·H(l)·W(l)) Where a' is a + I, a being an adjacency matrix derived from a multi-view map
Figure FDA0003090593940000016
And
Figure FDA0003090593940000017
i is an identity matrix representing self-connection, W(l)Is the weight matrix of the l-th layer, σ (-) represents the activation function, H(l)Initializing H for hidden layer representation of layer I node(0)=X,
Figure FDA0003090593940000018
For the regularized Laplacian matrix, adjacent matrixes of the multi-view image are respectively subjected to GCN coding to obtain three corresponding network embedded vectors
Figure FDA0003090593940000019
And
Figure FDA00030905939400000110
then, aggregating the network embedded vectors of three views by using a bidirectional gating cycle unit (BiGRU), wherein the formula is as follows:
Figure FDA00030905939400000111
Figure FDA00030905939400000112
wherein
Figure FDA00030905939400000113
Finally, obtaining a multi-view map embedding vector H of candidate argumentmpge
And 4, classifying and extracting the event arguments from the candidate arguments through a Softmax full-connection layer.
2. The argument extraction method of utilizing a topology dependency multiview encoder according to claim 1, wherein: in step 2, a dependency syntax tree of the candidate sentence is generated by using a Stanford Parser dependency syntax analysis tool, and the dependency syntax relation R (w) of the candidate primitive layer is usedi,wj) Building edge (w)i,wj) While adding a reversal edge (w)j,wi) And self-ringing side (w)i,wi) Constructing a candidate argument node-candidate argument node view angle information network graph ζww=(υww,εww) Wherein upsilon iswwIs a node, epsilonwwTo the side, use
Figure FDA0003090593940000021
Figure FDA0003090593940000022
The edges between the candidate arguments and the candidate arguments are constructed in a computing mode, and finally an adjacent matrix of the dependency relationship between the candidate arguments and the candidate argument nodes is obtained
Figure FDA0003090593940000023
Wherein n iswThe number of the candidate word layer nodes is obtained.
3. The argument extraction method of utilizing a topology dependency multiview encoder according to claim 1, wherein: in step 2, a candidate argument node-trigger word category node visual angle information network graph is constructed by judging whether a current candidate argument is a trigger word or not, if so, establishing an edge construction rule of one edge between the current candidate argument and the trigger word category node to which the current candidate argument is a trigger word, and constructing the candidate argument node-trigger word category node visual angle information network graph
Figure FDA0003090593940000024
Wherein upsilon iswtIs a node, epsilonwtIs an edge according to the figure
Figure FDA0003090593940000025
Obtaining an adjacency matrix of dependency relationships between candidate argument nodes and trigger word class nodes
Figure FDA0003090593940000026
Wherein n iswIs the number of candidate word level nodes, ntTo trigger word class level nodesAnd (4) the number.
4. The argument extraction method of utilizing a topology dependency multiview encoder according to claim 1, wherein: step 2, establishing a candidate argument node-entity category node visual angle information network graph by judging whether the current candidate argument belongs to a certain entity category or not and establishing an edge construction rule of one edge between a candidate argument node and an entity category node if an affiliation exists and according to BI labels
Figure FDA0003090593940000027
Figure FDA0003090593940000028
Wherein upsilon isweIs a node, epsilonweIs an edge according to the figure
Figure FDA0003090593940000029
The adjacency matrix for obtaining the dependency relationship between the candidate argument nodes and the entity class nodes is
Figure FDA00030905939400000210
Wherein n iswIs the number of candidate word level nodes, neThe number of the nodes of the entity category layer.
CN202110594279.0A 2021-05-28 2021-05-28 Argument extraction method for multi-view encoder by using topological dependency relationship Active CN113222119B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110594279.0A CN113222119B (en) 2021-05-28 2021-05-28 Argument extraction method for multi-view encoder by using topological dependency relationship

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110594279.0A CN113222119B (en) 2021-05-28 2021-05-28 Argument extraction method for multi-view encoder by using topological dependency relationship

Publications (2)

Publication Number Publication Date
CN113222119A true CN113222119A (en) 2021-08-06
CN113222119B CN113222119B (en) 2022-09-20

Family

ID=77099220

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110594279.0A Active CN113222119B (en) 2021-05-28 2021-05-28 Argument extraction method for multi-view encoder by using topological dependency relationship

Country Status (1)

Country Link
CN (1) CN113222119B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103530281A (en) * 2013-10-15 2014-01-22 苏州大学 Argument extraction method and system
CN110134757A (en) * 2019-04-19 2019-08-16 杭州电子科技大学 A kind of event argument roles abstracting method based on bull attention mechanism
CN111897908A (en) * 2020-05-12 2020-11-06 中国科学院计算技术研究所 Event extraction method and system fusing dependency information and pre-training language model
CN112084746A (en) * 2020-09-11 2020-12-15 广东电网有限责任公司 Entity identification method, system, storage medium and equipment
CN112084381A (en) * 2020-09-11 2020-12-15 广东电网有限责任公司 Event extraction method, system, storage medium and equipment
CN112163416A (en) * 2020-10-09 2021-01-01 北京理工大学 Event joint extraction method for merging syntactic and entity relation graph convolution network

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103530281A (en) * 2013-10-15 2014-01-22 苏州大学 Argument extraction method and system
CN110134757A (en) * 2019-04-19 2019-08-16 杭州电子科技大学 A kind of event argument roles abstracting method based on bull attention mechanism
CN111897908A (en) * 2020-05-12 2020-11-06 中国科学院计算技术研究所 Event extraction method and system fusing dependency information and pre-training language model
CN112084746A (en) * 2020-09-11 2020-12-15 广东电网有限责任公司 Entity identification method, system, storage medium and equipment
CN112084381A (en) * 2020-09-11 2020-12-15 广东电网有限责任公司 Event extraction method, system, storage medium and equipment
CN112163416A (en) * 2020-10-09 2021-01-01 北京理工大学 Event joint extraction method for merging syntactic and entity relation graph convolution network

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
SHIYAO CUI等: "Edge-Enhanced Graph Convolution Networks for Event Detection with Syntactic Relation", 《ARXIV》 *
XIANGBIN MENG等: "Multi-Graph Convolution Network with Jump Connection for Event Detection", 《2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI)》 *
潘丽敏等: "融合多级语义特征的双通道GAN事件检测方法", 《北京理工大学学报》 *
程思伟等: "BGCN:基于BERT和图卷积网络的触发词检测", 《计算机科学》 *
黄媛等: "一个基于语义的中文事件论元抽取方法", 《计算机科学》 *

Also Published As

Publication number Publication date
CN113222119B (en) 2022-09-20

Similar Documents

Publication Publication Date Title
Du et al. Stance classification with target-specific neural attention networks
Tian et al. Improving Chinese word segmentation with wordhood memory networks
CN111931506B (en) Entity relationship extraction method based on graph information enhancement
CN109255047A (en) Based on the complementary semantic mutual search method of image-text being aligned and symmetrically retrieve
CN111160471A (en) Method and device for processing point of interest data, electronic equipment and storage medium
CN115357719B (en) Power audit text classification method and device based on improved BERT model
CN112541355A (en) Few-sample named entity identification method and system with entity boundary class decoupling
CN111563384A (en) Evaluation object identification method and device for E-commerce products and storage medium
CN115081437B (en) Machine-generated text detection method and system based on linguistic feature contrast learning
CN113191148A (en) Rail transit entity identification method based on semi-supervised learning and clustering
CN113360582B (en) Relation classification method and system based on BERT model fusion multi-entity information
CN116661805B (en) Code representation generation method and device, storage medium and electronic equipment
CN112580330A (en) Vietnamese news event detection method based on Chinese trigger word guidance
CN114997288A (en) Design resource association method
CN113988075A (en) Network security field text data entity relation extraction method based on multi-task learning
CN112699685A (en) Named entity recognition method based on label-guided word fusion
CN113705222B (en) Training method and device for slot identification model and slot filling method and device
CN113947087B (en) Label-based relation construction method and device, electronic equipment and storage medium
CN113901224A (en) Knowledge distillation-based secret-related text recognition model training method, system and device
CN113901813A (en) Event extraction method based on topic features and implicit sentence structure
CN113222119B (en) Argument extraction method for multi-view encoder by using topological dependency relationship
Niu et al. Word embedding based edit distance
CN112507707A (en) Correlation degree analysis and judgment method for innovative technologies in different fields of power internet of things
Jiang et al. A Discourse Coherence Analysis Method Combining Sentence Embedding and Dimension Grid
CN110909547A (en) Judicial entity identification method based on improved deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant