CN113326371B - Event extraction method integrating pre-training language model and anti-noise interference remote supervision information - Google Patents

Event extraction method integrating pre-training language model and anti-noise interference remote supervision information Download PDF

Info

Publication number
CN113326371B
CN113326371B CN202110480675.0A CN202110480675A CN113326371B CN 113326371 B CN113326371 B CN 113326371B CN 202110480675 A CN202110480675 A CN 202110480675A CN 113326371 B CN113326371 B CN 113326371B
Authority
CN
China
Prior art keywords
training
model
remote supervision
text
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110480675.0A
Other languages
Chinese (zh)
Other versions
CN113326371A (en
Inventor
李书棋
高阳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd
Nanjing University
Original Assignee
Jiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd
Nanjing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd, Nanjing University filed Critical Jiangsu Wanwei Aisi Network Intelligent Industry Innovation Center Co ltd
Priority to CN202110480675.0A priority Critical patent/CN113326371B/en
Publication of CN113326371A publication Critical patent/CN113326371A/en
Application granted granted Critical
Publication of CN113326371B publication Critical patent/CN113326371B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/38Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/383Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention provides an event extraction method integrating a pre-training language model and anti-noise interference remote supervision information, and belongs to the technical field of computers. The method uses an integrated knowledge auxiliary model for judgment, is formed by introducing massive text pre-training, takes a pre-training language model containing a large amount of semantic grammar knowledge information as a network structure unit of an event extraction model, uses a model algorithm of a remote supervision feature for mixing anti-noise interference, and adds gradient direction anti-interference training under a circular constraint condition.

Description

Event extraction method integrating pre-training language model and anti-noise interference remote supervision information
Technical Field
The invention relates to an event extraction method integrating a pre-training language model and anti-noise interference remote supervision information, in particular to the technical field of computer data processing.
Background
With the continuous deepening of informatization construction in the Internet age, massive Internet information presents explosive growth, and how to utilize Internet information to assist in making industry decisions becomes an important point of current attention of enterprises and even government countries. Information from the Internet often appears in the form of words, usually from news manuscripts, forum replies and other channels, generally presents the characteristics of no structure and multiple redundancies, and needs to read, understand and position key information in words and filter irrelevant contents. The event extraction is to present the data content of the unstructured text in a structured form, extract the expression key intention in the text by taking the event as a unit, and convert unstructured text information into structured event information, so that the method is an important ring in the information extraction engineering in a series of works such as subsequent trend analysis, establishment of a priori knowledge map, public opinion message early warning and the like.
Traditional event extraction often relies on manual comprehensive participation, faces to massive internet information, and traditional event analysis reads and searches related information data in a huge article report by means of manual labor, and performs arrangement and recording, so that a large amount of human resources are consumed. In order to solve the problem of consuming a great deal of manpower in the process of information structuring, in recent years, recognition and extraction of event patterns by using a machine learning mode have been proposed. The machine learning mode is used for identifying the event mode in the text, and the text fragments conforming to the mode are extracted in a structured mode, so that the batched machine text processing can be realized, and the efficiency problem in extracting the text structured information by manual reading is greatly improved. However, the formulation of the traditional machine learning event mode template still needs to rely on the knowledge of experts in the field, and corresponding event mode features are automatically learned by means of deep learning through marked data, so that the method becomes a new direction of event structured extraction in recent years. Considering huge internet information and various content types, the migration and generalization capability of the deep learning model among different events is improved, and the method becomes a difficult problem of extracting the internet event information. It is common practice to introduce predictions of the external knowledge-aided model using a remote supervision approach. The remote supervision algorithm assumes: for a structured event in an existing knowledge-graph, it is assumed that any sentence in the external knowledge base that contains entities therein reflects this relationship to some extent. Based on the assumption, the remote supervision algorithm can label sentences in the external document library with relation labels based on a labeled small-sized knowledge graph, which is equivalent to automatic labeling of samples, so that the remote supervision algorithm is a semi-supervision algorithm. However, the remote supervision brings not only external knowledge information, but also erroneous guide information, and the accuracy of judgment of the model is affected by noise interference. Deficiencies in the text representation capabilities of RNNs and CNNs also affect the predictive extraction of events. Therefore, it is a problem to be considered how to use neural network models with more expressive power and use external knowledge to assist in the depth model in event structured extraction, while reducing error noise interference.
Disclosure of Invention
The invention aims to: one objective is to provide an event extraction method integrating a pre-training language model and anti-noise interference remote supervision information, so as to solve the problems in the prior art, enrich text information, and increase the resistance of the model to noise errors through anti-interference training.
The technical scheme is as follows: the event extraction method for fusing the pre-training language model and the anti-noise interference remote supervision information is provided, and comprises the following steps:
1. an event extraction method integrating a pre-training language model and anti-noise interference remote supervision information is characterized by comprising the following steps:
step 1, training data corpus acquisition, namely storing text-form data in a txt file form through a crawler according to internet text data acquired by the crawler;
step 2, preprocessing the marked data;
and 3, marking the text according to the event definition, adding the marked data into a remote supervision knowledge base in a supplementing manner, finishing the marking of the data, and according to 7:1: the proportion of 2 is divided into a training set, a verification set and a test set;
step 4, coding texts based on a pre-training model structure of massive text training, carrying out language modeling through a self-attention mechanism, fully utilizing multi-head attention to capture characteristic information of different angles of the texts, carrying out characteristic transformation and extraction through a double-layer neural network FFN, using a ReLU as an activation function and carrying out characteristic normalization through layer normalization, organically combining all layers of the pre-training model through a residual error connection method, and repeatedly iterating to obtain rich characteristic representations;
step 5, marking the position where the remote supervision trigger word appears by using the type number of the trigger word in the remote supervision library to obtain a discrete sequence feature, mapping the discrete sequence feature to a low-dimensional space through a remote supervision embedding layer, and splicing the feature extracted by the pre-training model and the remote supervision feature;
step 6, introducing disturbance in the gradient direction in the process of learning and parameter updating of the remote supervision embedded layer so as to calculate the counter-propagation gradient of the model in the remote supervision characteristic embedded layer; then, original parameters are reserved, antagonistic disturbance is introduced through disturbance step length, and forward and backward propagation is executed again, so that a new gradient is obtained; then, recovering original parameters of the remote supervision feature embedding layer, and updating parameters of the whole model by using the disturbed new gradient;
step 7, training the event extraction model by using training data, evaluating the quality of the training by verifying a data set and a test data set, and selecting the model with optimal performance for use by multiple rounds of iteration;
and 8, predicting and extracting the event of the new unlabeled internet text data by using the trained model, matching the new text with a remote supervision knowledge base after the new text is subjected to data preprocessing and cleaning, adding trigger words which appear in the remote knowledge base and the new text to be predicted into the remote characteristics of the text, and inputting the trigger type of the predicted event of the model and related event participation elements.
The beneficial effects are that: the invention provides an event extraction method integrating a pre-training language model and anti-noise interference remote supervision information, which is used for carrying out automatic event structured extraction by combining the pre-training language model and the remote supervision information, and improving the model effect by introducing external knowledge and effectively relieving error noise information brought by the remote supervision and improving the event extraction effect by adding a disturbance countermeasure training mode. The method not only can represent richer text information, but also can increase the resistance of the model to noise errors through anti-interference training when the remote supervision features are introduced.
Drawings
FIG. 1 is a block diagram of an event extraction method model incorporating a pre-training language model and anti-noise interference remote supervision information.
FIG. 2 is a schematic diagram of the feature extraction of the attention mechanism of an event extraction method integrating a pre-training language model with anti-noise interference remote supervision information.
FIG. 3 is a diagram of a remote supervision feature layer of an event extraction method incorporating a pre-training language model and anti-noise interference remote supervision information.
FIG. 4 is an experimental effect diagram of an event extraction method integrating a pre-training language model and anti-noise interference remote supervision information.
FIG. 5 is a schematic diagram of an event extraction method incorporating a pre-training language model and anti-noise interference remote supervision information into a circular constraint.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, the present application will be further described in detail with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the present application.
In one embodiment, the invention provides an event extraction method for fusing a pre-training language model and anti-noise interference remote supervision information, which comprises the following steps:
and 1, collecting training data corpus, namely storing text-form data in a txt file form through a crawler according to internet text data obtained through the crawler.
And 2, preprocessing the labeling data, namely removing html labels and special symbols, and dividing the text into short texts in the form of sentences or paragraphs.
Marking event trigger words, subjects, objects, time, places and event types of the text existing events according to event definition, and adding marked data into a remote supervision knowledge base in a supplementing manner to finish marking of the data; matching the marked data with a remote supervision knowledge base, adding successfully matched trigger words into remote supervision information of the sample, and according to 7:1: the scale of 2 is divided into a training set, a validation set and a test set.
And 4, respectively constructing a model in two stages of event detection and event participation element extraction of event extraction.
Specifically, a pre-training model and a remotely supervised event extraction model are combined, as shown in fig. 1, and the specific process is as follows: text is first encoded using a pretrained model structure based on massive text training, as shown in fig. 2, language modeling is performed through a self-attention mechanism:
wherein Q, K, V is a calculation matrix of attention, which is calculated by matrix operation from input data and corresponding parameters; d, d k The text vector representing the text after embedded matrix compression represents the dimension.
Secondly, the characteristic information of multiple angles in the multi-head attention capturing text is as follows:
MultiHead(Q,K,V)=Concat(head1,...,head h )W O
in which W is O Representing a linear transformation matrix, and participating in model training parameter updating; wherein:
in the method, in the process of the invention,a transformation mapping matrix representing Query vectors; />A transformation mapping matrix representing Key vectors; />A transformation mapping matrix representing Value x vectors;
and performing feature conversion extraction through a double-layer neural network FFN, and adopting a ReLU as an activation function layer:
FFN(x)=max(0,xW 1 +b 1 )W 2 +b 2
and performing feature normalization processing by using layer normalization:
wherein W is 1 A calculation parameter representing a first fully connected network in the feed forward network layer; w (W) 2 Representing a calculated parameter of a second fully connected network in the feed forward network layer; μ represents the average value of the vector represented by each word position of the sample, σ represents the variance of the vector represented by each word position of the sample, and α is a learnable weight parameter; e is a small value that prevents zero-divide errors from occurring; as indicated by the letter, the Element-wise Product was multiplied position by position.
From time to time, the layers of the pre-training model are combined by means of residual connection:
y=f(X)+x
x represents the input variable, corresponding to cross-layer addition; x represents the original input of the layer network; and y is the network layer output characteristic, and the characteristics extracted by the pre-training model are obtained by overlapping the above processes for a plurality of times.
And 5, splicing the features extracted by the pre-training model and the remote supervision features, marking the positions where the remote supervision trigger words appear by using the type numbers of the trigger words in the remote supervision library to obtain a discrete sequence feature, mapping the discrete sequence feature to a low-dimensional space through a remote supervision embedding layer, splicing the discrete sequence feature and the pre-training extracted features into a classifier to predict events, and carrying out classification prediction on the positions of each word to determine whether the trigger words are the trigger words.
In a further embodiment, in order to alleviate the false trigger word information caused by the remote supervision, in this patent, an anti-learning strategy is adopted, and in the process of learning and parameter updating by the remote supervision embedded layer, a disturbance in a gradient direction is added by adopting an anti-learning method, so that the anti-noise interference capability of the model is improved, and the method specifically comprises the following steps:
the computing model counter propagates gradients at the remote supervisory feature embedding layer:
the optimal point in disturbance offset constraint is prevented by means of spherical mapping, a disturbance radius epsilon is set, the size of a constraint range of disturbance is represented, and a disturbance step size is obtained:
wherein,for the constraint space of disturbance, +.>Is the step length of the small steps.
The original parameters of the remote supervision characteristic embedding layer are reserved, the obtained disturbance step length is used for adding anti-disturbance, and forward and backward propagation is carried out again to obtain a new gradient.
And recovering original parameters of the remote supervision feature embedding layer, and updating overall parameters of the model by using the new gradient added with disturbance.
The event element extraction network structure is similar to event prediction, only the remote supervision feature embedding layer is removed, and each type of element is predicted at the output layer. Tests prove that the method is superior to other machine learning event extraction methods in terms of accuracy, recall and f1 score, as shown in fig. 4.
And 7, training the event extraction model by using training data, evaluating the quality of the training by verifying the data set and the test data set, and selecting the model with the optimal performance for use by multiple rounds of iteration.
And 8, predicting and extracting the event of the new unlabeled internet text data by using the trained model, matching the new text with a remote supervision knowledge base after the new text is subjected to data preprocessing and cleaning, adding trigger words which appear in the remote knowledge base and the new text to be predicted into the remote characteristics of the text, and inputting the trigger type of the predicted event of the model and related event participation elements.
As described above, although the present invention has been shown and described with reference to certain preferred embodiments, it is not to be construed as limiting the invention itself. Various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (1)

1. An event extraction method integrating a pre-training language model and anti-noise interference remote supervision information is characterized by comprising the following steps:
step 1, training data corpus acquisition, namely storing text-form data in a txt file form through a crawler according to internet text data acquired by the crawler;
step 2, preprocessing the marked data;
and 3, marking the text according to the event definition, adding the marked data into a remote supervision knowledge base in a supplementing manner, finishing the marking of the data, and according to 7:1: the proportion of 2 is divided into a training set, a verification set and a test set;
step 4, coding texts based on a pre-training model structure of massive text training, carrying out language modeling through a self-attention mechanism, fully utilizing multi-head attention to capture characteristic information of different angles of the texts, carrying out characteristic transformation and extraction through a double-layer neural network FFN, using a ReLU as an activation function and carrying out characteristic normalization through layer normalization, organically combining all layers of the pre-training model through a residual error connection method, and repeatedly iterating to obtain rich characteristic representations;
step 5, marking the position where the remote supervision trigger word appears by using the type number of the trigger word in the remote supervision library to obtain a discrete sequence feature, mapping the discrete sequence feature to a low-dimensional space through a remote supervision embedding layer, and splicing the feature extracted by the pre-training model and the remote supervision feature;
step 6, introducing disturbance in the gradient direction in the process of learning and parameter updating of the remote supervision embedded layer so as to calculate the counter-propagation gradient of the model in the remote supervision characteristic embedded layer; then, original parameters are reserved, antagonistic disturbance is introduced through disturbance step length, and forward and backward propagation is executed again, so that a new gradient is obtained; then, recovering original parameters of the remote supervision feature embedding layer, and updating parameters of the whole model by using the disturbed new gradient;
step 7, training the event extraction model by using training data, evaluating the quality of the training by verifying a data set and a test data set, and selecting the model with optimal performance for use by multiple rounds of iteration;
and 8, predicting and extracting the event of the new unlabeled internet text data by using the trained model, matching the new text with a remote supervision knowledge base after the new text is subjected to data preprocessing and cleaning, adding trigger words which appear in the remote knowledge base and the new text to be predicted into the remote characteristics of the text, and inputting the trigger type of the predicted event of the model and related event participation elements.
CN202110480675.0A 2021-04-30 2021-04-30 Event extraction method integrating pre-training language model and anti-noise interference remote supervision information Active CN113326371B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110480675.0A CN113326371B (en) 2021-04-30 2021-04-30 Event extraction method integrating pre-training language model and anti-noise interference remote supervision information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110480675.0A CN113326371B (en) 2021-04-30 2021-04-30 Event extraction method integrating pre-training language model and anti-noise interference remote supervision information

Publications (2)

Publication Number Publication Date
CN113326371A CN113326371A (en) 2021-08-31
CN113326371B true CN113326371B (en) 2023-12-29

Family

ID=77414011

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110480675.0A Active CN113326371B (en) 2021-04-30 2021-04-30 Event extraction method integrating pre-training language model and anti-noise interference remote supervision information

Country Status (1)

Country Link
CN (1) CN113326371B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113762482B (en) * 2021-09-15 2024-04-16 智道网联科技(北京)有限公司 Training method and related device for neural network model for automatic driving

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106874410A (en) * 2017-01-22 2017-06-20 清华大学 Chinese microblogging text mood sorting technique and its system based on convolutional neural networks
CN108182295A (en) * 2018-02-09 2018-06-19 重庆誉存大数据科技有限公司 A kind of Company Knowledge collection of illustrative plates attribute extraction method and system
CN109063185A (en) * 2018-08-27 2018-12-21 电子科技大学 Social networks short text data filter method towards event detection
CN109582949A (en) * 2018-09-14 2019-04-05 阿里巴巴集团控股有限公司 Event element abstracting method, calculates equipment and storage medium at device
CN110188172A (en) * 2019-05-31 2019-08-30 清华大学 Text based event detecting method, device, computer equipment and storage medium
CN111125370A (en) * 2019-12-06 2020-05-08 南京中新赛克科技有限责任公司 Relation extraction method suitable for small samples
CN111339774A (en) * 2020-02-07 2020-06-26 腾讯科技(深圳)有限公司 Text entity relation extraction method and model training method
CN111694924A (en) * 2020-06-17 2020-09-22 合肥中科类脑智能技术有限公司 Event extraction method and system
CN111897908A (en) * 2020-05-12 2020-11-06 中国科学院计算技术研究所 Event extraction method and system fusing dependency information and pre-training language model
CN111914558A (en) * 2020-07-31 2020-11-10 湖北工业大学 Course knowledge relation extraction method and system based on sentence bag attention remote supervision
CN112016293A (en) * 2020-10-22 2020-12-01 浙江大学 Remote supervision relation extraction method based on multi-instance collaborative confrontation training
CN112052665A (en) * 2020-09-12 2020-12-08 广东工业大学 Remote monitoring event extraction method and application thereof
WO2020247616A1 (en) * 2019-06-07 2020-12-10 Raytheon Bbn Technologies Corp. Linguistically rich cross-lingual text event embeddings
CN112307130A (en) * 2020-10-21 2021-02-02 清华大学 Document-level remote supervision relation extraction method and system
AU2020103905A4 (en) * 2020-12-04 2021-02-11 Chongqing Normal University Unsupervised cross-domain self-adaptive medical image segmentation method based on deep adversarial learning
WO2021042503A1 (en) * 2019-09-06 2021-03-11 平安科技(深圳)有限公司 Information classification extraction method, apparatus, computer device and storage medium
CN112487203A (en) * 2019-01-25 2021-03-12 中译语通科技股份有限公司 Relation extraction system integrated with dynamic word vectors
CN112580328A (en) * 2020-12-11 2021-03-30 上海明略人工智能(集团)有限公司 Event information extraction method and device, storage medium and electronic equipment

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106874410A (en) * 2017-01-22 2017-06-20 清华大学 Chinese microblogging text mood sorting technique and its system based on convolutional neural networks
CN108182295A (en) * 2018-02-09 2018-06-19 重庆誉存大数据科技有限公司 A kind of Company Knowledge collection of illustrative plates attribute extraction method and system
CN109063185A (en) * 2018-08-27 2018-12-21 电子科技大学 Social networks short text data filter method towards event detection
CN109582949A (en) * 2018-09-14 2019-04-05 阿里巴巴集团控股有限公司 Event element abstracting method, calculates equipment and storage medium at device
CN112487203A (en) * 2019-01-25 2021-03-12 中译语通科技股份有限公司 Relation extraction system integrated with dynamic word vectors
WO2020239061A1 (en) * 2019-05-31 2020-12-03 腾讯科技(深圳)有限公司 Text-based event detection method and apparatus, computer device and storage medium
CN110188172A (en) * 2019-05-31 2019-08-30 清华大学 Text based event detecting method, device, computer equipment and storage medium
WO2020247616A1 (en) * 2019-06-07 2020-12-10 Raytheon Bbn Technologies Corp. Linguistically rich cross-lingual text event embeddings
WO2021042503A1 (en) * 2019-09-06 2021-03-11 平安科技(深圳)有限公司 Information classification extraction method, apparatus, computer device and storage medium
CN111125370A (en) * 2019-12-06 2020-05-08 南京中新赛克科技有限责任公司 Relation extraction method suitable for small samples
CN111339774A (en) * 2020-02-07 2020-06-26 腾讯科技(深圳)有限公司 Text entity relation extraction method and model training method
CN111897908A (en) * 2020-05-12 2020-11-06 中国科学院计算技术研究所 Event extraction method and system fusing dependency information and pre-training language model
CN111694924A (en) * 2020-06-17 2020-09-22 合肥中科类脑智能技术有限公司 Event extraction method and system
CN111914558A (en) * 2020-07-31 2020-11-10 湖北工业大学 Course knowledge relation extraction method and system based on sentence bag attention remote supervision
CN112052665A (en) * 2020-09-12 2020-12-08 广东工业大学 Remote monitoring event extraction method and application thereof
CN112307130A (en) * 2020-10-21 2021-02-02 清华大学 Document-level remote supervision relation extraction method and system
CN112016293A (en) * 2020-10-22 2020-12-01 浙江大学 Remote supervision relation extraction method based on multi-instance collaborative confrontation training
AU2020103905A4 (en) * 2020-12-04 2021-02-11 Chongqing Normal University Unsupervised cross-domain self-adaptive medical image segmentation method based on deep adversarial learning
CN112580328A (en) * 2020-12-11 2021-03-30 上海明略人工智能(集团)有限公司 Event information extraction method and device, storage medium and electronic equipment

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
A Unified Model for Financial Event Classification, Detection and Summarization;Quanzhi Li et al.;《Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence 》;4668-4674 *
A Unified Model for Financial Event Classification, Detection and Summarization;Quanzhi Li等;Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20);4668-4674 *
基于ELECTRA模型与词性特征的金融事件抽取方法研究;陈星月等;《数据分析与知识发现》;第5卷(第7期);36-47 *
基于ELECTRA模型与词性特征的金融事件抽取方法研究;陈星月等;数据分析与知识发现;第5卷(第07期);36-47 *

Also Published As

Publication number Publication date
CN113326371A (en) 2021-08-31

Similar Documents

Publication Publication Date Title
CN110990564B (en) Negative news identification method based on emotion calculation and multi-head attention mechanism
CN113434357B (en) Log anomaly detection method and device based on sequence prediction
CN109918505B (en) Network security event visualization method based on text processing
CN113742733B (en) Method and device for extracting trigger words of reading and understanding vulnerability event and identifying vulnerability type
CN111274817A (en) Intelligent software cost measurement method based on natural language processing technology
CN116303977B (en) Question-answering method and system based on feature classification
CN111984791A (en) Long text classification method based on attention mechanism
CN113836896A (en) Patent text abstract generation method and device based on deep learning
CN115292568B (en) Civil news event extraction method based on joint model
CN113326371B (en) Event extraction method integrating pre-training language model and anti-noise interference remote supervision information
CN114492460B (en) Event causal relationship extraction method based on derivative prompt learning
CN116089605A (en) Text emotion analysis method based on transfer learning and improved word bag model
CN114881172A (en) Software vulnerability automatic classification method based on weighted word vector and neural network
CN114881173A (en) Resume classification method and device based on self-attention mechanism
CN114648029A (en) Electric power field named entity identification method based on BiLSTM-CRF model
CN114491033A (en) Method for building user interest model based on word vector and topic model
CN114298052A (en) Entity joint labeling relation extraction method and system based on probability graph
Li et al. ADAN: An intelligent approach based on attentive neural network and relevant law articles for charge prediction
CN113535928A (en) Service discovery method and system of long-term and short-term memory network based on attention mechanism
Luo et al. A comparison of som based document categorization systems
CN116821349B (en) Literature analysis method and management system based on big data
CN114036946B (en) Text feature extraction and auxiliary retrieval system and method
Siddiqui et al. Poet Attribution of Urdu Ghazals using Deep Learning
Yang et al. An Anomaly Detection Algorithm for Logs Based on Self-attention Mechanism and BiGRU Model
CN117744657A (en) Medicine adverse event detection method and system based on neural network model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant