CN112084330A - Incremental relation extraction method based on course planning meta-learning - Google Patents

Incremental relation extraction method based on course planning meta-learning Download PDF

Info

Publication number
CN112084330A
CN112084330A CN202010806791.2A CN202010806791A CN112084330A CN 112084330 A CN112084330 A CN 112084330A CN 202010806791 A CN202010806791 A CN 202010806791A CN 112084330 A CN112084330 A CN 112084330A
Authority
CN
China
Prior art keywords
learning
training
task
meta
instance
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010806791.2A
Other languages
Chinese (zh)
Inventor
李学恺
吴桐桐
漆桂林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202010806791.2A priority Critical patent/CN112084330A/en
Publication of CN112084330A publication Critical patent/CN112084330A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses an incremental relation extraction method based on course planning meta-learning, which comprises the following steps of: constructing a course planning meta-learning incremental relation extraction framework; calculating the similarity among the relations, and replanning and memorizing the cache training sequence according to the similarity; training the relation extraction neural network model, and updating the parameters of the relation extraction neural network model by using a meta-learning parameter updating formula in combination with the meta-learning thought; selecting a batch of memory caches from the training data by using a clustering method; and loading the updated parameters into the relation extraction neural network model, and performing performance test. The invention introduces the ideas of course learning and meta learning into the task of extracting the increment relation, effectively improves the average accuracy and the overall accuracy of the model, and reduces the error range of the model, thereby effectively relieving the problems of 'catastrophic forgetting' and 'order sensitivity' in the increment learning and obtaining the effect superior to the traditional method on the mainstream data set.

Description

Incremental relation extraction method based on course planning meta-learning
Technical Field
The invention belongs to the field of computer natural language processing, and particularly relates to an incremental relation extraction method based on course planning meta-learning.
Background
Relationship extraction is an important task in the field of information extraction, and aims to extract semantic relationships between pairs of specified entities from natural language text. For example, in the example given in fig. 1, there is a relationship "located" between "southeast university" and "Nanjing". The relation extraction has wide application scenarios in subsequent applications such as automatic construction of knowledge graphs, automatic question answering and text mining.
The traditional supervised learning relationship extraction method has a good effect on the task and is widely used. However, in practical use scenarios, the construction of the knowledge-graph is often a continuous, iteratively updated, incremental process. New relationship categories will be added continuously, which puts higher demands on the relationship extraction system, namely the ability to process incremental data. Conventional supervised learning relationship extraction methods are still limited in this regard.
The traditional supervised learning relationship extraction method is set based on supervised learning, and all possible relationships are required to be defined in advance in model training and are trained by using a full data set. This makes the traditional relational extraction method require retraining the classification model using the full amount of data including incremental data when faced with new data, which takes a lot of effort and time cost. Some models have the capability of handling new relationship classes, such as metric learning models, but if incremental data is used directly to train the model to identify new classes, serious "catastrophic forgetting" problems can result, i.e., once an existing model is trained using a new data set, the model loses the capability of identifying the original data set, which is also a challenge in incremental learning.
Another important issue in incremental learning relationship extraction is "order sensitivity," i.e., the neural network of relationship extraction continuously processes a series of tasks during an update iteration, each task containing a different relationship class and training data. The performance of the final system is not only affected by the division of the training data, but also greatly affected by the training sequence of each task in the training process. This makes the performance of the relationship extraction system unstable, which makes the quality of the subsequent knowledge-graph construction work difficult to control.
Disclosure of Invention
The purpose of the invention is as follows: in order to overcome the defects in the prior art, the method for extracting the incremental relationship based on the course planning meta-learning is provided, the method is based on the meta-learning, the course planning method is creatively added, and the problems of 'catastrophic forgetting' and 'order sensitivity' can be better solved.
The technical scheme is as follows: in order to achieve the above object, the present invention provides an incremental relationship extraction method based on curriculum planning meta-learning, which comprises the following steps:
s1: constructing a course planning meta-learning incremental relation extraction framework; the incremental relation extraction framework comprises a relation extraction neural network model, a memory cache list, a network model training method, a meta learning parameter updating formula, a memory cache screening method and a memory training course planning method;
s2: calculating the similarity among the relations, and replanning and memorizing the cache training sequence according to the similarity;
s3: training the relation extraction neural network model, and updating the parameters of the relation extraction neural network model by using a meta-learning parameter updating formula in combination with the meta-learning thought;
s4: selecting a batch of memory caches from the training data in the step S3 by using a clustering method;
s5: loading the updated parameters into a relation extraction neural network model, and carrying out performance test;
s6: steps S3-S5 are repeated until no new tasks need to be processed.
Further, the updating the parameters of the relation extraction neural network model by the meta-learning parameter updating formula in step S3 specifically includes: and substituting the optimal parameters on the current task and the parameters of the previous task into a meta-learning parameter updating formula for updating.
Further, the selecting method of the memory cache in step S4 specifically includes:
a1: extracting a hidden layer of the neural network model from the relationship to obtain vector representations of natural language sentences of examples in all training data;
a2: clustering all vector representations into categories with required memory quantity by using a K-Means clustering algorithm, wherein each category has vector representation of sentences of at least one example; for the category with only one instance, taking the instance of the category as a memory; for a category of more than one instance, calculating the arithmetic mean of the vector representations of all instance sentences under the category as the vector centers of all instance sentences of the category, and taking the instance with the closest Euclidean distance of the instance sentence vector from the vector center as a memory instance.
Further, the calculation process of the inter-relationship similarity in step S2 is as follows: the relation label character is R, and the R is divided into a list W composed of a plurality of single characters by using a word segmentation tool1,W2,…,Wn]Obtaining word vector list of single word list [ E1,E2,…,En]Taking the mean value of the word vector list as the vector representation of the relation R, and calculating the relation R pair by pairiAnd RjCosine similarity between them, denoted Simi,jAnd storing the data in a similarity matrix Sim.
Further, the process of programming the memory buffer training sequence in step S2 is as follows: when the Task list is not empty, the first Task in the list is taken outiAll training data of
Figure BDA0002629427800000021
Contains k relation classes
Figure BDA0002629427800000022
Will be provided with
Figure BDA0002629427800000023
Added to senrrelations (sets of known relationships) that are used to deposit relationships that have already occurred.
Further, the training process of the relation extraction neural network model in step S3 is as follows:
each relationship class has NtrainStrip training deviceExample, NvalidStrip verification example and NtestThe test example strips labels which do not exist in the senrelationships in the error relationship label list in the three groups of examples; will NtrainStripe training instance partitioning into size NbatchThe training batch of (1), i.e. the training data is divided into a length of
Figure BDA0002629427800000031
The Ceil function represents rounding up;
task for the first Task0Circularly taking out each training batch TrainBatch, and extracting a neural network by training relations;
task for non-first TaskiCyclically fetching each training batch TrainBatch and cyclically fetching a memory instance group M of one task from the MemoryListjEach instance in the memory instance group comprises a relation category label, the similarity of the relation category label of each instance in the memory instance group and the relation category label of the instance in the training data batch is obtained from the similarity matrix Sim, the memory instance groups are arranged from small to large by using the similarity, and the memory instance group M 'is ordered by using the similarity'jTraining the relation extraction neural network, and then training the relation extraction neural network by using a training batch TrainBatch;
task for any TaskiAfter one round of training data lists per complete cycle, use the verification example
Figure BDA0002629427800000032
Evaluating the network training condition of the current task and saving the parameter theta of the round of network with the best performanceiWhen the network does not obtain better effect in multiple continuous turns on the current task, the network parameters are updated
Figure BDA0002629427800000033
Further, the meta learning parameter update formula in step S3 includes a linear parameter update formula, a fixed parameter update formula, and a square root parameter update formula. Different formulas can be adopted on different data sets, certain influence is exerted on the final performance, and the method belongs to one of adjustable hyper-parameters. Three formulas can be used for experiments according to actual application scenes to finally determine a specific formula form.
Further, the performance test in step S4 is performed in the form of average accuracy and overall accuracy of the test model on the test data of all previous tasks.
Further, the calculation formula of the average accuracy in step S4 is as follows:
Figure BDA0002629427800000034
wherein, acciTo be at TaskiThe accuracy on the test case set of (1);
the overall accuracy is calculated as follows:
Figure BDA0002629427800000041
wherein the content of the first and second substances,
Figure BDA0002629427800000042
is the accuracy on all test data.
Further, the error range of the performance index in step S4 is calculated by: re-dividing tasks or disordering task sequences, measuring performance under other task divisions or task sequences, counting average accuracy and overall accuracy under various task divisions and sequences, and calculating error range of performance index according to the following formula:
Figure BDA0002629427800000043
wherein the content of the first and second substances,
Figure BDA0002629427800000044
representing the confidence coefficient when the confidence is alpha, representing the statistical sampleAnd n is the number of statistical samples.
The incremental learning relationship extraction is a relationship extraction setting closer to the actual application scene. In incremental learning relationship extraction, data is divided according to tasks, each task comprises a group of relationship categories different from other tasks, and each relationship category comprises a group of training examples DtrainA set of verification instances DvalidAnd a set of test cases Dtest. Each instance is composed of a natural language sentence S, a pair of entities (H, T) contained in the sentence, correct relation labels (R) existing between the entities and error relation labels [ R ] not existing between a plurality of entities1,R2,…,Rs]And (4) forming. The artificial neural network model processes each task in sequence according to a certain sequence, and trains until no new task exists, and at the moment, the model can judge all the existing relations.
For the incremental learning relation extraction method, three indexes can measure the performance, the overall accuracy, the average accuracy and the error range. The overall accuracy rate is the accuracy rate obtained by the method on all the trained relation test sets, and reflects the performance level of the method; the average accuracy is the arithmetic average of the accuracies obtained on the test set of all processed tasks, reflecting the 'catastrophic forgetting' resistance of the method; the error range reflects the random fluctuation of the overall accuracy and the average accuracy of the method on a task sequence formed by dividing and sequencing a plurality of groups of different data, and reflects the order sensitivity resistance of the method.
Has the advantages that: compared with the prior art, the method introduces the curriculum learning and the meta learning ideas into the incremental relation extraction task, effectively improves the average accuracy and the overall accuracy of the model, and reduces the error range of the model, thereby effectively relieving the problems of 'catastrophic forgetting' and 'order sensitivity' in incremental learning, and obtaining the effect superior to the traditional method on the mainstream data set.
Drawings
FIG. 1 is a specific example of relationship extraction described in the background;
FIG. 2 is an alternative formula for updating neural network model parameters in the present invention;
FIG. 3 is a flow diagram of an incremental relationship extraction framework based on curriculum planning meta-learning, in accordance with an embodiment of the present invention;
FIG. 4 is a structure of a relationship extraction neural network model in an embodiment of the present invention;
FIG. 5 is a graph comparing the performance of the method of the present invention with currently existing methods.
Detailed Description
The present invention is further illustrated by the following figures and specific examples, which are to be understood as illustrative only and not as limiting the scope of the invention, which is to be given the full breadth of the appended claims and any and all equivalent modifications thereof which may occur to those skilled in the art upon reading the present specification.
In this embodiment, a data set, Lifelong FewRel, is extracted based on an incremental relationship, a BiLSTM model is used as a relationship extraction neural network model, the structure of which is shown in fig. 4, and GloVe word embedding vectors are used as pre-training word vectors to encode words in sentences.
In this embodiment, as shown in fig. 3, the incremental relationship extraction method based on course planning meta learning provided by the present invention specifically includes the following steps:
step 1) dividing the data set according to tasks.
Reading a Life Long FewRel data set, and listing the relation by using a word segmentation tool [ R ]1,R2,…,Rn]Dividing words in all relations, encoding each word in the relation by using GloVe word embedded vectors, averaging the divided words belonging to the relations one by one to obtain a vector representation list [ E ] of the relations1,E2,…,En]. The vector list of relationships is clustered into 10 different clusters using the K-Means algorithm, so that all relationships can be divided into 10 different clusters.
According to the division, training data, verification data and test data in the data set are divided into 10 tasks according to the correct relation labels of the examples.
Initializing BilSTM model parameters theta as relational extraction neural networks0
An empty list is initialized as a memory list MemoryList for storing the set of memory instances screened after each task is trained.
An empty list senrrelations is initialized to store the correct relationship labels contained in all the trained instances.
And 2) calculating the similarity among the relation categories.
Representing the list [ E ] by using the relation vector obtained in the step 1)1,E2,…,En]Get E pair by pairiAnd EjCalculating cosine similarity, and recording as Simi,jAnd storing the data in a similarity matrix Sim.
Step 3) when the Task list is not empty, taking out the first Task in the listiAll training data of
Figure BDA0002629427800000051
Contains k relation classes
Figure BDA0002629427800000052
Will be provided with
Figure BDA0002629427800000053
Added to SeenRelations. And eliminating labels which are not in SeneRelations in the error relation label lists in the training examples, the verification examples and the test examples, and dividing the training data into training batches with the size of 50.
Task for the first Task0And circularly taking out each training batch TrainBatch, segmenting the sentence S by using a segmentation tool for the training examples in one training batch, and encoding the sentence S by using a GloVe word vector. And taking the coded sentences and the correct relation class labels in the training examples as a group of training positive examples, and forming the training negative examples one by the coded sentences and the error relation labels in the training examples. Using the processed sentence code as the input of the relation extraction neural network BilSTM model, and using the positive example label or the negative example labelAs a label of the model, MarginLoss is used as a model loss function, and the Adam algorithm is used for carrying out back propagation optimization on the model.
Task for non-first TaskiCyclically fetching each training batch TrainBatch and cyclically fetching a memory instance group M of one task from the MemoryListjEach instance in the memory instance group comprises a relationship class label, the similarity between the relationship class label of each instance in the memory instance group and the relationship class label of the instance in the training data batch is obtained from the similarity matrix Sim, the memory instance groups are arranged from small to large by using the similarity, and the sorted memory instance group M 'is respectively encoded and processed according to the method'jThe BiSLTM model was trained using the processed data from a normal training batch TrainBatch.
Task for any TaskiAfter one round of training data lists per complete cycle, use the verification example
Figure BDA0002629427800000061
Evaluating the network training condition of the current task and saving the parameter theta of the round of network with the best performanceiWhen the network does not achieve better effect on the current task for multiple successive rounds, the parameters on the current task can be updated using one of the formulas shown in FIG. 2
Figure BDA0002629427800000062
In the embodiment, the linear parameter update formula in fig. 2 is used to update the parameters on the current task
Figure BDA0002629427800000063
Figure BDA0002629427800000064
Wherein the content of the first and second substances,
Figure BDA0002629427800000065
representing the parameters of the model after the last task is completed, wherein the epsilon is a fixed constant,n is the total number of tasks.
Step 4) Task from current completion trainingiTraining data of
Figure BDA0002629427800000066
In the following method, 50 examples are selected as the memory example set of the current task.
Extracting neural network BilSTM model from the relationship to obtain all training data
Figure BDA0002629427800000067
The vector representation of the natural language sentence S of (1), which may be obtained from the output of the last hidden layer of the model, all vector representations are clustered into 50 classes using the K-Means clustering algorithm. There will be a vector representation of the sentences of at least one instance in each category.
For the category with only one instance, taking the instance of the category as a memory; for a category of more than one instance, the arithmetic mean of the vector representations of all instance sentences under the category is calculated, and the instance with the closest euclidean distance to the mean sentence vector representation of the instance is taken as a memory, and each task will take 50 memory instances and add to the MemoryList.
Step 5) updating the network parameters
Figure BDA0002629427800000068
And loading the average accuracy and the overall accuracy of the test model on the test data of all the previous tasks in the relation extraction network BilSTM model.
The average accuracy is calculated as follows:
Figure BDA0002629427800000071
wherein, acciTo be at TaskiThe accuracy over the test case set.
The overall accuracy is calculated as follows:
Figure BDA0002629427800000072
wherein the content of the first and second substances,
Figure BDA0002629427800000073
is the accuracy on all test data.
Step 6) repeat steps 3) to 5) until no new tasks need to be processed. And measuring the average accuracy and the overall accuracy under the current task division.
Repartitioning the tasks or disordering the task order and then repeating steps 2) through 5). The performance under other task divisions or task orders is measured. The average accuracy and the overall accuracy under the division and the sequence of various tasks are counted, and the error range of the performance index is further calculated, wherein the calculation formula is as follows:
Figure BDA0002629427800000074
wherein the content of the first and second substances,
Figure BDA0002629427800000075
and the confidence coefficient when the confidence coefficient is alpha is shown, the standard deviation on the statistical samples is shown, and n is the number of the statistical samples.
In order to verify the effect of the method of the present invention, in this embodiment, the performance of the method of the present invention is compared with the traditional supervised learning method and other incremental relationship extraction methods, and the comparison result is shown in fig. 5, it can be seen that the method of the present invention extracts a data set in an incremental relationship: the average accuracy and the overall accuracy of the Life FewRel data set and the Life SimpleQuestion data set are superior to those of the traditional method and other incremental relation extraction methods, and have smaller error ranges, which shows that the method provided by the invention can effectively relieve the 'catastrophic forgetting' problem in incremental learning and has better stability.
In summary, the method provided by the invention is based on the Life FewRel dataset and the Life SimpleQuestion dataset, extracts the neural network by taking the BilSTM model as a relation, utilizes the Glove pre-training word vector, and combines the ideas of course learning and meta-learning, thereby providing the incremental relation extraction method with high accuracy and high stability. Based on the method, a relationship extraction neural network model with better effect can be trained in the application scenes of continuous updating and iteration, and an incremental relationship extraction system for continuous learning is constructed.

Claims (10)

1. An incremental relation extraction method based on course planning meta-learning is characterized by comprising the following steps:
s1: constructing a course planning meta-learning incremental relation extraction framework;
s2: calculating the similarity among the relations, and replanning and memorizing the cache training sequence according to the similarity;
s3: training the relation extraction neural network model, and updating the parameters of the relation extraction neural network model by using a meta-learning parameter updating formula in combination with the meta-learning thought;
s4: selecting a batch of memory caches from the training data in the step S3 by using a clustering method;
s5: loading the updated parameters into a relation extraction neural network model, and carrying out performance test;
s6: steps S3-S5 are repeated until no new tasks need to be processed.
2. The incremental relationship extraction method based on curriculum planning meta-learning according to claim 1, wherein the updating formula of meta-learning parameters in step S3 is specifically to update parameters of the relationship extraction neural network model by: and substituting the optimal parameters on the current task and the parameters of the previous task into a meta-learning parameter updating formula for updating.
3. The method as claimed in claim 1, wherein the selecting method of the memory cache in step S4 specifically comprises:
a1: extracting a hidden layer of the neural network model from the relationship to obtain vector representations of natural language sentences of examples in all training data;
a2: clustering all vector representations into categories with required memory quantity by using a K-Means clustering algorithm, wherein each category has vector representation of sentences of at least one example; for the category with only one instance, taking the instance of the category as a memory; for a category of more than one instance, calculating the arithmetic mean of the vector representations of all instance sentences under the category as the vector centers of all instance sentences of the category, and taking the instance with the closest Euclidean distance of the instance sentence vector from the vector center as a memory instance.
4. The method for extracting incremental relationship based on curriculum planning meta learning as claimed in claim 1, wherein the step S2 is performed by calculating the similarity between relationships as follows: the relation label character is R, and the R is divided into a list W composed of a plurality of single characters by using a word segmentation tool1,W2,…,Wn]Obtaining word vector list of single word list [ E1,E2,…,En]Taking the mean value of the word vector list as the vector representation of the relation R, and calculating the relation R pair by pairiAnd RjCosine similarity between them, denoted Simi,jAnd storing the data in a similarity matrix Sim.
5. The method as claimed in claim 1, wherein the step of learning the cached training sequence in step S2 comprises: when the Task list is not empty, the first Task in the list is taken outiAll training data of
Figure FDA0002629427790000011
Contains k relation classes
Figure FDA0002629427790000012
Will be provided with
Figure FDA0002629427790000013
Added to the set of known relationships SeenRelations.
6. The method of claim 5, wherein the training process of the neural network model for relationship extraction in step S3 is as follows:
each relationship class has NtrainBar training example, NvalidStrip verification example and NtestThe test example strips labels which do not exist in the senrelationships in the error relationship label list in the three groups of examples; will NtrainStripe training instance partitioning into size NbatchThe training batch of (1), i.e. the training data is divided into a length of
Figure FDA0002629427790000021
The Ceil function represents rounding up;
task for the first Task0Circularly taking out each training batch TrainBatch, and extracting a neural network by training relations;
task for non-first TaskiCyclically fetching each training batch TrainBatch and cyclically fetching a memory instance group M of one task from the MemoryListjEach instance in the memory instance group comprises a relation category label, the similarity of the relation category label of each instance in the memory instance group and the relation category label of the instance in the training data batch is obtained from the similarity matrix Sim, the memory instance groups are arranged from small to large by using the similarity, and the memory instance group M 'is ordered by using the similarity'jTraining the relation extraction neural network, and then training the relation extraction neural network by using a training batch TrainBatch;
task for any TaskiAfter one round of training data lists per complete cycle, use the verification example
Figure FDA0002629427790000022
Evaluating the current taskNetwork training situation and saving parameter theta of one round of network with best performanceiWhen the network does not obtain better effect in multiple continuous turns on the current task, the network parameters are updated
Figure FDA0002629427790000023
7. The method as claimed in claim 6, wherein the meta-learning parameter updating formulas in step S3 include linear parameter updating formulas, fixed parameter updating formulas and square root parameter updating formulas.
8. The method of claim 1, wherein the performance test in step S4 is performed by averaging accuracy of the test model over the test data of all previous tasks and calculating the error range of the performance index.
9. The method for extracting incremental relationship based on curriculum planning meta-learning as claimed in claim 8, wherein the average accuracy in step S4 is calculated as follows:
Figure FDA0002629427790000024
wherein, acciTo be at TaskiThe accuracy on the test case set of (1);
the overall accuracy is calculated as follows:
Figure FDA0002629427790000031
wherein the content of the first and second substances,
Figure FDA0002629427790000032
is the accuracy on all test data.
10. The method as claimed in claim 8, wherein the error range of the performance index in step S4 is calculated as follows: re-dividing tasks or disordering task sequences, measuring performance under other task divisions or task sequences, counting average accuracy and overall accuracy under various task divisions and sequences, and calculating error range of performance index according to the following formula:
Figure FDA0002629427790000033
wherein the content of the first and second substances,
Figure FDA0002629427790000034
and the confidence coefficient when the confidence coefficient is alpha is shown, the standard deviation on the statistical samples is shown, and n is the number of the statistical samples.
CN202010806791.2A 2020-08-12 2020-08-12 Incremental relation extraction method based on course planning meta-learning Pending CN112084330A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010806791.2A CN112084330A (en) 2020-08-12 2020-08-12 Incremental relation extraction method based on course planning meta-learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010806791.2A CN112084330A (en) 2020-08-12 2020-08-12 Incremental relation extraction method based on course planning meta-learning

Publications (1)

Publication Number Publication Date
CN112084330A true CN112084330A (en) 2020-12-15

Family

ID=73727877

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010806791.2A Pending CN112084330A (en) 2020-08-12 2020-08-12 Incremental relation extraction method based on course planning meta-learning

Country Status (1)

Country Link
CN (1) CN112084330A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112667732A (en) * 2021-01-25 2021-04-16 东南大学 Continuous relation extraction method based on task perception meta-learning
CN112698933A (en) * 2021-03-24 2021-04-23 中国科学院自动化研究所 Method and device for continuous learning in multitask data stream
CN112990318A (en) * 2021-03-18 2021-06-18 中国科学院深圳先进技术研究院 Continuous learning method, device, terminal and storage medium
CN113361645A (en) * 2021-07-03 2021-09-07 上海理想信息产业(集团)有限公司 Target detection model construction method and system based on meta-learning and knowledge memory

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160179945A1 (en) * 2014-12-19 2016-06-23 Universidad Nacional De Educación A Distancia (Uned) System and method for the indexing and retrieval of semantically annotated data using an ontology-based information retrieval model
CN105956197A (en) * 2016-06-15 2016-09-21 杭州量知数据科技有限公司 Social media graph representation model-based social risk event extraction method
US20180253529A1 (en) * 2017-03-02 2018-09-06 Drexel University Multi-temporal Information Object Incremental Learning Software System
CN108959305A (en) * 2017-05-22 2018-12-07 北京国信宏数科技有限公司 A kind of event extraction method and system based on internet big data

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160179945A1 (en) * 2014-12-19 2016-06-23 Universidad Nacional De Educación A Distancia (Uned) System and method for the indexing and retrieval of semantically annotated data using an ontology-based information retrieval model
CN105956197A (en) * 2016-06-15 2016-09-21 杭州量知数据科技有限公司 Social media graph representation model-based social risk event extraction method
US20180253529A1 (en) * 2017-03-02 2018-09-06 Drexel University Multi-temporal Information Object Incremental Learning Software System
CN108959305A (en) * 2017-05-22 2018-12-07 北京国信宏数科技有限公司 A kind of event extraction method and system based on internet big data

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
JATHUSHAN RAJASEGARAN 等: "iTAML: An Incremental Task-Agnostic Meta-learning Approach", 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 5 August 2020 (2020-08-05), pages 13585 - 13594 *
叶艳: "自适应元学习算法的研究", 中国优秀硕士学位论文全文数据库, no. 06, 15 June 2013 (2013-06-15), pages 1 - 68 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112667732A (en) * 2021-01-25 2021-04-16 东南大学 Continuous relation extraction method based on task perception meta-learning
CN112990318A (en) * 2021-03-18 2021-06-18 中国科学院深圳先进技术研究院 Continuous learning method, device, terminal and storage medium
WO2022193753A1 (en) * 2021-03-18 2022-09-22 中国科学院深圳先进技术研究院 Continuous learning method and apparatus, and terminal and storage medium
CN112990318B (en) * 2021-03-18 2024-05-07 中国科学院深圳先进技术研究院 Continuous learning method, device, terminal and storage medium
CN112698933A (en) * 2021-03-24 2021-04-23 中国科学院自动化研究所 Method and device for continuous learning in multitask data stream
CN113361645A (en) * 2021-07-03 2021-09-07 上海理想信息产业(集团)有限公司 Target detection model construction method and system based on meta-learning and knowledge memory
CN113361645B (en) * 2021-07-03 2024-01-23 上海理想信息产业(集团)有限公司 Target detection model construction method and system based on meta learning and knowledge memory

Similar Documents

Publication Publication Date Title
CN112084330A (en) Incremental relation extraction method based on course planning meta-learning
CN110084271B (en) Method and device for identifying picture category
CN112115721B (en) Named entity recognition method and device
CN108985617B (en) Product production flow scheduling method and system based on intelligent manufacturing
CN111860981B (en) Enterprise national industry category prediction method and system based on LSTM deep learning
CN110928981A (en) Method, system and storage medium for establishing and perfecting iteration of text label system
CN113177587B (en) Generalized zero sample target classification method based on active learning and variational self-encoder
CN112016313A (en) Spoken language element identification method and device and alarm situation analysis system
CN115391561A (en) Method and device for processing graph network data set, electronic equipment, program and medium
CN116304020A (en) Industrial text entity extraction method based on semantic source analysis and span characteristics
CN113177644A (en) Automatic modeling system based on word embedding and depth time sequence model
CN111783688B (en) Remote sensing image scene classification method based on convolutional neural network
Ding et al. Dirichlet process mixture models with shrinkage prior
CN115526300B (en) Sequence rearrangement method based on cyclic neural network
CN114969738B (en) Interface abnormal behavior monitoring method, system, device and storage medium
CN116910526A (en) Model training method, device, communication equipment and readable storage medium
CN114048320B (en) Multi-label international disease classification training method based on course learning
CN113051886B (en) Test question duplicate checking method, device, storage medium and equipment
CN112667732A (en) Continuous relation extraction method based on task perception meta-learning
CN113821571A (en) Food safety relation extraction method based on BERT and improved PCNN
CN116431757B (en) Text relation extraction method based on active learning, electronic equipment and storage medium
US11816427B1 (en) Automated data classification error correction through spatial analysis using machine learning
CN117910875B (en) System for evaluating stress resistance of elymus resource
CN115081439B (en) Multi-feature self-adaptive enhancement-based chemical classification method and system
CN117312138A (en) Software defect detection method, device, computer equipment, storage medium and product

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination