CN111191453A - Named entity recognition method based on confrontation training - Google Patents

Named entity recognition method based on confrontation training Download PDF

Info

Publication number
CN111191453A
CN111191453A CN201911358738.4A CN201911358738A CN111191453A CN 111191453 A CN111191453 A CN 111191453A CN 201911358738 A CN201911358738 A CN 201911358738A CN 111191453 A CN111191453 A CN 111191453A
Authority
CN
China
Prior art keywords
training
word
model
judicial
word vectors
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911358738.4A
Other languages
Chinese (zh)
Inventor
袁超逸
刘忠麟
王立才
张起闻
罗琪彬
郝韫宏
李孟书
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CETC 15 Research Institute
Original Assignee
CETC 15 Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CETC 15 Research Institute filed Critical CETC 15 Research Institute
Priority to CN201911358738.4A priority Critical patent/CN111191453A/en
Publication of CN111191453A publication Critical patent/CN111191453A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/355Class or cluster creation or modification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • Computational Linguistics (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a named entity recognition method based on confrontation training, which obtains the correlation characteristics between characters in the judicial field through RoBERTA model training and Bi-LSTM training respectively; then splicing the two relevance characteristics together, and predicting the training sample by using a conditional random field model to obtain a predicted result; the method can introduce the combination of external word vectors and word vectors with different dimensions and judicial domain text word mixed vectors with different dimensions, and can resist disturbance aiming at the mixed word vectors in the judicial domain text, thereby increasing the accuracy of model identification.

Description

Named entity recognition method based on confrontation training
Technical Field
The invention belongs to the technical field of named entity recognition, and particularly relates to a named entity recognition method based on countermeasure training.
Background
Named entity recognition has been widely applied in various fields, various fields are optimized in different degrees aiming at named entity recognition, a large number of personnel are required to be consumed in the traditional named entity recognition to perform feature extraction aiming at specific fields, a probabilistic graph model is used for named entity recognition, with the rise of deep learning in recent years, various fields are greatly explored for named entity recognition by using a deep learning method, at present, a large number of exploration and practice are performed in the financial, medical and legal fields, a large amount of labor cost is reduced, the accuracy is improved, how to use the information is particularly critical, entities with specific meanings in certain specific fields, such as in judicial texts (suspects, reports, original reports and the like), can be extracted for information later through the recognition of the entities, The question-answering system, the syntactic analysis, the knowledge reasoning, the construction of knowledge maps and other important tasks lay important foundations.
Currently, the main methods of named entity recognition in the judicial field fall into three main categories:
the first type is based on a probability map model, and the method mainly uses a Conditional Random Field (CRF) model which is a conditional probability distribution model of another group of output sequences under the condition of giving a group of input sequences, and manually extracts corresponding characteristics and sets corresponding rules by inputting labeled specific field data so as to identify unlabeled texts.
The second kind of deep learning-based method mainly uses a bidirectional long-time memory network (Bi-LSTM) model, utilizes word vector embedded information, and greatly reduces manual work by inputting tagged specific field data into the Bi-LSTM, and can obtain higher accuracy.
The third kind is based on deep learning and is combined with traditional method, the method utilizes Word vector training method (Word2Vec) or (GloVe) technology to give text of specific field, Word list of specific field, in specific fieldIn the field text, we build a language model through the text, and the language model is built through P (w)1,w2,…wn)=P(w1)P(w2|w1)P(wn|w1,…,wn-1) Converting the joint probability into conditional continuous multiplication, greatly reducing parameters by using Markov hypothesis, inputting a segment of word vector in a probability model corresponding to each word in a word list, outputting the joint probability of a text, learning the weight of the word vector, and constructing a simple neural network f (w)t-n+1…,wt)f(wt-n+1…,wt) To fit the conditional probability P (w)t|w1,…,wt-1) Inputting word vectors into a linear Embedding layer (Embedding) layer in the model, acquiring the word vectors of the text in the specific field by setting different sliding windows through the whole text in the specific field by using a trainable parameter matrix C, and acquiring corresponding word vectors (word vectors), wherein two methods are respectively used for training a Skip-word model (Skip-gram) or a continuous word bag model (CBOW), after the corresponding word vectors are acquired, the word vectors in the Bi-LSTM layer are input through a Bi-LSTM layer and pass through the hidden state of each time point, so that the representation of a context can be acquired, and the final characteristics utilize surrounding information through a CRF layer so as to effectively acquire corresponding labels, wherein the model is shown in figure 1.
The existing named entity model in a specific field utilizes a mode of combining Bi-LSTM and CRF models, but the capability of extracting features of the models is not strong enough, and modeling in Bi-LSTM is only simple to perform modeling from left to right or from right to left, and hidden states are spliced together, but the disadvantage of this is that only the information of the upper part or the lower part can be utilized, and the information of the upper part and the lower part cannot be utilized simultaneously. In addition, the number and quantity of texts in a specific field are limited, and a large amount of data is not available for improving the model performance.
With the appearance of the BERT model, applications have been gradually performed in various fields, but no corresponding applications have been obtained in specific fields, and words brought by BERT and a subsequent model RoBERTa are mutually independent, so that the disadvantages of loss of model performance and the like are brought during fine adjustment, the scale of data is large, and the accuracy of the model cannot be basically improved.
Disclosure of Invention
In view of the above, the invention provides a named entity recognition method based on countermeasure training, which can introduce external word vectors and word vectors of different dimensions to be combined with text word and word mixed vectors of different dimensions in the judicial field, and perform countermeasure disturbance on the mixed word vectors in the judicial field text, so as to increase the accuracy of model recognition.
The technical scheme for realizing the invention is as follows:
a named entity recognition method based on confrontation training comprises the following steps:
firstly, segmenting a referee document in the judicial field into single characters serving as training samples, and training through a RoBERTA model to obtain relevance characteristics among the characters in the judicial field;
cutting the judge text in the judicial field into single characters and phrases, converting the phrases into Word vectors by using a Word2Vec method, and converting the single characters into Word vectors based on characters by using a Fastext method; introducing Word vectors obtained by using a Word2Vec method outside the judicial field, and introducing Word vectors based on characters obtained by using a Fastext method outside the judicial field; mixing all the word vectors;
step two, disturbing the mixed word vector matrix, finding the disturbance in the worst case through the maximum value of a loss function, and obtaining the optimal robust parameter of the model by utilizing the minimization of the external experience risk so as to obtain the word vector after disturbance-resistant optimization;
step three, inputting the word vector obtained in the step two into the Bi-LSTM by using a sliding window with the length of a, and obtaining the relevance characteristics between the words in the judicial field through the training of the Bi-LSTM;
and step four, splicing the two relevance characteristics obtained in the step one and the step three together, and then predicting the training sample by using a conditional random field model to obtain a predicted result.
Further, 1000< a < 2000.
Has the advantages that:
1. the method introduces the combination of external word vectors and word vectors with different dimensions and text word mixed vectors in the judicial fields with different dimensions, and enriches training samples for the recognition of named entities in the judicial fields.
2. The method comprises the steps of extracting features of a text in the judicial field by using RoBERTA, fusing the extracted features with word vectors with different dimensions, and combining the fused features with Bi-LSTM features to obtain corresponding features, and obtaining a result by using CRF.
3. The method aims at the mixed word vector in the text in the judicial field to resist disturbance, and increases the generalization capability and robustness of the model.
Drawings
FIG. 1 is a diagram of the Bi-LSTM architecture.
Fig. 2 is a schematic diagram of RoBERTa model architecture.
FIG. 3 is a diagram of the word vector model according to the present invention.
FIG. 4 is a diagram of the named entity recognition model architecture of the present invention.
Detailed Description
The invention is described in detail below by way of example with reference to the accompanying drawings.
The invention provides a named entity recognition method based on countermeasure training, which comprises the following specific processes as shown in figure 4:
firstly, the invention introduces a RoBERTA model in the judicial field, firstly, corresponding word segmentation is carried out on each text in the judicial field, the words are input into the RoBERTA in the form of characters, different weights are distributed to different words through a self-attention mechanism (self-attention), namely, an input matrix is assumed to be X, the maximum word embedding vector is 512, and different weight matrixes W are adoptedq,Wk,WvFinally, obtaining a self-attention matrix Z through softmax, obtaining a plurality of expression subspaces of an attention layer through a multi-head mechanism, finally splicing different matrices Z, and extracting corresponding features C through dynamic masks of partial words, as shown in fig. 2:
in the judicial field, the text size of the corresponding judicial field is not so large, and only limited data existThe invention introduces Fastext based on words and Word2Vec based on words, constructs the Embedding layer, constructs the text of the judicial field into N-1 one-hot Word vectors, passes all the one-hot vectors through an NxV matrix, N is the dimension set by the user, V is the size of the dictionary, obtains the vector addition, the averaging and the multiplication by the output weight matrix to obtain the corresponding probability distribution, wherein the NxV matrix is a Word and Word vector matrix W1And W2Different word vectors and word vector dimensions are specified, the word vector matrix based on the words makes up the characteristic of less professional vocabularies in the judicial field, and the word vector matrix based on the words is the words in the judicial field, so that more accurate priori knowledge can be provided, and the external larger universal word vector matrix W is introduced3And word vector matrix W4Are spliced together [ W ]1,W2,W3,W4]The feature vector with rich information is obtained, the characteristics that the number of texts in the judicial field is small and a better effect cannot be obtained are overcome, and the model is shown as the graph 3:
step two, using the countermeasure training of the mixed word vector matrix to disturb the word vector, and assuming the mixed word vector matrix [ v [ ]1,v2,…vT]For x, perturb the mixed word vector matrix, γadv=∈·g/||g||2,
Figure BDA0002336630710000051
And optimize the function
Figure BDA0002336630710000052
Figure BDA0002336630710000053
Finding disturbance by an internal max function, finding an optimal robust parameter by an external min function, solving the non-convex constraint optimization problem of the internal max by a Fast Gradient Method in a formula with L as a loss function, and finally obtaining a corresponding result, wherein gamma isadvFor the value of the perturbation, e is the coefficient of the perturbation, g is the gradient over x,
Figure BDA0002336630710000054
is the range of the sample, y is the predicted value, θ is the parameter of the classifier, E is the empirical risk function, S is the range of the disturbance, fθA function mapped for a language model coder.
Step three, the Bi-LSTM model can increase the feature number of the context hidden vector through different windows, the spliced word vector is input into the Bi-LSTM, and the word vector passes through a forgetting gate ft=σ(Wf*[ht-1,xt]+bf) To judge whether to forget the old information, and then to input the old information through the input gate it=σ(Wi*[ht-1,xt]+bi) Updating the values using the sigmoid function and constructing new candidate values
Figure BDA0002336630710000055
Then through the refresh door
Figure BDA0002336630710000056
To decide whether to update the state, finally we need the output gate ot=σ(Wo*[ht-1,xt]+bo),ht=ot*tanh(Ct) To obtain a corresponding probability distribution, x in the formulatFor a matrix of vectors of words or words input in sequence, ftValue obtained through forget gate, itFor the value obtained through the input gate,
Figure BDA0002336630710000061
to pass the updated candidate after the gate, CtTo pass the updated state value after the gate, otIs the value of output, htFor the current hidden state, the sigmoid function is a function that maps variables between 0 and 1, σ is the sigmoid function, tanh function compresses values between-1 and 1, ht-1Is the hidden state at the previous moment, b is the bias term, W is the weight matrix, Ct-1In order to be in the last memory state,
Figure BDA0002336630710000062
for the current memory state, a hidden state h is obtained by constructing a language model from left to right and from right to leftt1And ht2The hidden state of the spliced two is Ht
Fourthly, splicing the characteristic C extracted by the RoBERTA model and the hidden state obtained by the Bi-LSTM to obtain a characteristic matrix [ C, Ht]C is the feature extracted between words, HtThe method solves the problem of independent hypothesis test brought by the RoBERTA model for the hidden state of post-splicing by the Bi-LSTM model, and supplements the characteristic loss caused by the word number limitation of the RoBERTA model. The CRF may obtain the named entity recognition result by using the viterbi algorithm in consideration of the constraint relationship between the labels.
In summary, the above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (2)

1. A named entity recognition method based on countermeasure training is characterized by comprising the following steps:
firstly, segmenting a referee document in the judicial field into single characters serving as training samples, and training through a RoBERTA model to obtain relevance characteristics among the characters in the judicial field;
cutting the judge text in the judicial field into single characters and phrases, converting the phrases into Word vectors by using a Word2Vec method, and converting the single characters into Word vectors based on characters by using a Fastext method; introducing Word vectors obtained by using a Word2Vec method outside the judicial field, and introducing Word vectors based on characters obtained by using a Fastext method outside the judicial field; mixing all the word vectors;
step two, disturbing the mixed word vector matrix, finding the disturbance in the worst case through the maximum value of a loss function, and obtaining the optimal robust parameter of the model by utilizing the minimization of the external experience risk so as to obtain the word vector after disturbance-resistant optimization;
step three, inputting the word vector obtained in the step two into the Bi-LSTM by using a sliding window with the length of a, and obtaining the relevance characteristics between the words in the judicial field through the training of the Bi-LSTM;
and step four, splicing the two relevance characteristics obtained in the step one and the step three together, and then predicting the training sample by using a conditional random field model to obtain a predicted result.
2. The method of claim 1, wherein 1000< a < 2000.
CN201911358738.4A 2019-12-25 2019-12-25 Named entity recognition method based on confrontation training Pending CN111191453A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911358738.4A CN111191453A (en) 2019-12-25 2019-12-25 Named entity recognition method based on confrontation training

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911358738.4A CN111191453A (en) 2019-12-25 2019-12-25 Named entity recognition method based on confrontation training

Publications (1)

Publication Number Publication Date
CN111191453A true CN111191453A (en) 2020-05-22

Family

ID=70709379

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911358738.4A Pending CN111191453A (en) 2019-12-25 2019-12-25 Named entity recognition method based on confrontation training

Country Status (1)

Country Link
CN (1) CN111191453A (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111680145A (en) * 2020-06-10 2020-09-18 北京百度网讯科技有限公司 Knowledge representation learning method, device, equipment and storage medium
CN111737952A (en) * 2020-06-24 2020-10-02 深圳前海微众银行股份有限公司 Training method and device for sequence labeling model
CN112115721A (en) * 2020-09-28 2020-12-22 青岛海信网络科技股份有限公司 Named entity identification method and device
CN112765319A (en) * 2021-01-20 2021-05-07 中国电子信息产业集团有限公司第六研究所 Text processing method and device, electronic equipment and storage medium
CN112784831A (en) * 2021-02-02 2021-05-11 电子科技大学 Character recognition method for enhancing attention mechanism by fusing multilayer features
CN112925918A (en) * 2021-02-26 2021-06-08 华南理工大学 Question-answer matching system based on disease field knowledge graph
CN112949291A (en) * 2021-03-02 2021-06-11 赛飞特工程技术集团有限公司 Report error correction system and method
CN113380418A (en) * 2021-06-22 2021-09-10 浙江工业大学 System for analyzing and identifying depression through dialog text
CN113987192A (en) * 2021-12-28 2022-01-28 中国电子科技网络信息安全有限公司 Hot topic detection method based on RoBERTA-WWM and HDBSCAN algorithm
WO2022078102A1 (en) * 2020-10-14 2022-04-21 腾讯科技(深圳)有限公司 Entity identification method and apparatus, device and storage medium
CN114399396A (en) * 2022-01-19 2022-04-26 中国平安人寿保险股份有限公司 Insurance product recommendation method and device, computer equipment and storage medium
CN114444506A (en) * 2022-01-11 2022-05-06 四川大学 Method for extracting relation triple fusing entity types
CN116702787A (en) * 2023-08-07 2023-09-05 四川隧唐科技股份有限公司 Long text entity identification method, device, computer equipment and medium
CN117057350A (en) * 2023-08-07 2023-11-14 内蒙古大学 Chinese electronic medical record named entity recognition method and system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108229582A (en) * 2018-02-01 2018-06-29 浙江大学 Entity recognition dual training method is named in a kind of multitask towards medical domain
CN110083831A (en) * 2019-04-16 2019-08-02 武汉大学 A kind of Chinese name entity recognition method based on BERT-BiGRU-CRF
CN110532377A (en) * 2019-05-13 2019-12-03 南京大学 A kind of semi-supervised file classification method based on dual training and confrontation learning network

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108229582A (en) * 2018-02-01 2018-06-29 浙江大学 Entity recognition dual training method is named in a kind of multitask towards medical domain
CN110083831A (en) * 2019-04-16 2019-08-02 武汉大学 A kind of Chinese name entity recognition method based on BERT-BiGRU-CRF
CN110532377A (en) * 2019-05-13 2019-12-03 南京大学 A kind of semi-supervised file classification method based on dual training and confrontation learning network

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
GEORGY KONOPLICH等: "Named Entity Recognition in Russian with Word Representation Learned by a Bidirectional Language Model", ARTIFICIAL INTELLIGENCE AND NATURAL LANGUAGE *
ZHENJIN DAI等: "Named Entity Recognition Using BERT BiLSTM CRF for Chinese Electronic Health Records", 12TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS *
冯建周等: "关于命名实体识别的生成式对抗网络的研究", 小型微型计算机系统 *
王文广等: "基于混合深度神经网络模型的司法文书智能化处理", 清华大学学报(自然科学版) *
郭宝震等: "采用词向量注意力机制的双路卷积神经网络句子分类模型", 浙江大学学报(工学版) *

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111680145B (en) * 2020-06-10 2023-08-15 北京百度网讯科技有限公司 Knowledge representation learning method, apparatus, device and storage medium
CN111680145A (en) * 2020-06-10 2020-09-18 北京百度网讯科技有限公司 Knowledge representation learning method, device, equipment and storage medium
CN111737952A (en) * 2020-06-24 2020-10-02 深圳前海微众银行股份有限公司 Training method and device for sequence labeling model
CN112115721A (en) * 2020-09-28 2020-12-22 青岛海信网络科技股份有限公司 Named entity identification method and device
CN112115721B (en) * 2020-09-28 2024-05-17 青岛海信网络科技股份有限公司 Named entity recognition method and device
WO2022078102A1 (en) * 2020-10-14 2022-04-21 腾讯科技(深圳)有限公司 Entity identification method and apparatus, device and storage medium
CN112765319A (en) * 2021-01-20 2021-05-07 中国电子信息产业集团有限公司第六研究所 Text processing method and device, electronic equipment and storage medium
CN112765319B (en) * 2021-01-20 2021-09-03 中国电子信息产业集团有限公司第六研究所 Text processing method and device, electronic equipment and storage medium
CN112784831A (en) * 2021-02-02 2021-05-11 电子科技大学 Character recognition method for enhancing attention mechanism by fusing multilayer features
CN112784831B (en) * 2021-02-02 2022-06-28 电子科技大学 Character recognition method for enhancing attention mechanism by fusing multilayer features
CN112925918B (en) * 2021-02-26 2023-03-24 华南理工大学 Question-answer matching system based on disease field knowledge graph
CN112925918A (en) * 2021-02-26 2021-06-08 华南理工大学 Question-answer matching system based on disease field knowledge graph
CN112949291A (en) * 2021-03-02 2021-06-11 赛飞特工程技术集团有限公司 Report error correction system and method
CN113380418A (en) * 2021-06-22 2021-09-10 浙江工业大学 System for analyzing and identifying depression through dialog text
CN113987192B (en) * 2021-12-28 2022-04-01 中国电子科技网络信息安全有限公司 Hot topic detection method based on RoBERTA-WWM and HDBSCAN algorithm
CN113987192A (en) * 2021-12-28 2022-01-28 中国电子科技网络信息安全有限公司 Hot topic detection method based on RoBERTA-WWM and HDBSCAN algorithm
CN114444506A (en) * 2022-01-11 2022-05-06 四川大学 Method for extracting relation triple fusing entity types
CN114444506B (en) * 2022-01-11 2023-05-02 四川大学 Relation triplet extraction method for fusing entity types
CN114399396A (en) * 2022-01-19 2022-04-26 中国平安人寿保险股份有限公司 Insurance product recommendation method and device, computer equipment and storage medium
CN116702787A (en) * 2023-08-07 2023-09-05 四川隧唐科技股份有限公司 Long text entity identification method, device, computer equipment and medium
CN117057350A (en) * 2023-08-07 2023-11-14 内蒙古大学 Chinese electronic medical record named entity recognition method and system
CN117057350B (en) * 2023-08-07 2024-05-10 内蒙古大学 Chinese electronic medical record named entity recognition method and system

Similar Documents

Publication Publication Date Title
CN111191453A (en) Named entity recognition method based on confrontation training
CN110929030B (en) Text abstract and emotion classification combined training method
CN108733792B (en) Entity relation extraction method
CN110532557B (en) Unsupervised text similarity calculation method
CN108563653B (en) Method and system for constructing knowledge acquisition model in knowledge graph
CN108628823B (en) Named entity recognition method combining attention mechanism and multi-task collaborative training
CN110334354B (en) Chinese relation extraction method
CN111160467B (en) Image description method based on conditional random field and internal semantic attention
CN106407333B (en) Spoken language query identification method and device based on artificial intelligence
CN108062388A (en) Interactive reply generation method and device
CN110909736B (en) Image description method based on long-term and short-term memory model and target detection algorithm
CN111666758B (en) Chinese word segmentation method, training device and computer readable storage medium
CN109800437A (en) A kind of name entity recognition method based on Fusion Features
CN109858041A (en) A kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries
CN109919175B (en) Entity multi-classification method combined with attribute information
CN112699685B (en) Named entity recognition method based on label-guided word fusion
CN114492441A (en) BilSTM-BiDAF named entity identification method based on machine reading understanding
CN110162789A (en) A kind of vocabulary sign method and device based on the Chinese phonetic alphabet
CN114648031B (en) Text aspect emotion recognition method based on bidirectional LSTM and multi-head attention mechanism
CN115062104A (en) Knowledge prompt-fused legal text small sample named entity identification method
CN114239612A (en) Multi-modal neural machine translation method, computer equipment and storage medium
CN115114409A (en) Civil aviation unsafe event combined extraction method based on soft parameter sharing
CN110309515B (en) Entity identification method and device
CN113642630B (en) Image description method and system based on double-path feature encoder
CN114239584A (en) Named entity identification method based on self-supervision learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200522