CN111191453A - Named entity recognition method based on confrontation training - Google Patents
Named entity recognition method based on confrontation training Download PDFInfo
- Publication number
- CN111191453A CN111191453A CN201911358738.4A CN201911358738A CN111191453A CN 111191453 A CN111191453 A CN 111191453A CN 201911358738 A CN201911358738 A CN 201911358738A CN 111191453 A CN111191453 A CN 111191453A
- Authority
- CN
- China
- Prior art keywords
- training
- word
- model
- judicial
- word vectors
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 32
- 238000012549 training Methods 0.000 title claims abstract description 23
- 239000013598 vector Substances 0.000 claims abstract description 60
- 239000011159 matrix material Substances 0.000 claims description 19
- 238000005457 optimization Methods 0.000 claims description 3
- 230000006870 function Effects 0.000 description 12
- 238000013135 deep learning Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 2
- 238000012935 Averaging Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
- G06F16/355—Class or cluster creation or modification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
- G06F16/367—Ontology
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Computational Linguistics (AREA)
- Machine Translation (AREA)
Abstract
The invention discloses a named entity recognition method based on confrontation training, which obtains the correlation characteristics between characters in the judicial field through RoBERTA model training and Bi-LSTM training respectively; then splicing the two relevance characteristics together, and predicting the training sample by using a conditional random field model to obtain a predicted result; the method can introduce the combination of external word vectors and word vectors with different dimensions and judicial domain text word mixed vectors with different dimensions, and can resist disturbance aiming at the mixed word vectors in the judicial domain text, thereby increasing the accuracy of model identification.
Description
Technical Field
The invention belongs to the technical field of named entity recognition, and particularly relates to a named entity recognition method based on countermeasure training.
Background
Named entity recognition has been widely applied in various fields, various fields are optimized in different degrees aiming at named entity recognition, a large number of personnel are required to be consumed in the traditional named entity recognition to perform feature extraction aiming at specific fields, a probabilistic graph model is used for named entity recognition, with the rise of deep learning in recent years, various fields are greatly explored for named entity recognition by using a deep learning method, at present, a large number of exploration and practice are performed in the financial, medical and legal fields, a large amount of labor cost is reduced, the accuracy is improved, how to use the information is particularly critical, entities with specific meanings in certain specific fields, such as in judicial texts (suspects, reports, original reports and the like), can be extracted for information later through the recognition of the entities, The question-answering system, the syntactic analysis, the knowledge reasoning, the construction of knowledge maps and other important tasks lay important foundations.
Currently, the main methods of named entity recognition in the judicial field fall into three main categories:
the first type is based on a probability map model, and the method mainly uses a Conditional Random Field (CRF) model which is a conditional probability distribution model of another group of output sequences under the condition of giving a group of input sequences, and manually extracts corresponding characteristics and sets corresponding rules by inputting labeled specific field data so as to identify unlabeled texts.
The second kind of deep learning-based method mainly uses a bidirectional long-time memory network (Bi-LSTM) model, utilizes word vector embedded information, and greatly reduces manual work by inputting tagged specific field data into the Bi-LSTM, and can obtain higher accuracy.
The third kind is based on deep learning and is combined with traditional method, the method utilizes Word vector training method (Word2Vec) or (GloVe) technology to give text of specific field, Word list of specific field, in specific fieldIn the field text, we build a language model through the text, and the language model is built through P (w)1,w2,…wn)=P(w1)P(w2|w1)P(wn|w1,…,wn-1) Converting the joint probability into conditional continuous multiplication, greatly reducing parameters by using Markov hypothesis, inputting a segment of word vector in a probability model corresponding to each word in a word list, outputting the joint probability of a text, learning the weight of the word vector, and constructing a simple neural network f (w)t-n+1…,wt)f(wt-n+1…,wt) To fit the conditional probability P (w)t|w1,…,wt-1) Inputting word vectors into a linear Embedding layer (Embedding) layer in the model, acquiring the word vectors of the text in the specific field by setting different sliding windows through the whole text in the specific field by using a trainable parameter matrix C, and acquiring corresponding word vectors (word vectors), wherein two methods are respectively used for training a Skip-word model (Skip-gram) or a continuous word bag model (CBOW), after the corresponding word vectors are acquired, the word vectors in the Bi-LSTM layer are input through a Bi-LSTM layer and pass through the hidden state of each time point, so that the representation of a context can be acquired, and the final characteristics utilize surrounding information through a CRF layer so as to effectively acquire corresponding labels, wherein the model is shown in figure 1.
The existing named entity model in a specific field utilizes a mode of combining Bi-LSTM and CRF models, but the capability of extracting features of the models is not strong enough, and modeling in Bi-LSTM is only simple to perform modeling from left to right or from right to left, and hidden states are spliced together, but the disadvantage of this is that only the information of the upper part or the lower part can be utilized, and the information of the upper part and the lower part cannot be utilized simultaneously. In addition, the number and quantity of texts in a specific field are limited, and a large amount of data is not available for improving the model performance.
With the appearance of the BERT model, applications have been gradually performed in various fields, but no corresponding applications have been obtained in specific fields, and words brought by BERT and a subsequent model RoBERTa are mutually independent, so that the disadvantages of loss of model performance and the like are brought during fine adjustment, the scale of data is large, and the accuracy of the model cannot be basically improved.
Disclosure of Invention
In view of the above, the invention provides a named entity recognition method based on countermeasure training, which can introduce external word vectors and word vectors of different dimensions to be combined with text word and word mixed vectors of different dimensions in the judicial field, and perform countermeasure disturbance on the mixed word vectors in the judicial field text, so as to increase the accuracy of model recognition.
The technical scheme for realizing the invention is as follows:
a named entity recognition method based on confrontation training comprises the following steps:
firstly, segmenting a referee document in the judicial field into single characters serving as training samples, and training through a RoBERTA model to obtain relevance characteristics among the characters in the judicial field;
cutting the judge text in the judicial field into single characters and phrases, converting the phrases into Word vectors by using a Word2Vec method, and converting the single characters into Word vectors based on characters by using a Fastext method; introducing Word vectors obtained by using a Word2Vec method outside the judicial field, and introducing Word vectors based on characters obtained by using a Fastext method outside the judicial field; mixing all the word vectors;
step two, disturbing the mixed word vector matrix, finding the disturbance in the worst case through the maximum value of a loss function, and obtaining the optimal robust parameter of the model by utilizing the minimization of the external experience risk so as to obtain the word vector after disturbance-resistant optimization;
step three, inputting the word vector obtained in the step two into the Bi-LSTM by using a sliding window with the length of a, and obtaining the relevance characteristics between the words in the judicial field through the training of the Bi-LSTM;
and step four, splicing the two relevance characteristics obtained in the step one and the step three together, and then predicting the training sample by using a conditional random field model to obtain a predicted result.
Further, 1000< a < 2000.
Has the advantages that:
1. the method introduces the combination of external word vectors and word vectors with different dimensions and text word mixed vectors in the judicial fields with different dimensions, and enriches training samples for the recognition of named entities in the judicial fields.
2. The method comprises the steps of extracting features of a text in the judicial field by using RoBERTA, fusing the extracted features with word vectors with different dimensions, and combining the fused features with Bi-LSTM features to obtain corresponding features, and obtaining a result by using CRF.
3. The method aims at the mixed word vector in the text in the judicial field to resist disturbance, and increases the generalization capability and robustness of the model.
Drawings
FIG. 1 is a diagram of the Bi-LSTM architecture.
Fig. 2 is a schematic diagram of RoBERTa model architecture.
FIG. 3 is a diagram of the word vector model according to the present invention.
FIG. 4 is a diagram of the named entity recognition model architecture of the present invention.
Detailed Description
The invention is described in detail below by way of example with reference to the accompanying drawings.
The invention provides a named entity recognition method based on countermeasure training, which comprises the following specific processes as shown in figure 4:
firstly, the invention introduces a RoBERTA model in the judicial field, firstly, corresponding word segmentation is carried out on each text in the judicial field, the words are input into the RoBERTA in the form of characters, different weights are distributed to different words through a self-attention mechanism (self-attention), namely, an input matrix is assumed to be X, the maximum word embedding vector is 512, and different weight matrixes W are adoptedq,Wk,WvFinally, obtaining a self-attention matrix Z through softmax, obtaining a plurality of expression subspaces of an attention layer through a multi-head mechanism, finally splicing different matrices Z, and extracting corresponding features C through dynamic masks of partial words, as shown in fig. 2:
in the judicial field, the text size of the corresponding judicial field is not so large, and only limited data existThe invention introduces Fastext based on words and Word2Vec based on words, constructs the Embedding layer, constructs the text of the judicial field into N-1 one-hot Word vectors, passes all the one-hot vectors through an NxV matrix, N is the dimension set by the user, V is the size of the dictionary, obtains the vector addition, the averaging and the multiplication by the output weight matrix to obtain the corresponding probability distribution, wherein the NxV matrix is a Word and Word vector matrix W1And W2Different word vectors and word vector dimensions are specified, the word vector matrix based on the words makes up the characteristic of less professional vocabularies in the judicial field, and the word vector matrix based on the words is the words in the judicial field, so that more accurate priori knowledge can be provided, and the external larger universal word vector matrix W is introduced3And word vector matrix W4Are spliced together [ W ]1,W2,W3,W4]The feature vector with rich information is obtained, the characteristics that the number of texts in the judicial field is small and a better effect cannot be obtained are overcome, and the model is shown as the graph 3:
step two, using the countermeasure training of the mixed word vector matrix to disturb the word vector, and assuming the mixed word vector matrix [ v [ ]1,v2,…vT]For x, perturb the mixed word vector matrix, γadv=∈·g/||g||2,And optimize the function Finding disturbance by an internal max function, finding an optimal robust parameter by an external min function, solving the non-convex constraint optimization problem of the internal max by a Fast Gradient Method in a formula with L as a loss function, and finally obtaining a corresponding result, wherein gamma isadvFor the value of the perturbation, e is the coefficient of the perturbation, g is the gradient over x,is the range of the sample, y is the predicted value, θ is the parameter of the classifier, E is the empirical risk function, S is the range of the disturbance, fθA function mapped for a language model coder.
Step three, the Bi-LSTM model can increase the feature number of the context hidden vector through different windows, the spliced word vector is input into the Bi-LSTM, and the word vector passes through a forgetting gate ft=σ(Wf*[ht-1,xt]+bf) To judge whether to forget the old information, and then to input the old information through the input gate it=σ(Wi*[ht-1,xt]+bi) Updating the values using the sigmoid function and constructing new candidate valuesThen through the refresh doorTo decide whether to update the state, finally we need the output gate ot=σ(Wo*[ht-1,xt]+bo),ht=ot*tanh(Ct) To obtain a corresponding probability distribution, x in the formulatFor a matrix of vectors of words or words input in sequence, ftValue obtained through forget gate, itFor the value obtained through the input gate,to pass the updated candidate after the gate, CtTo pass the updated state value after the gate, otIs the value of output, htFor the current hidden state, the sigmoid function is a function that maps variables between 0 and 1, σ is the sigmoid function, tanh function compresses values between-1 and 1, ht-1Is the hidden state at the previous moment, b is the bias term, W is the weight matrix, Ct-1In order to be in the last memory state,for the current memory state, a hidden state h is obtained by constructing a language model from left to right and from right to leftt1And ht2The hidden state of the spliced two is Ht。
Fourthly, splicing the characteristic C extracted by the RoBERTA model and the hidden state obtained by the Bi-LSTM to obtain a characteristic matrix [ C, Ht]C is the feature extracted between words, HtThe method solves the problem of independent hypothesis test brought by the RoBERTA model for the hidden state of post-splicing by the Bi-LSTM model, and supplements the characteristic loss caused by the word number limitation of the RoBERTA model. The CRF may obtain the named entity recognition result by using the viterbi algorithm in consideration of the constraint relationship between the labels.
In summary, the above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (2)
1. A named entity recognition method based on countermeasure training is characterized by comprising the following steps:
firstly, segmenting a referee document in the judicial field into single characters serving as training samples, and training through a RoBERTA model to obtain relevance characteristics among the characters in the judicial field;
cutting the judge text in the judicial field into single characters and phrases, converting the phrases into Word vectors by using a Word2Vec method, and converting the single characters into Word vectors based on characters by using a Fastext method; introducing Word vectors obtained by using a Word2Vec method outside the judicial field, and introducing Word vectors based on characters obtained by using a Fastext method outside the judicial field; mixing all the word vectors;
step two, disturbing the mixed word vector matrix, finding the disturbance in the worst case through the maximum value of a loss function, and obtaining the optimal robust parameter of the model by utilizing the minimization of the external experience risk so as to obtain the word vector after disturbance-resistant optimization;
step three, inputting the word vector obtained in the step two into the Bi-LSTM by using a sliding window with the length of a, and obtaining the relevance characteristics between the words in the judicial field through the training of the Bi-LSTM;
and step four, splicing the two relevance characteristics obtained in the step one and the step three together, and then predicting the training sample by using a conditional random field model to obtain a predicted result.
2. The method of claim 1, wherein 1000< a < 2000.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911358738.4A CN111191453A (en) | 2019-12-25 | 2019-12-25 | Named entity recognition method based on confrontation training |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911358738.4A CN111191453A (en) | 2019-12-25 | 2019-12-25 | Named entity recognition method based on confrontation training |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111191453A true CN111191453A (en) | 2020-05-22 |
Family
ID=70709379
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911358738.4A Pending CN111191453A (en) | 2019-12-25 | 2019-12-25 | Named entity recognition method based on confrontation training |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111191453A (en) |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111680145A (en) * | 2020-06-10 | 2020-09-18 | 北京百度网讯科技有限公司 | Knowledge representation learning method, device, equipment and storage medium |
CN111737952A (en) * | 2020-06-24 | 2020-10-02 | 深圳前海微众银行股份有限公司 | Training method and device for sequence labeling model |
CN112115721A (en) * | 2020-09-28 | 2020-12-22 | 青岛海信网络科技股份有限公司 | Named entity identification method and device |
CN112765319A (en) * | 2021-01-20 | 2021-05-07 | 中国电子信息产业集团有限公司第六研究所 | Text processing method and device, electronic equipment and storage medium |
CN112784831A (en) * | 2021-02-02 | 2021-05-11 | 电子科技大学 | Character recognition method for enhancing attention mechanism by fusing multilayer features |
CN112925918A (en) * | 2021-02-26 | 2021-06-08 | 华南理工大学 | Question-answer matching system based on disease field knowledge graph |
CN112949291A (en) * | 2021-03-02 | 2021-06-11 | 赛飞特工程技术集团有限公司 | Report error correction system and method |
CN113380418A (en) * | 2021-06-22 | 2021-09-10 | 浙江工业大学 | System for analyzing and identifying depression through dialog text |
CN113987192A (en) * | 2021-12-28 | 2022-01-28 | 中国电子科技网络信息安全有限公司 | Hot topic detection method based on RoBERTA-WWM and HDBSCAN algorithm |
WO2022078102A1 (en) * | 2020-10-14 | 2022-04-21 | 腾讯科技(深圳)有限公司 | Entity identification method and apparatus, device and storage medium |
CN114399396A (en) * | 2022-01-19 | 2022-04-26 | 中国平安人寿保险股份有限公司 | Insurance product recommendation method and device, computer equipment and storage medium |
CN114444506A (en) * | 2022-01-11 | 2022-05-06 | 四川大学 | Method for extracting relation triple fusing entity types |
CN116702787A (en) * | 2023-08-07 | 2023-09-05 | 四川隧唐科技股份有限公司 | Long text entity identification method, device, computer equipment and medium |
CN117057350A (en) * | 2023-08-07 | 2023-11-14 | 内蒙古大学 | Chinese electronic medical record named entity recognition method and system |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108229582A (en) * | 2018-02-01 | 2018-06-29 | 浙江大学 | Entity recognition dual training method is named in a kind of multitask towards medical domain |
CN110083831A (en) * | 2019-04-16 | 2019-08-02 | 武汉大学 | A kind of Chinese name entity recognition method based on BERT-BiGRU-CRF |
CN110532377A (en) * | 2019-05-13 | 2019-12-03 | 南京大学 | A kind of semi-supervised file classification method based on dual training and confrontation learning network |
-
2019
- 2019-12-25 CN CN201911358738.4A patent/CN111191453A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108229582A (en) * | 2018-02-01 | 2018-06-29 | 浙江大学 | Entity recognition dual training method is named in a kind of multitask towards medical domain |
CN110083831A (en) * | 2019-04-16 | 2019-08-02 | 武汉大学 | A kind of Chinese name entity recognition method based on BERT-BiGRU-CRF |
CN110532377A (en) * | 2019-05-13 | 2019-12-03 | 南京大学 | A kind of semi-supervised file classification method based on dual training and confrontation learning network |
Non-Patent Citations (5)
Title |
---|
GEORGY KONOPLICH等: "Named Entity Recognition in Russian with Word Representation Learned by a Bidirectional Language Model", ARTIFICIAL INTELLIGENCE AND NATURAL LANGUAGE * |
ZHENJIN DAI等: "Named Entity Recognition Using BERT BiLSTM CRF for Chinese Electronic Health Records", 12TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS * |
冯建周等: "关于命名实体识别的生成式对抗网络的研究", 小型微型计算机系统 * |
王文广等: "基于混合深度神经网络模型的司法文书智能化处理", 清华大学学报(自然科学版) * |
郭宝震等: "采用词向量注意力机制的双路卷积神经网络句子分类模型", 浙江大学学报(工学版) * |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111680145B (en) * | 2020-06-10 | 2023-08-15 | 北京百度网讯科技有限公司 | Knowledge representation learning method, apparatus, device and storage medium |
CN111680145A (en) * | 2020-06-10 | 2020-09-18 | 北京百度网讯科技有限公司 | Knowledge representation learning method, device, equipment and storage medium |
CN111737952A (en) * | 2020-06-24 | 2020-10-02 | 深圳前海微众银行股份有限公司 | Training method and device for sequence labeling model |
CN112115721A (en) * | 2020-09-28 | 2020-12-22 | 青岛海信网络科技股份有限公司 | Named entity identification method and device |
CN112115721B (en) * | 2020-09-28 | 2024-05-17 | 青岛海信网络科技股份有限公司 | Named entity recognition method and device |
WO2022078102A1 (en) * | 2020-10-14 | 2022-04-21 | 腾讯科技(深圳)有限公司 | Entity identification method and apparatus, device and storage medium |
CN112765319A (en) * | 2021-01-20 | 2021-05-07 | 中国电子信息产业集团有限公司第六研究所 | Text processing method and device, electronic equipment and storage medium |
CN112765319B (en) * | 2021-01-20 | 2021-09-03 | 中国电子信息产业集团有限公司第六研究所 | Text processing method and device, electronic equipment and storage medium |
CN112784831A (en) * | 2021-02-02 | 2021-05-11 | 电子科技大学 | Character recognition method for enhancing attention mechanism by fusing multilayer features |
CN112784831B (en) * | 2021-02-02 | 2022-06-28 | 电子科技大学 | Character recognition method for enhancing attention mechanism by fusing multilayer features |
CN112925918B (en) * | 2021-02-26 | 2023-03-24 | 华南理工大学 | Question-answer matching system based on disease field knowledge graph |
CN112925918A (en) * | 2021-02-26 | 2021-06-08 | 华南理工大学 | Question-answer matching system based on disease field knowledge graph |
CN112949291A (en) * | 2021-03-02 | 2021-06-11 | 赛飞特工程技术集团有限公司 | Report error correction system and method |
CN113380418A (en) * | 2021-06-22 | 2021-09-10 | 浙江工业大学 | System for analyzing and identifying depression through dialog text |
CN113987192B (en) * | 2021-12-28 | 2022-04-01 | 中国电子科技网络信息安全有限公司 | Hot topic detection method based on RoBERTA-WWM and HDBSCAN algorithm |
CN113987192A (en) * | 2021-12-28 | 2022-01-28 | 中国电子科技网络信息安全有限公司 | Hot topic detection method based on RoBERTA-WWM and HDBSCAN algorithm |
CN114444506A (en) * | 2022-01-11 | 2022-05-06 | 四川大学 | Method for extracting relation triple fusing entity types |
CN114444506B (en) * | 2022-01-11 | 2023-05-02 | 四川大学 | Relation triplet extraction method for fusing entity types |
CN114399396A (en) * | 2022-01-19 | 2022-04-26 | 中国平安人寿保险股份有限公司 | Insurance product recommendation method and device, computer equipment and storage medium |
CN116702787A (en) * | 2023-08-07 | 2023-09-05 | 四川隧唐科技股份有限公司 | Long text entity identification method, device, computer equipment and medium |
CN117057350A (en) * | 2023-08-07 | 2023-11-14 | 内蒙古大学 | Chinese electronic medical record named entity recognition method and system |
CN117057350B (en) * | 2023-08-07 | 2024-05-10 | 内蒙古大学 | Chinese electronic medical record named entity recognition method and system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111191453A (en) | Named entity recognition method based on confrontation training | |
CN110929030B (en) | Text abstract and emotion classification combined training method | |
CN108733792B (en) | Entity relation extraction method | |
CN110532557B (en) | Unsupervised text similarity calculation method | |
CN108563653B (en) | Method and system for constructing knowledge acquisition model in knowledge graph | |
CN108628823B (en) | Named entity recognition method combining attention mechanism and multi-task collaborative training | |
CN110334354B (en) | Chinese relation extraction method | |
CN111160467B (en) | Image description method based on conditional random field and internal semantic attention | |
CN106407333B (en) | Spoken language query identification method and device based on artificial intelligence | |
CN108062388A (en) | Interactive reply generation method and device | |
CN110909736B (en) | Image description method based on long-term and short-term memory model and target detection algorithm | |
CN111666758B (en) | Chinese word segmentation method, training device and computer readable storage medium | |
CN109800437A (en) | A kind of name entity recognition method based on Fusion Features | |
CN109858041A (en) | A kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries | |
CN109919175B (en) | Entity multi-classification method combined with attribute information | |
CN112699685B (en) | Named entity recognition method based on label-guided word fusion | |
CN114492441A (en) | BilSTM-BiDAF named entity identification method based on machine reading understanding | |
CN110162789A (en) | A kind of vocabulary sign method and device based on the Chinese phonetic alphabet | |
CN114648031B (en) | Text aspect emotion recognition method based on bidirectional LSTM and multi-head attention mechanism | |
CN115062104A (en) | Knowledge prompt-fused legal text small sample named entity identification method | |
CN114239612A (en) | Multi-modal neural machine translation method, computer equipment and storage medium | |
CN115114409A (en) | Civil aviation unsafe event combined extraction method based on soft parameter sharing | |
CN110309515B (en) | Entity identification method and device | |
CN113642630B (en) | Image description method and system based on double-path feature encoder | |
CN114239584A (en) | Named entity identification method based on self-supervision learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20200522 |