CN115293150A - Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism - Google Patents

Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism Download PDF

Info

Publication number
CN115293150A
CN115293150A CN202210959294.5A CN202210959294A CN115293150A CN 115293150 A CN115293150 A CN 115293150A CN 202210959294 A CN202210959294 A CN 202210959294A CN 115293150 A CN115293150 A CN 115293150A
Authority
CN
China
Prior art keywords
text
vector
local
feature
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210959294.5A
Other languages
Chinese (zh)
Inventor
赵岩
赵耕红
蔡巍
王忱玉
李佳宜
王跃
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Liaoning Cancer Hospital and Institute
Shenyang Neusoft Intelligent Medical Technology Research Institute Co Ltd
Original Assignee
Liaoning Cancer Hospital and Institute
Shenyang Neusoft Intelligent Medical Technology Research Institute Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Liaoning Cancer Hospital and Institute, Shenyang Neusoft Intelligent Medical Technology Research Institute Co Ltd filed Critical Liaoning Cancer Hospital and Institute
Priority to CN202210959294.5A priority Critical patent/CN115293150A/en
Publication of CN115293150A publication Critical patent/CN115293150A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a coding method, in particular to an automatic coding method for an operation record integrating a convolutional neural network and an attention mechanism. The operation type automatic coding is realized in the operation records by performing deep learning training on a large amount of operation record texts, so that the work of coding personnel is assisted. The method comprises the following steps: step 1, taking an automatic encoding task of the surgical operation as a text multi-label classification task. And 2, establishing a training model and realizing automatic coding through the training model.

Description

Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism
Technical Field
The invention relates to a coding method, in particular to an automatic coding method for surgery records integrating a convolutional neural network and a self-attention mechanism.
Background
The international disease classification (ICD) is the basis for realizing the payment according to DRGs and the reporting of medical information data such as a Hospital Quality Monitoring System (HQMS) and the like in hospitals, is an important tool for medical treatment, teaching and scientific research data retrieval, and the operation and operation records in medical record data are important resources for reflecting the treatment information of inpatients. The operation record is used for recording the detailed information of the whole operation process performed by a clinician, the encoding personnel need to accurately read the operation record on the basis of reading all the contents of the medical record, and the operation code can be directly influenced by one sentence in the operation record. Therefore, the work of the encoding personnel is a professional and technical work, the encoding classification principle needs to be mastered, the hard sleeve can not be moved by living study, and the encoding personnel with higher professional degree is relatively short of the prior art. Therefore, finding a method capable of assisting encoding personnel to perform efficient and accurate encoding has become an urgent need of hospitals. In recent years, with the gradual maturity of artificial intelligence and natural language processing technology, the method has been successfully applied in the medical field, and provides possibility for the task of automatically generating codes for surgical records.
Disclosure of Invention
The invention aims to overcome the defects in the prior art, and provides an automatic operation record coding method integrating a Convolutional Neural Network (CNN) and a Self Attention Mechanism (SAM).
In order to achieve the purpose, the invention adopts the following technical scheme, which comprises the following steps:
step 1, taking an automatic encoding task of the surgical operation as a text multi-label classification task.
And 2, establishing a training model and realizing automatic coding through the training model.
Further, in step 1, the task of automatically encoding the surgical operation as a text multi-label classification task comprises that a given surgical record text and a corresponding label { (c) 1 ,y 1 ),(c 2 ,y 2 ),...,(c m ,y m ) Using the training set as a training set;
wherein, c m ={w 1 ,w 2 ,...,w n Denotes the mth text, w n Denotes c m N-th word in (b), y m The label category corresponding to the mth text; and y is m ∈{0,1} k Where k represents the total number of tag types contained in all texts.
Further, in step 2, the establishing a training model and implementing automatic coding through the training model includes:
step 2.1, vector representation is carried out on the text by adopting a pre-training language model BERT, and each word in the text can be represented as a multi-dimensional dense vector:
w i =f(w i ) (1)
step 2.2, capturing the context bidirectional semantic dependency relationship of the text through a bidirectional long and short term memory network BilSTM; each word is caused to acquire a representation having contextual semantic information.
And 2.3, fusing local text feature extraction and global text feature extraction, and modeling the text information from two angles of local and global respectively.
Step 2.4, feature fusion and output: after the text local and global text feature representation is obtained, the text local and global text feature representation are fused to obtain a semantic feature vector X containing the local and the global, the semantic feature vector X is predicted through a multilayer perceptron, the semantic feature vector X is converted into a corresponding label prediction probability, and cross entropy loss training is performed.
Further, in step 2.2, the capturing the text context bidirectional semantic dependency relationship by using the bidirectional long-short term memory network BiLSTM includes: when the text vector input at the moment t is w t The implicit t time in the one-way LSTM sequenceLayer state h t The calculation process is as follows:
f t =σ(W f ·[h t-1 ,w t ]+b f ) (2)
i t =σ(W i ·[h t-1 ,w t ]+b i ) (3)
Figure BDA0003791665030000031
Figure BDA0003791665030000032
o t =σ(W o ·[h t-1 ,w t ]+b o ) (6)
h t =o t *tanh(c t ) (7)
in the formula, f t 、i t
Figure BDA0003791665030000033
c t 、o t Respectively a forgetting gate, an input gate, a temporary cell state, a cell state and an output gate value at the time t, wherein W and b are respectively a corresponding weight matrix and a bias term, sigma is a Sigmoid activation function, and tanh is a hyperbolic tangent activation function;
representing the positive and negative information of the text vector at each moment
Figure BDA0003791665030000034
And
Figure BDA0003791665030000035
and (3) splicing to obtain an output representation H of the BilSTM, so that each word obtains a representation with context semantic information:
Figure BDA0003791665030000036
further, in step 2.3, the local text feature extraction includes extracting text local feature information using the convolutional neural network CNN.
Furthermore, the extracting of the text local feature information by using the convolutional neural network CNN includes selecting 3 convolutional kernels with different lengths and widths equal to the word vector length to perform local feature extraction on the output H of the BiLSTM through the movement of a sliding window, and the feature map is represented as:
v i =f(w c ·h i:i+m-1 +b) (9)
v=[v 1 ,v 2 ,...,v n ] (10)
compressing the characteristic diagram by utilizing the maximum pooling layer and extracting characteristic information thereof, and then splicing the characteristic vectors extracted by different convolution kernels to form a characteristic sequence e c
u=max{v} (11)
e c =[u 1 ,u 2 ,...,u n ] (12)
Further, in step 2.3, the global text feature extraction includes extracting the text local feature information using an attention mechanism SAM.
Further, the output H of the BiLSTM is used as the input of the attention mechanism SAM, and then multiplied by three parameter matrices respectively to obtain a query matrix Q, a queried matrix K and an actual feature matrix V:
Figure BDA0003791665030000041
performing inner product operation on the Q and the V matrix and performing normalization processing to obtain a score corresponding to each word vector; then using softmax activation function to process to obtain the weight proportion of each word, multiplying the weight proportion with the actual characteristic vector V matrix to obtain the vector representation e with global information a
Figure BDA0003791665030000042
Wherein d is k In the form of a vector of dimensions, the vector,
Figure BDA0003791665030000043
the main purpose is to prevent the inner product from increasing with the increase of the dimension of the vector, and then to stabilize the gradient.
Further, in step 2.4, the vector X is:
X=add(e a ,e c ) (15)
and (3) predicting X by a multi-layer perceptron:
Figure BDA0003791665030000044
wherein, W 1 、W 2 F is a ReLu activation function as a training parameter, normalization processing is carried out through a sigmoid function, and the normalization processing is converted into corresponding label prediction probability;
the training by cross entropy loss is:
Figure BDA0003791665030000045
compared with the prior art, the invention has the beneficial effects.
The invention regards the automatic encoding task of the operation record as a text multi-label classification task, uses CNN to extract local features, uses SAM to extract global features, and obtains more comprehensive feature vector representation by fusing the features of CNN and SAM, thereby solving the problems of text feature information loss and global label dependence loss to a certain extent. Experiments show that the method provided by the invention not only can automatically generate the operation code for the operation record, but also can obtain a more accurate prediction result compared with a model which only uses CNN to perform local characteristics.
Drawings
The invention is further described with reference to the following figures and detailed description. The scope of the invention is not limited to the following expressions.
Figure 1 is a problem description schematic.
Fig. 2 is a model structure diagram.
Fig. 3 is a schematic diagram of a local information extraction layer structure.
Detailed Description
As shown in fig. 1-3, the specific embodiment: the method comprises the following steps: 1. problem description: as shown in FIG. 1, automatically extracting the mentioned surgical operations and obtaining codes from the text description of a surgical record can be described as a text multi-label classification task, i.e., a given surgical record text and its corresponding label { (c) 1 ,y 1 ),(c 2 ,y 2 ),...,(c m ,y m ) As a training set, where c m ={w 1 ,w 2 ,...,w n Denotes the mth text, w n Denotes c m N-th word in (b), y m For the tag class corresponding to the mth text, y m ∈{0,1} k Where k represents the total number of tag categories contained in all texts.
The multi-label classification model is obtained by training the training set, and new unlabeled samples can be classified into k semantic labels.
2. Introduction of a model:
1) A text embedding layer.
The invention adopts a pre-training language model BERT (Bidirectional Encoder reproduction from Transformers) to carry out vector Representation on a text, in the process, the BERT is only used as a vector Representation tool to serve downstream tasks, and each word in the text can be represented as a multi-dimensional dense vector:
w i =f(w i ) (1)
2) A BilSTM layer.
A bidirectional Long Short-Term Memory network (Bi-directional Long Short-Term Memory, bilSTM) can capture the bidirectional semantic dependency relationship of the text context. When the text vector input at the time t is w t Then the hidden layer state h at t time in the one-way LSTM sequence t The calculation process is as follows:
f t =σ(W f ·[h t-1 ,w t ]+b f ) (2)
i t =σ(W i ·[h t-1 ,w t ]+b i ) (3)
Figure BDA0003791665030000061
Figure BDA0003791665030000062
o t =σ(Wo·[h t-1 ,w t ]+b o ) (6)
h t =o t *tanh(c t ) (7)
in the formula (f) t 、i t
Figure BDA0003791665030000063
c t 、o t The cell state is a cell state, W and b are respectively a weight matrix and a bias term corresponding to the cell state, σ is a Sigmoid activation function, and tanh is a hyperbolic tangent activation function.
Representing the positive and negative information of the text vector at each moment
Figure BDA0003791665030000064
And
Figure BDA0003791665030000065
and (3) splicing to obtain an output representation H of the BilSTM, so that each word obtains a representation with context semantic information:
Figure BDA0003791665030000066
3) And fusing local and global text feature extraction layers.
In order to extract text characteristic information more comprehensively, the model integrates respective advantages of multilayer CNN and SAM, and models the text information from two angles of local and global.
a) A local information extraction layer.
The CNN is widely applied to various fields of NLP (non line segment) by virtue of the advantages of strong adaptability, simple structure, low computational complexity and the like, the method extracts local main characteristic information of a text by utilizing the CNN, and a model structure is shown as a figure 3:
the invention selects 3 convolution kernels with different lengths and widths equal to the length of a word vector to extract local features of the output H of the BilSTM through the movement of a sliding window, and the feature map is shown as follows:
v i =f(w c ·h i:i+m-1 +b) (9)
v=[v 1 ,v 2 ,...,v n ] (10)
compressing the characteristic diagram by utilizing the maximum pooling layer and extracting main characteristic information of the characteristic diagram, and then splicing the characteristic vectors extracted by different convolution kernels to form a characteristic sequence e c
u=max{v} (11)
e c =[u 1 ,u 2 ,...,u n ] (12)
b) And a global information extraction layer.
The essence of SAM is to give each word in the text a corresponding weight after weighing the context global information, and the word with the higher weight value plays a larger role in the classification task. The output H of the BilSTM is used as the input of SAM by the model, and then the SAM is multiplied by three parameter matrixes respectively to obtain a query matrix Q, a queried matrix K and an actual characteristic matrix V:
Figure BDA0003791665030000071
and performing inner product operation on the Q and the V matrix and performing normalization processing to obtain a score corresponding to each word vector. Then using softmax activation function to process to obtain the weight proportion of each word, multiplying the weight proportion by the actual characteristic vector V matrix to obtain the vector representation e with global information a
Figure BDA0003791665030000072
Wherein, d k In the form of a vector of dimensions, the vector,
Figure BDA0003791665030000073
the main purpose is to prevent the inner product from increasing with the increase of the dimension of the vector, and then to stabilize the gradient.
4) And (5) feature fusion and output.
After the local and global feature representations of the text are obtained, the local and global feature representations are fused to obtain a semantic feature vector X containing the local and global features:
X=add(e a ,e c ) (15)
finally, the model predicts X through a multilayer perceptron:
Figure BDA0003791665030000081
wherein, W 1 、W 2 F is a ReLu activation function as a training parameter, normalization processing is carried out through a sigmoid function, the normalization processing is converted into corresponding label prediction probability, and cross entropy loss training is carried out:
Figure BDA0003791665030000082
the invention regards the automatic coding task of the operation as a text multi-label classification task, and realizes automatic coding by using a BERT + BilSTM + CNN + Self Attention frame training model. Text local features and global feature information are extracted through the CNN and the SAM respectively and are fused to obtain more comprehensive feature expression vectors, and therefore a better classification effect is achieved.
It should be understood that the detailed description of the present invention is only for illustrating the present invention and is not limited by the technical solutions described in the embodiments of the present invention, and those skilled in the art should understand that the present invention can be modified or substituted equally to achieve the same technical effects; as long as the use requirements are met, the method is within the protection scope of the invention.

Claims (9)

1. An automatic encoding method for operation records fusing a convolutional neural network and a self-attention mechanism is characterized in that: the method comprises the following steps:
step 1, taking an automatic encoding task of the surgical operation as a text multi-label classification task.
And 2, establishing a training model and realizing automatic coding through the training model.
2. The method of claim 1, wherein: in step 1, the operation automatic coding task as a text multi-label classification task comprises a given operation record text and a corresponding label { (c) 1 ,y 1 ),(c 2 ,y 2 ),...,(c m ,y m ) Using the training set as a training set;
wherein, c m ={w 1 ,w 2 ,...,w n Denotes the mth text, w n Is shown by c m The nth word in (1), y m The label category corresponding to the mth text; and y is m ∈{0,1} k Where k represents the total number of tag types contained in all texts.
3. The method of claim 1, wherein: in step 2, the establishing of the training model and the realization of the automatic coding through the training model include:
step 2.1, vector representation is carried out on the text by adopting a pre-training language model BERT, and each word in the text can be represented as a multi-dimensional dense vector:
w i =f(w i ) (1)
step 2.2, capturing the context bidirectional semantic dependency relationship of the text through a bidirectional long and short term memory network BilSTM; each word is caused to acquire a representation having contextual semantic information.
2.3, fusing local text feature extraction and global text feature extraction, and modeling the text information from two angles of local and global respectively;
step 2.4, feature fusion and output: after the text local and global text feature representation is obtained, the text local and global text feature representation is fused to obtain a semantic feature vector X containing the local and the global, the semantic feature vector X is predicted through a multilayer perceptron, the semantic feature vector X is converted into a corresponding label prediction probability, and cross entropy loss training is performed.
4. The method of claim 3, wherein: in step 2.2, the capturing the context bidirectional semantic dependency relationship of the text by using the bidirectional long-short term memory network BilSTM comprises: when the text vector input at the time t is w t Then the hidden layer state h at t time in the one-way LSTM sequence t The calculation process is as follows:
f t =σ(W f ·[h t-1 ,w t ]+b f ) (2)
i t =σ(W i ·[h t-1 ,w t ]+b i ) (3)
Figure FDA0003791665020000021
Figure FDA0003791665020000022
o t =σ(W o ·[h t-1 ,w t ]+b o ) (6)
h t =o t *tanh(c t ) (7)
in the formula (f) t 、i t
Figure FDA0003791665020000023
c t 、o t A forgetting gate, an input gate and a temporary cell at the time tState, cell state and output gate value, W and b are respectively corresponding weight matrix and bias term, sigma is Sigmoid activation function, and tanh is hyperbolic tangent activation function;
representing the positive and negative information of the text vector at each moment
Figure FDA0003791665020000024
And with
Figure FDA0003791665020000025
And (3) splicing to obtain an output representation H of the BilSTM, so that each word obtains a representation with context semantic information:
Figure FDA0003791665020000026
5. the method of claim 4, wherein: in step 2.3, the local text feature extraction includes extracting text local feature information using the convolutional neural network CNN.
6. The method of claim 5, wherein: the method for extracting text local feature information by using the convolutional neural network CNN comprises the steps of selecting 3 convolutional kernels with different lengths and widths equal to the word vector length, and performing local feature extraction on the output H of the BilSTM through the movement of a sliding window, wherein the feature map is shown as follows:
v i =f(w c ·h i:i+m-1 +b) (9)
v=[v 1 ,v 2 ,...,v n ] (10)
compressing the characteristic graph by using the maximum pooling layer and extracting the characteristic information thereof, and then splicing the characteristic vectors extracted by different convolution kernels to form a characteristic sequence e c
u=max{v} (11)
e c =[u 1 ,u 2 ,...,u n ] (12)
7. The method of claim 6, wherein: in step 2.3, the global text feature extraction includes extracting text local feature information using an attention mechanism SAM.
8. The method of claim 7, wherein: taking the output H of the BilSTM as the input of an attention mechanism SAM, and multiplying the output H by three parameter matrixes respectively to obtain a query matrix Q, a queried matrix K and an actual characteristic matrix V:
Figure FDA0003791665020000031
performing inner product operation on the Q matrix and the V matrix and performing normalization processing to obtain a score corresponding to each word vector; then using softmax activation function to process to obtain the weight proportion of each word, multiplying the weight proportion by the actual characteristic vector V matrix to obtain the vector representation with global information e a
Figure FDA0003791665020000032
Wherein, d k In the form of a vector of dimensions, the vector,
Figure FDA0003791665020000033
the main purpose is to prevent the inner product from increasing with the increase of the dimension of the vector, and further to stabilize the gradient.
9. The method of claim 8, wherein: in step 2.4, vector X is:
X=add(e a ,e c ) (15)
and (3) predicting X by a multi-layer perceptron:
Figure FDA0003791665020000034
wherein, W 1 、W 2 F is a ReLu activation function as a training parameter, normalization processing is carried out through a sigmoid function, and the normalization processing is converted into corresponding label prediction probability;
training by cross entropy loss is:
Figure FDA0003791665020000041
CN202210959294.5A 2022-08-10 2022-08-10 Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism Pending CN115293150A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210959294.5A CN115293150A (en) 2022-08-10 2022-08-10 Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210959294.5A CN115293150A (en) 2022-08-10 2022-08-10 Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism

Publications (1)

Publication Number Publication Date
CN115293150A true CN115293150A (en) 2022-11-04

Family

ID=83828051

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210959294.5A Pending CN115293150A (en) 2022-08-10 2022-08-10 Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism

Country Status (1)

Country Link
CN (1) CN115293150A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116703928A (en) * 2023-08-08 2023-09-05 宁德市天铭新能源汽车配件有限公司 Automobile part production detection method and system based on machine learning
CN117807956A (en) * 2023-12-29 2024-04-02 兰州理工大学 ICD automatic coding method based on clinical text tree structure
CN118052627A (en) * 2024-04-15 2024-05-17 辽宁省网联数字科技产业有限公司 Intelligent filling method and system for bidding scheme

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116703928A (en) * 2023-08-08 2023-09-05 宁德市天铭新能源汽车配件有限公司 Automobile part production detection method and system based on machine learning
CN116703928B (en) * 2023-08-08 2023-10-27 宁德市天铭新能源汽车配件有限公司 Automobile part production detection method and system based on machine learning
CN117807956A (en) * 2023-12-29 2024-04-02 兰州理工大学 ICD automatic coding method based on clinical text tree structure
CN118052627A (en) * 2024-04-15 2024-05-17 辽宁省网联数字科技产业有限公司 Intelligent filling method and system for bidding scheme

Similar Documents

Publication Publication Date Title
CN111897908B (en) Event extraction method and system integrating dependency information and pre-training language model
CN109471895B (en) Electronic medical record phenotype extraction and phenotype name normalization method and system
CN109697285B (en) Hierarchical BilSt Chinese electronic medical record disease coding and labeling method for enhancing semantic representation
CN107330032B (en) Implicit discourse relation analysis method based on recurrent neural network
US8504361B2 (en) Deep neural networks and methods for using same
CN115293150A (en) Automatic encoding method for operation record by fusing convolutional neural network and self-attention mechanism
CN109858041B (en) Named entity recognition method combining semi-supervised learning with user-defined dictionary
CN109214006B (en) Natural language reasoning method for image enhanced hierarchical semantic representation
CN112989835B (en) Extraction method of complex medical entities
CN111651983B (en) Causal event extraction method based on self-training and noise model
CN113254675B (en) Knowledge graph construction method based on self-adaptive few-sample relation extraction
CN113822026A (en) Multi-label entity labeling method
CN111340006B (en) Sign language recognition method and system
CN111145914B (en) Method and device for determining text entity of lung cancer clinical disease seed bank
CN116628258A (en) Feature extraction and cross-modal matching retrieval method, device, equipment and medium
CN114638228A (en) Chinese named entity recognition method based on word set self-attention
CN114428860A (en) Pre-hospital emergency case text recognition method and device, terminal and storage medium
CN114582448A (en) Epidemic case information extraction framework construction method based on pre-training language model
CN112069825B (en) Entity relation joint extraction method for alert condition record data
CN117909918A (en) Monitor fault prediction method and system based on fusion characteristics
CN112863695A (en) Bidirectional long-short term memory prediction model based on quantum attention mechanism and extraction method
CN114970497B (en) Text classification method and word sense disambiguation method based on pre-training feature embedding
CN115859989A (en) Entity identification method and system based on remote supervision
CN114757183B (en) Cross-domain emotion classification method based on comparison alignment network
CN115295133A (en) Code checking method for surgical operation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination