WO2020135337A1 - 实体语义关系分类 - Google Patents
实体语义关系分类 Download PDFInfo
- Publication number
- WO2020135337A1 WO2020135337A1 PCT/CN2019/127449 CN2019127449W WO2020135337A1 WO 2020135337 A1 WO2020135337 A1 WO 2020135337A1 CN 2019127449 W CN2019127449 W CN 2019127449W WO 2020135337 A1 WO2020135337 A1 WO 2020135337A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- entity
- semantic relationship
- training sample
- vector corresponding
- text
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H10/00—ICT specially adapted for the handling or processing of patient-related medical or healthcare data
- G16H10/60—ICT specially adapted for the handling or processing of patient-related medical or healthcare data for patient-specific data, e.g. for electronic patient records
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/289—Phrasal analysis, e.g. finite state techniques or chunking
- G06F40/295—Named entity recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/70—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
Definitions
- Deep learning is a method of machine representation learning in machine learning. In the practical application of deep learning, a deep learning model needs to be trained in advance.
- the sample data used for training the deep learning model includes feature data of multiple dimensions, and the deep learning model is continuously trained according to the sample data to obtain an accurate prediction model.
- the prediction model is used to perform data prediction operations online.
- FIG. 1 is a schematic structural block diagram of an electronic device provided by an embodiment of this application.
- FIG. 2 is a schematic structural diagram of an entity semantic relationship classification model provided by an embodiment of this application.
- FIG. 3 is a schematic flowchart of a training method for an entity semantic relationship classification model provided by an embodiment of the present application
- FIG. 4 is a schematic flowchart of the sub-steps of S204 in FIG. 3;
- FIG. 5 is another schematic flowchart of the sub-step of S204 in FIG. 3;
- FIG. 6 is a schematic flowchart of an entity semantic relationship classification method provided by an embodiment of the present application.
- FIG. 7 is a schematic flowchart of sub-steps of S303 in FIG. 6;
- FIG. 8 is a schematic flowchart of the sub-steps of S3031 in FIG. 7;
- FIG. 9 is a schematic structural diagram of an entity semantic relationship classification model training device provided by an embodiment of this application.
- FIG. 10 is a schematic structural diagram of an entity semantic relationship classification device according to an embodiment of the present application.
- a deep learning model can be used to deepen the text information on the basis of entity recognition, thereby promoting unstructured sentence structure.
- the entity is a naming reference item, such as the name of a person, place, equipment, disease, etc.
- naming reference item such as the name of a person, place, equipment, disease, etc.
- a classification method based on a neural network model is generally used to classify the entity semantic relationship.
- the specific method is to use a large amount of corpus that has been classified as entity semantic relations as input to the neural network model to train the neural network model, and then use the trained neural network model to classify the entity semantic relationship of the new corpus .
- RNTN Recursive Neural Tensor Network, Recurrent Neural Tensor Network
- PCNN Pulse Coupled Neural Network, etc.
- EMR Electronic Medical Record
- FIG. 1 is a schematic structural block diagram of an electronic device 100 according to an embodiment of the present application.
- the electronic device 100 may be, but not limited to, a server, a personal computer (Personal Computer, PC), a tablet computer, a smart phone, a personal digital assistant (Personal Digital Assistant, PDA), and so on.
- the electronic device 100 includes a memory 101, a processor 102, and a communication interface 103.
- the memory 101, the processor 102, and the communication interface 103 are directly or indirectly electrically connected to each other to implement data transmission or interaction.
- the memory 101, the processor 102, and the communication interface 103 may be electrically connected to each other through one or more communication buses or signal lines.
- the memory 101 may be used to store program instructions and modules, such as the program instructions/modules corresponding to the entity semantic relationship classification model training device 400 and the program instructions/modules corresponding to the entity semantic relationship classification device 500 provided in the embodiments of the present application, and the processor 102 By executing the program instructions and modules stored in the memory 101, various functional applications and data processing are executed.
- the communication interface 103 can be used for signaling or data communication with other node devices.
- the memory 101 may be, but not limited to, random access memory (Random Access Memory, RAM), read-only memory (Read Only Memory, ROM), programmable read-only memory (Programmable Read-Only Memory, PROM), erasable In addition to read-only memory (Erasable Programmable Read-Only Memory, EPROM), electrically erasable read-only memory (Electric Erasable Programmable Read-Only Memory, EEPROM), etc.
- RAM Random Access Memory
- ROM read-only memory
- PROM Programmable Read-Only Memory
- EPROM Erasable Programmable Read-Only Memory
- EEPROM Electrically erasable read-only memory
- the processor 102 may be an integrated circuit chip with signal processing capabilities.
- the processor 102 may be a general-purpose processor, including but not limited to a central processing unit (Central Processing Unit, CPU), a digital signal processor (Digital Signal Processing, DSP), an NPU (Neural-network Processing Units, embedded neural network processing
- the processor 102 can also be an application specific integrated circuit (Application Specific Integrated Circuit, ASIC), field programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic device, discrete gate or transistor logic device , Discrete hardware components, etc.
- ASIC Application Specific Integrated Circuit
- FPGA Field-Programmable Gate Array
- FIG. 1 is only an illustration, and the electronic device 100 may include more or fewer components than those shown in FIG. 1 or have a configuration different from that shown in FIG. 1.
- Each component shown in FIG. 1 may be implemented by hardware, software, or a combination thereof.
- the following takes an example of mining entities in a corpus and classifying entity semantic relationships as an example to describe a method for training an entity semantic relationship classification model provided by the embodiments of the present application, so as to use the trained entity semantic relationship classification model, Classify the semantic relations of entities in the corpus.
- FIG. 2 is a schematic structural diagram of an entity semantic relationship classification model provided by an embodiment of the present application.
- the entity semantic relationship classification model may use bidirectional A model of Gated Recurrent Neural Network (BiGated Recurrent Unit, BiGRU) combined with Attention mechanism.
- BiGRU Gated Recurrent Neural Network
- the entity semantic relationship classification model is to add an attention layer before the output layer of the BiGRU model.
- the level of the GRU layer is set to 1, and the number of neurons in the GRU layer is 230.
- the user can also set the level of the GRU layer to 2 or other levels according to actual needs, and correspondingly set the number of neurons in the GRU layer to other numbers.
- the embodiments of the present application only provide one possible implementation manner, and do not limit the specific number.
- the dropout parameter of the BiGRU+Attention model is set to 1, that is, each neuron of the GRU layer in the BiGRU+Attention model is not discarded during training.
- the user can also set the Dropout parameter to other values as needed to determine inactivation of some neurons in the GRU layer.
- the entity relationship semantic classification model may also adopt other models, such as the GRU model.
- the embodiments of the present application are exemplified based on the BiGRU+Attention model.
- FIG. 3 is a schematic flowchart of an entity semantic relationship classification model training method provided by an embodiment of the present application.
- the entity semantic relationship classification model training method is applied to the electronic device 100 shown in FIG.
- the electronic device 100 uses an entity semantic relationship classification model to classify the semantic relationship of entities in the corpus.
- the training method of the entity semantic relation classification model includes the following steps:
- S201 Receive at least one training sample and identify the first entity and the second entity of each training sample in the at least one training sample.
- the training samples may be received multiple times at a time, or may be received one at a time, and the user may adjust according to actual needs in the practice process.
- S202 Obtain a first position distance between each text in each training sample and the corresponding first entity and a second position distance from the corresponding second entity.
- Step S202 is specifically: for each training sample, obtain a first position distance between each text in the training sample and the first entity of the training sample, and obtain a first position distance between each text in the training sample and the training sample The second position distance of the two entities.
- S203 Combine feature vectors corresponding to all characters in each training sample to obtain a model input vector corresponding to each training sample.
- Step S203 is specifically: for each training sample, combine feature vectors corresponding to all characters in the training sample to obtain a model input vector corresponding to the training sample.
- the feature vector corresponding to each text is obtained by combining the word vector corresponding to each text and the position embedding vector.
- the position embedding vector corresponding to each text includes the vector corresponding to the first position distance of each text, each text The second position distance corresponds to the vector. That is, for each character, the feature vector corresponding to the character is obtained by combining the character vector corresponding to the character and the position embedding vector.
- the position embedding vector corresponding to the character includes the vector corresponding to the first position distance of the character, the character The second position distance corresponds to the vector.
- the model input vector corresponding to each training sample is used as the input of the entity semantic relationship classification model to train the entity semantic relationship classification model.
- the training parameters can be set as follows: batch_size is set to 50; epochs is set to 100, that is, each training sample is used 100 times; after each training 100 times, Save the model.
- the significance of the training parameter is that each time 50 training samples are used to train the entity semantic relationship classification model, and each training sample is used 100 times, and the entity semantic relationship classification model is saved after each training 100 times.
- the electronic device when using each batch of training samples to train the entity semantic relationship classification model, the electronic device needs to receive 50 training samples and identify the first entity and the second entity in each training sample, where, The first entity and the second entity in each training sample constitute an entity pair.
- each entity in the training sample has an entity identifier, and the electronic device recognizes the entity identifier in each training sample to obtain each The first entity and the second entity in the training sample.
- a training sample is: there is no ⁇ e1> bulging symptom ⁇ /e1> in the precordial area, depression, normal apical beats, no pericardial friction, and normal relative voiced voice, ⁇ e2 >Heart rate test ⁇ /e2>70 beats/min, the heart rhythm is uniform, no pathological murmur is heard in the auscultation area of each heart valve.
- the electronic device recognizes ⁇ e1> ⁇ /e1> when receiving the training sample,
- the obtained first entity is "uplift”, the type is symptom, recognition ⁇ e2> ⁇ /e2>, the obtained second entity is "heart rate”, the type is test.
- other methods may also be used to identify the first entity and the second entity in the training sample, for example, by presetting an entity library and storing multiple entities in the entity library, Therefore, the training samples are identified based on the index of the preset entity library to obtain the first entity and the second entity.
- each word in a single training sample has a different position relative to the first entity and the second entity, then each word has an entity semantic relationship to both the first entity and the second entity
- the contribution of type recognition is also different.
- each text in each training sample obtained is the first position distance from each corresponding first entity, and The second position distance from the corresponding second entity.
- the entity is generally composed of multiple words, such as “swelling” and “heart rate” in the above example
- the first word in the entity is used as the calculation standard for the distance of the second position,
- the positional distance relative to the same entity may be the same, but the difference is that one word is in front of the entity in word order, and the other word is behind the entity, for example in the above exemplary training sample
- the word “zone” is at a distance of 2 from the first position of the "bulge” of the first entity, but the word “concave” in the “depression” is also the same as the first position of the "bulge” of the first entity.
- a location distance is 2.
- positive and negative values can be used to distinguish the direction of the position distance.
- the position distance of the word before the word order of the entity is represented by a negative value
- the position distance of the entity The positional distance of the word after the word order is expressed by a positive value.
- the word "zone" in the "heart front zone” is located before the word order of the first entity "uplift”
- the word “concave” in “depression” is located in the first entity "uplift”
- the distance of the first position of the word "region” is -2
- the distance of the first position of the word "recess” is 2.
- the word vector obtained after vectorization of the "front” word is [0.1, 0.5, 0.4, 0.3]
- the vector corresponding to the first position distance of the first word is [0.4,0.6]
- the vector corresponding to the distance of the second position is [0.6,0.4]
- the feature vector corresponding to the first word is [0.1,0.5,0.4,0.3 , 0.4, 0.6, 0.6, 0.4].
- the word vector obtained after vectorization of the above "previous" word is a 4-dimensional vector [0.1, 0.5, 0.4, 0.3] is only for illustration. In some other implementations of the embodiments of the present application, it may also be pre-stored in an electronic device
- the word vector table of other dimension numbers enables word vectors with different dimension numbers to be obtained after vectorization of the "previous" word. For example, using a word vector table with a number of dimensions pre-stored in an electronic device of 100, the word vector with 100 dimensions is obtained after vectorization of the "previous" word.
- the vector corresponding to the first position distance and the vector corresponding to the second position distance are both schematic illustrations, and use 2 dimensions. In some other implementations of the embodiments of the present application, other pre-stored electronic devices may also be used.
- the position of the number of dimensions is embedded in the vector table, for example, the position embedding vector table with 4 dimensions.
- a model input vector corresponding to each training sample may be recorded in a two-dimensional matrix arrangement.
- the feature vector corresponding to the word “heart” in the “heart front zone” is used as the first line of the model input vector; the feature corresponding to the word “front” in the “heart front zone” The vector is used as the second line of the model input vector; and so on, until the feature vectors corresponding to all the characters in the training sample are combined to obtain the model input vector corresponding to the training sample.
- the model input vector corresponding to each training sample is used as the input of the entity semantic relationship classification model to train the entity semantic relationship classification model.
- the corresponding model input vectors of multiple training samples can be used as the input of the entity semantic relationship classification model, for example, according to the above
- the input of the entity semantic relationship classification model is the model input vector corresponding to 50 training samples.
- each model input vector has the same dimension. For example, taking a word vector with a number of dimensions of 100 and a position embedding vector with a number of dimensions of 4 as an example, the number of dimensions of the model input vector of training sample 1 is 70*108, then the number of dimensions of the model input vector of training sample 2 is also 70*108.
- 70 represents that the training sample contains 70 texts
- 108 represents that the feature vector corresponding to each text contains 108 elements
- these 108 elements include 100 elements of the word vector
- the first position is away from the corresponding position of the embedding vector. 4 elements, and 4 elements of the vector embedded in the position corresponding to the second position distance.
- training sample 1 contains 60 words
- training sample 2 contains 70 words
- Sample 3 contains 73 words
- the model input vectors corresponding to multiple training samples need to be unified. That is, the number of dimensions of the model input vector corresponding to multiple training samples is unified.
- the training sample 1 can be The part of less than 70 words is filled with a pre-set vector, such as the 0 vector, and the 70*108-dimensional model input vector corresponding to training sample 1 is obtained; in addition, the combination of each text in training sample 3 is combined.
- the feature vector gets a 73*108 dimension vector, 73>70, you can remove the part of the training sample 3 that exceeds 70 words, and only retain the feature vector corresponding to the first 70 words in word order, and then get the corresponding corresponding to the training sample 3 70*108 dimension model input vector.
- the above training sample may use an electronic medical record, and the model input vector corresponding to the training sample is a combination of n feature vectors.
- the dimension of the model input vector is set to 70*108, which means that the model input vector contains 70 character corresponding feature vectors, and the dimension of each feature vector is 108.
- n is the average number of words contained in at least one electronic medical record (that is, the above training sample). For example, if the training sample uses 50 electronic medical records in total, and the average number of words contained in the 50 electronic medical records is 70, then n is equal to 70.
- n may also be set to a fixed value, for example, n is set to 100.
- corpora other than electronic medical records may also be used as training samples. For example, intelligent customer service dialogue or consultation information, etc.
- the model input vector corresponding to the single training sample is also used as the input of the entity semantic relationship classification model.
- batch_size is set to 1
- the input of the entity semantic relationship classification model is a model input vector corresponding to 1 training sample.
- the input layer of the model obtains a 70*108-dimensional model input vector, after preprocessing by the feature embedding layer, it is input to the GRU layer for calculation; GRU layer output 108 predicted entity semantic relationship types to the attention layer; the attention layer calculates the respective probability value of each predicted entity semantic relationship type according to the obtained 108 predicted entity semantic types, and the obtained 108 predicted entities Among the semantic relationship types, the entity semantic relationship type with the largest probability value is used as the entity semantic relationship type corresponding to the training sample.
- the entity semantic relationship classification model will give the predicted entity semantic relationship type for each training sample. For example, in the above example, enter the model input vector corresponding to 50 training samples, then the entity semantic relationship classification model will get 50 training The prediction entity semantic relationship type corresponding to each sample.
- FIG. 4 is a schematic flowchart of the sub-steps of S204 in FIG. 3.
- S204 includes the following sub-steps:
- Step S2041 is to use the model input vector corresponding to each training sample as the input of the entity semantic relationship classification model, to obtain the predicted entity semantic relationship type corresponding to each training sample obtained by the entity semantic relationship classification model, and to predict the entity semantic relationship
- the type is the predicted entity semantic relationship type of both the first entity and the second entity in each training sample.
- S2042 Obtain the deviation value of the entity semantic relationship type in each training sample and the entity semantic relationship type of the first entity and the second entity stored in advance for each training sample.
- Step S2042 is to obtain the deviation value between the predicted entity semantic relationship type and the preset entity semantic relationship type corresponding to each training sample, and the preset entity semantic relationship type corresponds to each pre-stored first entity and second entity corresponding to each training sample The type of entity semantic relationship between the two.
- the cross entropy function can be used to calculate the deviation value of each training sample.
- each training sample corresponds to the predicted entity semantic relationship type and the preset entity semantic relationship type as the input of the cross entropy function, and then each The cross entropy function value corresponding to the training sample is used as the deviation value corresponding to each training sample; then the deviation value of each training sample in a training process is added to obtain the deviation value of each training sample in the training process And, the sum of the deviation values characterizes the overall deviation degree of the training process. For example, in the above example where batch_size is set to 50, the sum of the deviation values is the result obtained by adding the deviation values of 50 training samples.
- the overall deviation characterizing the training process is large, and the entity semantic relationship type predicted by the entity semantic relationship classification model differs greatly from the actual entity semantic relationship type. Adjust the entity semantic relationship The parameters in the classification model to train the classification model of the semantic relationship of the entity. Conversely, if the sum of the deviation values does not exceed the first deviation threshold, the entity semantic relationship type predicted by the entity semantic relationship classification model is closer to the actual entity semantic relationship type. This training result meets the training requirements, and the model training is judged to end. .
- the above training process is to adjust the parameters of the entity semantic relationship classification model by using the overall deviation degree of a single training of the entity semantic relationship classification model.
- the output of a single training sample can also be used to adjust the parameters of the entity semantic relationship classification model.
- FIG. 5 is another schematic flowchart of the sub-steps of S204 in FIG. 3.
- S204 may further include the following sub-steps:
- S2042 Obtain the sum of the entity semantic relationship types in each training sample, and correspond to the pre-stored deviation values of the entity semantic relationship types of the first entity and the second entity for each training sample.
- S2045 Determine whether the deviation value of the target training sample exceeds the second deviation threshold; if yes, adjust the parameters in the entity semantic relation classification model to train the entity semantic relation classification model; if not, end the training.
- the cross entropy function is used to calculate the deviation value of each training sample; among all the training samples of the input entity semantic relationship classification model, the target training sample is determined. If the deviation value of the target training sample exceeds the second deviation threshold, it means that the training process does not meet the training requirements, and then the parameters in the entity semantic relationship classification model are adjusted to train the entity semantic relationship classification model. Conversely, if the deviation value of the target training sample does not exceed the second deviation threshold, it indicates that the training result meets the training requirements, and it is determined that the model training is ended.
- the target training sample may be any of all training samples of the input entity semantic relationship classification model, may be any training sample whose deviation value exceeds the second deviation threshold, or may be a traversal of the input entity semantic relationship classification model in sequence For all training samples of, each training sample is used as the target training sample for judgment. In some other implementation manners of the embodiments of the present application, the training sample with the largest deviation value among all the training samples of the input model may also be used as the target training sample.
- the method corresponding to FIG. 4 is to adjust the parameters of the entity semantic relationship classification model according to the overall deviation degree of a single training
- the method corresponding to FIG. 5 is to adjust the parameters of the entity semantic relationship classification model using the output result of a single training sample
- the parameters in the entity semantic relationship classification model are adjusted to train the entity semantic relationship classification During the model, the weight coefficient and bias coefficient of the GRU layer in the BiGRU+Attention model and the attention matrix of the attention layer are adjusted to achieve the purpose of training the entity semantic relationship classification model.
- the following takes mining entities in a corpus and classifying entity semantic relationships as an example. Based on the entity semantic relationship classification model obtained after the training of the above entity semantic relationship classification model training method, an entity semantic relationship classification method provided by the embodiments of the present application is provided. Be explained.
- FIG. 6 is a schematic flowchart of a method for classifying entity semantic relationship according to an embodiment of the present application.
- the method for classifying entity semantic relationship may be applied to an electronic device shown in FIG.
- the relationship classification method includes the following steps:
- S301 Determine the first entity and the second entity in the corpus.
- S302 Obtain a first position distance between each text in the corpus and the first entity and a second position distance from the second entity.
- S303 Combine feature vectors corresponding to all characters in the corpus to obtain a model input vector corresponding to the corpus.
- the feature vector corresponding to each text is obtained by combining the word vector corresponding to each text in the corpus with the position embedding vector.
- the position embedding vector corresponding to each text includes the vector corresponding to the first position distance of each text, each The distance between the second position of each text is the corresponding vector.
- S304 Use the model input vector corresponding to the corpus as the input of the preset entity semantic relationship classification model to determine the entity semantic relationship type of both the first entity and the second entity.
- Step S304 is to use the model input vector corresponding to the corpus as the input of the entity semantic relationship classification model to determine the entity semantic relationship type of both the first entity and the second entity.
- the corpus may use electronic medical records.
- the anterior region has no ⁇ e1>protruding symptom ⁇ /e1>, depression, normal apical beats, no pericardial friction, and the heart is relatively dull and normal
- ⁇ e2>heart rate test ⁇ /e2>70 times /Minute the heart rhythm is equal
- the heart valve auscultation area is unheard and pathological murmurs
- the electronic device obtains the type of entity semantic relationship between the entity pairs in the corpus, as a possible implementation, it can be based on the entity identifiers " ⁇ e1> ⁇ /e1>” and " ⁇ e2> ⁇ / e2>”, it is determined that the first entity in the corpus is “uplift” and the second entity is “heart rate”.
- an entity library preset in the electronic device may also be used, and multiple entities are pre-stored in the entity library, so as to index based on the preset entity library
- the corpus is identified to obtain the above-mentioned first entity "swell” and second entity "heart rate".
- each text is obtained separately from the first The first position distance of an entity and the second position distance of the second entity, so that the word vector corresponding to each text, the vector corresponding to the first position distance of each text, and the second position distance of each text correspond
- the vectors are combined to obtain the feature vector corresponding to each text, and then the feature vectors corresponding to all the words in the corpus are combined to obtain the model input vector corresponding to the corpus, and the model input vector is used as the entity semantic relationship in the electronic device
- the input of the classification model determines the type of entity semantic relationship between the first entity "uplift” and the second entity "heart rate" in the corpus.
- an entity semantic relationship classification model training method obtains the first entity and the second entity in the corpus by determining, and according to the first position of each text in the corpus and the first entity Distance and the second position distance between each text in the corpus and the second entity to obtain the feature vector corresponding to each text, and then combine the feature vectors corresponding to all the text in the corpus to obtain the model input vector corresponding to the corpus, thus
- the model input vector corresponding to the corpus is used as the input of the entity semantic relationship classification model to obtain the entity semantic relationship type corresponding to the corpus.
- it can improve the classification accuracy of entity semantic relations.
- FIG. 7 is a schematic flowchart of the sub-steps of S303 in FIG. 6.
- S303 includes the following sub-steps:
- S3031 Obtain a word vector corresponding to each character in the corpus, and a first position embedding vector and a second position embedding vector corresponding to the first position distance and the second position distance of each character.
- Step S3031 is to obtain a word vector corresponding to each text in the corpus, and obtain a first position embedding vector corresponding to the first position distance of each text, and a second position embedding vector corresponding to the second position of each text.
- S3032 Combine the word vector corresponding to each character, the first position embedding vector, and the second position embedding vector to obtain a feature vector corresponding to each character.
- the electronic device obtains the model input vector corresponding to the corpus, the above corpus "there is no ⁇ e1>protrusion symptom ⁇ /e1> and depression in the precordial area, the apex beats are normal, and there is no pericardial friction feeling, and the heart is relatively dull and normal, ⁇ e2> "Heart rate test ⁇ /e2> 70 beats/min, the heart rhythm is uniform, no pathological murmurs are heard in the auscultation area of the heart valves" as an example.
- the feature vector corresponding to each text is obtained according to the above-mentioned "heart" word, and then the corpus is obtained in the manner of obtaining the model input vector corresponding to the training sample in the above step of training the model.
- the feature vectors corresponding to all the characters in are combined to obtain the model input vector corresponding to the corpus.
- FIG. 8 is a schematic flowchart of the sub-steps of S3031 in FIG. 7.
- S3031 includes the following sub-steps:
- S30312 Determine, in the position embedding vector table, the first position embedding vector and the second position embedding vector corresponding to the first position distance and the second position distance, respectively.
- Step S30312 is: in the position embedding vector table, the first position embedding vector corresponding to the first position distance of each character and the second position embedding vector corresponding to the second position distance of each character are determined respectively.
- a position embedding vector table is stored in the electronic device, and the position embedding vector table has a correspondence relationship between position distance and position embedding vector. According to the position embedding vector table, the first position embedding vector can be converted from the first position distance, and the second position embedding vector can be converted from the second position distance.
- the position embedding vector table may be a vector with the number of dimensions m*n, and each column of elements in the position embedding vector table constitutes a specific position embedding vector, using specific values of the first position distance and the second position distance, Query the number of corresponding columns in the position embedding vector table, take all elements at the first position distance from the corresponding column as the first position embedding vector corresponding to the first position distance, and all elements at the second position distance from the corresponding column as the second position The second position corresponding to the distance embeds the vector.
- the third column of the position embedding vector table is queried, and all elements contained in the third column of the position embedding vector table are used as the first position embedding vector;
- the second position distance When it is "33” the 33rd column of the position embedding vector table is queried, and all elements contained in the 33rd column of the position embedding vector table are used as the second position embedding vector.
- the position embedding vector can also be directly represented by the size of the position distance value, for example, in the above example, the first position distance is “3” and the second position distance is “33", Then the embedding vector in the first position is "3" and the embedding vector in the second position is "33".
- the size of the position distance value is used to directly represent the position embedding vector, which can be regarded as a method of representing the position embedding vector using a one-dimensional vector.
- the position embedding vector table may be the entity semantic relationship classification model described above. Before being used to identify the entity semantic relationship type of both the first entity and the second entity in the corpus, back propagation (BackPropagation) , BP) algorithm to generate.
- BackPropagation BackPropagation
- BP back propagation
- the randomly generated initial vector table is continuously optimized by the BP algorithm to obtain the position embedding vector table.
- a 4-layer neural network structure is set, including 1 input layer L 1 , two hidden layers L 2 and L 3 , and 1 output layer L 4 , and neurons of input layer L 1
- the number is set to 10
- the number of neurons in the hidden layers L 2 and L 3 is set to 256
- the neurons in the output layer L 4 The number of is set to 3
- stop learning When the global error is less than the preset threshold, stop learning, and use the output result of the output layer in the last learning as the position embedded in the vector table; or when the global error is not less than the preset threshold, but the number of learning reaches 20,000 times, Stop learning, and embed the output result of the output layer in the last learning as the position into the vector table.
- FIG. 9 is a schematic structural diagram of an entity semantic relationship classification model training device 400 provided by an embodiment of the present application, and is applied to an electronic device.
- the electronic device presets an entity semantic relationship classification model .
- the entity semantic relationship classification model training device 400 includes a transceiver module 401, a second processing module 402, and a training module 403.
- the transceiver module 401 is used to receive at least one training sample and identify the first entity and the second entity of each training sample in the at least one training sample.
- the second processing module 402 is used to obtain, for each training sample, the first position distance between each text in the training sample and the first entity of the training sample, and to obtain each text in the training sample separately from the training sample The second position distance of the second entity.
- the second processing module 402 is also used to combine the feature vectors corresponding to all the text in each training sample to obtain the model input vector corresponding to each training sample, where the feature vector corresponding to each text is composed of each training sample.
- the word vector corresponding to each text in the text is combined with the position embedding vector.
- the position embedding vector corresponding to each text includes a vector corresponding to the first position distance of each text and a vector corresponding to the second position distance of each text. .
- the training module 403 is used to input the model input vector corresponding to each training sample as the input of the entity semantic relationship classification model to train the entity semantic relationship classification model.
- the training module 403 may specifically be used for:
- the predicted entity semantic relationship type is the predicted entity The type of entity semantic relationship between the first entity and the second entity in the training samples;
- the preset entity semantic relationship type is the entity semantics of each pre-stored first entity and second entity corresponding to each training sample Relationship type
- the parameters in the entity semantic relationship classification model are adjusted to train the entity semantic relationship classification model.
- the training module 403 may specifically be used for:
- the predicted entity semantic relationship type is the predicted entity The type of entity semantic relationship between the first entity and the second entity in the training samples;
- the preset entity semantic relationship type is the entity semantics of each pre-stored first entity and second entity corresponding to each training sample Relationship type
- the parameters in the entity semantic relationship classification model are adjusted according to the target training sample deviation value to train the entity semantic relationship classification model.
- the entity semantic relationship classification model is a model combining BiGRU and Attention mechanism
- the training module 403 may specifically be used to:
- each neuron of the GRU layer in the entity semantic relationship classification model during training is not discarded.
- the entity semantic relationship classification model is a model combining BiGRU and Attention mechanism
- At least one training sample is at least one electronic medical record
- the model input vector corresponding to the training sample is a combination of n feature vectors, where n is the average number of words contained in at least one electronic medical record.
- FIG. 10 is a schematic structural diagram of an entity semantic relationship classification apparatus 500 according to an embodiment of the present application, which is applied to an electronic device.
- the electronic device presets an entity semantic relationship classification model.
- the relationship classification device 500 includes a first processing module 501 and an identification module 502.
- the first processing module 501 is used to determine the first entity and the second entity in the corpus.
- the first processing module 501 is further used to obtain the first position distance of each text in the corpus from the first entity and the second position distance from the second entity.
- the recognition module 502 is used to combine the feature vectors corresponding to all the words in the corpus to obtain the model input vector corresponding to the corpus, where the feature vector corresponding to each text is performed by the word vector corresponding to each text in the corpus and the position embedding vector It is obtained after combination that the position embedding vector corresponding to each text includes a vector corresponding to the first position distance of each text and a vector corresponding to the second position distance of each text.
- the first processing module 501 may specifically be used for:
- the feature vectors corresponding to all the words in the corpus are combined to obtain the model input vector corresponding to the corpus.
- the first processing module 501 may specifically be used for:
- the first position embedding vector corresponding to the first position distance of each character and the second position embedding vector corresponding to the second position distance of each character are determined respectively.
- the entity semantic relationship classification model is a model combining BiGRU and Attention mechanism
- the corpus is an electronic medical record.
- An embodiment of the present application provides an electronic device.
- the electronic device includes a memory and a processor.
- the memory is used to store one or more programs and a preset entity semantic relationship classification model
- For each training sample obtain the first positional distance between each text in the training sample and the first entity of the training sample, and obtain the second position of each text in the training sample and the second entity of the training sample Location distance
- the feature vector corresponding to each text is made by the word vector corresponding to each text and the position embedding vector It is obtained after combination that the position embedding vector corresponding to each text includes the vector corresponding to the first position distance of each text and the vector corresponding to the second position distance of each text;
- the model input vector corresponding to each training sample is used as the input of the entity semantic relationship classification model to train the entity semantic relationship classification model.
- the specific implementation may be:
- the predicted entity semantic relationship type is the predicted entity The type of entity semantic relationship between the first entity and the second entity in the training samples;
- the preset entity semantic relationship type is the entity semantics of each pre-stored first entity and second entity corresponding to each training sample Relationship type
- the parameters in the entity semantic relationship classification model are adjusted to train the entity semantic relationship classification model.
- the specific implementation may be:
- the predicted entity semantic relationship type is the predicted entity The type of entity semantic relationship between the first entity and the second entity in the training samples;
- the preset entity semantic relationship type is the entity semantics of each pre-stored first entity and second entity corresponding to each training sample Relationship type
- the parameters in the entity semantic relationship classification model are adjusted to train the entity semantic relationship classification model.
- the entity semantic relationship classification model is a model combining BiGRU and Attention mechanism
- each neuron of the GRU layer in the entity semantic relationship classification model during training is not discarded.
- the entity semantic relationship classification model is a model combining BiGRU and Attention mechanism
- At least one training sample is at least one electronic medical record
- the model input vector corresponding to the training sample is a combination of n feature vectors, where n is the average number of words contained in at least one electronic medical record.
- An embodiment of the present application provides an electronic device.
- the electronic device includes a memory and a processor.
- the memory is used to store one or more programs and a preset entity semantic relationship classification model
- the feature vectors corresponding to all the words in the corpus are combined to obtain the model input vector corresponding to the corpus, where the feature vector corresponding to each text is obtained by combining the word vector corresponding to each text in the corpus with the position embedding vector.
- the position embedding vector corresponding to each text includes a vector corresponding to the first position distance of each text and a vector corresponding to the second position distance of each text;
- the model input vector corresponding to the corpus is used as the input of the entity semantic relationship classification model to determine the entity semantic relationship type of both the first entity and the second entity.
- the specific implementation may be:
- the feature vectors corresponding to all the characters in the corpus are combined to obtain the model input vector corresponding to the corpus.
- the specific implementation may be:
- the first position embedding vector corresponding to the first position distance of each character and the second position embedding vector corresponding to the second position distance of each character are determined respectively.
- the entity semantic relationship classification model is a model combining BiGRU and Attention mechanism
- the corpus is an electronic medical record.
- An embodiment of the present application also provides a computer-readable storage medium on which a computer program and a preset entity semantic relationship classification model are stored.
- the computer program is implemented when executed by a processor:
- the feature vectors corresponding to all the words in the corpus are combined to obtain the model input vector corresponding to the corpus, where the feature vector corresponding to each text is obtained by combining the word vector corresponding to each text in the corpus with the position embedding vector.
- the position embedding vector corresponding to each text includes a vector corresponding to the first position distance of each text and a vector corresponding to the second position distance of each text;
- the model input vector corresponding to the corpus is used as the input of the entity semantic relationship classification model to determine the entity semantic relationship type of both the first entity and the second entity.
- An embodiment of the present application also provides a computer-readable storage medium on which a computer program and a preset entity semantic relationship classification model are stored.
- the computer program is implemented when executed by a processor:
- For each training sample obtain the first positional distance between each text in the training sample and the first entity of the training sample, and obtain the second position of each text in the training sample and the second entity of the training sample Location distance
- the feature vector corresponding to each text is made by the word vector corresponding to each text and the position embedding vector It is obtained after combination that the position embedding vector corresponding to each text includes the vector corresponding to the first position distance of each text and the vector corresponding to the second position distance of each text;
- the model input vector corresponding to each training sample is used as the input of the entity semantic relationship classification model to train the entity semantic relationship classification model.
- An embodiment of the present application also provides a computer program, which is implemented when the computer program is executed by a processor:
- the feature vectors corresponding to all the words in the corpus are combined to obtain the model input vector corresponding to the corpus, where the feature vector corresponding to each text is obtained by combining the word vector corresponding to each text in the corpus with the position embedding vector.
- the position embedding vector corresponding to each text includes a vector corresponding to the first position distance of each text and a vector corresponding to the second position distance of each text;
- the model input vector corresponding to the corpus is used as the input of the preset entity semantic relationship classification model to determine the entity semantic relationship type of both the first entity and the second entity.
- An embodiment of the present application also provides a computer program, which is implemented when the computer program is executed by a processor:
- For each training sample obtain the first positional distance between each text in the training sample and the first entity of the training sample, and obtain the second position of each text in the training sample and the second entity of the training sample Location distance
- the feature vector corresponding to each text is made by the word vector corresponding to each text and the position embedding vector It is obtained after combination that the position embedding vector corresponding to each text includes the vector corresponding to the first position distance of each text and the vector corresponding to the second position distance of each text;
- the model input vector corresponding to each training sample is used as the input of the preset entity semantic relationship classification model to train the entity semantic relationship classification model.
- each block in the flowchart or block diagram may represent a module, a program segment, or a part of code, and the module, program segment, or part of the code contains one or more executables for implementing prescribed logical functions instruction.
- the functions noted in the block may occur out of the order noted in the figures.
- each block in the block diagrams and/or flowcharts, and combinations of blocks in the block diagrams and/or flowcharts can be implemented with a dedicated hardware-based system that performs specified functions or actions Or, it can be realized by a combination of dedicated hardware and computer instructions.
- the functional modules in the embodiments of the present application may be integrated together to form an independent part, or each module may exist alone, or two or more modules may be integrated to form an independent part.
- an entity semantic relationship classification method, model training method, device, electronic device, storage medium, and computer program provided by the embodiments of the present application obtain the first entity and the second entity in the corpus by determining, and according to the corpus The first position distance between each text in the first entity and the second position distance between each text in the second entity, to obtain the feature vector corresponding to each text, and then combine the feature vectors corresponding to all the characters in the corpus, Obtain the model input vector corresponding to the corpus, so that the model input vector corresponding to the corpus is used as the input of the entity semantic relationship classification model to obtain the entity semantic relationship type corresponding to the corpus.
- it can improve the classification accuracy of entity semantic relations.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Public Health (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- Biomedical Technology (AREA)
- Databases & Information Systems (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Pathology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Machine Translation (AREA)
Abstract
Description
Claims (15)
- 一种实体语义关系分类方法,应用于一电子设备,所述电子设备中预设有一实体语义关系分类模型,所述方法包括:确定出语料中的第一实体与第二实体;获得所述语料中每个文字各自与所述第一实体的第一位置距离以及与所述第二实体的第二位置距离;将所述语料中所有文字各自对应的特征向量进行组合,得到所述语料对应的模型输入向量,其中,每个文字对应的特征向量由所述语料中每个文字对应的字向量与位置嵌入向量进行组合后获得,每个文字对应的位置嵌入向量包括每个文字的第一位置距离对应的向量、每个文字的第二位置距离对应的向量;将所述语料对应的模型输入向量作为所述实体语义关系分类模型的输入,确定出所述第一实体与所述第二实体两者的实体语义关系类型。
- 如权利要求1所述的方法,将所述语料中所有文字各自对应的特征向量进行组合,得到所述语料对应的模型输入向量的步骤,包括:获得所述语料中每个文字对应的字向量,以及获得每个文字的第一位置距离对应的第一位置嵌入向量,每个文字的第二位置距离对应的第二位置嵌入向量;将所述语料中每个文字对应的字向量、第一位置嵌入向量和第二位置嵌入向量进行组合,获得每个文字对应的特征向量;将所述语料中所有文字各自对应的特征向量进行组合,获得所述语料对应的模型输入向量。
- 如权利要求2所述的方法,获得每个文字的第一位置距离对应的第一位置嵌入向量,每个文字的第二位置距离对应的第二位置嵌入向量的步骤,包括:获得位置嵌入向量表,其中,所述位置嵌入向量表记录有位置距离与位置嵌入向量的对应关系;在所述位置嵌入向量表中,分别确定出每个文字的第一位置距离对应的第一位置嵌入向量,以及每个文字的第二位置距离对应的第二位置嵌入向量。
- 一种实体语义关系分类模型训练方法,应用于一电子设备,所述电子设备中预设有一实体语义关系分类模型,所述方法包括:接收至少一个训练样本,识别所述至少一个训练样本中每个训练样本的第一实体和第二实体;针对所述每个训练样本,获得该训练样本中每个文字各自与该训练样本的第一实体的第一位置距离,以及获得该训练样本中每个文字各自与该训练样本的第二实体的第二位置距离;将所述每个训练样本中所有文字各自对应的特征向量进行组合,得到所述每个训练样本各自对应的模型输入向量,其中,每个文字对应的特征向量由每个文字对应的字向量与位置嵌入向量进行组合后获得,每个文字对应的位置嵌入向量包括每个文字的第一位置距离对应的向量、每个文字的第二位置距离对应的向量;将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,以对所述实体语义关系分类模型进行训练。
- 如权利要求4所述的方法,将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,以对所述实体语义关系分类模型进行训练的步骤,包括:将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,获得通过所述实体语义关系分类模型得到的所述每个训练样本对应的预测实体语义关系类型,所述预测实体语义关系类型为预测的所述每个训练样本中第一实体与第二实体两者的实体语义关系类型;获得所述每个训练样本对应的预测实体语义关系类型和预设实体语义关系类型的偏差值,所述预设实体语义关系类型为所述每个训练样本对应预先存储的第一实体和第二实体两者的实体语义关系类型;获得所述每个训练样本的偏差值之和;当所述偏差值之和超过第一偏差阈值,则调整所述实体语义关系分类模型中的参数,以训练所述实体语义关系分类模型。
- 如权利要求4所述的方法,将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,以对所述实体语义关系分类模 型进行训练的步骤,包括:将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,获得通过所述实体语义关系分类模型得到的所述每个训练样本对应的预测实体语义关系类型,所述预测实体语义关系类型为预测的所述每个训练样本中第一实体与第二实体两者的实体语义关系类型;获得所述每个训练样本对应的预测实体语义关系类型和预设实体语义关系类型的偏差值,所述预设实体语义关系类型为所述每个训练样本对应预先存储的第一实体和第二实体两者的实体语义关系类型;每当所述至少一个训练样本中目标训练样本的偏差值超过第二偏差阈值时,则调整所述实体语义关系分类模型中的参数,以训练所述实体语义关系分类模型。
- 如权利要求5所述的方法,所述实体语义关系分类模型为双向门控循环神经网络BiGRU结合注意力Attention机制的模型,调整所述实体语义关系分类模型中的参数的步骤,包括:调整所述实体语义关系分类模型中门控循环神经网络GRU层的权重系数及偏置系数,以及注意力层的注意力矩阵。
- 如权利要求4所述的方法,所述实体语义关系分类模型为双向门控循环神经网络BiGRU结合注意力Attention机制的模型;所述至少一个训练样本为至少一个电子病历,所述训练样本对应的模型输入向量为n个特征向量的组合,其中,所述n为所述至少一个电子病历中包含的平均字数。
- 一种电子设备,包括:存储器,用于存储一个或多个程序和预设的一实体语义关系分类模型;处理器;当所述一个或多个程序被所述处理器执行时,实现:确定出语料中的第一实体与第二实体;获得所述语料中每个文字各自与所述第一实体的第一位置距离以及与所述第二实体的第二位置距离;将所述语料中所有文字各自对应的特征向量进行组合,得到所述语料对 应的模型输入向量,其中,每个文字对应的特征向量由所述语料中每个文字对应的字向量与位置嵌入向量进行组合后获得,每个文字对应的位置嵌入向量包括每个文字的第一位置距离对应的向量、每个文字的第二位置距离对应的向量;将所述语料对应的模型输入向量作为所述实体语义关系分类模型的输入,确定出所述第一实体与所述第二实体两者的实体语义关系类型。
- 如权利要求9所述的电子设备,当所述一个或多个程序被所述处理器执行时,具体实现:获得所述语料中每个文字对应的字向量,以及获得每个文字的第一位置距离对应的第一位置嵌入向量,每个文字的第二位置距离对应的第二位置嵌入向量;将所述语料中每个文字对应的字向量、第一位置嵌入向量和第二位置嵌入向量进行组合,获得每个文字对应的特征向量;将所述语料中所有文字各自对应的特征向量进行组合,获得所述语料对应的模型输入向量。
- 如权利要求10所述的电子设备,当所述一个或多个程序被所述处理器执行时,具体实现:获得位置嵌入向量表,其中,所述位置嵌入向量表记录有位置距离与位置嵌入向量的对应关系;在所述位置嵌入向量表中,分别确定出每个文字的第一位置距离对应的第一位置嵌入向量,以及每个文字的第二位置距离对应的第二位置嵌入向量。
- 一种电子设备,包括:存储器,用于存储一个或多个程序和预设的一实体语义关系分类模型;处理器;当所述一个或多个程序被所述处理器执行时,实现:接收至少一个训练样本,识别所述至少一个训练样本中每个训练样本的第一实体和第二实体;针对所述每个训练样本,获得该训练样本中每个文字各自与该训练样本的第一实体的第一位置距离,以及获得该训练样本中每个文字各自与该训练 样本的第二实体的第二位置距离;将所述每个训练样本中所有文字各自对应的特征向量进行组合,得到所述每个训练样本各自对应的模型输入向量,其中,每个文字对应的特征向量由每个文字对应的字向量与位置嵌入向量进行组合后获得,每个文字对应的位置嵌入向量包括每个文字的第一位置距离对应的向量、每个文字的第二位置距离对应的向量;将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,以对所述实体语义关系分类模型进行训练。
- 如权利要求12所述的电子设备,当所述一个或多个程序被所述处理器执行时,具体实现:将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,获得通过所述实体语义关系分类模型得到的所述每个训练样本对应的预测实体语义关系类型,所述预测实体语义关系类型为预测的所述每个训练样本中第一实体与第二实体两者的实体语义关系类型;获得所述每个训练样本对应的预测实体语义关系类型和预设实体语义关系类型的偏差值,所述预设实体语义关系类型为所述每个训练样本对应预先存储的第一实体和第二实体两者的实体语义关系类型;获得所述每个训练样本的偏差值之和;当所述偏差值之和超过第一偏差阈值,则调整所述实体语义关系分类模型中的参数,以训练所述实体语义关系分类模型。
- 如权利要求12所述的电子设备,当所述一个或多个程序被所述处理器执行时,具体实现:将所述每个训练样本各自对应的模型输入向量作为所述实体语义关系分类模型的输入,获得通过所述实体语义关系分类模型得到的所述每个训练样本对应的预测实体语义关系类型,所述预测实体语义关系类型为预测的所述每个训练样本中第一实体与第二实体两者的实体语义关系类型;获得所述每个训练样本对应的预测实体语义关系类型和预设实体语义关系类型的偏差值,所述预设实体语义关系类型为所述每个训练样本对应预先存储的第一实体和第二实体两者的实体语义关系类型;每当所述至少一个训练样本中目标训练样本的偏差值超过第二偏差阈值时,则调整所述实体语义关系分类模型中的参数,以训练所述实体语义关系分类模型。
- 如权利要求12所述的电子设备,所述实体语义关系分类模型为双向门控循环神经网络BiGRU结合注意力Attention机制的模型;所述至少一个训练样本为至少一个电子病历,所述训练样本对应的模型输入向量为n个特征向量的组合,其中,所述n为所述至少一个电子病历中包含的平均字数。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021534922A JP7202465B2 (ja) | 2018-12-29 | 2019-12-23 | エンティティ意味関係分類 |
US17/287,476 US20210391080A1 (en) | 2018-12-29 | 2019-12-23 | Entity Semantic Relation Classification |
EP19901450.7A EP3985559A4 (en) | 2018-12-29 | 2019-12-23 | CLASSIFICATION OF SEMANTIC ENTITY RELATIONSHIPS |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811641958.3A CN109754012A (zh) | 2018-12-29 | 2018-12-29 | 实体语义关系分类方法、模型训练方法、装置及电子设备 |
CN201811641958.3 | 2018-12-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020135337A1 true WO2020135337A1 (zh) | 2020-07-02 |
Family
ID=66404511
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/127449 WO2020135337A1 (zh) | 2018-12-29 | 2019-12-23 | 实体语义关系分类 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20210391080A1 (zh) |
EP (1) | EP3985559A4 (zh) |
JP (1) | JP7202465B2 (zh) |
CN (1) | CN109754012A (zh) |
WO (1) | WO2020135337A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113326698A (zh) * | 2021-06-18 | 2021-08-31 | 深圳前海微众银行股份有限公司 | 检测实体关系的方法、模型训练方法及电子设备 |
CN115270801A (zh) * | 2022-09-28 | 2022-11-01 | 浙江太美医疗科技股份有限公司 | 文本信息抽取模型的训练方法、文本信息抽取方法和应用 |
CN117057343A (zh) * | 2023-10-10 | 2023-11-14 | 腾讯科技(深圳)有限公司 | 道路事件识别方法、装置、设备及存储介质 |
Families Citing this family (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109754012A (zh) * | 2018-12-29 | 2019-05-14 | 新华三大数据技术有限公司 | 实体语义关系分类方法、模型训练方法、装置及电子设备 |
CN110111901B (zh) * | 2019-05-16 | 2023-04-18 | 湖南大学 | 基于rnn神经网络的可迁移病人分类系统 |
CN110275928B (zh) * | 2019-06-24 | 2022-11-22 | 复旦大学 | 迭代式实体关系抽取方法 |
CN110442725B (zh) * | 2019-08-14 | 2022-02-25 | 科大讯飞股份有限公司 | 实体关系抽取方法及装置 |
CN110928889A (zh) * | 2019-10-23 | 2020-03-27 | 深圳市华讯方舟太赫兹科技有限公司 | 训练模型更新方法、设备以及计算机存储介质 |
CN112989032A (zh) * | 2019-12-17 | 2021-06-18 | 医渡云(北京)技术有限公司 | 实体关系分类方法、装置、介质及电子设备 |
US11893060B2 (en) * | 2020-02-06 | 2024-02-06 | Naver Corporation | Latent question reformulation and information accumulation for multi-hop machine reading |
CN111429985B (zh) * | 2020-03-02 | 2023-10-27 | 北京嘉和海森健康科技有限公司 | 电子病历数据处理方法及系统 |
CN113591886B (zh) * | 2020-04-30 | 2023-11-07 | 北京百度网讯科技有限公司 | 用于信息分类的方法、装置、设备及计算机可读存储介质 |
CN112199508B (zh) * | 2020-08-10 | 2024-01-19 | 淮阴工学院 | 一种基于远程监督的参数自适应农业知识图谱推荐方法 |
CN114254106A (zh) * | 2020-09-25 | 2022-03-29 | 北京灵汐科技有限公司 | 文本分类方法、装置、设备及存储介质 |
CN112416931A (zh) * | 2020-11-18 | 2021-02-26 | 脸萌有限公司 | 信息生成方法、装置和电子设备 |
CN112786132B (zh) * | 2020-12-31 | 2023-03-24 | 北京懿医云科技有限公司 | 病历文本数据分割方法、装置、可读存储介质及电子设备 |
CN113408588B (zh) * | 2021-05-24 | 2023-02-14 | 上海电力大学 | 一种基于注意力机制的双向gru轨迹预测方法 |
CN114398943B (zh) * | 2021-12-09 | 2023-04-07 | 北京百度网讯科技有限公司 | 样本增强方法及其装置 |
CN114881038B (zh) * | 2022-07-12 | 2022-11-11 | 之江实验室 | 基于跨度和注意力机制的中文实体与关系抽取方法及装置 |
CN116151243B (zh) * | 2023-04-23 | 2023-06-23 | 昆明理工大学 | 一种基于类型相关性表征的实体关系抽取方法 |
CN116665228B (zh) * | 2023-07-31 | 2023-10-13 | 恒生电子股份有限公司 | 图像处理方法及装置 |
CN117496542B (zh) * | 2023-12-29 | 2024-03-15 | 恒生电子股份有限公司 | 文档信息提取方法、装置、电子设备和存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106407211A (zh) * | 2015-07-30 | 2017-02-15 | 富士通株式会社 | 对实体词的语义关系进行分类的方法和装置 |
CN107305543A (zh) * | 2016-04-22 | 2017-10-31 | 富士通株式会社 | 对实体词的语义关系进行分类的方法和装置 |
CN108268643A (zh) * | 2018-01-22 | 2018-07-10 | 北京邮电大学 | 一种基于多粒度lstm网络的深层语义匹配实体链接方法 |
CN109754012A (zh) * | 2018-12-29 | 2019-05-14 | 新华三大数据技术有限公司 | 实体语义关系分类方法、模型训练方法、装置及电子设备 |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102167719B1 (ko) * | 2014-12-08 | 2020-10-19 | 삼성전자주식회사 | 언어 모델 학습 방법 및 장치, 음성 인식 방법 및 장치 |
CN107220237A (zh) * | 2017-05-24 | 2017-09-29 | 南京大学 | 一种基于卷积神经网络的企业实体关系抽取的方法 |
US11328006B2 (en) * | 2017-10-26 | 2022-05-10 | Mitsubishi Electric Corporation | Word semantic relation estimation device and word semantic relation estimation method |
CN108875809A (zh) * | 2018-06-01 | 2018-11-23 | 大连理工大学 | 联合attention机制与神经网络的生物医学实体关系分类方法 |
CN108763555A (zh) * | 2018-06-01 | 2018-11-06 | 北京奇虎科技有限公司 | 基于需求词的画像数据获取方法及装置 |
-
2018
- 2018-12-29 CN CN201811641958.3A patent/CN109754012A/zh active Pending
-
2019
- 2019-12-23 WO PCT/CN2019/127449 patent/WO2020135337A1/zh active Application Filing
- 2019-12-23 US US17/287,476 patent/US20210391080A1/en active Pending
- 2019-12-23 EP EP19901450.7A patent/EP3985559A4/en active Pending
- 2019-12-23 JP JP2021534922A patent/JP7202465B2/ja active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106407211A (zh) * | 2015-07-30 | 2017-02-15 | 富士通株式会社 | 对实体词的语义关系进行分类的方法和装置 |
CN107305543A (zh) * | 2016-04-22 | 2017-10-31 | 富士通株式会社 | 对实体词的语义关系进行分类的方法和装置 |
CN108268643A (zh) * | 2018-01-22 | 2018-07-10 | 北京邮电大学 | 一种基于多粒度lstm网络的深层语义匹配实体链接方法 |
CN109754012A (zh) * | 2018-12-29 | 2019-05-14 | 新华三大数据技术有限公司 | 实体语义关系分类方法、模型训练方法、装置及电子设备 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3985559A4 |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113326698A (zh) * | 2021-06-18 | 2021-08-31 | 深圳前海微众银行股份有限公司 | 检测实体关系的方法、模型训练方法及电子设备 |
CN115270801A (zh) * | 2022-09-28 | 2022-11-01 | 浙江太美医疗科技股份有限公司 | 文本信息抽取模型的训练方法、文本信息抽取方法和应用 |
CN115270801B (zh) * | 2022-09-28 | 2022-12-30 | 浙江太美医疗科技股份有限公司 | 文本信息抽取模型的训练方法、文本信息抽取方法和应用 |
CN117057343A (zh) * | 2023-10-10 | 2023-11-14 | 腾讯科技(深圳)有限公司 | 道路事件识别方法、装置、设备及存储介质 |
CN117057343B (zh) * | 2023-10-10 | 2023-12-12 | 腾讯科技(深圳)有限公司 | 道路事件识别方法、装置、设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
US20210391080A1 (en) | 2021-12-16 |
CN109754012A (zh) | 2019-05-14 |
JP2022514842A (ja) | 2022-02-16 |
EP3985559A4 (en) | 2022-10-05 |
JP7202465B2 (ja) | 2023-01-11 |
EP3985559A1 (en) | 2022-04-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020135337A1 (zh) | 实体语义关系分类 | |
US11227118B2 (en) | Methods, devices, and systems for constructing intelligent knowledge base | |
CN108597519B (zh) | 一种话单分类方法、装置、服务器和存储介质 | |
US10824653B2 (en) | Method and system for extracting information from graphs | |
WO2020147428A1 (zh) | 交互内容生成方法、装置、计算机设备及存储介质 | |
WO2020232877A1 (zh) | 一种问题答案选取方法、装置、计算机设备及存储介质 | |
CN111602128A (zh) | 计算机实现的确定方法和系统 | |
US20220083605A1 (en) | Predictive system for generating clinical queries | |
CN110866124B (zh) | 基于多数据源的医学知识图谱融合方法及装置 | |
US20220075958A1 (en) | Missing semantics complementing method and apparatus | |
CN112257449B (zh) | 命名实体识别方法、装置、计算机设备和存储介质 | |
JP2022524830A (ja) | 機械学習アプリケーションにおけるカテゴリフィールド値の取り扱い | |
KR102298330B1 (ko) | 음성인식과 자연어 처리 알고리즘을 통해 의료 상담 요약문과 전자 의무 기록을 생성하는 시스템 | |
WO2021120779A1 (zh) | 一种基于人机对话的用户画像构建方法、系统、终端及存储介质 | |
WO2021017300A1 (zh) | 问题生成方法、装置、计算机设备及存储介质 | |
CN112418059A (zh) | 一种情绪识别的方法、装置、计算机设备及存储介质 | |
WO2021051877A1 (zh) | 人工智能面试中获取输入文本和相关装置 | |
CN113094478A (zh) | 表情回复方法、装置、设备及存储介质 | |
US10902215B1 (en) | Social hash for language models | |
CN108009157B (zh) | 一种语句归类方法及装置 | |
WO2021063089A1 (zh) | 规则匹配方法、规则匹配装置、存储介质及电子设备 | |
CN115358817A (zh) | 基于社交数据的智能产品推荐方法、装置、设备及介质 | |
CN111506764B (zh) | 音频数据筛选方法、计算机设备和存储介质 | |
WO2021082570A1 (zh) | 基于人工智能的语义识别方法、装置和语义识别设备 | |
CN113761126A (zh) | 文本内容的识别方法、装置、设备及可读存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19901450 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021534922 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2019901450 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2019901450 Country of ref document: EP Effective date: 20210729 |