WO2021072852A1 - Procédé et système de marquage de séquences, et dispositif informatique - Google Patents

Procédé et système de marquage de séquences, et dispositif informatique Download PDF

Info

Publication number
WO2021072852A1
WO2021072852A1 PCT/CN2019/117403 CN2019117403W WO2021072852A1 WO 2021072852 A1 WO2021072852 A1 WO 2021072852A1 CN 2019117403 W CN2019117403 W CN 2019117403W WO 2021072852 A1 WO2021072852 A1 WO 2021072852A1
Authority
WO
WIPO (PCT)
Prior art keywords
sequence
word
target text
vector
labeling
Prior art date
Application number
PCT/CN2019/117403
Other languages
English (en)
Chinese (zh)
Inventor
金戈
徐亮
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021072852A1 publication Critical patent/WO2021072852A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Definitions

  • the embodiments of the present application relate to the field of sequence labeling, and in particular, to a sequence labeling method, system, computer equipment, and non-volatile computer-readable storage medium.
  • named entity recognition is the most basic and most widely used one. It refers to identifying entities with specific meanings in the text, including names of persons, places, organizations, proper nouns, etc.
  • the named entity recognition application is an important basic tool for other applications such as information extraction, question answering system, syntactic analysis, machine translation, semantic web-oriented metadata annotation and other application fields.
  • a natural language model can be constructed, which can understand, analyze and answer the results of natural language like humans.
  • existing models often fail to consider long-term contextual information, which results in a technical problem that limits the accuracy of recognition.
  • the embodiment of the present application provides a sequence labeling method, and the method steps include:
  • the first tag sequence includes a plurality of first n-dimensional vectors, each first n-dimensional vector corresponds to a word in the target text sequence, and the first n-dimensional vector represents the corresponding word belonging to n first The first probability of each first tag in the tags;
  • the first annotation sequence is input to the fully connected layer, and the second annotation sequence is output through the fully connected layer, wherein the second annotation sequence includes a plurality of second n-dimensional vectors, and each second n-dimensional vector corresponds to For a word in the target text sequence, the second n-dimensional vector represents the second probability of the corresponding word belonging to each of the n second tags;
  • a named entity sequence is generated according to the tag sequence, and the named entity sequence is output.
  • an embodiment of the present application also provides a sequence labeling system, including:
  • a receiving text module for receiving a target text sequence, and converting the target text sequence into a corresponding sentence vector, a word vector of each word, and a position vector of each word;
  • the first labeling module is used to input the sentence vector of the target text sequence, the word vector of each word and the position vector of each word into the trained BERT model, and the output of the BERT model corresponds to the target text sequence
  • the first tagging sequence wherein the first tagging sequence includes a plurality of first n-dimensional vectors, each first n-dimensional vector corresponds to a word in the target text sequence, and the first n-dimensional vector represents the corresponding The first probability of the word belonging to each of the n first tags;
  • the second labeling module is configured to input the first labeling sequence to the fully connected layer, and output a second labeling sequence through the fully connected layer, wherein the second labeling sequence includes a plurality of second n-dimensional vectors, each A second n-dimensional vector corresponds to a word in the target text sequence, and the second n-dimensional vector represents a second probability of the corresponding word belonging to each of the n second tags;
  • the output entity module is used to generate a named entity sequence according to the tag sequence, and output the named entity sequence.
  • an embodiment of the present application further provides a computer device, the computer device including a memory, a processor, and computer-readable instructions stored in the memory and running on the processor, the When the computer-readable instructions are executed by the processor, the following steps are implemented:
  • the first tag sequence includes a plurality of first n-dimensional vectors, each first n-dimensional vector corresponds to a word in the target text sequence, and the first n-dimensional vector represents the corresponding word belonging to n first The first probability of each first tag in the tags;
  • the first annotation sequence is input to the fully connected layer, and the second annotation sequence is output through the fully connected layer, wherein the second annotation sequence includes a plurality of second n-dimensional vectors, and each second n-dimensional vector corresponds to For a word in the target text sequence, the second n-dimensional vector represents the second probability of the corresponding word belonging to each of the n second tags;
  • a named entity sequence is generated according to the tag sequence, and the named entity sequence is output.
  • the embodiments of the present application also provide a non-volatile computer-readable storage medium, the non-volatile computer-readable storage medium stores computer-readable instructions, and the computer-readable instructions may Is executed by at least one processor, so that the at least one processor executes the following steps:
  • the first tag sequence includes a plurality of first n-dimensional vectors, each first n-dimensional vector corresponds to a word in the target text sequence, and the first n-dimensional vector represents the corresponding word belonging to n first The first probability of each first tag in the tags;
  • the first annotation sequence is input to the fully connected layer, and the second annotation sequence is output through the fully connected layer, wherein the second annotation sequence includes a plurality of second n-dimensional vectors, and each second n-dimensional vector corresponds to For a word in the target text sequence, the second n-dimensional vector represents the second probability of the corresponding word belonging to each of the n second tags;
  • a named entity sequence is generated according to the tag sequence, and the named entity sequence is output.
  • the sequence labeling method, system, computer equipment, and non-volatile computer-readable storage medium provided by the embodiments of the application provide an effective sequence labeling method for text sequences; the embodiments of the application solve the problem that the model in the prior art cannot consider the long-term Contextual information, which limits the accuracy of recognition technology. It is possible to extract the named entities in the sentence by directly inputting the original sentence into the model. It has strong adaptability and wide application, which improves the accuracy of sequence labeling for entity recognition. The technical effect of the rate.
  • FIG. 1 is a schematic flowchart of a sequence labeling method according to an embodiment of the application.
  • Fig. 2 is a schematic diagram of program modules of the second embodiment of the sequence labeling system of this application.
  • FIG. 3 is a schematic diagram of the hardware structure of the third embodiment of the computer equipment of this application.
  • the computer device 2 will be used as an execution subject for exemplary description.
  • FIG. 1 shows a flowchart of the steps of a sequence labeling method according to an embodiment of the present application. It can be understood that the flowchart in this method embodiment is not used to limit the order of execution of the steps.
  • the following is an exemplary description with the computer device 2 as the execution subject. details as follows.
  • Step S100 Receive a target text sequence, and convert the target text sequence into a corresponding sentence vector, a word vector of each word, and a position vector of each word.
  • the step S100 may further include:
  • Step S100a input the target text sequence to an embedding layer, and output a plurality of word vectors corresponding to the target text sequence through the embedding layer, and the plurality of word vectors includes at least one punctuation vector.
  • Step S100b Input the plurality of word vectors into the segmentation layer, and divide the plurality of word vectors according to the at least one punctuation vector to obtain n word vector sets, and the n word vector sets correspond to n divisions code.
  • the target text sequence is [Curie was born in Poland, lives in the United States] is divided into A sentence [Curie is born in Poland] and B sentence [residence in the United States], the first half of the sentence will be added with the segmentation code A, and then The segmentation code B will be added to the half sentence.
  • step S100c an encoding operation is performed on each segmentation code by position encoding, and the position information encoding of each segmentation code is determined, so as to obtain the position vector of each word in the target text sequence.
  • the position information encoding may be used to determine the position of each word in the target text sequence.
  • Step S100d generating a sentence vector of the target text sequence according to the word vector of each word in the target text sequence and the position vector of each word.
  • Step S102 Input the sentence vector of the target text sequence, the word vector of each word and the position vector of each word into the trained BERT model, and output the first annotation corresponding to the target text sequence through the BERT model Sequence, wherein the first annotation sequence includes a plurality of first n-dimensional vectors, each of the first n-dimensional vectors corresponds to a word in the target text sequence, and the first n-dimensional vector indicates that the corresponding word belongs to n The first probability of each first tag in the first tags.
  • the n first tags may be multiple location tags and multiple semantic tags, and the n first tags may also be multiple location tags and multiple part-of-speech tags.
  • BERT is an existing pre-training model.
  • the full name of BERT is Bidirectional Encoder Representations from Transformers, that is, a two-way Transformer encoder (Encoder); wherein, the Transformer is a type that completely relies on self-attention.
  • Encoder a two-way Transformer encoder
  • BERT aims to pre-train deep bidirectional representations by jointly adjusting the context in all layers. Therefore, the pre-trained BERT can be fine-tuned through an additional output layer, which is suitable for the construction of state-of-the-art models for a wide range of tasks, such as question answering tasks and language inference, without requiring major architectural modifications for specific tasks.
  • the BERT model can be obtained by capturing words through a masked language model (MLM) method and expressing the sentence level through a "Next Sentence Prediction” method; wherein the masked language model randomly masks the model input Some of the words in the (token), the goal is to predict the original vocabulary id based only on the context of the masked word.
  • MLM masked language model
  • the training target of the masked language model allows the representation and fusion of the left and right sides of the language Context, so as to pre-train a deep two-way Transformer;
  • Next Sentence Prediction that is, Next Sentence Prediction refers to the selection of two sentences in two situations when pre-training the language model.
  • training the pre-trained BERT model may include: acquiring multiple training text sequences, using the multiple training text sequences as a training set of the BERT model, and inputting the training set to the pre-trained BERT In the model, the pre-trained BERT model is trained through the training set to obtain a trained BERT model.
  • the step S102 may further include:
  • Step S102a Perform feature extraction on the sentence vector of the target text sequence, the word vector of each word, and the position vector of each word through the BERT model to obtain the first label of each word in the target text sequence. First probability.
  • Step S102b Generate a first labeling sequence according to the first probability of each first tag of each word in the target text sequence.
  • Step S104 Input the first labeling sequence to the fully connected layer, and output a second labeling sequence through the fully connected layer, where the second labeling sequence includes a plurality of second n-dimensional vectors, and each second n-dimensional vector The dimensional vector corresponds to a word in the target text sequence, and the second n-dimensional vector represents the second probability of the corresponding word belonging to each of the n second tags.
  • the n second tags may be multiple location tags and multiple semantic tags, and the n second tags may also be multiple location tags and multiple part-of-speech tags.
  • the step S104 may further include:
  • Step S104a Input the first label sequence into the neural network structure of the fully connected layer, and perform additional feature extraction to obtain the second probability of each label of each word in the target text sequence.
  • additional features of the i-th word sequence extracted calculation formula B i wX i + b, where, X i is the probability of each of the first tag of the first sequence of the first tagging word i, w and b is the learning parameter of the BERT model;
  • the neural network structure of the fully connected layer of this embodiment may be a multi-layer transformer structure.
  • the multi-layer transformer structure further includes an attention mechanism. After the first annotation sequence is processed by the attention mechanism, the input To the feedforward fully connected neural network structure for additional feature extraction, to obtain the second probability of each second label of each word in the target text sequence; that is, to obtain each of the target text sequences through the operation of wx+b The second probability of each second label of the word, where x is the sequence, and w and b are the model learning parameters.
  • Step S104b Generate a second tagging sequence according to the second probability of each second tag of each word in the target text sequence.
  • the step S106 may further include:
  • Step S106a input the second annotation sequence into the CRF model
  • Step S106b Viterbi solving the second annotation sequence by the Viterbi algorithm to obtain an optimal solution path in the second annotation sequence, where the optimal solution path is that the label sequence is an integer.
  • this step is to determine the output object that the target text sequence should correspond to according to the probability value of the second probability of each second label of each word in the target text sequence; here it is implemented by the Viterbi algorithm, so The Viterbi algorithm does not output the highest label probability among the second probabilities of each second label of each word in the target text sequence, but outputs the highest probability label sequence of the entire target text sequence.
  • the Viterbi algorithm may include: when the path with the second highest probability of each second tag of each word in the target text sequence passes through a certain point of the fence network, then from the starting point to the The sub-path of a point must also be the path with the greatest probability from the beginning to the point; when there are k states at the i-th moment, there are k shortest paths from the beginning to the k states at the i time, and the final shortest path must pass One of them.
  • Step S106c generating a label sequence according to the optimal solution path.
  • the highest probability labeling sequence of the entire target text sequence is calculated by the Viterbi algorithm.
  • the shortest path of the i+1th state only the k state values from the beginning to the current The shortest path and the shortest path from the current state value to the i+1th state value are sufficient.
  • Step S108 Generate a named entity sequence according to the tag sequence, and output the named entity sequence.
  • a named entity sequence can be generated according to the tag sequence, and the named entity sequence is a target text sequence predicted by the tagging system.
  • the named entity includes place name, person name, etc.; sequence labeling adopts the form of BIOES, where B is the beginning of the entity, I is the middle of the entity, O is the non-entity, E is the end of the entity, and S is the single word entity; and each named entity The label corresponds to the entity category, which can be refined into similar forms such as B-place name: the beginning of the place name entity.
  • B-place name the beginning of the place name entity.
  • the sentence "Curie was born in Warsaw" as an example, this sentence will be split into a sequence of words.
  • the home character is marked as B name
  • the inner character is marked as E name
  • the new character is marked as O
  • the word Yu is marked as O
  • the wave character is marked as B place name
  • the blue character is marked as E- place name.
  • Fig. 2 is a schematic diagram of program modules of the second embodiment of the sequence labeling system of this application.
  • the sequence labeling system 20 may include or be divided into one or more program modules.
  • One or more program modules are stored in a storage medium and executed by one or more processors to complete the application and realize the above Sequence labeling method.
  • the program module referred to in the embodiments of the present application refers to a series of computer-readable instruction segments that can complete specific functions. The following description will specifically introduce the functions of each program module in this embodiment:
  • the receiving text module 200 is configured to receive a target text sequence, and convert the target text sequence into a corresponding sentence vector, a word vector of each word, and a position vector of each word.
  • the receiving text module 200 is further configured to: input the target text sequence into an embedding layer, and output multiple word vectors corresponding to the target text sequence through the embedding layer, the multiple word vectors Including at least one punctuation vector; input the plurality of word vectors to the segmentation layer, and divide the plurality of word vectors according to the at least one punctuation vector to obtain n word vector sets, the n word vector sets Corresponding to n segmentation codes; perform an encoding operation on each segmentation code by position encoding, and determine the position information encoding of each segmentation code to obtain the position vector of each word in the target text sequence; and according to the target text sequence The word vector of each word in and the position vector of each word in the, generate the sentence vector of the target text sequence.
  • the first labeling module 202 is configured to input the sentence vector of the target text sequence, the word vector of each word, and the position vector of each word into the trained BERT model, and the output of the BERT model is related to the target text sequence.
  • the corresponding first annotation sequence wherein the first annotation sequence includes a plurality of first n-dimensional vectors, each first n-dimensional vector corresponds to a word in the target text sequence, and the first n-dimensional vector represents The first probability that the corresponding word belongs to each of the n first tags.
  • the first tagging module 202 is further configured to: use the BERT model to perform feature extraction on the sentence vector of the target text sequence, the word vector of each word, and the position vector of each word to obtain the target The first probability of each first label of each word in the text sequence; and the first labeling sequence is generated according to the first probability of each first label of each word in the target text sequence.
  • the second labeling module 204 is configured to input the first labeling sequence to the fully connected layer, and output a second labeling sequence through the fully connected layer, wherein the second labeling sequence includes a plurality of second n-dimensional vectors, Each second n-dimensional vector corresponds to a word in the target text sequence, and the second n-dimensional vector represents the second probability of the corresponding word belonging to each of the n second tags.
  • the second labeling module 204 is further configured to: input the first labeling sequence into the neural network structure of the fully connected layer, and perform additional feature extraction to obtain the information of each word in the target text sequence.
  • the first probability of each first label of each word, w and b are BERT model learning parameters; according to the second probability of each second label of each word in the target text sequence, a second labeling sequence is generated.
  • the output tag module 206 is further configured to: input the second annotation sequence into the CRF model; perform Viterbi solution on the second annotation sequence by the Viterbi algorithm to obtain the second annotation
  • the optimal solution path in the sequence wherein the optimal solution path is the highest probability sequence in which the label sequence is the entire target text sequence; the label sequence is generated according to the optimal solution path.
  • the output entity module 208 is configured to generate a named entity sequence according to the tag sequence, and output the named entity sequence.
  • the computer device 2 is a device that can automatically perform numerical calculation and/or information processing in accordance with pre-set or stored instructions.
  • the computer device 2 may be a rack server, a blade server, a tower server, or a cabinet server (including an independent server or a server cluster composed of multiple servers).
  • the computer device 2 at least includes, but is not limited to, a memory 21, a processor 22, a network interface 23, and a sequence labeling system 20 that can communicate with each other through a system bus.
  • the memory 21 includes at least one type of non-volatile computer-readable storage medium, and the readable storage medium includes flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), Random access memory (RAM), static random access memory (SRAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), programmable read only memory (PROM), magnetic memory, magnetic disk, optical disk Wait.
  • the memory 21 may be an internal storage unit of the computer device 2, for example, a hard disk or a memory of the computer device 2.
  • the memory 21 may also be an external storage device of the computer device 2, such as a plug-in hard disk, a smart media card (SMC), and a secure digital (Secure Digital, SD) card, flash card (Flash Card), etc.
  • the memory 21 may also include both the internal storage unit of the computer device 2 and its external storage device.
  • the memory 21 is generally used to store an operating system and various application software installed in the computer device 2, for example, the program code of the sequence labeling system 20 in the second embodiment.
  • the memory 21 can also be used to temporarily store various types of data that have been output or will be output.
  • the processor 22 may be a central processing unit (Central Processing Unit, CPU), a controller, a microcontroller, a microprocessor, or other data processing chips in some embodiments.
  • the processor 22 is generally used to control the overall operation of the computer device 2.
  • the processor 22 is used to run the program code or process data stored in the memory 21, for example, to run the sequence labeling system 20 to implement the sequence labeling method of the first embodiment.
  • the network interface 23 may include a wireless network interface or a wired network interface, and the network interface 23 is generally used to establish a communication connection between the computer device 2 and other electronic devices.
  • the network interface 23 is used to connect the computer device 2 to an external terminal through a network, and to establish a data transmission channel and a communication connection between the computer device 2 and the external terminal.
  • the network may be an intranet, the Internet, a global system of mobile communication (GSM), a wideband code division multiple access (WCDMA), a 4G network, and a 5G Network, Bluetooth (Bluetooth), Wi-Fi and other wireless or wired networks.
  • FIG. 3 only shows the computer device 2 with components 20-23, but it should be understood that it is not required to implement all the components shown, and more or fewer components may be implemented instead.
  • sequence labeling system 20 stored in the memory 21 may also be divided into one or more program modules, and the one or more program modules are stored in the memory 21 and processed by one or more The processor (in this embodiment, the processor 22) is executed to complete the application.
  • FIG. 2 shows a schematic diagram of program modules for implementing the sequence labeling system 20 according to the second embodiment of the present application.
  • the sequence labeling system 20 can be divided into a text receiving module 200 and a first labeling module 202. , The second labeling module 204, the output label module 206, and the output entity module 208.
  • the program module referred to in this application refers to a series of computer-readable instruction segments that can complete specific functions. The specific functions of the program modules 200-208 have been described in detail in the second embodiment, and will not be repeated here.
  • This embodiment also provides a non-volatile computer-readable storage medium, such as flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), random access memory (RAM), static random access memory ( SRAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), programmable read-only memory (PROM), magnetic memory, magnetic disks, optical disks, servers, App application malls, etc., on which storage There are computer-readable instructions, and the corresponding functions are realized when the program is executed by the processor.
  • the non-volatile computer-readable storage medium of this embodiment is used in the sequence labeling system 20, and the processor executes the following steps:
  • the first tag sequence includes a plurality of first n-dimensional vectors, each first n-dimensional vector corresponds to a word in the target text sequence, and the first n-dimensional vector represents the corresponding word belonging to n first The first probability of each first tag in the tags;
  • the first annotation sequence is input to the fully connected layer, and the second annotation sequence is output through the fully connected layer, where the second annotation sequence includes a plurality of second n-dimensional vectors, and each second n-dimensional vector corresponds to For a word in the target text sequence, the second n-dimensional vector represents the second probability of the corresponding word belonging to each of the n second tags;
  • a named entity sequence is generated according to the tag sequence, and the named entity sequence is output.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Machine Translation (AREA)

Abstract

L'invention concerne un procédé de marquage de séquences, comprenant: la réception d'une séquence textuelle cible, et la conversion de la séquence textuelle cible en un vecteur de phrase correspondant, un vecteur de mot de chaque mot et un vecteur de position de chaque mot (S100) ; la saisie du vecteur de phrase de la séquence textuelle cible, du vecteur de mot de chaque mot et du vecteur de position de chaque mot dans un modèle BERT entraîné, et l'émission en sortie, au moyen du modèle BERT, d'une première séquence de marquage correspondant à la séquence textuelle cible; la saisie de la première séquence de marquage dans une couche complètement connectée, et l'émission en sortie d'une seconde séquence de marquage au moyen de la couche complètement connectée; la prise de la seconde séquence de marquage en tant que séquence d'entrée d'un champ aléatoire conditionnel (CRF) modèle afin d'émettre en sortie une séquence d'étiquettes Y = (y1, y2,..., ym) au moyen du modèle de champ CRF (S106) ; et la génération d'une séquence d'entités nommées en fonction de la séquence d'étiquettes, et l'émission en sortie la séquence d'entités nommées (S108). Selon le procédé, le problème de l'incapacité des modèles existants à prendre en compte une relation d'information de contexte à long terme est résolu, et ainsi les effets techniques d'extraction directe d'une entité nommée dans un texte au moyen du modèle, et l'amélioration de la précision d'identification d'une entité sont réalisés.
PCT/CN2019/117403 2019-10-16 2019-11-12 Procédé et système de marquage de séquences, et dispositif informatique WO2021072852A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910983279.2A CN111222317B (zh) 2019-10-16 2019-10-16 序列标注方法、系统和计算机设备
CN201910983279.2 2019-10-16

Publications (1)

Publication Number Publication Date
WO2021072852A1 true WO2021072852A1 (fr) 2021-04-22

Family

ID=70827510

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/117403 WO2021072852A1 (fr) 2019-10-16 2019-11-12 Procédé et système de marquage de séquences, et dispositif informatique

Country Status (2)

Country Link
CN (1) CN111222317B (fr)
WO (1) WO2021072852A1 (fr)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113392659A (zh) * 2021-06-25 2021-09-14 携程旅游信息技术(上海)有限公司 机器翻译方法、装置、电子设备和存储介质
CN113516196A (zh) * 2021-07-20 2021-10-19 云知声智能科技股份有限公司 命名实体识别数据增强的方法、装置、电子设备和介质
CN113537346A (zh) * 2021-07-15 2021-10-22 思必驰科技股份有限公司 医疗领域数据标注模型训练方法及医疗领域数据标注方法
CN113569574A (zh) * 2021-07-16 2021-10-29 阳光电源股份有限公司 工单类型的识别方法、终端和存储介质
CN113626608A (zh) * 2021-10-12 2021-11-09 深圳前海环融联易信息科技服务有限公司 增强语义的关系抽取方法、装置、计算机设备及存储介质
CN113673247A (zh) * 2021-05-13 2021-11-19 江苏曼荼罗软件股份有限公司 基于深度学习的实体识别方法、装置、介质及电子设备
CN114048288A (zh) * 2021-11-10 2022-02-15 北京明略软件系统有限公司 细粒度情感分析方法、系统、计算机设备和存储介质
CN114386419A (zh) * 2022-01-11 2022-04-22 平安科技(深圳)有限公司 实体识别模型训练方法、装置、设备以及存储介质
CN114580424A (zh) * 2022-04-24 2022-06-03 之江实验室 一种用于法律文书的命名实体识别的标注方法和装置
CN115879473A (zh) * 2022-12-26 2023-03-31 淮阴工学院 基于改进图注意力网络的中文医疗命名实体识别方法

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111859858B (zh) * 2020-07-22 2024-03-01 智者四海(北京)技术有限公司 从文本中提取关系的方法及装置
CN112148856B (zh) * 2020-09-22 2024-01-23 北京百度网讯科技有限公司 建立标点预测模型的方法、装置
CN112541341A (zh) * 2020-12-18 2021-03-23 广东电网有限责任公司 一种文本事件元素提取方法
CN113064992A (zh) * 2021-03-22 2021-07-02 平安银行股份有限公司 投诉工单结构化处理方法、装置、设备及存储介质
CN113157883A (zh) * 2021-04-07 2021-07-23 浙江工贸职业技术学院 一种基于双模型结构的中文意见目标边界预测方法
CN113553824A (zh) * 2021-07-07 2021-10-26 临沂中科好孕智能技术有限公司 一种句子向量模型训练方法

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170206897A1 (en) * 2016-01-18 2017-07-20 Alibaba Group Holding Limited Analyzing textual data
CN109815952A (zh) * 2019-01-24 2019-05-28 珠海市筑巢科技有限公司 品牌名称识别方法、计算机装置及计算机可读存储介质
CN109994201A (zh) * 2019-03-18 2019-07-09 浙江大学 一种基于深度学习的糖尿病与高血压概率计算方法
CN110083831A (zh) * 2019-04-16 2019-08-02 武汉大学 一种基于BERT-BiGRU-CRF的中文命名实体识别方法
CN110147452A (zh) * 2019-05-17 2019-08-20 北京理工大学 一种基于层级bert神经网络的粗粒度情感分析方法
CN110297913A (zh) * 2019-06-12 2019-10-01 中电科大数据研究院有限公司 一种电子公文实体抽取方法

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB201614958D0 (en) * 2016-09-02 2016-10-19 Digital Genius Ltd Message text labelling
CN109635279B (zh) * 2018-11-22 2022-07-26 桂林电子科技大学 一种基于神经网络的中文命名实体识别方法
CN110223742A (zh) * 2019-06-14 2019-09-10 中南大学 中文电子病历数据的临床表现信息抽取方法和设备

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170206897A1 (en) * 2016-01-18 2017-07-20 Alibaba Group Holding Limited Analyzing textual data
CN109815952A (zh) * 2019-01-24 2019-05-28 珠海市筑巢科技有限公司 品牌名称识别方法、计算机装置及计算机可读存储介质
CN109994201A (zh) * 2019-03-18 2019-07-09 浙江大学 一种基于深度学习的糖尿病与高血压概率计算方法
CN110083831A (zh) * 2019-04-16 2019-08-02 武汉大学 一种基于BERT-BiGRU-CRF的中文命名实体识别方法
CN110147452A (zh) * 2019-05-17 2019-08-20 北京理工大学 一种基于层级bert神经网络的粗粒度情感分析方法
CN110297913A (zh) * 2019-06-12 2019-10-01 中电科大数据研究院有限公司 一种电子公文实体抽取方法

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113673247A (zh) * 2021-05-13 2021-11-19 江苏曼荼罗软件股份有限公司 基于深度学习的实体识别方法、装置、介质及电子设备
CN113392659A (zh) * 2021-06-25 2021-09-14 携程旅游信息技术(上海)有限公司 机器翻译方法、装置、电子设备和存储介质
CN113537346B (zh) * 2021-07-15 2023-08-15 思必驰科技股份有限公司 医疗领域数据标注模型训练方法及医疗领域数据标注方法
CN113537346A (zh) * 2021-07-15 2021-10-22 思必驰科技股份有限公司 医疗领域数据标注模型训练方法及医疗领域数据标注方法
CN113569574A (zh) * 2021-07-16 2021-10-29 阳光电源股份有限公司 工单类型的识别方法、终端和存储介质
CN113569574B (zh) * 2021-07-16 2024-02-09 阳光电源股份有限公司 工单类型的识别方法、终端和存储介质
CN113516196A (zh) * 2021-07-20 2021-10-19 云知声智能科技股份有限公司 命名实体识别数据增强的方法、装置、电子设备和介质
CN113516196B (zh) * 2021-07-20 2024-04-12 云知声智能科技股份有限公司 命名实体识别数据增强的方法、装置、电子设备和介质
CN113626608A (zh) * 2021-10-12 2021-11-09 深圳前海环融联易信息科技服务有限公司 增强语义的关系抽取方法、装置、计算机设备及存储介质
CN114048288A (zh) * 2021-11-10 2022-02-15 北京明略软件系统有限公司 细粒度情感分析方法、系统、计算机设备和存储介质
CN114386419B (zh) * 2022-01-11 2023-07-25 平安科技(深圳)有限公司 实体识别模型训练方法、装置、设备以及存储介质
CN114386419A (zh) * 2022-01-11 2022-04-22 平安科技(深圳)有限公司 实体识别模型训练方法、装置、设备以及存储介质
CN114580424B (zh) * 2022-04-24 2022-08-05 之江实验室 一种用于法律文书的命名实体识别的标注方法和装置
CN114580424A (zh) * 2022-04-24 2022-06-03 之江实验室 一种用于法律文书的命名实体识别的标注方法和装置
CN115879473A (zh) * 2022-12-26 2023-03-31 淮阴工学院 基于改进图注意力网络的中文医疗命名实体识别方法
CN115879473B (zh) * 2022-12-26 2023-12-01 淮阴工学院 基于改进图注意力网络的中文医疗命名实体识别方法

Also Published As

Publication number Publication date
CN111222317B (zh) 2022-04-29
CN111222317A (zh) 2020-06-02

Similar Documents

Publication Publication Date Title
WO2021072852A1 (fr) Procédé et système de marquage de séquences, et dispositif informatique
US11574122B2 (en) Method and system for joint named entity recognition and relation extraction using convolutional neural network
US11893345B2 (en) Inducing rich interaction structures between words for document-level event argument extraction
US11288593B2 (en) Method, apparatus and device for extracting information
US11651163B2 (en) Multi-turn dialogue response generation with persona modeling
CN111931517B (zh) 文本翻译方法、装置、电子设备以及存储介质
CN112101041B (zh) 基于语义相似度的实体关系抽取方法、装置、设备及介质
US20190073351A1 (en) Generating dependency parses of text segments using neural networks
WO2021143206A1 (fr) Procédé et appareil de traitement en langage naturel à énoncé individuel, dispositif informatique et support de stockage lisible par ordinateur
CN113704460B (zh) 一种文本分类方法、装置、电子设备和存储介质
CN112613322B (zh) 文本处理方法、装置、设备及存储介质
CN113204611A (zh) 建立阅读理解模型的方法、阅读理解方法及对应装置
CN113743101A (zh) 文本纠错方法、装置、电子设备和计算机存储介质
CN113626608B (zh) 增强语义的关系抽取方法、装置、计算机设备及存储介质
CN115146068A (zh) 关系三元组的抽取方法、装置、设备及存储介质
CN111241843B (zh) 基于复合神经网络的语义关系推断系统和方法
CN112599211B (zh) 一种医疗实体关系抽取方法及装置
CN114220505A (zh) 病历数据的信息抽取方法、终端设备及可读存储介质
CN111967253A (zh) 一种实体消歧方法、装置、计算机设备及存储介质
US20230070966A1 (en) Method for processing question, electronic device and storage medium
CN113704466B (zh) 基于迭代网络的文本多标签分类方法、装置及电子设备
CN115203388A (zh) 机器阅读理解方法、装置、计算机设备和存储介质
CN114490935A (zh) 异常文本的检测方法、装置、计算机可读介质及电子设备
CN114491030A (zh) 技能标签的抽取、候选短语分类模型的训练方法及装置
US12038955B2 (en) Method for generating query statement, electronic device and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19948946

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19948946

Country of ref document: EP

Kind code of ref document: A1