CN110837558A - Judgment document entity relation extraction method and system - Google Patents

Judgment document entity relation extraction method and system Download PDF

Info

Publication number
CN110837558A
CN110837558A CN201911082739.0A CN201911082739A CN110837558A CN 110837558 A CN110837558 A CN 110837558A CN 201911082739 A CN201911082739 A CN 201911082739A CN 110837558 A CN110837558 A CN 110837558A
Authority
CN
China
Prior art keywords
entity
vector
training
model bert
relationship
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911082739.0A
Other languages
Chinese (zh)
Other versions
CN110837558B (en
Inventor
不公告发明人
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Xingyun Law Technology Co Ltd
Original Assignee
Chengdu Xingyun Law Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Xingyun Law Technology Co Ltd filed Critical Chengdu Xingyun Law Technology Co Ltd
Priority to CN201911082739.0A priority Critical patent/CN110837558B/en
Publication of CN110837558A publication Critical patent/CN110837558A/en
Application granted granted Critical
Publication of CN110837558B publication Critical patent/CN110837558B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/355Class or cluster creation or modification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/353Clustering; Classification into predefined classes

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a method and a system for extracting the entity relationship of a referee document, wherein the method comprises the following steps: setting a format of sample data; converting the format of the training sample into the format of the sample data to obtain the training sample after format conversion; training a pre-training model bert by using the training sample after format conversion to obtain a trained model bert; inputting a referee document sentence of which the entity relationship is to be extracted into the trained model bert, and outputting the relationship between two entities in the referee document sentence; under the condition of the same data volume, the method and the system have better effect than other relation extraction models; the method and the system use sentence information, entity position information and entity type information to splice in the middle of the model, and accord with the relationship to extract the natural structure of the triple.

Description

Judgment document entity relation extraction method and system
Technical Field
The invention relates to the field of extraction of natural language processing information, in particular to a method and a system for extracting the entity relationship of a referee document.
Background
Currently, common relationship extraction methods include rule-based relationship extraction, relationship extraction based on methods such as deep learning CNN, and the like. Due to the diversity of legal relationships (only civil cases up to 800 cases) and the diversity of textual expressions, rule-based relationship extraction cannot be applied in the field of official documents. Relation extraction based on deep learning CNN and other methods depends on a large amount of labeled data, a large amount of professionals in different case field subdivision are needed, efficiency is low, and requirements for labeling personnel are high. At present, no typical case for extracting entity relation in the referee document field exists.
Disclosure of Invention
The invention aims to extract the relationship between two entities in the referee document. The official document describes case facts in a specific area, and converts unstructured case descriptions into a description mode which can be identified by a program, so that the official document is an important step for realizing case matching. The invention realizes the extraction of entity relations in the case situation facts of the referee documents, and is used for supporting the description of case situations in a triple form.
In order to achieve the above object, one aspect of the present invention provides a method for extracting an entity relationship of a referee document, the method comprising:
setting a format of sample data;
converting the format of the training sample into the format of the sample data to obtain the training sample after format conversion;
training a pre-training model bert by using the training sample after format conversion to obtain a trained model bert;
inputting the referee document sentence with the entity relationship to be extracted into the trained model bert, and outputting the relationship between two entities in the referee document sentence.
Preferably, the structure of the pre-training model bert is as follows:
the input to the model is a token sequence that unambiguously represents a single referee's sentence or a pair of referee's sentences. For a given token, its input representation is constructed by summing the corresponding token, segment, and position elements, resulting in a 768-dimensional vector for the token length representation through a 12-layer bi-directional self-orientation process.
Preferably, the format of the sample data is: (sentence, (entity 1, start position 1, end position 1, entity type1), (entity 2, start position 2, end position 2, entity type2), relationship between entity 1 and entity 2) represents the relationship of a known pair of entities in a sentence, and the entity position information is added to disambiguate when a plurality of identical entities exist in the sentence. For example, sample data: (Zhang III opens a three-bought driving car on the big avenue of Tianfu to have a tsunami, (Zhang III, 1, 2, natural people), (Ben Car, 15, 18, motor vehicle), drive), the relation that the sentence "Zhang III opens a three-bought driving car on the big avenue of Tianfu to have a tsunami" corresponding entity "Zhang III" and "Ben car" is "drive", there are two "Zhang III" in the sentence, the position information is used to distinguish the two "Zhang III", the first "Zhang III" and "Ben car" are "drive" relation, the second "Zhang III" and "Ben car" are "purchase" relation.
Preferably, the training process of the model is as follows:
inputting the training sample after format conversion into a pre-training model bert;
extracting an output vector of the CLS position of the pre-training model bert as a vector representation of a sentence of the referee document, and recording the vector representation as sentenceEmbedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 1 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 1 and is marked as ent1 Embedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 2 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 2 and is marked as ent2 Embedding;
sequentially splicing the sentenceEmbedding vector, the ent1Embedding vector and the ent2Embedding vector to obtain a vector after first splicing;
transforming the vectors subjected to the first splicing to obtain transformed vector representation, and recording the transformed vector representation as vector 1;
converting the entity type corresponding to the entity 1 into vector representation, and marking as ent1type embedding;
converting the entity type corresponding to the entity 2 into vector representation, and marking as ent2type embedding;
sequentially splicing vector1, ent1type embedding and ent2type embedding vectors to obtain a vector after secondary splicing;
and classifying the vectors subjected to the second splicing by using a softmax layer to obtain a trained model bert.
Preferably, the sentence in the method is a sentence of a referee document.
On the other hand, the invention also provides a referee document entity relationship extraction system, which comprises:
a sample data format setting unit for setting a format of sample data;
the training sample format conversion unit is used for converting the format of the training sample into the format of the sample data to obtain the training sample with the processed format;
the model training unit is used for training the pre-training model bert by using the training sample after format conversion to obtain a trained model bert;
and the entity relationship extraction unit is used for inputting the referee document sentence of the entity relationship to be extracted into the trained model bert and outputting the relationship between two entities in the referee document sentence.
Further, the structure of the pre-training model bert is as follows:
the input to the model is a token sequence that unambiguously represents a single referee's sentence or a pair of referee's sentences. For a given token, its input representation is constructed by summing the corresponding token, segment, and position elements, resulting in a 768-dimensional vector for the token length representation through a 12-layer bi-directional self-orientation process. Further, the format of the sample data is: (sentence, (entity 1, start position 1, end position 1, entity type1), (entity 2, start position 2, end position 2, entity type2), relationship between entity 1 and entity 2) represents the relationship of a known pair of entities in a sentence, and the entity position information is added to disambiguate when a plurality of identical entities exist in the sentence. For example, sample data: (Zusan opens a three-bought driving car on the big avenue of Tianfu to have a tsunami, (Zusan, 1, 2, natural people), (Zusan, 15, 18, motor vehicle), driving), the relation of a first entity 'Zusan' corresponding to a sentence 'Zusan opens a three-bought driving car on the big avenue of Tianfu to have a tsunami' and a 'driving car' is 'driving', two 'Zusan' exist in the sentence, position information is used for distinguishing the two 'Zusan', the first 'Zusan' and the 'driving car' are 'driving' relation, and the second 'Zusan' and the 'driving car' are 'purchasing' relation.
Further, the process of training the model by the model training unit is as follows:
inputting the training sample after format conversion into a pre-training model bert;
extracting an output vector of the CLS position of the pre-training model bert as a vector representation of a sentence of the referee document, and recording the vector representation as sentenceEmbedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 1 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 1 and is marked as ent1 Embedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 2 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 2 and is marked as ent2 Embedding;
sequentially splicing the sentenceEmbedding vector, the ent1Embedding vector and the ent2Embedding vector to obtain a vector after first splicing;
transforming the vectors subjected to the first splicing to obtain transformed vector representation, and recording the transformed vector representation as vector 1;
converting the entity type corresponding to the entity 1 into vector representation, and marking as ent1type embedding;
converting the entity type corresponding to the entity 2 into vector representation, and marking as ent2type embedding;
sequentially splicing vector1, ent1type embedding and ent2type embedding vectors to obtain a vector after secondary splicing;
and classifying the vectors subjected to the second splicing by using a softmax layer to obtain a trained model bert.
Further, the sentences in the system are sentences in the referee document.
One or more technical schemes provided by the invention at least have the following technical effects or advantages:
the method and the system can achieve better effect under the condition of small samples. The method and the system adopt a transfer learning method, use a pre-training model bert, and have the experiment effect far better than other relation extraction models (CNN, PCNN and LSTM) under the condition of a small sample.
The method and the system have better effect than other relation extraction models (Bert + CNN, CNN, PCNN, LSTM) under the condition of the same data volume (including small samples, medium samples and large samples). The method and the system use sentence information, entity position information and entity type information to splice in the middle of the model, and accord with the relation to extract the natural structure of the triple.
Drawings
The accompanying drawings, which are included to provide a further understanding of the embodiments of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the principles of the invention;
FIG. 1 is a flow chart of a method for extracting the entity relationship of a referee document according to the present invention;
FIG. 2 is a schematic structural diagram of a pre-training model bert in the present invention;
FIG. 3 is a schematic diagram of a system for extracting the entity relationship of official document according to the present invention.
Detailed Description
In order that the above objects, features and advantages of the present invention can be more clearly understood, a more particular description of the invention will be rendered by reference to the appended drawings. It should be noted that the embodiments and features of the embodiments of the present application may be combined with each other without conflicting with each other.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention, however, the present invention may be practiced in other ways than those specifically described and thus the scope of the present invention is not limited by the specific embodiments disclosed below.
Referring to fig. 1, an embodiment of the present invention provides a method for extracting an entity relationship of a referee document, where the method includes:
setting a format of sample data;
converting the format of the training sample into the format of the sample data to obtain the training sample after format conversion;
training a pre-training model bert by using the training sample after format conversion to obtain a trained model bert;
inputting the sentence of the referee document with the entity relationship to be extracted into the trained model bert, and outputting the relationship between the two entities in the sentence.
The method aims to extract the relationship between two entities in the sentence of the referee document. The method comprises the following three steps:
step 1: and (4) preparing data. The format of sample data is defined as (content, (ent1, start1, end1, type1), (ent2, start2, end2, type2), relationship). Wherein, the sensor represents a sentence, ent1 represents an entity 1, ent2 represents an entity 2, start1 and start2 represent the start position of the entity in the sentence, end1 and end2 represent the end position of the entity in the sentence, type1 and type2 represent the entity type, and relationship represents the relationship between the entity 1 and the entity 2.
Step 2: and (6) data processing. The input data is converted into the input format required by the pre-trained model bert.
And step 3: model training and entity relationship extraction. Detailed structure of the model referring to fig. 2, the training process comprises the following steps:
1. and inputting the input data after format conversion into a pre-training model bert.
2. And extracting an output vector of the CLS position of the model bert as a vector representation of the sensor, and marking the vector representation as sensor embedding, wherein the vector representation represents the meaning of a sentence.
3. Vectors corresponding to start1 to end1 of ent1 in the output sequence of the model bert are extracted, and then the extracted vectors are averaged to be used as vector representation of ent1, which is denoted as ent1 Embedding.
4. Vectors corresponding to start2 to end2 of ent2 in the output sequence of the model bert are extracted, and then the extracted vectors are averaged to be used as vector representation of ent2, which is denoted as ent2 Embedding.
5. And sequentially splicing the sentenceEmbedding, the ent1Embedding and the ent2Embedding to obtain a vector after first splicing.
6. And transforming the vectors subjected to the first splicing by a plurality of full connection layers to obtain a new vector representation vector 1.
7. The entity type1 corresponding to ent1 is converted into a vector representation ent1type embedding.
8. The entity type2 corresponding to ent2 is converted into a vector representation ent2type embedding.
9. Vector1, ent1TypeEmbedding and ent2TypeEmbedding are sequentially spliced to obtain a vector after second splicing.
10. And adding a softmax layer behind the second spliced vector for classifying the second spliced vector to obtain a corresponding relation.
Referring to fig. 3, an embodiment of the present invention provides a system for extracting an entity relationship of a referee document, including:
a sample data format setting unit for setting a format of sample data;
the training sample format conversion unit is used for converting the format of the training sample into the format of the sample data to obtain the training sample after format conversion;
the model training unit is used for training the pre-training model bert by using the training sample after format conversion to obtain a trained model bert;
and the entity relationship extraction unit is used for inputting the referee document sentence of the entity relationship to be extracted into the trained model bert and outputting the relationship between two entities in the referee document sentence.
While preferred embodiments of the present invention have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all such alterations and modifications as fall within the scope of the invention.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present invention without departing from the spirit and scope of the invention. Thus, if such modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.

Claims (8)

1. A method for extracting the entity relationship of a referee document, which is characterized by comprising the following steps:
setting a format of sample data;
converting the format of the training sample into the format of the sample data to obtain the training sample after format conversion;
training a pre-training model bert by using the training sample after format conversion to obtain a trained model bert;
inputting the referee document sentence with the entity relationship to be extracted into the trained model bert, and outputting the relationship between two entities in the referee document sentence.
2. The method for extracting the entity relationship of the official document according to claim 1, wherein the structure of the pre-trained model bert is as follows:
inputting a pre-training model bert into a token sequence which clearly represents a single referee document sentence or a pair of referee document sentences; for a given token, its input representation is constructed by summing the corresponding token, segment, and positionednumbers, resulting in a 768-dimensional vector for the token length representation through a 12-layer bi-directional self-orientation process.
3. The method according to claim 1, wherein the sample data format is: (sentence, (entity 1, start position 1, end position 1, entity type1), (entity 2, start position 2, end position 2, entity type2), relationship between entity 1 and entity 2).
4. The method for extracting the relationship between official document entities as claimed in claim 3, wherein the training process of the pre-training model bert is as follows:
inputting the training sample after format conversion into a pre-training model bert;
extracting an output vector of the CLS position of the pre-training model bert as a vector representation of a sentence of the referee document, and recording the vector representation as sentenceEmbedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 1 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 1 and is marked as ent1 Embedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 2 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 2 and is marked as ent2 Embedding;
splicing the sentenceEmbedding vector, the ent1Embedding vector and the ent2Embedding vector in sequence to obtain a vector after first splicing;
transforming the vectors subjected to the first splicing to obtain transformed vector representation, and recording the transformed vector representation as vector 1;
converting the entity type corresponding to the entity 1 into vector representation, and marking as ent1type embedding;
converting the entity type corresponding to the entity 2 into vector representation, and marking as ent2type embedding;
splicing vector1, ent1type embedding vector and ent2type embedding vector in sequence to obtain a vector after second splicing;
and classifying the vectors subjected to the second splicing by using a softmax layer to obtain a trained model bert.
5. A referee document entity relationship extraction system, the system comprising:
a sample data format setting unit for setting a format of sample data;
the training sample format conversion unit is used for converting the format of the training sample into the format of the sample data to obtain the training sample after format conversion;
the model training unit is used for training the pre-training model bert by using the training sample after format conversion to obtain a trained model bert;
and the entity relationship extraction unit is used for inputting the referee document sentence of the entity relationship to be extracted into the trained model bert and outputting the relationship between two entities in the referee document sentence.
6. The system for extracting referee document entity relationship according to claim 5, wherein the pre-trained model bert has the structure:
inputting a pre-training model bert into a token sequence which clearly represents a single referee document sentence or a pair of referee document sentences; for a given token, its input representation is constructed by summing the corresponding token, segment, and position elements, resulting in a 768-dimensional vector for the token length representation through a 12-layer bi-directional self-orientation process.
7. The system according to claim 5, wherein the sample data format is: (sentence, (entity 1, start position 1, end position 1, entity type1), (entity 2, start position 2, end position 2, entity type2), relationship between entity 1 and entity 2).
8. The system for extracting the relationship between official document entities as claimed in claim 7, wherein the process of training the pre-trained model bert by the model training unit is as follows:
inputting the training sample after format conversion into a pre-training model bert;
extracting an output vector of the CLS position of the pre-training model bert as a vector representation of a sentence of the referee document, and recording the vector representation as sentenceEmbedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 1 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 1 and is marked as ent1 Embedding;
extracting a vector sequence corresponding to the starting position to the ending position of the entity 2 in an output sequence of the pre-training model bert, wherein the average value of the orientation quantity sequence is used as the vector representation of the entity 2 and is marked as ent2 Embedding;
sequentially splicing the sentenceEmbedding vector, the ent1Embedding vector and the ent2Embedding vector to obtain a vector after first splicing;
transforming the vectors subjected to the first splicing to obtain transformed vector representation, and recording the transformed vector representation as vector 1;
converting the entity type corresponding to the entity 1 into vector representation, and marking as ent1type embedding;
converting the entity type corresponding to the entity 2 into vector representation, and marking as ent2type embedding;
sequentially splicing vector1, ent1type embedding and ent2type embedding vectors to obtain a vector after secondary splicing;
and classifying the vectors subjected to the second splicing by using a softmax layer to obtain a trained model bert.
CN201911082739.0A 2019-11-07 2019-11-07 Judgment document entity relation extraction method and system Active CN110837558B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911082739.0A CN110837558B (en) 2019-11-07 2019-11-07 Judgment document entity relation extraction method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911082739.0A CN110837558B (en) 2019-11-07 2019-11-07 Judgment document entity relation extraction method and system

Publications (2)

Publication Number Publication Date
CN110837558A true CN110837558A (en) 2020-02-25
CN110837558B CN110837558B (en) 2022-04-15

Family

ID=69574630

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911082739.0A Active CN110837558B (en) 2019-11-07 2019-11-07 Judgment document entity relation extraction method and system

Country Status (1)

Country Link
CN (1) CN110837558B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111476034A (en) * 2020-04-07 2020-07-31 同方赛威讯信息技术有限公司 Legal document information extraction method and system based on combination of rules and models
CN111539490A (en) * 2020-06-19 2020-08-14 支付宝(杭州)信息技术有限公司 Business model training method and device
CN111563374A (en) * 2020-03-23 2020-08-21 北京交通大学 Personnel social relationship extraction method based on judicial official documents
CN113569046A (en) * 2021-07-19 2021-10-29 北京华宇元典信息服务有限公司 Judgment document character relation identification method and device and electronic equipment

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10170114B2 (en) * 2013-05-30 2019-01-01 Promptu Systems Corporation Systems and methods for adaptive proper name entity recognition and understanding
CN109933789A (en) * 2019-02-27 2019-06-25 中国地质大学(武汉) A kind of judicial domain Relation extraction method and system neural network based
CN110032648A (en) * 2019-03-19 2019-07-19 微医云(杭州)控股有限公司 A kind of case history structuring analytic method based on medical domain entity
CN110059320A (en) * 2019-04-23 2019-07-26 腾讯科技(深圳)有限公司 Entity relation extraction method, apparatus, computer equipment and storage medium
CN110287334A (en) * 2019-06-13 2019-09-27 淮阴工学院 A kind of school's domain knowledge map construction method based on Entity recognition and attribute extraction model
CN110347802A (en) * 2019-07-17 2019-10-18 北京金山数字娱乐科技有限公司 A kind of text analyzing method and device
US20190325029A1 (en) * 2018-04-18 2019-10-24 HelpShift, Inc. System and methods for processing and interpreting text messages

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10170114B2 (en) * 2013-05-30 2019-01-01 Promptu Systems Corporation Systems and methods for adaptive proper name entity recognition and understanding
US20190325029A1 (en) * 2018-04-18 2019-10-24 HelpShift, Inc. System and methods for processing and interpreting text messages
CN109933789A (en) * 2019-02-27 2019-06-25 中国地质大学(武汉) A kind of judicial domain Relation extraction method and system neural network based
CN110032648A (en) * 2019-03-19 2019-07-19 微医云(杭州)控股有限公司 A kind of case history structuring analytic method based on medical domain entity
CN110059320A (en) * 2019-04-23 2019-07-26 腾讯科技(深圳)有限公司 Entity relation extraction method, apparatus, computer equipment and storage medium
CN110287334A (en) * 2019-06-13 2019-09-27 淮阴工学院 A kind of school's domain knowledge map construction method based on Entity recognition and attribute extraction model
CN110347802A (en) * 2019-07-17 2019-10-18 北京金山数字娱乐科技有限公司 A kind of text analyzing method and device

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
HAINAN CHEN等: "An automatic literature knowledge graph and reasoning network modeling framework based on ontology and natural language processing", 《ADVANCED ENGINEERING INFORMATICS》 *
杨飘等: "基于BERT嵌入的中文命名实体识别方法", 《计算机工程》 *
王立胜: "基于多级语义表示和混合注意力机制的机器阅读理解研究", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 *
祖木然提古丽 库尔班等: "中文命名实体识别模型对比分析", 《现代计算机》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111563374A (en) * 2020-03-23 2020-08-21 北京交通大学 Personnel social relationship extraction method based on judicial official documents
CN111476034A (en) * 2020-04-07 2020-07-31 同方赛威讯信息技术有限公司 Legal document information extraction method and system based on combination of rules and models
CN111539490A (en) * 2020-06-19 2020-08-14 支付宝(杭州)信息技术有限公司 Business model training method and device
CN111539490B (en) * 2020-06-19 2020-10-16 支付宝(杭州)信息技术有限公司 Business model training method and device
CN113569046A (en) * 2021-07-19 2021-10-29 北京华宇元典信息服务有限公司 Judgment document character relation identification method and device and electronic equipment

Also Published As

Publication number Publication date
CN110837558B (en) 2022-04-15

Similar Documents

Publication Publication Date Title
CN110837558B (en) Judgment document entity relation extraction method and system
CN107729309B (en) Deep learning-based Chinese semantic analysis method and device
CN110162636B (en) D-LSTM-based emotion reason identification method
CN112116075B (en) Event extraction model generation method and device, text event extraction method and device
CN112860888B (en) Attention mechanism-based bimodal emotion analysis method
CN108664474B (en) Resume analysis method based on deep learning
CN111259141A (en) Social media corpus emotion analysis method based on multi-model fusion
CN110276396B (en) Image description generation method based on object saliency and cross-modal fusion features
CN109783636A (en) A kind of car review subject distillation method based on classifier chains
CN114398889A (en) Video text summarization method, device and storage medium based on multi-modal model
CN115658955B (en) Cross-media retrieval and model training method, device, equipment and menu retrieval system
CN112749283A (en) Entity relationship joint extraction method for legal field
CN114170411A (en) Picture emotion recognition method integrating multi-scale information
CN115687571A (en) Depth unsupervised cross-modal retrieval method based on modal fusion reconstruction hash
CN112131881A (en) Information extraction method and device, electronic equipment and storage medium
CN115374786A (en) Entity and relationship combined extraction method and device, storage medium and terminal
CN116246279A (en) Graphic and text feature fusion method based on CLIP background knowledge
CN113761377A (en) Attention mechanism multi-feature fusion-based false information detection method and device, electronic equipment and storage medium
CN116595023A (en) Address information updating method and device, electronic equipment and storage medium
CN117115505A (en) Emotion enhancement continuous training method combining knowledge distillation and contrast learning
CN116450829A (en) Medical text classification method, device, equipment and medium
CN114611529B (en) Intention recognition method and device, electronic equipment and storage medium
CN115600561A (en) Webpage structuring method, equipment and storage medium fusing rules and small samples
WO2022251720A1 (en) Character-level attention neural networks
CN114510561A (en) Answer selection method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant