CN110968661A - Event extraction method and system, computer readable storage medium and electronic device - Google Patents

Event extraction method and system, computer readable storage medium and electronic device Download PDF

Info

Publication number
CN110968661A
CN110968661A CN202010141127.0A CN202010141127A CN110968661A CN 110968661 A CN110968661 A CN 110968661A CN 202010141127 A CN202010141127 A CN 202010141127A CN 110968661 A CN110968661 A CN 110968661A
Authority
CN
China
Prior art keywords
event
model
labeling
event extraction
labeled
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010141127.0A
Other languages
Chinese (zh)
Inventor
罗镇权
张发展
李焕
刘世林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Business Big Data Technology Co Ltd
Original Assignee
Chengdu Business Big Data Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Business Big Data Technology Co Ltd filed Critical Chengdu Business Big Data Technology Co Ltd
Priority to CN202010141127.0A priority Critical patent/CN110968661A/en
Publication of CN110968661A publication Critical patent/CN110968661A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/31Indexing; Data structures therefor; Storage structures
    • G06F16/313Selection or weighting of terms for indexing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • Computational Linguistics (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention relates to an event extraction method and system, a computer readable storage medium and an electronic device, wherein the method comprises the following steps: step 1, manually marking prepared linguistic data, including marking each element of an event and marking an event representation phrase of a trigger event; step 2, carrying out sequence labeling on the labeled linguistic data, wherein event representation phrases in the labeled linguistic data are labeled by event types; step 3, inputting the data processed in the step 2 into a model for training to obtain an event extraction model; and 4, inputting the text to be extracted into the event extraction model, and outputting to obtain an identification result. The invention provides a novel event extraction method, which reduces the work which is finished by two models originally to the work which can be finished by only one model, and greatly saves resources and time.

Description

Event extraction method and system, computer readable storage medium and electronic device
Technical Field
The present invention relates to the field of natural language processing technologies, and in particular, to an event extraction method and system, a computer-readable storage medium, and an electronic device capable of reducing event extraction computing resources.
Background
In the field of knowledge maps, an event is an occurrence or a change in state of one or more actions involving one or more characters occurring at a specific time point or time period and within a specific geographical area. Event extraction refers to extracting event information interested by a user from a natural language text and presenting the event information in a structured form, such as what person/organization, what time, where, what is done.
BERT (bidirectional Encoder retrieval from transformers) is a large-scale pre-training language model based on bidirectional transformers issued by Google, which can efficiently extract text information and apply to various NLP tasks, the research refreshes the current optimal performance records of 11 NLP tasks by virtue of the pre-training model, and the model can be understood as a language Encoder to convert input sentences or paragraphs into feature vectors.
Currently, event extraction is generally performed through the following processes: firstly, preparing a training corpus; secondly, judging whether the linguistic data (chapter level or sentence level) contains events or not and event types through a text classification model, namely classifying the events; and finally, extracting the event elements from the corpus containing the events by using an event element extraction model to complete the whole event extraction process.
From the above description, it can be seen that two models are required to complete event extraction: an event classification model and an event element extraction model. The deep learning model generally requires a large amount of computing resources, such as CPU computing resources, memory resources, GPU computing resources, GPU video memory, and the like, and especially in recent years, the pretraining language model represented by BERT is known to consume a large amount of video memory. Currently, when an event extraction task is performed, two BERT models need to be used simultaneously, so that a large amount of resources and time are consumed.
Disclosure of Invention
The present invention is directed to overcome the disadvantages of the prior art that event extraction requires a large amount of resources and time, and provides an event extraction method and system, a computer-readable storage medium, and an electronic device, which can reduce event extraction computing resources and time consumption.
In order to achieve the above object, the embodiments of the present invention provide the following technical solutions:
in one aspect, an embodiment of the present invention provides an event extraction method, including the following steps:
step 1, manually labeling the prepared text corpus, including labeling each element of an event and labeling an event representation phrase of a trigger event;
step 2, carrying out sequence labeling on the labeled linguistic data, wherein event representation phrases in the labeled linguistic data are labeled by event types;
step 3, inputting the data processed in the step 2 into a model for training to obtain an event extraction model;
and 4, inputting the text to be extracted into the event extraction model, and outputting to obtain an identification result.
According to the method, the elements of the event and the event representation phrases are labeled at the same time, then sequence labeling is carried out, but the event representation phrases are not labeled at the same time during sequence labeling, but event type labeling is carried out, the event type and the event elements can be simultaneously identified and extracted through an event extraction model obtained through the training after the processing, tasks completed by two models in the prior art can be realized through one model, the consumption of computing resources is greatly reduced, and the event extraction time is greatly reduced.
And in the step 2, a BEIO sequence labeling method is adopted to label the labeled corpus.
The model in the step 3 is a BERT model or a bilstm + crf model.
In the step 4, when the recognition result output by the event extraction model includes two or more event types, the first event type is taken as the standard.
On the other hand, an embodiment of the present invention further provides an event extraction system, including:
the corpus tagging module is used for tagging the prepared text corpus twice, wherein the tagging for one time comprises tagging each element of an event and tagging an event representation phrase triggering the event; the secondary labeling is to perform sequence labeling on the linguistic data subjected to the primary labeling, and the event representation phrases in the linguistic data subjected to the secondary labeling are labeled by event types;
the model training module is used for inputting the data subjected to secondary labeling into a model for training to obtain an event extraction model;
and the event extraction model is used for performing event element identification and event type identification on the input text to be extracted and outputting to obtain an identification result.
And the corpus labeling module is used for labeling the once labeled corpus by adopting a BEIO sequence labeling method.
The model used in the model training module is a BERT model or a bilstm + crf model.
In still another aspect, the present invention also provides a computer-readable storage medium including computer-readable instructions, which, when executed, cause a processor to perform the operations of the method described in the present invention.
In another aspect, an embodiment of the present invention also provides an electronic device, including: a memory storing program instructions; and the processor is connected with the memory and executes the program instructions in the memory to realize the steps of the method in the embodiment of the invention.
Compared with the prior art, the event extraction method is a novel event extraction method, the events can be extracted by only one model, and compared with the method that two models are needed, the method greatly reduces resource consumption and accelerates the event extraction efficiency. In addition, through test comparison, the method can also obviously improve the accuracy of the event extraction result.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the embodiments will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present invention and therefore should not be considered as limiting the scope, and for those skilled in the art, other related drawings can be obtained according to the drawings without inventive efforts.
Fig. 1 is a flowchart of an event extraction method described in the embodiment.
FIG. 2 is a diagram illustrating corpora labeled manually in the embodiment.
Fig. 3 is a schematic diagram of an event extraction system according to an embodiment.
Fig. 4 is a block diagram showing the components of the electronic apparatus described in the embodiment.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. The components of embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
As shown in fig. 1, the present embodiment schematically provides an event extraction method, including the following steps:
step 1, manually labeling the prepared text corpus, including labeling each element of an event and labeling an event representation phrase of a trigger event.
And 2, processing the labeled corpus into structured data, namely performing sequence labeling, wherein event representation phrases in the corpus are labeled by event types.
The BEIO is a sequence labeling method with wide application, so the sequence labeling method is adopted in the step for data conversion. The BEIO is a commonly used label representation method for sequence labeling, and is mainly used in named entity identification (NER) of sequence labeling. And (3) sequence labeling: in brief, the sequence notation is: given a sequence, each element in the sequence is marked or tagged. In general, a sequence refers to a sentence, and an element refers to a word in the sentence. B = begin, E = end, I = intermediate, representing the beginning, end and middle of a word, respectively, O = other, representing that the above three cases are not met. Such as "i am in all. ", an address is to be noted, i.e., as" OOBO EO ".
However, in this step, not all data in the corpus are BIEO converted, but the EVENT-characterized phrases are converted in an EVENT _ type manner. For example, for the FRAUD EVENT type, it is converted into "EVENT _ FRAUD" or "E _ FRAUD".
And 3, inputting the processed structured data into a model for training to obtain an event extraction model. And (4) the model predicts that the annotation symbol of the first event characterization phrase is the same as the annotation corpus, and then the event type prediction is correct.
The model here can be selected as long as the model can be used for sequence labeling, for example, a BERT model or a bilstm + crf model can be used, and the two models have the best effect at present.
And 4, inputting the text to BE extracted into the EVENT extraction model obtained after training for prediction, wherein the EVENT extraction model identifies EVENT elements of the sample corpus according to the BE, BIE, BIIE, B and the like, and identifies the EVENT type according to the EVENT _ type mode.
In the test process, the parameters of the trained model can be or may be adjusted by using the test sample, but in the practical application, the document to be extracted is directly input into the model, and the recognition result is output.
By the method, the two original models can be reduced into one model to achieve the purpose of event extraction.
Test examples
Referring to fig. 2, for example, the text corpus "in which the AA technology falsely increases the total profit by 8000 ten thousand yuan in a manner of fictitious customers and counterfeit contracts, and virtually increases the bank deposit by about 2.18 million yuan and virtually advances the engineering money by 3.1 ten thousand yuan, which leads to a false record in the year 2015 report" is taken as an example.
Firstly, manually marking, namely marking each element of an event, wherein if the 'AA science and technology' is marked as an event main body, the '2015 year' is marked as the occurrence time, and meanwhile, the '8000 ten thousand yuan of gross profit increase in an equivalent mode of counterfeit combination' is marked as an event representation phrase, and the 'financial fraud' is marked.
The artificially annotated text corpus is then processed into structured data resulting in "O | O | B _ Sub | I | E | O | O | O | O | O | O | O | O | O | O | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | E _ FRAUD | O | O | O | O | O | O | O | O | O | O | O | O | O | O | O | O | O |. Each word (including punctuation) corresponds to a label and "|" is a separator. For example, "AA technology" corresponds to "B _ Sub | I | E", "forgery-contract equally falsely increases the total profit by 8000 ten thousand yuan" corresponds to "E _ frame | E _ ud | E _ frame | 2015 year" corresponds to "B _ TIME | I _ TIME | E _ TIME".
It should be noted that although the event elements are all converted, the converted representations of different entity names are different. In addition, different indications are also used for different event types, such as E _ FRAUD for FRAUD type, and E _ evade tables for evasion type.
And finally, inputting the structured data obtained by processing into a model, and training to obtain an event extraction model.
And after the trained event extraction model is obtained, performing event extraction on the text.
The event extraction model is regarded as an event with the largest probability value of the output event, the event is labeled according to words, the label is ' E _1| E _1| E _1| E _1| E _1| E _1| E _1| E _1| assuming that the event representation phrase has 10 words, and the label is ' E _1| E _1| E _1| E _1| E _1| ' the event extraction model may not achieve the perfect label identification, and a plurality of event identification results may occur, wherein the first event is taken as a reference. For example, "E _1| 0| 0| E _2| 0| 0| 0| 0", so the first E _1 is considered to be an E _1 event, the fifth E _2 is considered to be an E _2 event, and the first E _1 event is considered to be an E _1 event, and experiments using the BERT model prove that the effect is good.
In addition, through test detection, the method and the device can not only greatly improve the event extraction efficiency, but also improve the accuracy of the extraction result. As shown in the table below, the accuracy of event extraction using the method of the present invention is significantly higher than that of the conventional method based on the same test data set.
The method of the invention Conventional methods
Event class average F1 value 85% 76%
Event element extraction average F1 value 55.64% 53.33%
Note: f1 value = correct rate recall 2/(correct rate + recall) is used to evaluate the merits of the different algorithms.
Referring to fig. 3, an event extraction system is also provided in the embodiment of the present invention, and includes a corpus tagging module, a model training module, and an event extraction model.
The corpus tagging module is used for tagging the prepared text corpus twice, wherein the tagging for one time comprises tagging each element of an event and tagging an event representation phrase triggering the event; and the secondary labeling is to perform sequence labeling on the linguistic data subjected to the primary labeling, and the event representation phrases in the linguistic data subjected to the labeling are labeled by event types. In the secondary labeling, EVENT type labeling is carried out on the EVENT representation phrases, such as "EVENT _ EVENT type" or "E _ EVENT type", and sequence labeling is carried out on characters (including characters, punctuation marks, English characters and numeric characters) outside the EVENT representation phrases, such as BIEO sequence labeling.
And the model training module is used for inputting the data subjected to secondary labeling into the model for training to obtain an event extraction model. The model used here is preferably the BERT model. The model used here is the BERT model or the bilstm + crf model.
The event extraction model is used for carrying out event element identification and event type identification on the input text to be extracted and outputting an identification result. The event extraction model is regarded as an event according to the maximum probability value of the output event, and when the output recognition result comprises two or more event types, the first event type is taken as the standard.
The event extraction system is based on the same inventive concept of the aforementioned event extraction method, therefore, the related description of the event extraction method can be referred to for what is not referred to in the description of the system, and the same applies to what is referred to in the description of the system.
As shown in fig. 4, the present embodiment also provides an electronic device, which may include a processor 51 and a memory 52, wherein the memory 52 is coupled to the processor 51. It is noted that the figure is exemplary and that other types of structures may be used in addition to or in place of the structure to implement text labeling, data conversion, communication, or other functionality.
As shown in fig. 4, the electronic device may further include: an input unit 53, a display unit 54, and a power supply 55. It is to be noted that the electronic device does not necessarily have to comprise all the components shown in fig. 4. Furthermore, the electronic device may also comprise components not shown in fig. 4, reference being made to the prior art.
The processor 51, also sometimes referred to as a controller or operational control, may comprise a microprocessor or other processor device and/or logic device, the processor 51 receiving input and controlling operation of the various components of the electronic device.
The memory 52 may be one or more of a buffer, a flash memory, a hard drive, a removable medium, a volatile memory, a non-volatile memory, or other suitable devices, and may store the configuration information of the processor 51, the instructions executed by the processor 51, the recorded table data, and other information. The processor 51 may execute a program stored in the memory 52 to realize information storage or processing, or the like. In one embodiment, a buffer memory, i.e., a buffer, is also included in the memory 52 to store the intermediate information.
The input unit 53 is for example used to provide the processor 51 with text data to be annotated. The display unit 54 is used for displaying various results in the process, such as input text data, the converted multi-dimensional vector, the calculated distance value, etc., and may be, for example, an LCD display, but the present invention is not limited thereto. The power supply 55 is used to provide power to the electronic device.
Embodiments of the present invention further provide a computer readable instruction, where when the instruction is executed in an electronic device, the program causes the electronic device to execute the operation steps included in the method of the present invention.
Embodiments of the present invention further provide a storage medium storing computer-readable instructions, where the computer-readable instructions cause an electronic device to execute the operation steps included in the method of the present invention.
Those of ordinary skill in the art will appreciate that the elements and algorithm steps of the examples described in connection with the embodiments disclosed herein may be embodied in electronic hardware, computer software, or combinations of both, and that the components and steps of the examples have been described in a functional general in the foregoing description for the purpose of illustrating clearly the interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention essentially or partially contributes to the prior art, or all or part of the technical solution can be embodied in the form of a software product stored in a storage medium and including instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
Those of ordinary skill in the art will appreciate that the various illustrative modules described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative components and steps have been described above generally in terms of their functionality in order to clearly illustrate the interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
In the several embodiments provided in the present application, it should be understood that the disclosed system may be implemented in other ways. For example, the above-described system embodiments are merely illustrative, and for example, the division of the modules is merely a logical division, and in actual implementation, there may be other divisions, for example, multiple modules or components may be combined or integrated into another system, or some features may be omitted, or not implemented.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (9)

1. An event extraction method, comprising the steps of:
step 1, manually labeling the prepared text corpus, including labeling each element of an event and labeling an event representation phrase of a trigger event;
step 2, carrying out sequence labeling on the labeled linguistic data, wherein event representation phrases in the labeled linguistic data are labeled by event types;
step 3, inputting the data processed in the step 2 into a model for training to obtain an event extraction model;
and 4, inputting the text to be extracted into the event extraction model, and outputting to obtain an identification result.
2. The event extraction method according to claim 1, wherein in the step 2, a BEIO sequence notation method is adopted to label the labeled corpus.
3. The event extraction method as claimed in claim 1, wherein the model in the step 3 is a BERT model or a bilstm + crf model.
4. The event extraction method according to claim 1, wherein in the step 4, when the recognition result output by the event extraction model includes two or more event types, the first event type is used as a criterion.
5. An event extraction system, comprising:
the corpus tagging module is used for tagging the prepared text corpus twice, wherein the tagging for one time comprises tagging each element of an event and tagging an event representation phrase triggering the event; the secondary labeling is to perform sequence labeling on the linguistic data subjected to the primary labeling, and the event representation phrases in the linguistic data subjected to the secondary labeling are labeled by event types;
the model training module is used for inputting the data subjected to secondary labeling into a model for training to obtain an event extraction model;
and the event extraction model is used for performing event element identification and event type identification on the input text to be extracted and outputting to obtain an identification result.
6. The event extraction system according to claim 5, wherein the corpus labeling module labels the once labeled corpus by adopting a BEIO sequence labeling method.
7. The event extraction system according to claim 5, wherein the model used in the model training module is a BERT model or a bilstm + crf model.
8. A computer readable storage medium comprising computer readable instructions that, when executed, cause a processor to perform the operations of the method of any of claims 1-4.
9. An electronic device, comprising:
a memory storing program instructions;
a processor coupled to the memory and executing the program instructions in the memory to implement the steps of the method of any of claims 1-4.
CN202010141127.0A 2020-03-04 2020-03-04 Event extraction method and system, computer readable storage medium and electronic device Pending CN110968661A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010141127.0A CN110968661A (en) 2020-03-04 2020-03-04 Event extraction method and system, computer readable storage medium and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010141127.0A CN110968661A (en) 2020-03-04 2020-03-04 Event extraction method and system, computer readable storage medium and electronic device

Publications (1)

Publication Number Publication Date
CN110968661A true CN110968661A (en) 2020-04-07

Family

ID=70038222

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010141127.0A Pending CN110968661A (en) 2020-03-04 2020-03-04 Event extraction method and system, computer readable storage medium and electronic device

Country Status (1)

Country Link
CN (1) CN110968661A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111931481A (en) * 2020-07-03 2020-11-13 北京新联财通咨询有限公司 Text emotion recognition method and device, storage medium and computer equipment
CN112800762A (en) * 2021-01-25 2021-05-14 上海犀语科技有限公司 Element content extraction method for processing text with format style
CN113342935A (en) * 2021-06-04 2021-09-03 北京捷通华声科技股份有限公司 Semantic recognition method and device, electronic equipment and readable storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109582949A (en) * 2018-09-14 2019-04-05 阿里巴巴集团控股有限公司 Event element abstracting method, calculates equipment and storage medium at device
CN109670172A (en) * 2018-12-06 2019-04-23 桂林电子科技大学 A kind of scenic spot anomalous event abstracting method based on complex neural network
US20190278754A1 (en) * 2012-05-18 2019-09-12 Splunk Inc. Query handling for field searchable raw machine data using a field searchable datastore and an inverted index
CN110297913A (en) * 2019-06-12 2019-10-01 中电科大数据研究院有限公司 A kind of electronic government documents entity abstracting method
CN110489514A (en) * 2019-07-23 2019-11-22 成都数联铭品科技有限公司 Promote system and method, the event extraction method and system of event extraction annotating efficiency
CN110633409A (en) * 2018-06-20 2019-12-31 上海财经大学 Rule and deep learning fused automobile news event extraction method
CN110765774A (en) * 2019-10-08 2020-02-07 北京三快在线科技有限公司 Training method and device of information extraction model and information extraction method and device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190278754A1 (en) * 2012-05-18 2019-09-12 Splunk Inc. Query handling for field searchable raw machine data using a field searchable datastore and an inverted index
CN110633409A (en) * 2018-06-20 2019-12-31 上海财经大学 Rule and deep learning fused automobile news event extraction method
CN109582949A (en) * 2018-09-14 2019-04-05 阿里巴巴集团控股有限公司 Event element abstracting method, calculates equipment and storage medium at device
CN109670172A (en) * 2018-12-06 2019-04-23 桂林电子科技大学 A kind of scenic spot anomalous event abstracting method based on complex neural network
CN110297913A (en) * 2019-06-12 2019-10-01 中电科大数据研究院有限公司 A kind of electronic government documents entity abstracting method
CN110489514A (en) * 2019-07-23 2019-11-22 成都数联铭品科技有限公司 Promote system and method, the event extraction method and system of event extraction annotating efficiency
CN110765774A (en) * 2019-10-08 2020-02-07 北京三快在线科技有限公司 Training method and device of information extraction model and information extraction method and device

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111931481A (en) * 2020-07-03 2020-11-13 北京新联财通咨询有限公司 Text emotion recognition method and device, storage medium and computer equipment
CN112800762A (en) * 2021-01-25 2021-05-14 上海犀语科技有限公司 Element content extraction method for processing text with format style
CN113342935A (en) * 2021-06-04 2021-09-03 北京捷通华声科技股份有限公司 Semantic recognition method and device, electronic equipment and readable storage medium

Similar Documents

Publication Publication Date Title
CN107729309B (en) Deep learning-based Chinese semantic analysis method and device
CN110765763A (en) Error correction method and device for speech recognition text, computer equipment and storage medium
CN110597997B (en) Military scenario text event extraction corpus iterative construction method and device
CN110580308B (en) Information auditing method and device, electronic equipment and storage medium
CN110968661A (en) Event extraction method and system, computer readable storage medium and electronic device
US20220147814A1 (en) Task specific processing of regulatory content
CN110502742B (en) Complex entity extraction method, device, medium and system
CN111062217A (en) Language information processing method and device, storage medium and electronic equipment
CN111339260A (en) BERT and QA thought-based fine-grained emotion analysis method
CN111274829A (en) Sequence labeling method using cross-language information
CN113204967A (en) Resume named entity identification method and system
CN112597366A (en) Encoder-Decoder-based event extraction method
CN115759119A (en) Financial text emotion analysis method, system, medium and equipment
CN109635289B (en) Entry classification method and audit information extraction method
Khorjuvenkar et al. Parts of speech tagging for Konkani language
CN113553853B (en) Named entity recognition method and device, computer equipment and storage medium
CN115688703A (en) Specific field text error correction method, storage medium and device
Lai et al. Web information extraction based on hidden Markov model
CN113688233A (en) Text understanding method for semantic search of knowledge graph
CN112133308A (en) Method and device for multi-label classification of voice recognition text
CN111341404A (en) Electronic medical record data set analysis method and system based on ernie model
CN111259650A (en) Text automatic generation method based on class mark sequence generation type countermeasure model
CN117494806B (en) Relation extraction method, system and medium based on knowledge graph and large language model
CN115618968B (en) New idea discovery method and device, electronic device and storage medium
CN111783471B (en) Semantic recognition method, device, equipment and storage medium for natural language

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200407