CN112633014B - Neural network-based long text reference digestion method and device - Google Patents
Neural network-based long text reference digestion method and device Download PDFInfo
- Publication number
- CN112633014B CN112633014B CN202011437239.7A CN202011437239A CN112633014B CN 112633014 B CN112633014 B CN 112633014B CN 202011437239 A CN202011437239 A CN 202011437239A CN 112633014 B CN112633014 B CN 112633014B
- Authority
- CN
- China
- Prior art keywords
- information
- vector
- head attention
- attention model
- indicated
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 26
- 230000029087 digestion Effects 0.000 title claims abstract description 24
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 17
- 239000013598 vector Substances 0.000 claims description 84
- 230000007246 mechanism Effects 0.000 claims description 16
- 238000004364 calculation method Methods 0.000 claims description 8
- 238000000605 extraction Methods 0.000 claims description 7
- 238000013507 mapping Methods 0.000 claims description 4
- XEBWQGVWTUSTLN-UHFFFAOYSA-M phenylmercury acetate Chemical compound CC(=O)O[Hg]C1=CC=CC=C1 XEBWQGVWTUSTLN-UHFFFAOYSA-M 0.000 claims description 4
- 230000000694 effects Effects 0.000 abstract description 5
- 238000003058 natural language processing Methods 0.000 abstract description 4
- 238000012549 training Methods 0.000 description 8
- 238000012545 processing Methods 0.000 description 6
- 238000013135 deep learning Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 2
- 239000000047 product Substances 0.000 description 2
- 238000011084 recovery Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000013078 crystal Substances 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000009795 derivation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 230000008092 positive effect Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
- G06F40/35—Discourse or dialogue representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- General Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Machine Translation (AREA)
Abstract
The invention relates to the field of natural language processing learning, and discloses a long text reference digestion method and a long text reference digestion device based on a neural network. The method is helpful for directly judging the reference relationship of the reference part and the referred part in the current context, and has better effect on both dominant reference and zero reference.
Description
Technical Field
The invention relates to the field of natural language processing learning, in particular to a long text reference digestion method and device based on a neural network.
Background
Natural language understanding means that a human uses a computer to process information such as the form, pronunciation, and semantics of a natural language expressed in the form of text, voice, and the like.
The reference and omission are language phenomena widely existing in natural language, have positive effects of simplifying expression, connecting context and the like, but can cause ambiguity problem of sentences, and bring great difficulty to understanding natural language, so that the content or default part of the pronoun reference needs to be recovered and supplemented.
Reference to resolution, in a broad sense, is a question of determining in chapters to which noun phrase a pronoun points. According to the direction, it can be divided into back finger and pre-finger. The back-meaning is that the antecedent of the pronoun precedes the pronoun, and the pre-meaning is that the antecedent of the pronoun follows the pronoun. Three categories can be distinguished by the type of reference being made: human pronouns, indicated pronouns, definite descriptions, omissions, partial-whole designations, and common noun phrases.
The index digestion has a longer research history, and the performance of the index digestion system is continuously improved from the research of early manual rules and other theoretical methods to the derivation of computer automatic processing technology in a large-scale corpus afterwards and the introduction of various machine learning methods at present. However, the understanding and representing method of the semantics in the natural language is still not mature enough, and the deep language knowledge and the semantic features are simpler to use, so that the deep mining of different characteristics of multiple levels of words, sentences and chapters is not carried out, and the effective utilization of the context information is not carried out. The reference resolution is a key difficulty in natural language processing, and has important significance for information extraction in the field of natural language processing.
Disclosure of Invention
In view of the foregoing drawbacks of the prior art, an object of the present invention is to provide a neural network-based long text reference resolution method and apparatus, which introduces deep learning techniques into the pronoun resolution task to recover and supplement the reference and default portions of the long text.
In order to achieve the above purpose, the invention provides a long text reference digestion method based on a neural network, which introduces a deep learning technology into a pronouncing digestion task to realize Chinese pronouncing digestion and omit a recovery task. In particular, the present invention employs an attention network, which is essentially a multi-layered forward neural network, for increasing or decreasing the degree of attention of the network to certain words by computing the probability value between the target and the source as the attention, and for adjusting in error feedback. The time complexity of the attention network is much smaller than that of the cyclic neural network (RNN network) and the like, and the application is applicable.
The invention provides a long text reference digestion method based on a neural network, which comprises the following steps:
step S1: inputting a long text;
step S2: extracting the information of the indicated part, the information of the surrounding words of the indicated part, the original text information, the information of the indicated part and the information of the surrounding words of the indicated part in the long text, and carrying out vectorization representation; embedding the indicating part information, the indicating part surrounding word information, the original text information, the indicated part information and the indicated part surrounding word information of the vectorization representation into position information respectively, and correspondingly outputting an indicating part vector, an indicating part surrounding information vector, an original text part vector, an indicated part vector and an indicated part surrounding information vector;
step S3: calculating the multi-head attention mechanism of the information vector around the pointing part and the pointing part, and then continuously calculating the obtained result and the original text part vector by adopting the multi-head attention mechanism to obtain a first characteristic result;
step S4: carrying out multi-head attention mechanism calculation on the pointed part vector and the information vector around the pointed part, and then, continuing to carry out multi-head attention mechanism calculation on the obtained result and the original part vector to obtain a second characteristic result;
step S5: and (3) connecting the first characteristic result and the second characteristic result obtained in the step (S3) and the step (S4) to form a comprehensive result, and mapping the comprehensive result in a discrimination space by utilizing a Softmax layer to judge whether the reference is indicated or not.
Further, the step S3 specifically includes: the method comprises the steps of taking a reference part vector as a query of a first multi-head attention model, taking a reference part surrounding information vector as a key and a value, taking output of the first multi-head attention model as a query of a second multi-head attention model, taking an original Wen Bufen vector as the key and the value of the second multi-head attention model, and taking the second multi-head attention model as an output of a first characteristic result.
Further, the step S4 specifically includes: the method comprises the steps of taking a designated part vector as a query of a third multi-head attention model, taking a designated part surrounding information vector as a key and a value, taking output of the third multi-head attention model as a query of a fourth multi-head attention model, taking an original Wen Bufen vector as the key and the value of the fourth multi-head attention model, and taking the fourth multi-head attention model as an output of a second characteristic result.
The invention also discloses a long text reference digestion device based on the neural network, which comprises a long text extraction module, a connecting layer, a Softmax layer and four multi-head attention models;
the long text extraction module is used for extracting the information of the indicated part, the information of the words around the indicated part, the original text information, the information of the indicated part and the information of the words around the indicated part from the long text and carrying out vectorization representation; embedding the indicating part information, the indicating part surrounding word information, the original text information, the indicated part information and the indicated part surrounding word information of the vectorization representation into position information respectively, and outputting an indicating part vector, an indicating part surrounding information vector, an original text part vector, an indicated part vector and an indicated part surrounding information vector;
the connection relation of the input and the output of the four multi-head attention models is as follows:
the input of the query of the first multi-head attention model refers to a partial vector, and the input of keys and values refers to a partial surrounding information vector; the output of the first multi-headed attention model is the input of a query of the second multi-headed attention model; the key and value inputs of the second multi-head attention model are original text partial vectors; outputting a first characteristic result by the second multi-head attention model;
the input of the query of the third multi-head attention model is a designated partial vector, the input of keys and values of the key and the value of the key and the value of the key are the designated partial vector, the output of the third multi-head attention model is the input of the query of the fourth multi-head attention model, the key and the value of the fourth multi-head attention model are the original partial vector, and the fourth multi-head attention model outputs a second characteristic result;
the connection layer is used for connecting the first characteristic result and the second characteristic result to form a comprehensive result;
the Softmax layer is used for mapping the comprehensive result in a discrimination space to judge whether the result is referred or not.
The technical effects are as follows:
the long text reference digestion method realizes the processing of different layers of information in natural language by constructing a multi-layer attention model. Vector representations of the indicated part and the indicated part under the surrounding information and original conditions are calculated according to the attention mechanism, and whether the indicated relation exists or not is further calculated. The method is helpful for directly judging the reference relationship of the reference part and the referred part in the current context, and has better effect on both dominant reference and zero reference.
Drawings
FIG. 1 is a schematic diagram of an algorithm architecture for an inventive reference digestion method;
FIG. 2 is a diagram of a multi-headed attention mechanism model;
FIG. 3 is a schematic diagram of a digestion unit in accordance with the present invention.
Detailed Description
For further illustration of the various embodiments, the invention is provided with the accompanying drawings. The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate embodiments and together with the description, serve to explain the principles of the embodiments. With reference to these matters, one of ordinary skill in the art will understand other possible embodiments and advantages of the present invention. The components in the figures are not drawn to scale and like reference numerals are generally used to designate like components.
The invention will now be further described with reference to the drawings and detailed description.
Example 1
As shown in fig. 1 and fig. 2, the present invention provides a neural network-based long text reference digestion method, which introduces a deep learning technique into a pronouncing digestion task, thereby optimizing chinese pronouncing digestion and omitting completion of a recovery task.
The invention relates to a plurality of specialized words referring to digestion, which are described as follows:
(1) The word "back" (anapora) is in turn interpreted as an upward, forward or corresponding meaning and refers to the unit or meaning mentioned above (Crystal, 1985) by a term (transfer from Hu Zhuanglin, 1994: 48). Wherein, the term referred to in the foregoing is called antecedent (antecedent), and the term referred to is called return term or term (anaphor). Generally speaking, when something is mentioned in an utterance, it is necessary to make a second time, a form of back finger (anaphoric) is used to make the contexts correspond to each other (Chen Ping, 1987).
In many cases, the antecedent in the index is the same as the index, and both are in a co-index (coreferential) relationship. The main checking means is the return of the language, i.e. see if the look-ahead can be placed in place of the return item without changing meaning, and if the expression of the original sentence meaning is not affected, it can be judged to be the dominant return.
In explicit echo, the antecedent has obvious tracing property (tracking) when chapter appears, and echo item can be replaced by antecedent without changing sentence meaning. Whereas implicit echo refers to the reference to an implicit (immediate) antecedent. The implicit echo term does not necessarily refer to a specific word or a syntactic component in the speech, but it must be an entity that stands out in the psychological characterization established by the parties to the interaction from the utterance. In this case, the callback term recognizes only the meaning (identity of sense) associated with the preceding word, and the two terms are not identical. Therefore, in implicit echo, if the echo item is replaced by the antecedent item, the meaning will be changed.
Refer to partial information, and to a word. Pronouns refer to words replacing nouns, verbs, adjectives, quantity words, adverbs, including: alpha) human pronouns such as "I, you, he, we, zans, oneself, people", b) query pronouns such as "who, what, where, how many places, how many, c) indicates a pronoun such as" this, here, so, that, there, then, that, then "etc.
(2) Multi-head attention mechanism (Multi-head attention)
Multiple head attention is directed to using multiple queries to compute multiple information choices from input information in parallel. Each focusing on a different part of the input information.
The structure of the multi-headed attention model is shown in fig. 2, where the query Q, key K, and the value V are first linearly transformed (linear) and then input to the scaled dot product (scaled dot-produact attention), where h times are taken, the so-called multi-headed, and the heads are calculated one at a time. And the parameters W for the linear transformation of Q, K, V are different each time. Then, the h times of scaling point product position results are spliced (concat), and a value obtained by linear transformation (linear) is used as a multi-head position result.
The invention discloses a long text reference digestion method based on a neural network, which specifically comprises the following steps:
(1) Inputting long text, wherein the long text can be a sentence or a section of characters;
(2) Extracting the information of the indicated part, the information of the surrounding words of the indicated part, the original text information, the information of the indicated part and the information of the surrounding words of the indicated part of the long text, and carrying out vectorization representation; and embedding the indicating part information, the indicating part surrounding word information, the original text information, the indicated part information and the indicated part surrounding word information of the vectorization representation into position information respectively, and correspondingly outputting an indicating part vector, an indicating part surrounding information vector, an original text part vector, an indicated part vector and an indicated part surrounding information vector.
(3) And carrying out multi-head attention mechanism calculation on the index part vector and the information vector around the index part, and then enabling the obtained result to continue to adopt multi-head attention mechanism calculation with the original text part vector to obtain a first characteristic result representing the index part. The result can reflect the effect of the word information surrounding the original text and the reference part on the reference part under the condition of a multi-layer attention mechanism. More specifically, a Query (Q) with a part of vectors as a first multi-head attention model, a Key (K) and a Value (V) with a part of surrounding information vectors as keys, an output of the first multi-head attention model is used as a Query Q of a second multi-head attention model, an original part of vectors is used as a Key K and a Value V of the second multi-head attention model, and the second multi-head attention model outputs a first feature result.
(4) And performing multi-head attention mechanism calculation on the reference part vector, the surrounding information vector of the reference part and the original part vector in the same way to obtain a second characteristic result representing the reference part. More specifically, the designated partial vector is taken as a query Q of a third multi-head attention model, the designated partial surrounding information vector is taken as a key K and a value V, the output of the third multi-head attention model is taken as a query Q of a fourth multi-head attention model, the original partial vector is taken as the key K and the value V of the fourth multi-head attention model, and the fourth multi-head attention model outputs a second characteristic result.
(5) Finally, the first characteristic result and the second characteristic result are connected to form a comprehensive result, and the comprehensive result is mapped in a discrimination space by utilizing a Softmax layer to judge whether the reference is indicated or not.
Example 2
The first step: before training the model, training data is collected and data pre-processed.
And a second step of: the training data is transmitted to the steps (2) to (5) mentioned in embodiment 1, and since the training data is transmitted in the form of arrays after being transmitted to the model, the output result of each layer cannot be visually interpreted, and the above 4 steps (data processing, vectorization processing, array conversion and calculation of the neural network layer, etc. of the long text) are all passed from the processing of the long text to the training of the model or the model reasoning, the steps (2) to (5) can be divided into two parts of a model training module and a model reasoning module according to the model requirement. Generating a model file in a model training module; the model inference module generates an inference result.
And a third step of: after the training data of the first step is processed and the model of the second step is trained and stored, the data reasoning stage is entered.
Fourth step: and outputting an reasoning result.
The long text reference digestion method based on the neural network is an echo digestion method, and the processing of different layers of information of natural language is realized by constructing a multi-layer attention model. Vector representations of the indicated part and the indicated part under the surrounding information and original conditions are calculated according to the attention mechanism, and whether the indicated relation exists or not is further calculated. The method is helpful for directly judging the reference relationship of the reference part and the referred part in the current context, and has better effect on both dominant reference and zero reference.
As shown in fig. 3, the invention also discloses a long text reference digestion device based on a neural network, which comprises a long text extraction module, a connection layer, a Softmax layer and four multi-head attention models;
the long text extraction module is used for extracting the information of the indicated part, the information of the words around the indicated part, the original text information, the information of the indicated part and the information of the words around the indicated part from the long text and carrying out vectorization representation; embedding the indicating part information, the indicating part surrounding word information, the original text information, the indicated part information and the indicated part surrounding word information of the vectorization representation into position information respectively, and outputting an indicating part vector, an indicating part surrounding information vector, an original text part vector, an indicated part vector and an indicated part surrounding information vector;
the connection relation of the input and the output of the four multi-head attention models is as follows:
the input of the query of the first multi-head attention model refers to a partial vector, and the input of keys and values refers to a partial surrounding information vector; the output of the first multi-headed attention model is the input of a query of the second multi-headed attention model; the key and value inputs of the second multi-head attention model are original text partial vectors; outputting a first characteristic result by the second multi-head attention model;
the input of the query of the third multi-head attention model is a designated partial vector, the input of keys and values of the key and the value of the key and the value of the key are the designated partial vector, the output of the third multi-head attention model is the input of the query of the fourth multi-head attention model, the key and the value of the fourth multi-head attention model are the original partial vector, and the fourth multi-head attention model outputs a second characteristic result;
the connection layer is used for connecting the first characteristic result and the second characteristic result to form a comprehensive result;
the Softmax layer is used for mapping the comprehensive result in a discrimination space to judge whether the result is referred or not.
While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.
Claims (2)
1. A neural network-based long text reference digestion method, characterized in that: the method comprises the following steps:
step S1: inputting a long text;
step S2: extracting the information of the indicated part, the information of the surrounding words of the indicated part, the original text information, the information of the indicated part and the information of the surrounding words of the indicated part in the long text, and carrying out vectorization representation; embedding the indicating part information, the indicating part surrounding word information, the original text information, the indicated part information and the indicated part surrounding word information of the vectorization representation into position information respectively, and correspondingly outputting an indicating part vector, an indicating part surrounding information vector, an original text part vector, an indicated part vector and an indicated part surrounding information vector;
step S3: calculating the multi-head attention mechanism of the information vector around the pointing part and the pointing part, and then continuously calculating the obtained result and the original text part vector by adopting the multi-head attention mechanism to obtain a first characteristic result;
step S4: carrying out multi-head attention mechanism calculation on the pointed part vector and the information vector around the pointed part, and then, continuing to carry out multi-head attention mechanism calculation on the obtained result and the original part vector to obtain a second characteristic result;
step S5: the first characteristic result and the second characteristic result obtained in the step S3 and the step S4 are connected to form a comprehensive result, and the comprehensive result is mapped in a discrimination space by utilizing a Softmax layer to judge whether the reference is indicated or not;
the step S3 specifically includes: taking the reference partial vector as the query of the first multi-head attention model, taking the reference partial surrounding information vector as the key and the value, taking the output of the first multi-head attention model as the query of the second multi-head attention model, taking the original Wen Bufen vector as the key and the value of the second multi-head attention model, and taking the second multi-head attention model to output the first characteristic result;
the step S4 specifically includes: the method comprises the steps of taking a designated part vector as a query of a third multi-head attention model, taking a designated part surrounding information vector as a key and a value, taking output of the third multi-head attention model as a query of a fourth multi-head attention model, taking an original Wen Bufen vector as the key and the value of the fourth multi-head attention model, and taking the fourth multi-head attention model as an output of a second characteristic result.
2. A neural network-based long text reference resolution device, characterized in that: the system comprises a long text extraction module, a connection layer, a Softmax layer and four multi-head attention models;
the long text extraction module is used for extracting the information of the indicated part, the information of the words around the indicated part, the original text information, the information of the indicated part and the information of the words around the indicated part from the long text and carrying out vectorization representation; embedding the indicating part information, the indicating part surrounding word information, the original text information, the indicated part information and the indicated part surrounding word information of the vectorization representation into position information respectively, and outputting an indicating part vector, an indicating part surrounding information vector, an original text part vector, an indicated part vector and an indicated part surrounding information vector;
the connection relation of the input and the output of the four multi-head attention models is as follows:
the input of the query of the first multi-head attention model refers to a partial vector, and the input of keys and values refers to a partial surrounding information vector; the output of the first multi-headed attention model is the input of a query of the second multi-headed attention model; the key and value inputs of the second multi-head attention model are original text partial vectors; outputting a first characteristic result by the second multi-head attention model;
the input of the query of the third multi-head attention model is a designated partial vector, the input of keys and values of the key and the value of the key and the value of the key are the designated partial vector, the output of the third multi-head attention model is the input of the query of the fourth multi-head attention model, the key and the value of the fourth multi-head attention model are the original partial vector, and the fourth multi-head attention model outputs a second characteristic result;
the connection layer is used for connecting the first characteristic result and the second characteristic result to form a comprehensive result;
the Softmax layer is used for mapping the comprehensive result in a discrimination space to judge whether the result is referred or not.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011437239.7A CN112633014B (en) | 2020-12-11 | 2020-12-11 | Neural network-based long text reference digestion method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011437239.7A CN112633014B (en) | 2020-12-11 | 2020-12-11 | Neural network-based long text reference digestion method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112633014A CN112633014A (en) | 2021-04-09 |
CN112633014B true CN112633014B (en) | 2024-04-05 |
Family
ID=75309601
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011437239.7A Active CN112633014B (en) | 2020-12-11 | 2020-12-11 | Neural network-based long text reference digestion method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112633014B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020073664A1 (en) * | 2018-10-08 | 2020-04-16 | 平安科技(深圳)有限公司 | Anaphora resolution method and electronic device and computer-readable storage medium |
CN111401035A (en) * | 2020-02-18 | 2020-07-10 | 平安科技(深圳)有限公司 | Zero-reference resolution method, device, equipment and medium based on big data |
CN111428490A (en) * | 2020-01-17 | 2020-07-17 | 北京理工大学 | Reference resolution weak supervised learning method using language model |
CN111611361A (en) * | 2020-04-01 | 2020-09-01 | 西南电子技术研究所(中国电子科技集团公司第十研究所) | Intelligent reading, understanding, question answering system of extraction type machine |
-
2020
- 2020-12-11 CN CN202011437239.7A patent/CN112633014B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020073664A1 (en) * | 2018-10-08 | 2020-04-16 | 平安科技(深圳)有限公司 | Anaphora resolution method and electronic device and computer-readable storage medium |
CN111428490A (en) * | 2020-01-17 | 2020-07-17 | 北京理工大学 | Reference resolution weak supervised learning method using language model |
CN111401035A (en) * | 2020-02-18 | 2020-07-10 | 平安科技(深圳)有限公司 | Zero-reference resolution method, device, equipment and medium based on big data |
CN111611361A (en) * | 2020-04-01 | 2020-09-01 | 西南电子技术研究所(中国电子科技集团公司第十研究所) | Intelligent reading, understanding, question answering system of extraction type machine |
Also Published As
Publication number | Publication date |
---|---|
CN112633014A (en) | 2021-04-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11586829B2 (en) | Natural language text generation from a set of keywords using machine learning and templates | |
Adelia et al. | Indonesian abstractive text summarization using bidirectional gated recurrent unit | |
CN111930942B (en) | Text classification method, language model training method, device and equipment | |
Pala et al. | DESAM—annotated corpus for Czech | |
Xie et al. | Topic enhanced deep structured semantic models for knowledge base question answering | |
CN111414481A (en) | Chinese semantic matching method based on pinyin and BERT embedding | |
CN112232053A (en) | Text similarity calculation system, method and storage medium based on multi-keyword pair matching | |
Vinnarasu et al. | Speech to text conversion and summarization for effective understanding and documentation | |
Nguyen et al. | Language-oriented Sentiment Analysis based on the Grammar Structure and Improved Self-attention Network. | |
Yu et al. | Make it directly: event extraction based on tree-LSTM and Bi-GRU | |
Wei et al. | Enhance understanding and reasoning ability for image captioning | |
CN110781666A (en) | Natural language processing text modeling based on generative countermeasure networks | |
CN112633014B (en) | Neural network-based long text reference digestion method and device | |
CN112579739A (en) | Reading understanding method based on ELMo embedding and gating self-attention mechanism | |
CN116956925A (en) | Electronic medical record named entity identification method and device, electronic equipment and storage medium | |
Hu et al. | Nlire: A natural language inference method for relation extraction | |
CN116414988A (en) | Graph convolution aspect emotion classification method and system based on dependency relation enhancement | |
Keezhatta | Understanding EFL Linguistic Models through Relationship between Natural Language Processing and Artificial Intelligence Applications. | |
Duong et al. | Leveraging semantic representations combined with contextual word representations for recognizing textual entailment in vietnamese | |
Lv et al. | StyleBERT: Chinese pretraining by font style information | |
Pisat et al. | Synonym Suggestion System using Word Embeddings | |
CN113569124A (en) | Medical title matching method, device, equipment and storage medium | |
Jeyasheeli et al. | Sentence Generation for Indian Sign Language Using NLP. | |
Chandarana et al. | Natural Language Sentence to SQL Query Converter | |
Kaur et al. | Text Generator using Natural Language Processing Methods |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |