CN112464673B - Language meaning understanding method for fusing meaning original information - Google Patents

Language meaning understanding method for fusing meaning original information Download PDF

Info

Publication number
CN112464673B
CN112464673B CN202011431776.0A CN202011431776A CN112464673B CN 112464673 B CN112464673 B CN 112464673B CN 202011431776 A CN202011431776 A CN 202011431776A CN 112464673 B CN112464673 B CN 112464673B
Authority
CN
China
Prior art keywords
sense
word
original
decoder
steps
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011431776.0A
Other languages
Chinese (zh)
Other versions
CN112464673A (en
Inventor
王念滨
汪先慈
张耘
周连科
王红滨
张毅
崔琎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Engineering University
Original Assignee
Harbin Engineering University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Engineering University filed Critical Harbin Engineering University
Priority to CN202011431776.0A priority Critical patent/CN112464673B/en
Publication of CN112464673A publication Critical patent/CN112464673A/en
Application granted granted Critical
Publication of CN112464673B publication Critical patent/CN112464673B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Machine Translation (AREA)

Abstract

Language meaning understanding method of fusion sense original information, belonging to languageThe technical field of information processing. The method aims to solve the problems that the existing language modeling method is high in complexity and cannot achieve effects. Firstly, processing a language by taking each word as a unit according to two paths; left path: word encoder+rnn+word decoder, left path output is noted as w l The method comprises the steps of carrying out a first treatment on the surface of the Right path: sense original encoder+RNN+sense original decoder+word decoder+sigmoid, right path output is denoted as w r The method comprises the steps of carrying out a first treatment on the surface of the The outputs of the two paths are then fused. Mainly for language meaning understanding.

Description

Language meaning understanding method for fusing meaning original information
Technical Field
The invention relates to a language meaning understanding method, and belongs to the technical field of language information processing.
Background
Language Modeling (LM) is a core task for Natural Language Processing (NLP) and language understanding. The purpose of language modeling is to show the probability distribution of a word sequence. LMs play a key role in text generation tasks, such as machine translation, document summarization.
The model always predicts words based on context. In a simple N-gram language model, it assumes: each word is only N-1 words relative to the previous word. In recent years, more and more neural network models have been built. They have obtained the most advanced performance and are still making progress. The classical neural network language model consists of an encoder and a decoder. In the encoder, the NN receives a sequence of languages and then encodes the left context of each word into a vector. In the decoder, the neural network takes the vector from the encoder and attempts to predict the word. The invention refers to all words to the left of a word as "left context".
These LMs rely on Recurrent Neural Networks (RNNs) or other modern neural networks to learn the context of words. Then, a similar decoder is used to obtain the probability distribution of the word. As computing power increases, many have focused on creating a complex structure that can better understand context. These efforts play a role, but more effort should be put into how the context is utilized.
Some expert annotated data helps to model the context and increase the interpretability of the context. HowNet is annotation data that annotates each concept in chinese with one or more related origins, revealing relationships between origins. The source of meaning is a semantic unit that is indivisible to human language defined by linguists. HowNet contains definitions, emotions, examples, etc. of a word, but only definitions are always filled in. The lower layer is based on a more accurate definition of the upper layer. Except for the first layer, there is only one supersense source for each sense source. The invention considers that HowNet can be fused into a neural network, and brings about obvious improvement.
In recent years, some have done much work on knowledge networks, such as emotion analysis and word similarity calculation. The present invention has seen that some works have good results, such as sense-driven language model (SDLM) and token learning. These works are instructive to the present invention. They use the sense original information for language modeling, attention, and sense original prediction. However, compared with the quality and cost of the known network, the method is not advanced enough, is complex and is difficult to popularize.
Disclosure of Invention
The invention aims to solve the problems that the existing language modeling method has higher complexity and cannot give consideration to effects.
The language meaning understanding method of the fusion meaning original information comprises the following steps:
processing the language by taking each word as a unit according to two paths; left path: word encoder+rnn+word decoder, left path output is noted as w l The method comprises the steps of carrying out a first treatment on the surface of the Right path: sense original encoder+RNN+sense original decoder+word decoder+sigmoid, right path output is denoted as w r The method comprises the steps of carrying out a first treatment on the surface of the Then fusing the outputs of the two paths;
the processing procedure of the original encoder comprises the following steps:
creating a trainable matrix having N rows, where N is the number of words, i represents the sequence number of the word, and corresponding to the ith row of the trainable matrix, taking each sense element as a vector:
s=select(w,M)
emb=∑s j *EMB(j)
where M is a constant matrix of sense primitive weights for words,
Figure BDA0002826756970000021
h1 and H2 represent the number of words and the number of sense origins; select is a function of extracting the sense-original weight vector represented by word w from M, s is an s j A composed sense origin weight vector representing the probability distribution of sense origins contained in the word; />
Figure BDA0002826756970000022
Weights representing the sense origins contained in a word, EMB is a trainable embedding layer that converts each sense origin into a corresponding vector;
the processing procedure of the word decoder comprises the following steps:
according to the base Yu Yiyuan decoder, the sense original probability needs to be converted into words:
w=sM T
wherein M is T Is the transpose of M;
the process of fusing the outputs of the two paths comprises the following steps:
according to the output w of the left path l And the output w of the right path r Obtaining fused output:
p(w|g)=softmax(w l w r )
where p (w|g) represents the probability of word w with respect to context;
further, the weight of the sense source contained in the word
Figure BDA0002826756970000023
Wherein word w has l at the h-th level in the sense original tree h A sense source, j represents one of them;
further, when the number of layers in the sense tree is 2 or more,
Figure BDA0002826756970000024
g' the total number of layers of the original tree;
further, the processing procedure of the sense original decoder comprises the following steps:
the output of RNN in the right path is first processed as follows:
q=σ(g T V+b)
wherein V and b are trainable parameters, σ represents an activation function, g is the output of the RNN;
then taking the hidden state q as input to calculate the sense original score; for the sense origin k in vector s, the sense origin is scored as:
s k =q T U k
wherein U is k Is a conversion matrix;
further, the conversion matrix U k The following are provided:
Figure BDA0002826756970000031
wherein the method comprises the steps of
Figure BDA0002826756970000032
α k,r >0 is a trainable parameter and +.>
Figure BDA0002826756970000033
Further, the sigmoid processing procedure in the right path includes the following steps:
firstly, aiming at the output of a word decoder, processing by using a Sigmoid function to obtain w r
Then pair w using the following formula r Updating:
w r =w r ×(1-X)+X
where X determines a constant parameter of the offset.
The beneficial effects are that:
the invention provides a new data expansion method which is realized by taking a meaning source as an additional input. Thus, for the input of one word, there are two types of inputs, one is the original word and the other is the sense source. The two types of inputs are modeled by the same intermediate layer, and at the last layer of the model, the outputs of the two paths are fused together. This is a simple generation method that fuses the sense primitive knowledge with the neural network. The present invention refers to this approach as data enhancement (SBDE) of the base Yu Yi source. Thus, the integration of SBDE as a model may make the model more robust. The effectiveness of SBDE was demonstrated by language modeling experiments and by downstream application experiments of title generation.
Drawings
FIG. 1 is a schematic diagram of word and source relation;
FIG. 2 is a sense-original generation model diagram;
FIG. 3 is a schematic diagram of a language model of the present invention;
fig. 4 is a diagram of an example of an artificial original embedding layer.
Detailed Description
The first embodiment is as follows:
the meaning sources are unique concepts in HowNet, and the design purpose is to represent basic semantic information by using a fixed part of words, so that any other words can be represented by the meaning sources. These representations are defined by the authors of HowNet, are numerous in terms of vocabulary, are structurally rich, and contain a large amount of human experience information. However, as a database defined by people, limited by the increase of expert level and knowledge amount, there are inevitably some subjectivity and timeliness related problems. Although the errors caused by these problems may not be obvious for the current deep learning method, the fusion of the problems with the deep learning method can complement each other, and combine the advantages of different methods.
The meaning source is designed and constructed by the father and son persons Dong Zhendong and Dong Jiang, and takes concepts (Sense) represented by words (Word) of English and Chinese as description objects to explain the relation between concepts and attributes of the concepts as an attempt knowledge base of basic content. The idea is mainly from the theory of reduction, i.e. any meaning of a word can consist of one or more semantic units, which are called sense origins (sememes) in the knowledge network. The meaning sources here are the smallest, indivisible semantic units, which of course not only illustrate the meaning sources comprised by the concept, but also there are relations and structures between these meaning sources, which as a whole represent a tree shape. The nodes of the tree represent the sense sources, while the edges of the tree represent the relationships between the sense sources, such as modifications, compositions, etc.
The source of meaning was pursued by researchers of various traditional methods in the early years, but with the sinking of traditional statistical methods, the source of meaning appears in papers less and less frequently. In the current era of deep learning model flooding, researchers mostly pursue much larger amounts of data and parameters than expert fine-labeling data. In the present invention, it is desirable to fuse such data with the neural network model. Since the sense origins are complicated and various in information, and it is difficult to form a matrix format commonly used in a neural network, in the present invention, only the hierarchical relationship between sense origins and sense origins in HowNet is extracted and no modifications between sense origins are recorded in order to reduce the complexity of the model. And since in written language words of different semantics may contain the same sense origin, the invention chooses to take only the highest one in the tree.
As shown in fig. 1, a word may contain multiple semantics, for each of which a relationship diagram between different origins is designed according to its semantics. For these relationships to appear tree-like, the semantics of the deeper hierarchical representation tend to be more subdivided.
Language models are the basis for various NLP tasks. Modern language models typically analyze the probability distribution of words in a given context. When a statistical language model is presented, the model assigns probabilities to a sequence S with N words, as shown in the following formula:
P(s)=P(w 1 w 2 …w N )=P(w 1 w 2 …w N-1 )P(w N |w 1 w 2 …w N-1 )=P(w 1 )P(w 2 |w 1 )…P(w N |w 1 w 2 …w N-1
wherein w is i Representing the ith word in the sequence s, the joint probability of the N words can be decomposed into the product of the joint probability of the N-1 words and the conditional probability of the nth word.
With the popularity of neural networks, mikolov et al propose for the first time that RNNs are applied to language models. They rely on recursive structures to capture longer context information and train through back propagation of time. While RNN can theoretically fit any function, it is difficult to learn very long dependencies. Some reasons are gradient explosion or vanishing problems. To solve this problem, a model was developed and many regularization and optimization methods were introduced. AWD-LSTM is an outstanding model that has reached the most advanced level on the Pennsylvania Tree Bank (PTB) and WikiText-2 data sets.
Recently, however, the effectiveness of pre-trained language models has been remarkable, and has led to research interest. These pre-trained models often contain many parameters and are then trained on a vast corpus.
With the development of deep learning, it is desirable to train a network with a large corpus. In recent years, the importance of expert manual annotation knowledge base is more and more paid attention to, and some research results on sense origin prediction are also presented. Niu et al use the word's sense primitive information to improve the quality of word embedding and accurately capture the meaning of the word in a particular context. Ruobin Xie uses pre-training word embedding and matrix decomposition to predict the meaning source of the word, and Jin and the like further consider Chinese characters and position information on the basis.
There has also been a prior attempt to combine a sense original and a language model, model name SDLM (SememeDrive Language Model), SDLM, which is a valuable attempt to fuse sense original information into a neural network, proving it to be beneficial to the language model. These inputs and encoders are the same as the usual language model, except that the decoder is changed. As shown in fig. 2, the SDLM uses two paths to generate semantic vectors and calculates by a constant matrix of semantics. While successful, the encoder of the SDLM receives a word as input, the decoder needs to generate a context of the sense origin. Thus, the encoder is to capture not only the context of the word, but also the context of the sense origin. But there is a problem here: the encoder cannot directly obtain the information of the sense origin. To make up for the gap between encoder and decoder, the present invention attempts to solve this problem by the following method, and shows different research perspectives in this field.
The language meaning understanding method of the fusion sense original information according to the embodiment comprises the following steps:
generating a language based on a sense source:
in this section, the present invention proposes a simple method to use the semblance information to refine the results of the NLP model. Some words are not candidates if their sense origins are too far from the context. This method will be described in detail in the subsequent experiments and examples of the present invention. In a general language model, an input sequence is fed into a neural network and the next word is predicted from its preceding word. As a description of the knowledge network, each word may be defined as a structured meaning source. Based on the theory, the invention can take the meaning source of the context as input, and further predict the probability of the meaning source contained in the current word.
In fact, the main efficiency of the known network is to provide a simpler model, which helps the model learn it easily. Since the sense origins are much smaller in space than words, different words may represent the same sense origins. There are two sentences: "apple is very good at" and "orange is very good at". Obviously, learning "fruit good-eating" is much easier than learning "apple good-eating" and "orange good-eating" respectively, so that neural networks can benefit from the specific information of the sense origin.
The method proposed by the invention is to establish a parallel path from input to output. If details of the encoder and decoder are ignored, the parallel path is identical to the other path. Word in encoder encodingConverted into a sense source and a coded vector is derived therefrom. In the decoder there is a new path to determine which sense in the current word is not present. The language model of the present invention is shown in fig. 3, and the present invention includes two paths, a left path: word encoder (abbreviated emb) +rnn+ word decoder (word decoder), left path output is denoted w l The method comprises the steps of carrying out a first treatment on the surface of the Right path: sense original encoder (sememe emudding, abbreviated sememe emub) +RNN+sense original decoder (sememe decoder) +word decoder) +sigmoid, and right path output is denoted as w r The method comprises the steps of carrying out a first treatment on the surface of the The outputs of the two paths are then fused (MUL).
An Embedding part:
in the original LM (language model), embedding is a step of taking words as vectors. Because vectors can obtain more information than a single number. In the left path, most models of the NLP region design an embedded component. The embedding is accomplished by creating a trainable matrix having N rows, where N is the number of words and i represents the sequence number of the word, corresponding to the ith row of the trainable matrix, as with the original LM, the right path of the model of the present invention takes each sense as a vector. The method comprises the following steps:
s=select(w,M)
emb=∑s j *EMB(j)
where M is a constant matrix of sense primitive weights for words,
Figure BDA0002826756970000061
h1 and H2 represent the number of words and the number of sense origins. Select is a function of extracting the sense-original weight vector represented by word w from M, s is an s j A composed sense primitive weight vector representing the probability distribution of sense primitives contained within the word, where each s j The weight of the sense origin is indicated.
Figure BDA0002826756970000064
The EMB, representing the weight of the sense sources contained in a word, is a trainable embedding layer that converts each sense source into a corresponding vector.
Obviously, these wordsThe sense sources play a different role in the context, with some important and some unimportant. Thus, if the word w has l at the h-th level in the sense original tree h The sense origin, j is one of them, use
Figure BDA0002826756970000062
To distinguish between the differences between these different layer origins. However, since the meaning sources are mostly located in the first layer and the second layer, the present invention regards all meaning sources of h > 2 as h=2. In other words, when the number of layers is 2 or more, the formula is +.>
Figure BDA0002826756970000063
g' means the total number of layers of the original tree. The above formula can also be represented as fig. 4, from which the meaning source and hierarchy information of each word can be obtained.
For each sense element, the invention can do the same sense element embedding operation as the word embedding.
RNN part:
in the right path, the present invention uses the same intermediate layer (RNN) between the encoder and decoder as the left path. The right path of the invention is completely independent of the left path, so that the performance of the original network is not affected.
A Decoder section:
the function of the decoder is to convert the hidden representation into word probabilities. The structure of the knowledge network does not contain hidden rules which are not known or written out by human expert. But such rules contain word-to-word inherent relationships and can be used to eliminate word uncertainty.
1) Original meaning decoder
The present invention designs the sense original decoder in two steps, the first step is to decode the output of the intermediate layer and the second step is to select a parameterized matrix for each sense original k.
The first step: a linear decoder is designed with an activation function:
q=σ(g T V+b)
where V and b are trainable parameters, σ represents the activation function, and g is the output of the RNN.
And a second step of: the hidden state q is used as input to calculate the sense origin score. The invention uses a unique linear function to describe the probability of each source. For the sense origin k in vector s, the sense origin score is:
s k =q T U k
wherein U is k Is a conversion matrix;
the invention regards each sense primitive as an expert and predicts probability distributions according to context, which reflect the prediction of sense primitive probabilities of all other words to the sense primitive of the current word. The sense origin decoder portion calculates the sense origin probability of the word.
2) Word decoder
The final output of the language model is the word probability. Therefore, the sense original probability needs to be converted into words:
w=sM T
where s is the probability of the sense origins, each sense origin being associated with a plurality of words. M is M T Is of (1)
Figure BDA0002826756970000071
And represent the sense-original probability distribution of a word. As described above, M T There are H2 rows, each with H1 numbers, representing the probability defined by HowNet. When the invention obtains the probability of the source meaning and the relation matrix of the source meaning and the word, the sum of the probability distribution of the word can be calculated by s.
Fusion:
although the net is known to be an excellent word dataset that annotates most of the meaning of words, some information may be lost or outdated so the present invention attempts to propose a fusion method to avoid the above-mentioned problems. The invention merges the original neural network (left path) and the additional path (right path). The invention selects multiplication and makes some corrections.
For the above reasons, the present invention treats the language model as a multi-label classification task. Thus, the present invention uses a sigmoid function on the output of the word decoder in the right path and refers to its output as w r . At the same time, know the netOn the basis, w can be utilized r Some words with low likelihood are pruned.
In the left path, w l Is obtained by the fully connected part of the standard word decoder; in the right path, w r Derived from the last sigmoid function, finally using w r Multiplying w l The effect of pruning the left path result can be achieved.
Thus, overall the model output principle is p (w|g), i.e. the probability distribution of word w with respect to context g, can be derived from the results of the left and right paths:
p(w|g)=softmax(w l w r )
additional details:
sigmoid: since the scope of the Sigmoid function is (0, 1). In a given context, there will always be more words that are irrelevant than there are words that are relevant, these words letting w l w r Becomes w l 0. Therefore, it is very difficult to obtain a counter-propagating gradient for the left path, and it is very difficult to train better with the model, many times. The invention herein uses a sigmoid function and modifies the output of the sigmoid (the whole can be regarded as a modified sigmoid function), modifying its range to (X, 1); the final formula is w r X (1-X) +x, where X determines the constant parameter of the offset.
Basis matrix: a conversion matrix U is arranged for each sense element k k Is extremely space consuming, a skill is designed here to achieve this. Modeling U using R matrices and their weighted sums k The formula is as follows:
Figure BDA0002826756970000081
wherein the method comprises the steps of
Figure BDA0002826756970000082
α k,r >0 is a trainable parameter and +.>
Figure BDA0002826756970000083
/>
The present invention is capable of other and further embodiments and its several details are capable of modification and variation in light of the present invention, as will be apparent to those skilled in the art, without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (6)

1. The language meaning understanding method of the fusion meaning original information is characterized by comprising the following steps of:
processing the language by taking each word as a unit according to two paths; left path: word encoder+rnn+word decoder, left path output is noted as w l The method comprises the steps of carrying out a first treatment on the surface of the Right path: sense original encoder+RNN+sense original decoder+word decoder+sigmoid, right path output is denoted as w r The method comprises the steps of carrying out a first treatment on the surface of the Then fusing the outputs of the two paths;
the processing procedure of the original encoder comprises the following steps:
creating a trainable matrix having N rows, where N is the number of words, i represents the sequence number of the word, and corresponding to the ith row of the trainable matrix, taking each sense element as a vector:
s=select(w,M)
emb=∑s j *EMB(j)
where M is a constant matrix of sense primitive weights for words,
Figure FDA0002826756960000011
h1 H2 represents the number of words and the number of sense origins; select is a function of extracting the sense-original weight vector represented by word w from M, s is an s j A composed sense origin weight vector representing the probability distribution of sense origins contained in the word; />
Figure FDA0002826756960000014
Weights representing the sense origins contained in a word, EMB is a trainable embedding layer that converts each sense origin into a corresponding vector;
the processing procedure of the word decoder comprises the following steps:
according to the base Yu Yiyuan decoder, the sense original probability needs to be converted into words:
w=sM T
wherein M is T Is the transpose of M;
the process of fusing the outputs of the two paths comprises the following steps:
according to the output w of the left path l And the output w of the right path r Obtaining fused output:
p(w|g)=softmax(w l w r )
where p (w|g) represents the probability of word w with respect to context.
2. The method for understanding language meaning of merged sense original information according to claim 1, wherein the one word contains weight of sense original
Figure FDA0002826756960000012
Wherein word w has l at the h-th level in the sense original tree h The meaning source, j, represents one of them.
3. The language meaning understanding method of merging sense original information according to claim 2, wherein when the number of layers in the sense original tree is 2 or more,
Figure FDA0002826756960000013
g' means the total number of layers of the original tree.
4. A method for understanding language meaning of fused sense original information according to claim 1, 2 or 3, wherein the process of the sense original decoder comprises the steps of:
the output of RNN in the right path is first processed as follows:
q=σ(g T V+b)
wherein V and b are trainable parameters, σ represents an activation function, g is the output of the RNN;
then taking the hidden state q as input to calculate the sense original score; for the sense origin k in vector s, the sense origin is scored as:
s k =q T U k
wherein U is k Is a conversion matrix.
5. The method for understanding language meaning of fusion of raw information according to claim 4, wherein the transformation matrix U k The following are provided:
Figure FDA0002826756960000021
wherein the method comprises the steps of
Figure FDA0002826756960000022
α k,r > 0 is a trainable parameter and +.>
Figure FDA0002826756960000023
6. The language meaning understanding method of fusion of raw information according to claim 5, wherein the sigmoid processing procedure in the right path comprises the steps of:
firstly, aiming at the output of a word decoder, processing by using a Sigmoid function to obtain w r
Then pair w using the following formula r Updating:
w r =w r ×(1-X)+X
where X determines a constant parameter of the offset.
CN202011431776.0A 2020-12-09 2020-12-09 Language meaning understanding method for fusing meaning original information Active CN112464673B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011431776.0A CN112464673B (en) 2020-12-09 2020-12-09 Language meaning understanding method for fusing meaning original information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011431776.0A CN112464673B (en) 2020-12-09 2020-12-09 Language meaning understanding method for fusing meaning original information

Publications (2)

Publication Number Publication Date
CN112464673A CN112464673A (en) 2021-03-09
CN112464673B true CN112464673B (en) 2023-05-26

Family

ID=74801137

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011431776.0A Active CN112464673B (en) 2020-12-09 2020-12-09 Language meaning understanding method for fusing meaning original information

Country Status (1)

Country Link
CN (1) CN112464673B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113468884B (en) * 2021-06-10 2023-06-16 北京信息科技大学 Chinese event trigger word extraction method and device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109446518A (en) * 2018-10-09 2019-03-08 清华大学 The coding/decoding method and decoder of language model
CN109597988A (en) * 2018-10-31 2019-04-09 清华大学 The former prediction technique of vocabulary justice, device and electronic equipment across language
CN110555203A (en) * 2018-05-31 2019-12-10 北京百度网讯科技有限公司 Text replying method, device, server and storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180107940A1 (en) * 2010-04-27 2018-04-19 Jeremy Lieberman Artificial intelligence method and apparatus

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110555203A (en) * 2018-05-31 2019-12-10 北京百度网讯科技有限公司 Text replying method, device, server and storage medium
CN109446518A (en) * 2018-10-09 2019-03-08 清华大学 The coding/decoding method and decoder of language model
CN109597988A (en) * 2018-10-31 2019-04-09 清华大学 The former prediction technique of vocabulary justice, device and electronic equipment across language

Also Published As

Publication number Publication date
CN112464673A (en) 2021-03-09

Similar Documents

Publication Publication Date Title
CN110717334B (en) Text emotion analysis method based on BERT model and double-channel attention
CN110134771B (en) Implementation method of multi-attention-machine-based fusion network question-answering system
CN111177394B (en) Knowledge map relation data classification method based on syntactic attention neural network
CN110795556B (en) Abstract generation method based on fine-grained plug-in decoding
CN108829722A (en) A kind of Dual-Attention relationship classification method and system of remote supervisory
CN111858932A (en) Multiple-feature Chinese and English emotion classification method and system based on Transformer
CN110929030A (en) Text abstract and emotion classification combined training method
CN109086270B (en) Automatic poetry making system and method based on ancient poetry corpus vectorization
CN111125333B (en) Generation type knowledge question-answering method based on expression learning and multi-layer covering mechanism
CN110688862A (en) Mongolian-Chinese inter-translation method based on transfer learning
CN114881042B (en) Chinese emotion analysis method based on graph-convolution network fusion of syntactic dependency and part of speech
CN110688861A (en) Multi-feature fusion sentence-level translation quality estimation method
CN115422939B (en) Fine granularity commodity named entity identification method based on big data
CN113657123A (en) Mongolian aspect level emotion analysis method based on target template guidance and relation head coding
CN112784604A (en) Entity linking method based on entity boundary network
CN114880461A (en) Chinese news text summarization method combining contrast learning and pre-training technology
CN114969304A (en) Case public opinion multi-document generation type abstract method based on element graph attention
CN115329088B (en) Robustness analysis method of graph neural network event detection model
CN114757184B (en) Method and system for realizing knowledge question and answer in aviation field
CN114254645A (en) Artificial intelligence auxiliary writing system
CN114972848A (en) Image semantic understanding and text generation based on fine-grained visual information control network
CN115496072A (en) Relation extraction method based on comparison learning
CN115238691A (en) Knowledge fusion based embedded multi-intention recognition and slot filling model
CN112464673B (en) Language meaning understanding method for fusing meaning original information
CN114781356B (en) Text abstract generation method based on input sharing

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant