CN114218928A - Abstract text summarization method based on graph knowledge and theme perception - Google Patents
Abstract text summarization method based on graph knowledge and theme perception Download PDFInfo
- Publication number
- CN114218928A CN114218928A CN202111654105.5A CN202111654105A CN114218928A CN 114218928 A CN114218928 A CN 114218928A CN 202111654105 A CN202111654105 A CN 202111654105A CN 114218928 A CN114218928 A CN 114218928A
- Authority
- CN
- China
- Prior art keywords
- topic
- sentence
- model
- graph
- document
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/253—Grammatical analysis; Style critique
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/211—Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/216—Parsing using statistical methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Machine Translation (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses an abstract text summarization method based on graph knowledge and topic perception. The invention provides a model based on BERT, a neural topic model and a graph neural network, which is called GTASum. At the input end of a document encoder, using BERT to obtain a hidden word vector of a document; at the input of the topic encoder, obtaining a word-topic distribution vector of the document using a neural topic model; inputting the two vectors into a graph neural network for training to obtain context content fused with topic knowledge, and generating a text by a transform-based decoder; meanwhile, the LN layer with the normalized condition can cooperatively train a neural topic model and a decoder, and effectively select the characteristics. The result shows that the method has better robustness and adaptive capacity.
Description
Technical Field
The invention belongs to the technical field of natural language processing, and relates to a text abstract generation method, in particular to an abstract text abstract method based on graph knowledge and topic perception, and a text abstract generation method based on a pre-training language model, a neural topic model and a graph neural network.
Background
With the development of computer performance and large-scale language models, natural language processing tasks (NLPs) have made significant development. The abstract task is one of core problems of natural language processing tasks and aims to enable people to quickly master important information in texts. Text summaries have been widely used in a number of areas such as news, finance, conferencing, and medicine. At present, there are two main methods for the summarization task: an abstraction method and an abstract method. The extraction method mainly copies important information from the original text and then aggregates them into text. The abstract generated by the method usually retains the significant information of the source text and has correct grammar, but inevitably generates a large amount of redundant information easily. Abstract methods are to form a summary based on an understanding of the source text input. It tries to understand the text content, can generate words not in the original text, more closely approaches the essence of the abstract, and has the potential of generating high-quality abstract. To sum up, in order to better generate news summaries to help readers quickly and efficiently grasp daily messages, the present study focuses on abstract summaries.
In abstract abstracts, sequence-to-sequence has become the dominant framework for a variety of architectures. In the early days, the abstracted text excerpts were primarily RNN-based codecs. Because of the "long range dependency" problem, the RNN has lost much information by the time the word is entered at the last time step, and bahdana et al apply the Attention mechanism to NLP. Since then, text summarization has advanced greatly, and various applications have emerged endlessly. Among these, the most notable sibling transform architecture. This model has achieved surprising performance in many areas. Current SOTA abstraction models, including BART, PEGASUS, and ProphetNet, all employ a transform-based architecture. Also, with the contribution of attention mechanism, the transform-based model can capture well the grammatical and contextual information between tokens. However, higher levels of semantic understanding do not perform well.
To address this problem, researchers have tried a variety of improvements, one of which is the way of exploiting topic perception. Topic models such as LDA, PFA, NVDM, NTM can all provide additional information for document understanding. For text summarization, we believe that it can improve model performance by incorporating the topic model properties into the summarization model. Furthermore, in recent years, Graph Neural Networks (GNNs) have been widely used for cross-sentence relational modeling of summarization tasks. Some studies have created document graphs based on linguistic analysis. However, this approach relies on external tools, possibly leading to semantically fragmented outputs. Wang and Liu et al have constructed word and sentence document graphs, but this approach has difficulty capturing semantic level relationships, and therefore, how to efficiently construct documents into abstractable graphs is also a difficult problem.
Disclosure of Invention
The invention aims to provide an abstract text summarization method based on graph knowledge and theme perception, aiming at the defects of the prior art.
The technical scheme adopted by the invention for solving the technical problem comprises the following steps:
step (1): given an original input document D, a [ CLS ] is inserted into the beginning and end of each sentence in the original input document D]And [ SEP ]]Then, the processed input document D is put into a pre-training language model BERT, and the feature representation H of the sentence is learnedB;
Step (2): inputting original input document D into neural topic model NTM, learning topic representation H of documentT;
And (3): representing the characteristics of the sentence HBAnd a topic representation H of the documentTInputting the information into a destination attention network GAT and initializing; generating sentence characteristics h' with topic information after the attention network GAT coding; the GAT coding process is to construct a heterogeneous document graph with topics and sentences, and continuously update the feature representation HBAnd a topic representation HTA constructed node representation;
and (4): sending the sentence characteristics h' with the topic information into a Transformer-based decoder for decoding; generating a text abstract after normalization;
and (5): training the GTASum model on the CNN/DailyMail data set and the XSum data set, selecting the optimal GTASum model, inputting any text into the trained GTASum model, and outputting corresponding abstract contents;
the GTASum model is composed of a pre-training language model BERT, a neural topic model NTM, a graph attention network GAT and a decoder.
The invention has the following beneficial effects:
the invention provides a topic model and a graph neural network which can help a pre-training language model to better speak. During training, the document theme information and the document embedded information are sent to a graph neural network for fusion and updating. Through the operation, when the pre-training language model faces downstream tasks, feature selection can be carried out by better referring to subject information. The results show that this approach has better robustness and adaptability. And testing according to standard performance indexes in the field of text summarization, wherein the evaluation indexes comprise ROUGE _1/ROUGE _2/ROUGE _ L. Tests were performed on news text summary datasets CNN/DailyMail and XSum and yielded results that were of a leading level in the field.
Drawings
FIG. 1 is the overall flow framework of the model;
FIG. 2 is a neural topic model portion framework;
FIG. 3 is an example of the present model usage;
Detailed Description
The invention is further illustrated by the following figures and examples.
The invention provides an abstract text summarization method based on graph knowledge and theme perception. First, we encode the input document with a pre-trained language model BERT to learn the contextual sentence representation while using the Neural Topic Model (NTM) to find potential topics. Then, we build a heterogeneous document graph composed of sentences and topic nodes and update the representation by using a modified graph attention network (GAT). Thirdly, the expression form of the sentence node is obtained, and the latent semantic meaning is calculated. Finally, the latent semantics are fed into a transform-based decoder for decoding to generate a final result. We performed extensive experiments on two real-world datasets CNN/DailyMail and XSum.
The model based on the BERT, the neural topic model and the graph neural network is called GTASum. At the input end of a document encoder, using BERT to obtain a hidden word vector of a document; at the input of the topic encoder, obtaining a word-topic distribution vector of the document using a neural topic model; inputting the two vectors into a graph neural network for training to obtain context content fused with topic knowledge, and generating a text by a transform-based decoder; meanwhile, the LN layer with the normalized condition can cooperatively train a neural topic model and a decoder, and effectively select the characteristics. The result shows that the model can well capture the key information of the abstract text abstract and has better robustness and adaptive capacity.
As shown in fig. 1 and 2, an abstract text summarization method based on graph knowledge and topic perception includes the following steps:
step (1): given an original input document D, a [ CLS ] is inserted into the beginning and end of each sentence in the original input document D]And [ SEP ]]Then, the processed input document D is put into a pre-training language model BERT, and the feature representation H of the sentence is learnedB;
Step (2): inputting original input document D into neural topic model NTM, learning topic representation H of documentT;
And (3): representing the characteristics of the sentence HBAnd a topic representation H of the documentTInputting the information into a destination attention network GAT and initializing; generating sentence characteristics h' with topic information after the attention network GAT coding; the GAT coding process is to construct a heterogeneous document graph with topics and sentences, and continuously update the feature representation HBAnd a topic representation HTA constructed node representation;
and (4): sending the sentence characteristics h' with the topic information into a Transformer-based decoder for decoding; generating a text abstract after normalization;
and (5): training the GTASum model on the CNN/DailyMail data set and the XSum data set, selecting the optimal GTASum model, inputting any text into the trained GTASum model, and outputting corresponding abstract contents;
the GTASum model consists of a pre-training language model BERT, a neural topic model NTM, a drawing attention network GAT and a decoder;
further, the step (1) is specifically implemented as follows:
1-1 inserting special marks into the beginning and end of each sentence in the original input document D<CLS>And<SEP>set of sentences W ═ W1,w2,…,wn}; wherein, wiRepresenting the ith sentence; [ CLS]Put at the beginning of each sentence, [ SEP]Put at the end of each sentence;
1-2 putting the sentence set W into a pre-trained language model BERT, as shown in equation 1, generating a hidden state representation H of the sequenceBAnd represents the hidden state as HBThe features considered as corresponding sentences represent:
HB={h1,h2,...,hi,...,hn}=BERT({w1,w2,...,wi,...,wn}) (1)
further, the step (2) is specifically implemented as follows:
2-1, inputting an original input document D into a neural topic model NTM for coding; during the encoding process, the average value μ ═ f is generatedμ(x) Sum variance log σ ═ fσ(x) (ii) a Wherein the function fμAnd fσAre all linear transformation functions;
the 2-2 decoding process comprises three steps:
firstly, the method comprises the following steps: a gaussian distribution is used to describe the topic distribution, i.e., z- Ν (μ, σ) and θ ═ softmax (z); where z is a potential subject variable, θ ∈ RKIs the result of z normalization, K is the subject dimension;
secondly, the method comprises the following steps: by pw=softmax(WφTheta) to learn the predicted word pw∈RVThe probability of occurrence of (c); wherein, Wφ∈RV×KIs a topic-word distribution matrix in a similar LDA topic model;
thirdly, the method comprises the following steps: from predicted words pwExtracting each word to construct a bag of words xbox;
2-3 extracting neural topic model intermediate parameter WφConstructing a topic representation H using equation 2T;
Wherein the content of the first and second substances,is a set of data having a predefined dimension of dtSubject representation of fφIs a linear transformation function;
further, the step (3) is specifically implemented as follows:
3-1 constructs an undirected graph G ═ V, E }, where V ═ VS∪VTIs a set of nodes, E is a set of edges; wherein VS={S1,S2,...,SNIndicates N sentence nodes, and the sentence feature indicates h1,h2,...,hi,...,hnIs corresponding to, VT={T1,T2,...,TKDenotes K topic nodes, and the topic representation of the document { t }1,t2,...,tj,...,tkCorresponding to the obtained result; e ═ E11,...,eNKRepresents the weight between the ith sentence node and the jth topic node;
3-2 representation of H by sentence featuresBAnd a topic representation H of the documentTInitializing nodes, carrying out graph coding, and obtaining weighted representation S of each sentence node through a graph attention networki';
Wherein, WbAnd WcIs a trainable parameter; LeakyReLU is an activation function; siIs the ith sentence node, TjIs the jth subject node; si' is the ith sentence node, and contains subject information through the weighting of the subject node;is a feed-forward neural network;
3-3 considering that sentence nodes and topic nodes represent differently, a heterogeneous document graph needs to be constructed, so equation 3 is rewritten into equation 6 as follows:
the formula removes the trainable parameter WbUsing a non-linear transformation function fSAnd fTMapping sentences and topics to an implicit common space and recalculating eij;
3-4, splicing n sentence nodes containing subject information to generate sentence characteristics h ' ═ { S ' with the subject information '1,S'2,...,S'i,...,S'n};
Further, the step (4) is specifically implemented as follows:
4-1, sending the sentence characteristics h' into a Transformer-based decoder for decoding, and then predicting through a multi-layer attention mechanism; as shown in fig. 3.
Further, the step (5) is specifically implemented as follows:
5-1, performing end-to-end training, setting epoch as 10, learning rate as 0.00001 and batch size as 16, and gradually adjusting the learning rate by using an Adam optimizer; jointly training the neural topic model and the decoder to reduce loss, wherein the loss function is as follows:
LNTM=DKL(p(z)||q(z|x))-Eq(z|x)[p(x|z)] (7)
wherein the first term represents KL-divergence loss and the second term represents reconstruction loss; q (z | x) and p (x | z) represent the encoder and decoder networks of the NTM, respectively;
LTrans=-∑log p(y|x;θ) (8)
where x represents the input document, y represents the reference summary, and θ is the model parameter
L=LTrans+λLNTM (9)
Wherein λ is a balance parameter, ranging from [0,1 ];
5-2, inputting the test document into the GTASum model to obtain the summary content.
Claims (6)
1. An abstract text summarization method based on graph knowledge and topic perception is characterized by comprising the following steps:
step (1): given an original input document D, a [ CLS ] is inserted into the beginning and end of each sentence in the original input document D]And [ SEP ]]Then, the processed input document D is put into a pre-training language model BERT, and the feature representation H of the sentence is learnedB;
Step (2): inputting original input document D into neural topic model NTM, learning topic representation H of documentT;
And (3): representing the characteristics of the sentence HBAnd a topic representation H of the documentTInputting the information into a destination attention network GAT and initializing; generating sentence characteristics h' with topic information after the attention network GAT coding; the GAT coding process is to construct a heterogeneous document graph with topics and sentences, and continuously update the feature representation HBAnd a topic representation HTA constructed node representation;
and (4): sending the sentence characteristics h' with the topic information into a Transformer-based decoder for decoding; generating a text abstract after normalization;
and (5): training the GTASum model on the CNN/DailyMail data set and the XSum data set, selecting the optimal GTASum model, inputting any text into the trained GTASum model, and outputting corresponding abstract contents;
the GTASum model is composed of a pre-training language model BERT, a neural topic model NTM, a graph attention network GAT and a decoder.
2. The abstract text summarization method based on graph knowledge and topic perception according to claim 1, wherein the step (1) is implemented as follows:
1-1 inserting special marks into the beginning and end of each sentence in the original input document D<CLS>And<SEP>set of sentences W ═ W1,w2,…,wn}; wherein, wiRepresenting the ith sentence; [ CLS]Put at the beginning of each sentence, [ SEP]Put at the end of each sentence;
1-2 putting the sentence set W into a pre-trained language model BERT, as shown in equation 1, generating a hidden state representation H of the sequenceBAnd represents the hidden state as HBThe features considered as corresponding sentences represent:
HB={h1,h2,...,hi,...,hn}=BERT({w1,w2,...,wi,...,wn}) (1)。
3. the abstract text summarization method based on graph knowledge and topic perception according to claim 2, wherein the step (2) is implemented as follows:
2-1, inputting an original input document D into a neural topic model NTM for coding; during the encoding process, the average value μ ═ f is generatedμ(x) Sum variance log σ ═ fσ(x) (ii) a Wherein the function fμAnd fσAre all linear transformation functions;
the 2-2 decoding process comprises three steps:
firstly, the method comprises the following steps: a gaussian distribution is used to describe the topic distribution, i.e., z- Ν (μ, σ) and θ ═ softmax (z); where z is a potential subject variable, θ ∈ RKIs the result of z normalization, K is the subject dimension;
secondly, the method comprises the following steps: by pw=softmax(WφTheta) to learn the predicted word pw∈RVThe probability of occurrence of (c); wherein, Wφ∈RV×KIs a topic-word distribution matrix in a similar LDA topic model;
thirdly, the method comprises the following steps: from predicted words pwExtracting each word to construct a bag of words xbox;
2-3 extracting neural topic model intermediate parameter WφConstructing a topic representation H using equation 2T;
4. The abstract text summarization method based on graph knowledge and topic perception according to claim 3, wherein the step (3) is implemented as follows:
3-1 constructs an undirected graph G ═ V, E }, where V ═ VS∪VTIs a set of nodes, E is a set of edges; wherein VS={S1,S2,...,SNIndicates N sentence nodes, and the sentence feature indicates h1,h2,...,hi,...,hnIs corresponding to, VT={T1,T2,...,TKDenotes K topic nodes, and the topic representation of the document { t }1,t2,...,tj,...,tkCorresponding to the obtained result; e ═ E11,...,eNKDenotes the ith sentence node and the jth topic nodeWeight in between;
3-2 representation of H by sentence featuresBAnd a topic representation H of the documentTInitializing nodes, carrying out graph coding, and obtaining weighted representation S of each sentence node through a graph attention networki';
S'i=σ(∑αijWcTj) (5)
Wherein, WbAnd WcIs a trainable parameter; LeakyReLU is an activation function; siIs the ith sentence node, TjIs the jth subject node; s'iIs the ith sentence node and contains topic information through the weighting of the topic node;is a feed-forward neural network;
3-3 considering that sentence nodes and topic nodes represent differently, a heterogeneous document graph needs to be constructed, so equation 3 is rewritten into equation 6 as follows:
the formula removes the trainable parameter WbUsing a non-linear transformation function fSAnd fTMapping sentences and topics to an implicit common space and recalculating eij;
3-4, splicing n sentence nodes containing subject information to generate sentence characteristics h ' ═ { S ' with the subject information '1,S'2,...,S'i,...,S'n}。
5. The abstract text summarization method based on graph knowledge and topic perception according to claim 3, wherein the step (4) is implemented as follows:
4-1 feeding sentence characteristics h' into a Transformer-based decoder; the prediction is then made by a multi-layer attention mechanism.
6. The abstract text summarization method based on graph knowledge and topic perception according to claim 5, wherein the step (5) is implemented as follows:
5-1, performing end-to-end training, setting epoch as 10, learning rate as 0.00001, setting batch size as 16, and gradually adjusting the learning rate by using an Adam optimizer; jointly training the neural topic model and the decoder to reduce loss, wherein the loss function is as follows:
LNTM=DKL(p(z)||q(z|x))-Eq(z|x)[p(x|z)] (7)
wherein the first term represents KL-divergence loss and the second term represents reconstruction loss; q (z | x) and p (x | z) represent the encoder and decoder networks of the NTM, respectively;
LTrans=-∑log p(y|x;θ) (8)
where x represents the input document, y represents the reference summary, and θ is the model parameter
L=LTrans+λLNTM (9)
Wherein λ is a balance parameter, ranging from [0,1 ];
5-2, inputting the test document into the GTASum model to obtain the summary content.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111654105.5A CN114218928A (en) | 2021-12-30 | 2021-12-30 | Abstract text summarization method based on graph knowledge and theme perception |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111654105.5A CN114218928A (en) | 2021-12-30 | 2021-12-30 | Abstract text summarization method based on graph knowledge and theme perception |
Publications (1)
Publication Number | Publication Date |
---|---|
CN114218928A true CN114218928A (en) | 2022-03-22 |
Family
ID=80707059
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111654105.5A Withdrawn CN114218928A (en) | 2021-12-30 | 2021-12-30 | Abstract text summarization method based on graph knowledge and theme perception |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114218928A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115033683A (en) * | 2022-06-17 | 2022-09-09 | 平安科技(深圳)有限公司 | Abstract generation method, device, equipment and storage medium |
CN115496061A (en) * | 2022-09-30 | 2022-12-20 | 内蒙古财经大学 | Neural network title generation model |
CN117763140A (en) * | 2024-02-22 | 2024-03-26 | 神州医疗科技股份有限公司 | Accurate medical information conclusion generation method based on computing feature network |
CN117875273A (en) * | 2024-03-13 | 2024-04-12 | 中南大学 | News abstract automatic generation method, device and medium based on large language model |
-
2021
- 2021-12-30 CN CN202111654105.5A patent/CN114218928A/en not_active Withdrawn
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115033683A (en) * | 2022-06-17 | 2022-09-09 | 平安科技(深圳)有限公司 | Abstract generation method, device, equipment and storage medium |
CN115033683B (en) * | 2022-06-17 | 2024-05-07 | 平安科技(深圳)有限公司 | Digest generation method, digest generation device, digest generation equipment and storage medium |
CN115496061A (en) * | 2022-09-30 | 2022-12-20 | 内蒙古财经大学 | Neural network title generation model |
CN117763140A (en) * | 2024-02-22 | 2024-03-26 | 神州医疗科技股份有限公司 | Accurate medical information conclusion generation method based on computing feature network |
CN117763140B (en) * | 2024-02-22 | 2024-05-28 | 神州医疗科技股份有限公司 | Accurate medical information conclusion generation method based on computing feature network |
CN117875273A (en) * | 2024-03-13 | 2024-04-12 | 中南大学 | News abstract automatic generation method, device and medium based on large language model |
CN117875273B (en) * | 2024-03-13 | 2024-05-28 | 中南大学 | News abstract automatic generation method, device and medium based on large language model |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Tan et al. | Neural machine translation: A review of methods, resources, and tools | |
CN108733792B (en) | Entity relation extraction method | |
CN110678881B (en) | Natural language processing using context-specific word vectors | |
CN113158665B (en) | Method for improving dialog text generation based on text abstract generation and bidirectional corpus generation | |
CN114218928A (en) | Abstract text summarization method based on graph knowledge and theme perception | |
CN109543017B (en) | Legal question keyword generation method and system | |
CN109684452A (en) | A kind of neural network problem generation method based on answer Yu answer location information | |
CN111125333B (en) | Generation type knowledge question-answering method based on expression learning and multi-layer covering mechanism | |
WO2023137911A1 (en) | Intention classification method and apparatus based on small-sample corpus, and computer device | |
Xu et al. | Multi-task learning for abstractive text summarization with key information guide network | |
Gambhir et al. | Deep learning-based extractive text summarization with word-level attention mechanism | |
Yan et al. | Leveraging contextual sentences for text classification by using a neural attention model | |
CN114881042A (en) | Chinese emotion analysis method based on graph convolution network fusion syntax dependence and part of speech | |
CN112818698A (en) | Fine-grained user comment sentiment analysis method based on dual-channel model | |
Mathur et al. | A scaled‐down neural conversational model for chatbots | |
Xiao et al. | FusionSum: Abstractive summarization with sentence fusion and cooperative reinforcement learning | |
Madhyastha et al. | Learning task-specific bilexical embeddings | |
CN113961706A (en) | Accurate text representation method based on neural network self-attention mechanism | |
CN115600582B (en) | Controllable text generation method based on pre-training language model | |
Zhao et al. | Guiding the training of distributed text representation with supervised weighting scheme for sentiment analysis | |
Hsiao et al. | [Retracted] Construction of an Artificial Intelligence Writing Model for English Based on Fusion Neural Network Model | |
CN112287641B (en) | Synonym sentence generating method, system, terminal and storage medium | |
CN114548117A (en) | Cause-and-effect relation extraction method based on BERT semantic enhancement | |
CN113449517A (en) | Entity relationship extraction method based on BERT (belief propagation) gating multi-window attention network model | |
Cui et al. | Aspect level sentiment classification based on double attention mechanism |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WW01 | Invention patent application withdrawn after publication | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20220322 |