CN114218928A - Abstract text summarization method based on graph knowledge and theme perception - Google Patents

Abstract text summarization method based on graph knowledge and theme perception Download PDF

Info

Publication number
CN114218928A
CN114218928A CN202111654105.5A CN202111654105A CN114218928A CN 114218928 A CN114218928 A CN 114218928A CN 202111654105 A CN202111654105 A CN 202111654105A CN 114218928 A CN114218928 A CN 114218928A
Authority
CN
China
Prior art keywords
topic
sentence
model
graph
document
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202111654105.5A
Other languages
Chinese (zh)
Inventor
姜明
邹一凡
张旻
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202111654105.5A priority Critical patent/CN114218928A/en
Publication of CN114218928A publication Critical patent/CN114218928A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/253Grammatical analysis; Style critique
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses an abstract text summarization method based on graph knowledge and topic perception. The invention provides a model based on BERT, a neural topic model and a graph neural network, which is called GTASum. At the input end of a document encoder, using BERT to obtain a hidden word vector of a document; at the input of the topic encoder, obtaining a word-topic distribution vector of the document using a neural topic model; inputting the two vectors into a graph neural network for training to obtain context content fused with topic knowledge, and generating a text by a transform-based decoder; meanwhile, the LN layer with the normalized condition can cooperatively train a neural topic model and a decoder, and effectively select the characteristics. The result shows that the method has better robustness and adaptive capacity.

Description

Abstract text summarization method based on graph knowledge and theme perception
Technical Field
The invention belongs to the technical field of natural language processing, and relates to a text abstract generation method, in particular to an abstract text abstract method based on graph knowledge and topic perception, and a text abstract generation method based on a pre-training language model, a neural topic model and a graph neural network.
Background
With the development of computer performance and large-scale language models, natural language processing tasks (NLPs) have made significant development. The abstract task is one of core problems of natural language processing tasks and aims to enable people to quickly master important information in texts. Text summaries have been widely used in a number of areas such as news, finance, conferencing, and medicine. At present, there are two main methods for the summarization task: an abstraction method and an abstract method. The extraction method mainly copies important information from the original text and then aggregates them into text. The abstract generated by the method usually retains the significant information of the source text and has correct grammar, but inevitably generates a large amount of redundant information easily. Abstract methods are to form a summary based on an understanding of the source text input. It tries to understand the text content, can generate words not in the original text, more closely approaches the essence of the abstract, and has the potential of generating high-quality abstract. To sum up, in order to better generate news summaries to help readers quickly and efficiently grasp daily messages, the present study focuses on abstract summaries.
In abstract abstracts, sequence-to-sequence has become the dominant framework for a variety of architectures. In the early days, the abstracted text excerpts were primarily RNN-based codecs. Because of the "long range dependency" problem, the RNN has lost much information by the time the word is entered at the last time step, and bahdana et al apply the Attention mechanism to NLP. Since then, text summarization has advanced greatly, and various applications have emerged endlessly. Among these, the most notable sibling transform architecture. This model has achieved surprising performance in many areas. Current SOTA abstraction models, including BART, PEGASUS, and ProphetNet, all employ a transform-based architecture. Also, with the contribution of attention mechanism, the transform-based model can capture well the grammatical and contextual information between tokens. However, higher levels of semantic understanding do not perform well.
To address this problem, researchers have tried a variety of improvements, one of which is the way of exploiting topic perception. Topic models such as LDA, PFA, NVDM, NTM can all provide additional information for document understanding. For text summarization, we believe that it can improve model performance by incorporating the topic model properties into the summarization model. Furthermore, in recent years, Graph Neural Networks (GNNs) have been widely used for cross-sentence relational modeling of summarization tasks. Some studies have created document graphs based on linguistic analysis. However, this approach relies on external tools, possibly leading to semantically fragmented outputs. Wang and Liu et al have constructed word and sentence document graphs, but this approach has difficulty capturing semantic level relationships, and therefore, how to efficiently construct documents into abstractable graphs is also a difficult problem.
Disclosure of Invention
The invention aims to provide an abstract text summarization method based on graph knowledge and theme perception, aiming at the defects of the prior art.
The technical scheme adopted by the invention for solving the technical problem comprises the following steps:
step (1): given an original input document D, a [ CLS ] is inserted into the beginning and end of each sentence in the original input document D]And [ SEP ]]Then, the processed input document D is put into a pre-training language model BERT, and the feature representation H of the sentence is learnedB
Step (2): inputting original input document D into neural topic model NTM, learning topic representation H of documentT
And (3): representing the characteristics of the sentence HBAnd a topic representation H of the documentTInputting the information into a destination attention network GAT and initializing; generating sentence characteristics h' with topic information after the attention network GAT coding; the GAT coding process is to construct a heterogeneous document graph with topics and sentences, and continuously update the feature representation HBAnd a topic representation HTA constructed node representation;
and (4): sending the sentence characteristics h' with the topic information into a Transformer-based decoder for decoding; generating a text abstract after normalization;
and (5): training the GTASum model on the CNN/DailyMail data set and the XSum data set, selecting the optimal GTASum model, inputting any text into the trained GTASum model, and outputting corresponding abstract contents;
the GTASum model is composed of a pre-training language model BERT, a neural topic model NTM, a graph attention network GAT and a decoder.
The invention has the following beneficial effects:
the invention provides a topic model and a graph neural network which can help a pre-training language model to better speak. During training, the document theme information and the document embedded information are sent to a graph neural network for fusion and updating. Through the operation, when the pre-training language model faces downstream tasks, feature selection can be carried out by better referring to subject information. The results show that this approach has better robustness and adaptability. And testing according to standard performance indexes in the field of text summarization, wherein the evaluation indexes comprise ROUGE _1/ROUGE _2/ROUGE _ L. Tests were performed on news text summary datasets CNN/DailyMail and XSum and yielded results that were of a leading level in the field.
Drawings
FIG. 1 is the overall flow framework of the model;
FIG. 2 is a neural topic model portion framework;
FIG. 3 is an example of the present model usage;
Detailed Description
The invention is further illustrated by the following figures and examples.
The invention provides an abstract text summarization method based on graph knowledge and theme perception. First, we encode the input document with a pre-trained language model BERT to learn the contextual sentence representation while using the Neural Topic Model (NTM) to find potential topics. Then, we build a heterogeneous document graph composed of sentences and topic nodes and update the representation by using a modified graph attention network (GAT). Thirdly, the expression form of the sentence node is obtained, and the latent semantic meaning is calculated. Finally, the latent semantics are fed into a transform-based decoder for decoding to generate a final result. We performed extensive experiments on two real-world datasets CNN/DailyMail and XSum.
The model based on the BERT, the neural topic model and the graph neural network is called GTASum. At the input end of a document encoder, using BERT to obtain a hidden word vector of a document; at the input of the topic encoder, obtaining a word-topic distribution vector of the document using a neural topic model; inputting the two vectors into a graph neural network for training to obtain context content fused with topic knowledge, and generating a text by a transform-based decoder; meanwhile, the LN layer with the normalized condition can cooperatively train a neural topic model and a decoder, and effectively select the characteristics. The result shows that the model can well capture the key information of the abstract text abstract and has better robustness and adaptive capacity.
As shown in fig. 1 and 2, an abstract text summarization method based on graph knowledge and topic perception includes the following steps:
step (1): given an original input document D, a [ CLS ] is inserted into the beginning and end of each sentence in the original input document D]And [ SEP ]]Then, the processed input document D is put into a pre-training language model BERT, and the feature representation H of the sentence is learnedB
Step (2): inputting original input document D into neural topic model NTM, learning topic representation H of documentT
And (3): representing the characteristics of the sentence HBAnd a topic representation H of the documentTInputting the information into a destination attention network GAT and initializing; generating sentence characteristics h' with topic information after the attention network GAT coding; the GAT coding process is to construct a heterogeneous document graph with topics and sentences, and continuously update the feature representation HBAnd a topic representation HTA constructed node representation;
and (4): sending the sentence characteristics h' with the topic information into a Transformer-based decoder for decoding; generating a text abstract after normalization;
and (5): training the GTASum model on the CNN/DailyMail data set and the XSum data set, selecting the optimal GTASum model, inputting any text into the trained GTASum model, and outputting corresponding abstract contents;
the GTASum model consists of a pre-training language model BERT, a neural topic model NTM, a drawing attention network GAT and a decoder;
further, the step (1) is specifically implemented as follows:
1-1 inserting special marks into the beginning and end of each sentence in the original input document D<CLS>And<SEP>set of sentences W ═ W1,w2,…,wn}; wherein, wiRepresenting the ith sentence; [ CLS]Put at the beginning of each sentence, [ SEP]Put at the end of each sentence;
1-2 putting the sentence set W into a pre-trained language model BERT, as shown in equation 1, generating a hidden state representation H of the sequenceBAnd represents the hidden state as HBThe features considered as corresponding sentences represent:
HB={h1,h2,...,hi,...,hn}=BERT({w1,w2,...,wi,...,wn}) (1)
further, the step (2) is specifically implemented as follows:
2-1, inputting an original input document D into a neural topic model NTM for coding; during the encoding process, the average value μ ═ f is generatedμ(x) Sum variance log σ ═ fσ(x) (ii) a Wherein the function fμAnd fσAre all linear transformation functions;
the 2-2 decoding process comprises three steps:
firstly, the method comprises the following steps: a gaussian distribution is used to describe the topic distribution, i.e., z- Ν (μ, σ) and θ ═ softmax (z); where z is a potential subject variable, θ ∈ RKIs the result of z normalization, K is the subject dimension;
secondly, the method comprises the following steps: by pw=softmax(WφTheta) to learn the predicted word pw∈RVThe probability of occurrence of (c); wherein, Wφ∈RV×KIs a topic-word distribution matrix in a similar LDA topic model;
thirdly, the method comprises the following steps: from predicted words pwExtracting each word to construct a bag of words xbox
2-3 extracting neural topic model intermediate parameter WφConstructing a topic representation H using equation 2T
Figure BDA0003447775050000051
Wherein the content of the first and second substances,
Figure BDA0003447775050000052
is a set of data having a predefined dimension of dtSubject representation of fφIs a linear transformation function;
further, the step (3) is specifically implemented as follows:
3-1 constructs an undirected graph G ═ V, E }, where V ═ VS∪VTIs a set of nodes, E is a set of edges; wherein VS={S1,S2,...,SNIndicates N sentence nodes, and the sentence feature indicates h1,h2,...,hi,...,hnIs corresponding to, VT={T1,T2,...,TKDenotes K topic nodes, and the topic representation of the document { t }1,t2,...,tj,...,tkCorresponding to the obtained result; e ═ E11,...,eNKRepresents the weight between the ith sentence node and the jth topic node;
3-2 representation of H by sentence featuresBAnd a topic representation H of the documentTInitializing nodes, carrying out graph coding, and obtaining weighted representation S of each sentence node through a graph attention networki';
Figure BDA0003447775050000053
Figure BDA0003447775050000054
Figure BDA0003447775050000055
Wherein, WbAnd WcIs a trainable parameter; LeakyReLU is an activation function; siIs the ith sentence node, TjIs the jth subject node; si' is the ith sentence node, and contains subject information through the weighting of the subject node;
Figure BDA0003447775050000056
is a feed-forward neural network;
3-3 considering that sentence nodes and topic nodes represent differently, a heterogeneous document graph needs to be constructed, so equation 3 is rewritten into equation 6 as follows:
Figure BDA0003447775050000061
the formula removes the trainable parameter WbUsing a non-linear transformation function fSAnd fTMapping sentences and topics to an implicit common space and recalculating eij
3-4, splicing n sentence nodes containing subject information to generate sentence characteristics h ' ═ { S ' with the subject information '1,S'2,...,S'i,...,S'n};
Further, the step (4) is specifically implemented as follows:
4-1, sending the sentence characteristics h' into a Transformer-based decoder for decoding, and then predicting through a multi-layer attention mechanism; as shown in fig. 3.
Further, the step (5) is specifically implemented as follows:
5-1, performing end-to-end training, setting epoch as 10, learning rate as 0.00001 and batch size as 16, and gradually adjusting the learning rate by using an Adam optimizer; jointly training the neural topic model and the decoder to reduce loss, wherein the loss function is as follows:
LNTM=DKL(p(z)||q(z|x))-Eq(z|x)[p(x|z)] (7)
wherein the first term represents KL-divergence loss and the second term represents reconstruction loss; q (z | x) and p (x | z) represent the encoder and decoder networks of the NTM, respectively;
LTrans=-∑log p(y|x;θ) (8)
where x represents the input document, y represents the reference summary, and θ is the model parameter
L=LTrans+λLNTM (9)
Wherein λ is a balance parameter, ranging from [0,1 ];
5-2, inputting the test document into the GTASum model to obtain the summary content.

Claims (6)

1. An abstract text summarization method based on graph knowledge and topic perception is characterized by comprising the following steps:
step (1): given an original input document D, a [ CLS ] is inserted into the beginning and end of each sentence in the original input document D]And [ SEP ]]Then, the processed input document D is put into a pre-training language model BERT, and the feature representation H of the sentence is learnedB
Step (2): inputting original input document D into neural topic model NTM, learning topic representation H of documentT
And (3): representing the characteristics of the sentence HBAnd a topic representation H of the documentTInputting the information into a destination attention network GAT and initializing; generating sentence characteristics h' with topic information after the attention network GAT coding; the GAT coding process is to construct a heterogeneous document graph with topics and sentences, and continuously update the feature representation HBAnd a topic representation HTA constructed node representation;
and (4): sending the sentence characteristics h' with the topic information into a Transformer-based decoder for decoding; generating a text abstract after normalization;
and (5): training the GTASum model on the CNN/DailyMail data set and the XSum data set, selecting the optimal GTASum model, inputting any text into the trained GTASum model, and outputting corresponding abstract contents;
the GTASum model is composed of a pre-training language model BERT, a neural topic model NTM, a graph attention network GAT and a decoder.
2. The abstract text summarization method based on graph knowledge and topic perception according to claim 1, wherein the step (1) is implemented as follows:
1-1 inserting special marks into the beginning and end of each sentence in the original input document D<CLS>And<SEP>set of sentences W ═ W1,w2,…,wn}; wherein, wiRepresenting the ith sentence; [ CLS]Put at the beginning of each sentence, [ SEP]Put at the end of each sentence;
1-2 putting the sentence set W into a pre-trained language model BERT, as shown in equation 1, generating a hidden state representation H of the sequenceBAnd represents the hidden state as HBThe features considered as corresponding sentences represent:
HB={h1,h2,...,hi,...,hn}=BERT({w1,w2,...,wi,...,wn}) (1)。
3. the abstract text summarization method based on graph knowledge and topic perception according to claim 2, wherein the step (2) is implemented as follows:
2-1, inputting an original input document D into a neural topic model NTM for coding; during the encoding process, the average value μ ═ f is generatedμ(x) Sum variance log σ ═ fσ(x) (ii) a Wherein the function fμAnd fσAre all linear transformation functions;
the 2-2 decoding process comprises three steps:
firstly, the method comprises the following steps: a gaussian distribution is used to describe the topic distribution, i.e., z- Ν (μ, σ) and θ ═ softmax (z); where z is a potential subject variable, θ ∈ RKIs the result of z normalization, K is the subject dimension;
secondly, the method comprises the following steps: by pw=softmax(WφTheta) to learn the predicted word pw∈RVThe probability of occurrence of (c); wherein, Wφ∈RV×KIs a topic-word distribution matrix in a similar LDA topic model;
thirdly, the method comprises the following steps: from predicted words pwExtracting each word to construct a bag of words xbox
2-3 extracting neural topic model intermediate parameter WφConstructing a topic representation H using equation 2T
Figure FDA0003447775040000021
Wherein the content of the first and second substances,
Figure FDA0003447775040000022
is a set of data having a predefined dimension of dtSubject representation of fφIs a linear transformation function.
4. The abstract text summarization method based on graph knowledge and topic perception according to claim 3, wherein the step (3) is implemented as follows:
3-1 constructs an undirected graph G ═ V, E }, where V ═ VS∪VTIs a set of nodes, E is a set of edges; wherein VS={S1,S2,...,SNIndicates N sentence nodes, and the sentence feature indicates h1,h2,...,hi,...,hnIs corresponding to, VT={T1,T2,...,TKDenotes K topic nodes, and the topic representation of the document { t }1,t2,...,tj,...,tkCorresponding to the obtained result; e ═ E11,...,eNKDenotes the ith sentence node and the jth topic nodeWeight in between;
3-2 representation of H by sentence featuresBAnd a topic representation H of the documentTInitializing nodes, carrying out graph coding, and obtaining weighted representation S of each sentence node through a graph attention networki';
Figure FDA0003447775040000023
Figure FDA0003447775040000024
S'i=σ(∑αijWcTj) (5)
Wherein, WbAnd WcIs a trainable parameter; LeakyReLU is an activation function; siIs the ith sentence node, TjIs the jth subject node; s'iIs the ith sentence node and contains topic information through the weighting of the topic node;
Figure FDA0003447775040000025
is a feed-forward neural network;
3-3 considering that sentence nodes and topic nodes represent differently, a heterogeneous document graph needs to be constructed, so equation 3 is rewritten into equation 6 as follows:
Figure FDA0003447775040000031
the formula removes the trainable parameter WbUsing a non-linear transformation function fSAnd fTMapping sentences and topics to an implicit common space and recalculating eij
3-4, splicing n sentence nodes containing subject information to generate sentence characteristics h ' ═ { S ' with the subject information '1,S'2,...,S'i,...,S'n}。
5. The abstract text summarization method based on graph knowledge and topic perception according to claim 3, wherein the step (4) is implemented as follows:
4-1 feeding sentence characteristics h' into a Transformer-based decoder; the prediction is then made by a multi-layer attention mechanism.
6. The abstract text summarization method based on graph knowledge and topic perception according to claim 5, wherein the step (5) is implemented as follows:
5-1, performing end-to-end training, setting epoch as 10, learning rate as 0.00001, setting batch size as 16, and gradually adjusting the learning rate by using an Adam optimizer; jointly training the neural topic model and the decoder to reduce loss, wherein the loss function is as follows:
LNTM=DKL(p(z)||q(z|x))-Eq(z|x)[p(x|z)] (7)
wherein the first term represents KL-divergence loss and the second term represents reconstruction loss; q (z | x) and p (x | z) represent the encoder and decoder networks of the NTM, respectively;
LTrans=-∑log p(y|x;θ) (8)
where x represents the input document, y represents the reference summary, and θ is the model parameter
L=LTrans+λLNTM (9)
Wherein λ is a balance parameter, ranging from [0,1 ];
5-2, inputting the test document into the GTASum model to obtain the summary content.
CN202111654105.5A 2021-12-30 2021-12-30 Abstract text summarization method based on graph knowledge and theme perception Withdrawn CN114218928A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111654105.5A CN114218928A (en) 2021-12-30 2021-12-30 Abstract text summarization method based on graph knowledge and theme perception

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111654105.5A CN114218928A (en) 2021-12-30 2021-12-30 Abstract text summarization method based on graph knowledge and theme perception

Publications (1)

Publication Number Publication Date
CN114218928A true CN114218928A (en) 2022-03-22

Family

ID=80707059

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111654105.5A Withdrawn CN114218928A (en) 2021-12-30 2021-12-30 Abstract text summarization method based on graph knowledge and theme perception

Country Status (1)

Country Link
CN (1) CN114218928A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115033683A (en) * 2022-06-17 2022-09-09 平安科技(深圳)有限公司 Abstract generation method, device, equipment and storage medium
CN115496061A (en) * 2022-09-30 2022-12-20 内蒙古财经大学 Neural network title generation model
CN117763140A (en) * 2024-02-22 2024-03-26 神州医疗科技股份有限公司 Accurate medical information conclusion generation method based on computing feature network
CN117875273A (en) * 2024-03-13 2024-04-12 中南大学 News abstract automatic generation method, device and medium based on large language model

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115033683A (en) * 2022-06-17 2022-09-09 平安科技(深圳)有限公司 Abstract generation method, device, equipment and storage medium
CN115033683B (en) * 2022-06-17 2024-05-07 平安科技(深圳)有限公司 Digest generation method, digest generation device, digest generation equipment and storage medium
CN115496061A (en) * 2022-09-30 2022-12-20 内蒙古财经大学 Neural network title generation model
CN117763140A (en) * 2024-02-22 2024-03-26 神州医疗科技股份有限公司 Accurate medical information conclusion generation method based on computing feature network
CN117763140B (en) * 2024-02-22 2024-05-28 神州医疗科技股份有限公司 Accurate medical information conclusion generation method based on computing feature network
CN117875273A (en) * 2024-03-13 2024-04-12 中南大学 News abstract automatic generation method, device and medium based on large language model
CN117875273B (en) * 2024-03-13 2024-05-28 中南大学 News abstract automatic generation method, device and medium based on large language model

Similar Documents

Publication Publication Date Title
Tan et al. Neural machine translation: A review of methods, resources, and tools
CN108733792B (en) Entity relation extraction method
CN110678881B (en) Natural language processing using context-specific word vectors
CN113158665B (en) Method for improving dialog text generation based on text abstract generation and bidirectional corpus generation
CN114218928A (en) Abstract text summarization method based on graph knowledge and theme perception
CN109543017B (en) Legal question keyword generation method and system
CN109684452A (en) A kind of neural network problem generation method based on answer Yu answer location information
CN111125333B (en) Generation type knowledge question-answering method based on expression learning and multi-layer covering mechanism
WO2023137911A1 (en) Intention classification method and apparatus based on small-sample corpus, and computer device
Xu et al. Multi-task learning for abstractive text summarization with key information guide network
Gambhir et al. Deep learning-based extractive text summarization with word-level attention mechanism
Yan et al. Leveraging contextual sentences for text classification by using a neural attention model
CN114881042A (en) Chinese emotion analysis method based on graph convolution network fusion syntax dependence and part of speech
CN112818698A (en) Fine-grained user comment sentiment analysis method based on dual-channel model
Mathur et al. A scaled‐down neural conversational model for chatbots
Xiao et al. FusionSum: Abstractive summarization with sentence fusion and cooperative reinforcement learning
Madhyastha et al. Learning task-specific bilexical embeddings
CN113961706A (en) Accurate text representation method based on neural network self-attention mechanism
CN115600582B (en) Controllable text generation method based on pre-training language model
Zhao et al. Guiding the training of distributed text representation with supervised weighting scheme for sentiment analysis
Hsiao et al. [Retracted] Construction of an Artificial Intelligence Writing Model for English Based on Fusion Neural Network Model
CN112287641B (en) Synonym sentence generating method, system, terminal and storage medium
CN114548117A (en) Cause-and-effect relation extraction method based on BERT semantic enhancement
CN113449517A (en) Entity relationship extraction method based on BERT (belief propagation) gating multi-window attention network model
Cui et al. Aspect level sentiment classification based on double attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20220322