CN116484858A - Text abstract generation method based on diffusion model - Google Patents

Text abstract generation method based on diffusion model Download PDF

Info

Publication number
CN116484858A
CN116484858A CN202310419493.1A CN202310419493A CN116484858A CN 116484858 A CN116484858 A CN 116484858A CN 202310419493 A CN202310419493 A CN 202310419493A CN 116484858 A CN116484858 A CN 116484858A
Authority
CN
China
Prior art keywords
text
vector
model
diffusion model
generated
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310419493.1A
Other languages
Chinese (zh)
Inventor
魏建香
陈之航
陈宇行
陈佳华
周钰锦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Posts and Telecommunications
Original Assignee
Nanjing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Posts and Telecommunications filed Critical Nanjing University of Posts and Telecommunications
Priority to CN202310419493.1A priority Critical patent/CN116484858A/en
Publication of CN116484858A publication Critical patent/CN116484858A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a text abstract generating method based on a diffusion model, which comprises the steps of firstly, performing word segmentation operation on a source text by using a natural language processing model; then, pre-encoding and converting the text data into sentence representation by the pre-processed source text through an input module by utilizing a word embedding method, and then mapping the text vector onto a space dimension capable of being continuously transformed by utilizing an evaluation module to realize controllable linear continuous transformation of the text vector; and finally, acquiring a target text through a condition guiding module, generating samples through gradient guiding in the sampling process of the diffusion model, reducing errors of generated text vectors and real text vectors by using a control module, and finally generating a text abstract through an output module according to the acquired sample vectors. The method realizes the generation of the text abstract based on the generated model, is beneficial to improving the accuracy of the text abstract and improves the semantic generalization of the model.

Description

Text abstract generation method based on diffusion model
Technical Field
The invention relates to a text abstract generation method based on a diffusion model, and belongs to the technical field of computer application.
Background
In recent years, due to the growth of various data generated by the internet and the improvement of computer performance, deep learning technologies represented by neural networks are rapidly developed and have achieved great success in various fields, such as the field of natural language processing, the field of computer vision and the like, which have triggered a new round of artificial intelligence research. Text generation is an important branch of the natural language processing field, and is a subdivision field studied by many scholars. It is a process of generating human-readable and understandable text satisfying specific constraints given input information. Has extremely high research value and application value, and has influence on human life in a plurality of fields such as machine translation, automatic abstract, dialogue generation, image description and the like.
In order to improve the quality of text generation, researchers in recent years select a plurality of generation type deep neural network models to apply to text generation tasks and have made certain progress, and the advantage is that the deep neural network can learn the semantic mapping from input data to output text end to end without manual participation in feature engineering. However, deep neural network models tend to have a large number of parameters, and most text generation tasks can use very small data sets, so deep neural networks are very easy to overfit on these data sets, making it impractical to generalize in practical applications. Along with the continuous increase of the difficulty of the text generation task, higher requirements are provided for obtaining the hidden semantic relation between the theme and the text, simulating the human style writing and the like. The traditional generation model comprises an autoregressive model, a variational self-encoder model, a neural network model for generating an countermeasure network and the like, but the likelihood-based model and the recessive generation model have different limitations, such as the variational self-encoder model cannot directly calculate accurate maximum likelihood due to model design, and the phenomenon of unstable training such as mode collapse and the like easily occurs in the training of the countermeasure network model, so the traditional generation model cannot fully cope with the fine-grained task.
In view of this, a text abstract generation method based on a diffusion model is provided, which has important meaning for solving the above-mentioned text abstract generation problem.
Disclosure of Invention
The invention aims to provide a text abstract generating method based on a diffusion model, which can fully utilize the context information of an input text to generate a high-quality text abstract, so that the quality of the text abstract is improved, the text abstract has good generalization capability, can adapt to various input texts of different types, including long texts and short texts, and has good adaptability.
In order to achieve the above object, the present invention provides a text abstract generating method based on a diffusion model, which mainly includes the following steps:
step 1, preprocessing text word segmentation: preliminary preprocessing is carried out on the corpus by using a natural language processing model in an input module, and the corpus is precoded by using a word embedding mode to obtain sentence representation W= [ W ] 1 ,w 2 ,w 3 ,...,w n ]Where W is the encoded set of sentence representations, W n A code representing an nth word segment in the sentence;
step 2, generating a text vector: using an auxiliary function in the evaluation module to represent the sentence generated in step 1 with w= [ W ] 1 ,w 2 ,w 3 ,...,w n ]Conversion to word embedding vector EMB (W) = [ EMB (W) 1 ),...,EMB(w n )]∈R d Mapping each text vector in the word-embedded vector to a vector space R d In each text vector can be represented in a vector space R d Wherein EMB (W) is a set of word-embedded vectors transformed from sentence-representation encoding W, EMB (W) n ) Embedding vectors for words generated by the code conversion of nth segmented words in sentences;
step 3, generating a condition guide: inputting the text vectors generated by the source text into the diffusion model by means of the guidance module in order to introduce additional conditional inputs co into the diffusion model θ (x t Y), then at diffusionFor each moment X in the model sampling process t-1 Gradient update of a network of individual run condition guided modules, where X t Representing the predicted distribution of the diffusion model generated at time t, y representing the conditional input of the diffusion model introduced at time t, co θ Representing the model at the extra conditional input y and the predictive distribution X t A conditional distribution generated by a conditional model parameter θ;
step 4, sampling generation: in the diffusion model sampling process, the initial input is vector space R d A random Gaussian noise in the sample, for each moment X in the whole sampling process t-1 Running gradient updates to generate a prediction vector f θ (x t-1 T), where t represents the time t, x after the model samples t steps t-1 Representing the sampling distribution at time t-1 generated by the diffusion model at time t, f θ Representing a model by x t-1 Distributing the prediction vector generated at the time T, and executing the whole sampling process within the range of T:0;
step 5, controlling vector errors: the control module is used for generating the prediction vector f finally generated in the step 4 according to the control function θ (x 0 T) mapping to its nearest real word embedding vector, generating a final prediction vectorWherein x is 0 Sampling distribution at 0 moment generated for a final diffusion model in a sampling process;
step 6, generating a summary: final prediction vector generated according to step 5And (3) word embedding pre-coding of the input module is converted into text information, and a text abstract is output according to a text sequence.
In step 1, a trained fine-tuning bert-base-Chinese model is used in an input module to divide sentences and words of a corpus, and a word embedding mode is used to pre-encode the corpus to obtain sentence representation W= [ W ] 1 ,w 2 ,w 3 ,...,w n ]。
As a further improvement of the present invention, in step 2, the vector space R d Is determined by the word embedding vector.
As a further improvement of the present invention, step 3 is specifically: converting the source text expected to execute abstract into text vector c through the guiding module, inputting the text vector c into a guiding network, and enabling the text vector c to be micro-processedControl is performed by the director network for each time X t-1 Run gradient update generated by association of a director network parameterized control conditions text and a target text, where p (c|X t-1 ) Representing a sampling distribution X at the moment of generating t-1 by a diffusion model t-1 The probability that the target text is c under the condition,representing the diffusion model at X when the step length is the t-th step in the process of sampling t-1 And performing gradient updating.
As a further improvement of the invention, the bootstrap network used by the bootstrap module is a fine-tuning bert-base-chipset model.
As a further improvement of the invention, the diffusion model in step 4 is input as vector space R at initial time T during sampling d Random gaussian noise of (a), for each moment X during the whole sampling process t-1 Running gradient update, finally ending at sampling X 0 The digest text vector fθ (x) input in the time generation step 5 0 ,t)。
As a further improvement of the invention, the sampling mode used in the sampling process in the step 4 is DDIM sampling, and the final X is generated by a denoising diffusion implicit model sampling method 0 Abstract text vector f of moment θ (x 0 ,t)。
As a further improvement of the present invention, in step 5, the control module is configured to control the prediction vector fθ (x t-1 T) mapping to its nearest real word embedding vector, generating a final prediction vectorTo reduce the predictive vector->Differences from the real word embedding vector.
As a further development of the invention, in step 5, the control function Squeeze (fθ (x t T)) to obtain a predictive vectorLet predictive vector +.>Approximating a real word embedding vector.
The beneficial effects of the invention are as follows: according to the method, the input text is encoded and modeled, so that semantic information in the text can be captured better, the problem of omission or misunderstanding possibly occurring in a traditional keyword extraction-based method is avoided, the accuracy of the text abstract is improved, and the semantic generalization of the model is improved.
Drawings
FIG. 1 is a technical roadmap of the text abstract generation method based on the diffusion model of the invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in detail with reference to the accompanying drawings and specific embodiments.
In this case, in order to avoid obscuring the present invention due to unnecessary details, only the structures and/or processing steps closely related to the aspects of the present invention are shown in the drawings, and other details not greatly related to the present invention are omitted.
In addition, it should be further noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus.
The invention provides a text abstract generating method based on a diffusion model, which comprises the steps of firstly, utilizing a natural language processing model to perform word segmentation operation on a source text; then, pre-coding and converting the text data into sentence representation by the input module through the pre-processed source text by utilizing a word embedding method, and then mapping the text vector onto a space dimension capable of being continuously transformed by utilizing the evaluation module to realize controllable linear continuous transformation of the text vector; and finally, acquiring a target text through a condition guiding module, generating samples through gradient guiding in the sampling process of the diffusion model, reducing errors of generated text vectors and real text vectors by using a control module, and finally generating a text abstract through an output module according to the acquired sample vectors.
As shown in fig. 1, the text abstract generating method based on the diffusion model mainly comprises the following steps:
step 1, preprocessing text word segmentation: preliminary preprocessing is carried out on the corpus by using a natural language processing model in an input module, and the corpus is precoded by using a word embedding mode to obtain sentence representation W= [ W ] 1 ,w 2 ,w 3 ,...,w n ]Where W is the encoded set of sentence representations, W n A code representing an nth word segment in the sentence;
step 2, generating a text vector: using an auxiliary function in the evaluation module to represent the sentence generated in step 1 with w= [ W ] 1 ,w 2 ,w 3 ,...,w n ]Conversion to word embedding vector EMB (W) = [ EMB (W) 1 ),...,EMB(w n )]∈R d Mapping each text vector in the word-embedded vector to a vector space R d In each text vector can be represented in a vector space R d Wherein EMB (W) is a set of word-embedded vectors transformed from sentence-representation encoding W, EMB (W) n ) Embedding vectors for words generated by the code conversion of nth segmented words in sentences;
step 3, generating a condition guide: will be guided by the guiding moduleThe text vectors generated by the source text are input into the diffusion model in order to introduce additional conditional inputs co into the diffusion model θ (x t Y) and then for each moment X during diffusion model sampling t-1 Gradient update of a network of individual run condition guided modules, where X t Representing the predicted distribution of the diffusion model generated at time t, y representing the conditional input of the diffusion model introduced at time t, co θ Representing the model at the extra conditional input y and the predictive distribution X t A conditional distribution generated by a conditional model parameter θ;
step 4, sampling generation: in the diffusion model sampling process, the initial input is vector space R d A random Gaussian noise in the sample, for each moment X in the whole sampling process t-1 Running gradient updates to generate a prediction vector f θ (x t-1 T), where t represents the time t, x after the model samples t steps t-1 Representing the sampling distribution at time t-1 generated by the diffusion model at time t, f θ Representing a model by x t-1 Distributing the prediction vector generated at the time T, and executing the whole sampling process within the range of T:0;
step 5, controlling vector errors: the control module is used for generating the prediction vector f finally generated in the step 4 according to the control function θ (x 0 T) mapping to its nearest real word embedding vector, generating a final prediction vectorWherein x is 0 Sampling distribution at 0 moment generated for a final diffusion model in a sampling process;
step 6, generating a summary: final prediction vector generated according to step 5And (3) word embedding pre-coding of the input module is converted into text information, and a text abstract is output according to a text sequence.
The steps 1 to 6 will be described in detail below.
In step 1, a trained fine-tuning bert-base-lane is used in the input modulePreliminary preprocessing such as sentence segmentation and word segmentation is carried out on the corpus by the model, and the corpus is precoded by a word embedding mode to obtain sentence representation W= [ W ] 1 ,w 2 ,w 3 ,...,w n ]。
In step 2, vector space R d Is determined by the word embedding vector.
The step 3 is specifically as follows: converting the source text expected to execute abstract into text vector c through the guiding module, inputting the text vector c into a guiding network, and enabling the text vector c to be micro-processedControl is performed by the director network for each time X t-1 Run gradient update generated by association of a director network parameterized control conditions text and a target text, where p (c|X t-1 ) Representing a sampling distribution X at the moment of generating t-1 by a diffusion model t-1 Probability of target text being c under the condition, +.>Representing the diffusion model at X when the step length is the t-th step in the process of sampling t-1 And performing gradient updating on the data, wherein a guide network used by the guide module is a fine-tuning bert-base-Chinese model.
In the sampling process of the diffusion model in the step 4, the initial T moment is input into a vector space R d Random gaussian noise of (a), for each moment X during the whole sampling process t-1 Running gradient update, finally ending at sampling X 0 The abstract text vector f input in the time generation step 5 θ (x 0 T). The sampling mode used in the sampling process is DDIM sampling, and the final X is generated by a denoising diffusion implicit model sampling method 0 Abstract text vector f of moment θ (x 0 ,t)。
In step 5, the control module is used to control the prediction vector fθ (x) generated in step 4 according to the control function t-1 T) mapping to its nearest real word embedding vector, generating a final prediction vectorTo reduce the predictive vector->Difference from the real word embedding vector, and prediction vector fθ (x t-1 The difference between t) and the real word embedding vector is controlled by a control function. The process is controlled by a control function Squeeze (fθ (x t T) to obtain a predictive vector ∈>Causing a predictive vectorThe vector is embedded by approximating the true word, thereby completing the task of controlling the accuracy of generating the vector.
The step 6 is specifically as follows: embedding the text vector according to the final predicted word generated in step 5And according to word embedding pre-coding of the input module, converting the word embedding pre-coding into text information, and outputting a text abstract according to a text sequence.
Table 1 below is text information generated after input of source text and a text summary obtained from the source text:
TABLE 1
In summary, the invention uses a natural language processing model to perform word segmentation operation on the source text; then, pre-encoding and converting the text data into sentence representation by the pre-processed source text through an input module by utilizing a word embedding method, and then mapping the text vector onto a space dimension capable of being continuously transformed by utilizing an evaluation module to realize controllable linear continuous transformation of the text vector; and finally, acquiring a target text through a guiding module, guiding by using a gradient in the sampling process of the diffusion model to generate a sample, reducing errors of a generated text vector and a real text vector by using a control module, and finally generating a text abstract through an output module according to the acquired sample vector. Compared with the prior art, the method and the device realize text abstract generation based on the generation model, are beneficial to improving the accuracy of the text abstract and promote the semantic generalization of the model.
In summary, the invention can fully utilize the context information of the input text to generate the high-quality text abstract, thereby improving the quality of the text abstract; meanwhile, the method has strong generalization capability, can be suitable for various input texts with different types, including long texts and short texts, and has strong adaptability; the semantic information of the input text can be accurately captured, the problem of omission or misunderstanding possibly occurring in the traditional keyword-based extraction method is avoided, and the accuracy of the text abstract is improved, so that a user can obtain text abstract data with higher quality.
The above embodiments are only for illustrating the technical solution of the present invention and not for limiting the same, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications and equivalents may be made thereto without departing from the spirit and scope of the technical solution of the present invention.

Claims (9)

1. A text abstract generation method based on a diffusion model is characterized by mainly comprising the following steps:
step 1, preprocessing text word segmentation: preliminary preprocessing is carried out on the corpus by using a natural language processing model in an input module, and the corpus is precoded by using a word embedding mode to obtain sentence representation W= [ W ] 1 ,w 2 ,w 3 ,...,w n ]Where W is the encoded set of sentence representations, W n A code representing an nth word segment in the sentence;
step 2, generating a text vector: using an auxiliary function in the evaluation module to represent the sentence generated in step 1 with w= [ W ] 1 ,w 2 ,w 3 ,...,w n ]Conversion to word embedding vector EMB (W) = [ EMB (W) 1 ),...,EMB(w n )]∈R d Mapping each text vector in the word-embedded vector to a vector space R d In each text vector can be represented in a vector space R d Wherein EMB (W) is a set of word-embedded vectors transformed from sentence-representation encoding W, EMB (W) n ) Embedding vectors for words generated by the code conversion of nth segmented words in sentences;
step 3, generating a condition guide: inputting the text vectors generated by the source text into the diffusion model by means of the guidance module in order to introduce additional conditional inputs co into the diffusion model θ (x t Y) and then for each moment X during diffusion model sampling t-1 Gradient update of a network of individual run condition guided modules, where X t Representing the predicted distribution of the diffusion model generated at time t, y representing the conditional input of the diffusion model introduced at time t, co θ Representing the model at the extra conditional input y and the predictive distribution X t A conditional distribution generated by a conditional model parameter θ;
step 4, sampling generation: in the diffusion model sampling process, the initial input is vector space R d A random Gaussian noise in the sample, for each moment X in the whole sampling process t-1 Running gradient updates to generate a prediction vector f θ (x t-1 T), where t represents the time t, x after the model samples t steps t-1 Representing the sampling distribution at time t-1 generated by the diffusion model at time t, f θ Representing a model by x t-1 Distributing the prediction vector generated at the time T, and executing the whole sampling process within the range of T:0;
step 5, controlling vector errors: the control module is used for generating the prediction vector f finally generated in the step 4 according to the control function θ (x 0 T) mapping to its nearest real word embedding vector, generating a final prediction vectorWherein x is 0 Final spreading for the sampling processSampling distribution of 0 moment generated by the scattered model;
step 6, generating a summary: final prediction vector generated according to step 5And (3) word embedding pre-coding of the input module is converted into text information, and a text abstract is output according to a text sequence.
2. The text abstract generation method based on the diffusion model according to claim 1, wherein: in step 1, a trained fine-tuning bert-base-Chinese model is used in an input module to perform sentence segmentation and word segmentation on a corpus, and the corpus is precoded in a word embedding mode to obtain sentence representation W= [ W ] 1 ,w 2 ,w 3 ,...,w n ]。
3. The text abstract generation method based on the diffusion model according to claim 1, wherein: in step 2, vector space R d Is determined by the word embedding vector.
4. The text abstract generation method based on the diffusion model according to claim 1, wherein the step 3 is specifically: converting the source text expected to execute abstract into text vector c through the guiding module, inputting the text vector c into a guiding network, and enabling the text vector c to be micro-processedControl is performed by the director network for each time X t-1 Run gradient update generated by association of a director network parameterized control conditions text and a target text, where p (c|X t-1 ) Representing a sampling distribution X at the moment of generating t-1 by a diffusion model t-1 Probability of target text being c under the condition, +.>Representing the diffusion model at X when the step length is the t-th step in the process of sampling t-1 And performing gradient updating.
5. The text abstract generation method based on the diffusion model according to claim 4, wherein: the bootstrap network used by the bootstrap module is a fine-tuning bert-base-chipset model.
6. The text abstract generation method based on the diffusion model according to claim 1, wherein: in the sampling process of the diffusion model in the step 4, the initial T moment is input into a vector space R d Random gaussian noise of (a), for each moment X during the whole sampling process t-1 Running gradient update, finally ending at sampling X 0 The abstract text vector f input in the time generation step 5 θ (x 0 ,t)。
7. The diffusion model-based text summarization method of claim 6 wherein: the sampling mode used in the sampling process in the step 4 is DDIM sampling, and the final X is generated by a denoising diffusion implicit model sampling method 0 Abstract text vector f of moment θ (x 0 ,t)。
8. The text abstract generation method based on the diffusion model according to claim 1, wherein: in step 5, the control module is used for converting the prediction vector f finally generated in step 4 according to the control function θ (x t-1 T) mapping to its nearest real word embedding vector, generating a final prediction vectorTo reduce the predictive vector->Differences from the real word embedding vector.
9. The text abstract generating method based on the diffusion model according to claim 8,the method is characterized in that: in step 5, the control function is used to control the sequence (f θ (x t T)) to obtain a predictive vectorLet predictive vector +.>Approximating a real word embedding vector.
CN202310419493.1A 2023-04-19 2023-04-19 Text abstract generation method based on diffusion model Pending CN116484858A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310419493.1A CN116484858A (en) 2023-04-19 2023-04-19 Text abstract generation method based on diffusion model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310419493.1A CN116484858A (en) 2023-04-19 2023-04-19 Text abstract generation method based on diffusion model

Publications (1)

Publication Number Publication Date
CN116484858A true CN116484858A (en) 2023-07-25

Family

ID=87226310

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310419493.1A Pending CN116484858A (en) 2023-04-19 2023-04-19 Text abstract generation method based on diffusion model

Country Status (1)

Country Link
CN (1) CN116484858A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117131187A (en) * 2023-10-26 2023-11-28 中国科学技术大学 Dialogue abstracting method based on noise binding diffusion model

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117131187A (en) * 2023-10-26 2023-11-28 中国科学技术大学 Dialogue abstracting method based on noise binding diffusion model
CN117131187B (en) * 2023-10-26 2024-02-09 中国科学技术大学 Dialogue abstracting method based on noise binding diffusion model

Similar Documents

Publication Publication Date Title
CN111522839B (en) Deep learning-based natural language query method
CN113987209A (en) Natural language processing method and device based on knowledge-guided prefix fine tuning, computing equipment and storage medium
CN111177376A (en) Chinese text classification method based on BERT and CNN hierarchical connection
CN110162766B (en) Word vector updating method and device
CN110442880B (en) Translation method, device and storage medium for machine translation
CN115599901B (en) Machine question-answering method, device, equipment and storage medium based on semantic prompt
WO2023137911A1 (en) Intention classification method and apparatus based on small-sample corpus, and computer device
CN115618022B (en) Low-resource relation extraction method based on data synthesis and two-stage self-training
CN116484858A (en) Text abstract generation method based on diffusion model
CN114692621A (en) Method for explaining influence function from sequence to sequence task based on sample in NLP
Elbedwehy et al. Efficient Image Captioning Based on Vision Transformer Models.
CN110609895B (en) Sample automatic generation method for actively selecting examples to conduct efficient text classification
CN116303966A (en) Dialogue behavior recognition system based on prompt learning
CN116485962A (en) Animation generation method and system based on contrast learning
CN114239575B (en) Statement analysis model construction method, statement analysis method, device, medium and computing equipment
CN114743539A (en) Speech synthesis method, apparatus, device and storage medium
CN115080736A (en) Model adjusting method and device of discriminant language model
CN113220892A (en) BERT-based self-adaptive text classification method and device
CN112015921A (en) Natural language processing method based on learning-assisted knowledge graph
CN110619118A (en) Automatic text generation method
CN111967580B (en) Low-bit neural network training method and system based on feature migration
CN116975344B (en) Chinese character library generation method and device based on Stable diffration
CN116070638B (en) Training updating method and system for Chinese sentence feature construction
CN114492332A (en) Controlled text generation method based on hidden variable manipulation of variational self-encoder
US20240005905A1 (en) End-to-end natural and controllable emotional speech synthesis methods

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination