CN112906385A - Text abstract generation method, computer equipment and storage medium - Google Patents

Text abstract generation method, computer equipment and storage medium Download PDF

Info

Publication number
CN112906385A
CN112906385A CN202110489771.1A CN202110489771A CN112906385A CN 112906385 A CN112906385 A CN 112906385A CN 202110489771 A CN202110489771 A CN 202110489771A CN 112906385 A CN112906385 A CN 112906385A
Authority
CN
China
Prior art keywords
text
sentence
calculating
vector
segment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110489771.1A
Other languages
Chinese (zh)
Other versions
CN112906385B (en
Inventor
杨德杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ping An Technology Shenzhen Co Ltd
Original Assignee
Ping An Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ping An Technology Shenzhen Co Ltd filed Critical Ping An Technology Shenzhen Co Ltd
Priority to CN202110489771.1A priority Critical patent/CN112906385B/en
Publication of CN112906385A publication Critical patent/CN112906385A/en
Application granted granted Critical
Publication of CN112906385B publication Critical patent/CN112906385B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention relates to the technical field of artificial intelligence, and provides a text abstract generating method, computer equipment and a storage medium, wherein the text abstract generating method comprises the following steps: performing word segmentation processing on the text to obtain target keywords; generating statement vectors of statements in the text and segment vectors of text segments according to the word vectors of the target keywords; calculating statement weight of the statement in the text according to the statement vector; coding the segment vector to obtain a hidden state of the text segment; obtaining the attention weight of the target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each moment; calculating according to the attention weight of the target keyword and the corresponding proxy weight to obtain vocabulary probability distribution at the time t, wherein the vocabulary probability distribution represents the target keyword appearing at the kth position, and the proxy weight is the sum of the sentence weights; and generating a text abstract according to the vocabulary probability distribution. The method and the device can accurately generate the text abstract, and the generated text abstract has strong readability and good effect on long texts.

Description

Text abstract generation method, computer equipment and storage medium
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to a text abstract generating method, computer equipment and a storage medium.
Background
Training courses are often long, e.g., typically over an hour, and can assist a user in quick review and review after class by generating a text summary of the training course.
The inventor finds that the existing text abstract generation mode is divided into a generation mode and an extraction mode, the generation mode text abstract method is used for generating a new abstract after understanding an original document, readability is strong, but correctness of the generated abstract is low, the existing generation mode abstract method is usually focused on a short-course text and has poor application effect on a long-course text, and the extraction mode text abstract method is used for extracting key words or key sentences of the original document and combining the key words or key sentences according to importance degree to form the abstract, but readability of the generated abstract is poor, and contained information amount is small.
Disclosure of Invention
In view of the above, there is a need for a method, a computer device and a storage medium for generating a text abstract, which can accurately generate a text abstract, and the generated text abstract has high readability and good effect on long texts.
The first aspect of the present invention provides a text summary generating method, including:
performing word segmentation processing on a text to obtain a target keyword, and acquiring a word vector of the target keyword;
generating statement vectors of statements in the text according to the word vectors, and generating segment vectors of text segments in the text according to the word vectors;
calculating statement weight of the statement in the text according to the statement vector;
coding the segment vector to obtain the hidden state of the text segment;
obtaining attention weight of a target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each moment;
calculating according to the attention weight of the target keyword and a corresponding proxy weight to obtain vocabulary probability distribution at the time t, wherein the vocabulary probability distribution represents the target keyword appearing at the kth position of the text abstract, and the proxy weight is the sum of statement weights of statements in the text segment;
and generating a text abstract according to the vocabulary probability distribution.
In an optional embodiment, the calculating the sentence weight of the sentence in the text according to the sentence vector comprises:
acquiring a first maximum boundary correlation degree of the statement;
acquiring a second maximum boundary correlation degree of other sentences except the sentence in the text;
and obtaining the sentence weight of the sentence according to the first maximum boundary correlation degree and the second maximum boundary correlation degree.
In an optional embodiment, the calculating of the first maximum boundary correlation includes:
calculating a first similarity between the sentence and the text according to the sentence vector of the sentence;
calculating a second similarity between the statement and the rest statements according to the statement vectors of the statement and the statement vectors of the rest statements;
and calculating to obtain the first maximum boundary correlation degree according to the first similarity degree and the second similarity degree.
In an optional embodiment, the calculating the first similarity of the sentence and the text according to the sentence vector of the sentence includes: calculating a first feature representation of the sentence according to the word vector in the sentence; calculating a second feature representation of the text according to the word vectors in the text; and calculating to obtain a first similarity according to the first feature representation and the second feature representation by adopting a similarity calculation model.
In an optional embodiment, the calculating the second similarity between the sentence and the rest of the sentences according to the sentence vectors of the sentences and the sentence vectors of the rest of the sentences includes: calculating third feature representations of the rest sentences according to the word vectors in the rest sentences; and calculating to obtain a second similarity according to the first characteristic representation and the third characteristic representation by adopting the similarity calculation model.
In an optional implementation manner, the calculating the first maximum boundary correlation according to the first similarity and the second similarity includes:
obtaining a first value according to a preset hyper-parameter and the first similarity;
determining a maximum value of the second similarity;
obtaining a second value according to the preset hyper-parameter and the maximum value;
and obtaining a first maximum boundary correlation degree according to the first value and the second value.
In an optional embodiment, the calculating the sentence weight of the sentence in the text according to the sentence vector comprises:
calculating a sum of the second maximum boundary correlations;
calculating a ratio of the first maximum boundary correlation to the sum;
and mapping the ratio by using a preset function to obtain the statement weight of the statement.
In an alternative embodiment of the method according to the invention,
the encoding the segment vector to obtain the hidden state of the text segment comprises:
coding the segment vector through a first-stage bidirectional LSTM model to obtain a hidden state of a target keyword in the text segment;
and coding the hidden state of the target keyword through a second-stage bidirectional LSTM model to obtain the hidden state of the text segment.
In an optional embodiment, the generating the text excerpt according to the lexical probability distribution includes:
for any moment, acquiring the maximum probability of the probability distribution at the any moment;
determining the target keywords corresponding to the maximum probability as the target keywords in the text abstract;
and combining the target keywords in the text abstract according to a time sequence to obtain the text abstract.
A second aspect of the invention provides a computer device comprising a processor for implementing the text summary generation method when executing a computer program stored in a memory.
A third aspect of the present invention provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the text digest generation method.
In summary, according to the text abstract generation method, the computer device, and the storage medium of the present invention, for a text requiring text abstract generation, a word vector of each target keyword in the text is extracted, a sentence vector is generated for each sentence in the text and a segment vector is generated for each text segment according to the word vector, so that a sentence weight of each sentence in the whole text is obtained according to the sentence vector, then the segment vector is encoded to obtain a hidden state of the text segment, an attention weight of the target keyword in the text segment is obtained according to the hidden state of the text segment and the hidden state vector at each time, and finally a vocabulary probability distribution at time t is obtained according to the attention weight of the target keyword and a corresponding proxy weight, where the vocabulary probability distribution represents the target keyword appearing at the k-th position. The generated text abstract not only considers the similarity between sentences and texts, but also considers the redundancy of abstract sentences, the weight of text segments formed by sentences with large meaning repeatability is relatively low, the importance of information in the process of a decoder module is also low, the possibility of repeated meaning target keywords in the finally obtained abstract is also lower, and the generated text abstract is more accurate, clear and concise in meaning.
Drawings
Fig. 1 is a flowchart of a text summary generating method according to an embodiment of the present invention.
Fig. 2 is a schematic diagram of a network architecture of a generative model provided in an embodiment of the present invention.
Fig. 3 is a schematic structural diagram of an agent in the generative model according to the second embodiment of the present invention.
Fig. 4 is a block diagram of a text abstract generating apparatus according to a second embodiment of the present invention.
Fig. 5 is a schematic structural diagram of a computer device according to a third embodiment of the present invention.
Detailed Description
In order that the above objects, features and advantages of the present invention can be more clearly understood, a detailed description of the present invention will be given below with reference to the accompanying drawings and specific embodiments. It should be noted that the embodiments of the present invention and features of the embodiments may be combined with each other without conflict.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. The terminology used in the description of the invention herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention.
The text abstract generating method provided by the embodiment of the invention is executed by computer equipment, and correspondingly, the text abstract generating device runs in the computer equipment.
Fig. 1 is a flowchart of a text summary generating method according to an embodiment of the present invention. The text abstract generating method specifically comprises the following steps, and the sequence of the steps in the flowchart can be changed and some steps can be omitted according to different requirements.
S11, performing word segmentation processing on the text to obtain a target keyword, and obtaining a word vector of the target keyword.
The text refers to the expression form of written language, and may be a sentence, a paragraph, or an article. If the method is applied to a training scene, the text is a training course text, and the training course text can be derived from a training video.
In an optional implementation manner, the text is a training course text, the performing word segmentation processing on the text to obtain a target keyword, and obtaining a word vector of the target keyword includes:
according to a pre-established training course text non-use word library, drying the training course text to obtain a standard text;
performing word segmentation processing on the standard text to obtain a target keyword;
and extracting a word vector of the target keyword by using the trained word2vec model.
A computer device may first obtain a training video, extract a training voice of the training video, and recognize a voice text of the training voice using a voice recognition technique (e.g., a Sphinx system) to obtain a training course text.
The computer device may perform a word segmentation process on the text using a crust word segmentation tool to obtain a plurality of target keywords.
Since there is a lot of environmental noise during the actual training, and thus the training course text is obtained with a lot of noise information, the computer device creates a training course text dead word library in advance, the training course text dead word library including: and (3) matching the training course text with the training course text stop word library by using stop words such as word-atmosphere words, personal names, exclamation words and the like and nonsense words, and filtering words successfully matched with the training course text stop word library in the training course text, so that the effect of removing dryness of the training course text is realized, and the standard text is obtained.
Since the training course text contains many professional vocabularies, the training related professional vocabularies (e.g., insurance related vocabularies) need to be added on the basis of the common Chinese lexicon to obtain the course corpus lexicon. The computer equipment can train a word2vec model according to words in a course corpus lexicon by adopting a skip-gram training mode, and each target keyword in the standard text is represented into one by the trained word2vec model
Figure DEST_PATH_IMAGE001
A word vector of dimensions.
The computer device may use the textDictoryCorpus tool in genim to manage the created course text deactivation thesaurus and the course corpus thesaurus.
S12, generating statement vectors of the sentences in the text according to the word vectors, and generating segment vectors of the text segments in the text according to the word vectors.
And for any statement, combining the word vectors of all target keywords in the statement to obtain the statement vector of the statement. For example, for statements
Figure 645199DEST_PATH_IMAGE002
To describe the sentence
Figure 562339DEST_PATH_IMAGE002
The word vectors of all the target keywords in the sentence are combined to obtain the sentence
Figure 342077DEST_PATH_IMAGE002
Statement vector of
Figure DEST_PATH_IMAGE003
Wherein, in the step (A),
Figure 737286DEST_PATH_IMAGE004
presentation statement
Figure 551658DEST_PATH_IMAGE002
To middle
Figure DEST_PATH_IMAGE005
A word vector for each of the target keywords,
Figure 813269DEST_PATH_IMAGE005
as a sentence
Figure 80302DEST_PATH_IMAGE002
The number of target keywords included.
In generalIn other words, the training course text belongs to a long text, and the computer device divides the training course text into a plurality of text segments in order to better extract a text abstract of the training course text. In specific implementation, periods of the training course text may be randomly used as division points, and the training course text may be divided into a plurality of text segments, where each text segment may include one sentence, or may include two or more sentences. Multiple text segments can be combined into a text segment sequence, denoted as
Figure 279202DEST_PATH_IMAGE006
Wherein, in the step (A),
Figure DEST_PATH_IMAGE007
representing the first in training course text
Figure 948081DEST_PATH_IMAGE008
And (4) each text segment. Since the text segment is composed of one or more sentences each including a plurality of target keywords, the text segment can be regarded as a sequence composed of a plurality of target keywords, and is denoted as
Figure DEST_PATH_IMAGE009
Wherein, in the step (A),
Figure 879128DEST_PATH_IMAGE010
a segment of text is represented that is,
Figure DEST_PATH_IMAGE011
a target keyword is represented by a target keyword,
Figure 633457DEST_PATH_IMAGE012
representing text segments
Figure 370469DEST_PATH_IMAGE010
The number of target keywords included, each target keyword
Figure 893854DEST_PATH_IMAGE011
Is one
Figure 323698DEST_PATH_IMAGE001
Word vector of dimension
Figure DEST_PATH_IMAGE013
The segment vector is
Figure 503007DEST_PATH_IMAGE012
An
Figure 778131DEST_PATH_IMAGE001
A vector of dimensions.
And S13, calculating the sentence weight of the sentence in the text according to the sentence vector.
In this embodiment, by means of the idea of the abstract mode, text information may be extracted from the text as much as possible, that is, a sentence with a large amount of information is extracted from the text, so as to generate a text abstract by means of the idea of the abstract mode.
The computer device calculates the sentence weight of each sentence in the text according to the sentence vector of each sentence, the larger the sentence weight is, the more important the corresponding sentence is in the text, and the smaller the sentence weight is, the more unimportant the sentence is in the text.
In an alternative embodiment, calculating the sentence weight of the sentence in the text from the sentence vector comprises:
acquiring a first maximum boundary correlation degree of the statement;
acquiring a second maximum boundary correlation degree of other sentences except the sentence in the text;
and obtaining the sentence weight of the sentence according to the first maximum boundary correlation degree and the second maximum boundary correlation degree.
Wherein the calculation process of the first maximum boundary correlation degree comprises: calculating a first similarity between the sentence and the text according to the sentence vector of the sentence, calculating a second similarity between the sentence and the other sentences according to the sentence vector of the sentence and the sentence vectors of the other sentences, and calculating the first maximum boundary correlation according to the first similarity and the second similarity.
Wherein the sentence is relative to the rest sentences, and the rest sentences refer to all sentences except the sentence in the text. For example, when the sentence is the first sentence in the text, the remaining sentences are the second sentence to the last sentence in the text. And when the sentence is the second sentence in the text, the other sentences are the first sentence, the third sentence and the last sentence in the text.
It should be understood that, since the sentence is relative to the rest of the sentences, the calculation process of the second maximum boundary correlation and the calculation process of the first maximum boundary correlation are not elaborated.
In an alternative embodiment, the computing, by the computer device, a first similarity of the sentence to the text from the sentence vector of the sentence comprises: calculating a first feature representation of the sentence according to the word vector in the sentence; calculating a second feature representation of the text according to the word vectors in the text; and calculating to obtain a first similarity according to the first feature representation and the second feature representation by adopting a similarity calculation model.
Illustratively, assume that a sentence includes 10 target keywords, a text includes 1000 target keywords, and a word vector for each target keyword is
Figure 421601DEST_PATH_IMAGE014
Dimension, the first mean vector obtained by adding and re-averaging the word vectors of the 10 target keywords is still
Figure 756768DEST_PATH_IMAGE014
Similarly, the second mean vector obtained by adding and averaging the word vectors of the 1000 target keywords is still
Figure 485689DEST_PATH_IMAGE014
And (5) maintaining. The first mean vector is a first feature representation of the sentence, and the second mean vector is a second feature representation of the text, that is, after the sentence and the text are respectively converted into feature representations, a first similarity between the sentence and the text can be calculated.
In an alternative embodiment, the computing, by the computer device, the second similarity of the sentence to the remaining sentences according to the sentence vectors of the sentence and the sentence vectors of the remaining sentences comprises: calculating third feature representations of the rest sentences according to the word vectors in the rest sentences; and calculating to obtain a second similarity according to the first characteristic representation and the third characteristic representation by adopting the similarity calculation model.
And taking a third mean value vector obtained by adding and averaging the word vectors of the target keywords in the rest sentences as a third feature representation of the rest sentences.
The similarity calculation model may be a preconfigured similarity calculation function, such as an euclidean distance, a cosine angle, and the like.
In an optional implementation manner, the calculating the first maximum boundary correlation according to the first similarity and the second similarity includes:
obtaining a first value according to a preset hyper-parameter and the first similarity;
determining a maximum value of the second similarity;
obtaining a second value according to the preset hyper-parameter and the maximum value;
and obtaining a first maximum boundary correlation degree according to the first value and the second value.
The computer device stores a first formula in advance, and the first maximum boundary correlation obtained by calculation according to the first similarity and the second similarity can be represented by the following first formula:
Figure DEST_PATH_IMAGE015
wherein the content of the first and second substances,
Figure 564504DEST_PATH_IMAGE016
presentation statement
Figure 164DEST_PATH_IMAGE002
The maximum degree of correlation of the boundary of (c),
Figure DEST_PATH_IMAGE017
a feature representation (i.e., the second mean vector) representing text,
Figure 771811DEST_PATH_IMAGE018
as a sentence
Figure 722450DEST_PATH_IMAGE002
The feature representation of (i.e., the first mean vector),
Figure DEST_PATH_IMAGE019
as a sentence
Figure 604955DEST_PATH_IMAGE020
The feature representation of (i.e., the third mean vector),
Figure DEST_PATH_IMAGE021
is a preset hyper-parameter and is continuously optimized in the subsequent training process,
Figure 393657DEST_PATH_IMAGE022
presentation statement
Figure 336206DEST_PATH_IMAGE002
A first degree of similarity to the text,
Figure DEST_PATH_IMAGE023
presentation statement
Figure 774140DEST_PATH_IMAGE002
And sentences
Figure 194757DEST_PATH_IMAGE020
Second phase ofAnd (4) degree.
Figure 667327DEST_PATH_IMAGE022
Indicating the relevance of a single sentence to the entire text,
Figure 780776DEST_PATH_IMAGE024
representing sentences in text
Figure 643690DEST_PATH_IMAGE002
And sentences
Figure 867998DEST_PATH_IMAGE020
The correlation between them.
Figure 929495DEST_PATH_IMAGE024
The larger, the sentence
Figure 213846DEST_PATH_IMAGE002
And sentences
Figure 626373DEST_PATH_IMAGE020
The stronger the correlation between them, the statement is indicated
Figure 654372DEST_PATH_IMAGE002
And sentences
Figure 835954DEST_PATH_IMAGE020
The larger the redundancy is, the more redundant the generated text abstract is;
Figure 963310DEST_PATH_IMAGE024
the smaller, the sentence
Figure 863133DEST_PATH_IMAGE002
And sentences
Figure 429244DEST_PATH_IMAGE020
The weaker the correlation between them, the statement
Figure 730912DEST_PATH_IMAGE002
And sentences
Figure 91486DEST_PATH_IMAGE020
The less redundant the text excerpt generated, the less redundant.
In an optional embodiment, the calculating the sentence weight of the sentence in the text according to the sentence vector comprises:
calculating a sum of the second maximum boundary correlations;
calculating a ratio of the first maximum boundary correlation to the sum;
and mapping the ratio by using a preset function to obtain the statement weight of the statement.
The computer device stores a second formula in advance, and the statement weight of the statement calculated according to the first maximum boundary correlation and the second maximum boundary correlation can be represented by the following second formula:
Figure DEST_PATH_IMAGE025
wherein the content of the first and second substances,
Figure 478605DEST_PATH_IMAGE016
presentation statement
Figure 848407DEST_PATH_IMAGE002
The maximum degree of correlation of the boundary of (c),
Figure 676685DEST_PATH_IMAGE026
presentation statement
Figure 473740DEST_PATH_IMAGE020
The maximum degree of correlation of the boundary of (c),
Figure DEST_PATH_IMAGE027
presentation statement
Figure 82576DEST_PATH_IMAGE002
The sentence weight of (1).
After calculating the maximum boundary correlation degree of each sentence in the text, the computer device calculates the sum of the maximum boundary correlation degrees of other sentences except the any sentence in the text for any sentence, calculates the ratio of the maximum boundary correlation degree of the any sentence to the sum, and finally performs softmax function processing to obtain the sentence weight (or called probability value) of the any sentence. The larger the ratio, the greater the relevance of the corresponding sentence to the text, and the higher the redundancy. The smaller the ratio, the weaker the correlation of the corresponding sentence with the text, and the lower the redundancy.
Combining the sentence weights of all sentences in the text to obtain a sentence weight vector representation which is recorded as
Figure 256068DEST_PATH_IMAGE028
Wherein N is the number of sentences in the text,
Figure DEST_PATH_IMAGE029
as a sentence
Figure 266750DEST_PATH_IMAGE002
The sentence prediction probability is used as a text abstract in the text.
S14, encoding the segment vector to obtain the hidden state of the text segment.
The segment vectors may be encoded using an agent in an encoder module that generates a model, resulting in a hidden state of the text segment.
In this embodiment, the network architecture of the generative model is as shown in fig. 2 below, and the generative model includes an Encoder module (e.g., an Encoder part in fig. 2) located at a bottom layer and a Decoder module (e.g., a Decoder part in fig. 2) located at an upper layer, where the Encoder module includes a plurality of agents (e.g., Agent1, Agent2, Agent m in fig. 2) independent of each other, each Agent includes a bidirectional LSTM (bilstm) model, and the Decoder module includes a multi-layer Attention mechanism layer (e.g., an extraction model Attention in fig. 2) and an Agent Attention mechanism layer (e.g., an Agent Attention in fig. 2) and an LSTM model.
The computer device configures the number of agents in the encoder module according to the number of text segments, that is, how many text segments exist, and how many agents are correspondingly configured in the encoder module, so that the text segments correspond to the agents one to one. The computer equipment inputs each text segment into one agent in the encoder module, encodes the text segment through the agent to obtain an intermediate vector (namely a hidden state), and transmits the intermediate vector to other agents, so that different agents can share global context information of different contents in the text. Each agent outputs an intermediate vector, and all intermediate vectors form an intermediate vector sequence. And finally, decoding the intermediate vectors output by all the agents through a decoder module, namely decoding the intermediate vector sequence to obtain the probability distribution of the text abstract, thereby obtaining the final text abstract. Referring to FIG. 2, a computer device segments text
Figure 673854DEST_PATH_IMAGE030
Inputting Agent1 in encoder module, and text segment by Agent1
Figure 769985DEST_PATH_IMAGE030
Coding to obtain text segments
Figure 481590DEST_PATH_IMAGE030
Hidden state of (2), segmenting text
Figure DEST_PATH_IMAGE031
Inputting Agent2 in encoder module, and text segment by Agent2
Figure 346777DEST_PATH_IMAGE031
Coding to obtain text segments
Figure 485635DEST_PATH_IMAGE031
Hidden state of (2), segmenting text
Figure 69063DEST_PATH_IMAGE007
Inputting proxy AgentM in encoder module, and aligning text segment by proxy AgentM
Figure 256462DEST_PATH_IMAGE007
Coding to obtain text segments
Figure 241735DEST_PATH_IMAGE007
Hidden state of (1).
In an alternative embodiment, as shown in fig. 3, the proxy includes two-stage bidirectional LSTM models (a first-stage bidirectional LSTM model and a second-stage bidirectional LSTM model), and the encoding the segment vector to obtain the hidden state of the text segment includes:
coding the segment vector through a first-stage bidirectional LSTM model to obtain a hidden state of a target keyword in the text segment;
and coding the hidden state of the target keyword through a second-stage bidirectional LSTM model to obtain the hidden state of the text segment.
The BilSTM in each agent is a two-stage bi-directional LSTM model, where the first stage is a single-layer bi-directional LSTM (e.g., Local Encoder in FIG. 3), the output of each Local Encoder is input to a second stage context coder (e.g., context Encoder in FIG. 3), which is a multi-layer bi-directional LSTM structure called context Encoder.
The procedure of the first stage is as follows:
passing word vector
Figure 551494DEST_PATH_IMAGE013
Combined text segments
Figure 622218DEST_PATH_IMAGE010
And is fed as a whole into the first stage single layer bi-directional LSTM for encoding. The coding formula of the first-stage single-layer bidirectional LSTM is as follows:
hidden state in the middle:
Figure 675625DEST_PATH_IMAGE032
hidden state of output:
Figure DEST_PATH_IMAGE033
wherein the content of the first and second substances,
Figure 515405DEST_PATH_IMAGE034
as target key words
Figure 933748DEST_PATH_IMAGE011
Is a hidden state, H is a hidden state
Figure 226189DEST_PATH_IMAGE034
The dimension (c) of (a) is,
Figure DEST_PATH_IMAGE035
continuously optimizing the parameters in the training process;
Figure 83286DEST_PATH_IMAGE036
representing target keywords
Figure 777573DEST_PATH_IMAGE011
The intermediate hidden state resulting from the encoding of the bi-directional LSTM,
Figure DEST_PATH_IMAGE037
and the latter target keyword
Figure 429134DEST_PATH_IMAGE038
Is added to the intermediate hidden state of (1).
As can be seen from the above formula, each target keyword in the text segment
Figure 412134DEST_PATH_IMAGE011
The corresponding hidden states are from the middle hidden state of the previous and next target keywords and the word vector representation of the target keywords themselves.
The process of the second stage is as follows:
Figure DEST_PATH_IMAGE039
Figure 807343DEST_PATH_IMAGE040
each agent is at
Figure DEST_PATH_IMAGE041
Layer(s)
Figure 356136DEST_PATH_IMAGE042
In the context encoder, information received from an upper layer is jointly encoded. As can be seen from the above formulas
Figure 444178DEST_PATH_IMAGE041
The input to the layer context encoder comprises three parts: hidden state from contextual target keywords: (
Figure DEST_PATH_IMAGE043
Figure 711211DEST_PATH_IMAGE044
) (ii) a Hidden state of the previous layer output (
Figure DEST_PATH_IMAGE045
) (ii) a Information from other agents: (
Figure 346329DEST_PATH_IMAGE046
Output of
Figure DEST_PATH_IMAGE047
Wherein, in the step (A),
Figure 749629DEST_PATH_IMAGE048
is the other agent (1)
Figure DEST_PATH_IMAGE049
) Average value of layer output
Figure 8572DEST_PATH_IMAGE050
Figure DEST_PATH_IMAGE051
For other agents in (1)
Figure 700584DEST_PATH_IMAGE049
) The hidden state vector of the last target keyword generated by the layer at each moment. In the text segment
Figure 437596DEST_PATH_IMAGE010
In each target keyword
Figure 960981DEST_PATH_IMAGE011
In addition to learning the own contextual target keyword information, overall information from other text segments is also accepted.
The encoder module output
Figure 390826DEST_PATH_IMAGE052
Representing a text segment
Figure 366872DEST_PATH_IMAGE010
Is/are as follows
Figure 907575DEST_PATH_IMAGE012
A hidden state vector sequence for each target keyword.
And S15, obtaining the attention weight of the target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each moment.
In an alternative embodiment, the attention extracting mechanism layer in the decoder module of the generative model may be used to calculate the attention weight of the target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each time.
And S16, calculating the vocabulary probability distribution at the time t according to the attention weight of the target keyword and the corresponding proxy weight.
The vocabulary probability distribution at the time t can be calculated by using an agent attention mechanism layer in the decoder module according to the attention weight of the target keyword and the corresponding agent weight.
The vocabulary probability distribution represents target keywords appearing at the kth position of the text abstract, and the agent weight is the sum of statement weights of statements in the text segment.
The decoder module is of a one-way LSTM structure so that the last target keyword of each text segment can contain the state information of all previous target keywords. Initial input state of decoder module
Figure DEST_PATH_IMAGE053
Output hidden state vector for last layer of last target key of all agents of encoder module
Figure 488729DEST_PATH_IMAGE054
The hidden state vector contains all the text information encoded by the encoder module.
In specific implementation, a hidden state vector sequence generated by the last hidden layer of the agent is utilized
Figure 823895DEST_PATH_IMAGE052
The attention weight of each target keyword in the text segment is calculated by the following formula
Figure DEST_PATH_IMAGE055
Figure 552817DEST_PATH_IMAGE056
. Wherein the content of the first and second substances,
Figure DEST_PATH_IMAGE057
Figure 631631DEST_PATH_IMAGE058
all of which are hyper-parameters,
Figure DEST_PATH_IMAGE059
representing the state vector at time t. For each agent a in the agent attention mechanism layer, the attention weight of the agent is represented by the weight of each target keyword in the text segment:
Figure 129609DEST_PATH_IMAGE060
segmenting text
Figure 573359DEST_PATH_IMAGE010
The sentence weight (i.e., the sentence prediction probability) of the middle sentence is added as the proxy weight
Figure DEST_PATH_IMAGE061
All agents will be
Figure 789577DEST_PATH_IMAGE062
Weighted summation of the attention distributions to obtain the attention weight of the t-th word of the abstract
Figure DEST_PATH_IMAGE063
. Will be provided with
Figure 672082DEST_PATH_IMAGE064
Putting the hidden state vector of each moment into a multilayer perceptron to obtain the vocabulary probability distribution of the t moment
Figure DEST_PATH_IMAGE065
Figure 287933DEST_PATH_IMAGE066
Representing the target keyword appearing at the k-th position of the text excerpt.
Each time step t of the LSTM represents the location of the t-th target keyword of the generated text excerpt. Generating a hidden state at the current time
Figure 230481DEST_PATH_IMAGE059
Predicted new target keyword at the t-th position in the abstract
Figure 668416DEST_PATH_IMAGE066
Probability distribution of (2). In this process, a layered attention mechanism is introduced inside the LSTM structure, including the attention weight word attention of the target keyword and the attention weight agent attention of the agent.
The probability distribution of the obtained text abstract is
Figure DEST_PATH_IMAGE067
Wherein, in the step (A),
Figure 89033DEST_PATH_IMAGE068
representing the probability of each word in the dictionary appearing at the moment t, wherein V is the number of target keywords in the course corpus thesaurus, and each element
Figure DEST_PATH_IMAGE069
In this alternative embodiment, the weight calculated by the generative model not only takes into account the similarity between the sentence and the text, but also takes into account the redundancy of the abstract sentence, and this weight is combined with the weight of the text segment of the decoder module of the generative model, and compared with a common decoder, the probability finally calculated also includes the information of redundancy. Namely, the weight of the text segment composed of sentences with large meaning repeatability is relatively low, the importance of the information in the process of the decoder module is also low, the probability of repeated meaning target keywords in the finally obtained abstract is also lower, the generated text abstract is more accurate, and the meaning is clear and concise.
And S17, generating a text abstract according to the vocabulary probability distribution.
Since the probability distribution represents the probability of each target keyword in the text appearing in the text abstract, the higher the probability that the corresponding target keyword appears in the text abstract, and the lower the probability, the lower the probability that the corresponding target keyword appears in the text abstract. The computer device may determine a target keyword appearing in the text excerpt according to the probability in the probability distribution, thereby generating the text excerpt according to the target keyword.
In an optional embodiment, the generating the text excerpt according to the lexical probability distribution includes:
for any moment, acquiring the maximum probability of the probability distribution at the any moment;
determining the target keywords corresponding to the maximum probability as the target keywords in the text abstract;
and combining the target keywords in the text abstract according to a time sequence to obtain the text abstract.
Illustratively, for a certain time
Figure 561603DEST_PATH_IMAGE070
Selecting a time
Figure 347156DEST_PATH_IMAGE070
The target keyword corresponding to the maximum probability of being the first of the text abstract to be generated
Figure 272387DEST_PATH_IMAGE070
Generating text abstract finally according to the target key words
Figure DEST_PATH_IMAGE071
,
Figure 496695DEST_PATH_IMAGE072
Is a time of day
Figure DEST_PATH_IMAGE073
The maximum probability of the lower case corresponds to the target keyword,
Figure 823771DEST_PATH_IMAGE074
as a summary of the text
Figure DEST_PATH_IMAGE075
The number of target keywords contained therein.
After the computer equipment generates the text abstract, the generated text abstract can be displayed for a learner to review and review after a training class according to the generated text abstract, and a key course is selected for learning according to target keywords (serving as keywords) in the text abstract.
In summary, for a text requiring text summarization generation, the text is first segmented to obtain target keywords, then word vectors of each target keyword in the text are extracted, sentence vectors are generated for each sentence in the text and segment vectors are generated for each text segment in the text according to the word vectors, so as to obtain sentence weights of each sentence in the whole text according to the sentence vectors, then the segment vectors are encoded by adopting a trained LSTM model to obtain hidden states of the text segments, attention weights of the target keywords in the text segments are obtained according to the hidden states of the text segments and the hidden state vectors at each moment, vocabulary probability distribution at t moment is obtained according to the attention weights of the target keywords and corresponding proxy weights, and the vocabulary probability distribution represents the target keywords appearing at the kth position, and the proxy weight is the sum of the sentence weights, and finally, the text abstract is generated according to the vocabulary probability distribution. The generated text abstract not only considers the similarity between sentences and texts, but also considers the redundancy of abstract sentences, the weight of text segments formed by sentences with large meaning repeatability is relatively low, the importance of information in the process of a decoder module is also low, the possibility of repeated meaning target keywords in the finally obtained abstract is also lower, and the generated text abstract is more accurate, clear and concise in meaning.
It is emphasized that the text may be stored in a node of the blockchain in order to further ensure privacy and security of the text.
Fig. 2 is a block diagram of a text abstract generating apparatus according to a second embodiment of the present invention.
In some embodiments, the text summary generating device 40 may include a plurality of functional modules composed of computer program segments. The computer program of each program segment in the text summary generation apparatus 40 may be stored in a memory of a computer device and executed by at least one processor to perform the functions of text summary generation (described in detail in fig. 1).
In this embodiment, the text abstract generating device 40 may be divided into a plurality of functional modules according to the functions performed by the device. The functional module may include: a word segmentation processing module 401, a vector generation module 402, a weight calculation module 403, a vector encoding module 404, an attention calculation module 405, a probability distribution module 406, and a summary generation module 407. The module referred to herein is a series of computer program segments capable of being executed by at least one processor and capable of performing a fixed function and is stored in memory. In the present embodiment, the functions of the modules will be described in detail in the following embodiments.
The word segmentation processing module 401 is configured to perform word segmentation processing on a text to obtain a target keyword, and obtain a word vector of the target keyword.
The text refers to the expression form of written language, and may be a sentence, a paragraph, or an article. If the method is applied to a training scene, the text is a training course text, and the training course text can be derived from a training video.
In an optional embodiment, the text is a training course text, the word segmentation processing module 401 performs word segmentation processing on the text to obtain a target keyword, and obtaining a word vector of the target keyword includes:
according to a pre-established training course text non-use word library, drying the training course text to obtain a standard text;
performing word segmentation processing on the standard text to obtain a target keyword;
and extracting a word vector of the target keyword by using the trained word2vec model.
A computer device may first obtain a training video, extract a training voice of the training video, and recognize a voice text of the training voice using a voice recognition technique (e.g., a Sphinx system) to obtain a training course text.
The computer device may perform a word segmentation process on the text using a crust word segmentation tool to obtain a plurality of target keywords.
Since there is a lot of environmental noise during the actual training, and thus the training course text is obtained with a lot of noise information, the computer device creates a training course text dead word library in advance, the training course text dead word library including: and (3) matching the training course text with the training course text stop word library by using stop words such as word-atmosphere words, personal names, exclamation words and the like and nonsense words, and filtering words successfully matched with the training course text stop word library in the training course text, so that the effect of removing dryness of the training course text is realized, and the standard text is obtained.
Since the training course text contains many professional vocabularies, the training related professional vocabularies (e.g., insurance related vocabularies) need to be added on the basis of the common Chinese lexicon to obtain the course corpus lexicon. The computer equipment can train a word2vec model according to words in a course corpus lexicon by adopting a skip-gram training mode, and then each target keyword in the standard text is represented into an n-dimensional word vector through the trained word2vec model.
The computer device may use the textDictoryCorpus tool in genim to manage the created course text deactivation thesaurus and the course corpus thesaurus.
The vector generating module 402 is configured to generate a statement vector of a statement in the text according to the word vector, and generate a segment vector of a text segment in the text according to the word vector.
And for any statement, combining the word vectors of all target keywords in the statement to obtain the statement vector of the statement. For example, for statements
Figure 422636DEST_PATH_IMAGE002
To describe the sentence
Figure 772846DEST_PATH_IMAGE002
The word vectors of all the target keywords in the sentence are combined to obtain the sentence
Figure 535265DEST_PATH_IMAGE002
Statement vector of
Figure 716848DEST_PATH_IMAGE003
Wherein, in the step (A),
Figure 172100DEST_PATH_IMAGE004
presentation statement
Figure 71923DEST_PATH_IMAGE002
To middle
Figure 638033DEST_PATH_IMAGE005
A word vector for each of the target keywords,
Figure 611806DEST_PATH_IMAGE005
as a sentence
Figure 237959DEST_PATH_IMAGE002
The number of target keywords included.
Generally, the training course text belongs to a long text, and in order to better extract a text abstract of the training course text, the computer device divides the training course text into a plurality of text segments. In specific implementation, periods of the training course text may be randomly used as division points, and the training course text may be divided into a plurality of text segments, where each text segment may include one sentence, or may include two or more sentences. Multiple text segments can be combined into a text segment sequence, denoted as
Figure 625078DEST_PATH_IMAGE006
Wherein, in the step (A),
Figure 729300DEST_PATH_IMAGE007
representing the first in training course text
Figure 885475DEST_PATH_IMAGE008
And (4) each text segment. Since the text segment is composed of one or more sentences each including a plurality of target keywords, the text segment may be composed of a plurality of target keywordsConsidered as a sequence consisting of a plurality of target keywords, denoted as
Figure 682530DEST_PATH_IMAGE009
Wherein, in the step (A),
Figure 229049DEST_PATH_IMAGE010
a segment of text is represented that is,
Figure 402541DEST_PATH_IMAGE011
a target keyword is represented by a target keyword,
Figure 413222DEST_PATH_IMAGE012
representing text segments
Figure 381178DEST_PATH_IMAGE010
The number of target keywords included, each target keyword
Figure 477310DEST_PATH_IMAGE011
Is one
Figure 923335DEST_PATH_IMAGE001
Word vector of dimension
Figure 493250DEST_PATH_IMAGE013
The segment vector is
Figure 632108DEST_PATH_IMAGE012
An
Figure 215536DEST_PATH_IMAGE001
A vector of dimensions.
The weight calculating module 403 is configured to calculate a sentence weight of the sentence in the text according to the sentence vector.
In this embodiment, by means of the idea of the abstract mode, text information may be extracted from the text as much as possible, that is, a sentence with a large amount of information is extracted from the text, so as to generate a text abstract by means of the idea of the abstract mode.
The computer device calculates the sentence weight of each sentence in the text according to the sentence vector of each sentence, the larger the sentence weight is, the more important the corresponding sentence is in the text, and the smaller the sentence weight is, the more unimportant the sentence is in the text.
In an alternative embodiment, the calculating the sentence weight of the sentence in the text by the weight calculating module 403 according to the sentence vector includes:
acquiring a first maximum boundary correlation degree of the statement;
acquiring a second maximum boundary correlation degree of other sentences except the sentence in the text;
and obtaining the sentence weight of the sentence according to the first maximum boundary correlation degree and the second maximum boundary correlation degree.
Wherein the calculation process of the first maximum boundary correlation degree comprises: calculating a first similarity between the sentence and the text according to the sentence vector of the sentence, calculating a second similarity between the sentence and the other sentences according to the sentence vector of the sentence and the sentence vectors of the other sentences, and calculating the first maximum boundary correlation according to the first similarity and the second similarity.
Wherein the sentence is relative to the rest sentences, and the rest sentences refer to all sentences except the sentence in the text. For example, when the sentence is the first sentence in the text, the remaining sentences are the second sentence to the last sentence in the text. And when the sentence is the second sentence in the text, the other sentences are the first sentence, the third sentence and the last sentence in the text.
It should be understood that, since the sentence is relative to the rest of the sentences, the calculation process of the second maximum boundary correlation and the calculation process of the first maximum boundary correlation are not elaborated.
In an alternative embodiment, the computing, by the computer device, a first similarity of the sentence to the text from the sentence vector of the sentence comprises: calculating a first feature representation of the sentence according to the word vector in the sentence; calculating a second feature representation of the text according to the word vectors in the text; and calculating to obtain a first similarity according to the first feature representation and the second feature representation by adopting a similarity calculation model.
Illustratively, assume that a sentence includes 10 target keywords, a text includes 1000 target keywords, and a word vector for each target keyword is
Figure 465251DEST_PATH_IMAGE014
Dimension, the first mean vector obtained by adding and re-averaging the word vectors of the 10 target keywords is still
Figure 450525DEST_PATH_IMAGE014
Similarly, the second mean vector obtained by adding and averaging the word vectors of the 1000 target keywords is still
Figure 760284DEST_PATH_IMAGE014
And (5) maintaining. The first mean vector is a first feature representation of the sentence, and the second mean vector is a second feature representation of the text, that is, after the sentence and the text are respectively converted into feature representations, a first similarity between the sentence and the text can be calculated.
In an alternative embodiment, the computing, by the computer device, the second similarity of the sentence to the remaining sentences according to the sentence vectors of the sentence and the sentence vectors of the remaining sentences comprises: calculating third feature representations of the rest sentences according to the word vectors in the rest sentences; and calculating to obtain a second similarity according to the first characteristic representation and the third characteristic representation by adopting the similarity calculation model.
And taking a third mean value vector obtained by adding and averaging the word vectors of the target keywords in the rest sentences as a third feature representation of the rest sentences.
The similarity calculation model may be a preconfigured similarity calculation function, such as an euclidean distance, a cosine angle, and the like.
In an optional implementation manner, the calculating the first maximum boundary correlation according to the first similarity and the second similarity includes:
obtaining a first value according to a preset hyper-parameter and the first similarity;
determining a maximum value of the second similarity;
obtaining a second value according to the preset hyper-parameter and the maximum value;
and obtaining a first maximum boundary correlation degree according to the first value and the second value.
The computer device stores a first formula in advance, and the first maximum boundary correlation obtained by calculation according to the first similarity and the second similarity can be represented by the following first formula:
Figure 768691DEST_PATH_IMAGE015
wherein the content of the first and second substances,
Figure 822098DEST_PATH_IMAGE016
presentation statement
Figure 661878DEST_PATH_IMAGE002
The maximum degree of correlation of the boundary of (c),
Figure 142537DEST_PATH_IMAGE017
a feature representation (i.e., the second mean vector) representing text,
Figure 434979DEST_PATH_IMAGE018
as a sentence
Figure 292076DEST_PATH_IMAGE002
The feature representation of (i.e., the first mean vector),
Figure 658467DEST_PATH_IMAGE019
as a sentence
Figure 310028DEST_PATH_IMAGE020
The feature representation of (i.e., the third mean vector),
Figure 355344DEST_PATH_IMAGE021
is a preset hyper-parameter and is continuously optimized in the subsequent training process,
Figure 750553DEST_PATH_IMAGE022
presentation statement
Figure 299346DEST_PATH_IMAGE002
A first degree of similarity to the text,
Figure 387388DEST_PATH_IMAGE024
presentation statement
Figure 857684DEST_PATH_IMAGE002
And sentences
Figure 791005DEST_PATH_IMAGE020
Of the second similarity degree.
Figure 459883DEST_PATH_IMAGE022
Indicating the relevance of a single sentence to the entire text,
Figure 718826DEST_PATH_IMAGE024
representing sentences in text
Figure 207577DEST_PATH_IMAGE002
And sentences
Figure 646386DEST_PATH_IMAGE020
The correlation between them.
Figure 169771DEST_PATH_IMAGE024
The larger, the sentence
Figure 334036DEST_PATH_IMAGE002
And sentences
Figure 575662DEST_PATH_IMAGE020
The stronger the correlation between them, the statement is indicated
Figure 116364DEST_PATH_IMAGE002
And sentences
Figure 494256DEST_PATH_IMAGE020
The larger the redundancy is, the more redundant the generated text abstract is;
Figure 32685DEST_PATH_IMAGE024
the smaller, the sentence
Figure 496027DEST_PATH_IMAGE002
And sentences
Figure 840421DEST_PATH_IMAGE020
The weaker the correlation between them, the statement
Figure 338398DEST_PATH_IMAGE002
And sentences
Figure 844466DEST_PATH_IMAGE020
The less redundant the text excerpt generated, the less redundant.
In an optional embodiment, the calculating the sentence weight of the sentence in the text according to the sentence vector comprises:
calculating a sum of the second maximum boundary correlations;
calculating a ratio of the first maximum boundary correlation to the sum;
and mapping the ratio by using a preset function to obtain the statement weight of the statement.
The computer device stores a second formula in advance, and the statement weight of the statement calculated according to the first maximum boundary correlation and the second maximum boundary correlation can be represented by the following second formula:
Figure 60684DEST_PATH_IMAGE025
wherein the content of the first and second substances,
Figure 880872DEST_PATH_IMAGE016
presentation statement
Figure 233356DEST_PATH_IMAGE002
The maximum degree of correlation of the boundary of (c),
Figure 910325DEST_PATH_IMAGE026
presentation statement
Figure 348260DEST_PATH_IMAGE020
The maximum degree of correlation of the boundary of (c),
Figure 34456DEST_PATH_IMAGE027
presentation statement
Figure 241446DEST_PATH_IMAGE002
The sentence weight of (1).
After calculating the maximum boundary correlation degree of each sentence in the text, the computer device calculates the sum of the maximum boundary correlation degrees of other sentences except the any sentence in the text for any sentence, calculates the ratio of the maximum boundary correlation degree of the any sentence to the sum, and finally performs softmax function processing to obtain the sentence weight (or called probability value) of the any sentence. The larger the ratio, the greater the relevance of the corresponding sentence to the text, and the higher the redundancy. The smaller the ratio, the weaker the correlation of the corresponding sentence with the text, and the lower the redundancy.
The vector encoding module 404 is configured to encode the segment vector to obtain a hidden state of the text segment.
The segment vectors may be encoded using an agent in an encoder module that generates a model, resulting in a hidden state of the text segment.
In this embodiment, the network architecture of the generative model is as shown in fig. 2 below, and the generative model includes an Encoder module (e.g., an Encoder part in fig. 2) located at a bottom layer and a Decoder module (e.g., a Decoder part in fig. 2) located at an upper layer, where the Encoder module includes a plurality of agents (e.g., Agent1, Agent2, Agent m in fig. 2) independent of each other, each Agent includes a bidirectional LSTM (bilstm) model, and the Decoder module includes a multi-layer Attention mechanism layer (e.g., an extraction model Attention in fig. 2) and an Agent Attention mechanism layer (e.g., an Agent Attention in fig. 2) and an LSTM model.
The computer device configures the number of agents in the encoder module according to the number of text segments, that is, how many text segments exist, and how many agents are correspondingly configured in the encoder module, so that the text segments correspond to the agents one to one. The computer equipment inputs each text segment into one agent in the encoder module, encodes the text segment through the agent to obtain an intermediate vector (namely a hidden state), and transmits the intermediate vector to other agents, so that different agents can share global context information of different contents in the text. Each agent outputs an intermediate vector, and all intermediate vectors form an intermediate vector sequence. And finally, decoding the intermediate vectors output by all the agents through a decoder module, namely decoding the intermediate vector sequence to obtain the probability distribution of the text abstract, thereby obtaining the final text abstract. Referring to FIG. 2, a computer device segments text
Figure 27000DEST_PATH_IMAGE030
Inputting Agent1 in encoder module, and text segment by Agent1
Figure 952231DEST_PATH_IMAGE030
Coding to obtain text segments
Figure 176538DEST_PATH_IMAGE030
Hidden state of (2), segmenting text
Figure 503615DEST_PATH_IMAGE031
Inputting Agent2 in encoder module, and text segment by Agent2
Figure 787965DEST_PATH_IMAGE031
Coding to obtain text segments
Figure 934913DEST_PATH_IMAGE031
Hidden state of (2), segmenting text
Figure 655920DEST_PATH_IMAGE007
Inputting proxy AgentM in encoder module, and aligning text segment by proxy AgentM
Figure 837503DEST_PATH_IMAGE007
Coding to obtain text segments
Figure 292755DEST_PATH_IMAGE007
Hidden state of (1).
In an alternative embodiment, as shown in fig. 3, the agent includes two-stage bidirectional LSTM models (a first-stage bidirectional LSTM model and a second-stage bidirectional LSTM model), and the vector encoding module 404 encodes the segment vector to obtain the hidden state of the text segment includes:
coding the segment vector through a first-stage bidirectional LSTM model to obtain a hidden state of a target keyword in the text segment;
and coding the hidden state of the target keyword through a second-stage bidirectional LSTM model to obtain the hidden state of the text segment.
The BilSTM in each agent is a two-stage bi-directional LSTM model, where the first stage is a single-layer bi-directional LSTM (e.g., Local Encoder in FIG. 3), the output of each Local Encoder is input to a second stage context coder (e.g., context Encoder in FIG. 3), which is a multi-layer bi-directional LSTM structure called context Encoder.
The procedure of the first stage is as follows:
passing word vector
Figure 192578DEST_PATH_IMAGE013
Combined text segments
Figure 758688DEST_PATH_IMAGE010
And is fed as a whole into the first stage single layer bi-directional LSTM for encoding. The coding formula of the first-stage single-layer bidirectional LSTM is as follows:
hidden state in the middle:
Figure 732461DEST_PATH_IMAGE032
hidden state of output:
Figure 358614DEST_PATH_IMAGE033
wherein the content of the first and second substances,
Figure 480154DEST_PATH_IMAGE034
as target key words
Figure 849955DEST_PATH_IMAGE011
Is a hidden state, H is a hidden state
Figure 6130DEST_PATH_IMAGE034
The dimension (c) of (a) is,
Figure 475289DEST_PATH_IMAGE035
continuously optimizing the parameters in the training process;
Figure 84125DEST_PATH_IMAGE036
representing target keywords
Figure 992038DEST_PATH_IMAGE011
The intermediate hidden state resulting from the encoding of the bi-directional LSTM,
Figure 2719DEST_PATH_IMAGE037
and the latter target keyword
Figure 970675DEST_PATH_IMAGE076
Is added to the intermediate hidden state of (1).
As can be seen from the above formula, each target keyword in the text segment
Figure 270069DEST_PATH_IMAGE011
The corresponding hidden states are from the middle hidden state of the previous and next target keywords and the word vector representation of the target keywords themselves.
The process of the second stage is as follows:
Figure 981673DEST_PATH_IMAGE039
Figure 112441DEST_PATH_IMAGE040
each agent is at
Figure 251298DEST_PATH_IMAGE041
Layer(s)
Figure 569147DEST_PATH_IMAGE042
In the context encoder, information received from an upper layer is jointly encoded. As can be seen from the above formulas
Figure 255081DEST_PATH_IMAGE041
The input to the layer context encoder comprises three parts: hidden state from contextual target keywords: (
Figure 240354DEST_PATH_IMAGE043
Figure 550113DEST_PATH_IMAGE044
) (ii) a Hidden state of the previous layer output (
Figure 355258DEST_PATH_IMAGE045
) (ii) a Information from other agents: (
Figure 408664DEST_PATH_IMAGE046
Output of
Figure 451707DEST_PATH_IMAGE047
Wherein, in the step (A),
Figure 932367DEST_PATH_IMAGE048
is the other agent (1)
Figure 490387DEST_PATH_IMAGE049
) Average value of layer output
Figure 81905DEST_PATH_IMAGE050
Figure 41771DEST_PATH_IMAGE051
For other agents in (1)
Figure 693332DEST_PATH_IMAGE049
) The layer generates a hidden state vector for the last target keyword. In the text segment
Figure 410753DEST_PATH_IMAGE010
In each target keyword
Figure 805962DEST_PATH_IMAGE011
In addition to learning the own contextual target keyword information, overall information from other text segments is also accepted.
The encoder module output
Figure DEST_PATH_IMAGE077
Representing a text segment
Figure 620334DEST_PATH_IMAGE010
Is/are as follows
Figure 442797DEST_PATH_IMAGE012
Hidden state vector order of individual target keywordsAnd (4) columns.
The attention calculating module 405 is configured to obtain the attention weight of the target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each time.
In an alternative embodiment, the attention extracting mechanism layer in the decoder module of the generative model may be used to calculate the attention weight of the target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each time.
The probability distribution module 406 calculates the vocabulary probability distribution at the time t according to the attention weight of the target keyword and the corresponding agent weight.
The vocabulary probability distribution at the time t can be calculated by using an agent attention mechanism layer in the decoder module according to the attention weight of the target keyword and the corresponding agent weight.
The vocabulary probability distribution represents target keywords appearing at the kth position of the text abstract, and the agent weight is the sum of statement weights of statements in the text segment.
The decoder module is of a one-way LSTM structure so that the last target keyword of each text segment can contain the state information of all previous target keywords. Initial input state of decoder module
Figure 709830DEST_PATH_IMAGE053
Output hidden state vector for last layer of last target key of all agents of encoder module
Figure 908730DEST_PATH_IMAGE054
The hidden state vector contains all the text information encoded by the encoder module.
In specific implementation, a hidden state vector sequence generated by the last hidden layer of the agent is utilized
Figure 249713DEST_PATH_IMAGE077
The text segment is calculated by the following formulaAttention weight of each target keyword in (1)
Figure 243076DEST_PATH_IMAGE055
Figure 997406DEST_PATH_IMAGE056
. Wherein the content of the first and second substances,
Figure 999997DEST_PATH_IMAGE057
Figure 523382DEST_PATH_IMAGE058
all of which are hyper-parameters,
Figure 687647DEST_PATH_IMAGE059
representing the state vector at time t. For each agent a in the agent attention mechanism layer, the attention weight of the agent is represented by the weight of each target keyword in the text segment:
Figure 634000DEST_PATH_IMAGE060
segmenting text
Figure 174703DEST_PATH_IMAGE010
The sentence weight (i.e., the sentence prediction probability) of the middle sentence is added as the proxy weight
Figure 552594DEST_PATH_IMAGE061
All agents will be
Figure 153340DEST_PATH_IMAGE062
Weighted summation of the attention distributions to obtain the attention weight of the t-th word of the abstract
Figure 616682DEST_PATH_IMAGE063
. Will be provided with
Figure 961076DEST_PATH_IMAGE064
Putting the hidden state vector of each moment into a multilayer perceptron to obtain the vocabulary probability distribution of the t moment
Figure 193474DEST_PATH_IMAGE065
Figure 637225DEST_PATH_IMAGE066
Representing the target keyword appearing at the k-th position of the text excerpt.
Each time step t of the LSTM represents the location of the t-th target keyword of the generated text excerpt. Generating a hidden state at the current time
Figure 587864DEST_PATH_IMAGE059
Predicted new target keyword at the t-th position in the abstract
Figure 470369DEST_PATH_IMAGE066
Probability distribution of (2). In this process, a layered attention mechanism is introduced inside the LSTM structure, including the attention weight word attention of the target keyword and the attention weight agent attention of the agent.
The probability distribution of the obtained text abstract is
Figure 822853DEST_PATH_IMAGE067
Wherein, in the step (A),
Figure 765401DEST_PATH_IMAGE068
representing the probability of each word in the dictionary appearing at the moment t, wherein V is the number of target keywords in the course corpus thesaurus, and each element
Figure 203336DEST_PATH_IMAGE069
In this alternative embodiment, the weight calculated by the generative model not only takes into account the similarity between the sentence and the text, but also takes into account the redundancy of the abstract sentence, and this weight is combined with the weight of the text segment of the decoder module of the generative model, and compared with a common decoder, the probability finally calculated also includes the information of redundancy. Namely, the weight of the text segment composed of sentences with large meaning repeatability is relatively low, the importance of the information in the process of the decoder module is also low, the probability of repeated meaning target keywords in the finally obtained abstract is also lower, the generated text abstract is more accurate, and the meaning is clear and concise.
The abstract generating module 407 is configured to generate a text abstract according to the vocabulary probability distribution.
Since the probability distribution represents the probability of each target keyword in the text appearing in the text abstract, the higher the probability that the corresponding target keyword appears in the text abstract, and the lower the probability, the lower the probability that the corresponding target keyword appears in the text abstract. The computer device may determine a target keyword appearing in the text excerpt according to the probability in the probability distribution, thereby generating the text excerpt according to the target keyword.
In an alternative embodiment, the digest generation module 407 generating the text digest according to the vocabulary probability distribution includes:
for any moment, acquiring the maximum probability of the probability distribution at the any moment;
determining the target keywords corresponding to the maximum probability as the target keywords in the text abstract;
and combining the target keywords in the text abstract according to a time sequence to obtain the text abstract.
Illustratively, for a certain time
Figure 889532DEST_PATH_IMAGE070
Selecting a time
Figure 34205DEST_PATH_IMAGE070
The target keyword corresponding to the maximum probability of being the first of the text abstract to be generated
Figure 147655DEST_PATH_IMAGE070
Generating text abstract finally according to the target key words
Figure 72886DEST_PATH_IMAGE071
,
Figure 297194DEST_PATH_IMAGE072
Is a time of day
Figure 624270DEST_PATH_IMAGE073
The maximum probability of the lower case corresponds to the target keyword,
Figure 643041DEST_PATH_IMAGE074
as a summary of the text
Figure 55568DEST_PATH_IMAGE075
The number of target keywords contained therein.
After the computer equipment generates the text abstract, the generated text abstract can be displayed for a learner to review and review after a training class according to the generated text abstract, and a key course is selected for learning according to target keywords (serving as keywords) in the text abstract.
In summary, the text abstract generating device provided by the present invention, for a text requiring text abstract generation, performs word segmentation on the text to obtain target keywords, extracts word vectors of each target keyword in the text, generates a sentence vector for each sentence in the text and a segment vector for each text segment in the text according to the word vectors, thereby obtaining a sentence weight of each sentence in the whole text according to the sentence vectors, then encodes the segment vectors by using a trained LSTM model to obtain a hidden state of the text segment, obtains an attention weight of the target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each time, obtains a vocabulary probability distribution at time t according to the attention weight of the target keyword and a corresponding proxy weight, and the vocabulary probability distribution represents the target keyword appearing at the kth position, and the proxy weight is the sum of the sentence weights, and finally, the text abstract is generated according to the vocabulary probability distribution. The generated text abstract not only considers the similarity between sentences and texts, but also considers the redundancy of abstract sentences, the weight of text segments formed by sentences with large meaning repeatability is relatively low, the importance of information in the process of a decoder module is also low, the possibility of repeated meaning target keywords in the finally obtained abstract is also lower, and the generated text abstract is more accurate, clear and concise in meaning.
It is emphasized that the text may be stored in a node of the blockchain in order to further ensure privacy and security of the text.
Fig. 5 is a schematic structural diagram of a computer device according to a third embodiment of the present invention. In the preferred embodiment of the present invention, the computer device 5 comprises a memory 51, at least one processor 51, at least one communication bus 53 and a transceiver 54.
It will be appreciated by those skilled in the art that the configuration of the computer device shown in fig. 5 is not limiting to the embodiments of the present invention, and may be a bus-type configuration or a star-type configuration, and that the computer device 5 may include more or less hardware or software than those shown, or a different arrangement of components.
In some embodiments, the computer device 5 is a device capable of automatically performing numerical calculation and/or information processing according to instructions set or stored in advance, and the hardware includes but is not limited to a microprocessor, an application specific integrated circuit, a programmable gate array, a digital processor, an embedded device, and the like. The computer device 5 may also include a client device, which includes, but is not limited to, any electronic product capable of interacting with a client through a keyboard, a mouse, a remote controller, a touch pad, or a voice control device, for example, a personal computer, a tablet computer, a smart phone, a digital camera, etc.
It should be noted that the computer device 5 is only an example, and other electronic products that are currently available or may come into existence in the future, such as electronic products that can be adapted to the present invention, should also be included in the scope of the present invention, and are incorporated herein by reference.
In some embodiments, the memory 51 has stored therein a computer program which, when executed by the at least one processor 51, implements all or part of the steps of the text summary generation method as described. The Memory 51 includes a Read-Only Memory (ROM), a Programmable Read-Only Memory (PROM), an Erasable Programmable Read-Only Memory (EPROM), a One-time Programmable Read-Only Memory (OTPROM), an electronically Erasable rewritable Read-Only Memory (Electrically-Erasable Programmable Read-Only Memory (EEPROM)), an optical Read-Only disk (CD-ROM) or other optical disk Memory, a magnetic disk Memory, a tape Memory, or any other medium readable by a computer capable of carrying or storing data.
Further, the computer-readable storage medium may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function, and the like; the storage data area may store data created according to the use of the blockchain node, and the like.
The block chain is a novel application mode of computer technologies such as distributed data storage, point-to-point transmission, a consensus mechanism, an encryption algorithm and the like. A block chain (Blockchain), which is essentially a decentralized database, is a series of data blocks associated by using a cryptographic method, and each data block contains information of a batch of network transactions, so as to verify the validity (anti-counterfeiting) of the information and generate a next block. The blockchain may include a blockchain underlying platform, a platform product service layer, an application service layer, and the like.
In some embodiments, the at least one processor 51 is a Control Unit (Control Unit) of the computer device 5, and is connected to various components of the whole computer device 5 by various interfaces and lines, and executes various functions and processes data of the computer device 5 by running or executing programs or modules stored in the memory 51 and calling data stored in the memory 51. For example, when the at least one processor 51 executes the computer program stored in the memory, all or part of the steps of the text summary generation method according to the embodiment of the present invention are implemented; or to implement all or part of the functions of the text digest generation apparatus. The at least one processor 51 may be composed of an integrated circuit, for example, a single packaged integrated circuit, or may be composed of a plurality of integrated circuits packaged with the same function or different functions, and includes one or more Central Processing Units (CPUs), microprocessors, digital Processing chips, graphics processors, and combinations of various control chips.
In some embodiments, the at least one communication bus 53 is arranged to enable connection communication between the memory 51 and the at least one processor 51, etc.
Although not shown, the computer device 5 may further include a power supply (such as a battery) for supplying power to each component, and preferably, the power supply may be logically connected to the at least one processor 51 through a power management device, so as to implement functions of managing charging, discharging, and power consumption through the power management device. The power supply may also include any component of one or more dc or ac power sources, recharging devices, power failure detection circuitry, power converters or inverters, power status indicators, and the like. The computer device 5 may further include various sensors, a bluetooth module, a Wi-Fi module, and the like, which are not described herein again.
The integrated unit implemented in the form of a software functional module may be stored in a computer-readable storage medium. The software functional module is stored in a storage medium and includes several instructions to enable a computer device (which may be a personal computer, a computer device, or a network device) or a processor (processor) to execute parts of the methods according to the embodiments of the present invention.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, the division of the modules is only one logical functional division, and other divisions may be realized in practice.
The modules described as separate parts may or may not be physically separate, and parts displayed as modules may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment.
In addition, functional modules in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, or in a form of hardware plus a software functional module.
It will be evident to those skilled in the art that the invention is not limited to the details of the foregoing illustrative embodiments, and that the present invention may be embodied in other specific forms without departing from the spirit or essential attributes thereof. The present embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein. Any reference sign in a claim should not be construed as limiting the claim concerned. Furthermore, it is obvious that the word "comprising" does not exclude other elements or that the singular does not exclude the plural. A plurality of units or means recited in the present invention can also be implemented by one unit or means through software or hardware. The terms first, second, etc. are used to denote names, but not any particular order.
Finally, it should be noted that the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting, and although the present invention is described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made on the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention.

Claims (10)

1. A text summary generation method, the method comprising:
performing word segmentation processing on a text to obtain a target keyword, and acquiring a word vector of the target keyword;
generating statement vectors of statements in the text according to the word vectors, and generating segment vectors of text segments in the text according to the word vectors;
calculating statement weight of the statement in the text according to the statement vector;
coding the segment vector to obtain the hidden state of the text segment;
obtaining attention weight of a target keyword in the text segment according to the hidden state of the text segment and the hidden state vector at each moment;
calculating according to the attention weight of the target keyword and a corresponding proxy weight to obtain vocabulary probability distribution at the time t, wherein the vocabulary probability distribution represents the target keyword appearing at the kth position of the text abstract, and the proxy weight is the sum of statement weights of statements in the text segment;
and generating a text abstract according to the vocabulary probability distribution.
2. The method of generating a text summary according to claim 1, wherein the calculating sentence weights of the sentences in the text according to the sentence vectors comprises:
acquiring a first maximum boundary correlation degree of the statement;
acquiring a second maximum boundary correlation degree of other sentences except the sentence in the text;
and obtaining the sentence weight of the sentence according to the first maximum boundary correlation degree and the second maximum boundary correlation degree.
3. The text summary generation method of claim 2, wherein the calculation of the first maximum boundary relevance includes:
calculating a first similarity between the sentence and the text according to the sentence vector of the sentence;
calculating a second similarity between the statement and the rest statements according to the statement vectors of the statement and the statement vectors of the rest statements;
and calculating to obtain the first maximum boundary correlation degree according to the first similarity degree and the second similarity degree.
4. The text summary generation method of claim 3, wherein calculating a first similarity of the sentence to the text from the sentence vector of the sentence comprises: calculating a first feature representation of the sentence according to the word vector in the sentence; calculating a second feature representation of the text according to the word vectors in the text; calculating to obtain a first similarity according to the first feature representation and the second feature representation by adopting a similarity calculation model;
the calculating the second similarity between the sentence and the rest of the sentences according to the sentence vectors of the sentences and the sentence vectors of the rest of the sentences comprises: calculating third feature representations of the rest sentences according to the word vectors in the rest sentences; and calculating to obtain a second similarity according to the first characteristic representation and the third characteristic representation by adopting the similarity calculation model.
5. The method of generating a text summary according to claim 4, wherein the calculating the first maximum boundary relevance according to the first similarity and the second similarity comprises:
obtaining a first value according to a preset hyper-parameter and the first similarity;
determining a maximum value of the second similarity;
obtaining a second value according to the preset hyper-parameter and the maximum value;
and obtaining a first maximum boundary correlation degree according to the first value and the second value.
6. The method of generating a text summary according to claim 4, wherein the calculating sentence weights of the sentences in the text according to the sentence vectors comprises:
calculating a sum of the second maximum boundary correlations;
calculating a ratio of the first maximum boundary correlation to the sum;
and mapping the ratio by using a preset function to obtain the statement weight of the statement.
7. The method according to any one of claims 1 to 6, wherein said encoding the segment vector to obtain the hidden state of the text segment comprises:
coding the segment vector through a first-stage bidirectional LSTM model to obtain a hidden state of a target keyword in the text segment;
and coding the hidden state of the target keyword through a second-stage bidirectional LSTM model to obtain the hidden state of the text segment.
8. The method of generating a text excerpt as claimed in claim 7, wherein said generating a text excerpt from said lexical probability distribution comprises:
for any moment, acquiring the maximum probability of the probability distribution at the any moment;
determining the target keywords corresponding to the maximum probability as the target keywords in the text abstract;
and combining the target keywords in the text abstract according to a time sequence to obtain the text abstract.
9. A computer device, characterized in that the computer device comprises a processor for implementing the text summary generation method according to any one of claims 1 to 8 when executing a computer program stored in a memory.
10. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, implements a text summary generation method according to any one of claims 1 to 8.
CN202110489771.1A 2021-05-06 2021-05-06 Text abstract generation method, computer equipment and storage medium Active CN112906385B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110489771.1A CN112906385B (en) 2021-05-06 2021-05-06 Text abstract generation method, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110489771.1A CN112906385B (en) 2021-05-06 2021-05-06 Text abstract generation method, computer equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112906385A true CN112906385A (en) 2021-06-04
CN112906385B CN112906385B (en) 2021-08-13

Family

ID=76108937

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110489771.1A Active CN112906385B (en) 2021-05-06 2021-05-06 Text abstract generation method, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112906385B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113204944A (en) * 2021-06-02 2021-08-03 云知声智能科技股份有限公司 Text generation method, device, equipment and storage medium
CN113420545A (en) * 2021-08-24 2021-09-21 平安科技(深圳)有限公司 Abstract generation method, device, equipment and storage medium
CN113449105A (en) * 2021-06-25 2021-09-28 上海明略人工智能(集团)有限公司 Work summary generation method, system, electronic device and medium
CN113704667A (en) * 2021-08-31 2021-11-26 北京百炼智能科技有限公司 Automatic extraction processing method and device for bidding announcement
CN113836938A (en) * 2021-09-23 2021-12-24 深圳市北科瑞声科技股份有限公司 Text similarity calculation method and device, storage medium and electronic device
CN115618857A (en) * 2022-09-09 2023-01-17 中国电信股份有限公司 Threat information processing method, threat information pushing method and device

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104375977A (en) * 2013-08-14 2015-02-25 腾讯科技(深圳)有限公司 Answer message processing method and device for question-answer communities
US20180300400A1 (en) * 2017-04-14 2018-10-18 Salesforce.Com, Inc. Deep Reinforced Model for Abstractive Summarization
CN110619043A (en) * 2019-08-30 2019-12-27 西南电子技术研究所(中国电子科技集团公司第十研究所) Automatic text abstract generation method based on dynamic word vector
CN111666756A (en) * 2020-05-26 2020-09-15 湖北工业大学 Sequence model text abstract generation method based on topic fusion
CN112069309A (en) * 2020-09-14 2020-12-11 腾讯科技(深圳)有限公司 Information acquisition method and device, computer equipment and storage medium
CN112417854A (en) * 2020-12-15 2021-02-26 北京信息科技大学 Chinese document abstraction type abstract method
US20210089622A1 (en) * 2019-09-23 2021-03-25 Hong Kong Applied Science and Technology Research Institute Company Limited Method of Summarizing Text with Sentence Extraction
CN112560456A (en) * 2020-11-03 2021-03-26 重庆安石泽太科技有限公司 Generation type abstract generation method and system based on improved neural network
CN112732899A (en) * 2020-12-31 2021-04-30 平安科技(深圳)有限公司 Abstract statement extraction method, device, server and computer readable storage medium

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104375977A (en) * 2013-08-14 2015-02-25 腾讯科技(深圳)有限公司 Answer message processing method and device for question-answer communities
US20180300400A1 (en) * 2017-04-14 2018-10-18 Salesforce.Com, Inc. Deep Reinforced Model for Abstractive Summarization
US20190311002A1 (en) * 2017-04-14 2019-10-10 Salesforce.Com, Inc. Deep reinforced model for abstractive summarization
CN110619043A (en) * 2019-08-30 2019-12-27 西南电子技术研究所(中国电子科技集团公司第十研究所) Automatic text abstract generation method based on dynamic word vector
US20210089622A1 (en) * 2019-09-23 2021-03-25 Hong Kong Applied Science and Technology Research Institute Company Limited Method of Summarizing Text with Sentence Extraction
CN111666756A (en) * 2020-05-26 2020-09-15 湖北工业大学 Sequence model text abstract generation method based on topic fusion
CN112069309A (en) * 2020-09-14 2020-12-11 腾讯科技(深圳)有限公司 Information acquisition method and device, computer equipment and storage medium
CN112560456A (en) * 2020-11-03 2021-03-26 重庆安石泽太科技有限公司 Generation type abstract generation method and system based on improved neural network
CN112417854A (en) * 2020-12-15 2021-02-26 北京信息科技大学 Chinese document abstraction type abstract method
CN112732899A (en) * 2020-12-31 2021-04-30 平安科技(深圳)有限公司 Abstract statement extraction method, device, server and computer readable storage medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
YAO MA;MA YAO;ZHUANG YUFENG;HU YANZHU;TIAN WENJIA;WANG SONG: ""Research on Logistics Text Information Abstract Generation Method"", 《JOURNAL OF PHYSICS: CONFERENCE SERIES》 *
YUWEN LI;YUZHUO FU;TING LIU: ""LTKW: An Improvement of Extractive Automatic Text Abstract model on Long Chinese Text"", 《2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC)》 *
滕少华,董谱,张巍: ""融合语义结构的注意力文本摘要模型"", 《广东工业大学学报》 *
王红斌,金子铃,毛存礼: ""结合层级注意力的抽取式新闻文本自动摘要"", 《计算机科学与探索》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113204944A (en) * 2021-06-02 2021-08-03 云知声智能科技股份有限公司 Text generation method, device, equipment and storage medium
CN113449105A (en) * 2021-06-25 2021-09-28 上海明略人工智能(集团)有限公司 Work summary generation method, system, electronic device and medium
CN113420545A (en) * 2021-08-24 2021-09-21 平安科技(深圳)有限公司 Abstract generation method, device, equipment and storage medium
CN113420545B (en) * 2021-08-24 2021-11-09 平安科技(深圳)有限公司 Abstract generation method, device, equipment and storage medium
CN113704667A (en) * 2021-08-31 2021-11-26 北京百炼智能科技有限公司 Automatic extraction processing method and device for bidding announcement
CN113704667B (en) * 2021-08-31 2023-06-27 北京百炼智能科技有限公司 Automatic extraction processing method and device for bid announcement
CN113836938A (en) * 2021-09-23 2021-12-24 深圳市北科瑞声科技股份有限公司 Text similarity calculation method and device, storage medium and electronic device
CN115618857A (en) * 2022-09-09 2023-01-17 中国电信股份有限公司 Threat information processing method, threat information pushing method and device
CN115618857B (en) * 2022-09-09 2024-03-01 中国电信股份有限公司 Threat information processing method, threat information pushing method and threat information pushing device

Also Published As

Publication number Publication date
CN112906385B (en) 2021-08-13

Similar Documents

Publication Publication Date Title
CN112906385B (en) Text abstract generation method, computer equipment and storage medium
WO2022078346A1 (en) Text intent recognition method and apparatus, electronic device, and storage medium
US20210232773A1 (en) Unified Vision and Dialogue Transformer with BERT
CN109670029A (en) For determining the method, apparatus, computer equipment and storage medium of problem answers
CN113052149B (en) Video abstract generation method and device, computer equipment and medium
CN112231485B (en) Text recommendation method and device, computer equipment and storage medium
CN112000805A (en) Text matching method, device, terminal and storage medium based on pre-training model
CN112417128B (en) Method and device for recommending dialect, computer equipment and storage medium
CN114007131A (en) Video monitoring method and device and related equipment
CN114021582B (en) Spoken language understanding method, device, equipment and storage medium combined with voice information
CN113435582B (en) Text processing method and related equipment based on sentence vector pre-training model
CN113096242A (en) Virtual anchor generation method and device, electronic equipment and storage medium
CN113722483A (en) Topic classification method, device, equipment and storage medium
CN113642316A (en) Chinese text error correction method and device, electronic equipment and storage medium
CN111611805A (en) Auxiliary writing method, device, medium and equipment based on image
CN113704410A (en) Emotion fluctuation detection method and device, electronic equipment and storage medium
CN112257860A (en) Model generation based on model compression
CN114077841A (en) Semantic extraction method and device based on artificial intelligence, electronic equipment and medium
CN113807973A (en) Text error correction method and device, electronic equipment and computer readable storage medium
CN113486659B (en) Text matching method, device, computer equipment and storage medium
CN113254814A (en) Network course video labeling method and device, electronic equipment and medium
CN114398902A (en) Chinese semantic extraction method based on artificial intelligence and related equipment
CN114417891B (en) Reply statement determination method and device based on rough semantics and electronic equipment
CN115510188A (en) Text keyword association method, device, equipment and storage medium
CN115658858A (en) Dialog recommendation method based on artificial intelligence and related equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant