CN113204956A - Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device - Google Patents

Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device Download PDF

Info

Publication number
CN113204956A
CN113204956A CN202110762240.5A CN202110762240A CN113204956A CN 113204956 A CN113204956 A CN 113204956A CN 202110762240 A CN202110762240 A CN 202110762240A CN 113204956 A CN113204956 A CN 113204956A
Authority
CN
China
Prior art keywords
training
abstract
text
model
sentences
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110762240.5A
Other languages
Chinese (zh)
Other versions
CN113204956B (en
Inventor
蒋志燕
吕少领
黄石磊
程刚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Raisound Technology Co ltd
Original Assignee
Shenzhen Raisound Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Raisound Technology Co ltd filed Critical Shenzhen Raisound Technology Co ltd
Priority to CN202110762240.5A priority Critical patent/CN113204956B/en
Publication of CN113204956A publication Critical patent/CN113204956A/en
Application granted granted Critical
Publication of CN113204956B publication Critical patent/CN113204956B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/34Browsing; Visualisation therefor
    • G06F16/345Summarisation for human users
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/126Character encoding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Machine Translation (AREA)

Abstract

The application relates to the technical field of artificial intelligence, and discloses a multi-model training method, which comprises the following steps: dividing the text in the training text set into single sentences to obtain a training single sentence set; extracting the characteristics of the training single sentence set to obtain a training single sentence vector set; extracting paragraph coding features and abstract coding features of training single sentences in the training single sentence vector set; and performing first training on the pre-constructed text segmentation model by using the training single sentence vector set and the abstract coding features, and performing second training on the pre-constructed text abstract extraction model by using the training single sentence vector set and the paragraph coding features to obtain a standard text segmentation model and a standard text abstract extraction model. In addition, the application also relates to a method for abstracting the abstract, a method for segmenting the text, a device, equipment and a storage medium. The method and the device can improve the model accuracy of the text segmentation model and the abstract extraction model obtained by training and the obtaining efficiency of the text segmentation model and the abstract extraction model obtained by training.

Description

Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device
Technical Field
The present application relates to the field of artificial intelligence technologies, and in particular, to a multi-model training method, an abstract segmentation method, a text segmentation method, an apparatus, an electronic device, and a storage medium.
Background
Text segmentation and abstract extraction are common processing modes for text processing, and in the prior art, a text abstract extraction model and a text segmentation model are usually obtained by training respectively. Specifically, the existing technology adopted for the text segmentation task is to use a Jaccard similarity analyzer to find the distance between consecutive sentences, and if the distance between them is less than a given value, the consecutive sentences are grouped into a paragraph. However, the segmentation method only performs segmentation from a single dimension of similarity between sentences, which easily results in inaccurate segmentation results.
The text abstract extracting model used in the existing text abstract extracting method usually only considers the position of a sentence and extracts the first few sentences of a text to form an abstract, the method is relatively suitable for news texts, the application range of the text abstract extracting model is not wide, and the extracting result is often inaccurate when paragraphs except specific types of texts (such as news texts) are extracted. In summary, the obtaining efficiency of the text segmentation model and the text abstract extraction model in the prior art is not high, and the model accuracy is not high.
Disclosure of Invention
In order to solve the technical problems or at least partially solve the technical problems, the application provides a multi-model training method, a summary segmentation method, a text segmentation method, a device, an electronic device and a storage medium.
In a first aspect, the present application provides a multi-model training method, including:
acquiring a training text set, and dividing texts in the training text set into single sentences to obtain a training single sentence set;
performing feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set;
extracting paragraph coding features and abstract coding features of the training single sentences in the training single sentence vector set;
and carrying out first training on a pre-constructed text segmentation model by using the training single sentence vector set and the abstract coding features of the training single sentences in the training single sentence vector set, and carrying out second training on a pre-constructed text abstract extraction model by using the training single sentence vector set and the paragraph coding features of the training single sentences in the training single sentence vector set to obtain a standard text segmentation model and a standard text abstract extraction model.
Optionally, the performing feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set includes:
extracting the mark embedding characteristics, distinguishing embedding characteristics and position embedding characteristics of the training single sentences in the training single sentence set through a first extraction network of a preset characteristic extraction model;
combining the mark embedding characteristics, the distinguishing embedding characteristics and the position embedding characteristics of the training single sentence and inputting the combination into a second extraction network of the characteristic extraction model to obtain a single sentence vector of the training single sentence;
and summarizing all the obtained single sentence vectors to obtain the training single sentence vector set.
Optionally, the feature extraction model is a BERT model.
Optionally, the performing, by using the training single sentence vector set and the paragraph coding features of the training single sentences in the training single sentence vector set, a second training on the pre-constructed text abstract extraction model includes:
inputting the training single sentence vector set and paragraph coding features of the training single sentences in the training single sentence vector set to a pre-constructed document feature extraction model to obtain an abstract training single sentence vector set;
and inputting the abstract training single sentence vector set into the text abstract extraction model for second training.
Optionally, after obtaining the training single sentence set, the method further includes:
and deleting stop words, tone words and repeated words in the training single sentence set.
In a second aspect, the present application provides a method for abstracting a summary, the method including:
acquiring a text to be processed;
and inputting the text to be processed into a standard text abstract extraction model for abstract extraction to obtain an abstract extraction result, wherein the standard text abstract extraction model is obtained by training by adopting the multi-model training method in the first aspect.
In a third aspect, the present application provides a text segmentation method, including:
acquiring a text to be processed;
and inputting the text to be processed into a standard text segmentation model for text segmentation to obtain a text segmentation result, wherein the standard text segmentation model is obtained by training by adopting the multi-model training method in the first aspect.
In a fourth aspect, the present application provides a multi-model training apparatus, the apparatus comprising:
the training text acquisition module is used for acquiring a training text set, dividing texts in the training text set into single sentences to obtain a training single sentence set;
the first feature extraction module is used for extracting features of the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set;
the second characteristic extraction module is used for extracting paragraph coding characteristics and abstract coding characteristics of the training single sentences in the training single sentence vector set;
and the training module is used for carrying out first training on the pre-constructed text segmentation model by utilizing the training single sentence vector set and the abstract coding characteristics of the training single sentences in the training single sentence vector set, and carrying out second training on the pre-constructed text abstract extraction model by utilizing the training single sentence vector set and the paragraph coding characteristics of the training single sentences in the training single sentence vector set to obtain a standard text segmentation model and a standard text abstract extraction model.
In a fifth aspect, the present application provides a digest extraction apparatus, including:
the abstract text acquisition module is used for acquiring a text to be processed;
and the abstract extraction module is used for inputting the text to be processed into a standard text abstract extraction model for abstract extraction to obtain an abstract extraction result, wherein the standard text abstract extraction model is obtained by training by adopting the multi-model training device in the fourth aspect.
In a sixth aspect, the present application provides a text segmentation apparatus, comprising:
the segmented text acquisition module is used for acquiring a text to be processed;
and the text segmentation module is used for inputting the text to be processed into a standard text segmentation model for text segmentation to obtain a text segmentation result, wherein the standard text segmentation model is obtained by training by adopting the multi-model training device according to any one of the fourth aspect.
In a seventh aspect, an electronic device is provided, which includes a processor, a communication interface, a memory, and a communication bus, where the processor, the communication interface, and the memory complete communication with each other through the communication bus;
a memory for storing a computer program;
a processor, configured to implement the steps of the multi-model training method according to any one of the embodiments of the first aspect, or implement the steps of the abstract extraction method according to the second aspect, or implement the steps of the text segmentation method according to the third aspect when executing the program stored in the memory.
In an eighth aspect, a computer-readable storage medium is provided, on which a computer program is stored, which, when being executed by a processor, implements the steps of the multi-model training method according to any one of the embodiments of the first aspect, or implements the steps of the abstract extraction method according to the second aspect, or implements the steps of the text segmentation method according to the third aspect.
Compared with the prior art, the technical scheme provided by the embodiment of the application has the following advantages:
according to the multi-model training method, the abstract segmentation method, the text segmentation method, the device, the electronic equipment and the computer readable storage medium, the text abstract extraction model and the text segmentation model can be trained simultaneously by using the training single sentence set obtained by the training text set, the abstract coding features of the training single sentences are used during the training of the text segmentation model, and the paragraph coding features are used during the training of the text abstract extraction model, so that the segmentation model obtained by training can be learned by combining the abstract coding features during the training of the segmentation model, the accuracy of the segmentation model obtained by training can be improved, the abstract extraction model can be learned by combining the paragraph coding features during the training of the abstract extraction model, and the accuracy of the abstract extraction model obtained by training can be improved. Therefore, the embodiment of the invention can not only improve the acquisition efficiency of the text segmentation model and the abstract extraction model obtained by training, but also improve the model accuracy of the text segmentation model and the abstract extraction model obtained by training.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description, serve to explain the principles of the invention.
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without inventive exercise.
Fig. 1 is a schematic flowchart of a multi-model training method according to an embodiment of the present disclosure;
fig. 2 is a schematic flowchart of a method for abstracting a summary according to an embodiment of the present application;
fig. 3 is a schematic flowchart of a text segmentation method according to an embodiment of the present application;
FIG. 4 is a block diagram of a multi-model training apparatus according to an embodiment of the present disclosure;
fig. 5 is a schematic block diagram of a summary extraction apparatus according to an embodiment of the present disclosure;
FIG. 6 is a block diagram of a text segmentation apparatus according to an embodiment of the present application;
fig. 7 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
Fig. 1 is a schematic flowchart of a multi-model training method according to an embodiment of the present disclosure. In this embodiment, the multi-model training method includes:
s11, obtaining a training text set, and dividing the text in the training text set into single sentences to obtain a training single sentence set.
In this embodiment, the training text set is data used for model training. Specifically, in this embodiment, the training text set is data of a text type, and specifically may be a chinese text, an english text, and the like.
In this embodiment, the training text set includes a plurality of paragraph texts, where the paragraph text data may be text data crawled from local or network, the paragraph text includes a plurality of single sentences, and includes paragraph tags and abstract tags for the single sentences in the paragraph text, and the paragraph tags and the abstract tags are used to mark whether the sentence is a paragraph and whether the sentence is an abstract.
For example, each sentence sentiAll contain respective labels
Figure 93118DEST_PATH_IMAGE001
And
Figure 410966DEST_PATH_IMAGE002
wherein
Figure 378791DEST_PATH_IMAGE001
Indicating whether the ith sentence is a paragraph boundary,
Figure 98486DEST_PATH_IMAGE002
indicating whether the ith sentence is included in the abstract, i.e., whether the ith sentence is an abstract, specifically,
Figure 142665DEST_PATH_IMAGE001
and
Figure 213389DEST_PATH_IMAGE002
can be 0 or 1 when
Figure 735637DEST_PATH_IMAGE001
When 0, it means that the ith sentence is not a paragraph boundary, when
Figure 575417DEST_PATH_IMAGE003
When 1, it means that the ith sentence is a paragraph boundary, when
Figure 790498DEST_PATH_IMAGE002
0 means that the ith sentence is not a summary, when
Figure 817360DEST_PATH_IMAGE002
A value of 1 indicates that the ith sentence is a summary.
In this embodiment, when the training text set is chinese, the text in the training text set may be divided into single sentences by punctuation marks. For example, when recognizing the occurrence of a symbol such as "a character symbol corresponding to a period", "a character symbol corresponding to an exclamation mark", "a character symbol corresponding to a question mark", the text before extracting the symbol is one sentence.
Further, after obtaining the training single sentence set, the method may further include:
and deleting stop words, tone words and repeated words in the training single sentence set.
Specifically, stop words are used herein to refer to common words (e.g., "and," "the," etc.), numbers, punctuation marks and other special characters, etc.
In the embodiment, the information of the training single sentence set is deleted, so that the single sentence can be simplified, the simple single sentence can be obtained, and the efficiency of model training is improved under the condition of ensuring the training accuracy.
And S12, performing feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set.
In this embodiment, the feature extraction of the training single sentence set may include performing word segmentation on the training single sentence in the training single sentence set to obtain word segmentation representation of the sentence, and performing feature extraction on the word segmentation representation of the sentence through a preset feature extraction model.
Specifically, a word segmenter (Tokenizer) may be employed to tag a sentence to obtain all words that appear in the sentence. And add [ CLS ] at the beginning and [ SEP ] at the end of each sentence.
In this embodiment, the feature extraction model is a model capable of performing sentence vector extraction, for example, the feature extraction model is an ELMo model.
Preferably, the feature extraction model is a BERT model. BERT is capable of capturing bi-directional context information, consisting of multiple bi-directional transform encoder layers.
For example, all the sent with the delimiters [ CLS ] and [ SEP ] are input as sequences into the model BERT, since the input length of the sequences is not fixed, the sequence length MAXLENN is set in advance, if the input sequence length is smaller than MAXLENN, the length uniformity is ensured by the character string padding, beyond MAXLENN, the sequences are truncated, and input is divided into a plurality of times (e.g., twice).
Preferably, the performing feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set includes:
extracting the mark embedding characteristics, distinguishing embedding characteristics and position embedding characteristics of the training single sentences in the training single sentence set through a first extraction network of a preset characteristic extraction model;
combining the mark embedding characteristics, the distinguishing embedding characteristics and the position embedding characteristics of the training single sentence and inputting the combination into a second extraction network of the characteristic extraction model to obtain a single sentence vector of the training single sentence;
and summarizing all the obtained single sentence vectors to obtain the training single sentence vector set.
In this embodiment, the feature extraction model includes a plurality of extraction networks, for example, the second extraction network is the last several (for example, three) transform layers in the BERT model, and the first extraction network is the other transform layers in the BERT model.
In this embodiment, the mark embedding feature is a word vector of a word in a single sentence, the distinguishing embedding feature is used to distinguish the single sentence from an adjacent sentence into different sentences, and the position embedding feature is used to identify an absolute position of the word in the single sentence.
For example, in distinguishing embedded features, E is usedAAnd EBTo distinguish two different sentences, i.e. for the sentence sentiIn other words, if i is odd, the distinguishing embedding feature is EAOtherwise, it is EB. For example: input 5 sentences Sent1, sent2, sent3, sent4, sent5The corresponding distinguishing embedding characteristic is { E }A, EB, EA, EB, EA}。
For example, the mark Embedding characteristic is Token Embedding, the distinguishing Embedding characteristic is Segment Embedding, and the Position Embedding characteristic is Position Embedding, so that the mark E of a sentence is Token Embedding + Segment Embedding + Position Embedding, E passes through multiple transform layers of BERT to obtain the context Embedding of each sentence, the context of each sentence is embedded through the last transform layer of BERT to obtain the vector representation of a single sentence, i.e., a single sentence vector, and all the single sentence vectors are summarized to obtain a training single sentence vector set.
Preferably, in an embodiment of the present invention, after obtaining the training single sentence vector set, the method further includes:
and dividing the training single sentence vector set into a training set, a verification set and a test set. The training set is used for training the model, the verification set is used for verifying the model, and the test set is used for testing the model.
Specifically, M% of data in the training single sentence vector set may be randomly taken as a training set, N% of texts may be taken as a verification set, and the last (100-M-N)% of texts may be taken as a test set, where the values of M and N may be preset.
And S13, extracting paragraph coding features and abstract coding features of the training single sentences in the training single sentence vector set.
In this embodiment, the paragraph coding feature of the training single sentence represents a paragraph feature of the training single sentence in the entire training text, and specifically may include a position of the single sentence in the paragraph, for example, the paragraph coding feature of a certain single sentence is N/M, M represents the number of sentences in the paragraph to which the single sentence belongs (M is different in value for different paragraphs), and N represents that the current single sentence is the second sentence in the paragraph to which the single sentence belongs.
The abstract coding characteristics of a single sentence indicate the abstract characteristics of the single sentence in a paragraph, and specifically may include whether the single sentence is an abstract sentence, and whether the single sentence is the first sentence of the abstract sentence. For example, a single sentence has a digest coding characteristic of K/P, where P represents the number of digests when the sentence is a digest sentence, K represents that the single sentence is the first sentence in the digests, and if the current sentence is not a digest sentence, it is represented as 0/0.
In this embodiment, the paragraph coding features and the abstract coding features of each single sentence may be pre-labeled, and then the paragraph coding features and the abstract coding features of the pre-labeled single sentences are directly extracted when the model is trained.
S14, performing first training on the pre-constructed text segmentation model by using the training single sentence vector set and the abstract coding features of the training single sentences in the training single sentence vector set, and performing second training on the pre-constructed text abstract extraction model by using the training single sentence vector set and the paragraph coding features of the training single sentences in the training single sentence vector set to obtain a standard text segmentation model and a standard text abstract extraction model.
For example, a single sentence vector of a single sentence in a training single sentence vector set is
Figure 159611DEST_PATH_IMAGE004
If the abstract code of the single sentence is K/P, the abstract code will be
Figure 588318DEST_PATH_IMAGE005
Splicing with K/P to obtain
Figure 974300DEST_PATH_IMAGE006
Then will be
Figure 754037DEST_PATH_IMAGE007
The first training is performed by inputting the first training data into a text segmentation model.
In this embodiment, the text segmentation model is a linear classifier.
For example, the text segmentation model is:
Figure 883667DEST_PATH_IMAGE008
wherein the content of the first and second substances,
Figure 166881DEST_PATH_IMAGE009
a value indicating a predictive tag for a single sentence, specifically,
Figure 989344DEST_PATH_IMAGE010
is a predicted value for a single sentence as a paragraph boundary,
Figure 256377DEST_PATH_IMAGE011
in order to activate the function sigmoid,
Figure 189698DEST_PATH_IMAGE012
representing the transpose of the preset parameter vector,
Figure 576686DEST_PATH_IMAGE013
representing an input sentence sentiThe vector representation of (i.e. the combination of the single sentence vector and the digest-coding features of the single sentence),
Figure 570049DEST_PATH_IMAGE014
indicating a preset bias.
In this embodiment, the first training is performed
Figure 58800DEST_PATH_IMAGE015
And
Figure 530232DEST_PATH_IMAGE014
the value of (1) then including
Figure 788038DEST_PATH_IMAGE016
And
Figure 952303DEST_PATH_IMAGE014
the text segmentation model of the value of (a) is a trained standard text segmentation model.
Furthermore, during the first training, the training set can be concentrated by using the training single sentence vectors, and then the trained model is verified and corrected through the verification set. In particular, predictive tag values may be utilized
Figure 662770DEST_PATH_IMAGE017
And actual tag value
Figure 937894DEST_PATH_IMAGE018
And (4) comparing the numerical values, determining whether the paragraph boundary prediction of the single sentence is accurate, and then counting the accurate proportion of the single sentence prediction to obtain the trained standard text segmentation model. And if the accuracy is lower than the preset accuracy, training the text segmentation model again until obtaining a model with the accuracy reaching the preset accuracy.
Similarly, for example, a single sentence vector of a single sentence in a set of training single sentence vectors is
Figure 66518DEST_PATH_IMAGE019
If the paragraph coding feature of the single sentence is N/M, then it will be
Figure 136105DEST_PATH_IMAGE020
Splicing with N/M to obtain
Figure 599448DEST_PATH_IMAGE021
Then will be
Figure 147104DEST_PATH_IMAGE022
And inputting the text abstract extraction model to perform second training.
In this embodiment, the text summarization extraction model may also be a linear classifier.
For example, the text summarization extraction model is:
Figure 379502DEST_PATH_IMAGE023
wherein the content of the first and second substances,
Figure 619990DEST_PATH_IMAGE024
representing the digest predictive tag value for a single sentence,
Figure 570629DEST_PATH_IMAGE025
in order to activate the function sigmoid,
Figure 436822DEST_PATH_IMAGE026
representing an input sentence sentiA vector representation of (i.e. a combination of a single sentence vector and paragraph coding features of the single sentence),
Figure 523727DEST_PATH_IMAGE027
and
Figure 200696DEST_PATH_IMAGE028
are parameters of the linear classifier.
In this embodiment, the second training is performed
Figure 638631DEST_PATH_IMAGE029
And
Figure 793669DEST_PATH_IMAGE030
the value of (1) then including
Figure 659DEST_PATH_IMAGE031
And
Figure 582950DEST_PATH_IMAGE030
the text abstract extraction model of the value of (1) is a trained standard text abstract extraction model.
Furthermore, during the second training, the training set can be concentrated by using the training single sentence vectors, and the trained model is verified and corrected through the verification set. In particular, predictive tag values may be utilized
Figure 242601DEST_PATH_IMAGE032
And actual tag value
Figure 201330DEST_PATH_IMAGE033
And comparing the numerical values to determine whether the abstract prediction of the single sentence is accurate, and then counting the accurate proportion of the abstract prediction of all the single sentences to obtain a trained standard text abstract extraction model. And if the accuracy is lower than the preset accuracy, training the text abstract extraction model again until a model with the accuracy reaching the preset accuracy is obtained.
Further, when training and verification processes occur, a lot of data is generated
Figure 13560DEST_PATH_IMAGE034
And
Figure 32331DEST_PATH_IMAGE014
the best 3 of the combinations (namely the ones with the most correct label prediction) are selected from the verification set
Figure 179279DEST_PATH_IMAGE035
And
Figure 941698DEST_PATH_IMAGE014
in the combination of (1), reserve these 3
Figure 857702DEST_PATH_IMAGE036
And
Figure 47375DEST_PATH_IMAGE014
combinations of (a) and (b).
Further, the performing a second training on the pre-constructed text abstract extraction model by using the training single sentence vector set and the paragraph coding features of the training single sentences in the training single sentence vector set includes:
inputting the training single sentence vector set and paragraph coding features of the training single sentences in the training single sentence vector set to a pre-constructed document feature extraction model to obtain an abstract training single sentence vector set;
and inputting the abstract training single sentence vector set into the text abstract extraction model for second training.
In this embodiment, the document feature extraction model may be a linear classification model, or a Long-Short Term Memory network (LSTM) model, or a transform model.
Specifically, one may be used first
Figure 681618DEST_PATH_IMAGE037
A function that combines a single sentence vector and paragraph coding features of the single sentence, namely:
Figure 982150DEST_PATH_IMAGE038
wherein
Figure 752659DEST_PATH_IMAGE039
It is shown that the initial value is,
Figure 362501DEST_PATH_IMAGE040
vector representation representing sentences, paragraph codingThe feature N/M may be immediately adjacent [ CLS ] of the sentence]After the delimiter, the sentence content is spaced apart with "@".
For example, will
Figure 749620DEST_PATH_IMAGE041
Inputting the data into a Transformer model, and extracting features at a document level by using a Multi-Head Attention mechanism (Multi-Head Attention) and Layer Normalization (Layer Normalization) in the Transformer model.
Is shown in the following formula, wherein
Figure 588263DEST_PATH_IMAGE042
The layer normalization operation is represented as a function of,
Figure 478859DEST_PATH_IMAGE043
a multi-head attention mechanism is shown,
Figure 10334DEST_PATH_IMAGE044
a feed-forward network is shown in which,
Figure 353591DEST_PATH_IMAGE045
representing the depth of the stack of layers of the transform model (e.g. the network layers of the multiple decoders and encoders in the transform model), i.e.
Figure 261504DEST_PATH_IMAGE045
Is an indication of the depth of the layer,
Figure 6606DEST_PATH_IMAGE046
after the input is input into a Transformer model, a multi-head attention mechanism and layer normalization are gradually carried out through each layer, and a final sentence vector is obtained.
Figure 708983DEST_PATH_IMAGE047
Figure 290268DEST_PATH_IMAGE048
Wherein the content of the first and second substances,
Figure 470714DEST_PATH_IMAGE049
is shown as
Figure 601481DEST_PATH_IMAGE050
The vector representation of the output of the layer,
Figure 474759DEST_PATH_IMAGE051
is shown as
Figure 792608DEST_PATH_IMAGE052
The vector output by the layer passes through
Figure 776744DEST_PATH_IMAGE053
The resulting vectors of the Layer Normalization operation of the layers,
Figure 496439DEST_PATH_IMAGE054
is shown as
Figure 540618DEST_PATH_IMAGE055
Vector representation of sentences output by layers, i.e. passing through
Figure 345763DEST_PATH_IMAGE056
Extracting information by a layer Transformer model to obtain a final vector of a certain sentence i and expressing the final vector as
Figure 382858DEST_PATH_IMAGE057
And then, the vector representation of the sentence is input into a text abstract extraction model for secondary classification, specifically, the text abstract extraction model may be:
Figure 222638DEST_PATH_IMAGE058
wherein the content of the first and second substances,
Figure 172140DEST_PATH_IMAGE059
representing the digest predictive tag value for a single sentence,
Figure 464581DEST_PATH_IMAGE060
in order to activate the function sigmoid,
Figure 56099DEST_PATH_IMAGE061
representing an input sentence sentiA vector representation of (i.e. a combination of a single sentence vector and paragraph coding features of the single sentence),
Figure 484806DEST_PATH_IMAGE062
and
Figure 870788DEST_PATH_IMAGE063
are parameters of the linear classifier.
According to the method and the device, the text abstract extraction model and the text segmentation model are trained simultaneously by using the training single sentence set obtained by the training text set, the abstract coding features of the training single sentence are used when the text segmentation model is trained, the paragraph coding features are used when the text abstract extraction model is extracted, the abstract coding features can be combined for learning when the segmentation model is trained, the accuracy of the segmentation model obtained by training is improved, the paragraph coding features can be combined for learning when the abstract extraction model is trained, and the accuracy of the abstract extraction model obtained by training is improved. Therefore, the embodiment of the invention can not only improve the acquisition efficiency of the text segmentation model and the abstract extraction model obtained by training, but also improve the model accuracy of the text segmentation model and the abstract extraction model obtained by training.
Fig. 2 is a flowchart illustrating a method for abstracting a summary according to an embodiment of the present disclosure. In this embodiment, the digest extraction method includes:
and S21, acquiring the text to be processed.
In this embodiment, the text to be processed may be input by a user, or may be a text to be segmented obtained in a local or network database, and the text to be processed may include a plurality of single sentences.
The type of the text to be processed can be a Chinese type, or other language types.
And S22, inputting the text to be processed into a standard text abstract extraction model for abstract extraction to obtain an abstract extraction result.
In this embodiment, the standard text abstract extraction model is obtained by training by using the multi-model training method described in the foregoing method embodiment.
In this embodiment, a single-sentence dividing process may be performed on the text to be processed, that is, the text to be processed is divided into single sentences, then the probability that a single sentence in the text to be processed is a summary sentence is identified by using the standard text summary extraction model, and a sentence with the probability greater than the preset probability is determined as the summary sentence.
Further, the abstract extraction result includes an abstract sentence identified as an abstract and an abstract probability of the abstract sentence, and after the abstract extraction result is obtained, the method further includes:
determining the number of the abstracts for generating the abstract paragraphs, and sequencing the abstract sentences according to the abstract probability;
judging whether at least two abstract sentences with similarity greater than preset similarity exist in the plurality of abstract sentences from front to back according to the sequence, wherein the total number of the plurality of abstract sentences is equal to the number of the abstract sentences;
if at least two abstract sentences with the similarity greater than the preset similarity exist, reserving any one of the at least two abstract sentences as a target abstract sentence, deleting abstract sentences except the target abstract sentence, and executing the operation of judging whether at least two abstract sentences with the similarity greater than the preset similarity exist in the plurality of abstract sentences from front to back according to the sequence;
and if at least two abstract sentences with the similarity larger than the preset similarity do not exist, forming the abstract sentences into abstract paragraphs.
In this embodiment, the summary paragraph is a summary paragraph composed of summaries. In the embodiment, a plurality of sentences are selected from the abstract extraction result to form the abstract paragraphs, so that the brief accuracy of generating the abstract is improved.
For example, it is determined that the number of abstracts for generating an abstract paragraph is 8, if it is determined that there are 12 sentences in the sentences with the probability greater than the preset probability, that is, there are 12 abstract sentences, the 12 sentences are sorted in the order of the abstract probability from large to small, and it is determined whether there are two sentences with a large similarity in the first 8 sentences, specifically, it is possible to determine whether similar or repeated phrases in the two sentences are greater than the preset number of terms (for example, it is determined whether similar or repeated phrases in the two sentences are greater than 3), and if similar or repeated phrases in the two sentences are greater than the preset number of terms, it is determined that the two sentences are similar sentences, and one of the two sentences is selected and deleted, preferably, a sentence with a low abstract probability is selected and deleted, and a sentence with a high abstract probability is retained; then judging whether a sentence with larger similarity exists in the 9 th sentence in the original sequence and the 7 sentences after deleting the sentence, if not, determining that the 7 th sentence after deleting the sentence and the 9 th sentence form a summary paragraph, if the 9 th sentence in the original sequence and any one of the 7 sentences after deleting the sentence are sentences with larger similarity, deleting the 9 th sentence, and taking the 10 th sentence in the original sequence and the 7 sentences to judge the similarity of the sentences, and so on until obtaining the summary paragraph without similar sentences, reducing redundant information of the summary paragraph, and enabling the summary paragraph to be more refined and the contained information to be more complete.
In this embodiment, since the standard text abstract extraction model is obtained by training by using the multi-model training method in the foregoing method embodiment, the method embodiment of the present invention can accurately judge the abstract sentence of the text to be processed, obtain an accurate abstract extraction result, and obtain a brief and accurate abstract paragraph.
Fig. 3 is a schematic flowchart of a text segmentation method according to an embodiment of the present application. In this embodiment, the text segmentation method includes:
and S31, acquiring the text to be processed.
In this embodiment, the text to be processed may be input by a user, or may be a text to be segmented obtained in a local or network database, and the text to be processed may include a plurality of single sentences.
The type of the text to be processed can be a Chinese type, or other language types.
And S32, inputting the text to be processed into a standard text segmentation model for text segmentation to obtain a text segmentation result.
In this embodiment, the standard text segmentation model is obtained by training using the multi-model training method described in the foregoing method embodiment.
In this embodiment, a text to be processed may be divided into single sentences, that is, the text to be processed is divided into single sentences, then labels of paragraph boundaries of different single sentences in the text to be processed are identified by using a standard text segmentation model, and which single sentences are the same paragraph and which single sentences are not the same paragraph are determined by the labels of the paragraph boundaries, so as to determine a segmentation result of the text to be processed.
In this embodiment, since the standard text segmentation model is obtained by training using the multi-model training method described in the foregoing method embodiment, the standard text segmentation model can accurately segment the text to be processed, so as to obtain an accurate segmentation result.
As shown in fig. 4, an embodiment of the present application provides a block diagram of a multi-model training apparatus 40, where the multi-model training apparatus 40 includes: a training text acquisition module 41, a first feature extraction module 42, a second feature extraction module 43, and a training module 44.
The training text acquisition module 41 is configured to acquire a training text set, and divide texts in the training text set into single sentences to obtain a training single sentence set;
the first feature extraction module 42 is configured to perform feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set;
the second feature extraction module 43 is configured to extract paragraph coding features and abstract coding features of a training single sentence in the training single sentence vector set;
the training module 44 is configured to perform first training on a pre-constructed text segmentation model by using the training single sentence vector set and the abstract coding features of the training single sentences in the training single sentence vector set, and perform second training on a pre-constructed text abstract extraction model by using the training single sentence vector set and the paragraph coding features of the training single sentences in the training single sentence vector set, so as to obtain a standard text segmentation model and a standard text abstract extraction model.
In detail, when the modules in the multi-model training device 40 in the embodiment of the present application are used, the same technical means as the multi-model training method described in fig. 1 above are adopted, and the same technical effects can be produced, which is not described herein again.
As shown in fig. 5, an embodiment of the present application provides a block diagram of a summary extraction apparatus 50, where the summary extraction apparatus 50 includes: a summary text acquisition module 51 and a summary extraction module 52.
The abstract text acquisition module 51 is configured to acquire a text to be processed;
the abstract extraction module 52 is configured to input the text to be processed into a standard text abstract extraction model for abstract extraction, so as to obtain an abstract extraction result, where the standard text abstract extraction model is obtained by training using the multi-model training apparatus described in the foregoing apparatus embodiment.
In detail, when the modules in the abstract extracting apparatus 50 in the embodiment of the present application are used, the same technical means as the abstract extracting method described in fig. 2 above is adopted, and the same technical effects can be produced, which is not described herein again.
As shown in fig. 6, an embodiment of the present application provides a block diagram of a text segmentation apparatus 60, where the text segmentation apparatus 60 includes: a segmented text acquisition module 61 and a text segmentation module 62.
The segmented text acquisition module 61 is configured to acquire a text to be processed;
the text segmentation module 62 is configured to input the text to be processed into a standard text segmentation model for text segmentation, so as to obtain a text segmentation result, where the standard text segmentation model is obtained by training using the multi-model training apparatus described in the foregoing apparatus embodiment.
In detail, when the modules in the text segmentation apparatus 60 in the embodiment of the present application are used, the same technical means as the text segmentation method described in fig. 3 above are adopted, and the same technical effects can be produced, which is not described herein again.
As shown in fig. 7, an electronic device according to an embodiment of the present application includes a processor 111, a communication interface 112, a memory 113, and a communication bus 114, where the processor 111, the communication interface 112, and the memory 113 complete communication with each other through the communication bus 114.
The memory 113 stores a computer program.
In an embodiment of the present application, the processor 111 is configured to, when executing the program stored in the memory 113, implement the multi-model training method provided in any one of the foregoing method embodiments, or implement the abstract extraction method provided in any one of the foregoing method embodiments, or implement the text segmentation method provided in any one of the foregoing method embodiments.
The multi-model training method comprises the following steps:
acquiring a training text set, and dividing texts in the training text set into single sentences to obtain a training single sentence set;
performing feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set;
extracting paragraph coding features and abstract coding features of the training single sentences in the training single sentence vector set;
and carrying out first training on a pre-constructed text segmentation model by using the training single sentence vector set and the abstract coding features of the training single sentences in the training single sentence vector set, and carrying out second training on a pre-constructed text abstract extraction model by using the training single sentence vector set and the paragraph coding features of the training single sentences in the training single sentence vector set to obtain a standard text segmentation model and a standard text abstract extraction model.
The abstract extraction method comprises the following steps:
acquiring a text to be processed;
and inputting the text to be processed into a standard text abstract extraction model for abstract extraction to obtain an abstract extraction result, wherein the standard text abstract extraction model is obtained by training by adopting the multi-model training method of any one of the method embodiments.
The text segmentation method comprises the following steps:
acquiring a text to be processed;
and inputting the text to be processed into a standard text segmentation model for text segmentation to obtain a text segmentation result, wherein the standard text segmentation model is obtained by training by adopting the multi-model training method of any one of the method embodiments.
The communication bus 114 may be a Peripheral Component Interconnect (PCI) bus, an Extended Industry Standard Architecture (EISA) bus, or the like. The communication bus 114 may be divided into an address bus, a data bus, a control bus, and the like. For ease of illustration, only one thick line is shown, but this does not mean that there is only one bus or one type of bus.
The communication interface 112 is used for communication between the above-described electronic apparatus and other apparatuses.
The memory 113 may include a Random Access Memory (RAM), and may also include a non-volatile memory (non-volatile memory), such as at least one disk memory. Optionally, the memory 113 may also be at least one storage device located remotely from the processor 111.
The processor 111 may be a general-purpose processor, and includes a Central Processing Unit (CPU), a Network Processor (NP), and the like; the integrated circuit may also be a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic device, or discrete hardware components.
Embodiments of the present application further provide a computer-readable storage medium, on which a computer program is stored, where the computer program, when executed by a processor, implements the steps of the multi-model training method as provided in any of the foregoing method embodiments, or implements the steps of the summarization extraction method as provided in any of the foregoing method embodiments, or implements the steps of the text segmentation method as provided in any of the foregoing method embodiments.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. The procedures or functions according to the embodiments of the present application are all or partially generated when the computer program instructions are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by wire (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wirelessly (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy disk, hard disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., solid state disk (ssd)), among others.
It is noted that, in this document, relational terms such as "first" and "second," and the like, may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The foregoing are merely exemplary embodiments of the present invention, which enable those skilled in the art to understand or practice the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (13)

1. A method of multi-model training, the method comprising:
acquiring a training text set, and dividing texts in the training text set into single sentences to obtain a training single sentence set;
performing feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set;
extracting paragraph coding features and abstract coding features of the training single sentences in the training single sentence vector set;
and carrying out first training on a pre-constructed text segmentation model by using the training single sentence vector set and the abstract coding features of the training single sentences in the training single sentence vector set, and carrying out second training on a pre-constructed text abstract extraction model by using the training single sentence vector set and the paragraph coding features of the training single sentences in the training single sentence vector set to obtain a standard text segmentation model and a standard text abstract extraction model.
2. The method of claim 1, wherein the performing feature extraction on the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set comprises:
extracting the mark embedding characteristics, distinguishing embedding characteristics and position embedding characteristics of the training single sentences in the training single sentence set through a first extraction network of a preset characteristic extraction model;
combining the mark embedding characteristics, the distinguishing embedding characteristics and the position embedding characteristics of the training single sentence and inputting the combination into a second extraction network of the characteristic extraction model to obtain a single sentence vector of the training single sentence;
and summarizing all the obtained single sentence vectors to obtain the training single sentence vector set.
3. The method of claim 2, wherein the feature extraction model is a BERT model.
4. The method of claim 1, wherein said second training of the pre-built text summarization extraction model using the set of training single sentence vectors and paragraph coding features of the training single sentences in the set of training single sentence vectors comprises:
inputting the training single sentence vector set and paragraph coding features of the training single sentences in the training single sentence vector set to a pre-constructed document feature extraction model to obtain an abstract training single sentence vector set;
and inputting the abstract training single sentence vector set into the text abstract extraction model for second training.
5. The method of claim 1, wherein after obtaining the set of training sentences, the method further comprises:
and deleting stop words, tone words and repeated words in the training single sentence set.
6. A method for abstracting a summary, the method comprising:
acquiring a text to be processed;
inputting the text to be processed into a standard text abstract extraction model for abstract extraction to obtain an abstract extraction result, wherein the standard text abstract extraction model is obtained by training by adopting the multi-model training method according to any one of claims 1 to 5.
7. The method of claim 6, wherein the digest extraction result includes a digest sentence identified as a digest and a digest probability of the digest sentence, and after obtaining the digest extraction result, the method further comprises:
determining the number of the abstracts for generating the abstract paragraphs, and sequencing the abstract sentences according to the abstract probability;
judging whether at least two abstract sentences with similarity greater than preset similarity exist in the plurality of abstract sentences from front to back according to the sequence, wherein the total number of the plurality of abstract sentences is equal to the number of the abstract sentences;
if at least two abstract sentences with the similarity greater than the preset similarity exist, reserving any one of the at least two abstract sentences as a target abstract sentence, deleting abstract sentences except the target abstract sentence, and executing the operation of judging whether at least two abstract sentences with the similarity greater than the preset similarity exist in the plurality of abstract sentences from front to back according to the sequence;
and if at least two abstract sentences with the similarity larger than the preset similarity do not exist, forming the abstract sentences into abstract paragraphs.
8. A method of text segmentation, the method comprising:
acquiring a text to be processed;
inputting the text to be processed into a standard text segmentation model for text segmentation to obtain a text segmentation result, wherein the standard text segmentation model is obtained by training by adopting the multi-model training method of any one of claims 1 to 5.
9. A multi-model training apparatus, the apparatus comprising:
the training text acquisition module is used for acquiring a training text set, dividing texts in the training text set into single sentences to obtain a training single sentence set;
the first feature extraction module is used for extracting features of the training single sentence set through a preset feature extraction model to obtain a training single sentence vector set;
the second characteristic extraction module is used for extracting paragraph coding characteristics and abstract coding characteristics of the training single sentences in the training single sentence vector set;
and the training module is used for carrying out first training on the pre-constructed text segmentation model by utilizing the training single sentence vector set and the abstract coding characteristics of the training single sentences in the training single sentence vector set, and carrying out second training on the pre-constructed text abstract extraction model by utilizing the training single sentence vector set and the paragraph coding characteristics of the training single sentences in the training single sentence vector set to obtain a standard text segmentation model and a standard text abstract extraction model.
10. An apparatus for abstracting a summary, the apparatus comprising:
the abstract text acquisition module is used for acquiring a text to be processed;
the abstract extraction module is used for inputting the text to be processed into a standard text abstract extraction model for abstract extraction to obtain an abstract extraction result, wherein the standard text abstract extraction model is obtained by training through the multi-model training device according to claim 9.
11. A text segmentation apparatus, characterized in that the apparatus comprises:
the segmented text acquisition module is used for acquiring a text to be processed;
a text segmentation module, configured to input the text to be processed into a standard text segmentation model for text segmentation, so as to obtain a text segmentation result, where the standard text segmentation model is obtained by training using the multi-model training apparatus according to claim 9.
12. An electronic device is characterized by comprising a processor, a communication interface, a memory and a communication bus, wherein the processor, the communication interface and the memory are communicated with each other through the communication bus;
a memory for storing a computer program;
a processor for implementing the steps of the multi-model training method of any one of claims 1 to 5, or implementing the steps of the summarization method of claim 6, or implementing the steps of the text segmentation method of claim 7, when executing a program stored in a memory.
13. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out the steps of the multi-model training method according to any one of claims 1 to 5, or the steps of the summarization extraction method according to claim 6, or the steps of the text segmentation method according to claim 7.
CN202110762240.5A 2021-07-06 2021-07-06 Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device Active CN113204956B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110762240.5A CN113204956B (en) 2021-07-06 2021-07-06 Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110762240.5A CN113204956B (en) 2021-07-06 2021-07-06 Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device

Publications (2)

Publication Number Publication Date
CN113204956A true CN113204956A (en) 2021-08-03
CN113204956B CN113204956B (en) 2021-10-08

Family

ID=77022760

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110762240.5A Active CN113204956B (en) 2021-07-06 2021-07-06 Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device

Country Status (1)

Country Link
CN (1) CN113204956B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113590963A (en) * 2021-08-04 2021-11-02 浙江新蓝网络传媒有限公司 Balanced text recommendation method
CN115098667A (en) * 2022-08-25 2022-09-23 北京聆心智能科技有限公司 Abstract generation method, device and equipment

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1894686A (en) * 2003-11-21 2007-01-10 皇家飞利浦电子股份有限公司 Text segmentation and topic annotation for document structuring
CN101446940A (en) * 2007-11-27 2009-06-03 北京大学 Method and device of automatically generating a summary for document set
CN103699525A (en) * 2014-01-03 2014-04-02 江苏金智教育信息技术有限公司 Method and device for automatically generating abstract on basis of multi-dimensional characteristics of text
CN106844340A (en) * 2017-01-10 2017-06-13 北京百度网讯科技有限公司 News in brief generation and display methods, apparatus and system based on artificial intelligence
CN109522402A (en) * 2018-10-22 2019-03-26 国家电网有限公司 A kind of abstract extraction method and storage medium based on power industry characteristic key words
CN109657051A (en) * 2018-11-30 2019-04-19 平安科技(深圳)有限公司 Text snippet generation method, device, computer equipment and storage medium
CN111191011A (en) * 2020-04-17 2020-05-22 郑州工程技术学院 Search matching method, device and equipment for text label and storage medium
CN111460789A (en) * 2020-05-15 2020-07-28 湖南工商大学 L STM sentence segmentation method, system and medium based on character embedding
CN111581341A (en) * 2020-04-21 2020-08-25 上海明略人工智能(集团)有限公司 Method for acquiring text abstract and language model generation method
CN111666764A (en) * 2020-06-02 2020-09-15 南京优慧信安科技有限公司 XLNET-based automatic summarization method and device
CN112507711A (en) * 2020-12-04 2021-03-16 南京擎盾信息科技有限公司 Text abstract extraction method and system
CN112836508A (en) * 2021-01-29 2021-05-25 平安科技(深圳)有限公司 Information extraction model training method and device, terminal equipment and storage medium

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1894686A (en) * 2003-11-21 2007-01-10 皇家飞利浦电子股份有限公司 Text segmentation and topic annotation for document structuring
CN101446940A (en) * 2007-11-27 2009-06-03 北京大学 Method and device of automatically generating a summary for document set
CN103699525A (en) * 2014-01-03 2014-04-02 江苏金智教育信息技术有限公司 Method and device for automatically generating abstract on basis of multi-dimensional characteristics of text
CN106844340A (en) * 2017-01-10 2017-06-13 北京百度网讯科技有限公司 News in brief generation and display methods, apparatus and system based on artificial intelligence
CN109522402A (en) * 2018-10-22 2019-03-26 国家电网有限公司 A kind of abstract extraction method and storage medium based on power industry characteristic key words
CN109657051A (en) * 2018-11-30 2019-04-19 平安科技(深圳)有限公司 Text snippet generation method, device, computer equipment and storage medium
CN111191011A (en) * 2020-04-17 2020-05-22 郑州工程技术学院 Search matching method, device and equipment for text label and storage medium
CN111581341A (en) * 2020-04-21 2020-08-25 上海明略人工智能(集团)有限公司 Method for acquiring text abstract and language model generation method
CN111460789A (en) * 2020-05-15 2020-07-28 湖南工商大学 L STM sentence segmentation method, system and medium based on character embedding
CN111666764A (en) * 2020-06-02 2020-09-15 南京优慧信安科技有限公司 XLNET-based automatic summarization method and device
CN112507711A (en) * 2020-12-04 2021-03-16 南京擎盾信息科技有限公司 Text abstract extraction method and system
CN112836508A (en) * 2021-01-29 2021-05-25 平安科技(深圳)有限公司 Information extraction model training method and device, terminal equipment and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113590963A (en) * 2021-08-04 2021-11-02 浙江新蓝网络传媒有限公司 Balanced text recommendation method
CN115098667A (en) * 2022-08-25 2022-09-23 北京聆心智能科技有限公司 Abstract generation method, device and equipment
CN115098667B (en) * 2022-08-25 2023-01-03 北京聆心智能科技有限公司 Abstract generation method, device and equipment

Also Published As

Publication number Publication date
CN113204956B (en) 2021-10-08

Similar Documents

Publication Publication Date Title
WO2023060795A1 (en) Automatic keyword extraction method and apparatus, and device and storage medium
WO2022105122A1 (en) Answer generation method and apparatus based on artificial intelligence, and computer device and medium
CN106951422B (en) Webpage training method and device, and search intention identification method and device
CN109858039B (en) Text information identification method and identification device
CN107341143B (en) Sentence continuity judgment method and device and electronic equipment
CN113204956B (en) Multi-model training method, abstract segmentation method, text segmentation method and text segmentation device
CN112686022A (en) Method and device for detecting illegal corpus, computer equipment and storage medium
CN110083832B (en) Article reprint relation identification method, device, equipment and readable storage medium
CN114218945A (en) Entity identification method, device, server and storage medium
CN112214984A (en) Content plagiarism identification method, device, equipment and storage medium
CN113590810A (en) Abstract generation model training method, abstract generation device and electronic equipment
CN114722832A (en) Abstract extraction method, device, equipment and storage medium
US20220101060A1 (en) Text partitioning method, text classifying method, apparatus, device and storage medium
CN114416981A (en) Long text classification method, device, equipment and storage medium
CN114003725A (en) Information annotation model construction method and information annotation generation method
CN112699671B (en) Language labeling method, device, computer equipment and storage medium
CN113515593A (en) Topic detection method and device based on clustering model and computer equipment
CN111783424A (en) Text clause dividing method and device
CN113255319B (en) Model training method, text segmentation method, abstract extraction method and device
CN110874408B (en) Model training method, text recognition device and computing equipment
CN113011162B (en) Reference digestion method, device, electronic equipment and medium
CN114911936A (en) Model training and comment recognition method and device, electronic equipment and medium
CN115129843A (en) Dialog text abstract extraction method and device
CN114186140A (en) Social interaction information processing method and device, electronic equipment and storage medium
CN116029280A (en) Method, device, computing equipment and storage medium for extracting key information of document

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant