WO2022141872A1 - Procédé et appareil de génération de résumé de document, dispositif informatique et support de stockage - Google Patents

Procédé et appareil de génération de résumé de document, dispositif informatique et support de stockage Download PDF

Info

Publication number
WO2022141872A1
WO2022141872A1 PCT/CN2021/084241 CN2021084241W WO2022141872A1 WO 2022141872 A1 WO2022141872 A1 WO 2022141872A1 CN 2021084241 W CN2021084241 W CN 2021084241W WO 2022141872 A1 WO2022141872 A1 WO 2022141872A1
Authority
WO
WIPO (PCT)
Prior art keywords
abstract
target
document
sentences
model
Prior art date
Application number
PCT/CN2021/084241
Other languages
English (en)
Chinese (zh)
Inventor
颜泽龙
王健宗
吴天博
程宁
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2022141872A1 publication Critical patent/WO2022141872A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/34Browsing; Visualisation therefor
    • G06F16/345Summarisation for human users
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3344Query execution using natural language analysis

Definitions

  • the present application relates to the technical field of artificial intelligence, and in particular, to a method, apparatus, computer equipment and storage medium for generating a literature abstract.
  • the embodiments of the present application provide a method, device, computer equipment and storage medium for generating a literature abstract, so as to solve the problem of low accuracy of the current method for obtaining abstracts.
  • a method for generating literature abstracts comprising:
  • the database is queried based on the abstract keywords, and obtained from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the original document is processed by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • a device for generating literature abstracts comprising:
  • An abstract generation request acquisition module is used to acquire an abstract generation request, where the abstract generation request includes abstract keywords
  • An original document acquisition module configured to query a database based on the abstract keywords, and acquire from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the target sentence acquisition module is used to process the original document by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • Directed acyclic graph acquisition module for converting M indivual
  • the target sentence is input into the trained target model combination, and the corresponding target sentence is obtained.
  • M*(M-1)/2 A directed acyclic graph;
  • Target summary acquisition module for M*(M-1)/2 A directed acyclic graph is obtained, and the target summary is obtained.
  • a computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, wherein the processor implements the following steps when executing the computer-readable instructions:
  • the database is queried based on the abstract keywords, and obtained from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the original document is processed by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
  • the database is queried based on the abstract keywords, and obtained from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the original document is processed by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • the original documents corresponding to the abstract keywords can be automatically determined, so as to ensure the accuracy of subsequent target abstracts and reduce labor costs.
  • the original document is processed by using a pre-trained extractive document abstract model to quickly obtain M indivual
  • the target sentence makes a strong connection between the target sentences and ensures that the subsequently generated target abstract records the important information of the original document. Will M indivual
  • the target sentence is input into the trained target model combination, and the corresponding target sentence is obtained.
  • M*(M-1)/2 A directed acyclic graph by determining the sequence between any two target sentences, it ensures that the subsequent process of sorting the target sentences is simple, effectively improves the accuracy, and ensures that the generated target abstracts are more coherent. based on M*(M-1)/2 A directed acyclic graph can quickly obtain a coherent target summary.
  • picture 1 is a schematic diagram of an application environment of the method for generating a document abstract in an embodiment of the present application
  • picture 2 is a flow chart of a method for generating a document abstract in an embodiment of the present application
  • picture 3 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 4 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 5 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 6 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 7 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 8 is a schematic block diagram of a document abstract generating device in an embodiment of the present application.
  • picture 9 is a topology diagram in an embodiment of the present application.
  • picture 10 It is a schematic diagram of a computer device in an embodiment of the present application.
  • the document abstract generation method provided in the embodiment of the present application can be applied in the application environment shown in FIG. 1 .
  • the literature abstract generation method is applied in an abstract generation system
  • the abstract generation system includes a client and a server as shown in FIG. 1 , the client and the server communicate through the network, and are used to determine any two through the target model combination.
  • the sequence between the target sentences improves the accuracy of generating the target summary and ensures that the generated target summary has better coherence.
  • the client also known as the client, refers to the program corresponding to the server and providing local services for the client.
  • Clients can be installed on, but not limited to, various personal computers, laptops, smartphones, tablets, and portable wearable devices.
  • the server can be implemented as an independent server or a server cluster composed of multiple servers.
  • a method for generating a document abstract is provided, and the method is applied to the server in FIG. 1 as an example for description, including the following steps:
  • the digest generation request is a request for generating a target digest.
  • the abstract keyword is a keyword for generating a target abstract required by the user, so that the corresponding original document can be obtained according to the abstract keyword.
  • the abstract keyword may be xx disease, xx medical treatment, or xx financial product.
  • abstract is also called summary and abstract. Abstracts are short texts that describe the important content of the literature concisely and precisely for the purpose of providing the outline of the content of the literature.
  • an original display interface for generating the target abstract is displayed in the client, the user clicks on the original display interface, and sends an instruction to fill in the abstract keywords to the server.
  • the server After the server obtains the instruction to fill in the abstract keywords, it controls the client to enter the abstract keywords.
  • a generation interface where the user fills in at least one abstract keyword in the abstract keyword generation interface. Understandably, in order to ensure that the generated target abstract meets the requirements of the user, when the number of abstract keywords filled in by the user is less than the preset number of keywords When the user fills in the summary keywords, more similar keywords are recommended for the user.
  • the client is controlled to display the confirmation button, and the user clicks the confirmation button to form an abstract.
  • the request is sent to the server, and when the server receives the abstract generation request, it parses the abstract generation request to obtain the abstract keywords, thereby realizing the automatic generation of the target abstract.
  • S202 Query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database.
  • the database is the library used to store the original literature.
  • the initial documents are documents pre-stored in the database. It is understood that the initial documents include abstracts of various types of abstracts, and the initial documents include documents corresponding to the abstract keywords and documents that do not correspond to the abstract keywords.
  • the initial documents can be It is the literature corresponding to the medical direction, the literature corresponding to the food direction, or the literature corresponding to the financial direction, etc.
  • the original document refers to the abstract corresponding to the abstract keywords.
  • the document abstracts in the database are classified in advance, and abstract documents of the same abstract type are obtained, and each abstract document carries at least one abstract type, so as to provide technical support for obtaining the corresponding abstract type according to the abstract keywords in the future.
  • the server obtains the abstract keywords, it uses a matching algorithm to query the database according to the abstract keywords, so as to obtain the original documents corresponding to the abstract keywords from the document abstracts, realize the automatic determination of the original documents of the same abstract type, and ensure subsequent target abstracts. accuracy and reduce labor costs.
  • the abstract type refers to the type corresponding to the literature abstract.
  • the extractive document abstract model refers to a model that directly extracts the required target sentences from N original documents.
  • the extractive document abstract model can better retain the information of the original documents and can effectively improve the accuracy of the target abstract. Reduce the grammatical and syntactic error rates of the subsequently generated target abstracts.
  • the extractive document abstract model is the NeuSUM model, which automatically extracts sentences with higher scores in the original documents as target sentences, reducing labor costs.
  • the model uses sentence benefit as a scoring method, taking into account the relationship between sentences, to ensure that the obtained target sentences are highly relevant, and the subsequently generated target summaries are more coherent.
  • the target sentence refers to the sentence used to form the target summary.
  • the original document is input into the pre-trained extractive document abstract model.
  • the original document is divided so that the original document is divided into multiple abstract sentences, and the abstract sentences are converted into sentence vectors in the embedding layer to convert them into computer Recognizable format; the sentence vector is encoded at the target coding layer to obtain the target coding vector containing semantic information to retain more information of the summary sentence; at the scoring coding layer, the target coding vector is scored according to the benefit of the sentence, and each target coding vector is obtained.
  • the score corresponding to the abstract sentence, using the sentence benefit as the scoring method, that is, with The ROUGE evaluation index is used as the index for scoring the summary sentences to consider the relationship between the summary sentences, and the top M summary sentences with higher scores are used as the target sentences, so that the target sentences have a strong connection, and the target sentence can be quickly obtained.
  • the training of the extractive document abstract model is to continuously adjust the weight of the initial model by using the back-propagation algorithm until the weight of the initial model converges, and then the extractive document abstract model is obtained.
  • S204 Input the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
  • the target model combination is the model used to rank any two target sentences.
  • the target combination model includes a BERT model and an MLP model.
  • the BERT model and the MLP model can be used to accurately determine the sequence of target sentences, and provide technical support for the subsequent generation of target abstracts.
  • the BERT model is used to process any two target sentences, and the transformer structure is used in the encoding layer and decoding layer of the BERT model to ensure that the output of the BERT model is a semantic information vector with semantic information. For example, any two Each target sentence is input into the BERT model in the form of target sentence 1; target sentence 2, and the corresponding output vector with semantic information is obtained, and the output vector is input into the MLP model to obtain the sequence between any two target sentences.
  • the training process is as follows: 1. Obtain a training corpus sample.
  • the training expected sample includes positive sample sentence pairs and negative sample sentence pairs. Understandably, positive sample sentence pairs exist between sentences. Context relationship, there is no context relationship between sentences for negative sample sentence pairs; 2.
  • the MLP model is a multi-layer perceptron model, which is used to perform binary classification processing on any two sentences to obtain the sequence of any two target sentences.
  • the MLP model is trained before the MLP model is used.
  • the training process is as follows: obtaining training samples and sequence labels corresponding to the training samples, wherein the training samples are original sentence pairs; input the training samples into the initial model, and obtain Predicted sentence sequence results; according to the sequence labels and sentence sequence results, the classification accuracy is calculated, and when the classification accuracy is greater than the preset value, the MLP model is obtained.
  • a directed acyclic graph is a directed graph without loops.
  • the directed acyclic graph is S1 ⁇ S2 ⁇ S3 ⁇ S4, where S1, S2, S3 and S4 are the target sentences.
  • M target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, each sentence combination is input into the BERT model to obtain a semantic information vector, and the semantic information vector is input into the MLP model to obtain the sequence of any two target sentences, and form a directed acyclic graph based on the sequence.
  • by determining the sequence between any two target sentences it ensures that the subsequent sequence of the target sentences is simple and effective.
  • the target sentences are S1, S2 and S3
  • the three sentences are combined in pairs to obtain three sentence combinations, namely S1 and S3, S1 and S2, S2 and S3, and each A sentence combination is input into the target model combination to obtain the sequence of any two target sentences, so as to ensure that the subsequent sorting process of the target sentences is simple, the accuracy is effectively improved, and the consistency of the generated target summaries is ensured.
  • the target summary refers to the summary required by the user.
  • M*(M-1)/2 directed acyclic graphs are processed to obtain a topology map
  • the breadth-first algorithm is used to process the topology map to obtain the current in-degree of each target sentence.
  • the in-degree sorts the target sentences and obtains the target summary.
  • the process is relatively simple, and a coherent target summary can be obtained quickly.
  • the target sentences are S1, S2, and S3, the directed acyclic graph is S1 ⁇ S2; S1 ⁇ S3; S2 ⁇ S3; the topological graph shown in Figure 9 is obtained by processing the directed acyclic graph.
  • the bottom element repeat this process until all target sentences are pushed into the stack queue, the stack queue output by this process is the target summary, the process is relatively simple, and a coherent target summary can be obtained quickly.
  • the in-degree derived from the graph theory algorithm, usually refers to the sum of the times that a certain point in a directed graph is used as the end point of an edge in the graph.
  • the current in-degree refers to the in-degree corresponding to each target sentence.
  • the method for generating a document abstract queries a database based on the abstract keywords, obtains N original documents corresponding to the abstract keywords from the initial documents stored in the database, realizes automatic determination of the original documents of the same abstract type, and ensures subsequent Accuracy of target summaries, reducing labor costs.
  • the pre-trained extractive literature abstract model is used to process the original literature, and quickly obtain M target sentences, which makes the target sentences have a strong connection, and ensures that the subsequently generated target abstracts record the important information of the original literature. Input the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
  • step S202 is to query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database, including:
  • S301 Query the classification table in the database based on the abstract keyword, and obtain the abstract type corresponding to the abstract keyword.
  • the classification table is a preset table, and the classification table is used to indicate the association relationship between preset keywords and abstract types.
  • the preset keywords are words corresponding to the abstract keywords.
  • the summary type refers to the type of summary, for example, the summary type can be medical type, financial type, mechanical type, etc. As an example, if the preset keyword is xx disease, the corresponding abstract type is medical type.
  • a matching algorithm is used to match the abstract keywords with the preset keywords in the classification table. If the matching is successful, it means that there are preset keywords corresponding to the abstract keywords. Therefore, according to the corresponding preset keywords The keyword can be used to obtain the abstract type corresponding to the abstract keyword, which provides technical support for the subsequent determination of the original literature.
  • S302 Query the initial documents in the database based on the abstract type, and determine the N initial documents including the abstract keywords as the N initial documents.
  • the initial documents in the database are classified in advance, and after the abstract type is determined, the abstract keywords are matched with the initial documents in the abstract type to obtain the initial documents corresponding to the abstract keywords, which is faster .
  • the method for generating a document abstract queries a classification table in a database based on the abstract keyword, obtains the abstract type corresponding to the abstract keyword, and provides technical support for subsequent determination of the original document.
  • the initial documents in the database are queried based on the abstract type, and the N initial documents containing the abstract keywords are determined as the N initial documents, which is faster.
  • a pre-trained extractive document abstract model is used to process the original document to obtain M target sentences, including:
  • S401 Segment the original document to obtain at least two abstract sentences.
  • the segmentation process refers to the process of dividing the original document into multiple sentences, so that the computer can process the abstract sentences.
  • the abstract sentence is a single sentence obtained by segmenting the original document.
  • the division is performed according to commas and periods in the original document. For example, if the original document is xxxx, yyyyy; zzz, the original document is divided into xxxx, yyyyy and zzz as three sentences by searching for commas and periods.
  • S402 Input all abstract sentences into the word embedding layer of the extractive document abstract model, and obtain sentence vectors corresponding to each abstract sentence.
  • the sentence vector refers to the vector obtained after the abstract sentence is processed by the word embedding layer.
  • the abstract sentence can be converted into a corresponding vector, which is convenient for computer recognition.
  • the word embedding layer is the layer used to convert summary sentences into computer-recognizable sentence vectors.
  • S403 Input each sentence vector into the target coding layer of the extractive document abstract model to obtain a target coding vector corresponding to each sentence vector.
  • the target encoding layer is used for sentence-level and document-level encoding of sentence vectors.
  • the sentence vector is firstly input into the sentence encoding layer to obtain the original encoding vector represented by the vector of the sentence, and the original encoding vector is input into the document encoding layer to obtain the target encoding vector.
  • the scoring result refers to the result of scoring the target coding vector corresponding to each summary sentence using the scoring coding layer. Understandably, the summary sentence with a higher score is determined as the target sentence, so that the target sentence contains important information. sentences, to ensure that the subsequently generated target abstracts record the important sentences of the original literature.
  • the first M summary sentences are selected in sequence from high to low, and are determined as M target sentences.
  • sentence scoring and sentence selection are combined using an extractive document summary model, so as to associate the information of the sentences and ensure that the target sentence has important information.
  • the original document is segmented to obtain at least two abstract sentences, so that the computer can process the abstract sentences.
  • Results According to the scoring results of multiple target encoding vectors, the first M summary sentences are selected in order from high to low, and determined as M target sentences.
  • the extractive literature abstract model is used to combine sentence scoring and sentence selection. Correlate the information of the sentences to ensure that the target sentence has important information.
  • step S403 is to input each sentence vector into the target coding layer of the extractive document abstract model to obtain the target coding vector corresponding to each sentence vector, including:
  • S501 Input each sentence vector into the sentence coding layer of the extractive document abstract model for coding, and obtain the original coding vector corresponding to the sentence vector;
  • S502 Input the original encoding vector into the document encoding layer of the extractive document abstract model for re-encoding to obtain a target encoding vector.
  • the sentence encoding layer is a bidirectional GRU sentence encoding layer, and the sentence-level encoding is obtained by using the bidirectional GRU sentence encoding layer.
  • the document encoding layer refers to the bidirectional GRU document encoding layer, and the document-level encoding is obtained by using the bidirectional GRU document encoding layer.
  • the target model combination includes a pre-trained BERT model and an MLP model; as shown in Figure 6, step S204 is to input M target sentences into the trained target model combination to obtain M*( M-1)/2 directed acyclic graphs, including:
  • the sentence combination refers to the combination formed by any two target sentences, so that the context before the two target sentences can be obtained subsequently.
  • the target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, which is conducive to simplifying the subsequent steps of determining the context between any two target sentences and ensuring that any two The accuracy of contextual relationships between target sentences ensures a coherent target summary.
  • the target sentence is 3, assuming that the target sentences are S1, S2 and S3, and the target sentences are combined in pairs, the sentence combinations can be obtained as S1 and S2, S1 and S3, S2 and S3.
  • S602 Input each sentence combination into the BERT model, and obtain a semantic information vector corresponding to each sentence combination.
  • the role of BERT is to obtain a vector representation of sentence combinations.
  • BERT mainly includes word embedding layer, encoding layer and decoding layer.
  • the role of the word embedding layer is to map documents to vectors, the input is a document, and the output is a vector.
  • Both the encoding layer and the decoding layer use the transformer structure to obtain semantic information vectors with semantic information.
  • S603 Input the semantic information vector into the MLP model to obtain a directed acyclic graph of any two target sentences.
  • the Bert model and the MLP model are used to extract and classify the abstract sentences, so as to obtain the target sentences, and determine the contextual dependencies between the target sentences, so as to solve the problem that in the prior art, only the Bert model is used for classification accuracy. low problem.
  • target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, which is beneficial to simplify the subsequent steps of determining the contextual relationship between any two target sentences And ensure the accuracy of determining the contextual relationship between any two target sentences, and ensure a coherent target summary.
  • Input each sentence combination into the BERT model to obtain the semantic information vector corresponding to each sentence combination; input the semantic information vector into the MLP model to obtain the directed acyclic graph of any two target sentences to obtain the target sentence and determine the target sentence In order to solve the problem of low classification accuracy using only Bert model in the prior art.
  • step S205 that is, based on M*(M-1)/2 directed acyclic graphs, obtain the target digest, including:
  • S701 Process M*(M-1)/2 directed acyclic graphs to obtain a topology graph.
  • the topological graph refers to the graph formed by the collection of all directed acyclic graphs, so that the subsequent breadth-first traversal can be performed to obtain the current in-degree of each target sentence.
  • S702 Use the breadth-first algorithm to traverse the topology map to obtain the current in-degree of each target sentence.
  • the breadth-first algorithm also known as breadth-first search, breadth-first search and horizontal-first search, is a graph search algorithm; the so-called breadth is a layer-by-layer traversal.
  • the breadth-first algorithm is used to process the topology map to obtain the current in-degree of each target sentence, and the target sentences are sorted according to the current in-degree to obtain the target abstract.
  • the process is relatively simple, and a coherent target abstract can be obtained quickly.
  • S703 Push all target sentences into the stack according to the current in-degree to obtain a stack queue.
  • the target sentence 2 is pushed into the stack queue as the bottom element of the stack, and this process is repeated until all target sentences are pushed into the stack queue.
  • the stack queue formed by this process is the target summary. The process is relatively simple and can be Get a coherent summary of your goals quickly.
  • the target sentence S1 is before the target sentence S2, the target sentence S2 is before the target sentence S3, and the target sentence S3 is before the target sentence S4; then S1 points to S2, S3 and S4 respectively, S2 points to S3 and S4 respectively, S3 points to S4, so , the current in-degree of S1 is 0; the current in-degree of S2 is 1; the current in-degree of S3 is 2; the current in-degree of S4 is 3.
  • the stack queue is S1 ⁇ S2 ⁇ S3 ⁇ S4.
  • the target abstract is obtained according to the order of each target sentence in the stack queue, so as to ensure that the generated target abstract is coherent.
  • the method for generating literature abstracts processes M*(M-1)/2 directed acyclic graphs to obtain a topology graph, so that breadth-first traversal can be performed subsequently to obtain the current in-degree of each target sentence.
  • the breadth-first algorithm is used to traverse the topology map to obtain the current in-degree of each target sentence; according to the current in-degree, all target sentences are pushed into the stack to obtain the stack queue; based on the stack queue, the target abstract is obtained to ensure that the generated target abstract is coherent.
  • an apparatus for generating a document abstract is provided, and the apparatus for generating a document abstract corresponds one-to-one with the method for generating a document abstract in the above-mentioned embodiment.
  • the document abstract generation device includes an abstract generation request acquisition module 801 , an original document acquisition module 802 , a target sentence acquisition module 803 , a directed acyclic graph acquisition module 804 and a target abstract acquisition module 805 .
  • the detailed description of each functional module is as follows:
  • the abstract generation request acquiring module 801 is configured to acquire an abstract generation request, where the abstract generation request includes abstract keywords.
  • the original document obtaining module 802 is configured to query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database.
  • the target sentence obtaining module 803 is used to process the original document by using the pre-trained extractive document abstract model to obtain M target sentences.
  • the directed acyclic graph obtaining module 804 is used for inputting the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
  • the target abstract obtaining module 805 is configured to obtain the target abstract based on M*(M-1)/2 directed acyclic graphs.
  • the original document acquisition module 802 includes: an abstract type acquisition unit and an original document acquisition unit.
  • the abstract type obtaining unit is used to query the classification table in the database based on the abstract keywords, and obtain the abstract types corresponding to the abstract keywords.
  • the original document acquisition unit is used to query the document abstracts in the database according to the abstract type, and determine N document abstracts including abstract keywords as N original documents.
  • the target sentence obtaining module 803 includes: a segmentation processing unit, a sentence vector obtaining unit, a target coding vector obtaining unit, and a scoring result obtaining unit.
  • the segmentation processing unit is used to segment the original document to obtain at least two abstract sentences.
  • the sentence vector obtaining unit is used to input all the abstract sentences into the word embedding layer of the extractive document abstract model, and obtain the sentence vector corresponding to each abstract sentence.
  • the target coding vector obtaining unit is used to input each sentence vector into the target coding layer of the extractive document abstract model, and obtain the target coding vector corresponding to each sentence vector.
  • the scoring result obtaining unit is used for inputting the target coding vector into the scoring coding layer of the extractive document summarization model to obtain the scoring result corresponding to each summary sentence.
  • the target sentence obtaining unit is used to select the first M summary sentences in order from high to low according to the scoring results of multiple target encoding vectors, and determine them as M target sentences.
  • the target coding vector obtaining unit includes: a first coding subunit and a second coding subunit.
  • the first encoding subunit is used to input each sentence vector into the sentence encoding layer of the extractive document abstract model for encoding, and obtain the original encoding vector corresponding to the sentence vector.
  • the second encoding subunit is used for inputting the original encoding vector into the document encoding layer of the extractive document abstract model for re-encoding to obtain the target encoding vector.
  • the target model combination includes a BERT model and an MLP model.
  • the directed acyclic graph obtaining module 804 includes: a sentence combination obtaining unit, a semantic information vector obtaining unit and a directed acyclic graph obtaining unit.
  • the sentence combination acquisition unit is used to combine the target sentences in pairs to obtain M*(M-1)/2 sentence combinations.
  • the semantic information vector obtaining unit is used to input each sentence combination into the BERT model, and obtain the semantic information vector corresponding to each sentence combination.
  • the directed acyclic graph acquisition unit is used to input the semantic information vector into the MLP model to obtain the directed acyclic graph of any two target sentences.
  • the target digest obtaining module 805 includes: a topology map obtaining unit, an in-degree obtaining unit, a stack queue obtaining unit and a target digest obtaining unit.
  • the topology map obtaining unit is used for processing M*(M-1)/2 directed acyclic graphs to obtain a topology map.
  • the in-degree obtaining unit is used to traverse the topology map using the breadth-first algorithm to obtain the current in-degree of each target sentence.
  • the stack queue obtaining unit is used to push all target sentences into the stack according to the current in-degree, and obtain the stack queue.
  • the target digest obtaining unit is used to obtain the target digest based on the stack queue.
  • Each module in the above-mentioned document abstract generating apparatus can be implemented in whole or in part by software, hardware and combinations thereof.
  • the above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
  • a computer device is provided, and the computer device may be a server, and its internal structure diagram may be as shown in FIG. 10 .
  • the computer device includes a processor, memory, a network interface, and a database connected by a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium, an internal memory.
  • the non-volatile storage medium stores an operating system, computer readable instructions and a database.
  • the internal memory provides an environment for the execution of the operating system and computer-readable instructions in the non-volatile storage medium.
  • the database of the computer equipment is used to store the directed acyclic graph.
  • the network interface of the computer device is used to communicate with an external terminal through a network connection.
  • the computer readable instructions when executed by a processor, implement a method for generating a literature abstract.
  • a computer device including a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, and the processor implements the documents in the above embodiments when the processor executes the computer-readable instructions
  • the steps of the abstract generating method such as steps S201-S205 shown in FIG. 2, or steps shown in FIG. 3 to FIG. 7, are not repeated here in order to avoid repetition.
  • the processor executes the computer-readable instructions, the functions of each module/unit in this embodiment of the document abstract generating apparatus are implemented, for example, the abstract generation request acquisition module 801, the original document acquisition module 802, the target sentence acquisition shown in FIG.
  • the functions of the module 803 , the directed acyclic graph obtaining module 804 and the target abstract obtaining module 805 are not repeated here in order to avoid repetition.
  • the readable storage medium provided by this embodiment includes a non-volatile readable storage medium and a volatile readable storage medium.
  • one or more readable storage media storing computer-readable instructions are provided, and the computer-readable instructions are stored on the readable storage medium, and the computer-readable instructions are executed by the processor to implement the above-mentioned implementation.
  • the steps of the method for generating a document abstract in the example such as steps S201-S205 shown in FIG. 2 , or steps shown in FIG. 3 to FIG. 7 , are not repeated here to avoid repetition.
  • the processor executes the computer-readable instructions
  • the functions of each module/unit in this embodiment of the document abstract generating apparatus are implemented, for example, the abstract generation request acquisition module 801, the original document acquisition module 802, the target sentence acquisition shown in FIG. 8
  • the functions of the module 803 , the directed acyclic graph obtaining module 804 and the target abstract obtaining module 805 are not repeated here in order to avoid repetition.
  • Nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in various forms such as static RAM (SRAM), dynamic RAM (MRAM), synchronous MRAM (SMRAM), double data rate SMRAM (MMRSMRAM), enhanced SMRAM (ESMRAM), synchronous chain Road (Synchlink) MRAM (SLMRAM), memory bus (Rambus) direct RAM (RMRAM), direct memory bus dynamic RAM (MRMRAM), and memory bus dynamic RAM (RMRAM), etc.

Abstract

La présente invention concerne un procédé et un appareil de génération de résumé de document, un dispositif informatique et un support de stockage qui se rapportent au domaine technique de l'intelligence artificielle. Le procédé consiste à : obtenir une requête de génération de résumé, la requête de génération de résumé comprenant un mot-clé de résumé (S201) ; interroger une base de données sur la base du mot-clé de résumé et obtenir N documents originaux correspondant au mot-clé de résumé à partir de documents initiaux stockés dans la base de données (S202) ; traiter les documents originaux à l'aide d'un modèle de résumé de document de type à extraction pré-entraîné pour obtenir M phrases cibles (S203) ; entrer les M phrases cibles dans une combinaison de modèles cibles entraînés pour obtenir M*(M-1)/2 graphes acycliques dirigés correspondant aux phrases cibles (S204) ; et obtenir un résumé cible sur la base des M*(M-1)/2 graphes acycliques dirigés (S205). Le procédé détermine la séquence de deux phrases cibles quelconques en utilisant la combinaison de modèles cibles, ce qui permet d'améliorer la précision de génération du résumé cible et d'assurer une bonne cohérence du résumé cible généré.
PCT/CN2021/084241 2020-12-30 2021-03-31 Procédé et appareil de génération de résumé de document, dispositif informatique et support de stockage WO2022141872A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011623844.3 2020-12-30
CN202011623844.3A CN112732898A (zh) 2020-12-30 2020-12-30 文献摘要生成方法、装置、计算机设备及存储介质

Publications (1)

Publication Number Publication Date
WO2022141872A1 true WO2022141872A1 (fr) 2022-07-07

Family

ID=75609644

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/084241 WO2022141872A1 (fr) 2020-12-30 2021-03-31 Procédé et appareil de génération de résumé de document, dispositif informatique et support de stockage

Country Status (2)

Country Link
CN (1) CN112732898A (fr)
WO (1) WO2022141872A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115809329A (zh) * 2023-01-30 2023-03-17 医智生命科技(天津)有限公司 一种长文本的摘要生成方法
CN116912047A (zh) * 2023-09-13 2023-10-20 湘潭大学 一种专利结构感知相似性检测方法

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113407710A (zh) * 2021-06-07 2021-09-17 维沃移动通信有限公司 信息显示方法、装置、电子设备及可读存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100228693A1 (en) * 2009-03-06 2010-09-09 phiScape AG Method and system for generating a document representation
US20150339288A1 (en) * 2014-05-23 2015-11-26 Codeq Llc Systems and Methods for Generating Summaries of Documents
CN108280112A (zh) * 2017-06-22 2018-07-13 腾讯科技(深圳)有限公司 摘要生成方法、装置及计算机设备
CN109657054A (zh) * 2018-12-13 2019-04-19 北京百度网讯科技有限公司 摘要生成方法、装置、服务器及存储介质
CN111414471A (zh) * 2020-03-20 2020-07-14 北京百度网讯科技有限公司 用于输出信息的方法和装置
CN111858913A (zh) * 2020-07-08 2020-10-30 北京嘀嘀无限科技发展有限公司 一种自动生成文本摘要的方法和系统

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100228693A1 (en) * 2009-03-06 2010-09-09 phiScape AG Method and system for generating a document representation
US20150339288A1 (en) * 2014-05-23 2015-11-26 Codeq Llc Systems and Methods for Generating Summaries of Documents
CN108280112A (zh) * 2017-06-22 2018-07-13 腾讯科技(深圳)有限公司 摘要生成方法、装置及计算机设备
CN109657054A (zh) * 2018-12-13 2019-04-19 北京百度网讯科技有限公司 摘要生成方法、装置、服务器及存储介质
CN111414471A (zh) * 2020-03-20 2020-07-14 北京百度网讯科技有限公司 用于输出信息的方法和装置
CN111858913A (zh) * 2020-07-08 2020-10-30 北京嘀嘀无限科技发展有限公司 一种自动生成文本摘要的方法和系统

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115809329A (zh) * 2023-01-30 2023-03-17 医智生命科技(天津)有限公司 一种长文本的摘要生成方法
CN116912047A (zh) * 2023-09-13 2023-10-20 湘潭大学 一种专利结构感知相似性检测方法
CN116912047B (zh) * 2023-09-13 2023-11-28 湘潭大学 一种专利结构感知相似性检测方法

Also Published As

Publication number Publication date
CN112732898A (zh) 2021-04-30

Similar Documents

Publication Publication Date Title
US20200342177A1 (en) Capturing rich response relationships with small-data neural networks
US11645314B2 (en) Interactive information retrieval using knowledge graphs
US10832011B2 (en) Question answering system using multilingual information sources
WO2019136993A1 (fr) Procédé et dispositif de calcul de similarité de texte, appareil informatique, et support de stockage
US10025819B2 (en) Generating a query statement based on unstructured input
US20200184307A1 (en) Utilizing recurrent neural networks to recognize and extract open intent from text inputs
US9318027B2 (en) Caching natural language questions and results in a question and answer system
WO2022141872A1 (fr) Procédé et appareil de génération de résumé de document, dispositif informatique et support de stockage
WO2021114810A1 (fr) Procédé de recommandation de document officiel à base de structure de graphe, appareil, dispositif informatique et support
CN111026319B (zh) 一种智能文本处理方法、装置、电子设备及存储介质
JP2021152963A (ja) 語義特徴の生成方法、モデルトレーニング方法、装置、機器、媒体及びプログラム
US11551437B2 (en) Collaborative information extraction
WO2020244065A1 (fr) Procédé, appareil et dispositif de définition de vecteur de caractère basés sur l'intelligence artificielle et support de stockage
WO2022088671A1 (fr) Procédé et appareil de réponse automatique à des questions, dispositif et support de mémoire
CN111026320B (zh) 多模态智能文本处理方法、装置、电子设备及存储介质
US11263400B2 (en) Identifying entity attribute relations
US20220300543A1 (en) Method of retrieving query, electronic device and medium
US11861918B2 (en) Image analysis for problem resolution
WO2023045187A1 (fr) Procédé et appareil de recherche sémantique basés sur un réseau neuronal, dispositif et support de stockage
WO2022174496A1 (fr) Procédé et appareil d'annotation de données basés sur un modèle génératif, dispositif et support de stockage
WO2023231331A1 (fr) Procédé, système et dispositif d'extraction de connaissances, et support de stockage
TW202001621A (zh) 語料庫產生方法及裝置、人機互動處理方法及裝置
CN112632258A (zh) 文本数据处理方法、装置、计算机设备和存储介质
US20220114361A1 (en) Multi-word concept tagging for images using short text decoder
CN111142728B (zh) 车载环境智能文本处理方法、装置、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21912642

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21912642

Country of ref document: EP

Kind code of ref document: A1