WO2022141872A1 - Document abstract generation method and apparatus, computer device, and storage medium - Google Patents

Document abstract generation method and apparatus, computer device, and storage medium Download PDF

Info

Publication number
WO2022141872A1
WO2022141872A1 PCT/CN2021/084241 CN2021084241W WO2022141872A1 WO 2022141872 A1 WO2022141872 A1 WO 2022141872A1 CN 2021084241 W CN2021084241 W CN 2021084241W WO 2022141872 A1 WO2022141872 A1 WO 2022141872A1
Authority
WO
WIPO (PCT)
Prior art keywords
abstract
target
document
sentences
model
Prior art date
Application number
PCT/CN2021/084241
Other languages
French (fr)
Chinese (zh)
Inventor
颜泽龙
王健宗
吴天博
程宁
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2022141872A1 publication Critical patent/WO2022141872A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/34Browsing; Visualisation therefor
    • G06F16/345Summarisation for human users
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3344Query execution using natural language analysis

Definitions

  • the present application relates to the technical field of artificial intelligence, and in particular, to a method, apparatus, computer equipment and storage medium for generating a literature abstract.
  • the embodiments of the present application provide a method, device, computer equipment and storage medium for generating a literature abstract, so as to solve the problem of low accuracy of the current method for obtaining abstracts.
  • a method for generating literature abstracts comprising:
  • the database is queried based on the abstract keywords, and obtained from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the original document is processed by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • a device for generating literature abstracts comprising:
  • An abstract generation request acquisition module is used to acquire an abstract generation request, where the abstract generation request includes abstract keywords
  • An original document acquisition module configured to query a database based on the abstract keywords, and acquire from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the target sentence acquisition module is used to process the original document by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • Directed acyclic graph acquisition module for converting M indivual
  • the target sentence is input into the trained target model combination, and the corresponding target sentence is obtained.
  • M*(M-1)/2 A directed acyclic graph;
  • Target summary acquisition module for M*(M-1)/2 A directed acyclic graph is obtained, and the target summary is obtained.
  • a computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, wherein the processor implements the following steps when executing the computer-readable instructions:
  • the database is queried based on the abstract keywords, and obtained from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the original document is processed by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
  • the database is queried based on the abstract keywords, and obtained from the initial documents stored in the database N indivual Original documents corresponding to the abstract keywords;
  • the original document is processed by using the pre-trained extractive document abstract model to obtain M indivual target sentence;
  • the original documents corresponding to the abstract keywords can be automatically determined, so as to ensure the accuracy of subsequent target abstracts and reduce labor costs.
  • the original document is processed by using a pre-trained extractive document abstract model to quickly obtain M indivual
  • the target sentence makes a strong connection between the target sentences and ensures that the subsequently generated target abstract records the important information of the original document. Will M indivual
  • the target sentence is input into the trained target model combination, and the corresponding target sentence is obtained.
  • M*(M-1)/2 A directed acyclic graph by determining the sequence between any two target sentences, it ensures that the subsequent process of sorting the target sentences is simple, effectively improves the accuracy, and ensures that the generated target abstracts are more coherent. based on M*(M-1)/2 A directed acyclic graph can quickly obtain a coherent target summary.
  • picture 1 is a schematic diagram of an application environment of the method for generating a document abstract in an embodiment of the present application
  • picture 2 is a flow chart of a method for generating a document abstract in an embodiment of the present application
  • picture 3 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 4 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 5 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 6 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 7 is another flow chart of the method for generating a document abstract in an embodiment of the present application.
  • picture 8 is a schematic block diagram of a document abstract generating device in an embodiment of the present application.
  • picture 9 is a topology diagram in an embodiment of the present application.
  • picture 10 It is a schematic diagram of a computer device in an embodiment of the present application.
  • the document abstract generation method provided in the embodiment of the present application can be applied in the application environment shown in FIG. 1 .
  • the literature abstract generation method is applied in an abstract generation system
  • the abstract generation system includes a client and a server as shown in FIG. 1 , the client and the server communicate through the network, and are used to determine any two through the target model combination.
  • the sequence between the target sentences improves the accuracy of generating the target summary and ensures that the generated target summary has better coherence.
  • the client also known as the client, refers to the program corresponding to the server and providing local services for the client.
  • Clients can be installed on, but not limited to, various personal computers, laptops, smartphones, tablets, and portable wearable devices.
  • the server can be implemented as an independent server or a server cluster composed of multiple servers.
  • a method for generating a document abstract is provided, and the method is applied to the server in FIG. 1 as an example for description, including the following steps:
  • the digest generation request is a request for generating a target digest.
  • the abstract keyword is a keyword for generating a target abstract required by the user, so that the corresponding original document can be obtained according to the abstract keyword.
  • the abstract keyword may be xx disease, xx medical treatment, or xx financial product.
  • abstract is also called summary and abstract. Abstracts are short texts that describe the important content of the literature concisely and precisely for the purpose of providing the outline of the content of the literature.
  • an original display interface for generating the target abstract is displayed in the client, the user clicks on the original display interface, and sends an instruction to fill in the abstract keywords to the server.
  • the server After the server obtains the instruction to fill in the abstract keywords, it controls the client to enter the abstract keywords.
  • a generation interface where the user fills in at least one abstract keyword in the abstract keyword generation interface. Understandably, in order to ensure that the generated target abstract meets the requirements of the user, when the number of abstract keywords filled in by the user is less than the preset number of keywords When the user fills in the summary keywords, more similar keywords are recommended for the user.
  • the client is controlled to display the confirmation button, and the user clicks the confirmation button to form an abstract.
  • the request is sent to the server, and when the server receives the abstract generation request, it parses the abstract generation request to obtain the abstract keywords, thereby realizing the automatic generation of the target abstract.
  • S202 Query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database.
  • the database is the library used to store the original literature.
  • the initial documents are documents pre-stored in the database. It is understood that the initial documents include abstracts of various types of abstracts, and the initial documents include documents corresponding to the abstract keywords and documents that do not correspond to the abstract keywords.
  • the initial documents can be It is the literature corresponding to the medical direction, the literature corresponding to the food direction, or the literature corresponding to the financial direction, etc.
  • the original document refers to the abstract corresponding to the abstract keywords.
  • the document abstracts in the database are classified in advance, and abstract documents of the same abstract type are obtained, and each abstract document carries at least one abstract type, so as to provide technical support for obtaining the corresponding abstract type according to the abstract keywords in the future.
  • the server obtains the abstract keywords, it uses a matching algorithm to query the database according to the abstract keywords, so as to obtain the original documents corresponding to the abstract keywords from the document abstracts, realize the automatic determination of the original documents of the same abstract type, and ensure subsequent target abstracts. accuracy and reduce labor costs.
  • the abstract type refers to the type corresponding to the literature abstract.
  • the extractive document abstract model refers to a model that directly extracts the required target sentences from N original documents.
  • the extractive document abstract model can better retain the information of the original documents and can effectively improve the accuracy of the target abstract. Reduce the grammatical and syntactic error rates of the subsequently generated target abstracts.
  • the extractive document abstract model is the NeuSUM model, which automatically extracts sentences with higher scores in the original documents as target sentences, reducing labor costs.
  • the model uses sentence benefit as a scoring method, taking into account the relationship between sentences, to ensure that the obtained target sentences are highly relevant, and the subsequently generated target summaries are more coherent.
  • the target sentence refers to the sentence used to form the target summary.
  • the original document is input into the pre-trained extractive document abstract model.
  • the original document is divided so that the original document is divided into multiple abstract sentences, and the abstract sentences are converted into sentence vectors in the embedding layer to convert them into computer Recognizable format; the sentence vector is encoded at the target coding layer to obtain the target coding vector containing semantic information to retain more information of the summary sentence; at the scoring coding layer, the target coding vector is scored according to the benefit of the sentence, and each target coding vector is obtained.
  • the score corresponding to the abstract sentence, using the sentence benefit as the scoring method, that is, with The ROUGE evaluation index is used as the index for scoring the summary sentences to consider the relationship between the summary sentences, and the top M summary sentences with higher scores are used as the target sentences, so that the target sentences have a strong connection, and the target sentence can be quickly obtained.
  • the training of the extractive document abstract model is to continuously adjust the weight of the initial model by using the back-propagation algorithm until the weight of the initial model converges, and then the extractive document abstract model is obtained.
  • S204 Input the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
  • the target model combination is the model used to rank any two target sentences.
  • the target combination model includes a BERT model and an MLP model.
  • the BERT model and the MLP model can be used to accurately determine the sequence of target sentences, and provide technical support for the subsequent generation of target abstracts.
  • the BERT model is used to process any two target sentences, and the transformer structure is used in the encoding layer and decoding layer of the BERT model to ensure that the output of the BERT model is a semantic information vector with semantic information. For example, any two Each target sentence is input into the BERT model in the form of target sentence 1; target sentence 2, and the corresponding output vector with semantic information is obtained, and the output vector is input into the MLP model to obtain the sequence between any two target sentences.
  • the training process is as follows: 1. Obtain a training corpus sample.
  • the training expected sample includes positive sample sentence pairs and negative sample sentence pairs. Understandably, positive sample sentence pairs exist between sentences. Context relationship, there is no context relationship between sentences for negative sample sentence pairs; 2.
  • the MLP model is a multi-layer perceptron model, which is used to perform binary classification processing on any two sentences to obtain the sequence of any two target sentences.
  • the MLP model is trained before the MLP model is used.
  • the training process is as follows: obtaining training samples and sequence labels corresponding to the training samples, wherein the training samples are original sentence pairs; input the training samples into the initial model, and obtain Predicted sentence sequence results; according to the sequence labels and sentence sequence results, the classification accuracy is calculated, and when the classification accuracy is greater than the preset value, the MLP model is obtained.
  • a directed acyclic graph is a directed graph without loops.
  • the directed acyclic graph is S1 ⁇ S2 ⁇ S3 ⁇ S4, where S1, S2, S3 and S4 are the target sentences.
  • M target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, each sentence combination is input into the BERT model to obtain a semantic information vector, and the semantic information vector is input into the MLP model to obtain the sequence of any two target sentences, and form a directed acyclic graph based on the sequence.
  • by determining the sequence between any two target sentences it ensures that the subsequent sequence of the target sentences is simple and effective.
  • the target sentences are S1, S2 and S3
  • the three sentences are combined in pairs to obtain three sentence combinations, namely S1 and S3, S1 and S2, S2 and S3, and each A sentence combination is input into the target model combination to obtain the sequence of any two target sentences, so as to ensure that the subsequent sorting process of the target sentences is simple, the accuracy is effectively improved, and the consistency of the generated target summaries is ensured.
  • the target summary refers to the summary required by the user.
  • M*(M-1)/2 directed acyclic graphs are processed to obtain a topology map
  • the breadth-first algorithm is used to process the topology map to obtain the current in-degree of each target sentence.
  • the in-degree sorts the target sentences and obtains the target summary.
  • the process is relatively simple, and a coherent target summary can be obtained quickly.
  • the target sentences are S1, S2, and S3, the directed acyclic graph is S1 ⁇ S2; S1 ⁇ S3; S2 ⁇ S3; the topological graph shown in Figure 9 is obtained by processing the directed acyclic graph.
  • the bottom element repeat this process until all target sentences are pushed into the stack queue, the stack queue output by this process is the target summary, the process is relatively simple, and a coherent target summary can be obtained quickly.
  • the in-degree derived from the graph theory algorithm, usually refers to the sum of the times that a certain point in a directed graph is used as the end point of an edge in the graph.
  • the current in-degree refers to the in-degree corresponding to each target sentence.
  • the method for generating a document abstract queries a database based on the abstract keywords, obtains N original documents corresponding to the abstract keywords from the initial documents stored in the database, realizes automatic determination of the original documents of the same abstract type, and ensures subsequent Accuracy of target summaries, reducing labor costs.
  • the pre-trained extractive literature abstract model is used to process the original literature, and quickly obtain M target sentences, which makes the target sentences have a strong connection, and ensures that the subsequently generated target abstracts record the important information of the original literature. Input the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
  • step S202 is to query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database, including:
  • S301 Query the classification table in the database based on the abstract keyword, and obtain the abstract type corresponding to the abstract keyword.
  • the classification table is a preset table, and the classification table is used to indicate the association relationship between preset keywords and abstract types.
  • the preset keywords are words corresponding to the abstract keywords.
  • the summary type refers to the type of summary, for example, the summary type can be medical type, financial type, mechanical type, etc. As an example, if the preset keyword is xx disease, the corresponding abstract type is medical type.
  • a matching algorithm is used to match the abstract keywords with the preset keywords in the classification table. If the matching is successful, it means that there are preset keywords corresponding to the abstract keywords. Therefore, according to the corresponding preset keywords The keyword can be used to obtain the abstract type corresponding to the abstract keyword, which provides technical support for the subsequent determination of the original literature.
  • S302 Query the initial documents in the database based on the abstract type, and determine the N initial documents including the abstract keywords as the N initial documents.
  • the initial documents in the database are classified in advance, and after the abstract type is determined, the abstract keywords are matched with the initial documents in the abstract type to obtain the initial documents corresponding to the abstract keywords, which is faster .
  • the method for generating a document abstract queries a classification table in a database based on the abstract keyword, obtains the abstract type corresponding to the abstract keyword, and provides technical support for subsequent determination of the original document.
  • the initial documents in the database are queried based on the abstract type, and the N initial documents containing the abstract keywords are determined as the N initial documents, which is faster.
  • a pre-trained extractive document abstract model is used to process the original document to obtain M target sentences, including:
  • S401 Segment the original document to obtain at least two abstract sentences.
  • the segmentation process refers to the process of dividing the original document into multiple sentences, so that the computer can process the abstract sentences.
  • the abstract sentence is a single sentence obtained by segmenting the original document.
  • the division is performed according to commas and periods in the original document. For example, if the original document is xxxx, yyyyy; zzz, the original document is divided into xxxx, yyyyy and zzz as three sentences by searching for commas and periods.
  • S402 Input all abstract sentences into the word embedding layer of the extractive document abstract model, and obtain sentence vectors corresponding to each abstract sentence.
  • the sentence vector refers to the vector obtained after the abstract sentence is processed by the word embedding layer.
  • the abstract sentence can be converted into a corresponding vector, which is convenient for computer recognition.
  • the word embedding layer is the layer used to convert summary sentences into computer-recognizable sentence vectors.
  • S403 Input each sentence vector into the target coding layer of the extractive document abstract model to obtain a target coding vector corresponding to each sentence vector.
  • the target encoding layer is used for sentence-level and document-level encoding of sentence vectors.
  • the sentence vector is firstly input into the sentence encoding layer to obtain the original encoding vector represented by the vector of the sentence, and the original encoding vector is input into the document encoding layer to obtain the target encoding vector.
  • the scoring result refers to the result of scoring the target coding vector corresponding to each summary sentence using the scoring coding layer. Understandably, the summary sentence with a higher score is determined as the target sentence, so that the target sentence contains important information. sentences, to ensure that the subsequently generated target abstracts record the important sentences of the original literature.
  • the first M summary sentences are selected in sequence from high to low, and are determined as M target sentences.
  • sentence scoring and sentence selection are combined using an extractive document summary model, so as to associate the information of the sentences and ensure that the target sentence has important information.
  • the original document is segmented to obtain at least two abstract sentences, so that the computer can process the abstract sentences.
  • Results According to the scoring results of multiple target encoding vectors, the first M summary sentences are selected in order from high to low, and determined as M target sentences.
  • the extractive literature abstract model is used to combine sentence scoring and sentence selection. Correlate the information of the sentences to ensure that the target sentence has important information.
  • step S403 is to input each sentence vector into the target coding layer of the extractive document abstract model to obtain the target coding vector corresponding to each sentence vector, including:
  • S501 Input each sentence vector into the sentence coding layer of the extractive document abstract model for coding, and obtain the original coding vector corresponding to the sentence vector;
  • S502 Input the original encoding vector into the document encoding layer of the extractive document abstract model for re-encoding to obtain a target encoding vector.
  • the sentence encoding layer is a bidirectional GRU sentence encoding layer, and the sentence-level encoding is obtained by using the bidirectional GRU sentence encoding layer.
  • the document encoding layer refers to the bidirectional GRU document encoding layer, and the document-level encoding is obtained by using the bidirectional GRU document encoding layer.
  • the target model combination includes a pre-trained BERT model and an MLP model; as shown in Figure 6, step S204 is to input M target sentences into the trained target model combination to obtain M*( M-1)/2 directed acyclic graphs, including:
  • the sentence combination refers to the combination formed by any two target sentences, so that the context before the two target sentences can be obtained subsequently.
  • the target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, which is conducive to simplifying the subsequent steps of determining the context between any two target sentences and ensuring that any two The accuracy of contextual relationships between target sentences ensures a coherent target summary.
  • the target sentence is 3, assuming that the target sentences are S1, S2 and S3, and the target sentences are combined in pairs, the sentence combinations can be obtained as S1 and S2, S1 and S3, S2 and S3.
  • S602 Input each sentence combination into the BERT model, and obtain a semantic information vector corresponding to each sentence combination.
  • the role of BERT is to obtain a vector representation of sentence combinations.
  • BERT mainly includes word embedding layer, encoding layer and decoding layer.
  • the role of the word embedding layer is to map documents to vectors, the input is a document, and the output is a vector.
  • Both the encoding layer and the decoding layer use the transformer structure to obtain semantic information vectors with semantic information.
  • S603 Input the semantic information vector into the MLP model to obtain a directed acyclic graph of any two target sentences.
  • the Bert model and the MLP model are used to extract and classify the abstract sentences, so as to obtain the target sentences, and determine the contextual dependencies between the target sentences, so as to solve the problem that in the prior art, only the Bert model is used for classification accuracy. low problem.
  • target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, which is beneficial to simplify the subsequent steps of determining the contextual relationship between any two target sentences And ensure the accuracy of determining the contextual relationship between any two target sentences, and ensure a coherent target summary.
  • Input each sentence combination into the BERT model to obtain the semantic information vector corresponding to each sentence combination; input the semantic information vector into the MLP model to obtain the directed acyclic graph of any two target sentences to obtain the target sentence and determine the target sentence In order to solve the problem of low classification accuracy using only Bert model in the prior art.
  • step S205 that is, based on M*(M-1)/2 directed acyclic graphs, obtain the target digest, including:
  • S701 Process M*(M-1)/2 directed acyclic graphs to obtain a topology graph.
  • the topological graph refers to the graph formed by the collection of all directed acyclic graphs, so that the subsequent breadth-first traversal can be performed to obtain the current in-degree of each target sentence.
  • S702 Use the breadth-first algorithm to traverse the topology map to obtain the current in-degree of each target sentence.
  • the breadth-first algorithm also known as breadth-first search, breadth-first search and horizontal-first search, is a graph search algorithm; the so-called breadth is a layer-by-layer traversal.
  • the breadth-first algorithm is used to process the topology map to obtain the current in-degree of each target sentence, and the target sentences are sorted according to the current in-degree to obtain the target abstract.
  • the process is relatively simple, and a coherent target abstract can be obtained quickly.
  • S703 Push all target sentences into the stack according to the current in-degree to obtain a stack queue.
  • the target sentence 2 is pushed into the stack queue as the bottom element of the stack, and this process is repeated until all target sentences are pushed into the stack queue.
  • the stack queue formed by this process is the target summary. The process is relatively simple and can be Get a coherent summary of your goals quickly.
  • the target sentence S1 is before the target sentence S2, the target sentence S2 is before the target sentence S3, and the target sentence S3 is before the target sentence S4; then S1 points to S2, S3 and S4 respectively, S2 points to S3 and S4 respectively, S3 points to S4, so , the current in-degree of S1 is 0; the current in-degree of S2 is 1; the current in-degree of S3 is 2; the current in-degree of S4 is 3.
  • the stack queue is S1 ⁇ S2 ⁇ S3 ⁇ S4.
  • the target abstract is obtained according to the order of each target sentence in the stack queue, so as to ensure that the generated target abstract is coherent.
  • the method for generating literature abstracts processes M*(M-1)/2 directed acyclic graphs to obtain a topology graph, so that breadth-first traversal can be performed subsequently to obtain the current in-degree of each target sentence.
  • the breadth-first algorithm is used to traverse the topology map to obtain the current in-degree of each target sentence; according to the current in-degree, all target sentences are pushed into the stack to obtain the stack queue; based on the stack queue, the target abstract is obtained to ensure that the generated target abstract is coherent.
  • an apparatus for generating a document abstract is provided, and the apparatus for generating a document abstract corresponds one-to-one with the method for generating a document abstract in the above-mentioned embodiment.
  • the document abstract generation device includes an abstract generation request acquisition module 801 , an original document acquisition module 802 , a target sentence acquisition module 803 , a directed acyclic graph acquisition module 804 and a target abstract acquisition module 805 .
  • the detailed description of each functional module is as follows:
  • the abstract generation request acquiring module 801 is configured to acquire an abstract generation request, where the abstract generation request includes abstract keywords.
  • the original document obtaining module 802 is configured to query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database.
  • the target sentence obtaining module 803 is used to process the original document by using the pre-trained extractive document abstract model to obtain M target sentences.
  • the directed acyclic graph obtaining module 804 is used for inputting the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
  • the target abstract obtaining module 805 is configured to obtain the target abstract based on M*(M-1)/2 directed acyclic graphs.
  • the original document acquisition module 802 includes: an abstract type acquisition unit and an original document acquisition unit.
  • the abstract type obtaining unit is used to query the classification table in the database based on the abstract keywords, and obtain the abstract types corresponding to the abstract keywords.
  • the original document acquisition unit is used to query the document abstracts in the database according to the abstract type, and determine N document abstracts including abstract keywords as N original documents.
  • the target sentence obtaining module 803 includes: a segmentation processing unit, a sentence vector obtaining unit, a target coding vector obtaining unit, and a scoring result obtaining unit.
  • the segmentation processing unit is used to segment the original document to obtain at least two abstract sentences.
  • the sentence vector obtaining unit is used to input all the abstract sentences into the word embedding layer of the extractive document abstract model, and obtain the sentence vector corresponding to each abstract sentence.
  • the target coding vector obtaining unit is used to input each sentence vector into the target coding layer of the extractive document abstract model, and obtain the target coding vector corresponding to each sentence vector.
  • the scoring result obtaining unit is used for inputting the target coding vector into the scoring coding layer of the extractive document summarization model to obtain the scoring result corresponding to each summary sentence.
  • the target sentence obtaining unit is used to select the first M summary sentences in order from high to low according to the scoring results of multiple target encoding vectors, and determine them as M target sentences.
  • the target coding vector obtaining unit includes: a first coding subunit and a second coding subunit.
  • the first encoding subunit is used to input each sentence vector into the sentence encoding layer of the extractive document abstract model for encoding, and obtain the original encoding vector corresponding to the sentence vector.
  • the second encoding subunit is used for inputting the original encoding vector into the document encoding layer of the extractive document abstract model for re-encoding to obtain the target encoding vector.
  • the target model combination includes a BERT model and an MLP model.
  • the directed acyclic graph obtaining module 804 includes: a sentence combination obtaining unit, a semantic information vector obtaining unit and a directed acyclic graph obtaining unit.
  • the sentence combination acquisition unit is used to combine the target sentences in pairs to obtain M*(M-1)/2 sentence combinations.
  • the semantic information vector obtaining unit is used to input each sentence combination into the BERT model, and obtain the semantic information vector corresponding to each sentence combination.
  • the directed acyclic graph acquisition unit is used to input the semantic information vector into the MLP model to obtain the directed acyclic graph of any two target sentences.
  • the target digest obtaining module 805 includes: a topology map obtaining unit, an in-degree obtaining unit, a stack queue obtaining unit and a target digest obtaining unit.
  • the topology map obtaining unit is used for processing M*(M-1)/2 directed acyclic graphs to obtain a topology map.
  • the in-degree obtaining unit is used to traverse the topology map using the breadth-first algorithm to obtain the current in-degree of each target sentence.
  • the stack queue obtaining unit is used to push all target sentences into the stack according to the current in-degree, and obtain the stack queue.
  • the target digest obtaining unit is used to obtain the target digest based on the stack queue.
  • Each module in the above-mentioned document abstract generating apparatus can be implemented in whole or in part by software, hardware and combinations thereof.
  • the above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
  • a computer device is provided, and the computer device may be a server, and its internal structure diagram may be as shown in FIG. 10 .
  • the computer device includes a processor, memory, a network interface, and a database connected by a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium, an internal memory.
  • the non-volatile storage medium stores an operating system, computer readable instructions and a database.
  • the internal memory provides an environment for the execution of the operating system and computer-readable instructions in the non-volatile storage medium.
  • the database of the computer equipment is used to store the directed acyclic graph.
  • the network interface of the computer device is used to communicate with an external terminal through a network connection.
  • the computer readable instructions when executed by a processor, implement a method for generating a literature abstract.
  • a computer device including a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, and the processor implements the documents in the above embodiments when the processor executes the computer-readable instructions
  • the steps of the abstract generating method such as steps S201-S205 shown in FIG. 2, or steps shown in FIG. 3 to FIG. 7, are not repeated here in order to avoid repetition.
  • the processor executes the computer-readable instructions, the functions of each module/unit in this embodiment of the document abstract generating apparatus are implemented, for example, the abstract generation request acquisition module 801, the original document acquisition module 802, the target sentence acquisition shown in FIG.
  • the functions of the module 803 , the directed acyclic graph obtaining module 804 and the target abstract obtaining module 805 are not repeated here in order to avoid repetition.
  • the readable storage medium provided by this embodiment includes a non-volatile readable storage medium and a volatile readable storage medium.
  • one or more readable storage media storing computer-readable instructions are provided, and the computer-readable instructions are stored on the readable storage medium, and the computer-readable instructions are executed by the processor to implement the above-mentioned implementation.
  • the steps of the method for generating a document abstract in the example such as steps S201-S205 shown in FIG. 2 , or steps shown in FIG. 3 to FIG. 7 , are not repeated here to avoid repetition.
  • the processor executes the computer-readable instructions
  • the functions of each module/unit in this embodiment of the document abstract generating apparatus are implemented, for example, the abstract generation request acquisition module 801, the original document acquisition module 802, the target sentence acquisition shown in FIG. 8
  • the functions of the module 803 , the directed acyclic graph obtaining module 804 and the target abstract obtaining module 805 are not repeated here in order to avoid repetition.
  • Nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in various forms such as static RAM (SRAM), dynamic RAM (MRAM), synchronous MRAM (SMRAM), double data rate SMRAM (MMRSMRAM), enhanced SMRAM (ESMRAM), synchronous chain Road (Synchlink) MRAM (SLMRAM), memory bus (Rambus) direct RAM (RMRAM), direct memory bus dynamic RAM (MRMRAM), and memory bus dynamic RAM (RMRAM), etc.

Abstract

A document abstract generation method and apparatus, a computer device, and a storage medium, relating to the technical field of artificial intelligence. The method comprises: obtaining an abstract generation request, the abstract generation request comprising an abstract keyword (S201); querying a database on the basis of the abstract keyword, and obtaining N original documents corresponding to the abstract keyword from initial documents stored in the database (S202); processing the original documents by using a pre-trained extraction type document abstract model to obtain M target sentences (S203); inputting the M target sentences into a trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences (S204); and obtaining a target abstract on the basis of the M*(M-1)/2 directed acyclic graphs (S205). The method determines the sequence of any two target sentences by using the target model combination, thereby improving the accuracy of generating the target abstract, and ensuring good coherence of the generated target abstract.

Description

文献摘要生成方法、装置、计算机设备及存储介质Method, device, computer equipment and storage medium for generating literature abstract
本申请要求于2020年12月30日提交中国专利局、申请号为202011623844.3,发明名称为“文献摘要生成方法、装置、计算机设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed on December 30, 2020 with the application number 202011623844.3 and the title of the invention is "Method, Apparatus, Computer Equipment and Storage Medium for Generating Document Abstracts", the entire contents of which are by reference Incorporated in this application.
  
技术领域technical field
本申请涉及人工智能技术领域,尤其涉及一种文献摘要生成方法、装置、计算机设备及存储介质。The present application relates to the technical field of artificial intelligence, and in particular, to a method, apparatus, computer equipment and storage medium for generating a literature abstract.
背景技术Background technique
随着互联网产生的文献数据越来越多,文献信息过载问题日益严重,用户需要花费大量时间从数量众多的文献数据得到关键信息,阅读效率低;但是发明人发现当前通常是利用单个模型提取文献中的句子得到摘要,但是目前得到摘要的方法准确率较低。With more and more document data generated by the Internet, the problem of document information overload is becoming more and more serious. Users need to spend a lot of time to obtain key information from a large number of document data, and the reading efficiency is low; however, the inventor found that currently, a single model is usually used to extract documents. Sentences in are summarized, but the current methods of obtaining summaries have low accuracy.
技术问题technical problem
本申请实施例提供一种文献摘要生成方法、装置、计算机设备及存储介质,以解决目前得到摘要的方法准确率较低的问题。 The embodiments of the present application provide a method, device, computer equipment and storage medium for generating a literature abstract, so as to solve the problem of low accuracy of the current method for obtaining abstracts.
技术解决方案technical solutions
一种文献摘要生成方法,包括:A method for generating literature abstracts, comprising:
获取摘要生成请求,所述摘要生成请求包括摘要关键词;Obtain an abstract generation request, where the abstract generation request includes abstract keywords;
基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取The database is queried based on the abstract keywords, and obtained from the initial documents stored in the database NN indivual 与所述摘要关键词对应的原始文献;Original documents corresponding to the abstract keywords;
采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到The original document is processed by using the pre-trained extractive document abstract model to obtain MM indivual 目标句子;target sentence;
Will MM indivual 所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的The target sentence is input into the trained target model combination, and the corresponding target sentence is obtained. M*(M-1)/2M*(M-1)/2 个有向无环图;A directed acyclic graph;
基于based on M*(M-1)/2M*(M-1)/2 个所述有向无环图,获取目标摘要。A directed acyclic graph is obtained, and the target summary is obtained.
一种文献摘要生成装置,包括:A device for generating literature abstracts, comprising:
摘要生成请求获取模块,用于获取摘要生成请求,所述摘要生成请求包括摘要关键词;An abstract generation request acquisition module is used to acquire an abstract generation request, where the abstract generation request includes abstract keywords;
原始文献获取模块,用于基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取An original document acquisition module, configured to query a database based on the abstract keywords, and acquire from the initial documents stored in the database NN indivual 与所述摘要关键词对应的原始文献;Original documents corresponding to the abstract keywords;
目标句子获取模块,用于采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到The target sentence acquisition module is used to process the original document by using the pre-trained extractive document abstract model to obtain MM indivual 目标句子;target sentence;
有向无环图获取模块,用于将Directed acyclic graph acquisition module for converting MM indivual 所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的The target sentence is input into the trained target model combination, and the corresponding target sentence is obtained. M*(M-1)/2M*(M-1)/2 个有向无环图;A directed acyclic graph;
目标摘要获取模块,用于基于Target summary acquisition module for M*(M-1)/2M*(M-1)/2 个所述有向无环图,获取目标摘要。A directed acyclic graph is obtained, and the target summary is obtained.
一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,其中,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, wherein the processor implements the following steps when executing the computer-readable instructions:
获取摘要生成请求,所述摘要生成请求包括摘要关键词;Obtain an abstract generation request, where the abstract generation request includes abstract keywords;
基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取The database is queried based on the abstract keywords, and obtained from the initial documents stored in the database NN indivual 与所述摘要关键词对应的原始文献;Original documents corresponding to the abstract keywords;
采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到The original document is processed by using the pre-trained extractive document abstract model to obtain MM indivual 目标句子;target sentence;
Will MM indivual 所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的The target sentence is input into the trained target model combination, and the corresponding target sentence is obtained. M*(M-1)/2M*(M-1)/2 个有向无环图;A directed acyclic graph;
基于based on M*(M-1)/2M*(M-1)/2 个所述有向无环图,获取目标摘要。A directed acyclic graph is obtained, and the target summary is obtained.
一个或多个存储有计算机可读指令的可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
获取摘要生成请求,所述摘要生成请求包括摘要关键词;Obtain an abstract generation request, where the abstract generation request includes abstract keywords;
基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取The database is queried based on the abstract keywords, and obtained from the initial documents stored in the database NN indivual 与所述摘要关键词对应的原始文献;Original documents corresponding to the abstract keywords;
采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到The original document is processed by using the pre-trained extractive document abstract model to obtain MM indivual 目标句子;target sentence;
Will MM indivual 所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的The target sentence is input into the trained target model combination, and the corresponding target sentence is obtained. M*(M-1)/2M*(M-1)/2 个有向无环图;A directed acyclic graph;
基于based on M*(M-1)/2M*(M-1)/2 个所述有向无环图,获取目标摘要。A directed acyclic graph is obtained, and the target summary is obtained.
有益效果beneficial effect
上述文献摘要生成方法、装置、计算机设备及存储介质,基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取The above-mentioned method, device, computer equipment and storage medium for generating a document abstract, query a database based on the abstract keyword, and obtain from the initial document stored in the database NN indivual 与所述摘要关键词对应的原始文献,实现自动化确定相同摘要类型的原始文献,确保后续的目标摘要的准确性,减少人工成本。采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,快速得到For the original documents corresponding to the abstract keywords, the original documents of the same abstract type can be automatically determined, so as to ensure the accuracy of subsequent target abstracts and reduce labor costs. The original document is processed by using a pre-trained extractive document abstract model to quickly obtain MM indivual 目标句子,使得目标句子之间具有较强的联系,确保后续生成的目标摘要记载原始文献的重要信息。将The target sentence makes a strong connection between the target sentences and ensures that the subsequently generated target abstract records the important information of the original document. Will MM indivual 所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的The target sentence is input into the trained target model combination, and the corresponding target sentence is obtained. M*(M-1)/2M*(M-1)/2 个有向无环图,通过确定任意两个目标句子之间的前后顺序,保证后续可以对目标句子进行排序过程简单,有效提高准确率,保证生成的目标摘要连贯性较佳。基于A directed acyclic graph, by determining the sequence between any two target sentences, it ensures that the subsequent process of sorting the target sentences is simple, effectively improves the accuracy, and ensures that the generated target abstracts are more coherent. based on M*(M-1)/2M*(M-1)/2 个所述有向无环图,可以快速得到连贯的目标摘要。A directed acyclic graph can quickly obtain a coherent target summary.
附图说明Description of drawings
为了更清楚地说明本申请实施例的技术方案,下面将对本申请实施例的描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the technical solutions of the embodiments of the present application more clearly, the following briefly introduces the drawings that are used in the description of the embodiments of the present application. Obviously, the drawings in the following description are only some embodiments of the present application. , for those of ordinary skill in the art, other drawings can also be obtained from these drawings without creative labor.
picture 11 是本申请一实施例中文献摘要生成方法的一应用环境示意图;is a schematic diagram of an application environment of the method for generating a document abstract in an embodiment of the present application;
picture 22 是本申请一实施例中文献摘要生成方法的一流程图;is a flow chart of a method for generating a document abstract in an embodiment of the present application;
picture 33 是本申请一实施例中文献摘要生成方法的另一流程图;is another flow chart of the method for generating a document abstract in an embodiment of the present application;
picture 44 是本申请一实施例中文献摘要生成方法的另一流程图;is another flow chart of the method for generating a document abstract in an embodiment of the present application;
picture 55 是本申请一实施例中文献摘要生成方法的另一流程图;is another flow chart of the method for generating a document abstract in an embodiment of the present application;
picture 66 是本申请一实施例中文献摘要生成方法的另一流程图;is another flow chart of the method for generating a document abstract in an embodiment of the present application;
picture 77 是本申请一实施例中文献摘要生成方法的另一流程图;is another flow chart of the method for generating a document abstract in an embodiment of the present application;
picture 88 是本申请一实施例中文献摘要生成装置的一原理框图;is a schematic block diagram of a document abstract generating device in an embodiment of the present application;
picture 99 是本申请一实施例中拓扑图;is a topology diagram in an embodiment of the present application;
picture 1010 是本申请一实施例中计算机设备的一示意图。It is a schematic diagram of a computer device in an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative work fall within the protection scope of the present application.
本申请实施例提供的文献摘要生成方法,该文献摘要生成方法可应用如图1所示的应用环境中。具体地,该文献摘要生成方法应用在摘要生成系统中,该摘要生成系统包括如图1所示的客户端和服务器,客户端与服务器通过网络进行通信,用于通过目标模型组合确定任意两个目标句子之间的前后顺序,提高生成目标摘要的准确率,保证生成的目标摘要连贯性较佳。其中,客户端又称为用户端,是指与服务器相对应,为客户提供本地服务的程序。客户端可安装在但不限于各种个人计算机、笔记本电脑、智能手机、平板电脑和便携式可穿戴设备上。服务器可以用独立的服务器或者是多个服务器组成的服务器集群来实现。The document abstract generation method provided in the embodiment of the present application can be applied in the application environment shown in FIG. 1 . Specifically, the literature abstract generation method is applied in an abstract generation system, the abstract generation system includes a client and a server as shown in FIG. 1 , the client and the server communicate through the network, and are used to determine any two through the target model combination. The sequence between the target sentences improves the accuracy of generating the target summary and ensures that the generated target summary has better coherence. Among them, the client, also known as the client, refers to the program corresponding to the server and providing local services for the client. Clients can be installed on, but not limited to, various personal computers, laptops, smartphones, tablets, and portable wearable devices. The server can be implemented as an independent server or a server cluster composed of multiple servers.
在一实施例中,如图2所示,提供一种文献摘要生成方法,以该方法应用在图1中的服务器为例进行说明,包括如下步骤:In one embodiment, as shown in FIG. 2, a method for generating a document abstract is provided, and the method is applied to the server in FIG. 1 as an example for description, including the following steps:
S201:获取摘要生成请求,摘要生成请求包括摘要关键词。S201: Obtain an abstract generation request, where the abstract generation request includes abstract keywords.
其中,摘要生成请求是用于生成目标摘要的请求。摘要关键词是生成用户所需的目标摘要的关键词,以便后续根据摘要关键词得到对应的原始文献,例如,摘要关键词可以为xx疾病、xx医疗或者xx金融产品等。其中,摘要又称概要和内容提要。摘要是以提供文献内容梗概为目的,简明、确切地记述文献重要内容的短文。The digest generation request is a request for generating a target digest. The abstract keyword is a keyword for generating a target abstract required by the user, so that the corresponding original document can be obtained according to the abstract keyword. For example, the abstract keyword may be xx disease, xx medical treatment, or xx financial product. Among them, abstract is also called summary and abstract. Abstracts are short texts that describe the important content of the literature concisely and precisely for the purpose of providing the outline of the content of the literature.
具体地,在客户端中显示用于生成目标摘要的原始显示界面,用户点击原始显示界面,发送摘要关键词填写指令给服务器,服务器获取该摘要关键词填写指令后,控制客户端进入摘要关键词生成界面,用户在该摘要关键词生成界面填写至少一个摘要关键词,可以理解地,为确保生成的目标摘要符合用户的要求,当用户所填写的摘要关键词的数量少于预设关键词数量时,则根据用户填写的摘要关键词为用户推荐更多的相似关键词,当用户填写完不小于预设关键词数量的摘要关键词则控制客户端显示确认按钮,用户点击确认按钮,形成摘要请求发送给服务器,服务器接收到摘要生成请求时,对摘要生成请求进行解析,即可得到摘要关键词,从而实现自动化生成目标摘要。Specifically, an original display interface for generating the target abstract is displayed in the client, the user clicks on the original display interface, and sends an instruction to fill in the abstract keywords to the server. After the server obtains the instruction to fill in the abstract keywords, it controls the client to enter the abstract keywords. A generation interface, where the user fills in at least one abstract keyword in the abstract keyword generation interface. Understandably, in order to ensure that the generated target abstract meets the requirements of the user, when the number of abstract keywords filled in by the user is less than the preset number of keywords When the user fills in the summary keywords, more similar keywords are recommended for the user. When the user fills in the summary keywords that are not less than the preset number of keywords, the client is controlled to display the confirmation button, and the user clicks the confirmation button to form an abstract. The request is sent to the server, and when the server receives the abstract generation request, it parses the abstract generation request to obtain the abstract keywords, thereby realizing the automatic generation of the target abstract.
S202:基于摘要关键词查询数据库,从数据库存储的初始文献中获取N个与摘要关键词对应的原始文献。S202: Query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database.
其中,数据库是用于存储初始文献的库。初始文献是预先存储在数据库中的文献,可以理解地,初始文献包括各种摘要类型的摘要,初始文献包括与摘要关键词对应的文献和不与摘要关键词对应的文献,例如,初始文献可以是医疗方向对应的文献、美食方向对应的文献或者金融方向对应的文献等。原始文献是指与摘要关键词对应的摘要。Among them, the database is the library used to store the original literature. The initial documents are documents pre-stored in the database. It is understood that the initial documents include abstracts of various types of abstracts, and the initial documents include documents corresponding to the abstract keywords and documents that do not correspond to the abstract keywords. For example, the initial documents can be It is the literature corresponding to the medical direction, the literature corresponding to the food direction, or the literature corresponding to the financial direction, etc. The original document refers to the abstract corresponding to the abstract keywords.
具体地,预先对数据库中的文献摘要进行分类,获取相同摘要类型的摘要文献,并使每一摘要文献携带有至少一个摘要类型,为后续根据摘要关键词得到对应的摘要类型提供技术支持。当服务器获取到摘要关键词时,则采用匹配算法根据摘要关键词查询数据库,以便从文献摘要中得到摘要关键词对应的原始文献,实现自动化确定下相同摘要类型的原始文献,确保后续的目标摘要的准确性,减少人工成本。其中,摘要类型是指文献摘要对应的类型。Specifically, the document abstracts in the database are classified in advance, and abstract documents of the same abstract type are obtained, and each abstract document carries at least one abstract type, so as to provide technical support for obtaining the corresponding abstract type according to the abstract keywords in the future. When the server obtains the abstract keywords, it uses a matching algorithm to query the database according to the abstract keywords, so as to obtain the original documents corresponding to the abstract keywords from the document abstracts, realize the automatic determination of the original documents of the same abstract type, and ensure subsequent target abstracts. accuracy and reduce labor costs. Among them, the abstract type refers to the type corresponding to the literature abstract.
S203:采用预先训练好的抽取式文献摘要模型对原始文献进行处理,得到M个目标句子。S203: Use a pre-trained extractive literature summary model to process the original literature to obtain M target sentences.
其中,抽取式文献摘要模型是指直接从N个原始文献中抽出所需的目标句子的模型,抽取式文献摘要模型能较好的保留原始文献的信息,可以有效地提高目标摘要的准确性,降低后续生成的目标摘要的语法和句法错误率,本实施例中,抽取式文献摘要模型为NeuSUM 模型,实现自动化将原始文献中分数较高的句子抽取出来作为目标句子,减少人力成本,该NeuSUM 模型使用句子受益作为打分方式,考虑到了句子之间的相互关系,保证得到的目标句子关联性较高,后续生成的目标摘要连贯性更强。目标句子是指用于形成目标摘要的句子。Among them, the extractive document abstract model refers to a model that directly extracts the required target sentences from N original documents. The extractive document abstract model can better retain the information of the original documents and can effectively improve the accuracy of the target abstract. Reduce the grammatical and syntactic error rates of the subsequently generated target abstracts. In this embodiment, the extractive document abstract model is the NeuSUM model, which automatically extracts sentences with higher scores in the original documents as target sentences, reducing labor costs. The model uses sentence benefit as a scoring method, taking into account the relationship between sentences, to ensure that the obtained target sentences are highly relevant, and the subsequently generated target summaries are more coherent. The target sentence refers to the sentence used to form the target summary.
具体地,将原始文献输入预先训练好的抽取式文献摘要模型,首先,对原始文件进行分割,使得原始文件拆分为多个摘要句子,在嵌入层将摘要句子转化句子向量,以转化为计算机可识别的格式;在目标编码层对句子向量进行编码,得到包含语义信息的目标编码向量,以保留摘要句子的更多信息; 在打分编码层根据句子受益对目标编码向量进行打分,获取每一摘要句子对应的分数,使用句子受益作为打分方式,即以 ROUGE 评价指标作为摘要句子打分的指标,以考虑摘要句子之间的相互关系,将分数较高的前M个摘要句子作为目标句子,使得目标句子之间具有较强的联系,实现快速得到目标句子。Specifically, the original document is input into the pre-trained extractive document abstract model. First, the original document is divided so that the original document is divided into multiple abstract sentences, and the abstract sentences are converted into sentence vectors in the embedding layer to convert them into computer Recognizable format; the sentence vector is encoded at the target coding layer to obtain the target coding vector containing semantic information to retain more information of the summary sentence; at the scoring coding layer, the target coding vector is scored according to the benefit of the sentence, and each target coding vector is obtained. The score corresponding to the abstract sentence, using the sentence benefit as the scoring method, that is, with The ROUGE evaluation index is used as the index for scoring the summary sentences to consider the relationship between the summary sentences, and the top M summary sentences with higher scores are used as the target sentences, so that the target sentences have a strong connection, and the target sentence can be quickly obtained. .
本实施例中,抽取式文献摘要模型的训练是采用反向传播算法不断的调整初始模型的权重,直至初始模型权重收敛,则得到抽取式文献摘要模型。In this embodiment, the training of the extractive document abstract model is to continuously adjust the weight of the initial model by using the back-propagation algorithm until the weight of the initial model converges, and then the extractive document abstract model is obtained.
S204:将M个目标句子输入训练好的目标模型组合,得到目标句子对应的M*(M-1)/2个有向无环图。S204: Input the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
其中 ,目标模型组合是用于对任意两个目标句子进行排序的模型。该目标组合模型包括BERT模型和MLP模型,本实施例中,采用BERT模型和MLP模型可以准确地确定目标句子的前后顺序,为后续生成目标摘要提供技术支持。其中,BERT模型用于任意两个目标句子进行处理,在BERT模型的编码层和解码层中使用了transformer结构,确保BERT模型输出的是具备语义信息的语义信息的语义信息向量,例如,任意两个目标句子以目标句子1;目标句子2的形式输入BERT模型,得到对应的具有语义信息的输出向量,将输出向量输入MLP模型,得到任意两个目标句子之间的前后顺序。在使用BERT模型之前,对BERT模型进行训练,训练过程如下:1、获取训练语料样本,该训练预料样本包括正样本句子对和负样本句子对,可以理解地,正样本句子对存在句子间的上下文关系,负样本句子对不存在句子间的上下文关系;2、采用[SEP]标签对句子对进行连接的连接句子,例如,句子1[SEP]句子2;在连接句子中采用[CLS]作为标签,句末采用[SEP]作为标签,以利用这些标签的形式对句子本身的位置、句子间的前后关系进行标记,能够帮助训练初始Bert模型时让初始Bert模型学习到这些特征;3、随机地掩盖连接句子,获取训练预料;4、将训练语料输入到初始Bert模型中进行训练,得到Bert模型。where , the target model combination is the model used to rank any two target sentences. The target combination model includes a BERT model and an MLP model. In this embodiment, the BERT model and the MLP model can be used to accurately determine the sequence of target sentences, and provide technical support for the subsequent generation of target abstracts. Among them, the BERT model is used to process any two target sentences, and the transformer structure is used in the encoding layer and decoding layer of the BERT model to ensure that the output of the BERT model is a semantic information vector with semantic information. For example, any two Each target sentence is input into the BERT model in the form of target sentence 1; target sentence 2, and the corresponding output vector with semantic information is obtained, and the output vector is input into the MLP model to obtain the sequence between any two target sentences. Before using the BERT model, train the BERT model. The training process is as follows: 1. Obtain a training corpus sample. The training expected sample includes positive sample sentence pairs and negative sample sentence pairs. Understandably, positive sample sentence pairs exist between sentences. Context relationship, there is no context relationship between sentences for negative sample sentence pairs; 2. Use [SEP] tags to connect sentence pairs to connected sentences, for example, sentence 1 [SEP] sentence 2; use [CLS] as the connected sentence in the connected sentence Label, use [SEP] as a label at the end of the sentence, and use these labels to mark the position of the sentence itself and the contextual relationship between sentences, which can help the initial Bert model learn these features when training the initial Bert model; 3. Randomly 4. Input the training corpus into the initial Bert model for training to obtain the Bert model.
MLP模型即多层感知机模型,用于对任意两个句子进行二分类处理,得到任意两个目标句子的前后顺序。本实施例采用在使用MLP模型之前,先对MLP模型进行训练,训练过程:获取训练样本和与训练样本对应的顺序标签,其中,训练样本为原始句子对;将训练样本输入初始模型中,获取预测的句子顺序结果;根据顺序标签和句子顺序结果,计算得到分类准确率,当分类准确率大于预设值,则得到MLP模型。The MLP model is a multi-layer perceptron model, which is used to perform binary classification processing on any two sentences to obtain the sequence of any two target sentences. In this embodiment, the MLP model is trained before the MLP model is used. The training process is as follows: obtaining training samples and sequence labels corresponding to the training samples, wherein the training samples are original sentence pairs; input the training samples into the initial model, and obtain Predicted sentence sequence results; according to the sequence labels and sentence sequence results, the classification accuracy is calculated, and when the classification accuracy is greater than the preset value, the MLP model is obtained.
有向无环图是一个无回路的有向图,假设有向无环图为S1→S2→S3→S4,其中,S1、S2、S3和S4为目标句子。A directed acyclic graph is a directed graph without loops. Suppose the directed acyclic graph is S1→S2→S3→S4, where S1, S2, S3 and S4 are the target sentences.
现有技术在给目标句子排序形成目标摘要的过程中,确定当前排序位置的句子需要先预测前一个位置的句子,这种方法在模型训练过程较为复杂,计算量大、训练时间长且准确率低,得到的目标摘要。本实施例,对M个目标句子进行两两组合,得到M*(M-1)/2个句子组合,将每一个句子组合输入到BERT模型中,得到语义信息向量,将语义信息向量输入MLP模型,得到任意两个目标句子的前后顺序,基于前后顺序形成有向无环图,本实施例通过确定任意两个目标句子之间的前后顺序,保证后续可以对目标句子进行排序过程简单,有效提高准确率,保证生成的目标摘要连贯性较佳。作为一示例,当M等于3,即目标句子为S1、S2和S3,对这个3个句子进行两两组合,得到3个句子组合,即S1和S3、S1和S2、S2和S3,将每个句子组合输入目标模型组合中得到任意两个目标句子的前后顺序,以保证后续可以对目标句子进行排序过程简单,有效提高准确率,保证生成的目标摘要连贯性较佳。In the prior art, in the process of sorting the target sentence to form the target summary, to determine the sentence in the current sorting position, it is necessary to predict the sentence in the previous position first. This method is relatively complex in the model training process, requires a large amount of calculation, takes a long training time, and has high accuracy. low to get the target summary. In this embodiment, M target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, each sentence combination is input into the BERT model to obtain a semantic information vector, and the semantic information vector is input into the MLP model to obtain the sequence of any two target sentences, and form a directed acyclic graph based on the sequence. In this embodiment, by determining the sequence between any two target sentences, it ensures that the subsequent sequence of the target sentences is simple and effective. Improve the accuracy and ensure that the generated target summaries are more coherent. As an example, when M is equal to 3, that is, the target sentences are S1, S2 and S3, the three sentences are combined in pairs to obtain three sentence combinations, namely S1 and S3, S1 and S2, S2 and S3, and each A sentence combination is input into the target model combination to obtain the sequence of any two target sentences, so as to ensure that the subsequent sorting process of the target sentences is simple, the accuracy is effectively improved, and the consistency of the generated target summaries is ensured.
S205:基于M*(M-1)/2个有向无环图,获取目标摘要。S205: Based on M*(M-1)/2 directed acyclic graphs, obtain a target summary.
其中,目标摘要是指用户需要的摘要。Among them, the target summary refers to the summary required by the user.
本实施例中,对M*(M-1)/2个有向无环图进行处理,得到拓扑图,采用广度优先算法对拓扑图进行处理,得到每一目标句子的当前入度,根据当前入度对目标句子进行排序,得到目标摘要,过程较为简单,可以快速得到连贯的目标摘要。作为一示例,假设目标句子为S1、S2和S3,则向无环图为S1→S2;S1→S3;S2→S3;对向无环图进行处理得到如图9所示的拓扑图,图9中S1的当前入度为0;S2的当前入度为1;S3的当前入度为2;将当前入度为0的目标句子,即S1压入栈队列作为栈底元素,将目标句子S1指向的其他的目标句子的入度减去1,则S2的当前入度为1-1=0;S3的当前入度为2-1=1;那么将目标句子S2压入栈队列作为栈底元素,重复这个过程直到所有目标句子都压入栈队列中,此过程输出的栈队列即为目标摘要,过程较为简单,可以快速得到连贯的目标摘要。其中,入度,源于图论算法,通常指有向图中某点作为图中边的终点的次数之和。当前入度是指每一目标句子对应的入度。In this embodiment, M*(M-1)/2 directed acyclic graphs are processed to obtain a topology map, and the breadth-first algorithm is used to process the topology map to obtain the current in-degree of each target sentence. The in-degree sorts the target sentences and obtains the target summary. The process is relatively simple, and a coherent target summary can be obtained quickly. As an example, assuming that the target sentences are S1, S2, and S3, the directed acyclic graph is S1→S2; S1→S3; S2→S3; the topological graph shown in Figure 9 is obtained by processing the directed acyclic graph. In 9, the current in-degree of S1 is 0; the current in-degree of S2 is 1; the current in-degree of S3 is 2; the target sentence whose current in-degree is 0, that is, S1, is pushed into the stack queue as the stack bottom element, and the target sentence is The in-degree of other target sentences pointed to by S1 is subtracted by 1, then the current in-degree of S2 is 1-1=0; the current in-degree of S3 is 2-1=1; then the target sentence S2 is pushed into the stack queue as the stack The bottom element, repeat this process until all target sentences are pushed into the stack queue, the stack queue output by this process is the target summary, the process is relatively simple, and a coherent target summary can be obtained quickly. Among them, the in-degree, derived from the graph theory algorithm, usually refers to the sum of the times that a certain point in a directed graph is used as the end point of an edge in the graph. The current in-degree refers to the in-degree corresponding to each target sentence.
本实施例所提供的文献摘要生成方法,基于摘要关键词查询数据库,从数据库存储的初始文献中获取N个与摘要关键词对应的原始文献,实现自动化确定相同摘要类型的原始文献,确保后续的目标摘要的准确性,减少人工成本。采用预先训练好的抽取式文献摘要模型对原始文献进行处理,快速得到M个目标句子,使得目标句子之间具有较强的联系,确保后续生成的目标摘要记载原始文献的重要信息。将M个目标句子输入训练好的目标模型组合,得到目标句子对应的M*(M-1)/2个有向无环图,通过确定任意两个目标句子之间的前后顺序,保证后续可以对目标句子进行排序过程简单,有效提高准确率,保证生成的目标摘要连贯性较佳。基于M*(M-1)/2个有向无环图,可以快速得到连贯的目标摘要。The method for generating a document abstract provided in this embodiment queries a database based on the abstract keywords, obtains N original documents corresponding to the abstract keywords from the initial documents stored in the database, realizes automatic determination of the original documents of the same abstract type, and ensures subsequent Accuracy of target summaries, reducing labor costs. The pre-trained extractive literature abstract model is used to process the original literature, and quickly obtain M target sentences, which makes the target sentences have a strong connection, and ensures that the subsequently generated target abstracts record the important information of the original literature. Input the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences. By determining the sequence between any two target sentences, it is guaranteed that the subsequent The sorting process of the target sentences is simple, the accuracy is effectively improved, and the consistency of the generated target summaries is ensured. Based on M*(M-1)/2 directed acyclic graphs, coherent target summaries can be quickly obtained.
在一实施例中,如图3所示,步骤S202,即基于摘要关键词查询数据库,从数据库存储的初始文献中获取N个与摘要关键词对应的原始文献,包括:In one embodiment, as shown in FIG. 3 , step S202 is to query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database, including:
S301:基于摘要关键词查询数据库中的分类表,获取摘要关键词对应的摘要类型。S301: Query the classification table in the database based on the abstract keyword, and obtain the abstract type corresponding to the abstract keyword.
其中,分类表是预先设置的表格,该分类表用于指示预设关键词与摘要类型之间的关联关系。其中,预设关键词是与摘要关键词相对应的词。摘要类型是指摘要的类型,例如,摘要类型可以是医疗类型、金融类型和机械类型等。作为一示例,预设关键词是xx疾病,则对应的摘要类型为医疗类型。The classification table is a preset table, and the classification table is used to indicate the association relationship between preset keywords and abstract types. The preset keywords are words corresponding to the abstract keywords. The summary type refers to the type of summary, for example, the summary type can be medical type, financial type, mechanical type, etc. As an example, if the preset keyword is xx disease, the corresponding abstract type is medical type.
本实施例中,采用匹配算法对摘要关键词与分类表中的预设关键词进行匹配,若匹配成功,则说明存在与摘要关键词相对应的预设关键词,因此,根据对应的预设关键词即可得到摘要关键词对应的摘要类型,为后续确定原始文献提供技术支持。In this embodiment, a matching algorithm is used to match the abstract keywords with the preset keywords in the classification table. If the matching is successful, it means that there are preset keywords corresponding to the abstract keywords. Therefore, according to the corresponding preset keywords The keyword can be used to obtain the abstract type corresponding to the abstract keyword, which provides technical support for the subsequent determination of the original literature.
S302:基于摘要类型查询数据库中的初始文献,将包含摘要关键词的N个初始文献,确定为N个原始文献。S302: Query the initial documents in the database based on the abstract type, and determine the N initial documents including the abstract keywords as the N initial documents.
本实施例中,预先对数据库中的初始文献进行分类,当确定摘要类型后,将摘要关键词与该摘要类型中的初始文献进行匹配,以得到包含摘要关键词对应的初始文献,速度较快。In this embodiment, the initial documents in the database are classified in advance, and after the abstract type is determined, the abstract keywords are matched with the initial documents in the abstract type to obtain the initial documents corresponding to the abstract keywords, which is faster .
本实施例所提供的文献摘要生成方法,基于摘要关键词查询数据库中的分类表,获取摘要关键词对应的摘要类型,为后续确定原始文献提供技术支持。基于摘要类型查询数据库中的初始文献,将包含摘要关键词的N个初始文献,确定为N个原始文献,速度较快。The method for generating a document abstract provided in this embodiment queries a classification table in a database based on the abstract keyword, obtains the abstract type corresponding to the abstract keyword, and provides technical support for subsequent determination of the original document. The initial documents in the database are queried based on the abstract type, and the N initial documents containing the abstract keywords are determined as the N initial documents, which is faster.
在一实施例中,如图4所示,步骤S203,即采用预先训练好的抽取式文献摘要模型对原始文献进行处理,得到M个目标句子,包括:In one embodiment, as shown in FIG. 4 , in step S203, a pre-trained extractive document abstract model is used to process the original document to obtain M target sentences, including:
S401:对原始文献进行分割处理,获取至少两个摘要句子。S401: Segment the original document to obtain at least two abstract sentences.
其中,分割处理是指将原始文献分割为多个句子的处理,以便计算机对摘要句子进行处理。摘要句子是对原始文献进行分割处理得到的单个句子。Among them, the segmentation process refers to the process of dividing the original document into multiple sentences, so that the computer can process the abstract sentences. The abstract sentence is a single sentence obtained by segmenting the original document.
作为一示例,根据原始文献中的逗号和句号进行分割,例如,原始文献为xxxx,yyyyy;zzz,则通过查找逗号和句号将原始文献划分为xxxx、yyyyy和zzz作为3个句子。As an example, the division is performed according to commas and periods in the original document. For example, if the original document is xxxx, yyyyy; zzz, the original document is divided into xxxx, yyyyy and zzz as three sentences by searching for commas and periods.
S402:将所有摘要句子输入抽取式文献摘要模型的词嵌入层,获取每一摘要句子对应的句子向量。S402: Input all abstract sentences into the word embedding layer of the extractive document abstract model, and obtain sentence vectors corresponding to each abstract sentence.
其中,句子向量是指摘要句子经过词嵌入层处理后得到的向量,经过词嵌入层即可将摘要句子转化为对应的向量,便于计算机识别。词嵌入层是用于将摘要句子转化为计算机可识别的句子向量的层。Among them, the sentence vector refers to the vector obtained after the abstract sentence is processed by the word embedding layer. After the word embedding layer, the abstract sentence can be converted into a corresponding vector, which is convenient for computer recognition. The word embedding layer is the layer used to convert summary sentences into computer-recognizable sentence vectors.
S403:将每一句子向量输入抽取式文献摘要模型的目标编码层,得到每一句子向量对应的目标编码向量。S403: Input each sentence vector into the target coding layer of the extractive document abstract model to obtain a target coding vector corresponding to each sentence vector.
其中,目标编码层是用于对句子向量进行句子级别和文档级别的编码。具体地,首先将句子向量输入句子编码层,得到句子的向量表示的原始编码向量,将原始编码向量输入文档编码层,得到目标编码向量。Among them, the target encoding layer is used for sentence-level and document-level encoding of sentence vectors. Specifically, the sentence vector is firstly input into the sentence encoding layer to obtain the original encoding vector represented by the vector of the sentence, and the original encoding vector is input into the document encoding layer to obtain the target encoding vector.
S404:将目标编码向量输入抽取式文献摘要模型的打分编码层,获取每一摘要句子对应的打分结果。S404: Input the target coding vector into the scoring coding layer of the extractive document summary model, and obtain the scoring result corresponding to each summary sentence.
其中,打分结果是指采用打分编码层对每一摘要句子对应的目标编码向量进行打分形成的结果,可以理解地,将分数较高的摘要句子确定为目标句子,使得目标句子为包含有重要信息的句子,保证后续生成的目标摘要记载原始文献的重要句子。The scoring result refers to the result of scoring the target coding vector corresponding to each summary sentence using the scoring coding layer. Understandably, the summary sentence with a higher score is determined as the target sentence, so that the target sentence contains important information. sentences, to ensure that the subsequently generated target abstracts record the important sentences of the original literature.
S405:将多个目标编码向量的打分结果,按照数值从高到低依次选取前M个摘要句子,确定为M个目标句子。S405: According to the scoring results of the multiple target coding vectors, the first M summary sentences are selected in sequence from high to low, and are determined as M target sentences.
本实施例中,利用抽取式文献摘要模型将句子打分以及句子选择联合在一起,以将句子的信息关联起来,保证目标句子具有重要信息。In this embodiment, sentence scoring and sentence selection are combined using an extractive document summary model, so as to associate the information of the sentences and ensure that the target sentence has important information.
本实施例所提供的文献摘要生成方法,对原始文献进行分割处理,获取至少两个摘要句子,以便计算机对摘要句子进行处理。将所有摘要句子输入抽取式文献摘要模型的词嵌入层,获取每一摘要句子对应的句子向量,便于计算机识别。将每一句子向量输入抽取式文献摘要模型的目标编码层,得到每一句子向量对应的目标编码向量;将目标编码向量输入抽取式文献摘要模型的打分编码层,获取每一摘要句子对应的打分结果;将多个目标编码向量的打分结果,按照数值从高到低依次选取前M个摘要句子,确定为M个目标句子,利用抽取式文献摘要模型将句子打分以及句子选择联合在一起,以将句子的信息关联起来,保证目标句子具有重要信息。In the method for generating a document abstract provided in this embodiment, the original document is segmented to obtain at least two abstract sentences, so that the computer can process the abstract sentences. Input all abstract sentences into the word embedding layer of the extractive document abstract model, and obtain the sentence vector corresponding to each abstract sentence, which is convenient for computer identification. Input each sentence vector into the target coding layer of the extractive document summarization model to obtain the target coding vector corresponding to each sentence vector; input the target coding vector into the scoring coding layer of the extractive document summarization model to obtain the corresponding score of each abstract sentence Results: According to the scoring results of multiple target encoding vectors, the first M summary sentences are selected in order from high to low, and determined as M target sentences. The extractive literature abstract model is used to combine sentence scoring and sentence selection. Correlate the information of the sentences to ensure that the target sentence has important information.
在一实施例中,如图5,步骤S403,即将每一句子向量输入抽取式文献摘要模型的目标编码层,得到每一句子向量对应的目标编码向量,包括:In one embodiment, as shown in FIG. 5 , step S403 is to input each sentence vector into the target coding layer of the extractive document abstract model to obtain the target coding vector corresponding to each sentence vector, including:
S501:将每一句子向量输入抽取式文献摘要模型的句子编码层进行编码,获取句子向量对应的原始编码向量;S501: Input each sentence vector into the sentence coding layer of the extractive document abstract model for coding, and obtain the original coding vector corresponding to the sentence vector;
S502:将原始编码向量输入抽取式文献摘要模型的文档编码层再编码,获取目标编码向量。S502: Input the original encoding vector into the document encoding layer of the extractive document abstract model for re-encoding to obtain a target encoding vector.
其中,句子编码层是双向GRU句子编码层,利用该双向GRU句子编码层获得句子级别的编码。文档编码层是指双向GRU文档编码层,利用该双向GRU文档编码层获得文档级别的编码。Among them, the sentence encoding layer is a bidirectional GRU sentence encoding layer, and the sentence-level encoding is obtained by using the bidirectional GRU sentence encoding layer. The document encoding layer refers to the bidirectional GRU document encoding layer, and the document-level encoding is obtained by using the bidirectional GRU document encoding layer.
在一实施例中,目标模型组合包括预先训练好的BERT模型和MLP模型;如图6所示,步骤S204,即将M个目标句子输入训练好的目标模型组合,得到目标句子对应的M*(M-1)/2个有向无环图,包括:In one embodiment, the target model combination includes a pre-trained BERT model and an MLP model; as shown in Figure 6, step S204 is to input M target sentences into the trained target model combination to obtain M*( M-1)/2 directed acyclic graphs, including:
S601:对目标句子进行两两组合,得到M*(M-1)/2个句子组合;S601: Combining target sentences in pairs to obtain M*(M-1)/2 sentence combinations;
其中,句子组合是指任意两个目标句子形成的组合,以便后续可以得到两个目标句子之前的前后关系。Among them, the sentence combination refers to the combination formed by any two target sentences, so that the context before the two target sentences can be obtained subsequently.
本实施例中,对目标句子进行两两组合,得到M*(M-1)/2个句子组合,有利于简化后续确定任意两个目标句子之间的前后关系的步骤并保证确定任意两个目标句子之间的前后关系的准确性,确保得到连贯的目标摘要。作为一示例,当目标句子为3时,假设目标句子为S1、S2和S3,对目标句子进行两两组合,则可以得到句子组合为S1和S2、S1和S3、S2和S3。In this embodiment, the target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, which is conducive to simplifying the subsequent steps of determining the context between any two target sentences and ensuring that any two The accuracy of contextual relationships between target sentences ensures a coherent target summary. As an example, when the target sentence is 3, assuming that the target sentences are S1, S2 and S3, and the target sentences are combined in pairs, the sentence combinations can be obtained as S1 and S2, S1 and S3, S2 and S3.
S602:将每一句子组合输入BERT模型,获取每一句子组合对应的语义信息向量。S602: Input each sentence combination into the BERT model, and obtain a semantic information vector corresponding to each sentence combination.
本实施例中,BERT的作用在于获得句子组合的向量表示。BERT主要包括词嵌入层,编码层和解码层。词嵌入层的作用是将文献映射到向量,输入是文献,输出是向量。编码层和解码层使用的都是transformer结构,以获得具有语义信息的语义信息向量。In this embodiment, the role of BERT is to obtain a vector representation of sentence combinations. BERT mainly includes word embedding layer, encoding layer and decoding layer. The role of the word embedding layer is to map documents to vectors, the input is a document, and the output is a vector. Both the encoding layer and the decoding layer use the transformer structure to obtain semantic information vectors with semantic information.
S603:将语义信息向量输入MLP模型,获取任意两个目标句子的有向无环图。S603: Input the semantic information vector into the MLP model to obtain a directed acyclic graph of any two target sentences.
本实施例中,采用Bert模型加MLP模型对摘要句子进行抽取和分类,以得到目标句子,并确定目标句子之间的前后依赖关系,以解决现有技术中,仅采用Bert模型进行分类准确率低的问题。In this embodiment, the Bert model and the MLP model are used to extract and classify the abstract sentences, so as to obtain the target sentences, and determine the contextual dependencies between the target sentences, so as to solve the problem that in the prior art, only the Bert model is used for classification accuracy. low problem.
本实施例所提供的文献摘要生成方法,对目标句子进行两两组合,得到M*(M-1)/2个句子组合,有利于简化后续确定任意两个目标句子之间的前后关系的步骤并保证确定任意两个目标句子之间的前后关系的准确性,确保得到连贯的目标摘要。将每一句子组合输入BERT模型,获取每一句子组合对应的语义信息向量;将语义信息向量输入MLP模型,获取任意两个目标句子的有向无环图,以得到目标句子,并确定目标句子之间的前后依赖关系,以解决现有技术中,仅采用Bert模型进行分类准确率低的问题。In the method for generating document abstracts provided in this embodiment, target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations, which is beneficial to simplify the subsequent steps of determining the contextual relationship between any two target sentences And ensure the accuracy of determining the contextual relationship between any two target sentences, and ensure a coherent target summary. Input each sentence combination into the BERT model to obtain the semantic information vector corresponding to each sentence combination; input the semantic information vector into the MLP model to obtain the directed acyclic graph of any two target sentences to obtain the target sentence and determine the target sentence In order to solve the problem of low classification accuracy using only Bert model in the prior art.
在一实施例中,如图7所示,步骤S205,即基于M*(M-1)/2个有向无环图,获取目标摘要,包括:In one embodiment, as shown in FIG. 7 , step S205, that is, based on M*(M-1)/2 directed acyclic graphs, obtain the target digest, including:
S701:对M*(M-1)/2个有向无环图进行处理,得到拓扑图。S701: Process M*(M-1)/2 directed acyclic graphs to obtain a topology graph.
其中,拓扑图是指将所有的有向无环图集合形成的图,以便后续进行广度优先遍历,得到每个目标句子的当前入度。Among them, the topological graph refers to the graph formed by the collection of all directed acyclic graphs, so that the subsequent breadth-first traversal can be performed to obtain the current in-degree of each target sentence.
S702:采用广度优先算法遍历拓扑图,获取每一目标句子的当前入度。S702: Use the breadth-first algorithm to traverse the topology map to obtain the current in-degree of each target sentence.
其中,广度优先算法又称广度优先搜索、宽度优先搜索和横向优先搜索,,是一种图形搜索演算法;所谓广度,就是一层一层的,向下遍历。Among them, the breadth-first algorithm, also known as breadth-first search, breadth-first search and horizontal-first search, is a graph search algorithm; the so-called breadth is a layer-by-layer traversal.
本实施中,采用广度优先算法对拓扑图进行处理,得到每一目标句子的当前入度,根据当前入度对目标句子进行排序,得到目标摘要,过程较为简单,可以快速得到连贯的目标摘要。假设,目标句子S1在目标句子S2之前,目标句子S2在目标句子S3之前,目标句子S3在目标句子S4之前;则S1分别指向S2、S3和S4,S2分别指向S3和S4,S3指向S4,因此,S1的当前入度为0;S2的当前入度为1;S3的当前入度为2;S4的当前入度为3。In this implementation, the breadth-first algorithm is used to process the topology map to obtain the current in-degree of each target sentence, and the target sentences are sorted according to the current in-degree to obtain the target abstract. The process is relatively simple, and a coherent target abstract can be obtained quickly. Suppose that the target sentence S1 is before the target sentence S2, the target sentence S2 is before the target sentence S3, and the target sentence S3 is before the target sentence S4; then S1 points to S2, S3 and S4 respectively, S2 points to S3 and S4 respectively, S3 points to S4, Therefore, the current in-degree of S1 is 0; the current in-degree of S2 is 1; the current in-degree of S3 is 2; the current in-degree of S4 is 3.
S703:依据当前入度对所有目标句子进行入栈,获取栈队列。S703: Push all target sentences into the stack according to the current in-degree to obtain a stack queue.
具体地,将第一个入度为0的目标句子压入栈队列作为栈底元素,将目标句子指向的其他的目标句子的入度减去1,则原来入度为1的目标句子变为入度变成0,那么将目标句子2压入栈队列作为栈底元素,重复这个过程直到所有目标句子都压入栈队列中,此过程形成的栈队列即为目标摘要,过程较为简单,可以快速得到连贯的目标摘要。Specifically, push the first target sentence with an in-degree of 0 into the stack queue as the bottom element of the stack, and subtract 1 from the in-degree of other target sentences pointed to by the target sentence, then the original target sentence with an in-degree of 1 becomes When the in-degree becomes 0, the target sentence 2 is pushed into the stack queue as the bottom element of the stack, and this process is repeated until all target sentences are pushed into the stack queue. The stack queue formed by this process is the target summary. The process is relatively simple and can be Get a coherent summary of your goals quickly.
假设目标句子S1在目标句子S2之前,目标句子S2在目标句子S3之前,目标句子S3在目标句子S4之前;则S1分别指向S2、S3和S4,S2分别指向S3和S4,S3指向S4,因此,S1的当前入度为0;S2的当前入度为1;S3的当前入度为2;S4的当前入度为3。则首先将S1压入栈队列作为栈底元素,S2的当前入度变为为0;S3的当前入度为1;S4的当前入度为2,将S2压入栈队列作为栈底元素,……,得到栈队列为S1→S2→S3→S4。Suppose the target sentence S1 is before the target sentence S2, the target sentence S2 is before the target sentence S3, and the target sentence S3 is before the target sentence S4; then S1 points to S2, S3 and S4 respectively, S2 points to S3 and S4 respectively, S3 points to S4, so , the current in-degree of S1 is 0; the current in-degree of S2 is 1; the current in-degree of S3 is 2; the current in-degree of S4 is 3. Then first push S1 into the stack queue as the stack bottom element, the current in-degree of S2 becomes 0; the current in-degree of S3 is 1; the current in-degree of S4 is 2, and S2 is pushed into the stack queue as the stack bottom element, ......, the stack queue is S1→S2→S3→S4.
S704:基于栈队列,获取目标摘要。S704: Obtain a target summary based on the stack queue.
本实施例,根据栈队列中每个目标句子的顺序得到目标摘要,确保生成的目标摘要连贯通顺。In this embodiment, the target abstract is obtained according to the order of each target sentence in the stack queue, so as to ensure that the generated target abstract is coherent.
本实施例所提供的文献摘要生成方法,对M*(M-1)/2个有向无环图进行处理,得到拓扑图,以便后续进行广度优先遍历,得到每个目标句子的当前入度。采用广度优先算法遍历拓扑图,获取每一目标句子的当前入度;依据当前入度对所有目标句子进行入栈,获取栈队列;基于栈队列,获取目标摘要,确保生成的目标摘要连贯通顺。The method for generating literature abstracts provided in this embodiment processes M*(M-1)/2 directed acyclic graphs to obtain a topology graph, so that breadth-first traversal can be performed subsequently to obtain the current in-degree of each target sentence. . The breadth-first algorithm is used to traverse the topology map to obtain the current in-degree of each target sentence; according to the current in-degree, all target sentences are pushed into the stack to obtain the stack queue; based on the stack queue, the target abstract is obtained to ensure that the generated target abstract is coherent.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the size of the sequence numbers of the steps in the above embodiments does not mean the sequence of execution, and the execution sequence of each process should be determined by its function and internal logic, and should not constitute any limitation to the implementation process of the embodiments of the present application.
在一实施例中,提供一种文献摘要生成装置,该文献摘要生成装置与上述实施例中文献摘要生成方法一一对应。如图8所示,该文献摘要生成装置包括摘要生成请求获取模块801、原始文献获取模块802、目标句子获取模块803、有向无环图获取模块804和目标摘要获取模块805。各功能模块详细说明如下:In one embodiment, an apparatus for generating a document abstract is provided, and the apparatus for generating a document abstract corresponds one-to-one with the method for generating a document abstract in the above-mentioned embodiment. As shown in FIG. 8 , the document abstract generation device includes an abstract generation request acquisition module 801 , an original document acquisition module 802 , a target sentence acquisition module 803 , a directed acyclic graph acquisition module 804 and a target abstract acquisition module 805 . The detailed description of each functional module is as follows:
摘要生成请求获取模块801,用于获取摘要生成请求,摘要生成请求包括摘要关键词。The abstract generation request acquiring module 801 is configured to acquire an abstract generation request, where the abstract generation request includes abstract keywords.
原始文献获取模块802,用于基于摘要关键词查询数据库,从数据库存储的初始文献中获取N个与摘要关键词对应的原始文献。The original document obtaining module 802 is configured to query the database based on the abstract keywords, and obtain N original documents corresponding to the abstract keywords from the initial documents stored in the database.
目标句子获取模块803,用于采用预先训练好的抽取式文献摘要模型对原始文献进行处理,得到M个目标句子。The target sentence obtaining module 803 is used to process the original document by using the pre-trained extractive document abstract model to obtain M target sentences.
有向无环图获取模块804,用于将M个目标句子输入训练好的目标模型组合,得到目标句子对应的M*(M-1)/2个有向无环图。The directed acyclic graph obtaining module 804 is used for inputting the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences.
目标摘要获取模块805,用于基于M*(M-1)/2个有向无环图,获取目标摘要。The target abstract obtaining module 805 is configured to obtain the target abstract based on M*(M-1)/2 directed acyclic graphs.
优选地,原始文献获取模块802,包括:摘要类型获取单元和原始文献获取单元。Preferably, the original document acquisition module 802 includes: an abstract type acquisition unit and an original document acquisition unit.
摘要类型获取单元,用于基于摘要关键词查询数据库中的分类表,获取摘要关键词对应的摘要类型。The abstract type obtaining unit is used to query the classification table in the database based on the abstract keywords, and obtain the abstract types corresponding to the abstract keywords.
原始文献获取单元,用于根据摘要类型查询数据库中的文献摘要,将包含摘要关键词的N个文献摘要,确定为N个原始文献。The original document acquisition unit is used to query the document abstracts in the database according to the abstract type, and determine N document abstracts including abstract keywords as N original documents.
优选地,目标句子获取模块803,包括:分割处理单元、句子向量获取单元、目标编码向量获取单元和打分结果获取单元。Preferably, the target sentence obtaining module 803 includes: a segmentation processing unit, a sentence vector obtaining unit, a target coding vector obtaining unit, and a scoring result obtaining unit.
分割处理单元,用于对原始文献进行分割处理,获取至少两个摘要句子。The segmentation processing unit is used to segment the original document to obtain at least two abstract sentences.
句子向量获取单元,用于将所有摘要句子输入抽取式文献摘要模型的词嵌入层,获取每一摘要句子对应的句子向量。The sentence vector obtaining unit is used to input all the abstract sentences into the word embedding layer of the extractive document abstract model, and obtain the sentence vector corresponding to each abstract sentence.
目标编码向量获取单元,用于将每一句子向量输入抽取式文献摘要模型的目标编码层,得到每一句子向量对应的目标编码向量。The target coding vector obtaining unit is used to input each sentence vector into the target coding layer of the extractive document abstract model, and obtain the target coding vector corresponding to each sentence vector.
打分结果获取单元,用于将目标编码向量输入抽取式文献摘要模型的打分编码层,获取每一摘要句子对应的打分结果。The scoring result obtaining unit is used for inputting the target coding vector into the scoring coding layer of the extractive document summarization model to obtain the scoring result corresponding to each summary sentence.
目标句子获取单元,用于将多个目标编码向量的打分结果,按照数值从高到低依次选取前M个摘要句子,确定为M个目标句子。The target sentence obtaining unit is used to select the first M summary sentences in order from high to low according to the scoring results of multiple target encoding vectors, and determine them as M target sentences.
优选地,目标编码向量获取单元,包括:第一编码子单元和第二编码子单元。Preferably, the target coding vector obtaining unit includes: a first coding subunit and a second coding subunit.
第一编码子单元,用于将每一句子向量输入抽取式文献摘要模型的句子编码层进行编码,获取句子向量对应的原始编码向量。The first encoding subunit is used to input each sentence vector into the sentence encoding layer of the extractive document abstract model for encoding, and obtain the original encoding vector corresponding to the sentence vector.
第二编码子单元,用于将原始编码向量输入抽取式文献摘要模型的文档编码层再编码,获取目标编码向量。The second encoding subunit is used for inputting the original encoding vector into the document encoding layer of the extractive document abstract model for re-encoding to obtain the target encoding vector.
优选地,目标模型组合包括BERT模型和MLP模型。有向无环图获取模块804,包括:句子组合获取单元、语义信息向量获取单元和有向无环图获取单元。Preferably, the target model combination includes a BERT model and an MLP model. The directed acyclic graph obtaining module 804 includes: a sentence combination obtaining unit, a semantic information vector obtaining unit and a directed acyclic graph obtaining unit.
句子组合获取单元,用于对目标句子进行两两组合,得到M*(M-1)/2个句子组合。The sentence combination acquisition unit is used to combine the target sentences in pairs to obtain M*(M-1)/2 sentence combinations.
语义信息向量获取单元,用于将每一句子组合输入BERT模型,获取每一句子组合对应的语义信息向量。The semantic information vector obtaining unit is used to input each sentence combination into the BERT model, and obtain the semantic information vector corresponding to each sentence combination.
有向无环图获取单元,用于将语义信息向量输入MLP模型,获取任意两个目标句子的有向无环图。The directed acyclic graph acquisition unit is used to input the semantic information vector into the MLP model to obtain the directed acyclic graph of any two target sentences.
优选地,目标摘要获取模块805,包括:拓扑图获取单元、入度获取单元、栈队列获取单元和目标摘要获取单元。Preferably, the target digest obtaining module 805 includes: a topology map obtaining unit, an in-degree obtaining unit, a stack queue obtaining unit and a target digest obtaining unit.
拓扑图获取单元,用于对M*(M-1)/2个有向无环图进行处理,得到拓扑图。The topology map obtaining unit is used for processing M*(M-1)/2 directed acyclic graphs to obtain a topology map.
入度获取单元,用于采用广度优先算法遍历拓扑图,获取每一目标句子的当前入度。The in-degree obtaining unit is used to traverse the topology map using the breadth-first algorithm to obtain the current in-degree of each target sentence.
栈队列获取单元,用于依据当前入度对所有目标句子进行入栈,获取栈队列。The stack queue obtaining unit is used to push all target sentences into the stack according to the current in-degree, and obtain the stack queue.
目标摘要获取单元,用于基于栈队列,获取目标摘要。The target digest obtaining unit is used to obtain the target digest based on the stack queue.
关于文献摘要生成装置的具体限定可以参见上文中对于文献摘要生成方法的限定,在此不再赘述。上述文献摘要生成装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。For the specific limitation of the document abstract generating apparatus, please refer to the above limitation on the document abstract generating method, which will not be repeated here. Each module in the above-mentioned document abstract generating apparatus can be implemented in whole or in part by software, hardware and combinations thereof. The above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
在一个实施例中,提供了一种计算机设备,该计算机设备可以是服务器,其内部结构图可以如图10所示。该计算机设备包括通过系统总线连接的处理器、存储器、网络接口和数据库。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统、计算机可读指令和数据库。该内存储器为非易失性存储介质中的操作系统和计算机可读指令的运行提供环境。该计算机设备的数据库用于存储有向无环图。该计算机设备的网络接口用于与外部的终端通过网络连接通信。该计算机可读指令被处理器执行时以实现一种文献摘要生成方法。In one embodiment, a computer device is provided, and the computer device may be a server, and its internal structure diagram may be as shown in FIG. 10 . The computer device includes a processor, memory, a network interface, and a database connected by a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities. The memory of the computer device includes a non-volatile storage medium, an internal memory. The non-volatile storage medium stores an operating system, computer readable instructions and a database. The internal memory provides an environment for the execution of the operating system and computer-readable instructions in the non-volatile storage medium. The database of the computer equipment is used to store the directed acyclic graph. The network interface of the computer device is used to communicate with an external terminal through a network connection. The computer readable instructions, when executed by a processor, implement a method for generating a literature abstract.
在一个实施例中,提供了一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机可读指令,处理器执行计算机可读指令时实现上述实施例中文献摘要生成方法的步骤,例如图2所示的步骤S201-S205,或者图3至图7中所示的步骤,为避免重复,这里不再赘述。或者,处理器执行计算机可读指令时实现文献摘要生成装置这一实施例中的各模块/单元的功能,例如图8所示的摘要生成请求获取模块801、原始文献获取模块802、目标句子获取模块803、有向无环图获取模块804和目标摘要获取模块805的功能,为避免重复,这里不再赘述。本实施例所提供的可读存储介质包括非易失性可读存储介质和易失性可读存储介质。In one embodiment, a computer device is provided, including a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, and the processor implements the documents in the above embodiments when the processor executes the computer-readable instructions The steps of the abstract generating method, such as steps S201-S205 shown in FIG. 2, or steps shown in FIG. 3 to FIG. 7, are not repeated here in order to avoid repetition. Or, when the processor executes the computer-readable instructions, the functions of each module/unit in this embodiment of the document abstract generating apparatus are implemented, for example, the abstract generation request acquisition module 801, the original document acquisition module 802, the target sentence acquisition shown in FIG. 8 The functions of the module 803 , the directed acyclic graph obtaining module 804 and the target abstract obtaining module 805 are not repeated here in order to avoid repetition. The readable storage medium provided by this embodiment includes a non-volatile readable storage medium and a volatile readable storage medium.
在一实施例中,提供了一个或多个存储有计算机可读指令的可读存储介质,该可读存储介质上存储有计算机可读指令,该计算机可读指令被处理器执行时实现上述实施例中文献摘要生成方法的步骤,例如图2所示的步骤S201-S205,或者图3至图7中所示的步骤,为避免重复,这里不再赘述。或者,处理器执行计算机可读指令时实现文献摘要生成装置这一实施例中的各模块/单元的功能,例如图8所示的摘要生成请求获取模块801、原始文献获取模块802、目标句子获取模块803、有向无环图获取模块804和目标摘要获取模块805的功能,为避免重复,这里不再赘述。In one embodiment, one or more readable storage media storing computer-readable instructions are provided, and the computer-readable instructions are stored on the readable storage medium, and the computer-readable instructions are executed by the processor to implement the above-mentioned implementation. The steps of the method for generating a document abstract in the example, such as steps S201-S205 shown in FIG. 2 , or steps shown in FIG. 3 to FIG. 7 , are not repeated here to avoid repetition. Or, when the processor executes the computer-readable instructions, the functions of each module/unit in this embodiment of the document abstract generating apparatus are implemented, for example, the abstract generation request acquisition module 801, the original document acquisition module 802, the target sentence acquisition shown in FIG. 8 The functions of the module 803 , the directed acyclic graph obtaining module 804 and the target abstract obtaining module 805 are not repeated here in order to avoid repetition.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机可读指令来指令相关的硬件来完成,所述的计算机可读指令可存储于一非易失性计算机可读取存储介质中,该计算机可读指令在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(MRAM)、同步MRAM(SMRAM)、双数据率SMRAM(MMRSMRAM)、增强型SMRAM(ESMRAM)、同步链路(Synchlink) MRAM(SLMRAM)、存储器总线(Rambus)直接RAM(RMRAM)、直接存储器总线动态RAM(MRMRAM)、以及存储器总线动态RAM(RMRAM)等。Those of ordinary skill in the art can understand that all or part of the processes in the methods of the above embodiments can be implemented by instructing relevant hardware through computer-readable instructions, and the computer-readable instructions can be stored in a non-volatile computer. In the readable storage medium, the computer-readable instructions, when executed, may include the processes of the foregoing method embodiments. Wherein, any reference to memory, storage, database or other medium used in the various embodiments provided in this application may include non-volatile and/or volatile memory. Nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory may include random access memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in various forms such as static RAM (SRAM), dynamic RAM (MRAM), synchronous MRAM (SMRAM), double data rate SMRAM (MMRSMRAM), enhanced SMRAM (ESMRAM), synchronous chain Road (Synchlink) MRAM (SLMRAM), memory bus (Rambus) direct RAM (RMRAM), direct memory bus dynamic RAM (MRMRAM), and memory bus dynamic RAM (RMRAM), etc.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。Those skilled in the art can clearly understand that, for the convenience and simplicity of description, only the division of the above-mentioned functional units and modules is used as an example. Module completion, that is, dividing the internal structure of the device into different functional units or modules to complete all or part of the functions described above.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, but not to limit them; although the present application has been described in detail with reference to the above-mentioned embodiments, those of ordinary skill in the art should understand that: it can still be used for the above-mentioned implementations. The technical solutions described in the examples are modified, or some technical features thereof are equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions in the embodiments of the application, and should be included in the within the scope of protection of this application.

Claims (20)

  1. 一种文献摘要生成方法,其中,包括: A method for generating literature abstracts, including:
    获取摘要生成请求,所述摘要生成请求包括摘要关键词;Obtain an abstract generation request, where the abstract generation request includes abstract keywords;
    基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取N个与所述摘要关键词对应的原始文献;Querying a database based on the abstract keywords, obtaining N original documents corresponding to the abstract keywords from the initial documents stored in the database;
    采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到M个目标句子;The original document is processed by using a pre-trained extractive document abstract model to obtain M target sentences;
    将M个所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的M*(M-1)/2个有向无环图;The M described target sentences are input into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences;
    基于M*(M-1)/2个所述有向无环图,获取目标摘要。Based on M*(M-1)/2 of the directed acyclic graphs, a target summary is obtained.
  2. 如权利要求1所述的文献摘要生成方法,其中,所述基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取N个与所述摘要关键词对应的原始文献,包括: The method for generating document abstracts according to claim 1, wherein the querying a database based on the abstract keywords, and obtaining N original documents corresponding to the abstract keywords from the initial documents stored in the database, comprising:
    基于所述摘要关键词查询数据库中的分类表,获取所述摘要关键词对应的摘要类型;Query the classification table in the database based on the abstract keywords, and obtain the abstract types corresponding to the abstract keywords;
    根据所述摘要类型查询数据库中的文献摘要,将包含所述摘要关键词的N个文献摘要,确定为N个原始文献。The document abstracts in the database are searched according to the abstract type, and N document abstracts including the abstract keywords are determined as N original documents.
  3. 如权利要求1所述的文献摘要生成方法,其中,所述采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到M个目标句子,包括: The method for generating a document abstract as claimed in claim 1, wherein the pre-trained extractive document abstract model is used to process the original document to obtain M target sentences, including:
    对所述原始文献进行分割处理,获取至少两个摘要句子;Segmenting the original document to obtain at least two abstract sentences;
    将所有所述摘要句子输入抽取式文献摘要模型的词嵌入层,获取每一所述摘要句子对应的句子向量;Input all the abstract sentences into the word embedding layer of the extractive document abstract model, and obtain the sentence vector corresponding to each of the abstract sentences;
    将每一所述句子向量输入抽取式文献摘要模型的目标编码层,得到每一所述句子向量对应的目标编码向量;Inputting each of the sentence vectors into the target coding layer of the extractive document abstract model to obtain a target coding vector corresponding to each of the sentence vectors;
    将所述目标编码向量输入抽取式文献摘要模型的打分编码层,获取每一所述摘要句子对应的打分结果;Inputting the target coding vector into the scoring coding layer of the extractive document abstract model, to obtain the scoring result corresponding to each of the summary sentences;
    将多个所述目标编码向量的打分结果,按照数值从高到低依次选取前M个摘要句子,确定为M个目标句子。The first M summary sentences are selected in order from high to low from the scoring results of the plurality of target coding vectors, and are determined as M target sentences.
  4. 如权利要求3所述的文献摘要生成方法,其中,所述将每一所述句子向量输入所述抽取式文献摘要模型的目标编码层,得到每一所述句子向量对应的目标编码向量,包括: The method for generating document abstracts according to claim 3, wherein each of the sentence vectors is input into the target coding layer of the extractive document abstract model to obtain a target coding vector corresponding to each sentence vector, comprising: :
    将每一所述句子向量输入所述抽取式文献摘要模型的句子编码层进行编码,获取所述句子向量对应的原始编码向量;Inputting each of the sentence vectors into the sentence coding layer of the extractive document abstract model for coding, and obtaining the original coding vector corresponding to the sentence vector;
    将所述原始编码向量输入所述抽取式文献摘要模型的文档编码层再编码,获取所述目标编码向量。The original encoding vector is input into the document encoding layer of the extractive document abstract model for re-encoding to obtain the target encoding vector.
  5. 如权利要求1所述的文献摘要生成方法,其中,所述目标模型组合包括BERT模型和MLP模型;The method for generating document abstracts according to claim 1, wherein the target model combination comprises a BERT model and an MLP model;
    所述将M个所述目标句子输入训练好的目标模型组合,得到任意两个所述目标句子的M*(M-1)/2个有向无环图,包括:Described inputting M described target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs of any two described target sentences, including:
    对所述目标句子进行两两组合,得到M*(M-1)/2个句子组合;The target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations;
    将每一所述句子组合输入所述BERT模型,获取每一所述句子组合对应的语义信息向量;Inputting each combination of sentences into the BERT model to obtain the semantic information vector corresponding to each combination of sentences;
    将所述语义信息向量输入所述MLP模型,获取任意两个所述目标句子的有向无环图。Input the semantic information vector into the MLP model to obtain a directed acyclic graph of any two target sentences.
  6. 如权利要求1所述的文献摘要生成方法,其中,基于M*(M-1)/2个所述有向无环图,获取目标摘要,包括: The method for generating a document abstract according to claim 1, wherein, based on M*(M-1)/2 of the directed acyclic graphs, obtaining the target abstract, comprising:
    对M*(M-1)/2个所述有向无环图进行处理,得到拓扑图;Process the M*(M-1)/2 directed acyclic graphs to obtain a topology graph;
    采用广度优先算法遍历所述拓扑图,获取每一所述目标句子的当前入度;The breadth-first algorithm is used to traverse the topology map to obtain the current in-degree of each target sentence;
    依据所述当前入度对所有目标句子进行入栈,获取栈队列;All target sentences are pushed into the stack according to the current in-degree, and the stack queue is obtained;
    基于所述栈队列,获取目标摘要。Based on the stack queue, a target digest is obtained.
  7. 一种文献摘要生成装置,其中,包括:A device for generating literature abstracts, comprising:
    摘要生成请求获取模块,用于获取摘要生成请求,所述摘要生成请求包括摘要关键词;An abstract generation request acquiring module is used to acquire an abstract generation request, where the abstract generation request includes an abstract keyword;
    原始文献获取模块,用于基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取N个与所述摘要关键词对应的原始文献;An original document acquisition module, configured to query a database based on the abstract keywords, and acquire N original documents corresponding to the abstract keywords from the initial documents stored in the database;
    目标句子获取模块,用于采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到M个目标句子;The target sentence acquisition module is used to process the original document by using the pre-trained extractive document abstract model to obtain M target sentences;
    有向无环图获取模块,用于将M个所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的M*(M-1)/2个有向无环图;The directed acyclic graph acquisition module is used to input the M target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences;
    目标摘要获取模块,用于基于M*(M-1)/2个所述有向无环图,获取目标摘要。The target abstract obtaining module is used for obtaining target abstracts based on M*(M-1)/2 of the directed acyclic graphs.
  8. 如权利要求7所述的文献摘要生成装置,其中,所述原始文献获取模块,包括: The device for generating document abstracts according to claim 7, wherein the original document acquisition module comprises:
    摘要类型获取单元,用于基于所述摘要关键词查询数据库中的分类表,获取所述摘要关键词对应的摘要类型;An abstract type obtaining unit is configured to query the classification table in the database based on the abstract keywords, and obtain the abstract types corresponding to the abstract keywords;
    原始文献获取单元,用于根据所述摘要类型查询数据库中的文献摘要,将包含所述摘要关键词的N个文献摘要,确定为N个原始文献。The original document acquisition unit is configured to query the document abstracts in the database according to the abstract type, and determine N document abstracts including the abstract keywords as N original documents.
  9. 一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,其中,所述处理器执行所述计算机可读指令时实现如下步骤: A computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, wherein the processor implements the following steps when executing the computer-readable instructions:
    获取摘要生成请求,所述摘要生成请求包括摘要关键词;Obtain an abstract generation request, where the abstract generation request includes abstract keywords;
    基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取N个与所述摘要关键词对应的原始文献;Querying a database based on the abstract keywords, obtaining N original documents corresponding to the abstract keywords from the initial documents stored in the database;
    采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到M个目标句子;The original document is processed by using a pre-trained extractive document abstract model to obtain M target sentences;
    将M个所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的M*(M-1)/2个有向无环图;The M described target sentences are input into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences;
    基于M*(M-1)/2个所述有向无环图,获取目标摘要。Based on M*(M-1)/2 of the directed acyclic graphs, a target summary is obtained.
  10. 如权利要求9所述的计算机设备,所述基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取N个与所述摘要关键词对应的原始文献,包括: The computer device according to claim 9, wherein querying a database based on the abstract keywords, and acquiring N original documents corresponding to the abstract keywords from the initial documents stored in the database, comprising:
    基于所述摘要关键词查询数据库中的分类表,获取所述摘要关键词对应的摘要类型;Query the classification table in the database based on the abstract keywords, and obtain the abstract types corresponding to the abstract keywords;
    根据所述摘要类型查询数据库中的文献摘要,将包含所述摘要关键词的N个文献摘要,确定为N个原始文献。The document abstracts in the database are searched according to the abstract type, and N document abstracts including the abstract keywords are determined as N original documents.
  11. 如权利要求9所述的计算机设备,所述采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到M个目标句子,包括: The computer device according to claim 9, wherein the original document is processed by using a pre-trained extractive document abstract model to obtain M target sentences, including:
    对所述原始文献进行分割处理,获取至少两个摘要句子;Segmenting the original document to obtain at least two abstract sentences;
    将所有所述摘要句子输入抽取式文献摘要模型的词嵌入层,获取每一所述摘要句子对应的句子向量;Input all the abstract sentences into the word embedding layer of the extractive document abstract model, and obtain the sentence vector corresponding to each of the abstract sentences;
    将每一所述句子向量输入抽取式文献摘要模型的目标编码层,得到每一所述句子向量对应的目标编码向量;Inputting each of the sentence vectors into the target coding layer of the extractive document abstract model to obtain a target coding vector corresponding to each of the sentence vectors;
    将所述目标编码向量输入抽取式文献摘要模型的打分编码层,获取每一所述摘要句子对应的打分结果;Inputting the target coding vector into the scoring coding layer of the extractive document abstract model, to obtain the scoring result corresponding to each of the summary sentences;
    将多个所述目标编码向量的打分结果,按照数值从高到低依次选取前M个摘要句子,确定为M个目标句子。The first M summary sentences are selected in order from high to low from the scoring results of the plurality of target coding vectors, and are determined as M target sentences.
  12. 如权利要求11所述的计算机设备,所述将每一所述句子向量输入所述抽取式文献摘要模型的目标编码层,得到每一所述句子向量对应的目标编码向量,包括: The computer device according to claim 11, wherein each of the sentence vectors is input into the target coding layer of the extractive document abstraction model to obtain a target coding vector corresponding to each of the sentence vectors, comprising:
    将每一所述句子向量输入所述抽取式文献摘要模型的句子编码层进行编码,获取所述句子向量对应的原始编码向量;Inputting each of the sentence vectors into the sentence coding layer of the extractive document abstract model for coding, and obtaining the original coding vector corresponding to the sentence vector;
    将所述原始编码向量输入所述抽取式文献摘要模型的文档编码层再编码,获取所述目标编码向量。The original encoding vector is input into the document encoding layer of the extractive document abstract model for re-encoding to obtain the target encoding vector.
  13. 如权利要求9所述的计算机设备,所述目标模型组合包括BERT模型和MLP模型; The computer device of claim 9, the target model combination comprising a BERT model and an MLP model;
    所述将M个所述目标句子输入训练好的目标模型组合,得到任意两个所述目标句子的M*(M-1)/2个有向无环图,包括:Described inputting M described target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs of any two described target sentences, including:
    对所述目标句子进行两两组合,得到M*(M-1)/2个句子组合;The target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations;
    将每一所述句子组合输入所述BERT模型,获取每一所述句子组合对应的语义信息向量;Inputting each combination of sentences into the BERT model to obtain the semantic information vector corresponding to each combination of sentences;
    将所述语义信息向量输入所述MLP模型,获取任意两个所述目标句子的有向无环图。Input the semantic information vector into the MLP model to obtain a directed acyclic graph of any two target sentences.
  14. 如权利要求9所述的计算机设备,基于M*(M-1)/2个所述有向无环图,获取目标摘要,包括: The computer device according to claim 9, based on M*(M-1)/2 of the directed acyclic graphs, obtaining a target summary, comprising:
    对M*(M-1)/2个所述有向无环图进行处理,得到拓扑图;Process the M*(M-1)/2 directed acyclic graphs to obtain a topology graph;
    采用广度优先算法遍历所述拓扑图,获取每一所述目标句子的当前入度;The breadth-first algorithm is used to traverse the topology map to obtain the current in-degree of each target sentence;
    依据所述当前入度对所有目标句子进行入栈,获取栈队列;All target sentences are pushed into the stack according to the current in-degree, and the stack queue is obtained;
    基于所述栈队列,获取目标摘要。Based on the stack queue, a target digest is obtained.
  15. 一个或多个存储有计算机可读指令的可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤: One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
    获取摘要生成请求,所述摘要生成请求包括摘要关键词;Obtain an abstract generation request, where the abstract generation request includes abstract keywords;
    基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取N个与所述摘要关键词对应的原始文献;Querying a database based on the abstract keywords, obtaining N original documents corresponding to the abstract keywords from the initial documents stored in the database;
    采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到M个目标句子;The original document is processed by using a pre-trained extractive document abstract model to obtain M target sentences;
    将M个所述目标句子输入训练好的目标模型组合,得到所述目标句子对应的M*(M-1)/2个有向无环图;The M described target sentences are input into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs corresponding to the target sentences;
    基于M*(M-1)/2个所述有向无环图,获取目标摘要。Based on M*(M-1)/2 of the directed acyclic graphs, a target summary is obtained.
  16. 如权利要求15所述的可读存储介质,其中,所述基于所述摘要关键词查询数据库,从所述数据库存储的初始文献中获取N个与所述摘要关键词对应的原始文献,包括: The readable storage medium according to claim 15, wherein the querying a database based on the abstract keywords, and obtaining N original documents corresponding to the abstract keywords from the initial documents stored in the database, comprising:
    基于所述摘要关键词查询数据库中的分类表,获取所述摘要关键词对应的摘要类型;Query the classification table in the database based on the abstract keywords, and obtain the abstract types corresponding to the abstract keywords;
    根据所述摘要类型查询数据库中的文献摘要,将包含所述摘要关键词的N个文献摘要,确定为N个原始文献。The document abstracts in the database are searched according to the abstract type, and N document abstracts including the abstract keywords are determined as N original documents.
  17. 如权利要求15所述的可读存储介质,所述采用预先训练好的抽取式文献摘要模型对所述原始文献进行处理,得到M个目标句子,包括: The readable storage medium according to claim 15, wherein the original document is processed by using a pre-trained extractive document abstract model to obtain M target sentences, including:
    对所述原始文献进行分割处理,获取至少两个摘要句子;Segmenting the original document to obtain at least two abstract sentences;
    将所有所述摘要句子输入抽取式文献摘要模型的词嵌入层,获取每一所述摘要句子对应的句子向量;Input all the abstract sentences into the word embedding layer of the extractive document abstract model, and obtain the sentence vector corresponding to each of the abstract sentences;
    将每一所述句子向量输入抽取式文献摘要模型的目标编码层,得到每一所述句子向量对应的目标编码向量;Inputting each of the sentence vectors into the target coding layer of the extractive document abstract model to obtain a target coding vector corresponding to each of the sentence vectors;
    将所述目标编码向量输入抽取式文献摘要模型的打分编码层,获取每一所述摘要句子对应的打分结果;Inputting the target coding vector into the scoring coding layer of the extractive document abstract model, to obtain the scoring result corresponding to each of the summary sentences;
    将多个所述目标编码向量的打分结果,按照数值从高到低依次选取前M个摘要句子,确定为M个目标句子。The first M summary sentences are selected in order from high to low from the scoring results of the plurality of target coding vectors, and are determined as M target sentences.
  18. 如权利要求17所述的可读存储介质,其中,所述将每一所述句子向量输入所述抽取式文献摘要模型的目标编码层,得到每一所述句子向量对应的目标编码向量,包括: The readable storage medium according to claim 17, wherein the inputting each of the sentence vectors into a target coding layer of the extractive document summarization model to obtain a target coding vector corresponding to each of the sentence vectors, comprising: :
    将每一所述句子向量输入所述抽取式文献摘要模型的句子编码层进行编码,获取所述句子向量对应的原始编码向量;Inputting each of the sentence vectors into the sentence coding layer of the extractive document abstract model for coding, and obtaining the original coding vector corresponding to the sentence vector;
    将所述原始编码向量输入所述抽取式文献摘要模型的文档编码层再编码,获取所述目标编码向量。The original encoding vector is input into the document encoding layer of the extractive document abstract model for re-encoding to obtain the target encoding vector.
  19. 如权利要求15所述的可读存储介质,其中,所述目标模型组合包括BERT模型和MLP模型; The readable storage medium of claim 15, wherein the target model combination comprises a BERT model and an MLP model;
    所述将M个所述目标句子输入训练好的目标模型组合,得到任意两个所述目标句子的M*(M-1)/2个有向无环图,包括:Described inputting the M described target sentences into the trained target model combination to obtain M*(M-1)/2 directed acyclic graphs of any two described target sentences, including:
    对所述目标句子进行两两组合,得到M*(M-1)/2个句子组合;The target sentences are combined in pairs to obtain M*(M-1)/2 sentence combinations;
    将每一所述句子组合输入所述BERT模型,获取每一所述句子组合对应的语义信息向量;Input each of the sentence combinations into the BERT model, and obtain the semantic information vector corresponding to each of the sentence combinations;
    将所述语义信息向量输入所述MLP模型,获取任意两个所述目标句子的有向无环图。Input the semantic information vector into the MLP model to obtain a directed acyclic graph of any two target sentences.
  20. 如权利要求15所述的可读存储介质,其中,基于M*(M-1)/2个所述有向无环图,获取目标摘要,包括: The readable storage medium of claim 15, wherein, based on M*(M-1)/2 of the directed acyclic graphs, obtaining the target digest comprises:
    对M*(M-1)/2个所述有向无环图进行处理,得到拓扑图;Process the M*(M-1)/2 directed acyclic graphs to obtain a topology graph;
    采用广度优先算法遍历所述拓扑图,获取每一所述目标句子的当前入度;The breadth-first algorithm is used to traverse the topology map to obtain the current in-degree of each of the target sentences;
    依据所述当前入度对所有目标句子进行入栈,获取栈队列;All target sentences are pushed into the stack according to the current in-degree, and the stack queue is obtained;
    基于所述栈队列,获取目标摘要。Based on the stack queue, a target digest is obtained.
      
PCT/CN2021/084241 2020-12-30 2021-03-31 Document abstract generation method and apparatus, computer device, and storage medium WO2022141872A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011623844.3 2020-12-30
CN202011623844.3A CN112732898A (en) 2020-12-30 2020-12-30 Document abstract generation method and device, computer equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2022141872A1 true WO2022141872A1 (en) 2022-07-07

Family

ID=75609644

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/084241 WO2022141872A1 (en) 2020-12-30 2021-03-31 Document abstract generation method and apparatus, computer device, and storage medium

Country Status (2)

Country Link
CN (1) CN112732898A (en)
WO (1) WO2022141872A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115809329A (en) * 2023-01-30 2023-03-17 医智生命科技(天津)有限公司 Method for generating abstract of long text
CN116912047A (en) * 2023-09-13 2023-10-20 湘潭大学 Patent structure perception similarity detection method

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113407710A (en) * 2021-06-07 2021-09-17 维沃移动通信有限公司 Information display method and device, electronic equipment and readable storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100228693A1 (en) * 2009-03-06 2010-09-09 phiScape AG Method and system for generating a document representation
US20150339288A1 (en) * 2014-05-23 2015-11-26 Codeq Llc Systems and Methods for Generating Summaries of Documents
CN108280112A (en) * 2017-06-22 2018-07-13 腾讯科技(深圳)有限公司 Abstraction generating method, device and computer equipment
CN109657054A (en) * 2018-12-13 2019-04-19 北京百度网讯科技有限公司 Abstraction generating method, device, server and storage medium
CN111414471A (en) * 2020-03-20 2020-07-14 北京百度网讯科技有限公司 Method and apparatus for outputting information
CN111858913A (en) * 2020-07-08 2020-10-30 北京嘀嘀无限科技发展有限公司 Method and system for automatically generating text abstract

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100228693A1 (en) * 2009-03-06 2010-09-09 phiScape AG Method and system for generating a document representation
US20150339288A1 (en) * 2014-05-23 2015-11-26 Codeq Llc Systems and Methods for Generating Summaries of Documents
CN108280112A (en) * 2017-06-22 2018-07-13 腾讯科技(深圳)有限公司 Abstraction generating method, device and computer equipment
CN109657054A (en) * 2018-12-13 2019-04-19 北京百度网讯科技有限公司 Abstraction generating method, device, server and storage medium
CN111414471A (en) * 2020-03-20 2020-07-14 北京百度网讯科技有限公司 Method and apparatus for outputting information
CN111858913A (en) * 2020-07-08 2020-10-30 北京嘀嘀无限科技发展有限公司 Method and system for automatically generating text abstract

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115809329A (en) * 2023-01-30 2023-03-17 医智生命科技(天津)有限公司 Method for generating abstract of long text
CN116912047A (en) * 2023-09-13 2023-10-20 湘潭大学 Patent structure perception similarity detection method
CN116912047B (en) * 2023-09-13 2023-11-28 湘潭大学 Patent structure perception similarity detection method

Also Published As

Publication number Publication date
CN112732898A (en) 2021-04-30

Similar Documents

Publication Publication Date Title
US20200342177A1 (en) Capturing rich response relationships with small-data neural networks
US11645314B2 (en) Interactive information retrieval using knowledge graphs
US10832011B2 (en) Question answering system using multilingual information sources
WO2019136993A1 (en) Text similarity calculation method and device, computer apparatus, and storage medium
US10025819B2 (en) Generating a query statement based on unstructured input
US20200184307A1 (en) Utilizing recurrent neural networks to recognize and extract open intent from text inputs
US9318027B2 (en) Caching natural language questions and results in a question and answer system
WO2022141872A1 (en) Document abstract generation method and apparatus, computer device, and storage medium
WO2021114810A1 (en) Graph structure-based official document recommendation method, apparatus, computer device, and medium
CN111026319B (en) Intelligent text processing method and device, electronic equipment and storage medium
JP2021152963A (en) Word meaning feature generating method, model training method, apparatus, device, medium, and program
US11551437B2 (en) Collaborative information extraction
WO2020244065A1 (en) Character vector definition method, apparatus and device based on artificial intelligence, and storage medium
WO2022088671A1 (en) Automated question answering method and apparatus, device, and storage medium
CN111026320B (en) Multi-mode intelligent text processing method and device, electronic equipment and storage medium
US11263400B2 (en) Identifying entity attribute relations
US20220300543A1 (en) Method of retrieving query, electronic device and medium
US11861918B2 (en) Image analysis for problem resolution
WO2023045187A1 (en) Semantic search method and apparatus based on neural network, device, and storage medium
WO2022174496A1 (en) Data annotation method and apparatus based on generative model, and device and storage medium
WO2023231331A1 (en) Knowledge extraction method, system and device, and storage medium
WO2020106644A1 (en) Transliteration of data records for improved data matching
US20220114361A1 (en) Multi-word concept tagging for images using short text decoder
CN111142728B (en) Vehicle-mounted environment intelligent text processing method and device, electronic equipment and storage medium
CN112632258A (en) Text data processing method and device, computer equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21912642

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21912642

Country of ref document: EP

Kind code of ref document: A1