WO2020237479A1 - 实时事件摘要的生成方法、装置、设备及存储介质 - Google Patents
实时事件摘要的生成方法、装置、设备及存储介质 Download PDFInfo
- Publication number
- WO2020237479A1 WO2020237479A1 PCT/CN2019/088630 CN2019088630W WO2020237479A1 WO 2020237479 A1 WO2020237479 A1 WO 2020237479A1 CN 2019088630 W CN2019088630 W CN 2019088630W WO 2020237479 A1 WO2020237479 A1 WO 2020237479A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- text
- representation
- event
- knowledge
- user query
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
Definitions
- the invention belongs to the field of computer technology, and in particular relates to a method, device, equipment and storage medium for generating a real-time event summary.
- Event Summarization is a very challenging task in the field of Natural Language Processing (NLP).
- NLP Natural Language Processing
- the purpose of the task is to generate an informative text summary for a given text stream, and to dynamically change the process of events. Update text summaries in real-time, and provide text summaries of events that people are interested in.
- NLP Natural Language Processing
- the text summaries are generated after parsing the text using static summarization methods, and only simple updates are made to the text summaries.
- the static summary method can only generate one summary at a time, and cannot infer the evolution of the event over time and update the summary in real time when new information appears. It is not suitable for large-scale dynamic streaming media applications.
- Real-time Event Summarization aims to generate a series of text summaries from a large number of real-time text streams, which can accurately describe the events of interest to users.
- Real-time event summaries are generally used in news and social media scenarios, and have very broad application prospects.
- streaming media applications including Twitter
- Twitter can provide users with summary push services of current popular or user-interested tweets.
- this is also a very challenging task.
- news texts are usually written by professional reporters or writers, with complete sentences and grammatical structures, and the extracted abstracts are of good quality.
- social media texts are usually short, with many spelling errors and grammatical errors, as well as many popular words and sentences on the Internet, which greatly hinders the summarization of social media texts.
- the purpose of the present invention is to provide a method, device, equipment and storage medium for generating and controlling real-time event summary, which aims to solve the problem of insufficient real-time event summary information, high redundancy, and poor performance of real-time event summary in the prior art. problem.
- the present invention provides a method for generating a real-time event summary.
- the method includes the following steps:
- the interactive learning text representation of the event text and the user query text representation are generated.
- Interactive learning text representation
- the specific text representation of the event text is input into a trained multi-task joint training model to generate a real-time event summary of the text stream.
- the multi-task joint training model includes a real-time event summary task model and a correlation prediction task model.
- the present invention provides a device for generating a real-time event summary, the device comprising:
- a text receiving module for receiving a text stream and user query text, the text stream including event text sorted by time;
- a knowledge-aware representation generation module configured to generate a knowledge-aware text representation of the event text and a knowledge-aware text representation of the user query text according to the event text, the user query text, and a preset knowledge base;
- An interactive representation generating module for generating interactive learning text of the event text based on the knowledge-aware text representation of the event text, the knowledge-aware text representation of the user query text and the trained interactive multi-head attention network An interactive learning text representation of the representation and the user query text;
- a specific representation generating module for generating a specific text representation of the event text based on the interactive learning text representation of the event text, the interactive learning text representation of the user query text, and the trained dynamic memory network;
- the real-time summary generation module is used to input the specific text representation of the event text into a trained multi-task joint training model to generate a real-time event summary of the text stream.
- the multi-task joint training model includes a real-time event summary task model and Relevance prediction task model.
- the present invention also provides a computer device, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor.
- the processor implements the computer program when the computer program is executed. The steps described in the above-mentioned method for generating real-time event summary.
- the present invention also provides a computer-readable storage medium, the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the above-mentioned method for generating a real-time event summary is implemented. step.
- the present invention receives a text stream and a user query text.
- the text stream includes event text sorted by time, generates event text and a knowledge-aware text representation of the user query text according to the knowledge base, and based on these knowledge-aware text representations and interactive multi-head attention networks, Generate interactive learning text representations of event text and user query text, generate specific text representations of event text based on these interactive learning text representations and dynamic memory networks, and input specific text representations into the multi-task joint training model to obtain real-time event summaries.
- the content of the real-time event summary is effectively enriched, the text representation is better learned through the interactive learning and attention mechanism, and the redundancy of the real-time event summary is effectively reduced through the dynamic memory network.
- the training model realizes the joint processing of real-time event summary tasks and correlation prediction tasks, improves the performance of real-time event summaries, and effectively generates real-time event summaries.
- FIG. 1 is an implementation flowchart of a method for generating a real-time event summary provided by Embodiment 1 of the present invention
- Embodiment 2 is an implementation flowchart of a method for generating a real-time event summary provided by Embodiment 2 of the present invention
- FIG. 3 is a schematic structural diagram of an apparatus for generating a real-time event summary provided by Embodiment 3 of the present invention.
- FIG. 4 is a schematic diagram of a preferred structure of an apparatus for generating a real-time event summary provided in Embodiment 3 of the present invention.
- FIG. 5 is a schematic diagram of the structure of a computer device according to Embodiment 4 of the present invention.
- Fig. 1 shows the implementation process of the method for generating a real-time event summary provided in the first embodiment of the present invention.
- Fig. 1 shows the implementation process of the method for generating a real-time event summary provided in the first embodiment of the present invention.
- Fig. 1 shows the implementation process of the method for generating a real-time event summary provided in the first embodiment of the present invention.
- Only the parts related to the embodiment of the present invention are shown, which are detailed as follows:
- step S101 a text stream and a user query text are received, and the text stream includes event text sorted by time.
- Event texts (such as social media texts) can be collected in real time on the network, and the text streams collected under these different timestamps can form a text stream.
- the user query text is the keyword text entered by the user. Each event text and user query text includes multiple words.
- step S102 a knowledge-aware text representation of the event text and a knowledge-aware text representation of the user query text are generated according to the event text, the user query text and the preset knowledge base.
- a knowledge base contains a large amount of knowledge, such as a Microsoft knowledge base or some knowledge bases built on Wikipedia.
- the knowledge base is used to represent event text and user query text. Effectively improve the richness of real-time time summary.
- the knowledge-aware text representation of the event text includes the initial context representation and initial knowledge representation of the event text
- the knowledge-aware text representation of the user query text includes the initial context representation and the initial knowledge representation of the user query text.
- step S103 based on the knowledge-aware text representation of the event text, the knowledge-aware text representation of the user query text and the trained interactive multi-head attention network, the interactive learning text representation of the event text and the interactive learning user query text are generated Text representation.
- an interactive multi-head attention network is constructed and trained in advance, and the knowledge-aware text representation of the event text and the user query text is input into the trained interactive multi-head attention network to obtain the attention of each event text Matrix, based on the attention matrix of the event text and the knowledge-aware text representation of the event text, the interactive learning text representation of the event text is calculated. Similarly, by inputting the knowledge-aware text representation of the event text and the user query text into the interactive multi-head attention network, the attention matrix of the user query text is obtained. The interactive learning text representation of the user query text is calculated based on the attention matrix of the user query text and the knowledge-aware text representation of the user query text.
- the calculation process of the attention matrix of the event text has the participation of the user query text
- the calculation process of the attention matrix of the user query text has the participation of the event text.
- the interactive multi-head attention network realizes the event text and
- the interactive learning between the user query text can effectively capture the interactive information between the event text and the user query text, and provide the performance of the text representation of the event text and the user query text.
- step S104 a specific text representation of the event text is generated based on the interactive learning text representation of the event text, the interactive learning text representation of the user query text and the trained dynamic memory network.
- the dynamic memory network is used to memorize the past event text, and adjust the current attention according to the memory content to avoid a large amount of redundant content in the real-time event summary.
- the dynamic memory network also includes methods for updating memory content The cyclic network obtains the memory content of the event text under the current time stamp according to the memory content of the event text under the previous time stamp and the interactive learning text representation of the event text under the current time stamp.
- step S105 the specific text representation of the event text is input into the trained multi-task joint training model to generate a real-time event summary of the text stream.
- the multi-task joint training model includes a real-time event summary task model and a correlation prediction task model.
- the specific text representation of each event text in the text stream is input to the trained multi-task joint training model, and each event in the text stream is calculated by the correlation prediction task model in the multi-task joint training model.
- the specific text of the text represents the relevance label relative to the user's query text.
- the real-time event summary in the multi-task joint training model is used to determine the text action of each event text in the text stream, and generate the text stream based on the text action of each event text Real-time event summary.
- the predicted relevance tags include highly relevant, relevant, and irrelevant, and the text actions include push and no push.
- the specific text representation of the event text is pushed to the real-time event summary.
- the knowledge-aware text representation of the event text and the user query text is generated by the knowledge base, and the knowledge-aware text is interactively learned through the interactive multi-head attention network to generate the event text and the user query text.
- Interactive learning text representation These interactive learning text representations are processed through a dynamic memory network to generate a specific text representation of the event text.
- the specific text representation of the event text is input into the multi-task joint training model to generate a real-time event summary of the text stream. This effectively improves the content richness and performance of the real-time event summary, reduces the redundancy of the real-time event summary, and further improves the generation effect of the real-time event summary.
- Figure 2 shows the implementation process of the method for generating a real-time event summary provided in the second embodiment of the present invention.
- Figure 2 shows the implementation process of the method for generating a real-time event summary provided in the second embodiment of the present invention.
- Only the parts related to the embodiment of the present invention are shown, which are described in detail as follows:
- step S201 a text stream and a user query text are received, and the text stream includes event text sorted by time.
- the event text under a timestamp. Every text in the text stream By l words Composition (the time suffix of the text is omitted here to simplify the mathematical representation of these parameters).
- the user query text can be expressed as From n words composition.
- step S202 the initial context representation of the event text is obtained by extracting the hidden state of the word in the event text, and the initial context representation of the user query text is obtained by extracting the hidden state of the word in the user query text.
- each word in the event text and each word in the user query text are respectively mapped to a low-dimensional word embedding vector through a preset word embedding layer.
- the low-dimensional word embedding vector of each word in the event text is input into the first gated recurrent unit (GRU), and the hidden state of each word in the event text is calculated.
- the first gating recirculation unit and the second gating recirculation unit are mutually independent gating recirculation units.
- the calculation formula for calculating the hidden state of the word through the gated loop unit is:
- h k GRU(h k-1 ,v k ), where v k represents the low-dimensional word embedding vector of the k-th word, h k represents the hidden state of the k- th word, and h k-1 represents the k-1 The hidden state of a word.
- the hidden states of all words in the event text are combined into the initial context representation of the event text
- the hidden states of all words in the user query text are combined into the initial context representation of the user query text.
- the initial context of the event text is expressed as
- the initial context of the user query text is expressed as among them, Is the i-th word in the event text Hidden state, Is the jth word in the event text Hidden state.
- step S203 an initial knowledge representation of the event text is generated according to the initial context representation, attention mechanism and knowledge base of the event text, and initial knowledge of the query text is generated according to the initial context representation, attention mechanism and knowledge base of the user query text Said.
- a candidate entity set composed of a preset number of embedded entities is selected in the knowledge base for each word in the event text and each user query text, and the candidate entity set is expressed as:
- N is the total number of embedded entities
- the knowledge representation of each word in the event text is learned by embedding the corresponding candidate entity set in the knowledge base, and the learning process can be expressed as:
- e ki is the i-th embedded entity in the set of candidate entities of the k-th word in the event text
- a ki is the context-guided attention weight of e ki
- ⁇ is the average pooling operation
- W kb and W c are the trained weight matrices
- b kb is the bias value.
- the initial knowledge estimate of the event text is formed by the knowledge representation of all words in the event text
- the knowledge base and the initial context representation of the event text the initial knowledge representation of the event text can be obtained.
- the knowledge base and the initial context representation of the user query text the initial knowledge representation of the user query text can be obtained.
- step S204 the knowledge-aware text representation of the event text is obtained by combining the initial context representation of the event text and the initial knowledge representation of the event text
- the user query text is obtained by combining the initial context representation of the user query text and the initial knowledge representation of the user query text
- the knowledge-aware textual representation is obtained by combining the initial context representation of the user query text and the initial knowledge representation of the user query text.
- the knowledge-aware text of the event text is expressed as:
- the knowledge-aware text of the user query text is expressed as:
- step S205 the knowledge-aware text representation of the event text and the knowledge-aware text representation of the user query text are input into the interactive multi-head attention network, and the attention matrix of the event text and the attention matrix of the user query text are calculated.
- the knowledge-aware text representation of the event text and the knowledge-aware text representation of the user query text are input into the interactive multi-head attention network, and the attention matrix of the event text and the attention matrix of the user query text are calculated.
- the calculation formula of the attention matrix of the event text is expressed as:
- A [A 1 ,A 2 ,...,A l ]
- ⁇ is the average pooling operation
- a i is the i-th row matrix in the attention matrix A of the event text
- ⁇ is the attention function
- U (1) and W (1) are the weight matrices trained for the interactive multi-head attention network.
- the calculation formula of the attention matrix of the user query text is expressed as:
- B i is the i-th row matrix in the attention matrix B of the user query text.
- step S206 the interactive learning text representation of the event text is calculated according to the attention matrix of the event text and the knowledge-aware text representation, and the interaction of the user query text is calculated according to the attention matrix of the user query text and the knowledge-aware text representation Learning textual representation.
- step S207 a specific text representation of the event text is generated based on the interactive learning text representation of the event text, the interactive learning text representation of the user query text and the trained dynamic memory network.
- each time stamp in the text stream is equivalent to each step of the dynamic memory network.
- obtain the memory content of the previous time stamp of the current time stamp and input the memory content, the interactive learning text representation of the event text under the current time stamp, and the user query text into the dynamic memory network through dynamic
- the attention mechanism in the network calculates the specific text representation of the event text under the current timestamp.
- the calculation formula of the specific text representation of the event text is:
- emb t is the specific text representation of the event text at timestamp t
- o tj d is the interactive learning text representation of the j-th word in the event text at timestamp t
- the attention function w tj is a forward neural network
- ⁇ flattened matrix for the vector function of the form W a, U a, V a w tj is the weight matrix
- b a w tj is the bias term
- m t-1 is the time stamp t- 1.
- the memory content corresponding to the event text under the last time stamp and the specific text representation of the event text under the current time stamp is calculated, so as to generate the corresponding memory content of the event text in the order of the time stamp.
- the memory content is stored in the dynamic memory network.
- the memory content corresponding to the event text under the last time stamp and the specific text representation of the event text under the current time stamp is calculated by the third gating loop unit, and the calculation formula is :
- the memory content corresponding to the time text under the initial timestamp is the interactive learning text representation corresponding to the last word in the user query text which is
- step S208 the specific text representation of the event text is input into the trained multi-task joint training model to generate a real-time event summary of the text stream.
- the multi-task joint training model includes a real-time event summary task model and a correlation prediction task model.
- the objective function of the correlation prediction task model in the training process of the multi-task joint training model, can be expressed as:
- the weight matrix of the correlation prediction task in a supervised way with For learning, the training data set is d t and q t are the event text and user query text under the time stamp t in the training data set, Is the true correlation label of d t relative to q t . Training is performed by minimizing the objective function (ie, the cross entropy between the predicted correlation label and the true correlation label).
- the objective function of the real-time event summary task model can be expressed as:
- the expected reward is a typical delayed reward
- r( ⁇ ) is the reward function
- ⁇ is the coefficient of control function EG( ⁇ ) and function nCG( ⁇ )
- Is the strategy function here the independent function approximator with parameter ⁇ in the stochastic strategy gradient algorithm is used to approximate the stochastic strategy ⁇ ⁇
- b s is the offset value
- a t ⁇ ⁇ 0,1 ⁇ is a text operation
- s t emb t.
- a reinforcement learning algorithm is used to optimize the objective function of the real-time event summary task model.
- a strategy gradient algorithm is used as a reinforcement learning algorithm for optimizing the real-time event summary task model to improve the training effect of the real-time event summary task model.
- the multi-task joint training model can be expressed as:
- L ⁇ 1 L 1 + ⁇ 2 L 2
- L 1 is the objective function of the correlation prediction task model
- L 2 is the objective function of the real-time event summary task model
- ⁇ 1 , ⁇ 2 are the weights of L 1 and L 2 respectively Coefficient
- training the multi-task joint training model means that the correlation prediction task model and the real-time event summary task model are synchronized training, taking into account the interdependence of the correlation prediction task and the real-time event summary task, effectively improving the real-time time summary The generation effect.
- the knowledge-aware text representation of the event text and the user query text is generated by the knowledge base, and the knowledge-aware text is interactively learned through the interactive multi-head attention network to generate the event text and the user query text.
- Interactive learning text representation These interactive learning text representations are processed through a dynamic memory network to generate a specific text representation of the event text.
- the specific text representation of the event text is input into the multi-task joint training model to generate a real-time event summary of the text stream. This effectively improves the content richness and performance of the real-time event summary, reduces the redundancy of the real-time event summary, and further improves the generation effect of the real-time event summary.
- Fig. 3 shows the structure of the device for generating a real-time event summary provided in the third embodiment of the present invention.
- Fig. 3 shows the structure of the device for generating a real-time event summary provided in the third embodiment of the present invention.
- parts related to the embodiment of the present invention including:
- the text receiving module 31 is used to receive a text stream and user query text, the text stream includes event text sorted by time;
- the knowledge-aware representation generating module 32 is used to generate the knowledge-aware text representation of the event text and the knowledge-aware text representation of the user query text according to the event text, the user query text and the preset knowledge base;
- Interactive representation generating module 33 used to generate interactive learning text representation of event text and user query text based on knowledge-aware text representation of event text, knowledge-aware text representation of user query text and trained interactive multi-head attention network Interactive learning text representation;
- the specific representation generation module 34 is used to generate a specific text representation of the event text based on the interactive learning text representation of the event text, the interactive learning text representation of the user query text and the trained dynamic memory network;
- the real-time summary generation module 35 is used to input the specific text representation of the event text into the trained multi-task joint training model to generate a real-time event summary of the text stream.
- the multi-task joint training model includes a real-time event summary task model and a correlation prediction task model .
- the knowledge-aware representation generating module 32 includes:
- the context generation module 321 is configured to obtain the initial context representation of the event text by extracting the hidden state of words in the event text, and obtain the initial context representation of the user query text by extracting the hidden state of the words in the user query text;
- the initial knowledge representation generating module 322 is used to generate the initial knowledge representation of the event text based on the initial context representation, attention mechanism and knowledge base of the event text, and generate queries based on the initial context representation, attention mechanism and knowledge base of the user query text The initial knowledge representation of the text;
- the knowledge-aware representation combination module 323 is used to combine the initial context representation of the event text and the initial knowledge representation of the event text to obtain the knowledge-aware text representation of the event text, and combine the initial context representation of the user query text and the initial knowledge representation of the user query text Get the knowledge-aware text representation of the user's query text.
- the interactive representation generating module 33 includes:
- the attention matrix calculation module is used to input the knowledge-aware text representation of the event text and the knowledge-aware text representation of the user query text into the interactive multi-head attention network, and calculate the attention matrix of the event text and the attention matrix of the user query text;
- the interactive representation generation sub-module is used to calculate the interactive learning text representation of the event text according to the attention matrix and knowledge perception text representation of the event text, and calculate the user according to the attention matrix and knowledge perception text representation of the user query text Interactive learning text representation of query text.
- the specific representation generating module 34 includes:
- the memory content acquisition module is used to acquire the memory content of the event text under the last time stamp in the text stream.
- the specific representation generation sub-module is used to input the memory content of the event text at the last time stamp, the interactive learning text representation of the event text at the current time stamp, and the interactive learning text representation of the user query text into the dynamic memory network to obtain the current time Stamp a specific textual representation of the event text.
- the specific representation generating module 34 further includes:
- the memory content calculation module is used to calculate the memory content of the event text under the current time stamp based on the specific text representation of the event text under the current time stamp and the memory content of the event text under the previous time stamp.
- the device for generating a real-time event summary further includes:
- the training module is used to obtain training data, and simultaneously train the real-time event summary task and the correlation prediction task according to the training data.
- the real-time event summary task is trained by strategy gradient algorithm, and the correlation prediction task is trained in a supervised manner.
- the knowledge-aware text representation of the event text and the user query text is generated by the knowledge base, and the knowledge-aware text is interactively learned through the interactive multi-head attention network to generate the event text and the user query text.
- Interactive learning text representation These interactive learning text representations are processed through a dynamic memory network to generate a specific text representation of the event text.
- the specific text representation of the event text is input into the multi-task joint training model to generate a real-time event summary of the text stream. This effectively improves the content richness and performance of the real-time event summary, reduces the redundancy of the real-time event summary, and further improves the generation effect of the real-time event summary.
- each unit of the apparatus for generating a real-time event summary can refer to the detailed description of the corresponding steps in the first embodiment and the second embodiment, which will not be repeated here.
- each unit of the device for generating a real-time event summary can be implemented by a corresponding hardware or software unit.
- Each unit can be an independent software and hardware unit, or can be integrated into a software and hardware unit. Limit the invention.
- FIG. 5 shows the structure of the computer device provided in the fourth embodiment of the present invention. For ease of description, only the parts related to the embodiment of the present invention are shown.
- the computer device 5 in the embodiment of the present invention includes a processor 50, a memory 51, and a computer program 52 stored in the memory 51 and running on the processor 50.
- the processor 50 implements the steps in the foregoing method embodiments when the computer program 52 is executed, such as steps S101 to S105 shown in FIG. 1.
- the processor 50 executes the computer program 52, the functions of the units in the above-mentioned device embodiments, such as the functions of the units 31 to 35 shown in FIG. 3, are realized.
- the knowledge-aware text representation of the event text and the user query text is generated by the knowledge base, and the knowledge-aware text is interactively learned through the interactive multi-head attention network to generate the event text and the user query text.
- Interactive learning text representation These interactive learning text representations are processed through a dynamic memory network to generate a specific text representation of the event text.
- the specific text representation of the event text is input into the multi-task joint training model to generate a real-time event summary of the text stream. This effectively improves the content richness and performance of the real-time event summary, reduces the redundancy of the real-time event summary, and further improves the generation effect of the real-time event summary.
- a computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps in the foregoing method embodiment are implemented, for example, as shown in FIG. Steps S101 to S105 shown.
- the computer program when executed by the processor, realizes the functions of the units in the above-mentioned device embodiment, for example, the functions of units 31 to 35 shown in FIG. 3.
- the knowledge-aware text representation of the event text and the user query text is generated by the knowledge base, and the knowledge-aware text is interactively learned through the interactive multi-head attention network to generate the event text and the user query text.
- Interactive learning text representation These interactive learning text representations are processed through a dynamic memory network to generate a specific text representation of the event text.
- the specific text representation of the event text is input into the multi-task joint training model to generate a real-time event summary of the text stream. This effectively improves the content richness and performance of the real-time event summary, reduces the redundancy of the real-time event summary, and further improves the generation effect of the real-time event summary.
- the computer-readable storage medium in the embodiment of the present invention may include any entity or device or recording medium capable of carrying computer program code, such as ROM/RAM, magnetic disk, optical disk, flash memory and other memories.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Machine Translation (AREA)
Abstract
Description
Claims (10)
- 一种实时事件摘要的生成方法,其特征在于,所述方法包括下述步骤:接收文本流和用户查询文本,所述文本流包括按时间排序的事件文本;依据所述事件文本、所述用户查询文本和预设的知识库,生成所述事件文本的知识感知文本表示和所述用户查询文本的知识感知文本表示;依据所述事件文本的知识感知文本表示、所述用户查询文本的知识感知文本表示和训练好的交互式多头注意力网络,生成所述事件文本的交互式学习文本表示和所述用户查询文本的交互式学习文本表示;依据所述事件文本的交互式学习文本表示、所述用户查询文本的交互式学习文本表示和训练好的动态记忆网络,生成所述事件文本的特定文本表示;将所述事件文本的特定文本表示输入训练好的多任务联合训练模型,生成所述文本流的实时事件摘要,所述多任务联合训练模型包括实时事件摘要任务模型和相关性预测任务模型。
- 如权利要求1所述的方法,其特征在于,所述生成所述事件文本的知识感知文本表示和所述用户查询文本的知识感知文本表示的步骤,包括:通过提取所述事件文本中单词的隐藏状态,得到所述事件文本的初始上下文表示,通过提取所述用户查询文本中单词的隐藏状态,得到所述用户查询文本的初始上下文表示;根据所述事件文本的初始上下文表示、注意力机制和所述知识库,生成所述事件文本的初始知识表示,根据所述用户查询文本的初始上下文表示、注意力机制和所述知识库,生成所述查询文本的初始知识表示;由所述事件文本的初始上下文表示和所述事件文本的初始知识表示组合得到所述事件文本的知识感知文本表示,由所述用户查询文本的初始上下文表示和所述用户查询文本的初始知识表示组合得到所述用户查询文本的知识感知文本表示。
- 如权利要求1所述的方法,其特征在于,所述生成所述事件文本的交互 式学习文本表示和所述用户查询文本的交互式学习文本表示的步骤,包括:将所述事件文本的知识感知文本表示和所述用户查询文本的知识感知文本表示输入所述交互式多头注意力网络,计算所述事件文本的注意力矩阵和所述用户查询文本的注意力矩阵;根据所述事件文本的注意力矩阵和知识感知文本表示,计算得到所述事件文本的交互式学习文本表示,根据所述用户查询文本的注意力矩阵和知识感知文本表示,计算得到所述用户查询文本的交互式学习文本表示。
- 如权利要求1所述的方法,其特征在于,所述生成所述事件文本的特定文本表示的步骤,包括:获取所述文本流中上一时间戳下的事件文本的记忆内容;将所述上一时间戳下事件文本的记忆内容、当前时间戳下事件文本的交互式学习文本表示和用户查询文本的交互式学习文本表示输入动态记忆网络,获得所述当前时间戳下事件文本的特定文本表示。
- 如权利要求4所述的方法,其特征在于,所述生成所述事件文本的特定文本表示的步骤,还包括:根据所述当前时间戳下事件文本的特定文本表示和所述上一时间戳下事件文本的记忆内容,计算所述当前时间戳下事件文本的记忆内容。
- 如权利要求1所述的方法,其特征在于,在所述接收文本流和用户查询文本的步骤之前,所述方法还包括:获取训练数据,根据所述训练数据对所述实时事件摘要任务与所述相关性预测任务进行同时训练,所述实时事件摘要任务采用策略梯度算法进行训练,所述相关性预测任务采用有监督方式进行训练。
- 一种实时事件摘要的生成装置,其特征在于,所述装置包括:文本接收模块,用于接收文本流和用户查询文本,所述文本流包括按时间排序的事件文本;知识感知表示生成模块,用于依据所述事件文本、所述用户查询文本和预 设的知识库,生成所述事件文本的知识感知文本表示和所述用户查询文本的知识感知文本表示;交互式表示生成模块,用于依据所述事件文本的知识感知文本表示、所述用户查询文本的知识感知文本表示和训练好的交互式多头注意力网络,生成所述事件文本的交互式学习文本表示和所述用户查询文本的交互式学习文本表示;特定表示生成模块,用于依据所述事件文本的交互式学习文本表示、所述用户查询文本的交互式学习文本表示和训练好的动态记忆网络,生成所述事件文本的特定文本表示;以及实时摘要生成模块,用于将所述事件文本的特定文本表示输入训练好的多任务联合训练模型,生成所述文本流的实时事件摘要,所述多任务联合训练模型包括实时事件摘要任务模型和相关性预测任务模型。
- 如权利要求6所述的装置,其特征在于,所述知识感知表示生成模块包括:上下文生成模块,用于通过提取所述事件文本中单词的隐藏状态,得到所述事件文本的初始上下文表示,通过提取所述用户查询文本中单词的隐藏状态,得到所述用户查询文本的初始上下文表示;初始知识表示生成模块,用于根据所述事件文本的初始上下文表示、注意力机制和所述知识库,生成所述事件文本的初始知识表示,根据所述用户查询文本的初始上下文表示、注意力机制和所述知识库,生成所述查询文本的初始知识表示;以及知识感知表示组合模块,用于由所述事件文本的初始上下文表示和所述事件文本的初始知识表示组合得到所述事件文本的知识感知文本表示,由所述用户查询文本的初始上下文表示和所述用户查询文本的初始知识表示组合得到所述用户查询文本的知识感知文本表示。
- 一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在 所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至6任一项所述方法的步骤。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至6任一项所述方法的步骤。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/088630 WO2020237479A1 (zh) | 2019-05-27 | 2019-05-27 | 实时事件摘要的生成方法、装置、设备及存储介质 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/088630 WO2020237479A1 (zh) | 2019-05-27 | 2019-05-27 | 实时事件摘要的生成方法、装置、设备及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020237479A1 true WO2020237479A1 (zh) | 2020-12-03 |
Family
ID=73553569
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/088630 WO2020237479A1 (zh) | 2019-05-27 | 2019-05-27 | 实时事件摘要的生成方法、装置、设备及存储介质 |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2020237479A1 (zh) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106484767A (zh) * | 2016-09-08 | 2017-03-08 | 中国科学院信息工程研究所 | 一种跨媒体的事件抽取方法 |
CN108763211A (zh) * | 2018-05-23 | 2018-11-06 | 中国科学院自动化研究所 | 融合蕴含知识的自动文摘方法及系统 |
CN109344391A (zh) * | 2018-08-23 | 2019-02-15 | 昆明理工大学 | 基于神经网络的多特征融合中文新闻文本摘要生成方法 |
-
2019
- 2019-05-27 WO PCT/CN2019/088630 patent/WO2020237479A1/zh active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106484767A (zh) * | 2016-09-08 | 2017-03-08 | 中国科学院信息工程研究所 | 一种跨媒体的事件抽取方法 |
CN108763211A (zh) * | 2018-05-23 | 2018-11-06 | 中国科学院自动化研究所 | 融合蕴含知识的自动文摘方法及系统 |
CN109344391A (zh) * | 2018-08-23 | 2019-02-15 | 昆明理工大学 | 基于神经网络的多特征融合中文新闻文本摘要生成方法 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110297885B (zh) | 实时事件摘要的生成方法、装置、设备及存储介质 | |
WO2019114512A1 (zh) | 用于客户服务的方法、装置、电子设备、计算机可读存储介质 | |
Zhang et al. | Cross-domain recommendation with semantic correlation in tagging systems | |
Cai et al. | Intelligent question answering in restricted domains using deep learning and question pair matching | |
RU2720074C2 (ru) | Способ и система создания векторов аннотации для документа | |
CN103886047A (zh) | 面向流式数据的分布式在线推荐方法 | |
CN108833933A (zh) | 一种使用支持向量机推荐视频流量的方法及系统 | |
CN111563158A (zh) | 文本排序方法、排序装置、服务器和计算机可读存储介质 | |
Xiao et al. | User behavior prediction of social hotspots based on multimessage interaction and neural network | |
CN115510226A (zh) | 一种基于图神经网络的情感分类方法 | |
Liu et al. | Heterogeneous relational graph neural networks with adaptive objective for end-to-end task-oriented dialogue | |
Yan et al. | Response selection from unstructured documents for human-computer conversation systems | |
CN112231554A (zh) | 一种搜索推荐词生成方法、装置、存储介质和计算机设备 | |
Pulikottil et al. | Onet–a temporal meta embedding network for mooc dropout prediction | |
Garg et al. | Reinforced approximate exploratory data analysis | |
WO2020237479A1 (zh) | 实时事件摘要的生成方法、装置、设备及存储介质 | |
Wang | College physical education and training in big data: a big data mining and analysis system | |
Yao et al. | Scalable algorithms for CQA post voting prediction | |
Xu et al. | A novel data-to-text generation model with transformer planning and a wasserstein auto-encoder | |
Zeng et al. | Multi-aspect attentive text representations for simple question answering over knowledge base | |
CN113849641A (zh) | 一种跨领域层次关系的知识蒸馏方法和系统 | |
Li et al. | Personalized education resource recommendation method based on deep learning in intelligent educational robot environments | |
Du et al. | Employ Multimodal Machine Learning for Content Quality Analysis | |
Matsuda et al. | Benchmark for Personalized Federated Learning | |
Wang et al. | MOOC resources recommendation based on heterogeneous information network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19931483 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19931483 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19931483 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 090622) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19931483 Country of ref document: EP Kind code of ref document: A1 |