WO2018218706A1 - Method and system for extracting news event based on neural network - Google Patents

Method and system for extracting news event based on neural network Download PDF

Info

Publication number
WO2018218706A1
WO2018218706A1 PCT/CN2017/089136 CN2017089136W WO2018218706A1 WO 2018218706 A1 WO2018218706 A1 WO 2018218706A1 CN 2017089136 W CN2017089136 W CN 2017089136W WO 2018218706 A1 WO2018218706 A1 WO 2018218706A1
Authority
WO
WIPO (PCT)
Prior art keywords
event
sentence
word
trigger word
neural network
Prior art date
Application number
PCT/CN2017/089136
Other languages
French (fr)
Chinese (zh)
Inventor
周勇
刘兵
陈斌
王重秋
Original Assignee
中国矿业大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国矿业大学 filed Critical 中国矿业大学
Publication of WO2018218706A1 publication Critical patent/WO2018218706A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Definitions

  • the invention relates to natural language processing, in particular to a news event extraction method and system based on a bidirectional long-term memory network (BiLSTM) and a convolutional neural network (CNN).
  • BiLSTM bidirectional long-term memory network
  • CNN convolutional neural network
  • Event extraction is produced in this context. As a sub-task of information extraction, event extraction is a research hotspot of information extraction. Its research content is to automatically find specific types of events and their event elements from natural text.
  • Extracting the corresponding event from the text is usually done by identifying the trigger word of the event, so the trigger word is the key to identifying the event instance.
  • the patent document with the patent number CN201210321193.1 discloses an event extraction method, which uses a trigger word morphological structure and a similarity degree to extend the trigger word, so that when the event instance is extracted, not only the event corresponding to the known trigger word can be extracted. For example, an event instance corresponding to the extended unknown trigger word may also be extracted, which improves the recall rate of the event extraction.
  • the patent document with the patent number CN201410108447.0 discloses a method for extracting news atomic events. Firstly, the initial fusion rule base and the information unit fusion rule base are used to fuse the part-of-speech and noun recognition results, and then the core vocabulary and event extraction rules are utilized. The library extracts events from the information unit fusion result of the news body.
  • the object of the present invention is to overcome the deficiencies in the prior art, and to provide a neural network based method and system for news event extraction to eliminate candidate trigger word ambiguity and to be able to process news events of irregular statements.
  • a neural network based method for news event extraction comprising the following steps:
  • Step S1 performing data preprocessing on the original text of the training corpus: segmenting the original text of the training corpus, obtaining the event sentence, and then performing segmentation and naming recognition on the event sentence; according to the manually marked news event information, The sentence is sequenced, the trigger word is marked according to its type, the non-trigger word is marked as no category, and the sequence of event sentences is obtained; and the sequence of event sentences is expressed in the form of a word vector;
  • Step S2 the event sentence sequence represented by the word vector is transmitted to the two-way long and short time memory network, and the semantic feature of each candidate trigger word is trained by using the two-way long and short time memory network;
  • Step S3 the event sentence sequence represented by the word vector is introduced into the convolutional neural network, and the global feature of the event sentence where the candidate trigger word is located is trained by using the convolutional neural network;
  • Step S4 according to the semantic feature of the candidate trigger word obtained in step S2 and the global feature of the sentence in which the candidate trigger word is obtained in step S3, use softmax as a classifier to classify each candidate trigger word to find a news event.
  • Step S1 is specifically:
  • Step S12 according to the word segmentation and the naming body recognition result, the event sentence is manually labeled, in the labeling process, the non-trigger word is marked as untyped, and the trigger word is marked according to the news event category to obtain the event sentence sequence;
  • Step S13 the word vector is obtained through the open source toolkit word2vec training, and each word in the sequence of the event sentence is expressed as a vector of 300 length according to the word vector obtained by the training using the Skip-gram model;
  • Step S2 is specifically:
  • the output result of the reverse sequence BW ⁇ bw 1 , bw 2 , . . . , bw i , . . . , bw n ⁇ , where bw i indicates that the ith candidate trigger word is extracted by the memory network in the reverse length Semantic features;
  • Step S3 is specifically as follows:
  • step S32 a convolution operation is performed on the event sentence, and the calculation formula is:
  • f is the activation function
  • C i is the feature obtained by convolution
  • w is the weight matrix
  • h is the convolution kernel size
  • i:i-h+1 is the i-th word to the i-h+1th word.
  • b indicates the offset;
  • step S33 using the maximum pooling, the feature map is pooled to obtain a global feature C o of the event sentence.
  • Step S4 is specifically:
  • step S32 the output vector O t is classified using softmax to obtain the type of news event prediction.
  • a system for news event extraction based on neural network comprising a text and processing module, a neural network training module, and a news event prediction module, wherein:
  • the text and processing module is configured to perform data preprocessing on the original text of the training corpus, including: segmenting the original text of the training corpus, obtaining an event sentence, and then performing segmentation and naming recognition on the event sentence; and manually marking the news event according to the manual
  • the information is sequenced by the event sentence, the trigger word is marked according to its type, the non-trigger word is marked as no category, and the sequence of the event sentence is obtained; and the sequence of the event sentence is expressed in the form of a word vector;
  • the neural network training module comprises a bidirectional long and short time memory network training module and a convolutional neural network training module, and the bidirectional long and short time memory network training module is configured to train the event sentence sequence represented by the word vector to obtain the semantics of each candidate trigger word.
  • the convolutional neural network training module is configured to train the sequence of event sentences represented by the word vector to obtain a global feature of the event sentence in which the candidate trigger word is located;
  • the news event prediction module is configured to classify each candidate trigger word by using softmax as a classifier according to the semantic feature of the candidate trigger word obtained by the neural network training module and the global feature of the sentence where the candidate trigger word is located, thereby finding the news.
  • the trigger word of the event and according to the type of the trigger word, determine the type of the event.
  • the present invention Since the above technical solution is adopted, the present invention has the following beneficial effects as compared with the prior art:
  • the present invention employs a bidirectional long and short time memory network (BiLSTM), which is capable of eliminating ambiguity of candidate trigger words based on context information of candidate trigger words. For example, “A car hits the guardrail of the expressway.” And “I went to dinner today and just hit a classmate who hasn't seen it for a long time.” The trigger words in the above two sentences are "collided", the former belongs to traffic. Accident-like incidents, the latter being incidents of encounters. When BiLSTM is used to extract the candidate trigger term meaning information, the actual meaning of the candidate trigger word can be judged according to the context information of the sentence, which can effectively avoid the ambiguity phenomenon of the vocabulary, thereby improving the accuracy of the news event classification.
  • BiLSTM bidirectional long and short time memory network
  • the global feature of the sentence extracted by the convolutional neural network (CNN) is used in the present invention.
  • the event category can be accurately determined according to the global feature of the sentence and the semantic features of the candidate trigger word. Therefore, the present invention can solve the problem of news event recognition of irregular statements.
  • FIG. 1 is a flowchart of a neural network based news event extraction method and system provided by the present invention
  • Figure 2 is a key step workflow for news event extraction based on bidirectional long-term memory network (BiLSTM) and convolutional neural network (CNN);
  • BiLSTM bidirectional long-term memory network
  • CNN convolutional neural network
  • FIG. 3 is a schematic diagram of the structure of a convolutional neural network (CNN).
  • CNN convolutional neural network
  • a neural network based news event extraction system includes a text and processing module, a neural network training module, and a news event prediction module, wherein:
  • the text and processing module is used for data preprocessing of the original text of the training corpus, including: segmenting the original text of the training corpus, obtaining the event sentence, and then performing segmentation and naming recognition on the event sentence; according to the manually marked news event information, The event sentence is sequenced, the trigger word is labeled according to its type, and the non-trigger word is marked as none. Category, get the sequence of event sentences; and express the sequence of event sentences in the form of a word vector;
  • the neural network training module comprises a two-way long-term memory network training module and a convolutional neural network training module, and the two-way long-term memory network training module is configured to train the sequence of event sentences represented by the word vector to obtain the semantic features of each candidate trigger word;
  • the convolutional neural network training module is configured to train the sequence of event sentences represented by the word vector to obtain global features of the event sentences in which the candidate trigger words are located;
  • the news event prediction module is configured to classify each candidate trigger word by using softmax as a classifier according to the semantic feature of the candidate trigger word obtained by the neural network training module and the global feature of the sentence in which the candidate trigger word is located, thereby finding a news event. Trigger the word and determine the type of the event based on the type of the trigger word.
  • a method based on neural network for news event extraction the example sentence is: "11:25, S20 outer ring Humin interchange has a 3 car rear-end collision.” It is known that the event trigger word in this sentence is "tail-catching" The news event category belongs to a traffic accident.
  • Step 1 Perform segmentation on the event sentence and identify the name. Available:
  • BiLSTM bidirectional long-term memory network
  • x t as the input word vector at time t
  • h t is the hidden layer state vector storing all useful information at time t
  • is the sigmoid regression layer
  • U i , U f , U c , U o are different for the input
  • the weight matrix of x t , W i , W f , W c , W o is the weight matrix of the hidden layer state h t
  • b i , b f , b c , b o are offset vectors;
  • i t represents the value at t, which determines the value to be updated, Indicates the information that determines the update.
  • o t is the output at time t
  • h t is the vector of the hidden layer at time t.
  • fw t o t , that is, the output of the tth word in the sentence is fw t .
  • CNN convolutional neural network
  • f is the activation function
  • C i is the feature obtained by convolution
  • w is the weight matrix
  • h is the convolution kernel size
  • i:i-h+1 is the i-th word to the i-h+1th word.
  • b indicates the offset;

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

A method and system for extracting a news event based on a neural network. The method comprises the steps of: carrying out data pre-processing on original text of a training corpus; introducing an event sentence sequence represented by a word vector into a bidirectional long- and short-term memory network, and using the bidirectional long- and short-term memory network to carry out training so as to obtain a semantic feature of each candidate trigger word; introducing the event sentence sequence represented by a word vector into a convolutional neural network, and using the convolutional neural network to carry out training so as to obtain a global feature of an event sentence in which the candidate trigger word is; and according to the semantic feature of the candidate trigger word and the global feature of the sentence in which the candidate trigger word is, using softmax as a classifier to classify each candidate trigger word, so as to find a trigger word of a news event, and according to the type of the trigger word, determining the type to which the event belongs. The method and system can quickly and accurately extract a news event and process a news event included in a non-standard sentence, and have the characteristics of high efficiency and universal applicability.

Description

一种基于神经网络的新闻事件抽取的方法及系统Method and system for news event extraction based on neural network 技术领域Technical field
本发明涉及自然语言处理,尤其涉及一种基于双向长短时记忆网络(BiLSTM)与卷积神经网络(CNN)结合的新闻事件抽取方法及系统。The invention relates to natural language processing, in particular to a news event extraction method and system based on a bidirectional long-term memory network (BiLSTM) and a convolutional neural network (CNN).
背景技术Background technique
随着计算机的发展和互联网的日益普及,大量信息以电子文本的形式出现在人们面前。在大量的网络文本中,如何发掘有价值的新闻事件已经成为迫切需要解决的问题,事件抽取正是在这种背景下产生的。作为信息抽取的一个子任务,事件抽取是信息抽取的研究热点,它的研究内容是自动的从自然文本中发现特定类型的事件及其事件元素。With the development of computers and the increasing popularity of the Internet, a large amount of information appears in the form of electronic text. In a large number of web texts, how to find valuable news events has become an urgent problem to be solved. Event extraction is produced in this context. As a sub-task of information extraction, event extraction is a research hotspot of information extraction. Its research content is to automatically find specific types of events and their event elements from natural text.
从文本中抽取相应的事件通常是通过识别事件的触发词来实现的,所以触发词是识别事件实例的关键。Extracting the corresponding event from the text is usually done by identifying the trigger word of the event, so the trigger word is the key to identifying the event instance.
专利号为CN201210321193.1的专利文献公开了一种事件抽取方法,利用触发词形态结构,结合义原相似度来扩展触发词,这样在抽取事件实例时,不仅可以抽取已知触发词对应的事件实例,还可以抽取扩展的未知触发词对应的事件实例,提高了事件抽取的召回率。专利号为CN201410108447.0的专利文献公开了一种新闻原子事件抽取方法,首先利用初步融合规则库和信息单元融合规则库对词性和命名体识别结果进行融合,然后利用核心词表和事件抽取规则库对新闻正文的信息单元融合结果进行事件抽取。The patent document with the patent number CN201210321193.1 discloses an event extraction method, which uses a trigger word morphological structure and a similarity degree to extend the trigger word, so that when the event instance is extracted, not only the event corresponding to the known trigger word can be extracted. For example, an event instance corresponding to the extended unknown trigger word may also be extracted, which improves the recall rate of the event extraction. The patent document with the patent number CN201410108447.0 discloses a method for extracting news atomic events. Firstly, the initial fusion rule base and the information unit fusion rule base are used to fuse the part-of-speech and noun recognition results, and then the core vocabulary and event extraction rules are utilized. The library extracts events from the information unit fusion result of the news body.
基于上述研究现状,针对新闻事件抽取主要存在以下问题:第一、对新闻事件的判别主要依赖于触发词本身,忽视了上下文关系,在遇到具有歧义的候选触发词时,容易造成事件类别的判断错误。第二,网络文本尤其是微博文本多为不规范语句,当前的事件抽取方法缺乏从不规范语句中提取事件的研究。Based on the above research status, there are mainly the following problems for news event extraction: First, the discrimination of news events mainly depends on the trigger word itself, neglecting the context relationship, and it is easy to cause event categories when encountering ambiguous candidate trigger words. Misjudgment. Second, the network text, especially the microblog text, is mostly an irregular statement. The current event extraction method lacks the research of extracting events from non-standard statements.
发明内容Summary of the invention
本发明的目的是克服已有技术中的不足之处,提供一种基于神经网络的新闻事件抽取的方法及系统,以消除候选触发词歧义并且能够处理不规范语句的新闻事件。The object of the present invention is to overcome the deficiencies in the prior art, and to provide a neural network based method and system for news event extraction to eliminate candidate trigger word ambiguity and to be able to process news events of irregular statements.
为实现上述目的,本发明采用的技术方案为:In order to achieve the above object, the technical solution adopted by the present invention is:
一种基于神经网络的新闻事件抽取的方法,包括以下步骤:A neural network based method for news event extraction, comprising the following steps:
步骤S1,对训练语料原始文本进行数据预处理:对训练语料原始文本进行分句,得到事件句,然后对事件句进行分词、命名体识别;根据人工标注的新闻事件信息,将事 件句进行序列标注,触发词根据其类型进行标注,非触发词标注为无类别,得到事件句序列;并将事件句序列以词向量的形式进行表述;Step S1: performing data preprocessing on the original text of the training corpus: segmenting the original text of the training corpus, obtaining the event sentence, and then performing segmentation and naming recognition on the event sentence; according to the manually marked news event information, The sentence is sequenced, the trigger word is marked according to its type, the non-trigger word is marked as no category, and the sequence of event sentences is obtained; and the sequence of event sentences is expressed in the form of a word vector;
步骤S2,将以词向量表示的事件句序列传入到双向长短时记忆网络,利用双向长短时记忆网络来训练得到每个候选触发词的语义特征;Step S2, the event sentence sequence represented by the word vector is transmitted to the two-way long and short time memory network, and the semantic feature of each candidate trigger word is trained by using the two-way long and short time memory network;
步骤S3,将以词向量表示的事件句序列传入到卷积神经网络中,利用卷积神经网络训练得到候选触发词所在事件句的全局特征;Step S3, the event sentence sequence represented by the word vector is introduced into the convolutional neural network, and the global feature of the event sentence where the candidate trigger word is located is trained by using the convolutional neural network;
步骤S4,根据步骤S2中得到的候选触发词的语义特征与步骤S3中获得的候选触发词所在句子的全局特征,利用softmax作为分类器来对每个候选触发词进行分类,从而找出新闻事件的触发词,并根据触发词类型,判断该事件所属类型。Step S4, according to the semantic feature of the candidate trigger word obtained in step S2 and the global feature of the sentence in which the candidate trigger word is obtained in step S3, use softmax as a classifier to classify each candidate trigger word to find a news event. The trigger word, and according to the trigger word type, determine the type of the event.
步骤S1具体为:Step S1 is specifically:
步骤S11,采用自然语言处理工具对训练语料原始文本进行分句,分词和命名体识别,使训练语料原始文本以事件句和每句包含若干个词汇呈现,则事件句表达为L={w1,w2,...,wi,...,wn},其中,wi为事件句中第i个词,n表示事件句长度;In step S11, the natural language processing tool is used to perform segmentation, word segmentation and naming body recognition on the original text of the training corpus, so that the original text of the training corpus is presented in the event sentence and each sentence includes several words, and the event sentence is expressed as L={w 1 , w 2 ,...,w i ,...,w n }, where w i is the i-th word in the event sentence, and n represents the length of the event sentence;
步骤S12,根据分词和命名体识别结果,对事件句进行人工标注,在标注过程中,将非触发词标记为无类型,触发词根据所属新闻事件类别进行标注,得到事件句序列;Step S12, according to the word segmentation and the naming body recognition result, the event sentence is manually labeled, in the labeling process, the non-trigger word is marked as untyped, and the trigger word is marked according to the news event category to obtain the event sentence sequence;
步骤S13,通过开源工具包word2vec训练得到词向量,采用Skip-gram模型,根据训练得到的词向量,将事件句序列中每个词表述为一个300长度的向量;Step S13, the word vector is obtained through the open source toolkit word2vec training, and each word in the sequence of the event sentence is expressed as a vector of 300 length according to the word vector obtained by the training using the Skip-gram model;
步骤S14,将每个事件句处理成词向量表述的序列形式,即每个候选触发词wi以一个300长度的词向量xi来表示,事件句表述为L={x1,x2,...,xi,...,xn}。Step S14, each event sentence is processed into a sequence form of a word vector representation, that is, each candidate trigger word w i is represented by a 300-length word vector x i , and the event sentence is expressed as L={x 1 , x 2 , ..., x i ,...,x n }.
步骤S2具体为:Step S2 is specifically:
步骤S21,假设事件句表述为L={x1,x2,...,xi,...,xn},其中xi为第i个候选触发词的词向量,n表示句子长度;Step S21, assuming that the event sentence is expressed as L={x 1 , x 2 , . . . , x i , . . . , x n }, where x i is the word vector of the ith candidate trigger word, and n is the length of the sentence. ;
步骤S22,将L作为一个序列传入长短时记忆网络,得到序列的输出结果FW={fw1,fw2,...,fwi,...,fwn},其中,fwi表示第i个候选触发词经过长短时记忆网络提取出来的语义特征;Step S22, passing L as a sequence into the long and short time memory network, and obtaining the output result of the sequence FW={fw 1 , fw 2 , . . . , fw i , . . . , fw n }, where fw i represents the first The semantic features extracted by the i candidate trigger words through the long and short time memory network;
步骤S23,将L进行反转即L′={xn,xn-1,...,xi,...,x1},将反向序列L′传入长短 时记忆网络,得到反向序列的输出结果BW={bw1,bw2,...,bwi,...,bwn},其中,bwi表示第i个候选触发词经过反向长短时记忆网络提取出来的语义特征;In step S23, L is inverted, that is, L'={x n , x n-1 , . . . , x i , . . . , x 1 }, and the reverse sequence L′ is transmitted to the long and short time memory network. The output result of the reverse sequence BW={bw 1 , bw 2 , . . . , bw i , . . . , bw n }, where bw i indicates that the ith candidate trigger word is extracted by the memory network in the reverse length Semantic features;
步骤S24,将双向长短时记忆网络获得输出结果的FW与BW进行拼接,得到句子L经过双向长短时记忆网络的输出结果,即O={r1,r2,...ri,...,rn},其中,ri=[fwi:bwi]。Step S24, splicing the FW and the BW obtained by the two-way long-term memory network to obtain the output result of the sentence L after the two-way length and long-term memory network, that is, O={r 1 , r 2 ,...r i ,.. ., r n }, where r i =[fw i :bw i ].
步骤S3具体为:Step S3 is specifically as follows:
步骤S31,假设事件句表述为L={x1,x2,...,xi,...,xn},其中xi为第i个单词的词向量,n表示句子长度;Step S31, assuming that the event sentence is expressed as L = {x 1 , x 2 , ..., x i , ..., x n }, where x i is the word vector of the ith word, and n represents the length of the sentence;
步骤S32,对事件句采取卷积操作,计算公式为:In step S32, a convolution operation is performed on the event sentence, and the calculation formula is:
Ci=f(wTxi:i-h+1+b)C i =f(w T x i:i-h+1 +b)
其中,f是激活函数,Ci表示卷积得到的特征,,w表示权重矩阵,h表示卷积核大小,i:i-h+1表示第i个词到第i-h+1个词,b表示偏置;Where f is the activation function, C i is the feature obtained by convolution, w is the weight matrix, h is the convolution kernel size, and i:i-h+1 is the i-th word to the i-h+1th word. , b indicates the offset;
通过滑动窗口,对所有词进行卷积得到特征图;By sliding the window, convolving all the words to obtain a feature map;
步骤S33,使用最大池化,对特征图进行池化,得到事件句的全局特征CoIn step S33, using the maximum pooling, the feature map is pooled to obtain a global feature C o of the event sentence.
步骤S4具体为:Step S4 is specifically:
步骤S31,将双向长短时记忆网络获得的候选触发词语义特征O={r1,r2,...,ri,...,rn}与卷积神经网络提取出的句子全局特征Co进行级联,得到输出向量Ot=[O:Co];Step S31, the candidate trigger term meanings O={r 1 , r 2 , . . . , r i , . . . , r n } obtained by the two-way long-term memory network and the global features of the sentence extracted by the convolutional neural network C o is cascaded to obtain an output vector O t =[O:C o ];
步骤S32,使用softmax对输出向量Ot进行分类,获得新闻事件预测的类型。In step S32, the output vector O t is classified using softmax to obtain the type of news event prediction.
一种基于神经网络的新闻事件抽取的系统,包括文本与处理模块、神经网络训练模块、新闻事件预测模块,其中:A system for news event extraction based on neural network, comprising a text and processing module, a neural network training module, and a news event prediction module, wherein:
所述文本与处理模块用于对训练语料原始文本进行数据预处理,包括:对训练语料原始文本进行分句,得到事件句,然后对事件句进行分词、命名体识别;根据人工标注的新闻事件信息,将事件句进行序列标注,触发词根据其类型进行标注,非触发词标注为无类别,得到事件句序列;并将事件句序列以词向量的形式进行表述; The text and processing module is configured to perform data preprocessing on the original text of the training corpus, including: segmenting the original text of the training corpus, obtaining an event sentence, and then performing segmentation and naming recognition on the event sentence; and manually marking the news event according to the manual The information is sequenced by the event sentence, the trigger word is marked according to its type, the non-trigger word is marked as no category, and the sequence of the event sentence is obtained; and the sequence of the event sentence is expressed in the form of a word vector;
所述神经网络训练模块包括双向长短时记忆网络训练模块和卷积神经网络训练模块,双向长短时记忆网络训练模块用于将以词向量表示的事件句序列训练,得到每个候选触发词的语义特征;卷积神经网络训练模块用于将以词向量表示的事件句序列训练,得到候选触发词所在事件句的全局特征;The neural network training module comprises a bidirectional long and short time memory network training module and a convolutional neural network training module, and the bidirectional long and short time memory network training module is configured to train the event sentence sequence represented by the word vector to obtain the semantics of each candidate trigger word. The convolutional neural network training module is configured to train the sequence of event sentences represented by the word vector to obtain a global feature of the event sentence in which the candidate trigger word is located;
所述新闻事件预测模块用于根据神经网络训练模块得到的候选触发词的语义特征与候选触发词所在句子的全局特征,利用softmax作为分类器来对每个候选触发词进行分类,从而找出新闻事件的触发词,并根据触发词类型,判断该事件所属类型。The news event prediction module is configured to classify each candidate trigger word by using softmax as a classifier according to the semantic feature of the candidate trigger word obtained by the neural network training module and the global feature of the sentence where the candidate trigger word is located, thereby finding the news. The trigger word of the event, and according to the type of the trigger word, determine the type of the event.
有益效果:由于采用了上述技术方案,本发明与现有技术相比具有如下有益效果:Advantageous Effects: Since the above technical solution is adopted, the present invention has the following beneficial effects as compared with the prior art:
1、本发明采用了双向长短时记忆网络(BiLSTM),能够根据候选触发词的上下文信息消除候选触发词的歧义。比如“一辆车撞上了高速公路的护栏。”和“今天我去吃饭的时候正好撞上了好久不见的同学。”,以上两个句子中触发词都为“撞上”,前者属于交通事故类事件,后者属于遇见类事件。采用BiLSTM提取候选触发词语义信息时,能够根据句子的上下文信息判断出候选触发词的实际意思,可有效避免词汇的歧义现象,从而提高新闻事件分类的准确率。1. The present invention employs a bidirectional long and short time memory network (BiLSTM), which is capable of eliminating ambiguity of candidate trigger words based on context information of candidate trigger words. For example, "A car hits the guardrail of the expressway." And "I went to dinner today and just hit a classmate who hasn't seen it for a long time." The trigger words in the above two sentences are "collided", the former belongs to traffic. Accident-like incidents, the latter being incidents of encounters. When BiLSTM is used to extract the candidate trigger term meaning information, the actual meaning of the candidate trigger word can be judged according to the context information of the sentence, which can effectively avoid the ambiguity phenomenon of the vocabulary, thereby improving the accuracy of the news event classification.
2、本发明采用卷积神经网络(CNN)抽取出的句子全局特征,当句子是不规范语句时,根据句子的全局特征与候选触发词的语义特征共同判断,可以准确的判断出事件类别。因此,本发明可以解决不规范语句的新闻事件识别问题。2. The global feature of the sentence extracted by the convolutional neural network (CNN) is used in the present invention. When the sentence is an irregular statement, the event category can be accurately determined according to the global feature of the sentence and the semantic features of the candidate trigger word. Therefore, the present invention can solve the problem of news event recognition of irregular statements.
附图说明DRAWINGS
图1是本发明提供的基于神经网络的新闻事件抽取方法及系统的流程图;1 is a flowchart of a neural network based news event extraction method and system provided by the present invention;
图2是基于双向长短时记忆网络(BiLSTM)和卷积神经网络(CNN)的新闻事件抽取的关键步骤工作流程;Figure 2 is a key step workflow for news event extraction based on bidirectional long-term memory network (BiLSTM) and convolutional neural network (CNN);
图3是卷积神经网络(CNN)结构示意图。Figure 3 is a schematic diagram of the structure of a convolutional neural network (CNN).
具体实施方式detailed description
下面通过具体实施方式对本发明做进一步的描述。The invention is further described below by way of specific embodiments.
如图1所示为一种基于神经网络的新闻事件抽取的系统,包括文本与处理模块、神经网络训练模块、新闻事件预测模块,其中:As shown in FIG. 1 , a neural network based news event extraction system includes a text and processing module, a neural network training module, and a news event prediction module, wherein:
文本与处理模块用于对训练语料原始文本进行数据预处理,包括:对训练语料原始文本进行分句,得到事件句,然后对事件句进行分词、命名体识别;根据人工标注的新闻事件信息,将事件句进行序列标注,触发词根据其类型进行标注,非触发词标注为无 类别,得到事件句序列;并将事件句序列以词向量的形式进行表述;The text and processing module is used for data preprocessing of the original text of the training corpus, including: segmenting the original text of the training corpus, obtaining the event sentence, and then performing segmentation and naming recognition on the event sentence; according to the manually marked news event information, The event sentence is sequenced, the trigger word is labeled according to its type, and the non-trigger word is marked as none. Category, get the sequence of event sentences; and express the sequence of event sentences in the form of a word vector;
神经网络训练模块包括双向长短时记忆网络训练模块和卷积神经网络训练模块,双向长短时记忆网络训练模块用于将以词向量表示的事件句序列训练,得到每个候选触发词的语义特征;卷积神经网络训练模块用于将以词向量表示的事件句序列训练,得到候选触发词所在事件句的全局特征;The neural network training module comprises a two-way long-term memory network training module and a convolutional neural network training module, and the two-way long-term memory network training module is configured to train the sequence of event sentences represented by the word vector to obtain the semantic features of each candidate trigger word; The convolutional neural network training module is configured to train the sequence of event sentences represented by the word vector to obtain global features of the event sentences in which the candidate trigger words are located;
新闻事件预测模块用于根据神经网络训练模块得到的候选触发词的语义特征与候选触发词所在句子的全局特征,利用softmax作为分类器来对每个候选触发词进行分类,从而找出新闻事件的触发词,并根据触发词类型,判断该事件所属类型。The news event prediction module is configured to classify each candidate trigger word by using softmax as a classifier according to the semantic feature of the candidate trigger word obtained by the neural network training module and the global feature of the sentence in which the candidate trigger word is located, thereby finding a news event. Trigger the word and determine the type of the event based on the type of the trigger word.
下面结合具体实例对本发明做进一步说明。The present invention will be further described below in conjunction with specific examples.
一种基于神经网络的新闻事件抽取的方法,本示例语句为:“11时25分,S20外圈沪渝立交发生一起3车追尾事故。”已知在该句子中事件触发词为“追尾”,所属的新闻事件类别为交通事故。A method based on neural network for news event extraction, the example sentence is: "11:25, S20 outer ring Humin interchange has a 3 car rear-end collision." It is known that the event trigger word in this sentence is "tail-catching" The news event category belongs to a traffic accident.
步骤一、对该事件句进行分词,命名体识别。可得:Step 1. Perform segmentation on the event sentence and identify the name. Available:
11时25分\O S20外圈\O沪渝立交\O发生\O一起\O 3车\O追尾\Y事故\O11:25\O S20 outer circle\O Hushen Interchange\O Occurrence\O together\O 3 car\O rear-end\Y accident\O
事件触发词“追尾”被标记为属于交通事故,其余候选触发词标记为无类别,获得该事件句的序列标注L={w1,w2,...,wi,...,wn},其中,wi为事件句中第i个词,n表示事件句长度。The event trigger word "tailing" is marked as belonging to a traffic accident, and the remaining candidate trigger words are marked as no category, and the sequence label of the event sentence is obtained L={w 1 , w 2 ,...,w i ,...,w n }, where w i is the ith word in the event sentence, and n is the length of the event sentence.
选取足够大的语料,使用开源工具包word2vec进行词向量的训练,选取Skip-gram模型,每个词表述为一个300长度的向量。Select a sufficiently large corpus, use the open source toolkit word2vec to train the word vector, select the Skip-gram model, and each word is expressed as a 300-length vector.
则最终该事件句可表述为:L={x1,x2,...,xi,...,xn},其中xi为事件句中第i个词的300维的向量,n表示事件句长度。Finally, the event sentence can be expressed as: L = {x 1 , x 2 , ..., x i , ..., x n }, where x i is a 300-dimensional vector of the ith word in the event sentence, n represents the length of the event sentence.
步骤二、将待训练的事件句L={x1,x2,...,xi,...,xn}传入到双向长短时记忆网络(BiLSTM),利用BiLSTM来训练得到的每个候选触发词的语义特征,如图2。Step 2: Pass the event sentence L={x 1 , x 2 ,..., x i ,..., x n } to be trained into the bidirectional long-term memory network (BiLSTM) and train it with BiLSTM. The semantic features of each candidate trigger word are shown in Figure 2.
将句子L作为一个序列传入长短时记忆网络(LSTM),得到序列的输出结果FW={fw1,fw2,...,fwi,...,fwn},fwi表示第i个候选触发词经过LSTM提取出来的语义特征,n表示事件句长度。其中对于fwi的计算过程如下: Passing the sentence L as a sequence into the long-short-time memory network (LSTM), and obtaining the output result of the sequence FW={fw 1 , fw 2 ,...,fw i ,...,fw n },fw i denotes the i The semantic features extracted by the candidate trigger words through LSTM, and n is the length of the event sentence. The calculation process for fw i is as follows:
定义xt为t时刻的输入词向量,ht为在t时刻存储所有有用信息的隐藏层状态向量,σ为sigmoid回归层,Ui,Uf,Uc,Uo为不同状态下针对输入xt的权重矩阵,Wi,Wf,Wc,Wo为隐藏层状态ht的权重矩阵,bi,bf,bc,bo为偏置向量;Define x t as the input word vector at time t , h t is the hidden layer state vector storing all useful information at time t, σ is the sigmoid regression layer, U i , U f , U c , U o are different for the input The weight matrix of x t , W i , W f , W c , W o is the weight matrix of the hidden layer state h t , and b i , b f , b c , b o are offset vectors;
(1)在t时刻遗忘门的计算如(1)式所示:(1) The calculation of the forgotten gate at time t is as shown in equation (1):
ft=σ(Wf·[ht-1,xt]+bf)          (1)f t =σ(W f ·[h t-1 ,x t ]+b f ) (1)
(2)在t时刻更新ht-1中存储的所有信息,计算公式如(2)、(3)式所示:(2) Update all the information stored in h t-1 at time t , and the calculation formula is as shown in equations (2) and (3):
it=σ(Wi·[ht-1,xt]+bi)            (2)i t =σ(W i ·[h t-1 ,x t ]+b i ) (2)
Figure PCTCN2017089136-appb-000001
Figure PCTCN2017089136-appb-000001
上式中,it表示t时刻决定需要更新的值、
Figure PCTCN2017089136-appb-000002
表示确定更新的信息。
In the above formula, i t represents the value at t, which determines the value to be updated,
Figure PCTCN2017089136-appb-000002
Indicates the information that determines the update.
(3)在t时刻将t-1时刻存储的信息更新为t时刻的存储信息,计算公式为式(4):(3) The information stored at time t-1 is updated to the storage information at time t at time t, and the calculation formula is equation (4):
Figure PCTCN2017089136-appb-000003
Figure PCTCN2017089136-appb-000003
(4)在t时刻的输出如式(5)所示,并更新ht,计算公式如式(6):(4) The output at time t is as shown in equation (5), and h t is updated, and the calculation formula is as shown in equation (6):
ot=σ(Wo·[ht-1,xt]+bo)             (5)o t =σ(W o ·[h t-1 ,x t ]+b o ) (5)
ht=ot*tanh(Ct)                 (6)h t =o t *tanh(C t ) (6)
其中ot为t时刻的输出;ht为t时刻隐藏层的向量。最终,fwt=ot,即句子中第t个词的输出为fwtWhere o t is the output at time t; h t is the vector of the hidden layer at time t. Finally, fw t = o t , that is, the output of the tth word in the sentence is fw t .
同理得出BW={bw1,bw2,...,bwi,...,bwn},将双向LSTM获得输出结果的FW与BW进行拼接,得到句子L经过双向长短时记忆网络(BiLSTM)的输出结果,即O={r1,r2,...,ri,...,rn},其中ri=[fwi:bwi]。Similarly, BW={bw 1 , bw 2 , . . . , bw i , . . . , bw n } is obtained, and the FW obtained by the bidirectional LSTM obtaining the output result is spliced with the BW, and the sentence L is obtained through the bidirectional long-term memory network. The output of (BiLSTM), that is, O = {r 1 , r 2 , ..., r i , ..., r n }, where r i = [fw i : bw i ].
步骤三、将待训练的事件句L={x1,x2,...,xi,...,xn}传入到卷积神经网络(CNN)中,利用CNN获取候选触发词所在句子的全局特征,如图3。Step 3: The event sentence L={x 1 , x 2 , . . . , x i , . . . , x n } to be trained is introduced into a convolutional neural network (CNN), and the candidate trigger word is obtained by using the CNN. The global characteristics of the sentence, as shown in Figure 3.
(1)对句子采取卷积操作,计算公式为(7): (1) Take a convolution operation on the sentence, the calculation formula is (7):
Ci=f(wTxi:i-h+1+b)            (7)C i =f(w T x i:i-h+1 +b) (7)
其中,f是激活函数,Ci表示卷积得到的特征,w表示权重矩阵,h表示卷积核大小,i:i-h+1表示第i个词到第i-h+1个词,b表示偏置;Where f is the activation function, C i is the feature obtained by convolution, w is the weight matrix, h is the convolution kernel size, and i:i-h+1 is the i-th word to the i-h+1th word. b indicates the offset;
通过滑动窗口,对所有词进行卷积得到特征图。By sliding the window, all words are convolved to obtain a feature map.
(2)使用最大池化,对特征图进行池化,得到句子特征Co(2) Using the maximum pooling, pool the feature map to get the sentence feature C o .
步骤四、根据步骤二所得事件句L={x1,x2,...,xi,...,xn}的语义特征O={r1,r2,...,ri,...,rn}(其中候选触发词xi对应语义特征为ri),和步骤三所得事件句L的全局特征Co,进行分类,从而所属的新闻事件类别。Step 4: According to the second step, the semantic features of the event sentence L={x 1 , x 2 , . . . , x i , . . . , x n } are O={r 1 , r 2 , . . . , r i , ..., r n } (where the candidate trigger word x i corresponds to the semantic feature is r i ), and the global feature C o of the event sentence L obtained in step 3 is classified, thereby belonging to the news event category.
将步骤二所得事件句L={x1,x2,...,xi,...,xn}的语义特征O={r1,r2,...,ri,...,rn}(其中候选触发词xi对应语义特征为ri),和步骤三所得事件句L的全局特征Co进行级联,得到输出向量Ot=[O:Co];使用softmax对输出向量Ot进行分类,获得新闻事件预测的类型。The semantic feature of the event sentence L={x 1 ,x 2 ,...,x i ,...,x n } obtained in step 2 is O={r 1 ,r 2 ,...,r i ,.. , r n } (where the candidate trigger word x i corresponds to the semantic feature r i ), and cascades with the global feature C o of the event sentence L obtained in step three to obtain an output vector O t =[O:C o ]; Softmax classifies the output vector O t to obtain the type of news event prediction.
以上所述仅是本发明的优选实施方式,应当指出:对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。 The above description is only a preferred embodiment of the present invention, and it should be noted that those skilled in the art can also make several improvements and retouchings without departing from the principles of the present invention. It should be considered as the scope of protection of the present invention.

Claims (6)

  1. 一种基于神经网络的新闻事件抽取的方法,其特征在于:包括以下步骤:A method for extracting news events based on a neural network, comprising: the following steps:
    步骤S1,对训练语料原始文本进行数据预处理:对训练语料原始文本进行分句,得到事件句,然后对事件句进行分词、命名体识别;根据人工标注的新闻事件信息,将事件句进行序列标注,触发词根据其类型进行标注,非触发词标注为无类别,得到事件句序列;并将事件句序列以词向量的形式进行表述;Step S1: performing data preprocessing on the original text of the training corpus: segmenting the original text of the training corpus, obtaining an event sentence, and then performing segmentation and naming recognition on the event sentence; and sequence the event sentence according to the manually marked news event information Labeling, the trigger word is marked according to its type, the non-trigger word is marked as no category, and the sequence of event sentences is obtained; and the sequence of event sentences is expressed in the form of a word vector;
    步骤S2,将以词向量表示的事件句序列传入到双向长短时记忆网络,利用双向长短时记忆网络来训练得到每个候选触发词的语义特征;Step S2, the event sentence sequence represented by the word vector is transmitted to the two-way long and short time memory network, and the semantic feature of each candidate trigger word is trained by using the two-way long and short time memory network;
    步骤S3,将以词向量表示的事件句序列传入到卷积神经网络中,利用卷积神经网络训练得到候选触发词所在事件句的全局特征;Step S3, the event sentence sequence represented by the word vector is introduced into the convolutional neural network, and the global feature of the event sentence where the candidate trigger word is located is trained by using the convolutional neural network;
    步骤S4,根据步骤S2中得到的候选触发词的语义特征与步骤S3中获得的候选触发词所在句子的全局特征,利用softmax作为分类器来对每个候选触发词进行分类,从而找出新闻事件的触发词,并根据触发词类型,判断该事件所属类型。Step S4, according to the semantic feature of the candidate trigger word obtained in step S2 and the global feature of the sentence in which the candidate trigger word is obtained in step S3, use softmax as a classifier to classify each candidate trigger word to find a news event. The trigger word, and according to the trigger word type, determine the type of the event.
  2. 根据权利要求1所述的基于神经网络的新闻事件抽取的方法,其特征在于:步骤S1具体为:The method for extracting news events based on a neural network according to claim 1, wherein the step S1 is specifically:
    步骤S11,采用自然语言处理工具对训练语料原始文本进行分句,分词和命名体识别,使训练语料原始文本以事件句和每句包含若干个词汇呈现,则事件句表达为L={w1,w2,...,wi,...,wn},其中,wi为事件句中第i个词,n表示事件句长度;In step S11, the natural language processing tool is used to perform segmentation, word segmentation and naming body recognition on the original text of the training corpus, so that the original text of the training corpus is presented in the event sentence and each sentence includes several words, and the event sentence is expressed as L={w 1 , w 2 ,...,w i ,...,w n }, where w i is the i-th word in the event sentence, and n represents the length of the event sentence;
    步骤S12,根据分词和命名体识别结果,对事件句进行人工标注,在标注过程中,将非触发词标记为无类型,触发词根据所属新闻事件类别进行标注,得到事件句序列;Step S12, according to the word segmentation and the naming body recognition result, the event sentence is manually labeled, in the labeling process, the non-trigger word is marked as untyped, and the trigger word is marked according to the news event category to obtain the event sentence sequence;
    步骤S13,通过开源工具包word2vec训练得到词向量,采用Skip-gram模型,根据训练得到的词向量,将事件句序列中每个词表述为一个300长度的向量;Step S13, the word vector is obtained through the open source toolkit word2vec training, and each word in the sequence of the event sentence is expressed as a vector of 300 length according to the word vector obtained by the training using the Skip-gram model;
    步骤S14,将每个事件句处理成词向量表述的序列形式,即每个候选触发词wi以一个300长度的词向量xi来表示,事件句表述为L={x1,x2,...,xi,...,xn}。Step S14, each event sentence is processed into a sequence form of a word vector representation, that is, each candidate trigger word w i is represented by a 300-length word vector x i , and the event sentence is expressed as L={x 1 , x 2 , ..., x i ,...,x n }.
  3. 根据权利要求1所述的基于神经网络的新闻事件抽取的方法,其特征在于:步骤S2具体为:The method for extracting news events based on a neural network according to claim 1, wherein the step S2 is specifically:
    步骤S21,假设事件句表述为L={x1,x2,...,xi,...,xn},其中xi为第i个候选触发词的词向量,n表示句子长度; Step S21, assuming that the event sentence is expressed as L={x 1 , x 2 , . . . , x i , . . . , x n }, where x i is the word vector of the ith candidate trigger word, and n is the length of the sentence. ;
    步骤S22,将L作为一个序列传入长短时记忆网络,得到序列的输出结果FW={fw1,fw2,...,fwi,...,fwn},其中,fwi表示第i个候选触发词经过长短时记忆网络提取出来的语义特征;Step S22, passing L as a sequence into the long and short time memory network, and obtaining the output result of the sequence FW={fw 1 , fw 2 , . . . , fw i , . . . , fw n }, where fw i represents the first The semantic features extracted by the i candidate trigger words through the long and short time memory network;
    步骤S23,将L进行反转即L′={xn,xn-1,...,xi,...,x1},将反向序列L′传入长短时记忆网络,得到反向序列的输出结果BW={bw1,bw2,...,bwi,...,bwn},其中,bwi表示第i个候选触发词经过反向长短时记忆网络提取出来的语义特征;In step S23, L is inverted, that is, L'={x n , x n-1 , . . . , x i , . . . , x 1 }, and the reverse sequence L′ is transmitted to the long and short time memory network. The output result of the reverse sequence BW={bw 1 , bw 2 , . . . , bw i , . . . , bw n }, where bw i indicates that the ith candidate trigger word is extracted by the memory network in the reverse length Semantic features;
    步骤S24,将双向长短时记忆网络获得输出结果的FW与BW进行拼接,得到句子L经过双向长短时记忆网络的输出结果,即O={r1,r2,...,ri,...,rn},其中,ri=[fwi:bwi]。Step S24, the FW and the BW obtained by the bidirectional long-term memory network are outputted, and the output result of the sentence L after the bidirectional length and length of the memory network is obtained, that is, O={r 1 , r 2 , . . . , r i ,. .., r n }, where r i =[fw i :bw i ].
  4. 根据权利要求1所述的基于神经网络的新闻事件抽取的方法,其特征在于:步骤S3具体为:The method for extracting news events based on a neural network according to claim 1, wherein the step S3 is specifically:
    步骤S31,假设事件句表述为L={x1,x2,...,xi,...,xn},其中xi为第i个单词的词向量,n表示句子长度;Step S31, assuming that the event sentence is expressed as L = {x 1 , x 2 , ..., x i , ..., x n }, where x i is the word vector of the ith word, and n represents the length of the sentence;
    步骤S32,对事件句采取卷积操作,计算公式为:In step S32, a convolution operation is performed on the event sentence, and the calculation formula is:
    Ci=f(wTxi:i-h+1+b)C i =f(w T x i:i-h+1 +b)
    其中,f是激活函数,Ci表示卷积得到的特征,w表示权重矩阵,h表示卷积核大小,i:i-h+1表示第i个词到第i-h+1个词,b表示偏置;Where f is the activation function, C i is the feature obtained by convolution, w is the weight matrix, h is the convolution kernel size, and i:i-h+1 is the i-th word to the i-h+1th word. b indicates the offset;
    通过滑动窗口,对所有词进行卷积得到特征图;By sliding the window, convolving all the words to obtain a feature map;
    步骤S33,使用最大池化,对特征图进行池化,得到事件句的全局特征CoIn step S33, using the maximum pooling, the feature map is pooled to obtain a global feature C o of the event sentence.
  5. 根据权利要求1所述的基于神经网络的新闻事件抽取的方法,其特征在于:步骤S4具体为:The method for extracting news events based on a neural network according to claim 1, wherein the step S4 is specifically:
    步骤S31,将双向长短时记忆网络获得的候选触发词语义特征O={r1,r2,...,ri,...,rn}与卷积神经网络提取出的句子全局特征Co进行级联,得到输出向量Ot=[O:Co]; Step S31, the candidate trigger term meanings O={r 1 , r 2 , . . . , r i , . . . , r n } obtained by the two-way long-term memory network and the global features of the sentence extracted by the convolutional neural network C o is cascaded to obtain an output vector O t =[O:C o ];
    步骤S32,使用softmax对输出向量Ot进行分类,获得新闻事件预测的类型。In step S32, the output vector O t is classified using softmax to obtain the type of news event prediction.
  6. 一种基于神经网络的新闻事件抽取的系统,其特征在于:包括文本与处理模块、神经网络训练模块、新闻事件预测模块,其中:A system for news event extraction based on neural network, comprising: a text and processing module, a neural network training module, and a news event prediction module, wherein:
    所述文本与处理模块用于对训练语料原始文本进行数据预处理,包括:对训练语料原始文本进行分句,得到事件句,然后对事件句进行分词、命名体识别;根据人工标注的新闻事件信息,将事件句进行序列标注,触发词根据其类型进行标注,非触发词标注为无类别,得到事件句序列;并将事件句序列以词向量的形式进行表述;The text and processing module is configured to perform data preprocessing on the original text of the training corpus, including: segmenting the original text of the training corpus, obtaining an event sentence, and then performing segmentation and naming recognition on the event sentence; and manually marking the news event according to the manual The information is sequenced by the event sentence, the trigger word is marked according to its type, the non-trigger word is marked as no category, and the sequence of the event sentence is obtained; and the sequence of the event sentence is expressed in the form of a word vector;
    所述神经网络训练模块包括双向长短时记忆网络训练模块和卷积神经网络训练模块,双向长短时记忆网络训练模块用于将以词向量表示的事件句序列训练,得到每个候选触发词的语义特征;卷积神经网络训练模块用于将以词向量表示的事件句序列训练,得到候选触发词所在事件句的全局特征;The neural network training module comprises a bidirectional long and short time memory network training module and a convolutional neural network training module, and the bidirectional long and short time memory network training module is configured to train the event sentence sequence represented by the word vector to obtain the semantics of each candidate trigger word. The convolutional neural network training module is configured to train the sequence of event sentences represented by the word vector to obtain a global feature of the event sentence in which the candidate trigger word is located;
    所述新闻事件预测模块用于根据神经网络训练模块得到的候选触发词的语义特征与候选触发词所在句子的全局特征,利用softmax作为分类器来对每个候选触发词进行分类,从而找出新闻事件的触发词,并根据触发词类型,判断该事件所属类型。 The news event prediction module is configured to classify each candidate trigger word by using softmax as a classifier according to the semantic feature of the candidate trigger word obtained by the neural network training module and the global feature of the sentence where the candidate trigger word is located, thereby finding the news. The trigger word of the event, and according to the type of the trigger word, determine the type of the event.
PCT/CN2017/089136 2017-05-27 2017-06-20 Method and system for extracting news event based on neural network WO2018218706A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710391227.7 2017-05-27
CN201710391227.7A CN107239445A (en) 2017-05-27 2017-05-27 The method and system that a kind of media event based on neutral net is extracted

Publications (1)

Publication Number Publication Date
WO2018218706A1 true WO2018218706A1 (en) 2018-12-06

Family

ID=59984603

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/089136 WO2018218706A1 (en) 2017-05-27 2017-06-20 Method and system for extracting news event based on neural network

Country Status (2)

Country Link
CN (1) CN107239445A (en)
WO (1) WO2018218706A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111382273A (en) * 2020-03-09 2020-07-07 西安理工大学 Text classification method based on feature selection of attraction factors
CN111460830A (en) * 2020-03-11 2020-07-28 北京交通大学 Method and system for extracting economic events in judicial texts
CN112905742A (en) * 2021-02-20 2021-06-04 厦门吉比特网络技术股份有限公司 Method, device, system and storage medium for recognizing new vocabulary based on semantic model neural network

Families Citing this family (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108415923B (en) * 2017-10-18 2020-12-11 北京邮电大学 Intelligent man-machine conversation system of closed domain
CN109947932B (en) * 2017-10-27 2021-06-04 中移(苏州)软件技术有限公司 Push information classification method and system
CN108021616B (en) * 2017-11-06 2020-08-14 大连理工大学 Community question-answer expert recommendation method based on recurrent neural network
CN107943788B (en) * 2017-11-17 2021-04-06 平安科技(深圳)有限公司 Enterprise abbreviation generation method and device and storage medium
CN110019807B (en) * 2017-12-27 2020-11-13 航天信息股份有限公司 Commodity classification method and device
CN109977393B (en) * 2017-12-28 2021-09-03 中国科学院计算技术研究所 Popular news prediction method and system based on content disputeness
CN108304530B (en) * 2018-01-26 2022-03-18 腾讯科技(深圳)有限公司 Knowledge base entry classification method and device and model training method and device
CN108446355B (en) * 2018-03-12 2022-05-20 深圳证券信息有限公司 Investment and financing event element extraction method, device and equipment
CN108829801B (en) * 2018-06-06 2020-11-20 大连理工大学 Event trigger word extraction method based on document level attention mechanism
CN110633409B (en) * 2018-06-20 2023-06-09 上海财经大学 Automobile news event extraction method integrating rules and deep learning
CN110209807A (en) * 2018-07-03 2019-09-06 腾讯科技(深圳)有限公司 A kind of method of event recognition, the method for model training, equipment and storage medium
CN109192300A (en) * 2018-08-17 2019-01-11 百度在线网络技术(北京)有限公司 Intelligent way of inquisition, system, computer equipment and storage medium
CN109325116B (en) * 2018-08-23 2020-11-27 武大吉奥信息技术有限公司 Urban event automatic classification dispatching method and device based on deep learning
CN109446299B (en) * 2018-08-27 2022-08-16 中国科学院信息工程研究所 Method and system for searching e-mail content based on event recognition
CN109522406A (en) * 2018-10-11 2019-03-26 平安科技(深圳)有限公司 Text semantic matching process, device, computer equipment and storage medium
CN109282837B (en) * 2018-10-24 2021-06-22 福州大学 Demodulation method of Bragg fiber grating staggered spectrum based on LSTM network
CN109299470B (en) * 2018-11-01 2024-02-09 成都数联铭品科技有限公司 Method and system for extracting trigger words in text bulletin
CN109446326B (en) * 2018-11-01 2021-04-20 大连理工大学 Biomedical event combined extraction method based on replication mechanism
CN109670168B (en) * 2018-11-14 2023-04-18 华南师范大学 Short answer automatic scoring method, system and storage medium based on feature learning
CN109558591A (en) * 2018-11-28 2019-04-02 中国科学院软件研究所 Chinese event detection method and device
CN109657058A (en) * 2018-11-29 2019-04-19 东莞理工学院 A kind of abstracting method of notice information
CN109670172A (en) * 2018-12-06 2019-04-23 桂林电子科技大学 A kind of scenic spot anomalous event abstracting method based on complex neural network
CN109948646A (en) * 2019-01-24 2019-06-28 西安交通大学 A kind of time series data method for measuring similarity and gauging system
CN110032641B (en) * 2019-02-14 2024-02-13 创新先进技术有限公司 Method and device for extracting event by using neural network and executed by computer
CN110276068B (en) * 2019-05-08 2020-08-28 清华大学 Legal case analysis method and device
CN110263323B (en) * 2019-05-08 2020-08-28 清华大学 Keyword extraction method and system based on barrier type long-time memory neural network
CN110134720B (en) * 2019-05-17 2021-02-09 苏州大学 Event joint extraction method integrating local features and deep learning
CN110210019A (en) * 2019-05-21 2019-09-06 四川大学 A kind of event argument abstracting method based on recurrent neural network
CN110287491B (en) * 2019-06-25 2024-01-12 北京百度网讯科技有限公司 Event name generation method and device
CN110334213B (en) * 2019-07-09 2021-05-11 昆明理工大学 Method for identifying time sequence relation of Hanyue news events based on bidirectional cross attention mechanism
CN110377738A (en) * 2019-07-15 2019-10-25 昆明理工大学 Merge the Vietnamese news event detecting method of interdependent syntactic information and convolutional neural networks
CN110704598B (en) * 2019-09-29 2023-01-17 北京明略软件系统有限公司 Statement information extraction method, extraction device and readable storage medium
CN110674303B (en) * 2019-09-30 2022-03-01 北京明略软件系统有限公司 Event statement processing method and device, computer equipment and readable storage medium
CN111125520B (en) * 2019-12-11 2023-04-21 东南大学 Event line extraction method based on deep clustering model for news text
CN111611218A (en) * 2020-04-24 2020-09-01 武汉大学 Distributed abnormal log automatic identification method based on deep learning
CN111967268B (en) * 2020-06-30 2024-03-19 北京百度网讯科技有限公司 Event extraction method and device in text, electronic equipment and storage medium
CN111950199A (en) * 2020-08-11 2020-11-17 杭州叙简科技股份有限公司 Earthquake data structured automation method based on earthquake news event
CN112163416B (en) * 2020-10-09 2021-11-02 北京理工大学 Event joint extraction method for merging syntactic and entity relation graph convolution network
CN112507680B (en) * 2020-11-13 2023-04-07 北京航空航天大学 Traffic operation information extraction and situation early warning method and device
CN112966525B (en) * 2021-03-31 2023-02-10 上海大学 Law field event extraction method based on pre-training model and convolutional neural network algorithm
CN113792545B (en) * 2021-11-16 2022-03-04 成都索贝数码科技股份有限公司 News event activity name extraction method based on deep learning
CN115292568B (en) * 2022-03-02 2023-11-17 内蒙古工业大学 Civil news event extraction method based on joint model

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160099010A1 (en) * 2014-10-03 2016-04-07 Google Inc. Convolutional, long short-term memory, fully connected deep neural networks
CN106096568A (en) * 2016-06-21 2016-11-09 同济大学 A kind of pedestrian's recognition methods again based on CNN and convolution LSTM network
CN106528528A (en) * 2016-10-18 2017-03-22 哈尔滨工业大学深圳研究生院 A text emotion analysis method and device
CN106569998A (en) * 2016-10-27 2017-04-19 浙江大学 Text named entity recognition method based on Bi-LSTM, CNN and CRF
CN106599198A (en) * 2016-12-14 2017-04-26 广东顺德中山大学卡内基梅隆大学国际联合研究院 Image description method for multi-stage connection recurrent neural network

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106599933A (en) * 2016-12-26 2017-04-26 哈尔滨工业大学 Text emotion classification method based on the joint deep learning model
CN106611055A (en) * 2016-12-27 2017-05-03 大连理工大学 Chinese hedge scope detection method based on stacked neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160099010A1 (en) * 2014-10-03 2016-04-07 Google Inc. Convolutional, long short-term memory, fully connected deep neural networks
CN106096568A (en) * 2016-06-21 2016-11-09 同济大学 A kind of pedestrian's recognition methods again based on CNN and convolution LSTM network
CN106528528A (en) * 2016-10-18 2017-03-22 哈尔滨工业大学深圳研究生院 A text emotion analysis method and device
CN106569998A (en) * 2016-10-27 2017-04-19 浙江大学 Text named entity recognition method based on Bi-LSTM, CNN and CRF
CN106599198A (en) * 2016-12-14 2017-04-26 广东顺德中山大学卡内基梅隆大学国际联合研究院 Image description method for multi-stage connection recurrent neural network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
FENG, XIAOCHENG ET AL.: "A Language-independent Neural Network for Event Detection", PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 12 August 2016 (2016-08-12), pages 67, 68, XP055548695 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111382273A (en) * 2020-03-09 2020-07-07 西安理工大学 Text classification method based on feature selection of attraction factors
CN111382273B (en) * 2020-03-09 2023-04-14 广州智赢万世市场管理有限公司 Text classification method based on feature selection of attraction factors
CN111460830A (en) * 2020-03-11 2020-07-28 北京交通大学 Method and system for extracting economic events in judicial texts
CN111460830B (en) * 2020-03-11 2022-04-12 北京交通大学 Method and system for extracting economic events in judicial texts
CN112905742A (en) * 2021-02-20 2021-06-04 厦门吉比特网络技术股份有限公司 Method, device, system and storage medium for recognizing new vocabulary based on semantic model neural network
CN112905742B (en) * 2021-02-20 2022-07-29 厦门吉比特网络技术股份有限公司 Method and device for recognizing new vocabulary based on semantic model neural network

Also Published As

Publication number Publication date
CN107239445A (en) 2017-10-10

Similar Documents

Publication Publication Date Title
WO2018218706A1 (en) Method and system for extracting news event based on neural network
CN108920622B (en) Training method, training device and recognition device for intention recognition
CN108628828B (en) Combined extraction method based on self-attention viewpoint and holder thereof
CN111274394B (en) Method, device and equipment for extracting entity relationship and storage medium
Du et al. Text classification research with attention-based recurrent neural networks
CN107193959B (en) Pure text-oriented enterprise entity classification method
US20210200961A1 (en) Context-based multi-turn dialogue method and storage medium
CN109858041B (en) Named entity recognition method combining semi-supervised learning with user-defined dictionary
WO2018028077A1 (en) Deep learning based method and device for chinese semantics analysis
CN113011533A (en) Text classification method and device, computer equipment and storage medium
CN110175246B (en) Method for extracting concept words from video subtitles
CN110807084A (en) Attention mechanism-based patent term relationship extraction method for Bi-LSTM and keyword strategy
CN108509521B (en) Image retrieval method for automatically generating text index
CN113505200B (en) Sentence-level Chinese event detection method combined with document key information
CN110134950B (en) Automatic text proofreading method combining words
CN113157916A (en) Civil aviation emergency extraction method based on deep learning
CN112395421B (en) Course label generation method and device, computer equipment and medium
CN109271624B (en) Target word determination method, device and storage medium
CN111027309A (en) Method for extracting entity attribute value based on bidirectional long-short term memory network
CN112257444B (en) Financial information negative entity discovery method, device, electronic equipment and storage medium
CN114417851A (en) Emotion analysis method based on keyword weighted information
CN113742733A (en) Reading understanding vulnerability event trigger word extraction and vulnerability type identification method and device
CN115859980A (en) Semi-supervised named entity identification method, system and electronic equipment
WO2023178802A1 (en) Named entity recognition method and apparatus, device, and computer readable storage medium
CN114756675A (en) Text classification method, related equipment and readable storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17911709

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17911709

Country of ref document: EP

Kind code of ref document: A1