CN106570148A - Convolutional neutral network-based attribute extraction method - Google Patents

Convolutional neutral network-based attribute extraction method Download PDF

Info

Publication number
CN106570148A
CN106570148A CN201610968810.5A CN201610968810A CN106570148A CN 106570148 A CN106570148 A CN 106570148A CN 201610968810 A CN201610968810 A CN 201610968810A CN 106570148 A CN106570148 A CN 106570148A
Authority
CN
China
Prior art keywords
vector
word
sentence
sequence
convolution
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610968810.5A
Other languages
Chinese (zh)
Other versions
CN106570148B (en
Inventor
汤斯亮
吴飞
张金剑
蒋焕剑
庄越挺
鲁伟明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201610968810.5A priority Critical patent/CN106570148B/en
Publication of CN106570148A publication Critical patent/CN106570148A/en
Application granted granted Critical
Publication of CN106570148B publication Critical patent/CN106570148B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification

Abstract

The invention discloses a convolutional neutral network-based attribute extraction method. The method comprises the following steps of (1) constructing an external knowledge library; (2) obtaining text data; (3) obtaining attribute-containing sentences by using a remote supervision method; (4) obtaining the sentences by utilizing a word vector method and performing vectorization; and (5) inputting the sentences to a convolutional neutral network, and performing training and classification. According to the method, the attribute-containing candidate sentences are extracted from a non-structured text data set based on artificially defined mapping by utilizing the external knowledge library in combination with remote supervision and convolutional neutral network models, and sentence classifications are classified in combination with the convolutional neutral network model, so that an attribute extraction task is finished.

Description

A kind of attribute extraction method based on convolutional neural networks
Technical field
The present invention relates to Text character extraction and attribute extraction, more particularly to a kind of attribute based on convolutional neural networks is taken out Take method.
Background technology
The world today is in the epoch of an information huge explosion, and the prevalence and high speed development of internet generate magnanimity Information resources.These resources are great for the meaning of development in science and technology, and scientific circles need the base for therefrom extracting scientific research This material, industrial quarters needs therefrom to excavate potential business opportunity, therefore how to be near using the information resources of these internets One of main flow direction of Technological research over year.
Although the information resources quantity in internet is huge, these resources often lack structurized characteristic. Structural data refers to row data, the data that can be expressed with bivariate table structure, and unstructured data[2]Field it is long Degree is variable, it has not been convenient to expressed using two-dimentional logical table.Because these resources are much destructurings or semi-structured Data, thus quickly and efficiently search and understand that these data are limited by very large.
Text data is the pith in internet information resource, and the most of text data on internet is also non-knot The data of structure, such as news, blog, Email, governmental documents, chat record, system journal etc..In order to efficiently Using these unstructured text datas, information extraction (Information Extraction) technology is arisen at the historic moment-automatically will Destructuring or semi-structured text in input page changes into structurized data.Information extraction task is by input Define with the target for extracting, input can be the non-structured document write using natural language, or half on webpage Structurized document;And extraction target is the relation of k- tuples (k is the number of attributes of a record) or the layer of a complexity The data object of secondaryization.
Traditional attribute extraction technology has many drawbacks, whether rule-based first to be also based on sorting algorithm, all More manpower intervention is needed, such as rule-based rule design and the data based on classification are marked and characteristic Design, and The cost of manpower intervention is expensive, and needs the mark of professional just to obtain more authoritative artificial data, while people Work also brings along certain error, and error also can be accumulated constantly in follow-up algorithm, and the deviation for ultimately resulting in result is excessive; Secondly, the training dataset of this class algorithm is limited in certain field, i.e., do not have versatility, such as train with regard to body Educating the attribute extraction grader in terms of news cannot use well in other news;The effect one that generally above-mentioned algorithm is obtained As it is also not ideal enough because the rule of engineer is limited in rule-based method, and be based on the mark of the method for classification Note data are also limited and the method compares the quality of the feature for relying on engineer.
The content of the invention
The purpose of the present invention is to overcome the deficiencies in the prior art, there is provided a kind of feature extraction side based on convolutional neural networks Method.
Based on the attribute extraction method of convolutional neural networks, comprise the steps
1) the information frame data of Wikipedia is obtained, external knowledge storehouse is obtained;
2) forum and news data are obtained, obtains text corpus;
3) using remote supervisory method, text corpus are scanned for, obtains the sentence comprising attribute;
4) based on term vector, vectorization description is first carried out to each word, then the term vector in sentence is flocked together, obtained Vectorization to sentence is described;
5) by sentence inputting convolutional neural networks, convolutional calculation is first carried out to sentence, then for the maximum of convolution results Pond is carried out, softmax functions are recently entered, classification results are obtained;
6) according to classification results correspondence to property value.
Each step can be using following preferred implementation in the present invention:
Described step 4) it is specific as follows:
4.1) the term vector model G trained in news data is obtained, is preserved by sequence form:
[word1 vector1 word2 vector2...wordN vectorN]
Wherein N is the word number in term vector model, wordNFor n-th word, vectorNFor n-th term vector;
4.2) file is read, sequence is switched to mapping format, its corresponding vector is mapped to by word:
{word1:vector1,word2:vector2...wordN:vectorN}
4.3) sentence is read, and carries out participle, the sequence that sentence is switched to be composed of words:
[wordi1,wordi2...wordil]
Wherein i represents i-th sentence, and l represents the word number in sentence, wordilRepresent l-th word of i-th sentence;
4.4) sequence is read, and inquires about mapping, the word in sequence is switched to into term vector:
[vectori1,vectori2...vectoril]
Wherein vectorilRepresent l-th term vector of i-th sentence.
Described step 5) it is specific as follows:
5.1) term vector in sequence is switched to into column vector:
X=[x1,x2,...,xl]
Wherein xlRepresent the column vector of the term vector of l-th word in sentence;
5.2) convolution operation is carried out to sentence, obtains the sequence after convolution:
S=[s1,s2,...,sl-ω+1]
Wherein the window size of convolutional layer is ω, and a convolution kernel is f=[f1,f2,...,fω], fiFor size and word The equally big column vector of vector;
5.3) sequence after convolution is calculated into convolution results sequence s after activation by ReLU activation primitivesi
Wherein, b is bias term, and g is ReLU activation primitives.
5.4) repeatedly 5.1)~5.3), computed repeatedly using the convolution kernel of different convolution kernels and different windows size, Final each convolution kernel obtains a convolution results sequence, for the original word vector sum position vector of abstract expression.
5.5) for result sequence s that each convolution kernel f is obtainedi, pond process is carried out using max functions, select this As a result the maximum in sequence all values is used as new result feature pf
pf=max { s }=max { s1,s2,...,sl-ω+1}
Same pond process is all carried out to all of convolution kernel, the equal spy of the quantity of a length and convolution kernel is obtained Vector is levied, using characteristic vector as the abstract sentence characteristics for obtaining of whole convolutional neural networks;
5.6) pond result is input into into softmax functions, obtains the classification results output (x of jth classj):
Wherein K represents classification number.
Various drawbacks in conventional method, the present invention proposes mark more than a kind of many examples based on convolutional neural networks The attribute extraction algorithm of label, using the method for Distant Supervision existing knowledge base is utilized, and automatically generates training number According to, and using some optimization method cleaning training datas, so as to save the tedious work of artificial mark.The present invention also uses convolution The method of neutral net while saving manual working, extracts more abstract and more automatically extracting the feature of text sentence The feature of tool expressiveness.Finally using the model of many example multi-tags, to solve and there may be various relations between two entities Problem.The square algorithm of some main flows of the method in effect better than traditional attribute extraction algorithm and in recent years.
Description of the drawings
Fig. 1 is pseudo- graph expression of the core used herein using convolutional neural networks model.Left side shows respectively in figure Two positive samples and negative sample, positive and negative sample standard deviation extracted using remote supervisory method.Mapped according to word and vector sum Relation, extracts the vector statement of sentence, after the operation of convolution pondization is carried out, then is input into softmax functions.
Fig. 2 is that Medical Data collection describes fragment with regard to the text of description medicine and side effect, is respectively medical side effect Positive class label and negative class label candidate sentence.
Specific embodiment
The present invention is further elaborated with reference to the accompanying drawings and detailed description.
Comprised the steps based on the attribute extraction method of convolutional neural networks:
1) the information frame data of Wikipedia is obtained, external knowledge storehouse is obtained.Comprise the following steps that:
1.1. the public data of wikipedia is downloaded
2.1. the data frame of entry in wikipedia is extracted, and data frame name is mapped to Property Name, and store attribute Value and entry name.
3.1. in all entries and all data frames, attribute-name identical property value and entry name are saved together.
2) forum and news data are obtained, obtains text corpus.Comprise the following steps that:
2.1. news data, the such as public data of the New York Times are downloaded.
2.2. text data is pre-processed, the labels such as HTML or XML is removed, character encoding format switchs to utf-8, Form switchs to plain text data.
2.3. natural language processing instrument, such as Stanford Core NLP instruments is used to carry out to plain text data point Word, and extract name entity information.
The public data of wikipedia and news data directly adopt the data sets of TAC-KBP 2015 in the present invention.
3) using remote supervisory method, text corpus are scanned for, obtains the sentence comprising attribute.Concrete steps are such as Under:
3.1. positive sample is built.Under same attribute-name, if physical name and property value are occurred in a certain sentence simultaneously, Then this sentence is labeled as positive sample.
3.2. negative sample is built.Under same attribute-name, if physical name is occurred in a certain sentence, property value does not go out Now in the sentence, but name entity information of the sentence comprising property value.Then the sentence is labeled as negative sample.
3.3. stochastical sampling is carried out to negative sample, makes negative sample number and positive sample number roughly equal.
4) based on term vector, vectorization description is first carried out to each word, then the term vector in sentence is flocked together, obtained Vectorization to sentence is described.Comprise the following steps that:
4.1) obtain Google and be disclosed in the term vector model G trained in news data, preserved by sequence form:
[word1 vector1 word2 vector2...wordN vectorN]
Wherein N is the word number in term vector model, wordNFor n-th word, vectorNFor n-th term vector;
4.2) file is read, sequence is switched to mapping format, its corresponding vector is mapped to by word:
{word1:vector1,word2:vector2...wordN:vectorN}
4.3) sentence is read, and carries out participle, the sequence that sentence is switched to be composed of words:
[wordi1,wordi2...wordil]
Wherein i represents i-th sentence, and l represents the word number in sentence, wordilRepresent l-th word of i-th sentence;
4.4) sequence is read, and inquires about mapping, the word in sequence is switched to into term vector:
[vectori1,vectori2...vectoril]
Wherein vectorilRepresent l-th term vector of i-th sentence.
5) by sentence inputting convolutional neural networks, convolutional calculation is first carried out to sentence, then for the maximum of convolution results Pond is carried out, softmax functions are recently entered, classification results are obtained.Comprise the following steps that:
5.1) term vector in sequence is switched to into column vector:
X=[x1,x2,...,xl]
Wherein xlRepresent the column vector of the term vector of l-th word in sentence;
5.2) convolution operation is carried out to sentence, obtains the sequence after convolution:
S=[s1,s2,...,sl-ω+1]
Wherein the window size of convolutional layer is ω, and a convolution kernel is f=[f1,f2,...,fω], fiFor size and word The equally big column vector of vector;
5.3) sequence after convolution is calculated into convolution results sequence s after activation by ReLU activation primitivesi
Wherein, b is bias term, and g is ReLU activation primitives.
5.4) repeatedly 5.1)~5.3), computed repeatedly using the convolution kernel of different convolution kernels and different windows size, Final each convolution kernel obtains a convolution results sequence, for the original word vector sum position vector of abstract expression.
5.5) for result sequence s that each convolution kernel f is obtainedi, pond process is carried out using max functions, select this As a result the maximum in sequence all values is used as new result feature pf
pf=max { s }=max { s1,s2,...,sl-ω+1}
Same pond process is all carried out to all of convolution kernel, the equal spy of the quantity of a length and convolution kernel is obtained Vector is levied, using characteristic vector as the abstract sentence characteristics for obtaining of whole convolutional neural networks;
5.6) pond result is input into into softmax functions, obtains the classification results output (x of jth classj):
Wherein K represents classification number.
6) according to classification results correspondence to property value:According to the sentence classification for getting, entity and property value pair are got, Obtain entity attributes value.
Embodiment 1
The present embodiment carries out attribute extraction by taking one section of newsletter archive that user submits to as an example using said method, each to implement Specific parameter and way are as follows in step:
1. the sentence of pair input, searches for whether contain entity, and the sentence containing entity is constituted into original statement set
{sentence1,sentence2,...sentencesN}
2. in original statement set, whether search contains property value, and the sentence containing property value is constituted into candidate sentences Set (as shown in Figure 2), sentence is simultaneously comprising entity and property value in Candidate Set.
{candidate1,candidate2,...candidateN1}
3. record in Candidate Set in candidate sentences, entity and property value pair, with sequence form storage, and sentence in Candidate Set Order it is corresponding.
{(entity1,slot filler 1),(entity2,slot filler 2),...(entityN1,slot filler N1)}
4. the term vector model of Google is downloaded, sequence form data are read.
[word1 vector1 word2 vector2...wordN vectorN]
5. sequence is switched to mapping format, its corresponding vector is mapped to by word.
{word1:vector1,word2:vector2...wordN:vectorN}
6. sentence carries out participle in pair Candidate Set, switchs to the sequence of word.
[wordi1,wordi2...wordis]
7. sentence in pair Candidate Set, reads the mapped file that Google provides term vector model, obtains the vectorization of sentence Statement.
[vectori1,vectori2...vectoris]
8., for (as shown in Figure 1) is stated in the vectorization of sentence, each term vector is switched to into the form of column vector, this When, sentence switchs to two-dimensional matrix form.
X=[x1,x2,...,xl]
9. pair sentence carries out the convolution operation that convolution kernel is 3, while adding biasing.
10. convolution kernel output result is input into into activation primitive, using ReLU activation primitives.
11. reuse convolution kernel 4,5, repeat 9 and operation 10, and input value is cascaded.
S=[s1,s2,...,sl-ω+1,s21,s22,...,sl2-ω2+1,s31,s32,...,sl3-ω3+1]
12. enter pond layer, the output of convolutional layer are operated, using maximum pond.
pf=max { s }=max { s1,s2,...,sl-ω+1}
13. as shown in Fig. 2 by the output result input softmax functions of pond layer, obtain classification results.
15. according to output result, according to sentence order, finds corresponding entity attribute pair, has also just obtained attribute extraction Result.
(entity1,slot filler 1)
As shown in table 1, on the data sets of TAC-KBP 2015, method and the main stream approach for pre-existing described in the present invention Comparing result shows that the present invention is respectively provided with obvious advantage on Precision, Recall and F1-Score evaluation criterions.
Table 1
Model Precision Recall F1-Score
LR-SF 0.4483 0.3652 0.4025
MIML-SF 0.5412 0.3893 0.4529
CNN-SF 0.5657 0.4067 0.4732
This paper models 0.6343 0.4136 0.5007

Claims (3)

1. a kind of attribute extraction method based on convolutional neural networks, it is characterised in that comprise the steps
1) the information frame data of Wikipedia is obtained, external knowledge storehouse is obtained;
2) forum and news data are obtained, obtains text corpus;
3) using remote supervisory method, text corpus are scanned for, obtains the sentence comprising attribute;
4) based on term vector, vectorization description is first carried out to each word, then the term vector in sentence is flocked together, obtain sentence The vectorization description of son;
5) by sentence inputting convolutional neural networks, first carry out convolutional calculation to sentence, then carry out for the maximum of convolution results Chi Hua, recently enters softmax functions, obtains classification results;
6) according to classification results correspondence to property value.
2. a kind of attribute extraction method based on convolutional neural networks according to described by claim 1, it is characterised in that described The step of 4) it is specific as follows:
4.1) the term vector model G trained in news data is obtained, is preserved by sequence form:
[word1 vector1 word2 vector2 ... wordN vectorN]
Wherein N is the word number in term vector model, wordNFor n-th word, vectorNFor n-th term vector;
4.2) file is read, sequence is switched to mapping format, its corresponding vector is mapped to by word:
{word1:vector1,word2:vector2...wordN:vectorN}
4.3) sentence is read, and carries out participle, the sequence that sentence is switched to be composed of words:
[wordi1,wordi2...wordil]
Wherein i represents i-th sentence, and l represents the word number in sentence, wordilRepresent l-th word of i-th sentence;
4.4) sequence is read, and inquires about mapping, the word in sequence is switched to into term vector:
[vectori1,vectori2...vectoril]
Wherein vectorilRepresent l-th term vector of i-th sentence.
3. a kind of attribute extraction method based on convolutional neural networks according to described by claim 1, it is characterised in that described The step of 5) it is specific as follows:
5.1) term vector in sequence is switched to into column vector:
X=[x1,x2,...,xl]
Wherein xlRepresent the column vector of the term vector of l-th word in sentence;
5.2) convolution operation is carried out to sentence, obtains the sequence after convolution:
S=[s1,s2,...,sl-ω+1]
Wherein the window size of convolutional layer is ω, and a convolution kernel is f=[f1,f2,...,fω], fiFor size and word vector Equally big column vector;
5.3) sequence after convolution is calculated into convolution results sequence s after activation by ReLU activation primitivesi
s i = g ( Σ j = 0 ω - 1 f j + 1 T x j + 1 + b )
Wherein, b is bias term, and g is ReLU activation primitives.
5.4) repeatedly 5.1)~5.3), computed repeatedly using the convolution kernel of different convolution kernels and different windows size, finally Each convolution kernel obtains a convolution results sequence, for the original word vector sum position vector of abstract expression.
5.5) for result sequence s that each convolution kernel f is obtainedi, pond process is carried out using max functions, select the result sequence Maximum in row all values is used as new result feature pf
pf=max { s }=max { s1,s2,...,sl-ω+1}
Same pond process is all carried out to all of convolution kernel, obtain the equal feature of quantity of a length and convolution kernel to Amount, using characteristic vector as the abstract sentence characteristics for obtaining of whole convolutional neural networks;
5.6) pond result is input into into softmax functions, obtains the classification results output (x of jth classj):
o u t p u t ( x j ) = e p j Σe p j , j = 1 ... , K
Wherein K represents classification number.
CN201610968810.5A 2016-10-27 2016-10-27 A kind of attribute extraction method based on convolutional neural networks Active CN106570148B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610968810.5A CN106570148B (en) 2016-10-27 2016-10-27 A kind of attribute extraction method based on convolutional neural networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610968810.5A CN106570148B (en) 2016-10-27 2016-10-27 A kind of attribute extraction method based on convolutional neural networks

Publications (2)

Publication Number Publication Date
CN106570148A true CN106570148A (en) 2017-04-19
CN106570148B CN106570148B (en) 2019-07-23

Family

ID=58541372

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610968810.5A Active CN106570148B (en) 2016-10-27 2016-10-27 A kind of attribute extraction method based on convolutional neural networks

Country Status (1)

Country Link
CN (1) CN106570148B (en)

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107180247A (en) * 2017-05-19 2017-09-19 中国人民解放军国防科学技术大学 Relation grader and its method based on selective attention convolutional neural networks
CN107220237A (en) * 2017-05-24 2017-09-29 南京大学 A kind of method of business entity's Relation extraction based on convolutional neural networks
CN107301170A (en) * 2017-06-19 2017-10-27 北京百度网讯科技有限公司 The method and apparatus of cutting sentence based on artificial intelligence
CN107301167A (en) * 2017-05-25 2017-10-27 中国科学院信息工程研究所 A kind of work(performance description information recognition methods and device
CN107506350A (en) * 2017-08-16 2017-12-22 京东方科技集团股份有限公司 A kind of method and apparatus of identification information
CN107590133A (en) * 2017-10-24 2018-01-16 武汉理工大学 The method and system that position vacant based on semanteme matches with job seeker resume
CN107943967A (en) * 2017-11-28 2018-04-20 华南理工大学 Algorithm of documents categorization based on multi-angle convolutional neural networks and Recognition with Recurrent Neural Network
CN108009284A (en) * 2017-12-22 2018-05-08 重庆邮电大学 Using the Law Text sorting technique of semi-supervised convolutional neural networks
CN108038543A (en) * 2017-10-24 2018-05-15 华南师范大学 It is expected and anti-desired depth learning method and nerve network system
CN108108354A (en) * 2017-06-18 2018-06-01 北京理工大学 A kind of microblog users gender prediction's method based on deep learning
CN108154234A (en) * 2017-12-04 2018-06-12 盈盛资讯科技有限公司 A kind of knowledge learning method and system based on template
CN108256583A (en) * 2018-01-25 2018-07-06 北京东方科诺科技发展有限公司 A kind of multi-tag classification learning method based on coupling learning
CN108280055A (en) * 2017-12-04 2018-07-13 盈盛资讯科技有限公司 A kind of knowledge learning method and system based on binary crelation
CN108287911A (en) * 2018-02-01 2018-07-17 浙江大学 A kind of Relation extraction method based on about fasciculation remote supervisory
CN108304530A (en) * 2018-01-26 2018-07-20 腾讯科技(深圳)有限公司 Knowledge base entry sorting technique and device, model training method and device
CN108520042A (en) * 2018-04-03 2018-09-11 公安部第三研究所 Realize the system and method through detecing case-involving the role's calibration and role's assessment of suspect in work
CN108664474A (en) * 2018-05-21 2018-10-16 众安信息技术服务有限公司 A kind of resume analytic method based on deep learning
CN108763353A (en) * 2018-05-14 2018-11-06 中山大学 Rule-based and remote supervisory Baidupedia relationship triple abstracting method
CN109063759A (en) * 2018-07-20 2018-12-21 浙江大学 A kind of neural network structure searching method applied to the more attribute forecasts of picture
CN109308304A (en) * 2018-09-18 2019-02-05 深圳和而泰数据资源与云技术有限公司 Information extraction method and device
CN109446299A (en) * 2018-08-27 2019-03-08 中国科学院信息工程研究所 The method and system of searching email content based on event recognition
CN109492230A (en) * 2019-01-11 2019-03-19 浙江大学城市学院 A method of insurance contract key message is extracted based on textview field convolutional neural networks interested
CN109657207A (en) * 2018-11-29 2019-04-19 爱保科技(横琴)有限公司 The formatting processing method and processing unit of clause
CN109670542A (en) * 2018-12-11 2019-04-23 田刚 A kind of false comment detection method based on comment external information
CN109815338A (en) * 2018-12-28 2019-05-28 北京市遥感信息研究所 Relation extraction method and system in knowledge mapping based on mixed Gauss model
CN109885677A (en) * 2018-12-26 2019-06-14 中译语通科技股份有限公司 A kind of multi-faceted big data acquisition clearing system and method
CN110516239A (en) * 2019-08-26 2019-11-29 贵州大学 A kind of segmentation pond Relation extraction method based on convolutional neural networks
CN110717047A (en) * 2019-10-22 2020-01-21 湖南科技大学 Web service classification method based on graph convolution neural network
CN111078895A (en) * 2019-12-18 2020-04-28 江南大学 Remote supervision entity relation extraction method based on denoising convolutional neural network
CN112445955A (en) * 2019-08-30 2021-03-05 珠海格力电器股份有限公司 Business opportunity information management method, system and storage medium
CN113535820A (en) * 2021-07-20 2021-10-22 贵州电网有限责任公司 Electrical operating personnel attribute presumption method based on convolutional neural network

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016054301A1 (en) * 2014-10-02 2016-04-07 Microsoft Technology Licensing, Llc Distant supervision relationship extractor
CN106055675A (en) * 2016-06-06 2016-10-26 杭州量知数据科技有限公司 Relation extracting method based on convolution neural network and distance supervision

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016054301A1 (en) * 2014-10-02 2016-04-07 Microsoft Technology Licensing, Llc Distant supervision relationship extractor
CN106055675A (en) * 2016-06-06 2016-10-26 杭州量知数据科技有限公司 Relation extracting method based on convolution neural network and distance supervision

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ZHEN WANG等: "Distantly Supervised Neural Network Model", 《14TH CHINA NATIONAL CONFERENCE, CCL 2015 AND THIRD INTERNATIONAL SYMPOSIUM, NLP-NABD 2015》 *

Cited By (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107180247A (en) * 2017-05-19 2017-09-19 中国人民解放军国防科学技术大学 Relation grader and its method based on selective attention convolutional neural networks
CN107220237A (en) * 2017-05-24 2017-09-29 南京大学 A kind of method of business entity's Relation extraction based on convolutional neural networks
CN107301167A (en) * 2017-05-25 2017-10-27 中国科学院信息工程研究所 A kind of work(performance description information recognition methods and device
CN108108354A (en) * 2017-06-18 2018-06-01 北京理工大学 A kind of microblog users gender prediction's method based on deep learning
CN108108354B (en) * 2017-06-18 2021-04-06 北京理工大学 Microblog user gender prediction method based on deep learning
CN107301170A (en) * 2017-06-19 2017-10-27 北京百度网讯科技有限公司 The method and apparatus of cutting sentence based on artificial intelligence
US10755048B2 (en) 2017-06-19 2020-08-25 Beijing Baidu Netcom Science And Technology Co., Ltd. Artificial intelligence based method and apparatus for segmenting sentence
CN107506350A (en) * 2017-08-16 2017-12-22 京东方科技集团股份有限公司 A kind of method and apparatus of identification information
US10747961B2 (en) 2017-08-16 2020-08-18 Boe Technology Group Co., Ltd. Method and device for identifying a sentence
CN108038543A (en) * 2017-10-24 2018-05-15 华南师范大学 It is expected and anti-desired depth learning method and nerve network system
CN108038543B (en) * 2017-10-24 2021-01-22 华南师范大学 Expectation and anti-expectation deep learning method and neural network system
CN107590133A (en) * 2017-10-24 2018-01-16 武汉理工大学 The method and system that position vacant based on semanteme matches with job seeker resume
CN107943967B (en) * 2017-11-28 2020-05-22 华南理工大学 Text classification algorithm based on multi-angle convolutional neural network and cyclic neural network
CN107943967A (en) * 2017-11-28 2018-04-20 华南理工大学 Algorithm of documents categorization based on multi-angle convolutional neural networks and Recognition with Recurrent Neural Network
CN108154234A (en) * 2017-12-04 2018-06-12 盈盛资讯科技有限公司 A kind of knowledge learning method and system based on template
CN108280055A (en) * 2017-12-04 2018-07-13 盈盛资讯科技有限公司 A kind of knowledge learning method and system based on binary crelation
CN108009284A (en) * 2017-12-22 2018-05-08 重庆邮电大学 Using the Law Text sorting technique of semi-supervised convolutional neural networks
CN108256583A (en) * 2018-01-25 2018-07-06 北京东方科诺科技发展有限公司 A kind of multi-tag classification learning method based on coupling learning
CN108304530B (en) * 2018-01-26 2022-03-18 腾讯科技(深圳)有限公司 Knowledge base entry classification method and device and model training method and device
CN108304530A (en) * 2018-01-26 2018-07-20 腾讯科技(深圳)有限公司 Knowledge base entry sorting technique and device, model training method and device
CN108287911B (en) * 2018-02-01 2020-04-24 浙江大学 Relation extraction method based on constrained remote supervision
CN108287911A (en) * 2018-02-01 2018-07-17 浙江大学 A kind of Relation extraction method based on about fasciculation remote supervisory
CN108520042A (en) * 2018-04-03 2018-09-11 公安部第三研究所 Realize the system and method through detecing case-involving the role's calibration and role's assessment of suspect in work
CN108520042B (en) * 2018-04-03 2022-02-08 公安部第三研究所 System and method for realizing suspect case-involved role calibration and role evaluation in detection work
CN108763353B (en) * 2018-05-14 2022-03-15 中山大学 Baidu encyclopedia relation triple extraction method based on rules and remote supervision
CN108763353A (en) * 2018-05-14 2018-11-06 中山大学 Rule-based and remote supervisory Baidupedia relationship triple abstracting method
CN108664474A (en) * 2018-05-21 2018-10-16 众安信息技术服务有限公司 A kind of resume analytic method based on deep learning
CN109063759A (en) * 2018-07-20 2018-12-21 浙江大学 A kind of neural network structure searching method applied to the more attribute forecasts of picture
CN109446299B (en) * 2018-08-27 2022-08-16 中国科学院信息工程研究所 Method and system for searching e-mail content based on event recognition
CN109446299A (en) * 2018-08-27 2019-03-08 中国科学院信息工程研究所 The method and system of searching email content based on event recognition
CN109308304A (en) * 2018-09-18 2019-02-05 深圳和而泰数据资源与云技术有限公司 Information extraction method and device
CN109657207A (en) * 2018-11-29 2019-04-19 爱保科技(横琴)有限公司 The formatting processing method and processing unit of clause
CN109657207B (en) * 2018-11-29 2023-11-03 爱保科技有限公司 Formatting processing method and processing device for clauses
CN109670542A (en) * 2018-12-11 2019-04-23 田刚 A kind of false comment detection method based on comment external information
CN109885677A (en) * 2018-12-26 2019-06-14 中译语通科技股份有限公司 A kind of multi-faceted big data acquisition clearing system and method
CN109815338A (en) * 2018-12-28 2019-05-28 北京市遥感信息研究所 Relation extraction method and system in knowledge mapping based on mixed Gauss model
CN109492230B (en) * 2019-01-11 2022-12-20 浙江大学城市学院 Method for extracting insurance contract key information based on interested text field convolutional neural network
CN109492230A (en) * 2019-01-11 2019-03-19 浙江大学城市学院 A method of insurance contract key message is extracted based on textview field convolutional neural networks interested
CN110516239A (en) * 2019-08-26 2019-11-29 贵州大学 A kind of segmentation pond Relation extraction method based on convolutional neural networks
CN112445955A (en) * 2019-08-30 2021-03-05 珠海格力电器股份有限公司 Business opportunity information management method, system and storage medium
CN112445955B (en) * 2019-08-30 2023-10-13 珠海格力电器股份有限公司 Business opportunity information management method, system and storage medium
CN110717047A (en) * 2019-10-22 2020-01-21 湖南科技大学 Web service classification method based on graph convolution neural network
CN110717047B (en) * 2019-10-22 2022-06-28 湖南科技大学 Web service classification method based on graph convolution neural network
CN111078895A (en) * 2019-12-18 2020-04-28 江南大学 Remote supervision entity relation extraction method based on denoising convolutional neural network
CN111078895B (en) * 2019-12-18 2023-04-18 江南大学 Remote supervision entity relation extraction method based on denoising convolutional neural network
CN113535820A (en) * 2021-07-20 2021-10-22 贵州电网有限责任公司 Electrical operating personnel attribute presumption method based on convolutional neural network

Also Published As

Publication number Publication date
CN106570148B (en) 2019-07-23

Similar Documents

Publication Publication Date Title
CN106570148B (en) A kind of attribute extraction method based on convolutional neural networks
CN108614875B (en) Chinese emotion tendency classification method based on global average pooling convolutional neural network
CN104391942B (en) Short essay eigen extended method based on semantic collection of illustrative plates
CN107463607B (en) Method for acquiring and organizing upper and lower relations of domain entities by combining word vectors and bootstrap learning
CN108287911A (en) A kind of Relation extraction method based on about fasciculation remote supervisory
CN108038205B (en) Viewpoint analysis prototype system for Chinese microblogs
CN107122413A (en) A kind of keyword extracting method and device based on graph model
CN106446526A (en) Electronic medical record entity relation extraction method and apparatus
CN110765260A (en) Information recommendation method based on convolutional neural network and joint attention mechanism
CN107943784A (en) Relation extraction method based on generation confrontation network
CN113239210B (en) Water conservancy literature recommendation method and system based on automatic completion knowledge graph
CN111767725B (en) Data processing method and device based on emotion polarity analysis model
CN106970912A (en) Chinese sentence similarity calculating method, computing device and computer-readable storage medium
CN105138665A (en) Online internet topic mining method based on improved LDA model
CN104484380A (en) Personalized search method and personalized search device
CN106372064A (en) Characteristic word weight calculating method for text mining
CN105893362A (en) A method for acquiring knowledge point semantic vectors and a method and a system for determining correlative knowledge points
CN104699797A (en) Webpage data structured analytic method and device
CN104133855A (en) Smart association method and device for input method
CN107247739A (en) A kind of financial publication text knowledge extracting method based on factor graph
CN105608075A (en) Related knowledge point acquisition method and system
CN109145288A (en) Based on variation from the text depth characteristic extracting method of encoding model
CN113312480A (en) Scientific and technological thesis level multi-label classification method and device based on graph convolution network
CN105956158A (en) Automatic extraction method of network neologism on the basis of mass microblog texts and use information
CN111814476A (en) Method and device for extracting entity relationship

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant