CN104915448B - A kind of entity based on level convolutional network and paragraph link method - Google Patents

A kind of entity based on level convolutional network and paragraph link method Download PDF

Info

Publication number
CN104915448B
CN104915448B CN201510372795.3A CN201510372795A CN104915448B CN 104915448 B CN104915448 B CN 104915448B CN 201510372795 A CN201510372795 A CN 201510372795A CN 104915448 B CN104915448 B CN 104915448B
Authority
CN
China
Prior art keywords
paragraph
vectorization
entity
feature
sentence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510372795.3A
Other languages
Chinese (zh)
Other versions
CN104915448A (en
Inventor
包红云
郑孙聪
许家铭
齐振宇
徐博
郝红卫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201510372795.3A priority Critical patent/CN104915448B/en
Publication of CN104915448A publication Critical patent/CN104915448A/en
Application granted granted Critical
Publication of CN104915448B publication Critical patent/CN104915448B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2458Special types of queries, e.g. statistical queries, fuzzy queries or distributed queries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Fuzzy Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Databases & Information Systems (AREA)
  • Machine Translation (AREA)

Abstract

A kind of entity based on level convolutional network and paragraph link method, including:Represent that changing into sentence vectorization represents by term vectorization using convolutional neural networks;Represent to again pass by convolutional neural networks using sentence vectorization and consider that the sentence order information obtains paragraph vectorization and represented;Sentence vectorization expression and paragraph vectorization represent to export by Softmax, carry out the training of the convolutional neural networks model as supervision message by existing entity;Meanwhile consider that the pair wise similarity informations between paragraph semantic vector feature and Entity Semantics vector characteristics further improve the training of convolutional neural networks model;A test description paragraph is given, carrying out Deep Semantics feature extraction using the neural network model trained obtains the vectorization expression of test paragraph, and being then based on this semantic expressiveness can be directly linked on target entity by Softmax outputs.

Description

A kind of entity based on level convolutional network and paragraph link method
Technical field
The present invention relates to construction of knowledge base technical field, relate more specifically to a kind of entity based on level convolutional network with Paragraph link method.
Background technology
Nowadays, the large-scale knowledge base being widely used has Freebase, WordNet and YAGO etc..They are devoted to A global resources bank is built, and allows machine more easily to access and obtains structuring public information.Meanwhile these are known Know storehouse and provide application structure (APIs) to be convenient for people to inquire about the information of related entities more horn of plenty.For example, work as us A city name is retrieved in YAGO databases, and " during Washington D.C. ", returning result is as shown in table 1 below:
Table 1
It can be seen that the object information of return is all the organizational information of some highly structurals.But these structured messages And actual context and semantic information it is appreciated that entity are not met.Different with YAGO databases, Freebase and WordNet are then The descriptive paragraph related to retrieval entity can be additionally returned while return structure information, it is as shown in table 2 below:
Table 2
It can be seen that descriptive paragraph as shown in table 2 be relatively beneficial to user understand query entity word specific linguistic context and Semantic information.However, Freebase and WordNet descriptive paragraph information is all by manually entering edlin, this can cause The limitation of paragraph description is carried out to entity under big data and taken a substantial amount of time and manpower.Therefore, a height how is designed The entity of effect and the task that descriptive paragraph AutoLink method is that big data epoch construction of knowledge base is urgently needed badly.
It can further be seen that descriptive content is not necessarily to include query entity word from the returning result of table 2, and only need Entity is carried out comprising some related terms to describe in many aspects.Therefore, in order to solve this problem, entity and paragraph side of link Method needs to set about in terms of two:1st, the subject information of text is caught from a given segment description paragraph;2nd, find and real The important descriptive content that body phase is closed.It is the subject information that paragraph is extracted based on topic model method mostly to compare traditional method, The latent semantic analysis (PLSA) of such as Di Li Crays distribution (LDA) and probability.The common problem of these methods is the theme letter extracted Breath is the Term co-occurrence information acquisition based on document level, and the high of short text character representation openness influences to compare in by social media Seriously, and it lost the word order information in text.
In recent years, with the rise of deep neural network, some researchers attempt to use depth model and term vector table Show that the deep layer latent semantic feature of the descriptive paragraph of method study is represented to solve the link problems of entity and paragraph.It is however, existing Some, when the semantic feature for solving descriptive paragraph extracts, is simply simply regarded whole paragraph as based on depth model method One long sentence, which is handled or directly multiple sentences are weighted, averagely obtains semantic vector.And in fact, sentence in paragraph Sub- order is also with Semantic logical relation.
On the other hand, it is also very important to catch descriptive clue closely related with entity in paragraph.Such as above-mentioned table 2 Although descriptive paragraph in returning result directly comprising query entity word " Washington D.C. ", but comprising Many related vocabulary or phrase, such as:" George Washington ", " United States " and " capital " etc..Cause This, carry out that vectorization character representation contributes to entity and descriptive paragraph to entity links work.
The content of the invention
For above-mentioned technical problem, it is a primary object of the present invention to provide a kind of entity based on level convolutional network with Paragraph link method, so as to need not participation be manually AutoLink by the entity word in internet and descriptive paragraph, have Help the structure of the semantic knowledge-base under big data.
To achieve these goals, the invention provides a kind of entity based on level convolutional network and paragraph side of link Method, comprise the following steps:
Represent that changing into sentence vectorization represents by term vectorization using convolutional neural networks, the convolutional network is favourable In important clue of the extraction query entity in paragraph is described;
The sentence vectorization represents to again pass by convolutional neural networks and consider that the sentence order information obtains paragraph Vectorization represents;
The sentence vectorization represents and the paragraph vectorization represents to export by Softmax, makees by existing entity The training of the convolutional neural networks model is carried out for supervision message;
The pair-wise similarities letter between the paragraph semantic vector feature and Entity Semantics vector characteristics is considered simultaneously Breath further improves the training of the convolutional neural networks model;
A test description paragraph is given, Deep Semantics feature extraction is carried out using the neural network model trained The vectorization for obtaining the test paragraph represents that mesh can be directly linked to by Softmax outputs by being then based on this semantic expressiveness Mark is physically.
Feature learning problem in the linking of entity and paragraph is divided into four by entity and the paragraph link method of the present invention Individual level, it is respectively:The eigenmatrix layer that urtext paragraph represents to obtain by term vector;Obtained by convolutional neural networks The sentence vectorization arrived represents characteristic layer;The paragraph vectorization obtained by convolutional neural networks represents characteristic layer;Using word to The vectorization that amount look-up table obtains entity word represents characteristic layer.Tabled look-up by convolution character network and term vector, side of the invention Method accuracy value ACC of entity and paragraph link method on two text data sets is significantly superior to other control methods, and phase For best control methods two, accuracy value of the inventive method in two datasets improves 12.4% He respectively 16.76%.
Brief description of the drawings
Fig. 1 is the flow of the entity and paragraph link method based on level convolutional network as one embodiment of the invention Figure;
Fig. 2 is the framework of the entity and paragraph link method based on level convolutional network as one embodiment of the invention Schematic diagram;
Fig. 3 is the performance of the entity and paragraph link method based on level convolutional network as one embodiment of the invention Schematic diagram.
Embodiment
For the object, technical solutions and advantages of the present invention are more clearly understood, below in conjunction with specific embodiment, and reference Accompanying drawing, the present invention is described in further detail.
, can be by internet the invention discloses a kind of entity based on level convolutional network and paragraph link method Entity word proceeds without artificial participation ground AutoLink with descriptive paragraph, and its general plotting is to pass through stratification convolutional Neural Network first carries out convolution to the term vector in paragraph and obtains the vectorization expression of sentence.Consider order letter of the sentence in paragraph Breath, and the vectorization to sentence represents to carry out convolution again and obtains the vectorization expression of paragraph.Then using substance feature as Supervision message instructs the parameter learning of convolutional neural networks model, at the same consider the deep semantic feature of paragraph and Entity Semantics to Pair-wise affinity informations between quantization means improve the study of convolutional neural networks model.Given one new descriptive Paragraph, then using convolutional neural networks model extraction its deep semantic feature trained, and export to obtain based on this feature Corresponding entity link.
More specifically, this method represents to change into sentence vectorization table first with convolutional neural networks by term vectorization Show.Then using sentence vectorization represent again pass by convolutional neural networks and consider the sentence order information obtain paragraph to Quantization means.Sentence vectorization expression and paragraph vectorization represent to export by Softmax, believe by existing entity as supervision Breath carries out the training of the convolutional neural networks model.Meanwhile consider paragraph semantic vector feature and Entity Semantics vector characteristics Between pair-wise similarity informations further improve the training of convolutional neural networks model.Give a test description section Fall, carrying out Deep Semantics feature extraction using the neural network model trained obtains the vectorization expression of test paragraph, then It can be directly linked to based on this semantic expressiveness by Softmax outputs on target entity.
The entity based on level convolutional network as one embodiment of the invention is linked with paragraph below in conjunction with the accompanying drawings Method is described in detail.
Fig. 1 is the flow of the entity and paragraph link method based on level convolutional network as one embodiment of the invention Figure.
Reference picture 1, in step S101, represented by convolutional neural networks model and term vectorization, extract pending paragraph In the vectorization of every sentence represent feature;
It is described to be represented by convolutional neural networks model and term vectorization according to one exemplary embodiment of the present invention, Extracting in pending paragraph the step of the vectorization of every sentence represents feature includes:
In step S1011, a sentence in pending paragraph is given, lexical item quantization means is obtained using look-up table and incites somebody to action Sentence is characterized into matrix form;
In step S1012, represent to carry out one-dimensional convolution in feature in the sentence matrixing, obtain the feature square after convolution Battle array;
Average sampling is carried out on step S1013, the eigenmatrix after the convolution to be compressed feature, obtains sentence The vectorization of son represents.
It is described to obtain lexical item quantization means and by sentence table using look-up table according to one exemplary embodiment of the present invention The step of levying into matrix form includes:
Give the term vector set that a word2vec is trainedWherein, | V | be dictionary size, d be word to The dimension of amount.Then length is that n sentence can be expressed as in any paragraph:
S=(x1;x2;...;xn) (1)
Wherein, xiIt is that vectorization corresponding to i-th of the word found using look-up table in term vector set is represented.Wherein, If word xiDo not appear in the term vector set trained, then directly it is carried out in the exemplary embodiment of the present invention Random initializtion represents.
It is described to represent to carry out one-dimensional convolution in feature in sentence matrixing in step S1012, obtain the feature square after convolution The step of battle array, includes:
Here, useRepresent in sentence s from the h of i-th of word startingsIndividual continuous word is special Sign.Give an one-dimensional convolution kernelThen hsEigenmatrix after individual continuous word feature convolution is:
Wherein, b(1)It is bias term, f is activation primitive,It is hsIndividual continuous word featureConvolution Eigenmatrix afterwards.Then the eigenmatrix of the sentence is after convolution:
Average sampling is carried out on step S1013, the eigenmatrix after convolution to be compressed feature, obtains sentence The step of vectorization of son represents includes:
The present invention the exemplary embodiment in, it is described use average sampling the step of for:
So far, each convolution kernelA d dimensional feature vector can be generatedIf having used k convolution kernel, pass through A convolutional layer is crossed, the vectorization that will eventually get sentence is expressed asThe dimension that then sentence vectorization represents Spend for dk.
In step S102, represented using convolutional neural networks structure and the sentence vectorization, learn the depth of the paragraph Spend semantic feature;
According to one exemplary embodiment of the present invention, the deep semantic feature learning method of the paragraph includes:
In step S1021, word order of the sentence in the paragraph is pressed by section using the sentence vector characteristics in the paragraph Fall to characterize into matrix form;
In step S1022, represent to carry out one-dimensional convolution in feature in the paragraph matrixing, obtain the feature square after convolution Battle array;
Average sampling is carried out on step S1023, the eigenmatrix after the convolution to be compressed feature and carry out Once linear converts, and the vectorization for obtaining paragraph represents.
According to one exemplary embodiment of the present invention, the sentence vector characteristics using in paragraph are by sentence described Word order in paragraph, which characterizes paragraph into the step of matrix form, to be included:
The vectorization for having obtained the l bar sentences of the paragraph represents that then paragraph can be expressed as:
T=(s1;s2;...;sl) (5)
It is described to represent to carry out one-dimensional convolution in feature in paragraph matrixing in step S1022, obtain the feature square after convolution The step of battle array, includes:
Here, useRepresent in paragraph t from the h of i-th of sentence startingtIndividual continuous sentence Feature.Give an one-dimensional convolution kernelThen htConvolution after individual continuous sentence characteristics convolution is characterized as:
Wherein, b(2)It is bias term, f is activation primitive,It is htIndividual continuous sentence characteristicsVolume Feature after product.Then the feature of the paragraph is after convolution:
Average sampling is carried out on step S1023, the eigenmatrix after convolution to be compressed feature and carry out Once linear convert, obtain paragraph vectorization represent the step of include:
The present invention the exemplary embodiment in, it is described use average sampling the step of for:
So far, by convolution kernel W(2)Generate a dk dimensional feature vectorIn order to facilitate calculating paragraph feature and reality The similitude of body characteristicses, the unification of vector dimension need to be ensured, then once linear conversion be carried out to paragraph vector:
Wherein,For the matrix of a linear transformation, and characteristic vector z is one exemplary embodiment of the present invention In final paragraph characteristic vector.
In step S103, the vectorization of the sentence represents and the vectorization of the paragraph represents to pass through Softmax respectively Output is fitted the affiliated entity of paragraph;
According to one exemplary embodiment of the present invention, the vectorization of the sentence and the paragraph represents to be fitted institute respectively The method for stating the affiliated entity of paragraph comprises the following steps:
In step S1031, linear transformation is carried out respectively to the sentence vector sum paragraph vector and obtains output vector, and is made Canonical is carried out with Dropout technologies;
In step S1032, the link probability of Softmax functions calculating candidate's entity is used;
According to one exemplary embodiment of the present invention, distich subvector and the paragraph vector carries out linear transformation and obtained Output vector, and included using the step of Dropout technologies progress canonical:
Distich subvector feature s and paragraph vector characteristics t carry out linear change respectively, obtain two output vectors:
Ys=W(4)·(sοr)+b(4) (10)
Y=W(5)·(zοr)+b(5) (11)
Wherein,WithIt is weight matrix, m is in one exemplary embodiment of the present invention Entity number, symbol.Representing matrix element multiplies operation, andIt is then a Bernoulli Jacob for obeying certain probability ρ point Cloth.Over-fitting can be prevented using Dropout technologies, can be with the robustness of strength neural network model.
In step S1032, the step of calculating the link probability of candidate's entity using the Softmax functions, includes:
Swashed respectively using Softmax in two output layers of the sentence vector characteristics and the paragraph vector characteristics Function living calculates the probable value for each corresponding to the entity word:
Then in formula (12) and formula (13), psiAnd piThe probable value of corresponding i-th of entity word is represented respectively.
In step S104, the vectorization for calculating the entity represents the pair-wise phases represented with the paragraph vectorization Like information;
Give an entity set of words E={ e1, e2..., em, the entity set of words is carried out just using word2vec Beginningization, then entity set of words E and the paragraph characteristic vector z similitude be:
Sim (z, E)={ ze1, ze2..., zem} (14)
Wherein, operator ze represents the paragraph characteristic vector z and the corresponding entity word e similitude.
In step S105, pass through Softmax fit objects entity word and paragraph characteristic vector and the pair- of target entity word Wise similarity informations carry out error back propagation training convolutional neural networks model;
It is described by Softmax fit objects entity word and the paragraph according to one exemplary embodiment of the present invention Characteristic vector and the pair-wise similarity informations of target entity word carry out training convolutional neural networks described in error back propagation The step of model, includes:
In step S1051, exported according to the sentence characteristics and paragraph feature, using the Softmax to the training The fitting result sets target function of target entity word in data set;
In step S1052, set according to the pair-wise similarity informations of the paragraph feature and the target entity word Object function;
In step S1053, global object constraint function is set;
In step S1054, the parameter in model is updated using stochastic gradient descent method;
It is described according to utilizing sentence characteristics and the output of paragraph feature according to one exemplary embodiment of the present invention Softmax concentrates the step of fitting result sets target function of target entity word to include to the training data:
Using formula (10), (11) and formula (12), (13), the sentence vectorization feature and paragraph vector are set Change feature goal constraint function be respectively:
Wherein, LsFor the goal constraint function of the sentence vectorization feature, Lp1For the mesh of the paragraph vectorization feature Mark constraint function,For paragraph set in all training corpusIn all sentence set,It is belonging to i-th of sentence Correct entity word andIt is the correct entity word belonging to i-th of paragraph.
It is described to be set according to the pair-wise similarity informations of paragraph feature and the target entity word in step S1052 The step of object function, includes:
In order to strengthen the semantic meaning representation ability of the paragraph and entity, the present invention strengthens institute by sets target constraint function The similitude of the corresponding affiliated entity word vectorization feature of paragraph vectorization feature is stated, and weakens the paragraph vector Change the similitude of the corresponding non-belonging entity word vectorization feature of feature, its goal constraint function is as follows:
Wherein, erIt is to give the correct entity word belonging to the paragraph z.
It is as follows in step S1053, described the step of setting global object constraint function:
L=Ls+(1-α)·Lp1+α·Lp2 (18)
Wherein, α is weight harmonic coefficient, for balancing two of paragraph vectorization feature constraint Lp1And Lp2
In step S1054, described the step of being updated using stochastic gradient descent method to the parameter in the model Including:
All model training parameters are collectively expressed as θ in the goal constraint function of setting:
θ=(x, W(1), b(1), W(2), b(2), α, W(3), W(4), b(4), W(5), b(5), E) and (19)
In one exemplary embodiment of the present invention, error back propagation is carried out to described using stochastic gradient descent method Object function optimizes.
In step S106, deep semantic feature is carried out to test description paragraph using convolutional neural networks model after renewal Extract, the vectorization for being then based on paragraph represents to be linked with corresponding entity word.
According to one exemplary embodiment of the present invention, the convolutional neural networks model with after renewal is to described Test description paragraph carries out deep semantic feature extraction, and the vectorization for being then based on the paragraph represents and the corresponding reality The step of pronouns, general term for nouns, numerals and measure words is linked includes:
In step S1061, a test paragraph text is given, is first passed through in formula (2), (3), (4) calculating paragraph The vectorization feature s of sentence;
In step S1062, the vectorization feature z of the paragraph is calculated by formula (6), (7), (8), (9);
In step S1063, using the vectorization feature z of the paragraph of generation, using the linear transformation without Dropout and The matching probability of the entity word corresponding to the output of Softmax functions:
Y=W(5)·z+b(5) (20)
Then matching probability highest entity word is the affiliated entity word of the test paragraph.
Fig. 2 is the framework of the entity and paragraph link method based on level convolutional network as one embodiment of the invention Schematic diagram.
Reference picture 2, the entity based on level convolutional network share the characteristic vector of four levels with paragraph link method Represent, be respectively:
Feature hierarchy one:The eigenmatrix that urtext paragraph represents to obtain by term vector;
Feature hierarchy two:The sentence vectorization obtained by convolutional neural networks represents feature;
Feature hierarchy three:The paragraph vectorization obtained by convolutional neural networks represents feature;
Feature hierarchy four:The vectorization that entity word is obtained using term vector look-up table represents feature;
The whole model training stage shares supervision message at three and instructed, and is respectively:
Supervision message one:The vectorization of sentence represents feature after linear change and Softmax outputs to affiliated entity The fitting information of word;
Supervision message two:The vectorization of paragraph represents feature after linear change and Softmax outputs to affiliated entity The fitting information of word;
Supervision message three:The vectorization of paragraph represents Pair-wise of the feature with affiliated entity word after linear change Similarity information;
For the entity of accurate evaluation the inventive method and the link performance of paragraph, the present invention passes through comparison entity and paragraph Link result and the uniformity of the true affiliated entity of paragraph obtain the precision (ACC) of the inventive method.It is given one descriptive section Fall sample x(i), the entity word of the inventive method link is e(i), and truly the entity word is paragraphThe then definition of precision It is as follows:
Wherein,The number of descriptive paragraph, δ (x, y) is indicator function, as x=y indicator function be 1, when x ≠ Indicator function is 0 during y.
Two kinds of disclosure data sets are used in the experiment of the present invention:
History:The data set includes 409 entities, 1704 paragraphs.
Literature:The data set includes 445 entities, 2247 paragraphs.
For these text data sets, the present invention is (including going the operation such as stop words and stem reduction) without any processing. Average each paragraph includes 4-6 bar sentences, and each paragraph only includes 1 entity word.The specific statistical information of data set such as table 3 It is shown:
Table 3
Following control methods is used in the experiment of the present invention:
Control methods one:Based on bag of words and this special homing method of logic, this method is directly in the bag of words of urtext Using this special homing method of logic on model;
Control methods two:Link method based on convolutional neural networks, this method use traditional convolutional neural networks mould Type simply regards entity and paragraph link problems as a classification problem.
It is as shown in table 4 using parameter setting in present invention experiment:
Table 4
Data set ρ hs ht d k
History 0.5 3 6 100 1
Literature 0.5 3 8 100 1
In table 4, using Dropout specific gravity factor, h when parameter ρ is model trainingsFor sentence vectorization character representation when The frame mouth size of convolution kernel, htFor paragraph vectorization character representation when convolution kernel frame mouth size, d is term vector dimension, and k is sentence The number of convolution kernel during subvector character representation.
In present invention experiment, all entities perform 50 times with paragraph link method and ask for its mean accuracy value (ACC), finally Result of the test it is as shown in table 5:
Table 5
Method History/ accuracy values (%) Literature/ accuracy values (%)
Control methods one 65.10±0.01 61.17±0.05
Control methods two 77.01±3.92 74.50±10.3
The inventive method 89.41±1.05 91.26±0.50
Table 5 is that the inventive method, control methods one, the entity on two text data sets of control methods two link with paragraph Accuracy value (ACC) evaluation result of method.Result of the test shows that the performance of the inventive method is significantly superior to other to analogy Method.And 12.4% is improved respectively relative to best control methods two, accuracy value of the inventive method in two datasets With 16.76%.
Meanwhile the slip word window size of verification experimental verification of the present invention convolution kernel when carrying out sentence characteristics expression is to the present invention Method carries out the influence for the accuracy value performance that entity links with paragraph, and result of the test is as shown in Figure 3.It can be seen that when word window When size is 3, the inventive method performance is all optimal in two datasets, and when word window size is more than 3, the present invention The accuracy value hydraulic performance decline of method.Thus the slip word window size for the sentence characteristics convolution kernel that the present invention uses in testing is 3。
Particular embodiments described above, the purpose of the present invention, technical scheme and beneficial effect are carried out further in detail Describe in detail bright, it should be understood that the foregoing is only the present invention specific embodiment, be not intended to limit the invention, it is all Within the spirit and principles in the present invention, any modification, equivalent substitution and improvements done etc., the protection of the present invention should be included in Within the scope of.

Claims (10)

1. a kind of entity based on level convolutional network and paragraph link method, comprise the following steps:
Represented by convolutional neural networks model and term vectorization, the vectorization for extracting every sentence in pending paragraph represents special Sign;
Represented using convolutional neural networks structure and sentence vectorization, learn the deep semantic feature of the paragraph;
The vectorization expression of the sentence and the vectorization of paragraph are represented true by Softmax output fitting paragraphs institute respectively Body;
Calculate the pair-wise analog informations that the vectorization expression of the entity represents with paragraph vectorization;
Pass through Softmax fit objects entity word and paragraph characteristic vector and the pair-wise similarity informations of target entity word Carry out error back propagation and train the convolutional neural networks model;
Deep semantic feature extraction is carried out to the pending paragraph using the convolutional neural networks model after renewal, then Vectorization based on the paragraph represents to be linked with corresponding entity word.
2. the entity according to claim 1 based on level convolutional network and paragraph link method, it is characterised in that described Represented by convolutional neural networks model and term vectorization, the vectorization for extracting every sentence in pending paragraph represents feature Step includes:
A sentence in pending paragraph is given, term vectorization expression is obtained using look-up table and characterizes the sentence into square Formation formula;
Represent to carry out one-dimensional convolution in feature in the sentence matrixing, obtain the eigenmatrix after convolution;
Average sampling is carried out in convolution feature after the convolution to be compressed feature, obtains the vectorization table of the sentence Show.
3. the entity according to claim 1 based on level convolutional network and paragraph link method, it is characterised in that described Represented using convolutional neural networks structure and the sentence vectorization, wrapped the step of the deep semantic feature for learning the paragraph Include:
Paragraph is characterized into matrix form by word order of the sentence in the paragraph using the sentence vector characteristics in the paragraph;
Represent to carry out one-dimensional convolution in feature in the paragraph matrixing, obtain the eigenmatrix after convolution;
Average sampling is carried out in convolution feature after the convolution to be compressed feature and carry out once linear conversion, is obtained The vectorization of the paragraph represents.
4. the entity according to claim 1 based on level convolutional network and paragraph link method, it is characterised in that described The vectorization of sentence represents and the vectorization of the paragraph represents to be fitted the affiliated entity of paragraph by Softmax outputs respectively The step of include:
Linear transformation is carried out respectively to the sentence vector sum paragraph vector and obtains output vector, and is entered using Dropout technologies Row canonical;
The link probability of candidate's entity is calculated using Softmax functions.
5. the entity according to claim 1 based on level convolutional network and paragraph link method, it is characterised in that described The vectorization for calculating the entity represents that the method for the pair-wise analog informations represented with the paragraph vectorization is as follows:
Give an entity set of words E={ e1, e2..., em, the entity set of words is carried out using word2vec initial Change, then entity set of words E and the paragraph characteristic vector z similitude are:
Sim (z, E)={ ze1, ze2..., zem}.
Wherein, operator ze represents the paragraph characteristic vector z and the corresponding entity word e similitude.
6. the entity according to claim 1 based on level convolutional network and paragraph link method, it is characterised in that described Pass through the Softmax fit objects entity word and the paragraph characteristic vector and the pair-wise similarities of target entity word The step of information progress error back propagation trains the convolutional neural networks model includes:
Exported according to the sentence characteristics and paragraph feature, target entity is concentrated to the training data using the Softmax The fitting result sets target function of word;
According to the paragraph feature and the pair-wise similarity information sets target functions of the target entity word;
Set global object constraint function and the object function is subjected to unified fusion;
The parameter in the convolutional neural networks model is updated using stochastic gradient descent method.
7. the entity according to claim 6 based on level convolutional network and paragraph link method, it is characterised in that described Included according to the step of paragraph feature and pair-wise similarity information sets target functions of the target entity word:
In order to strengthen the semantic meaning representation ability of the paragraph and entity, the paragraph vector is strengthened by sets target constraint function Change the similitude of the corresponding affiliated entity word vectorization feature of feature, and it is corresponding to weaken the paragraph vectorization feature Non-belonging entity word vectorization feature similitude, its described goal constraint function is as follows:
<mrow> <msub> <mi>L</mi> <mrow> <mi>p</mi> <mn>2</mn> </mrow> </msub> <mo>=</mo> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mrow> <mo>|</mo> <mi>c</mi> <mo>|</mo> </mrow> </munderover> <munder> <mo>&amp;Sigma;</mo> <mrow> <msub> <mi>e</mi> <mi>j</mi> </msub> <mo>&amp;Element;</mo> <mi>E</mi> <mrow> <mo>(</mo> <msub> <mi>e</mi> <mi>j</mi> </msub> <mo>&amp;NotEqual;</mo> <msub> <mi>e</mi> <mi>r</mi> </msub> <mo>)</mo> </mrow> </mrow> </munder> <mi>m</mi> <mi>a</mi> <mi>x</mi> <mrow> <mo>(</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>-</mo> <mi>s</mi> <mi>i</mi> <mi>m</mi> <mo>(</mo> <mrow> <msup> <mi>z</mi> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> </msup> <mo>,</mo> <msubsup> <mi>e</mi> <mi>r</mi> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> </msubsup> </mrow> <mo>)</mo> <mo>+</mo> <mi>s</mi> <mi>i</mi> <mi>m</mi> <mo>(</mo> <mrow> <msup> <mi>z</mi> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> </msup> <mo>,</mo> <msubsup> <mi>e</mi> <mi>j</mi> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> </msubsup> </mrow> <mo>)</mo> <mo>)</mo> </mrow> <mo>;</mo> </mrow>
Wherein, erIt is to give the correct entity word belonging to the paragraph z, er (i)For the true entity word of paragraph.
8. the entity according to claim 6 based on level convolutional network and paragraph link method, it is characterised in that described The step of object function is carried out unified fusion by setting global object constraint function includes:
It is as follows to set the global object constraint function:
L=Ls+(1-α)·Lp1+α·Lp2
Wherein, LsFor the goal constraint function of sentence vectorization feature, α is weight harmonic coefficient, for balancing the paragraph vector Change two constraints of feature, i.e. the fitting of target entity word is concentrated in paragraph feature output using Softmax to the training data Bound term Lp1With paragraph feature and the pair-wise similarity bound terms L of the target entity wordp2
9. the entity according to claim 1 based on level convolutional network and paragraph link method, it is characterised in that described Processing section is treated with the convolutional neural networks model after renewal and drops into row deep semantic feature extraction, is then based on the paragraph Vectorization represents to include the step of link with the corresponding entity word:
A pending paragraph text is given, the convolutional neural networks model trained is first passed through and calculates sentence in the paragraph Vectorization feature;
The vectorization feature of the paragraph is calculated by the convolutional neural networks model trained;
Using the vectorization feature of the paragraph of generation, the linear transformation without Dropout and the output pair of Softmax functions are used The matching probability for the entity word answered.
10. the entity according to claim 1 based on level convolutional network and paragraph link method, it is characterised in that institute State in convolutional neural networks model, the slip word window size of the sentence characteristics convolution kernel used is 3.
CN201510372795.3A 2015-06-30 2015-06-30 A kind of entity based on level convolutional network and paragraph link method Active CN104915448B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510372795.3A CN104915448B (en) 2015-06-30 2015-06-30 A kind of entity based on level convolutional network and paragraph link method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510372795.3A CN104915448B (en) 2015-06-30 2015-06-30 A kind of entity based on level convolutional network and paragraph link method

Publications (2)

Publication Number Publication Date
CN104915448A CN104915448A (en) 2015-09-16
CN104915448B true CN104915448B (en) 2018-03-27

Family

ID=54084511

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510372795.3A Active CN104915448B (en) 2015-06-30 2015-06-30 A kind of entity based on level convolutional network and paragraph link method

Country Status (1)

Country Link
CN (1) CN104915448B (en)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107220220A (en) * 2016-03-22 2017-09-29 索尼公司 Electronic equipment and method for text-processing
CN106326985A (en) * 2016-08-18 2017-01-11 北京旷视科技有限公司 Neural network training method, neural network training device, data processing method and data processing device
CN106339718A (en) * 2016-08-18 2017-01-18 苏州大学 Classification method based on neural network and classification device thereof
CN106446526B (en) * 2016-08-31 2019-11-15 北京千安哲信息技术有限公司 Electronic health record entity relation extraction method and device
CN106844765B (en) * 2017-02-22 2019-12-20 中国科学院自动化研究所 Significant information detection method and device based on convolutional neural network
CN107144569A (en) * 2017-04-27 2017-09-08 西安交通大学 The fan blade surface defect diagnostic method split based on selective search
CN107168956B (en) * 2017-05-26 2020-06-02 北京理工大学 Chinese chapter structure analysis method and system based on pipeline
CN109426664A (en) * 2017-08-30 2019-03-05 上海诺悦智能科技有限公司 A kind of sentence similarity calculation method based on convolutional neural networks
CN107704563B (en) * 2017-09-29 2021-05-18 广州多益网络股份有限公司 Question recommendation method and system
CN108304552B (en) * 2018-02-01 2021-01-08 浙江大学 Named entity linking method based on knowledge base feature extraction
CN108764233B (en) * 2018-05-08 2021-10-15 天津师范大学 Scene character recognition method based on continuous convolution activation
CN109344244B (en) * 2018-10-29 2019-11-08 山东大学 A kind of the neural network relationship classification method and its realization system of fusion discrimination information
CN109697288B (en) * 2018-12-25 2020-09-15 北京理工大学 Instance alignment method based on deep learning
CN110147533B (en) 2019-01-24 2023-08-29 腾讯科技(深圳)有限公司 Encoding method, apparatus, device and storage medium
CN109992629B (en) * 2019-02-28 2021-08-06 中国科学院计算技术研究所 Neural network relation extraction method and system fusing entity type constraints
CN112328800A (en) * 2019-08-05 2021-02-05 上海交通大学 System and method for automatically generating programming specification question answers
CN110674317B (en) * 2019-09-30 2022-04-12 北京邮电大学 Entity linking method and device based on graph neural network
CN110717339B (en) * 2019-12-12 2020-06-30 北京百度网讯科技有限公司 Semantic representation model processing method and device, electronic equipment and storage medium
CN111222314B (en) * 2020-01-03 2021-12-21 北大方正集团有限公司 Layout document comparison method, device, equipment and storage medium
CN113361261B (en) * 2021-05-19 2022-09-09 重庆邮电大学 Method and device for selecting legal case candidate paragraphs based on enhance matrix
CN115130435B (en) * 2022-06-27 2023-08-11 北京百度网讯科技有限公司 Document processing method, device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104317834A (en) * 2014-10-10 2015-01-28 浙江大学 Cross-media sorting method based on deep neural network
CN104462357A (en) * 2014-12-08 2015-03-25 百度在线网络技术(北京)有限公司 Method and device for realizing personalized search
CN104615767A (en) * 2015-02-15 2015-05-13 百度在线网络技术(北京)有限公司 Searching-ranking model training method and device and search processing method
CN104679863A (en) * 2015-02-28 2015-06-03 武汉烽火众智数字技术有限责任公司 Method and system for searching images by images based on deep learning

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130212049A1 (en) * 2012-02-15 2013-08-15 American Gnc Corporation Machine Evolutionary Behavior by Embedded Collaborative Learning Engine (eCLE)

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104317834A (en) * 2014-10-10 2015-01-28 浙江大学 Cross-media sorting method based on deep neural network
CN104462357A (en) * 2014-12-08 2015-03-25 百度在线网络技术(北京)有限公司 Method and device for realizing personalized search
CN104615767A (en) * 2015-02-15 2015-05-13 百度在线网络技术(北京)有限公司 Searching-ranking model training method and device and search processing method
CN104679863A (en) * 2015-02-28 2015-06-03 武汉烽火众智数字技术有限责任公司 Method and system for searching images by images based on deep learning

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
A Convolutional Neural Network for Modelling Sentences;N. Kalchbrenner;《Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics》;20140630;第655-665页 *
A Neural Network for Factoid Question Answering over Paragraphs;M Iyyer etal;《Conference on Empirical Methods in Natural Language Processing》;20141231;第633-644页 *
Convolutional neural networks for sentence classification;Y. Kim etal;《Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing》;20141231;第1746-1751页 *

Also Published As

Publication number Publication date
CN104915448A (en) 2015-09-16

Similar Documents

Publication Publication Date Title
CN104915448B (en) A kind of entity based on level convolutional network and paragraph link method
CN104834747B (en) Short text classification method based on convolutional neural networks
CN108595632B (en) Hybrid neural network text classification method fusing abstract and main body characteristics
CN107766324B (en) Text consistency analysis method based on deep neural network
CN109933670B (en) Text classification method for calculating semantic distance based on combined matrix
CN108399163A (en) Bluebeard compound polymerize the text similarity measure with word combination semantic feature
CN104951548B (en) A kind of computational methods and system of negative public sentiment index
CN109697285A (en) Enhance the hierarchical B iLSTM Chinese electronic health record disease code mask method of semantic expressiveness
WO2019080863A1 (en) Text sentiment classification method, storage medium and computer
CN109740154A (en) A kind of online comment fine granularity sentiment analysis method based on multi-task learning
CN109753660B (en) LSTM-based winning bid web page named entity extraction method
CN104216876B (en) Information text filter method and system
CN106484675A (en) Fusion distributed semantic and the character relation abstracting method of sentence justice feature
CN110245229A (en) A kind of deep learning theme sensibility classification method based on data enhancing
CN106980609A (en) A kind of name entity recognition method of the condition random field of word-based vector representation
CN106776562A (en) A kind of keyword extracting method and extraction system
Lei et al. Patent analytics based on feature vector space model: A case of IoT
Chang et al. Research on detection methods based on Doc2vec abnormal comments
CN104657350A (en) Hash learning method for short text integrated with implicit semantic features
CN109299462A (en) Short text similarity calculating method based on multidimensional convolution feature
CN106599054A (en) Method and system for title classification and push
CN102289522A (en) Method of intelligently classifying texts
CN107943824A (en) A kind of big data news category method, system and device based on LDA
CN103020167B (en) A kind of computer Chinese file classification method
CN107895000A (en) A kind of cross-cutting semantic information retrieval method based on convolutional neural networks

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant