CN103714178B - Automatic image marking method based on word correlation - Google Patents
Automatic image marking method based on word correlation Download PDFInfo
- Publication number
- CN103714178B CN103714178B CN201410008553.1A CN201410008553A CN103714178B CN 103714178 B CN103714178 B CN 103714178B CN 201410008553 A CN201410008553 A CN 201410008553A CN 103714178 B CN103714178 B CN 103714178B
- Authority
- CN
- China
- Prior art keywords
- word
- image
- mark
- training set
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
- G06F18/2155—Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling
Abstract
The invention discloses an automatic image marking method based on work correlation. A training set T comprises l images, n marking words are marked on each image of the training set T, the training set T is provided with a corresponding vision lemma, and the image to be marked is I. The method includes the steps that a semantic vector of each marking word w is calculated according to a formula, the marking word w is represented by the vector form w=<v1, v2,......, vm>, ci is an associated word in a context, and the associated words in the context total m; semantic similarity of the marking words is calculated according to a formula, and vector module calculated is achieved as is shown in the specification; p(A) is calculated according to the formula, wherein A is a marking word group in w1, w2,......wn, and n is the number of the marking word groups; the conditional probability p(I/wi) is calculated according to a formula; the marking word group A of the image I to be marked is calculated according to the formula A=arg maxAp(I/A) p(A).
Description
Technical field
The present invention relates to image processing field, the image automatic annotation method of particularly a kind of word-based correlation.
Background technology
With the high speed development of multimedia and Internet technology, daily life, work are believed to multimedias such as images
The dependence of breath is more and more stronger.It is intended to based on the retrieval that semantic image retrieval can not only accurately express user, be also convenient to use
The use at family, therefore this retrieval mode not only become a kind of important form of image retrieval, and become what researcher chased
Hot technology.
And automatic image annotation technology is an important and challenging job in Semantic Image Retrieval, image is certainly
The appearance of dynamic label technology is to automatically obtain the semantic information comprising in image vision content, and it attempts to regard in image bottom
Feel and build a bridge block, thus support is made to semantic retrieval in semantic level between feature and high-level semantic.Therefore, it is based on
The automatic marking algorithm research of image, semantic has become a very active research branch and crucial skill in field of image search
Art, has good application prospect and researching value.
Automatic image annotation is exactly to allow the image automatically to no mark for the computer add the semanteme that can reflect picture material
Keyword.It practises semantic concept space and vision spy automatically using the image collection of mark or other obtainable information
Levy the relational model in space, and mark the image of unknown semantics with this model.By in the high-layer semantic information of image and bottom
Set up a kind of mapping relations between feature, solve the problems, such as semantic gap to a certain extent.
The image automatic annotation method of joint media correlation model is to be currently based in generation model image labeling method to answer
With a kind of most commonly used image labeling algorithm, obtain the widely studied of scholar.The basic thought of this marking model is to utilize
The method of probability statistics sets up the probability relativity in Image Visual Feature space and semantic concept space, by both statistical learnings
Between the joint probability distribution that exists, it is maximum to find out the joint probability that one group of semantic tagger word is allowed to and picture material between, and
This group is marked word as the final mark of testing image.
But joint media correlation model belongs to one kind of probabilistic model, this class model mark word high to occurrence frequency has
Skewed popularity.Secondly, in joint media correlation model automatic image annotation mark method, different candidate's mark words are in annotation process
It is assumed to be it is separate, and the correlation marking between word is not fully excavated.In fact with piece image, no
There is multiple association such as symbiosis, level or space between mark word.
Such as one width contains the image of semantic objects such as " sun, sky, cloud, mountain, tree ", regards from image
Feel that " sun " can not depart from " sky " this language it can be seen that " sun " and " sky " object has certain spatial correlation in content
Adopted object and be individually present;Equally, for " mountain " and " tree " two semantic objects in picture material,
The presence of " tree " object is with " mountain " semantic object for vision content background, and the two is same in image vision content
There is the contact of interwoveness it is impossible to utterly assume that this two mark words are separate being labeled.Therefore, joint media
Correlation model Automatic image annotation algorithm thinks that in annotation process different candidates mark separate way between word and exist
Certain defect, may lead to because ignoring correlation between word mark semantic inconsistent phenomenon between word in annotation results.
Content of the invention
In view of this, the invention provides a kind of image automatic annotation method of word-based correlation, combined with overcoming
Media correlation model Automatic image annotation algorithm thinks that in annotation process different candidates mark separate way between word
The defect existing, solves the problems, such as to lead to because ignoring correlation between word in annotation results semanteme between mark word inconsistent.This
The technical scheme that invention proposes is:
A kind of image automatic annotation method of word-based correlation, training set t comprises l image, described l image
Constitute image collection p=[p1p2… pl];Each image labeling of described training set t has n mark word, trains set t
In all mark words constitute mark set of words w=[w1w2… ws];Each image of training set t has corresponding visual word
Unit, in training set t, all vision lemmas constitute visual word unit set b=[b1b2… by], image to be marked is i, the party
Method includes:
A. according to formulaCalculate the semantic vector of each of training set t mark word w, word w will be marked
It is expressed as vector form w=< v1,v2,…,vm>, wherein, ciFor context relation word, total m context relation word, p (ci)
For context relation word ciOverall distribution probability, p (ci/ w) represent context relation word ciWith mark word w in training set t
Co-occurrence number of times with mark word w training set t in appearance total degree ratio, that is,Wherein,
Described context relation word is the mark word in training set t;
B. according to formulaCalculate the semantic similarity between mark word, wherein | | | | it is
Vectorial mould calculates, wi·wjFor dot product computing;
C. according to formulaCalculate p (a), wherein a is mark phrase { w1,
w2,…wn, n is the number of mark phrase;
D. according to formulaDesign conditions Probability p (i/wi), wherein, p (wi)
For marking word wiThe ratio of total degree with training set t all marks word in the number of times occurring in training set t, that is,
p(wi,b1,…,bn) computational methods be:
Wherein p (j) represent in image collection p with
Machine extracts the probability of a width training image j;p(wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/j)
Represent in training image j that vision lemma b occurskPosterior probability;
E. basisCalculate p (i/a);
F. by formula a=arg maxaP (i/a) p (a) calculates the mark phrase a of image i to be marked.
In such scheme, p (w in step di/ j) and p (bk/ j) computational methods be respectively as follows:
Wherein, αjWith βjFor smoothing parameter, it is an empirical value;
#(wi, j) represent mark word wiWhether training image j occurs, if it is, # (wi, j)=1, otherwise # (wi,
J)=0;
#(wi, t) represent mark word wiWhether training set t occurs, if it is, # (wi, t)=1, otherwise # (wi,
T)=0;
#(bk, j) represent vision lemma bkWhether training image j occurs, if it is, # (bk, j)=1, otherwise #
(bk, j)=0;
| j | represents the total number of mark word and vision lemma in training image j;| t | represent training set t in mark word and
The total number of vision lemma.
In sum, technical scheme proposed by the present invention combining mark word and the image in joint media correlation model
The prior probability that probability calculation process is converted into the probability and mark phrase occurring in mark entry part hypograph two-part is asked
Solution, greatly reduces high frequency candidate and marks word for the impact of probability statistics model so that non-high frequency candidate mark word plays more
Big effect, improves recall ratio and the precision ratio that non-high frequency candidate marks word, is incorporated into semantic similitude language model simultaneously
In the middle of joint media correlation model, remove to estimate the prior probability of one group of mark word by semantic similitude language model, so more have
One group of higher mark word of semantic dependency may be produced.Thus improving the entirety mark effect of image.
Brief description
Fig. 1 is the flow chart of the embodiment of the present invention.
Specific embodiment
For making becoming more apparent of the object, technical solutions and advantages of the present invention expression, below in conjunction with the accompanying drawings and specifically
The present invention is further described in more detail for embodiment.
The technical scheme is that
A. according to formulaCalculate the semantic vector of each of training set t mark word w, word w will be marked
It is expressed as vector form w=< v1,v2,…,vm>, wherein, ciFor context relation word, total m context relation word, p (ci)
For context relation word ciOverall distribution probability, p (ci/ w) represent context relation word ciWith mark word w in training set t
Co-occurrence number of times with mark word w training set t in appearance total degree ratio, that is,Wherein,
Described context relation word is the mark word in training set t;
B. according to formulaCalculate the semantic similarity between mark word, wherein | | | | it is
Vectorial mould calculates, wi·wjFor dot product computing;
C. according to formulaCalculate p (a), wherein a is mark phrase { w1,
w2,…wn, n is the number of mark phrase;
D. according to formulaDesign conditions Probability p (i/wi), wherein, p (wi)
For marking word wiThe ratio of total degree with training set t all marks word in the number of times occurring in training set t, that is,
p(wi,b1,…,bn) computational methods be:
Wherein p (j) represent in image collection p with
Machine extracts the probability of a width training image j;p(wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/j)
Represent in training image j that vision lemma b occurskPosterior probability;
E. basisCalculate p (i/a);
F. by formula a=arg maxaP (i/a) p (a) calculates the mark phrase a of image i to be marked.
Image labeling problem can be defined as at present: a given training set t, this training set t comprise image collection p
With mark set of words w, and each image piAll complete to mark word mark, the mark word of all images constitutes mark set of words w,
From described mark set of words w, how to choose one group of mark word a therein one width new images i is labeled?
The image labeling method of the present invention adopts probabilistic model, and its target seeks to find mark phrase a, its conditional probability
P (a/i) maximum it may be assumed that
A=arg maxap(a/i) (3)
Wherein a is a mark phrase { w1,w2,…wn, image i is with one group of visual signature { b1,b2…,bmRepresent, lead to
Cross and image i is pre-processed with (the such as operation such as image segmentation, feature extraction, characteristic value normalization) and image block region
Sort out computing to obtain.P (a/i) can be rewritten as following form:
Because the prior probability of piece image is typically considered, obedience is equally distributed, and therefore p (i) can be regarded as one
Individual constant, and
P (a, i)=p (i/a) p (a) (5)
With formula (4), (5), formula (3) is simplified, obtains:
A=arg maxap(i/a)p(a) (6)
By p (i/a) and two probability of p (a) combine, solve maximum to find optimal mark phrase a.p(i/a)
Can obtain from original image marking model, and p (a) can obtain from language model.By giving not to two probability
With weight to represent original image model and language model to the final mark influential effect ability obtaining:
It is carried out with following form conversion:
A=arg maxa(λ1log p(i/a)+λ2log p(a)) (8)
Can be in the hope of going out to mark phrase a as long as calculating p (a) and p (i/a).Wherein, λ1With λ2It is in training image collection
Determine during machine learning and model foundation, be two constants during carrying out the automatic marking of testing image.
Technical solution of the present invention is illustrated below, image to be marked is taking the training set t comprising l image as a example
i.The l image construction image collection p=[p of training set t1p2… pl];Each image labeling of training set t has n
Mark word, in training set t, all mark words constitute mark set of words w=[w1w2… ws];Each image of training set t
There is corresponding vision lemma, in training set t, all vision lemmas constitute visual word unit set b=[b1b2… by].
Fig. 1 is the flow chart of the present embodiment, as shown in figure 1, comprising the following steps:
Step 101: image i to be marked is carried out with Image semantic classification and segmented areas sort out computing.
In this step, Image semantic classification (image segmentation, feature extraction, characteristic value normalization are carried out to image i to be marked
Deng), then carrying out image block region sorts out computing, using clustering algorithm, each image block region is sorted out, is used in combination
The combination of visual word unit represents image vision content: i={ i1i2… if}.The preparation method of vision lemma is prior art, this
Place is no longer described in detail.
Step 102: p (a) is calculated by semantic similitude language model.
In order to introduce correlation information between mark word in the similarity between mark word, present invention employs semanteme
Vector model represents each mark word w: context relation set of words c=[c1c2… cm], each element ciRepresent one
Individual context relation word, total m context relation word, all marks of mark set of words w in training set t can be chosen
Word is as context relation word, i.e. c=w.Each marks word w context relation word vector representation associated there, i.e. w
=< v1,v2,…,vm>, the calculating of each of which semanteme component vi is defined as context relation word ciWith respect to mark word w's
Conditional probability and context relation word ciProbability ratio:
Wherein p (ci) represent context relation word ciOverall distribution probability, for being uniformly distributed.Conditional probability p (ci/ w) table
Show context relation word ciWith the co-occurrence number of times marking during all image labelings of image collection p in training set t for the word w and mark
The ratio of the total degree that note word w occurs in all image labelings of image collection p:
p(ci/ w) represent the intensity distribution of vocabulary w and context relation Term co-occurrence, then divided by each context relation word
Overall probability be exactly in order to prevent semantic vector w=< v1,v2,…,vm> led by the high context relation word of the frequency of occurrences
Lead, because high-frequency conjunctive word will also tend to very big conditional probability.As shown in table 1, wherein " sky ", " sun ",
" clouds ", " town " represent one group of context relation word, and " tree ", " building ", " river " are one group of mark word, mark
The semantic vector of note word represents as shown in table 1.
Table 1
sky | sun | clouds | town | |
tree | 2.56 | 0.91 | 0.74 | 0.63 |
building | 5.01 | 0.57 | 2.41 | 21.19 |
river | 2.57 | 2.57 | 1.12 | 5.72 |
Then the semantic similarity between mark word will be calculated.The calculating of similarity is as shown in Equation 11:
Wherein | | | | calculate for vectorial mould.
wi·wjCalculating as shown in Equation 12:
Wherein ckRepresent context relation word.Between mark word, semantic similarity is as shown in table 2.Similarity span is
0 to 1, the higher similarity representing between two mark words of numerical value is higher, and the probability that they occur in same piece image is got over
Greatly.
Table 2
It is assumed that in same mark, mark vocabulary is semantic related to context relation word, then one group of mark word a=
{w1,w2,…,wnProbability p (a) can by calculate each mark word and other mark word between similarity obtain:
Formula 10,11,12 is updated in formula 13, you can calculate mark phrase Probability p (a):
Step 103: p (i/a) is calculated by joint media correlation model.
In this step, first according to formulaDesign conditions Probability p (i/
wi).Wherein,
p(wi) computational methods be:
With marking word wiThe ratio that the number of times occurring in training set t total degree with all mark words represents vocabulary
wiPrior probability p (wi):
p(wi,b1,…,bn) computational methods be:
P (j) represents the probability randomly drawing a width training image j in image collection p, generally assumes that as being uniformly distributed;p
(wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/ j) represent in training image j that vision lemma occurs
bkPosterior probability.The probable value of each is estimated as follows:
Wherein, αjWith βjFor smoothing parameter, it is an empirical value;#(wi, j) represent mark word wiTraining image j is
No appearance, if it is, # (wi, j)=1, otherwise # (wi, j)=0;#(wi, t) represent mark word wiIn training set t whether
Occur, if it is, # (wi, t)=1, otherwise # (wi, t)=0;#(bk, j) represent vision lemma bkIn training image j whether
Occur, if it is, # (bk, j)=1, otherwise # (bk, j)=0;| j | represents the total of mark word and vision lemma in training image j
Number;| t | represents the total number of mark word and vision lemma in training set t.
Then p (i/a) can with approximate evaluation go out for
Step 104: calculate the phrase to be marked of image i to be marked.
Below solved p (a) and p (i/a) respectively, according to a=arg maxa(λ1log p(i/a)+λ2Log p (a)) can
Calculate mark phrase a for piece image i
The foregoing is only presently preferred embodiments of the present invention, not in order to limit the present invention, all essences in the present invention
Any modification, equivalent substitution and improvement done within god and principle etc., should be included within the scope of protection of the invention.
Claims (2)
1. a kind of image automatic annotation method of word-based correlation is it is characterised in that training set t comprises l image, institute
State l image construction image collection p=[p1p2… pl];Each image labeling of described training set t has n mark word,
In training set t, all mark words constitute mark set of words w=[w1w2… ws];Each image of training set t has accordingly
Vision lemma, in training set t, all vision lemmas constitute the first set b=[b of visual word1b2… by], image to be marked
For i, the method includes:
A. according to formulaCalculate the semantic vector of each of training set t mark word w, word w will be marked and represent
For vector form w=< v1,v2,…,vm>, wherein, ciFor context relation word, total m context relation word, p (ci) it is upper
Hereafter conjunctive word ciOverall distribution probability, p (ci/ w) represent context relation word ciWith mark word w being total in training set t
The ratio of the total degree that occurrence number is occurred in training set t with mark word w, that is,Wherein, described
Context relation word is the mark word in training set t;
B. according to formulaCalculate the semantic similarity between mark word, wherein | | | | for vectorial mould
Calculate, wi·wjFor dot product computing;
C. according to formulaCalculate p (a), wherein a is mark phrase { w1,w2,…wn,
N is the number of mark phrase;
D. according to formulaDesign conditions Probability p (i/wi), wherein, p (wi) for marking
Note word wiThe ratio of total degree with training set t all marks word in the number of times occurring in training set t, that is,
p(wi,b1,…,bn) computational methods be:
Wherein p (j) expression is taken out in image collection p at random
Take the probability of a width training image j;p(wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/ j) represent
Vision lemma b occurs in training image jkPosterior probability;
E. basisCalculate p (i/a);
F. by formula a=argmaxaP (i/a) p (a) calculates the mark phrase a of image i to be marked.
2. method according to claim 1 is it is characterised in that p (w in step di/ j) and p (bk/ j) computational methods respectively
For:
Wherein, αjWith βjFor smoothing parameter, it is an empirical value;
#(wi, j) represent mark word wiWhether training image j occurs, if it is, # (wi, j)=1, otherwise # (wi, j)=
0;
#(wi, t) represent mark word wiWhether training set t occurs, if it is, # (wi, t)=1, otherwise # (wi, t)=
0;
#(bk, j) represent vision lemma bkWhether training image j occurs, if it is, # (bk, j)=1, otherwise # (bk,j)
=0;
| j | represents the total number of mark word and vision lemma in training image j;| t | represents mark word and vision in training set t
The total number of lemma.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410008553.1A CN103714178B (en) | 2014-01-08 | 2014-01-08 | Automatic image marking method based on word correlation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410008553.1A CN103714178B (en) | 2014-01-08 | 2014-01-08 | Automatic image marking method based on word correlation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103714178A CN103714178A (en) | 2014-04-09 |
CN103714178B true CN103714178B (en) | 2017-01-25 |
Family
ID=50407153
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410008553.1A Active CN103714178B (en) | 2014-01-08 | 2014-01-08 | Automatic image marking method based on word correlation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103714178B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104794183A (en) * | 2015-04-10 | 2015-07-22 | 浙江大学 | Picture labeling method based on multiple views and multiple labels |
WO2017137439A1 (en) * | 2016-02-08 | 2017-08-17 | Koninklijke Philips N.V. | Device for and method of determining clusters |
CN108268875B (en) * | 2016-12-30 | 2020-12-08 | 广东精点数据科技股份有限公司 | Image semantic automatic labeling method and device based on data smoothing |
CN110162644B (en) | 2018-10-10 | 2022-12-20 | 腾讯科技(深圳)有限公司 | Image set establishing method, device and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1920820A (en) * | 2006-09-14 | 2007-02-28 | 浙江大学 | Image meaning automatic marking method based on marking significance sequence |
CN101620615A (en) * | 2009-08-04 | 2010-01-06 | 西南交通大学 | Automatic image annotation and translation method based on decision tree learning |
CN101685464A (en) * | 2009-06-18 | 2010-03-31 | 浙江大学 | Method for automatically labeling images based on community potential subject excavation |
CN102298606A (en) * | 2011-06-01 | 2011-12-28 | 清华大学 | Random walking image automatic annotation method and device based on label graph model |
CN102542067A (en) * | 2012-01-06 | 2012-07-04 | 上海交通大学 | Automatic image semantic annotation method based on scale learning and correlated label dissemination |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101587478B (en) * | 2008-05-20 | 2013-07-24 | 株式会社理光 | Methods and devices for training, automatically labeling and searching images |
-
2014
- 2014-01-08 CN CN201410008553.1A patent/CN103714178B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1920820A (en) * | 2006-09-14 | 2007-02-28 | 浙江大学 | Image meaning automatic marking method based on marking significance sequence |
CN101685464A (en) * | 2009-06-18 | 2010-03-31 | 浙江大学 | Method for automatically labeling images based on community potential subject excavation |
CN101620615A (en) * | 2009-08-04 | 2010-01-06 | 西南交通大学 | Automatic image annotation and translation method based on decision tree learning |
CN102298606A (en) * | 2011-06-01 | 2011-12-28 | 清华大学 | Random walking image automatic annotation method and device based on label graph model |
CN102542067A (en) * | 2012-01-06 | 2012-07-04 | 上海交通大学 | Automatic image semantic annotation method based on scale learning and correlated label dissemination |
Non-Patent Citations (3)
Title |
---|
"Hidden-Concept Driven Multilabel Image Annotation and Label Ranking";Bing-Kun Bao et al.;《IEEE Transactions on Multimedia》;20120229;第14卷(第1期);第199-210页 * |
"Latent Semantic Analysis-based Image Auto Annotation";Mahdia Bakalem et al.;《IEEE Conf. on Machine and Web Intelligence》;20101231;第2010年卷;第460-463页 * |
"词间相关性的CMRM图像标注方法";刘咏梅等;《智能系统学报》;20110831;第6卷(第4期);第350-354页 * |
Also Published As
Publication number | Publication date |
---|---|
CN103714178A (en) | 2014-04-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103617157B (en) | Based on semantic Text similarity computing method | |
Pang et al. | Text matching as image recognition | |
WO2018010365A1 (en) | Cross-media search method | |
CN105005553B (en) | Short text Sentiment orientation analysis method based on sentiment dictionary | |
CN104391942B (en) | Short essay eigen extended method based on semantic collection of illustrative plates | |
CN106294593B (en) | In conjunction with the Relation extraction method of subordinate clause grade remote supervisory and semi-supervised integrated study | |
CN102902821B (en) | The image high-level semantics mark of much-talked-about topic Network Based, search method and device | |
CN103218444B (en) | Based on semantic method of Tibetan language webpage text classification | |
CN104199857B (en) | A kind of tax document hierarchy classification method based on multi-tag classification | |
CN103678564B (en) | Internet product research system based on data mining | |
CN104794169B (en) | A kind of subject terminology extraction method and system based on sequence labelling model | |
CN104881458B (en) | A kind of mask method and device of Web page subject | |
CN104778161A (en) | Keyword extracting method based on Word2Vec and Query log | |
CN103823859B (en) | Name recognition algorithm based on combination of decision-making tree rules and multiple statistic models | |
CN106202256A (en) | Propagate based on semanteme and mix the Web graph of multi-instance learning as search method | |
CN105320960A (en) | Voting based classification method for cross-language subjective and objective sentiments | |
CN109492105B (en) | Text emotion classification method based on multi-feature ensemble learning | |
CN106095829A (en) | Cross-media retrieval method based on degree of depth study with the study of concordance expression of space | |
CN102637192A (en) | Method for answering with natural language | |
Afzal et al. | Mayonlp at semeval-2016 task 1: Semantic textual similarity based on lexical semantic net and deep learning semantic model | |
CN103714178B (en) | Automatic image marking method based on word correlation | |
CN106294863A (en) | A kind of abstract method for mass text fast understanding | |
CN105843799B (en) | A kind of academic paper label recommendation method based on multi-source heterogeneous information graph model | |
CN102495865A (en) | Image annotation method combined with image internal space relation and visual symbiosis relation | |
CN108363725A (en) | A kind of method of the extraction of user comment viewpoint and the generation of viewpoint label |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |