CN103714178B - Automatic image marking method based on word correlation - Google Patents

Automatic image marking method based on word correlation Download PDF

Info

Publication number
CN103714178B
CN103714178B CN201410008553.1A CN201410008553A CN103714178B CN 103714178 B CN103714178 B CN 103714178B CN 201410008553 A CN201410008553 A CN 201410008553A CN 103714178 B CN103714178 B CN 103714178B
Authority
CN
China
Prior art keywords
word
image
mark
training set
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201410008553.1A
Other languages
Chinese (zh)
Other versions
CN103714178A (en
Inventor
安震
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Jingdong Century Trading Co Ltd
Beijing Jingdong Shangke Information Technology Co Ltd
Original Assignee
Beijing Jingdong Century Trading Co Ltd
Beijing Jingdong Shangke Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Jingdong Century Trading Co Ltd, Beijing Jingdong Shangke Information Technology Co Ltd filed Critical Beijing Jingdong Century Trading Co Ltd
Priority to CN201410008553.1A priority Critical patent/CN103714178B/en
Publication of CN103714178A publication Critical patent/CN103714178A/en
Application granted granted Critical
Publication of CN103714178B publication Critical patent/CN103714178B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2155Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling

Abstract

The invention discloses an automatic image marking method based on work correlation. A training set T comprises l images, n marking words are marked on each image of the training set T, the training set T is provided with a corresponding vision lemma, and the image to be marked is I. The method includes the steps that a semantic vector of each marking word w is calculated according to a formula, the marking word w is represented by the vector form w=<v1, v2,......, vm>, ci is an associated word in a context, and the associated words in the context total m; semantic similarity of the marking words is calculated according to a formula, and vector module calculated is achieved as is shown in the specification; p(A) is calculated according to the formula, wherein A is a marking word group in w1, w2,......wn, and n is the number of the marking word groups; the conditional probability p(I/wi) is calculated according to a formula; the marking word group A of the image I to be marked is calculated according to the formula A=arg maxAp(I/A) p(A).

Description

A kind of image automatic annotation method of word-based correlation
Technical field
The present invention relates to image processing field, the image automatic annotation method of particularly a kind of word-based correlation.
Background technology
With the high speed development of multimedia and Internet technology, daily life, work are believed to multimedias such as images The dependence of breath is more and more stronger.It is intended to based on the retrieval that semantic image retrieval can not only accurately express user, be also convenient to use The use at family, therefore this retrieval mode not only become a kind of important form of image retrieval, and become what researcher chased Hot technology.
And automatic image annotation technology is an important and challenging job in Semantic Image Retrieval, image is certainly The appearance of dynamic label technology is to automatically obtain the semantic information comprising in image vision content, and it attempts to regard in image bottom Feel and build a bridge block, thus support is made to semantic retrieval in semantic level between feature and high-level semantic.Therefore, it is based on The automatic marking algorithm research of image, semantic has become a very active research branch and crucial skill in field of image search Art, has good application prospect and researching value.
Automatic image annotation is exactly to allow the image automatically to no mark for the computer add the semanteme that can reflect picture material Keyword.It practises semantic concept space and vision spy automatically using the image collection of mark or other obtainable information Levy the relational model in space, and mark the image of unknown semantics with this model.By in the high-layer semantic information of image and bottom Set up a kind of mapping relations between feature, solve the problems, such as semantic gap to a certain extent.
The image automatic annotation method of joint media correlation model is to be currently based in generation model image labeling method to answer With a kind of most commonly used image labeling algorithm, obtain the widely studied of scholar.The basic thought of this marking model is to utilize The method of probability statistics sets up the probability relativity in Image Visual Feature space and semantic concept space, by both statistical learnings Between the joint probability distribution that exists, it is maximum to find out the joint probability that one group of semantic tagger word is allowed to and picture material between, and This group is marked word as the final mark of testing image.
But joint media correlation model belongs to one kind of probabilistic model, this class model mark word high to occurrence frequency has Skewed popularity.Secondly, in joint media correlation model automatic image annotation mark method, different candidate's mark words are in annotation process It is assumed to be it is separate, and the correlation marking between word is not fully excavated.In fact with piece image, no There is multiple association such as symbiosis, level or space between mark word.
Such as one width contains the image of semantic objects such as " sun, sky, cloud, mountain, tree ", regards from image Feel that " sun " can not depart from " sky " this language it can be seen that " sun " and " sky " object has certain spatial correlation in content Adopted object and be individually present;Equally, for " mountain " and " tree " two semantic objects in picture material, The presence of " tree " object is with " mountain " semantic object for vision content background, and the two is same in image vision content There is the contact of interwoveness it is impossible to utterly assume that this two mark words are separate being labeled.Therefore, joint media Correlation model Automatic image annotation algorithm thinks that in annotation process different candidates mark separate way between word and exist Certain defect, may lead to because ignoring correlation between word mark semantic inconsistent phenomenon between word in annotation results.
Content of the invention
In view of this, the invention provides a kind of image automatic annotation method of word-based correlation, combined with overcoming Media correlation model Automatic image annotation algorithm thinks that in annotation process different candidates mark separate way between word The defect existing, solves the problems, such as to lead to because ignoring correlation between word in annotation results semanteme between mark word inconsistent.This The technical scheme that invention proposes is:
A kind of image automatic annotation method of word-based correlation, training set t comprises l image, described l image Constitute image collection p=[p1p2… pl];Each image labeling of described training set t has n mark word, trains set t In all mark words constitute mark set of words w=[w1w2… ws];Each image of training set t has corresponding visual word Unit, in training set t, all vision lemmas constitute visual word unit set b=[b1b2… by], image to be marked is i, the party Method includes:
A. according to formulaCalculate the semantic vector of each of training set t mark word w, word w will be marked It is expressed as vector form w=< v1,v2,…,vm>, wherein, ciFor context relation word, total m context relation word, p (ci) For context relation word ciOverall distribution probability, p (ci/ w) represent context relation word ciWith mark word w in training set t Co-occurrence number of times with mark word w training set t in appearance total degree ratio, that is,Wherein, Described context relation word is the mark word in training set t;
B. according to formulaCalculate the semantic similarity between mark word, wherein | | | | it is Vectorial mould calculates, wi·wjFor dot product computing;
C. according to formulaCalculate p (a), wherein a is mark phrase { w1, w2,…wn, n is the number of mark phrase;
D. according to formulaDesign conditions Probability p (i/wi), wherein, p (wi) For marking word wiThe ratio of total degree with training set t all marks word in the number of times occurring in training set t, that is,
p(wi,b1,…,bn) computational methods be:
Wherein p (j) represent in image collection p with Machine extracts the probability of a width training image j;p(wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/j) Represent in training image j that vision lemma b occurskPosterior probability;
E. basisCalculate p (i/a);
F. by formula a=arg maxaP (i/a) p (a) calculates the mark phrase a of image i to be marked.
In such scheme, p (w in step di/ j) and p (bk/ j) computational methods be respectively as follows:
p ( w i | j ) = ( 1 - &alpha; j ) # ( w i , j ) | j | + &alpha; j # ( w i , t ) | t | - - - ( 1 )
p ( b k | j ) = ( 1 - &beta; j ) # ( b k , j ) | j | + &beta; j # ( b k , t ) | t | - - - ( 2 )
Wherein, αjWith βjFor smoothing parameter, it is an empirical value;
#(wi, j) represent mark word wiWhether training image j occurs, if it is, # (wi, j)=1, otherwise # (wi, J)=0;
#(wi, t) represent mark word wiWhether training set t occurs, if it is, # (wi, t)=1, otherwise # (wi, T)=0;
#(bk, j) represent vision lemma bkWhether training image j occurs, if it is, # (bk, j)=1, otherwise # (bk, j)=0;
| j | represents the total number of mark word and vision lemma in training image j;| t | represent training set t in mark word and The total number of vision lemma.
In sum, technical scheme proposed by the present invention combining mark word and the image in joint media correlation model The prior probability that probability calculation process is converted into the probability and mark phrase occurring in mark entry part hypograph two-part is asked Solution, greatly reduces high frequency candidate and marks word for the impact of probability statistics model so that non-high frequency candidate mark word plays more Big effect, improves recall ratio and the precision ratio that non-high frequency candidate marks word, is incorporated into semantic similitude language model simultaneously In the middle of joint media correlation model, remove to estimate the prior probability of one group of mark word by semantic similitude language model, so more have One group of higher mark word of semantic dependency may be produced.Thus improving the entirety mark effect of image.
Brief description
Fig. 1 is the flow chart of the embodiment of the present invention.
Specific embodiment
For making becoming more apparent of the object, technical solutions and advantages of the present invention expression, below in conjunction with the accompanying drawings and specifically The present invention is further described in more detail for embodiment.
The technical scheme is that
A. according to formulaCalculate the semantic vector of each of training set t mark word w, word w will be marked It is expressed as vector form w=< v1,v2,…,vm>, wherein, ciFor context relation word, total m context relation word, p (ci) For context relation word ciOverall distribution probability, p (ci/ w) represent context relation word ciWith mark word w in training set t Co-occurrence number of times with mark word w training set t in appearance total degree ratio, that is,Wherein, Described context relation word is the mark word in training set t;
B. according to formulaCalculate the semantic similarity between mark word, wherein | | | | it is Vectorial mould calculates, wi·wjFor dot product computing;
C. according to formulaCalculate p (a), wherein a is mark phrase { w1, w2,…wn, n is the number of mark phrase;
D. according to formulaDesign conditions Probability p (i/wi), wherein, p (wi) For marking word wiThe ratio of total degree with training set t all marks word in the number of times occurring in training set t, that is,
p(wi,b1,…,bn) computational methods be:
Wherein p (j) represent in image collection p with Machine extracts the probability of a width training image j;p(wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/j) Represent in training image j that vision lemma b occurskPosterior probability;
E. basisCalculate p (i/a);
F. by formula a=arg maxaP (i/a) p (a) calculates the mark phrase a of image i to be marked.
Image labeling problem can be defined as at present: a given training set t, this training set t comprise image collection p With mark set of words w, and each image piAll complete to mark word mark, the mark word of all images constitutes mark set of words w, From described mark set of words w, how to choose one group of mark word a therein one width new images i is labeled?
The image labeling method of the present invention adopts probabilistic model, and its target seeks to find mark phrase a, its conditional probability P (a/i) maximum it may be assumed that
A=arg maxap(a/i) (3)
Wherein a is a mark phrase { w1,w2,…wn, image i is with one group of visual signature { b1,b2…,bmRepresent, lead to Cross and image i is pre-processed with (the such as operation such as image segmentation, feature extraction, characteristic value normalization) and image block region Sort out computing to obtain.P (a/i) can be rewritten as following form:
p ( a / i ) = p ( a , i ) p ( i ) - - - ( 4 )
Because the prior probability of piece image is typically considered, obedience is equally distributed, and therefore p (i) can be regarded as one Individual constant, and
P (a, i)=p (i/a) p (a) (5)
With formula (4), (5), formula (3) is simplified, obtains:
A=arg maxap(i/a)p(a) (6)
By p (i/a) and two probability of p (a) combine, solve maximum to find optimal mark phrase a.p(i/a) Can obtain from original image marking model, and p (a) can obtain from language model.By giving not to two probability With weight to represent original image model and language model to the final mark influential effect ability obtaining:
a = arg max a p ( i / a ) &lambda; 1 p ( a ) &lambda; 2 - - - ( 7 )
It is carried out with following form conversion:
A=arg maxa1log p(i/a)+λ2log p(a)) (8)
Can be in the hope of going out to mark phrase a as long as calculating p (a) and p (i/a).Wherein, λ1With λ2It is in training image collection Determine during machine learning and model foundation, be two constants during carrying out the automatic marking of testing image.
Technical solution of the present invention is illustrated below, image to be marked is taking the training set t comprising l image as a example i.The l image construction image collection p=[p of training set t1p2… pl];Each image labeling of training set t has n Mark word, in training set t, all mark words constitute mark set of words w=[w1w2… ws];Each image of training set t There is corresponding vision lemma, in training set t, all vision lemmas constitute visual word unit set b=[b1b2… by].
Fig. 1 is the flow chart of the present embodiment, as shown in figure 1, comprising the following steps:
Step 101: image i to be marked is carried out with Image semantic classification and segmented areas sort out computing.
In this step, Image semantic classification (image segmentation, feature extraction, characteristic value normalization are carried out to image i to be marked Deng), then carrying out image block region sorts out computing, using clustering algorithm, each image block region is sorted out, is used in combination The combination of visual word unit represents image vision content: i={ i1i2… if}.The preparation method of vision lemma is prior art, this Place is no longer described in detail.
Step 102: p (a) is calculated by semantic similitude language model.
In order to introduce correlation information between mark word in the similarity between mark word, present invention employs semanteme Vector model represents each mark word w: context relation set of words c=[c1c2… cm], each element ciRepresent one Individual context relation word, total m context relation word, all marks of mark set of words w in training set t can be chosen Word is as context relation word, i.e. c=w.Each marks word w context relation word vector representation associated there, i.e. w =< v1,v2,…,vm>, the calculating of each of which semanteme component vi is defined as context relation word ciWith respect to mark word w's Conditional probability and context relation word ciProbability ratio:
v i = p ( c i / w ) p ( c i ) - - - ( 9 )
Wherein p (ci) represent context relation word ciOverall distribution probability, for being uniformly distributed.Conditional probability p (ci/ w) table Show context relation word ciWith the co-occurrence number of times marking during all image labelings of image collection p in training set t for the word w and mark The ratio of the total degree that note word w occurs in all image labelings of image collection p:
p ( c i / w ) = c o u n t ( c i , w ) c o u n t ( w ) - - - ( 10 )
p(ci/ w) represent the intensity distribution of vocabulary w and context relation Term co-occurrence, then divided by each context relation word Overall probability be exactly in order to prevent semantic vector w=< v1,v2,…,vm> led by the high context relation word of the frequency of occurrences Lead, because high-frequency conjunctive word will also tend to very big conditional probability.As shown in table 1, wherein " sky ", " sun ", " clouds ", " town " represent one group of context relation word, and " tree ", " building ", " river " are one group of mark word, mark The semantic vector of note word represents as shown in table 1.
Table 1
sky sun clouds town
tree 2.56 0.91 0.74 0.63
building 5.01 0.57 2.41 21.19
river 2.57 2.57 1.12 5.72
Then the semantic similarity between mark word will be calculated.The calculating of similarity is as shown in Equation 11:
s i m ( w i , w j ) = w i &centerdot; w j | | w i | | &centerdot; | | w j | | - - - ( 11 )
Wherein | | | | calculate for vectorial mould.
wi·wjCalculating as shown in Equation 12:
w i &centerdot; w j = &sigma; k = 1 m v w i , k v w j , k = &sigma; k = 1 m p ( c k / w i ) p ( c k ) &centerdot; p ( c k / w j ) p ( c k ) - - - ( 12 )
Wherein ckRepresent context relation word.Between mark word, semantic similarity is as shown in table 2.Similarity span is 0 to 1, the higher similarity representing between two mark words of numerical value is higher, and the probability that they occur in same piece image is got over Greatly.
Table 2
It is assumed that in same mark, mark vocabulary is semantic related to context relation word, then one group of mark word a= {w1,w2,…,wnProbability p (a) can by calculate each mark word and other mark word between similarity obtain:
p ( a ) &proportional; 1 n - 1 &sigma; w i &element; a &sigma; w j &element; a , j &notequal; i s i m ( w i , w j ) - - - ( 13 )
Formula 10,11,12 is updated in formula 13, you can calculate mark phrase Probability p (a):
p ( a ) &proportional; 1 n - 1 &sigma; w i &element; a &sigma; w j &element; a , j &notequal; i &sigma; k = 1 m c o u n t ( c k , w i ) c o u n t ( w i ) &centerdot; p ( c k ) &centerdot; c o u n t ( c k , w j ) c o u n t ( w j ) &centerdot; p ( c k ) | | w i | | &centerdot; | | w j | | - - - ( 14 )
Step 103: p (i/a) is calculated by joint media correlation model.
In this step, first according to formulaDesign conditions Probability p (i/ wi).Wherein,
p(wi) computational methods be:
With marking word wiThe ratio that the number of times occurring in training set t total degree with all mark words represents vocabulary wiPrior probability p (wi):
p ( w i ) = | w i | &sigma; w k &element; t | w k | - - - ( 15 )
p(wi,b1,…,bn) computational methods be:
p ( w i , b 1 , b 2 , ... , b n ) = &sigma; j &element; t p ( j ) p ( w i | j ) &pi; k = 1 n p ( b k | j ) - - - ( 16 )
P (j) represents the probability randomly drawing a width training image j in image collection p, generally assumes that as being uniformly distributed;p (wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/ j) represent in training image j that vision lemma occurs bkPosterior probability.The probable value of each is estimated as follows:
p ( w i | j ) = ( 1 - &alpha; j ) # ( w i , j ) | j | + &alpha; j # ( w i , t ) | t | - - - ( 17 )
p ( b k | j ) = ( 1 - &beta; j ) # ( b k , j ) | j | + &beta; j # ( b k , t ) | t | - - - ( 18 )
Wherein, αjWith βjFor smoothing parameter, it is an empirical value;#(wi, j) represent mark word wiTraining image j is No appearance, if it is, # (wi, j)=1, otherwise # (wi, j)=0;#(wi, t) represent mark word wiIn training set t whether Occur, if it is, # (wi, t)=1, otherwise # (wi, t)=0;#(bk, j) represent vision lemma bkIn training image j whether Occur, if it is, # (bk, j)=1, otherwise # (bk, j)=0;| j | represents the total of mark word and vision lemma in training image j Number;| t | represents the total number of mark word and vision lemma in training set t.
Then p (i/a) can with approximate evaluation go out for
Step 104: calculate the phrase to be marked of image i to be marked.
Below solved p (a) and p (i/a) respectively, according to a=arg maxa1log p(i/a)+λ2Log p (a)) can Calculate mark phrase a for piece image i
The foregoing is only presently preferred embodiments of the present invention, not in order to limit the present invention, all essences in the present invention Any modification, equivalent substitution and improvement done within god and principle etc., should be included within the scope of protection of the invention.

Claims (2)

1. a kind of image automatic annotation method of word-based correlation is it is characterised in that training set t comprises l image, institute State l image construction image collection p=[p1p2… pl];Each image labeling of described training set t has n mark word, In training set t, all mark words constitute mark set of words w=[w1w2… ws];Each image of training set t has accordingly Vision lemma, in training set t, all vision lemmas constitute the first set b=[b of visual word1b2… by], image to be marked For i, the method includes:
A. according to formulaCalculate the semantic vector of each of training set t mark word w, word w will be marked and represent For vector form w=< v1,v2,…,vm>, wherein, ciFor context relation word, total m context relation word, p (ci) it is upper Hereafter conjunctive word ciOverall distribution probability, p (ci/ w) represent context relation word ciWith mark word w being total in training set t The ratio of the total degree that occurrence number is occurred in training set t with mark word w, that is,Wherein, described Context relation word is the mark word in training set t;
B. according to formulaCalculate the semantic similarity between mark word, wherein | | | | for vectorial mould Calculate, wi·wjFor dot product computing;
C. according to formulaCalculate p (a), wherein a is mark phrase { w1,w2,…wn, N is the number of mark phrase;
D. according to formulaDesign conditions Probability p (i/wi), wherein, p (wi) for marking Note word wiThe ratio of total degree with training set t all marks word in the number of times occurring in training set t, that is,
p(wi,b1,…,bn) computational methods be:
Wherein p (j) expression is taken out in image collection p at random Take the probability of a width training image j;p(wi/ j) represent in training image j that vocabulary w occursiPosterior probability;And p (bk/ j) represent Vision lemma b occurs in training image jkPosterior probability;
E. basisCalculate p (i/a);
F. by formula a=argmaxaP (i/a) p (a) calculates the mark phrase a of image i to be marked.
2. method according to claim 1 is it is characterised in that p (w in step di/ j) and p (bk/ j) computational methods respectively For:
p ( w i | j ) = ( 1 - &alpha; j ) # ( w i , j ) | j | + &alpha; j # ( w i , t ) | t | - - - ( 1 )
p ( b k | j ) = ( 1 - &beta; j ) # ( b k , j ) | j | + &beta; j # ( b k , t ) | t | - - - ( 2 )
Wherein, αjWith βjFor smoothing parameter, it is an empirical value;
#(wi, j) represent mark word wiWhether training image j occurs, if it is, # (wi, j)=1, otherwise # (wi, j)= 0;
#(wi, t) represent mark word wiWhether training set t occurs, if it is, # (wi, t)=1, otherwise # (wi, t)= 0;
#(bk, j) represent vision lemma bkWhether training image j occurs, if it is, # (bk, j)=1, otherwise # (bk,j) =0;
| j | represents the total number of mark word and vision lemma in training image j;| t | represents mark word and vision in training set t The total number of lemma.
CN201410008553.1A 2014-01-08 2014-01-08 Automatic image marking method based on word correlation Active CN103714178B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410008553.1A CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410008553.1A CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Publications (2)

Publication Number Publication Date
CN103714178A CN103714178A (en) 2014-04-09
CN103714178B true CN103714178B (en) 2017-01-25

Family

ID=50407153

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410008553.1A Active CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Country Status (1)

Country Link
CN (1) CN103714178B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104794183A (en) * 2015-04-10 2015-07-22 浙江大学 Picture labeling method based on multiple views and multiple labels
WO2017137439A1 (en) * 2016-02-08 2017-08-17 Koninklijke Philips N.V. Device for and method of determining clusters
CN108268875B (en) * 2016-12-30 2020-12-08 广东精点数据科技股份有限公司 Image semantic automatic labeling method and device based on data smoothing
CN110162644B (en) 2018-10-10 2022-12-20 腾讯科技(深圳)有限公司 Image set establishing method, device and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920820A (en) * 2006-09-14 2007-02-28 浙江大学 Image meaning automatic marking method based on marking significance sequence
CN101620615A (en) * 2009-08-04 2010-01-06 西南交通大学 Automatic image annotation and translation method based on decision tree learning
CN101685464A (en) * 2009-06-18 2010-03-31 浙江大学 Method for automatically labeling images based on community potential subject excavation
CN102298606A (en) * 2011-06-01 2011-12-28 清华大学 Random walking image automatic annotation method and device based on label graph model
CN102542067A (en) * 2012-01-06 2012-07-04 上海交通大学 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101587478B (en) * 2008-05-20 2013-07-24 株式会社理光 Methods and devices for training, automatically labeling and searching images

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920820A (en) * 2006-09-14 2007-02-28 浙江大学 Image meaning automatic marking method based on marking significance sequence
CN101685464A (en) * 2009-06-18 2010-03-31 浙江大学 Method for automatically labeling images based on community potential subject excavation
CN101620615A (en) * 2009-08-04 2010-01-06 西南交通大学 Automatic image annotation and translation method based on decision tree learning
CN102298606A (en) * 2011-06-01 2011-12-28 清华大学 Random walking image automatic annotation method and device based on label graph model
CN102542067A (en) * 2012-01-06 2012-07-04 上海交通大学 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Hidden-Concept Driven Multilabel Image Annotation and Label Ranking";Bing-Kun Bao et al.;《IEEE Transactions on Multimedia》;20120229;第14卷(第1期);第199-210页 *
"Latent Semantic Analysis-based Image Auto Annotation";Mahdia Bakalem et al.;《IEEE Conf. on Machine and Web Intelligence》;20101231;第2010年卷;第460-463页 *
"词间相关性的CMRM图像标注方法";刘咏梅等;《智能系统学报》;20110831;第6卷(第4期);第350-354页 *

Also Published As

Publication number Publication date
CN103714178A (en) 2014-04-09

Similar Documents

Publication Publication Date Title
CN103617157B (en) Based on semantic Text similarity computing method
Pang et al. Text matching as image recognition
WO2018010365A1 (en) Cross-media search method
CN105005553B (en) Short text Sentiment orientation analysis method based on sentiment dictionary
CN104391942B (en) Short essay eigen extended method based on semantic collection of illustrative plates
CN106294593B (en) In conjunction with the Relation extraction method of subordinate clause grade remote supervisory and semi-supervised integrated study
CN102902821B (en) The image high-level semantics mark of much-talked-about topic Network Based, search method and device
CN103218444B (en) Based on semantic method of Tibetan language webpage text classification
CN104199857B (en) A kind of tax document hierarchy classification method based on multi-tag classification
CN103678564B (en) Internet product research system based on data mining
CN104794169B (en) A kind of subject terminology extraction method and system based on sequence labelling model
CN104881458B (en) A kind of mask method and device of Web page subject
CN104778161A (en) Keyword extracting method based on Word2Vec and Query log
CN103823859B (en) Name recognition algorithm based on combination of decision-making tree rules and multiple statistic models
CN106202256A (en) Propagate based on semanteme and mix the Web graph of multi-instance learning as search method
CN105320960A (en) Voting based classification method for cross-language subjective and objective sentiments
CN109492105B (en) Text emotion classification method based on multi-feature ensemble learning
CN106095829A (en) Cross-media retrieval method based on degree of depth study with the study of concordance expression of space
CN102637192A (en) Method for answering with natural language
Afzal et al. Mayonlp at semeval-2016 task 1: Semantic textual similarity based on lexical semantic net and deep learning semantic model
CN103714178B (en) Automatic image marking method based on word correlation
CN106294863A (en) A kind of abstract method for mass text fast understanding
CN105843799B (en) A kind of academic paper label recommendation method based on multi-source heterogeneous information graph model
CN102495865A (en) Image annotation method combined with image internal space relation and visual symbiosis relation
CN108363725A (en) A kind of method of the extraction of user comment viewpoint and the generation of viewpoint label

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant