CN103714178A - Automatic image marking method based on word correlation - Google Patents

Automatic image marking method based on word correlation Download PDF

Info

Publication number
CN103714178A
CN103714178A CN201410008553.1A CN201410008553A CN103714178A CN 103714178 A CN103714178 A CN 103714178A CN 201410008553 A CN201410008553 A CN 201410008553A CN 103714178 A CN103714178 A CN 103714178A
Authority
CN
China
Prior art keywords
word
mark
image
training set
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410008553.1A
Other languages
Chinese (zh)
Other versions
CN103714178B (en
Inventor
安震
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Jingdong Century Trading Co Ltd
Beijing Jingdong Shangke Information Technology Co Ltd
Original Assignee
Beijing Jingdong Century Trading Co Ltd
Beijing Jingdong Shangke Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Jingdong Century Trading Co Ltd, Beijing Jingdong Shangke Information Technology Co Ltd filed Critical Beijing Jingdong Century Trading Co Ltd
Priority to CN201410008553.1A priority Critical patent/CN103714178B/en
Publication of CN103714178A publication Critical patent/CN103714178A/en
Application granted granted Critical
Publication of CN103714178B publication Critical patent/CN103714178B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2155Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling

Abstract

The invention discloses an automatic image marking method based on work correlation. A training set T comprises l images, n marking words are marked on each image of the training set T, the training set T is provided with a corresponding vision lemma, and the image to be marked is I. The method includes the steps that a semantic vector of each marking word w is calculated according to a formula, the marking word w is represented by the vector form w=<v1, v2,......, vm>, ci is an associated word in a context, and the associated words in the context total m; semantic similarity of the marking words is calculated according to a formula, and vector module calculated is achieved as is shown in the specification; p(A) is calculated according to the formula, wherein A is a marking word group in w1, w2,......wn, and n is the number of the marking word groups; the conditional probability p(I/wi) is calculated according to a formula; the marking word group A of the image I to be marked is calculated according to the formula A=arg maxAp(I/A) p(A).

Description

A kind of image automatic annotation method based on correlativity between word
Technical field
The present invention relates to image processing field, particularly a kind of image automatic annotation method based on correlativity between word.
Background technology
Along with the high speed development of multimedia and Internet technology, daily life, work more and more stronger to the dependence of the multimedia messagess such as image.Image retrieval based on semantic can not only accurately be expressed user's retrieval intention, is also convenient to user's use, so this retrieval mode not only becomes a kind of important form of image retrieval, and becomes the hot technology that researchist chases.
And automatic image annotation technology is an important and challenging job in Semantic Image Retrieval, the appearance of automatic image annotation technology is the semantic information comprising in automatic acquisition image vision content, it attempts between image bottom visual signature and high-level semantic, to build a bridge block, thereby in semantic level, semantic retrieval is made to support.Therefore, the automatic marking algorithm research based on image, semantic has become very active research branch and a gordian technique in field of image search, has good application prospect and researching value.
Automatic image annotation is exactly to allow computing machine automatically to the image without mark, add the semantic key words that can reflect picture material.It utilizes the image collection or other the obtainable information that have marked automatically to practise the relational model in semantic concept space and visual signature space, and with this model, marks the image of unknown semantics.By setting up a kind of mapping relations between the high-layer semantic information at image and low-level image feature, solve to a certain extent semantic gap problem.
The image automatic annotation method of associating media correlation model is current a kind of image labeling algorithm based on being most widely used in generation model image labeling method, has obtained scholar's broad research.The basic thought of this marking model is to utilize the method for probability statistics to set up the probability relativity in Image Visual Feature space and semantic concept space, the joint probability distribution existing between the two by statistical learning, find out that one group of semantic tagger word makes it and picture material between joint probability maximum, and the final mark using this group mark word as testing image.
But associating media correlation model belongs to a kind of of probability model, this class model has skewed popularity to the high mark word of occurrence frequency.Secondly in associating media correlation model automatic image annotation mark method, different candidates mark word and in mark process, are assumed to be it is separate, and the correlativity marking between word is not fully excavated.In fact with in piece image, between different labeled word, exist the multiple associations such as symbiosis, level or space.
The image that has comprised semantic objects such as " sun, sky, cloud, mountain, tree " such as a width, image vision content, can find out that " sun " and " sky " object exists certain spatial correlation, " sun " can not depart from " sky " this semantic object and independent existence; Equally, for " mountain " in picture material and " tree " two semantic objects, the existence of " tree " object is to take " mountain " semantic object as vision content background, the two has equally inseparable contact in image vision content, can not suppose utterly that these two mark words are separate marks.Therefore, associating media correlation model automatic image annotation algorithm thinks that in mark process different candidates mark way separate between word and have certain defect, may cause marking in annotation results semantic inconsistent phenomenon between word because ignoring between word correlativity.
Summary of the invention
In view of this, the invention provides a kind of image automatic annotation method based on correlativity between word, to overcome associating media correlation model automatic image annotation algorithm, in mark process, think that different candidates mark the defect that way separate between word exists, solve because ignoring between word correlativity and cause marking in annotation results semantic inconsistent problem between word.The technical scheme that the present invention proposes is:
An image automatic annotation method based on correlativity between word, training set T comprises l image, described l image construction image collection P=[p 1p 2p l]; Each image labeling of described training set T has n mark word, and in training set T, all mark words form mark set of words W=[w 1w 2w s]; Each image of training set T has corresponding vision lemma, and in training set T, all vision lemmas form the set B=[b of visual word unit 1b 2b y], image to be marked is I, the method comprises:
A. according to formula
Figure BDA0000454408100000031
in calculation training set T, the semantic vector of each mark word w, is expressed as vector form w=< v by mark word w 1, v 2..., v m>, wherein, c ifor context relation word, total m context relation word, p (c i) be context relation word c ioverall distribution probability, p (c i/ w) represent context relation word c ithe ratio of the total degree that the co-occurrence number of times with mark word w in training set T and mark word w occur in training set T,
B. according to formula
Figure BDA0000454408100000033
calculate the semantic similarity between mark word, wherein || || for vectorial mould calculates;
C. according to formula calculate p (A), wherein A is mark phrase { w 1, w 2... w n, n is the number of mark phrase;
D. according to formula
Figure BDA0000454408100000035
design conditions Probability p (I/w i);
E. basis
Figure BDA0000454408100000036
calculate p (I/A);
F. by formula A=argmax ap (I/A) p (A) calculates the mark phrase A of image I to be marked.
In such scheme, step D further comprises:
P(w i) be mark word w ithe number of times occurring in training set T gathers with training the ratio that total degree appears in all mark words of T,
Figure BDA0000454408100000037
P(w i, b 1..., b n) computing method be:
P ( w i , b 1 , b 2 , . . . , b n ) = &Sigma; J &Element; T P ( J ) P ( w i | J ) &Pi; k = 1 n P ( b k | J ) , Wherein p (J) is illustrated in the probability of randomly drawing a width training image J in image collection P; p(w i/ J) represent to occur vocabulary w in training image J iposterior probability; And p (b k/ J) represent to occur vision lemma b in training image J kposterior probability; And
P ( w i | J ) = ( 1 - &alpha; J ) # ( w i , J ) | J | + &alpha; J # ( w i , T ) | T | - - - ( 1 )
P ( b k | J ) = ( 1 - &beta; J ) # ( b k , J ) | J | + &beta; J # ( b k , T ) | T | - - - ( 2 )
Wherein, α jwith β jfor smoothing parameter, it is an empirical value;
# (w i, J) represent mark word w iin training image J, whether occur, if so, # (w i, J)=1, otherwise # (w i, J)=0;
# (w i, T) represent mark word w iin training set T, whether occur, if so, # (w i, T)=1, otherwise # (w i, T)=0;
# (b k, J) represent vision lemma b kin training image J, whether occur, if so, # (b k, J)=1, otherwise # (b k, J)=0;
| J| represents to mark in training image J total number of word and vision lemma; | T| represents to mark in training set T total number of word and vision lemma.
In such scheme, described context relation word is the mark word in training set T.
In sum, the technical scheme that the present invention proposes converts the joint probability calculation process of the mark word in associating media correlation model and image the probability of mark entry part hypograph appearance and two-part the solving of prior probability of mark phrase to, greatly reduce high frequency candidate and mark word for the impact of probability statistics model, make non-high frequency candidate mark the larger effect of word performance, recall ratio and precision ratio that non-high frequency candidate marks word have been improved, the similar language model of semanteme is incorporated in the middle of associating media correlation model simultaneously, by the similar language model of semanteme, remove to estimate the prior probability of one group of mark word, so more likely produce one group of mark word that semantic dependency is stronger.Thereby improve the integral body mark effect of image.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the embodiment of the present invention.
Embodiment
Clearer for what the object, technical solutions and advantages of the present invention were expressed, below in conjunction with drawings and the specific embodiments, the present invention is further described in more detail.
Technical scheme of the present invention is:
A. according to formula
Figure BDA0000454408100000051
in calculation training set T, the semantic vector of each mark word w, is expressed as vector form w=< v by mark word w 1, v 2..., v m>, wherein, c ifor context relation word, total m context relation word, p (c i) be context relation word c ioverall distribution probability, p (c i/ w) represent context relation word c ithe ratio of the total degree that the co-occurrence number of times with mark word w in training set T and mark word w occur in training set T,
Figure BDA0000454408100000052
B. according to formula calculate the semantic similarity between mark word, wherein || || for vectorial mould calculates;
C. according to formula
Figure BDA0000454408100000054
calculate p (A), wherein A is mark phrase { w 1, w 2... w n, n is the number of mark phrase;
D. according to formula
Figure BDA0000454408100000055
design conditions Probability p (I/w i);
E. basis calculate p (I/A);
F. by formula A=arg max ap (I/A) p (A) calculates the mark phrase A of image I to be marked.
Image labeling problem can be defined as at present: a given training set T, this training set T comprises image collection P and mark set of words W, and every width image p iall completed mark word mark, the mark word of all images forms mark set of words W, does the one group of mark word A how choosing from described mark set of words W wherein mark a width new images I?
Image labeling method of the present invention adopts probability model, and its target is just to locate mark phrase A, its conditional probability p (A/I) maximum, that is:
A=arg?max Ap(A/I) (3)
Wherein A is a mark phrase { w 1, w 2... w n, one group of visual signature { b for image I 1, b 2, b mrepresent, by image I is carried out pre-service (such as image cut apart, the operation such as feature extraction, characteristic value normalization) and image block region sort out computing and obtain.P (A/I) can be rewritten as following form:
P ( A / I ) = p ( A , I ) P ( I ) - - - ( 4 )
Because the prior probability of piece image is considered to obey equally distributed conventionally, so p (I) can be regarded as a constant, and
p(A,I)=p(I/A)p(A) (5)
With formula (4), (5), formula (3) is simplified, is obtained:
A=arg?max Ap(I/A)p(A) (6)
By p (I/A) and two probability of p (A) are combined, solve maximal value and find best mark phrase A.P (I/A) can obtain from original image marking model, and p (A) can obtain from language model.By giving different weights to two probability, represent original image model and the mark influential effect ability of language model to final acquisition:
A = arg ma x A p ( I / A ) &lambda; 1 P ( A ) &lambda; 2 - - - ( 7 )
It is carried out to following formal transformation:
A=arg?max A1log?p(I/A)+λ 2log?p(A)) (8)
As long as calculate p (A) and p (I/A), just can obtain mark phrase A.Wherein, λ 1with λ 2be to determine in the machine learning of training plan image set and model process of establishing, carrying out in the automatic marking process of testing image is two constants.
The training set T that comprises l image of take below describes technical solution of the present invention as example, and image to be marked is I.L the image construction image collection P=[p of training set T 1p 2p l]; Each image labeling of training set T has n mark word, and in training set T, all mark words form mark set of words W=[w 1w 2w s]; Each image of training set T has corresponding vision lemma, and in training set T, all vision lemmas form the set B=[b of visual word unit 1b 2b y].
Fig. 1 is the process flow diagram of the present embodiment, as shown in Figure 1, comprises the following steps:
Step 101: image I to be marked is carried out to image pre-service and segmented areas classification computing.
In this step, image I to be marked is carried out to image pre-service (image is cut apart, feature extraction, characteristic value normalization etc.), then carry out image block region and sort out computing, utilize clustering algorithm to sort out each image block region, and combine presentation video vision content: I={i with visual word unit 1i 2i f.The preparation method of vision lemma is prior art, no longer describes in detail herein.
Step 102: calculate p (A) by the similar language model of semanteme.
In order to introduce correlation information between mark word in the similarity between mark word, the present invention has adopted each mark word w of semantic vector model representation: context relation set of words C=[c 1c 2c m], each element c irepresent a context relation word, total m context relation word, can choose all mark words of the mark set of words W in training set T as context relation word, i.e. C=W.Each the context relation word vector representation associated with it for w of mark word, i.e. w=< v 1, v 2..., v m>, wherein each semantic component v icalculating be defined as context relation word c iconditional probability and context relation word c with respect to mark word w ithe ratio of probability:
v i = p ( c i / w ) p ( c i ) - - - ( 9 )
P (c wherein i) expression context relation word c ioverall distribution probability, for being uniformly distributed.Conditional probability p (c i/ w) represent context relation word c ithe ratio of the total degree that co-occurrence number of times during all image labelings of image collection P with mark word w in training set T occurs during at all image labelings of image collection P with mark word w:
p ( c i / w ) = count ( c i , w ) count ( w ) - - - ( 10 )
P(c i/ w) represent the intensity distributions of vocabulary w and context relation word co-occurrence, then be exactly in order to prevent semantic vector w=< v divided by the whole probability of each context relation word 1, v 2..., v m> is dominated by the high context relation word of the frequency of occurrences, because high-frequency conjunctive word often also has very large conditional probability.As shown in table 1, wherein " sky ", " sun ", " clouds ", " town " represent a group context conjunctive word, and " tree ", " building ", " river " are one group of mark word, and the semantic vector of mark word represents as shown in table 1.
Table 1
? sky sun clouds town
tree 2.56 0.91 0.74 0.63
building 5.01 0.57 2.41 21.19
river 2.57 2.57 1.12 5.72
Then to calculate the semantic similarity between mark word.The calculating of similarity is as shown in Equation 11:
sim ( w i , w j ) = w i &CenterDot; w j | | w i | | &CenterDot; | | w j | | - - - ( 11 )
Wherein || || for vectorial mould calculates.
W iw jcalculating as shown in Equation 12:
w i &CenterDot; w j = &Sigma; k = 1 m v wi , k v w j , k = &Sigma; k = 1 m p ( c k / w i ) p ( c k ) &CenterDot; p ( c k / w j ) p ( c k ) - - - ( 12 )
C wherein krepresent context relation word.Between mark word, semantic similarity is as shown in table 2.Similarity span is 0 to 1, and the similarity between two mark words of the higher expression of numerical value is higher, and the probability that they appear in same piece image is just larger.
Table 2
? tree road sky wood
tree
1 0.1723 0.4311 0.2140
road 0.1723 1 0.1742 0.0021
sky 0.4311 0.1742 1 0.0383
wood 0.2140 0.0021 0.0383 1
Suppose in same mark, mark vocabulary is semantic relevant to context relation word, so one group of mark word A={w 1, w 2..., w nprobability p (A) can obtain with similarity that other mark between word by calculating each mark word:
p ( A ) &Proportional; 1 n - 1 &Sigma; w i &Element; A &Sigma; w j &Element; A , j &NotEqual; i sim ( w i , w j ) - - - ( 13 )
Formula 10,11,12 is updated in formula 13, can calculates the Probability p (A) of mark phrase:
p ( A ) &Proportional; 1 n - 1 &Sigma; w i &Element; A &Sigma; w j &Element; A , j &NotEqual; i &Sigma; k = 1 m count ( c k , w i ) count ( w i ) &CenterDot; p ( c k ) &CenterDot; count ( c k , w j ) count ( w j ) &CenterDot; p ( c k ) | | w i | | &CenterDot; | | w j | | - - - ( 14 )
Step 103: calculate p (I/A) by associating media correlation model.
In this step, first according to formula design conditions Probability p (I/w i).Wherein,
P(w i) computing method be:
With mark word w ithere is the ratio value representation vocabulary w of total degree in the number of times occurring and all mark words in training set T iprior probability p (w i):
p ( w i ) = | w i | &Sigma; w k &Element; T | w k | - - - ( 15 )
P(w i, b 1..., b n) computing method be:
P ( w i , b 1 , b 2 , . . . , b n ) = &Sigma; J &Element; T P ( J ) P ( w i | J ) &Pi; k = 1 n P ( b k | J ) - - - ( 16 )
P (J) is illustrated in the probability of randomly drawing a width training image J in image collection P, is generally assumed to be and is uniformly distributed; p(w i/ J) represent to occur vocabulary w in training image J iposterior probability; And p (b k/ J) represent to occur vision lemma b in training image J kposterior probability.The probable value of each is estimated as follows:
P ( w i | J ) = ( 1 - &alpha; J ) # ( w i , J ) | J | + &alpha; J # ( w i , T ) | T | - - - ( 17 )
P ( b k | J ) = ( 1 - &beta; J ) # ( b k , J ) | J | + &beta; J # ( b k , T ) | T | - - - ( 18 )
Wherein, α jwith β jfor smoothing parameter, it is an empirical value; # (w i, J) represent mark word w iin training image J, whether occur, if so, # (w i, J)=1, otherwise # (w i, J)=0; # (w i, T) represent mark word w iin training set T, whether occur, if so, # (w i, T)=1, otherwise # (w i, T)=0; # (b k, J) represent vision lemma b kin training image J, whether occur, if so, # (b k, J)=1, otherwise # (b k, J)=0; | J| represents to mark in training image J total number of word and vision lemma; | T| represents to mark in training set T total number of word and vision lemma.
P (I/A) can approximate evaluation go out for
Figure BDA0000454408100000101
Step 104: the phrase to be marked that calculates image I to be marked.
Below p (A) and p (I/A) have been solved respectively, according to A=arg max a1log p (I/A)+λ 2log p (A)) can be piece image I and calculate mark phrase A
The foregoing is only preferred embodiment of the present invention, not in order to limit the present invention, all any modifications of making within the spirit and principles in the present invention, be equal to replacement, improvement etc., within all should being included in the scope of protection of the invention.

Claims (3)

1. the image automatic annotation method based on correlativity between word, is characterized in that, training set T comprises l image, described l image construction image collection P=[p 1p 2p l]; Each image labeling of described training set T has n mark word, and in training set T, all mark words form mark set of words W=[w 1w 2w s]; Each image of training set T has corresponding vision lemma, and in training set T, all vision lemmas form the set B=[b of visual word unit 1b 2b y], image to be marked is I, the method comprises:
A. according to formula
Figure FDA0000454408090000011
in calculation training set T, the semantic vector of each mark word w, is expressed as vector form w=< v by mark word w 1, v 2..., v m>, wherein, c ifor context relation word, total m context relation word, p (c i) be context relation word c ioverall distribution probability, p (c i/ w) represent context relation word c ithe ratio of the total degree that the co-occurrence number of times with mark word w in training set T and mark word w occur in training set T,
Figure FDA0000454408090000012
B. according to formula
Figure FDA0000454408090000013
calculate the semantic similarity between mark word, wherein || || for vectorial mould calculates;
C. according to formula
Figure FDA0000454408090000014
calculate p (A), wherein A is mark phrase { w 1, w 2... w n, n is the number of mark phrase;
D. according to formula
Figure FDA0000454408090000015
design conditions Probability p (I/w i);
E. basis
Figure FDA0000454408090000016
calculate p (I/A);
F. by formula A=argmax ap (I/A) p (A) calculates the mark phrase A of image I to be marked.
2. method according to claim 1, is characterized in that, step D further comprises:
P(w i) be mark word w ithe number of times occurring in training set T gathers with training the ratio that total degree appears in all mark words of T,
Figure FDA0000454408090000021
P(w i, b 1..., b n) computing method be:
P ( w i , b 1 , b 2 , . . . , b n ) = &Sigma; J &Element; T P ( J ) P ( w i | J ) &Pi; k = 1 n P ( b k | J ) , Wherein p (J) is illustrated in the probability of randomly drawing a width training image J in image collection P; p(w i/ J) represent to occur vocabulary w in training image J iposterior probability; And p (b k/ J) represent to occur vision lemma b in training image J kposterior probability; And
P ( w i | J ) = ( 1 - &alpha; J ) # ( w i , J ) | J | + &alpha; J # ( w i , T ) | T | - - - ( 1 )
P ( b k | J ) = ( 1 - &beta; J ) # ( b k , J ) | J | + &beta; J # ( b k , T ) | T | - - - ( 2 )
Wherein, α jwith β jfor smoothing parameter, it is an empirical value;
# (w i, J) represent mark word w iin training image J, whether occur, if so, # (w i, J)=1, otherwise # (w i, J)=0;
# (w i, T) represent mark word w iin training set T, whether occur, if so, # (w i, T)=1, otherwise # (w i, T)=0;
# (b k, J) represent vision lemma b kin training image J, whether occur, if so, # (b k, J)=1, otherwise # (b k, J)=0;
| J| represents to mark in training image J total number of word and vision lemma; | T| represents to mark in training set T total number of word and vision lemma.
3. method according to claim 1, is characterized in that, described context relation word is the mark word in training set T.
CN201410008553.1A 2014-01-08 2014-01-08 Automatic image marking method based on word correlation Active CN103714178B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410008553.1A CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410008553.1A CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Publications (2)

Publication Number Publication Date
CN103714178A true CN103714178A (en) 2014-04-09
CN103714178B CN103714178B (en) 2017-01-25

Family

ID=50407153

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410008553.1A Active CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Country Status (1)

Country Link
CN (1) CN103714178B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104794183A (en) * 2015-04-10 2015-07-22 浙江大学 Picture labeling method based on multiple views and multiple labels
CN108268875A (en) * 2016-12-30 2018-07-10 广东精点数据科技股份有限公司 A kind of image meaning automatic marking method and device based on data smoothing
CN108604902A (en) * 2016-02-08 2018-09-28 皇家飞利浦有限公司 Determine the device and method of cluster
WO2020073952A1 (en) * 2018-10-10 2020-04-16 腾讯科技(深圳)有限公司 Method and apparatus for establishing image set for image recognition, network device, and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920820A (en) * 2006-09-14 2007-02-28 浙江大学 Image meaning automatic marking method based on marking significance sequence
US20090289942A1 (en) * 2008-05-20 2009-11-26 Timothee Bailloeul Image learning, automatic annotation, retrieval method, and device
CN101620615A (en) * 2009-08-04 2010-01-06 西南交通大学 Automatic image annotation and translation method based on decision tree learning
CN101685464A (en) * 2009-06-18 2010-03-31 浙江大学 Method for automatically labeling images based on community potential subject excavation
CN102298606A (en) * 2011-06-01 2011-12-28 清华大学 Random walking image automatic annotation method and device based on label graph model
CN102542067A (en) * 2012-01-06 2012-07-04 上海交通大学 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920820A (en) * 2006-09-14 2007-02-28 浙江大学 Image meaning automatic marking method based on marking significance sequence
US20090289942A1 (en) * 2008-05-20 2009-11-26 Timothee Bailloeul Image learning, automatic annotation, retrieval method, and device
CN101685464A (en) * 2009-06-18 2010-03-31 浙江大学 Method for automatically labeling images based on community potential subject excavation
CN101620615A (en) * 2009-08-04 2010-01-06 西南交通大学 Automatic image annotation and translation method based on decision tree learning
CN102298606A (en) * 2011-06-01 2011-12-28 清华大学 Random walking image automatic annotation method and device based on label graph model
CN102542067A (en) * 2012-01-06 2012-07-04 上海交通大学 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
BING-KUN BAO ET AL.: ""Hidden-Concept Driven Multilabel Image Annotation and Label Ranking"", 《IEEE TRANSACTIONS ON MULTIMEDIA》 *
MAHDIA BAKALEM ET AL.: ""Latent Semantic Analysis-based Image Auto Annotation"", 《IEEE CONF. ON MACHINE AND WEB INTELLIGENCE》 *
刘咏梅等: ""词间相关性的CMRM图像标注方法"", 《智能系统学报》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104794183A (en) * 2015-04-10 2015-07-22 浙江大学 Picture labeling method based on multiple views and multiple labels
CN108604902A (en) * 2016-02-08 2018-09-28 皇家飞利浦有限公司 Determine the device and method of cluster
CN108268875A (en) * 2016-12-30 2018-07-10 广东精点数据科技股份有限公司 A kind of image meaning automatic marking method and device based on data smoothing
CN108268875B (en) * 2016-12-30 2020-12-08 广东精点数据科技股份有限公司 Image semantic automatic labeling method and device based on data smoothing
WO2020073952A1 (en) * 2018-10-10 2020-04-16 腾讯科技(深圳)有限公司 Method and apparatus for establishing image set for image recognition, network device, and storage medium
US11853352B2 (en) 2018-10-10 2023-12-26 Tencent Technology (Shenzhen) Company Limited Method and apparatus for establishing image set for image recognition, network device, and storage medium

Also Published As

Publication number Publication date
CN103714178B (en) 2017-01-25

Similar Documents

Publication Publication Date Title
Gao et al. Database saliency for fast image retrieval
CN103617157B (en) Based on semantic Text similarity computing method
CN102902821B (en) The image high-level semantics mark of much-talked-about topic Network Based, search method and device
CN104391942B (en) Short essay eigen extended method based on semantic collection of illustrative plates
CN103678670B (en) Micro-blog hot word and hot topic mining system and method
CN104199857B (en) A kind of tax document hierarchy classification method based on multi-tag classification
CN103544697B (en) A kind of image partition method based on hypergraph analysis of spectrum
CN104881458B (en) A kind of mask method and device of Web page subject
CN105844424A (en) Product quality problem discovery and risk assessment method based on network comments
CN101950284A (en) Chinese word segmentation method and system
CN100573557C (en) A kind of SAR image partition method of short annealing based on MRF
CN103425757A (en) Cross-medial personage news searching method and system capable of fusing multi-mode information
Tran et al. Cluster-based similarity aggregation for ontology matching
CN104484380A (en) Personalized search method and personalized search device
CN103714178A (en) Automatic image marking method based on word correlation
CN102521368A (en) Similarity matrix iteration based cross-media semantic digesting and optimizing method
WO2013118435A1 (en) Semantic similarity level computation method, system and program
CN105653640A (en) Collaborative filtering recommendation method based on trust mechanism
CN102637199B (en) Image marking method based on semi-supervised subject modeling
CN110390022A (en) A kind of professional knowledge map construction method of automation
Zhao et al. Retrieving Social Flooding Images Based on Multimodal Information.
CN103927730A (en) Image noise reduction method based on Primal Sketch correction and matrix filling
CN107301426A (en) A kind of multi-tag clustering method of shoe sole print image
CN103064907A (en) System and method for topic meta search based on unsupervised entity relation extraction
Ma et al. PSVM: a preference-enhanced SVM model using preference data for classification

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant