CN102542067A - Automatic image semantic annotation method based on scale learning and correlated label dissemination - Google Patents

Automatic image semantic annotation method based on scale learning and correlated label dissemination Download PDF

Info

Publication number
CN102542067A
CN102542067A CN2012100023165A CN201210002316A CN102542067A CN 102542067 A CN102542067 A CN 102542067A CN 2012100023165 A CN2012100023165 A CN 2012100023165A CN 201210002316 A CN201210002316 A CN 201210002316A CN 102542067 A CN102542067 A CN 102542067A
Authority
CN
China
Prior art keywords
image
keyword
yardstick
study
label
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2012100023165A
Other languages
Chinese (zh)
Inventor
王斌
肖建力
刘允才
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaotong University
Original Assignee
Shanghai Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaotong University filed Critical Shanghai Jiaotong University
Priority to CN2012100023165A priority Critical patent/CN102542067A/en
Publication of CN102542067A publication Critical patent/CN102542067A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an automatic image semantic annotation method based on scale learning and correlated label dissemination, which comprises the following steps: firstly, the global and partial feature descriptor of each image is extracted after the image library is read; the feature descriptor is sent to a model based on a structured support vector machine for learning the distance scale between the images, actually the Mahalanobis distance; a model about the internal relation between key words is built; the learned Mahalanobis distance is embedded in a built label dissemination model so as to obtain the confidence degree score of each key word belonging to the image to be labeled; and a threshold value is set for the confidence degree score of each key word, and the key words of which the scores are higher than the threshold value are distributed to the images to be labeled, thereby completing labeling. The learning algorithm model based on the structured support vector machine can effectively solve the measuring problem of similarity between the images, the internal relation between the key words is fully excavated through the embedded-type correlated label dissemination model, and the accuracy of the image annotation and image retrieval is effectively improved.

Description

Automated graphics semanteme marking method based on yardstick study and the propagation of related label
Technical field
The present invention relates to image retrieval and automatic image annotation technical field, be specifically related to a kind of automated graphics semanteme marking method of propagating based on yardstick study and related label.
Background technology
In the last few years, along with the develop rapidly of Internet technology, multimedia image and memory device, the amount of images that the user touches was explosive increase.How fast and effeciently from mass data, searching the needed information of user is an important research project, and this also is the research contents of field of image search.
Initial image retrieval is based on text based retrieval mode (text-based image retrieval; Be called for short TBIR); It is at first described with a series of keyword the content of every width of cloth image with artificial mode; According to these keywords image is set up index then, image retrieval problem has just changed into the matching problem of text message like this.The mode of this artificial mark both wasted time and energy, and can not guarantee to give the objectivity of the keyword of image.The 1980s, content-based image labeling (content-based image retrieval is called for short CBIR) has appearred.This technology is at first extracted every width of cloth level image characteristic, according to these characteristics the image in the image data base is set up index then.The user only need provide a secondary example image, and the image with similarity characteristic just can be searched by system.The existence of semantic wide gap has greatly limited the application of CBIR in image retrieval.Automatic image annotation is in order to set up the contact between picture material and the keyword, automatically to give image to be marked with keyword assignment.Because the limitation of artificial mask method, increasing research steering utilizes the automatic image annotation technology to accomplish the task of image retrieval.
The purpose of image labeling is to let system automatically distribute the keyword that can describe its semantic content to image.The theory of existing image labeling technology general using machine learning goes out the relational model between feature space and the keyword space from marking the image focusing study, and utilizes this model to distribute the keyword of waiting to mark image as guidance.At present, the image labeling technology generally is divided three classes: based on the method for classification; Method based on probability model; Method based on internet hunt.Method based on classification is to regard image labeling as a kind of classification problem; A sorter all learnt out in each keyword from the image library that has marked; And utilize these different sorters to treat the mark image respectively and classify, the output result is that positive example shows that then the corresponding keyword of this sorter can be used for marking this width of cloth image.This method is one by one handled single keyword, does not have the inner link between the taking into account critical speech.Based on the method for probability model, be devoted to set up the characteristic of image and the joint ensemble between the keyword, only the characteristic of waiting to mark image need be provided, just can draw the probability that correspondent keyword occurs.The limitation of this method is that model extension property is relatively poor, the new keyword of coming in that adds, and all models need training again.Based on the method for internet hunt, accomplish image labeling through the extensive resource of excavating on the internet, improve the result of image labeling around literal such as utilizing around the internet epigraph.This method depends on the surrounding environment and the contextual information of internet epigraph, and the accuracy of annotation results is to be determined by the whether true of environmental information, and this has limited image labeling result's stability to a certain extent.
Summary of the invention
The objective of the invention is to deficiency to existing method; A kind of automated graphics semanteme marking method of propagating based on yardstick study and related label is proposed; Overcome the distance of only utilizing Euclidean distance etc. to define in advance in the prior art and weighed the defective of the Semantic Similarity between image; While has fully been excavated the correlativity between the keyword, makes that the result of mark is more accurate and effective.
For realizing above-mentioned purpose; The present invention at first extracts its content local and global characteristics fully describes image to every width of cloth image; And then with a kind of based on the distance scale-mahalanobis distance between the model learning image of structuring SVMs (Mahalanobis distance), then to the direct modeling of the inner link between the keyword of image, carry out label and propagate; And then the mahalanobis distance that will succeed in school is embedded into to be set up in the good label propagation model; Draw each keyword and belong to the probability of waiting to mark image, last setting threshold obtains the keyword of image.
The automated graphics semanteme marking method that the present invention is based on yardstick study and label propagation specifically comprises following step:
1. the color moment characteristic of image and Wavelet Texture are extracted the global characteristics descriptor of yardstick invariant features (scale-invariant feature transform is called for short SIFT) as image as its local feature description's in the extraction image library.
2. local feature and global characteristics are distinguished normalization, and be fused into a long proper vector as the characteristics of image descriptor.
3. proper vector and the label information with image library is integrated into the yardstick learning algorithm based on the structuring SVMs, draws the Ma Shi yardstick tolerance of similarity between image.
4. the relation between the image keyword is directly carried out modeling, fully excavate the correlativity between notion.
5. the scaling function of learning out is embedded in the keyword relationship modeling, finally draws the degree of confidence score of each keyword of waiting to mark image.
6. setting threshold is given image to be marked with the degree of confidence score greater than the keyword assignment of threshold value.
Particularly, according to an aspect of the present invention, a kind of automated graphics semanteme marking method of propagating based on yardstick study and related label is provided, comprises following concrete steps:
Step 1: the reading images storehouse, the feature description of extracting every width of cloth image is sub;
Step 2: draw similarity measurement between image through yardstick study;
Step 3: the inner link between the keyword of image is carried out relationship modeling;
Step 4: the similarity measurement of learning in the said step 2 is embedded in the process of keyword modeling;
Step 5: draw the degree of confidence score of the keyword of waiting to mark image, setting threshold is given image to be marked with the degree of confidence score greater than the keyword assignment of threshold value, accomplishes label and propagates.
Preferably, together, wherein, said global characteristics comprises that color moment and 62 dimensions are based on Gabor wave filter textural characteristics to feature description in the said step 1 with global characteristics and local Feature Fusion; Said local feature comprises the SIFT characteristic; With these characteristic normalization, be in turn connected into feature description of a long vector more respectively as image.
Preferably, the similarity measurement in the said step 2 realizes that through the yardstick study based on the structuring SVMs this implementation procedure may further comprise the steps:
Step 201: to the every width of cloth image x in the image library i, according to its key word information generation ordering r i, and design of graphics picture-sequence is right
Figure BDA0000128988700000031
As training set, utilize the modeling of structuring SVMs;
Step 202: the restrictive condition that provides optimization problem according to following principle: for x i, the discriminant score of its correct sample ordering? M, ⌒ (x i, r i)/with the score of other any incorrect orderings
Figure BDA0000128988700000032
Satisfy inequality
Figure BDA0000128988700000033
Wherein
Figure BDA0000128988700000034
Be loss function, ⌒ is the mapping that the input space arrives output region,
Figure BDA0000128988700000035
Be about x in the image library iOther any orderings; Under above-mentioned restrictive condition, find the solution feasible The M and the σ that set up, wherein, σ is a slack variable, μ is an adjustable parameter, is used for controlling lax degree;
Step 203: the Optimization Model of utilizing the said step 202 of cutting planes method iterative to propose; This method circulation makes up working set
Figure BDA0000128988700000042
wherein;
Figure BDA0000128988700000043
is the set of all restrictive condition; initial value is an empty set; Iteration begins: under current circulation; Algorithm is found the solution M and the σ under the work at present collection
Figure BDA0000128988700000045
, changes step 204 then over to;
Step 204: Find the biggest upheaval restrictions
Figure BDA0000128988700000046
and add it into the working set
Figure BDA0000128988700000047
Step 205: setting threshold; If the wavy amplitude of
Figure BDA0000128988700000049
then stops algorithm less than the threshold value that configures; Otherwise, return said step 203, the circulation of beginning next round.
Preferably, in the beginning of yardstick study, taken into full account the key word information of image set, for every width of cloth image generates the relevant ordering of image, and design of graphics picture-ordering is to gathering as training sample.
Preferably, keyword modeling in the said step 3 and label communication process are to accomplish through the mahalanobis distance yardstick that obtains in the said step 2 is embedded, and detailed process is following:
Step 301: the label communication process of taking into account critical speech mutual relationship is modeled as:
Figure BDA00001289887000000410
Wherein,
Figure BDA00001289887000000411
is that k keyword belongs to the degree of confidence score of waiting to mark image; δ (E) is an indicator function; And if only if, and incident E is that its value of true time is 1, otherwise is 0;
Step 302: for a keyword set Q; Provide its 0-1 vector representation form:
Figure BDA00001289887000000412
and if only if keyword i in set Q the time
Figure BDA00001289887000000413
value be 1, otherwise be 0;
Step 303: the δ indicator function is generalized to a recessed kernel function H; Obtaining a series of submodule function
Figure BDA00001289887000000414
is described below:
Figure BDA00001289887000000415
here,
Figure BDA00001289887000000416
is a kernel function;
Step 304: wait to mark image x for a pair u, the degree of confidence score of keyword draws through following greedy algorithm:
Figure BDA00001289887000000417
Here,
Figure BDA00001289887000000418
It is a keyword set;
Step 305: obtain an x uThe degree of confidence score vector of keyword
Step 306: threshold value θ is set 0, when The time, k keyword passed to x u
Preferably; After yardstick study finishes, only can preceding
Figure BDA0000128988700000052
width of cloth image of resulting image ordering be joined in keyword modeling and the label communication process.
Compare with existing method, superiority of the present invention is embodied in:
(1) not only extracts local feature but also extracted global characteristics, more fully portrayed the vision content of image, helped improving the accuracy of mark.
(2) similarity measurement between image is learnt to come out from image library, therefore is that sample is relevant, has overcome classic method and has used the distance (like Euclidean distance) that defines in advance to weigh the limitation of similarity between image.
(3) in the process of similarity yardstick tolerance study, added the key word information of image library, make that the yardstick of learning out can be better from the similarity between semantic level dimensioned plan picture.
(4) directly the inner link between the image keyword is carried out modeling, and be embedded into the scaling function of learning in this model, these two problems of the excavation that between study that has solved similarity measurement under the same framework simultaneously and keyword, concerns.
Description of drawings
Fig. 1 is the process flow diagram based on the automated graphics semantic tagger of yardstick study and the propagation of related label.
Fig. 2 by embodiment the example image of employing image library.
Fig. 3 by the present invention embodiment the annotation results on the employing image library.
Fig. 4 is the present invention and the mark comparing result of classical model on the part key speech.
Fig. 5 by the present invention embodiment the result for retrieval on the employing image library.
Embodiment
Below in conjunction with specific embodiment technical scheme of the present invention is done more detailed elaboration, the process flow diagram of operation is as shown in Figure 1.
Present embodiment is an example with the image labeling on the Core15k database.The example image of this database is as shown in Figure 2, comprises 374 keywords altogether, and every width of cloth image has been assigned with 1-5 keyword.
Below in conjunction with accompanying drawing embodiment of the present invention is done explanation more specifically, details are as follows:
Step 1, reading images storehouse, the feature description of extracting every width of cloth image is sub.
Image in the image library is extracted following three kinds of characteristics: (1) color moment characteristic; (2) 62 dimensions are based on Gabor wave filter textural characteristics; (3) SIFT characteristic.Preceding two kinds of characteristics are used for the global characteristics of picture engraving, a kind of local feature that comes picture engraving in back.With the normalization respectively of three kinds of characteristics, be fused into a long proper vector.
Step 2, based on the study of the yardstick of structuring SVMs.
Regard yardstick study the problem of category information retrieval as, come modeling and find the solution with the structuring SVMs.The image ordering that this yardstick of learning out based on the yardstick learning algorithm of structuring SVMs obtains is optimum.With X InputThe presentation video storehouse, Y OutputThe set of the image sequence of all images in the presentation video storehouse.For any two width of cloth image x in the image library iAnd x j,
Figure BDA0000128988700000061
Be mahalanobis distance (similarity size just) between the two, wherein M is a scaling function.For any sub-picture, r is its optimum image ordering,
Figure BDA0000128988700000062
Be Y OutputIn other orderings arbitrarily, the order
Figure BDA0000128988700000063
Be the loss function between the different images ordering.
Yardstick learning algorithm step based on the structuring SVMs is specific as follows:
Step 201: to the every width of cloth image x in the image library i,, generate its optimum image ordering r according to its key word information i, and design of graphics picture-sequence is right
Figure BDA0000128988700000064
As training set, N is the number of total image.Utilize the structuring SVMs to the modeling of this problem;
Step 202: the restrictive condition that provides optimization problem according to following principle: for x i, the discriminant score of its correct sample ordering? M, ⌒ (x i, r i)/with the score of other any incorrect orderings
Figure BDA0000128988700000065
Satisfy inequality
Figure BDA0000128988700000066
Wherein
Figure BDA0000128988700000067
Be loss function, ⌒ is the mapping that the input space arrives output region, Be about x in the image library iOther any orderings.Under above-mentioned restrictive condition, find the solution M and σ that feasible
Figure BDA0000128988700000069
sets up.Here, σ is a slack variable, and μ is an adjustable parameter, is used for controlling lax degree;
Step 203: the Optimization Model of utilizing the said step 202 of cutting planes method iterative to propose; This method circulation makes up working set
Figure BDA00001289887000000610
(set of all restrictive conditions;
Figure BDA00001289887000000611
initial value is an empty set); Iteration begins: under current circulation; Find the solution M and σ under the work at present collection
Figure BDA00001289887000000612
based on the structuring algorithm of support vector machine, change step 204 then over to;
Step 204: Find the biggest upheaval restrictions and add it into the working set
Figure BDA0000128988700000072
Figure BDA0000128988700000073
Step 205: setting threshold; If the wavy amplitude of
Figure BDA0000128988700000074
then stops algorithm less than the threshold value that configures; Otherwise, return said step 203.
The scaling function of learning out through above-mentioned algorithm makes based on the image ordering of distance optimum; For pair image to be marked; Only get preceding
Figure BDA0000128988700000075
sub-picture of image sequence and participate in the following processes modeling; This has removed the redundant information in the image library to a certain extent, has saved the computing cost.
Step 3, the inner link between the keyword is carried out relationship modeling.
The process of modeling is followed such principle: one by one pass to the degree of confidence score of image to the element in the keyword set, must not be higher than the degree of confidence score that passes to certain sub-set of set image.Under the guidance of this principle, concrete modeling process is following:
Under the tolerance of scaling function M, the process model building that keyword is transmitted one by one is:
Figure BDA0000128988700000076
Wherein,
Figure BDA0000128988700000077
Be the degree of confidence score of this process,
Figure BDA0000128988700000078
L is the key word number, Q iBe image x iKeyword set; δ (E) is an indicator function, and value was 1 when and if only if incident E was genuine.The process that a certain subclass Q of keyword set is passed to image can be modeled as:
Figure BDA0000128988700000079
Wherein, q u(Q) be the degree of confidence score of this process.Particularly,
Step 301: according to modeling principle, the label communication process of mutual relationship is modeled as between the taking into account critical speech:
The concrete derivation algorithm of this model is following:
Step 302: for a keyword set Q; Provide its 0-1 vector representation form: and if only if keyword i in set Q the time
Figure BDA00001289887000000712
value be 1, otherwise be 0;
Step 303: the δ indicator function is generalized to a recessed kernel function H; Obtaining a series of submodule function
Figure BDA0000128988700000081
is described below:
Figure BDA0000128988700000082
here,
Figure BDA0000128988700000083
is a kernel function;
Step 304: wait to mark image x for a pair u, the degree of confidence score of keyword draws through following greedy algorithm:
Figure BDA0000128988700000084
Here,
Figure BDA0000128988700000086
is a set of keywords;
Step 305: obtain one about x uThe degree of confidence score vector of keyword
Figure BDA0000128988700000087
Step 306: threshold value θ is set 0, when
Figure BDA0000128988700000088
The time, k keyword passed to x u
Through top five steps, obtained the highest keyword set of degree of confidence score of test pattern.Fig. 3 has provided the result of the present invention's automatic image annotation on the embodiment image library.Can find out that from this figure method of the present invention has been accomplished the task of image labeling effectively.Though some keyword that the present invention provides in an embodiment (providing with oblique font) is not included among the result of artificial mark, can find out that these keywords are meaningful and appropriate for the description of image, semantic content.Why the present invention can excavate abundanter keyword set, is because considered the mutual relationship between the keyword, and to direct modeling.
In order to verify validity of the present invention, on embodiment, calculated the size of accuracy rate (Precision) and precision ratio (Recall), and contrasted with several kinds of classical image labeling algorithms.These algorithms comprise Machine Translation Model (machine translation model, MT), stride the medium correlation model (cross-media relevance model, CMR), and mutually the symbiosis model (co-occurrence model, COM).Fig. 4 is the present invention and the comparing result of its excess-three kind model on the part key speech.Can find out that from this figure mark accuracy rate and the precision ratio of technology of the present invention on most keywords all is better than other three kinds of models.In addition, we have done quantitative test to the mark ability of different models, table 1 be the present invention on Embodiment C orel5K with the annotation results comparison diagram of above-mentioned different models.
Table 1
Wherein, N+ is the number with keyword of non-zero precision ratio.With respect to several classical models, the present invention has raising in various degree in accuracy rate and precision ratio.Fig. 5 is the instance that the present invention is applied in image retrieval.From top to bottom, term is: buildings, bear, flower, railway.For each term, return the higher image of the first five secondary score.

Claims (6)

1. an automated graphics semanteme marking method of propagating based on yardstick study and related label is characterized in that, comprises following concrete steps:
Step 1: the reading images storehouse, the feature description of extracting every width of cloth image is sub;
Step 2: draw similarity measurement between image through yardstick study;
Step 3: the inner link between the keyword of image is carried out relationship modeling;
Step 4: the similarity measurement of learning in the said step 2 is embedded in the process of keyword modeling;
Step 5: draw the degree of confidence score of the keyword of waiting to mark image, setting threshold is given image to be marked with the degree of confidence score greater than the keyword assignment of threshold value, accomplishes label and propagates.
2. the automated graphics semanteme marking method of propagating based on yardstick study and related label according to claim 1; It is characterized in that; Feature description in the said step 1 with global characteristics and local Feature Fusion together; Wherein, said global characteristics comprises that color moment and 62 dimensions are based on Gabor wave filter textural characteristics; Said local feature comprises the SIFT characteristic; With these characteristic normalization, be in turn connected into feature description of a long vector more respectively as image.
3. the automated graphics semanteme marking method of propagating based on yardstick study and related label according to claim 1 and 2; It is characterized in that; Similarity measurement in the said step 2 realizes that through the yardstick study based on the structuring SVMs this implementation procedure may further comprise the steps:
Step 201: to the every width of cloth image x in the image library i, according to its key word information generation ordering r i, and design of graphics picture-sequence is right
Figure FDA0000128988690000011
As training set, utilize the modeling of structuring SVMs;
Step 202: the restrictive condition that provides optimization problem according to following principle: for x i, the discriminant score of its correct sample ordering<m, Ψ (x i, r i)>Score with other any incorrect orderings Satisfy inequality
Figure FDA0000128988690000013
Wherein
Figure FDA0000128988690000014
Be loss function, Ψ is the mapping that the input space arrives output region,
Figure FDA0000128988690000015
Be about x in the image library iOther any orderings; Under above-mentioned restrictive condition, find the solution and make mintr{ (M TM)+and μ * σ } M and the σ that set up, wherein, σ is a slack variable, μ is an adjustable parameter, is used for controlling lax degree;
Step 203: the Optimization Model of utilizing the said step 202 of cutting planes method iterative to propose; The circulation of this method makes up working set Ω, and wherein, Ω is the set of all restrictive condition; The Ω initial value is an empty set; Iteration begins: under current circulation, algorithm is found the solution M and the σ under the work at present collection Ω, changes step 204 then over to;
Step 204: find out the maximum restrictive condition of wavy amplitude and add into working set Ω to it:
r ^ i = arg max r &Element; Y output &Delta; ( r i , r ) + < M , &Psi; ( x i , r ) > , i = 1,2 , &CenterDot; , N
Step 205: setting threshold; If the wavy amplitude of then stops algorithm less than the threshold value that configures; Otherwise, return said step 203, the circulation of beginning next round.
4. according to each described automated graphics semanteme marking method of propagating based on yardstick study and related label in the claim 1 to 3; It is characterized in that; Beginning in yardstick study; Taken into full account the key word information of image set, for every width of cloth image generates the relevant ordering of image, and design of graphics picture-ordering is to gathering as training sample.
5. according to each described automated graphics semanteme marking method of propagating based on yardstick study and related label in the claim 1 to 4; It is characterized in that; Keyword modeling in the said step 3 and label communication process are to accomplish through the mahalanobis distance yardstick that obtains in the said step 2 is embedded, and detailed process is following:
Step 301: the label communication process of taking into account critical speech mutual relationship is modeled as:
&Sigma; k = 1 l p k u &delta; ( k &Element; Q ) &le; &Sigma; i = 1 N &OverBar; d M ( x u , x i ) &delta; ( Q &cap; Q i )
Wherein,
Figure FDA0000128988690000025
is that k keyword belongs to the degree of confidence score of waiting to mark image; δ (E) is an indicator function; And if only if, and incident E is that its value of true time is 1, otherwise is 0;
Step 302: for a keyword set Q; Provide its 0-1 vector representation form:
Figure FDA0000128988690000026
and if only if keyword i in set Q the time
Figure FDA0000128988690000027
value be 1, otherwise be 0;
Step 303: the δ indicator function is generalized to a recessed kernel function H, obtains a series of submodule function T, be described below: Here, H (x)=1-2 -α xIt is a kernel function;
Step 304: wait to mark image x for a pair u, the degree of confidence score of keyword draws through following greedy algorithm: p k u = &Delta; T ( C k ) - T ( C k - 1 ) , k = 1,2 , &CenterDot; , l , Here, It is a keyword set;
Step 305: obtain an x uThe degree of confidence score vector of keyword
Figure FDA0000128988690000031
Step 306: threshold value θ is set 0, when
Figure FDA0000128988690000032
The time, k keyword passed to x u
6. according to each described automated graphics semanteme marking method of propagating based on yardstick study and related label in the claim 1 to 5; It is characterized in that; After yardstick study finishes, only can preceding
Figure FDA0000128988690000033
width of cloth image of resulting image ordering be joined in keyword modeling and the label communication process.
CN2012100023165A 2012-01-06 2012-01-06 Automatic image semantic annotation method based on scale learning and correlated label dissemination Pending CN102542067A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2012100023165A CN102542067A (en) 2012-01-06 2012-01-06 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2012100023165A CN102542067A (en) 2012-01-06 2012-01-06 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Publications (1)

Publication Number Publication Date
CN102542067A true CN102542067A (en) 2012-07-04

Family

ID=46348946

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2012100023165A Pending CN102542067A (en) 2012-01-06 2012-01-06 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Country Status (1)

Country Link
CN (1) CN102542067A (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103345481A (en) * 2013-06-19 2013-10-09 新疆大学 Method for labeling Uyghur image files
CN103440352A (en) * 2013-09-24 2013-12-11 中国科学院自动化研究所 Method and device for analyzing correlation among objects based on deep learning
CN103473275A (en) * 2013-08-23 2013-12-25 中山大学 Automatic image labeling method and automatic image labeling system by means of multi-feature fusion
CN103631889A (en) * 2013-11-15 2014-03-12 北京奇虎科技有限公司 Image recognizing method and device
CN103714178A (en) * 2014-01-08 2014-04-09 北京京东尚科信息技术有限公司 Automatic image marking method based on word correlation
CN104036021A (en) * 2014-06-26 2014-09-10 广西师范大学 Method for semantically annotating images on basis of hybrid generative and discriminative learning models
WO2015070678A1 (en) * 2013-11-15 2015-05-21 北京奇虎科技有限公司 Image recognition method, and method and device for mining main body information about image
CN105808752A (en) * 2016-03-10 2016-07-27 大连理工大学 CCA and 2PKNN based automatic image annotation method
CN106682060A (en) * 2015-11-11 2017-05-17 奥多比公司 Structured Knowledge Modeling, Extraction and Localization from Images
CN107316042A (en) * 2017-07-18 2017-11-03 盛世贞观(北京)科技有限公司 A kind of pictorial image search method and device
CN107423697A (en) * 2017-07-13 2017-12-01 西安电子科技大学 Activity recognition method based on non-linear fusion depth 3D convolution description
CN107423318A (en) * 2017-03-27 2017-12-01 北京珠穆朗玛移动通信有限公司 A kind of method and mobile terminal of picture mark
CN108182443A (en) * 2016-12-08 2018-06-19 广东精点数据科技股份有限公司 A kind of image automatic annotation method and device based on decision tree
CN108304847A (en) * 2017-11-30 2018-07-20 腾讯科技(深圳)有限公司 Image classification method and device, personalized recommendation method and device
CN108665055A (en) * 2017-03-28 2018-10-16 上海荆虹电子科技有限公司 A kind of figure says generation method and device
CN109784359A (en) * 2018-11-27 2019-05-21 北京邮电大学 Image generating method, device, equipment and readable storage medium storing program for executing
WO2020191706A1 (en) * 2019-03-28 2020-10-01 香港纺织及成衣研发中心有限公司 Active learning automatic image annotation system and method
CN111753861A (en) * 2019-03-28 2020-10-09 香港纺织及成衣研发中心有限公司 Automatic image annotation system and method for active learning
CN113139378A (en) * 2021-03-18 2021-07-20 杭州电子科技大学 Image description method based on visual embedding and condition normalization
CN113408633A (en) * 2021-06-29 2021-09-17 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for outputting information

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
BIN WANG等: "INTEGRATING DISTANCE METRIC LEARNING INTO LABEL PROPAGATION MODEL FOR MULTI-LABEL IMAGE ANNOTATION", 《2011 18TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING》 *
朱松豪等: "基于语义相似性的自动图片标注研究", 《和谐人机环境》 *

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103345481B (en) * 2013-06-19 2016-08-24 新疆大学 A kind of mask method of Uighur image file
CN103345481A (en) * 2013-06-19 2013-10-09 新疆大学 Method for labeling Uyghur image files
CN103473275A (en) * 2013-08-23 2013-12-25 中山大学 Automatic image labeling method and automatic image labeling system by means of multi-feature fusion
CN103440352A (en) * 2013-09-24 2013-12-11 中国科学院自动化研究所 Method and device for analyzing correlation among objects based on deep learning
CN103440352B (en) * 2013-09-24 2017-04-19 中国科学院自动化研究所 Method and device for analyzing correlation among objects based on deep learning
WO2015070678A1 (en) * 2013-11-15 2015-05-21 北京奇虎科技有限公司 Image recognition method, and method and device for mining main body information about image
CN103631889B (en) * 2013-11-15 2017-04-12 北京奇虎科技有限公司 Image recognizing method and device
CN103631889A (en) * 2013-11-15 2014-03-12 北京奇虎科技有限公司 Image recognizing method and device
CN103714178B (en) * 2014-01-08 2017-01-25 北京京东尚科信息技术有限公司 Automatic image marking method based on word correlation
CN103714178A (en) * 2014-01-08 2014-04-09 北京京东尚科信息技术有限公司 Automatic image marking method based on word correlation
CN104036021A (en) * 2014-06-26 2014-09-10 广西师范大学 Method for semantically annotating images on basis of hybrid generative and discriminative learning models
CN106682060A (en) * 2015-11-11 2017-05-17 奥多比公司 Structured Knowledge Modeling, Extraction and Localization from Images
CN105808752B (en) * 2016-03-10 2018-04-10 大连理工大学 A kind of automatic image marking method based on CCA and 2PKNN
CN105808752A (en) * 2016-03-10 2016-07-27 大连理工大学 CCA and 2PKNN based automatic image annotation method
CN108182443B (en) * 2016-12-08 2020-08-07 广东精点数据科技股份有限公司 Automatic image labeling method and device based on decision tree
CN108182443A (en) * 2016-12-08 2018-06-19 广东精点数据科技股份有限公司 A kind of image automatic annotation method and device based on decision tree
CN107423318A (en) * 2017-03-27 2017-12-01 北京珠穆朗玛移动通信有限公司 A kind of method and mobile terminal of picture mark
CN108665055A (en) * 2017-03-28 2018-10-16 上海荆虹电子科技有限公司 A kind of figure says generation method and device
CN108665055B (en) * 2017-03-28 2020-10-23 深圳荆虹科技有限公司 Method and device for generating graphic description
CN107423697A (en) * 2017-07-13 2017-12-01 西安电子科技大学 Activity recognition method based on non-linear fusion depth 3D convolution description
CN107423697B (en) * 2017-07-13 2020-09-08 西安电子科技大学 Behavior identification method based on nonlinear fusion depth 3D convolution descriptor
CN107316042A (en) * 2017-07-18 2017-11-03 盛世贞观(北京)科技有限公司 A kind of pictorial image search method and device
CN108304847B (en) * 2017-11-30 2021-09-28 腾讯科技(深圳)有限公司 Image classification method and device and personalized recommendation method and device
CN108304847A (en) * 2017-11-30 2018-07-20 腾讯科技(深圳)有限公司 Image classification method and device, personalized recommendation method and device
US11238315B2 (en) 2017-11-30 2022-02-01 Tencent Technology (Shenzhen) Company Limited Image classification method, personalized recommendation method, computer device and storage medium
CN109784359A (en) * 2018-11-27 2019-05-21 北京邮电大学 Image generating method, device, equipment and readable storage medium storing program for executing
WO2020191706A1 (en) * 2019-03-28 2020-10-01 香港纺织及成衣研发中心有限公司 Active learning automatic image annotation system and method
CN111753861A (en) * 2019-03-28 2020-10-09 香港纺织及成衣研发中心有限公司 Automatic image annotation system and method for active learning
CN111753861B (en) * 2019-03-28 2024-04-30 香港纺织及成衣研发中心有限公司 Automatic image labeling system and method for active learning
CN113139378A (en) * 2021-03-18 2021-07-20 杭州电子科技大学 Image description method based on visual embedding and condition normalization
CN113408633A (en) * 2021-06-29 2021-09-17 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for outputting information

Similar Documents

Publication Publication Date Title
CN102542067A (en) Automatic image semantic annotation method based on scale learning and correlated label dissemination
US7788265B2 (en) Taxonomy-based object classification
CN110502621A (en) Answering method, question and answer system, computer equipment and storage medium
CN111291210B (en) Image material library generation method, image material recommendation method and related devices
Wilkinson et al. Neural Ctrl-F: segmentation-free query-by-string word spotting in handwritten manuscript collections
Unar et al. Detected text‐based image retrieval approach for textual images
US20090144277A1 (en) Electronic table of contents entry classification and labeling scheme
CN108959559B (en) Question and answer pair generation method and device
CN104376406A (en) Enterprise innovation resource management and analysis system and method based on big data
CN104156433B (en) Image retrieval method based on semantic mapping space construction
CN103425687A (en) Retrieval method and system based on queries
US11861925B2 (en) Methods and systems of field detection in a document
CN103488724A (en) Book-oriented reading field knowledge map construction method
CN104517112A (en) Table recognition method and system
CN1936892A (en) Image content semanteme marking method
CN109272440B (en) Thumbnail generation method and system combining text and image content
CN116450834A (en) Archive knowledge graph construction method based on multi-mode semantic features
CN111737477A (en) Intellectual property big data-based intelligence investigation method, system and storage medium
Lu et al. Web multimedia object classification using cross-domain correlation knowledge
Pengcheng et al. Fast Chinese calligraphic character recognition with large-scale data
CN115661846A (en) Data processing method and device, electronic equipment and storage medium
Chang et al. Enhancing POI search on maps via online address extraction and associated information segmentation
US20230138491A1 (en) Continuous learning for document processing and analysis
US20230134218A1 (en) Continuous learning for document processing and analysis
CN115186240A (en) Social network user alignment method, device and medium based on relevance information

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20120704