CN102542067A - Automatic image semantic annotation method based on scale learning and correlated label dissemination - Google Patents
Automatic image semantic annotation method based on scale learning and correlated label dissemination Download PDFInfo
- Publication number
- CN102542067A CN102542067A CN2012100023165A CN201210002316A CN102542067A CN 102542067 A CN102542067 A CN 102542067A CN 2012100023165 A CN2012100023165 A CN 2012100023165A CN 201210002316 A CN201210002316 A CN 201210002316A CN 102542067 A CN102542067 A CN 102542067A
- Authority
- CN
- China
- Prior art keywords
- image
- keyword
- yardstick
- study
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Landscapes
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to an automatic image semantic annotation method based on scale learning and correlated label dissemination, which comprises the following steps: firstly, the global and partial feature descriptor of each image is extracted after the image library is read; the feature descriptor is sent to a model based on a structured support vector machine for learning the distance scale between the images, actually the Mahalanobis distance; a model about the internal relation between key words is built; the learned Mahalanobis distance is embedded in a built label dissemination model so as to obtain the confidence degree score of each key word belonging to the image to be labeled; and a threshold value is set for the confidence degree score of each key word, and the key words of which the scores are higher than the threshold value are distributed to the images to be labeled, thereby completing labeling. The learning algorithm model based on the structured support vector machine can effectively solve the measuring problem of similarity between the images, the internal relation between the key words is fully excavated through the embedded-type correlated label dissemination model, and the accuracy of the image annotation and image retrieval is effectively improved.
Description
Technical field
The present invention relates to image retrieval and automatic image annotation technical field, be specifically related to a kind of automated graphics semanteme marking method of propagating based on yardstick study and related label.
Background technology
In the last few years, along with the develop rapidly of Internet technology, multimedia image and memory device, the amount of images that the user touches was explosive increase.How fast and effeciently from mass data, searching the needed information of user is an important research project, and this also is the research contents of field of image search.
Initial image retrieval is based on text based retrieval mode (text-based image retrieval; Be called for short TBIR); It is at first described with a series of keyword the content of every width of cloth image with artificial mode; According to these keywords image is set up index then, image retrieval problem has just changed into the matching problem of text message like this.The mode of this artificial mark both wasted time and energy, and can not guarantee to give the objectivity of the keyword of image.The 1980s, content-based image labeling (content-based image retrieval is called for short CBIR) has appearred.This technology is at first extracted every width of cloth level image characteristic, according to these characteristics the image in the image data base is set up index then.The user only need provide a secondary example image, and the image with similarity characteristic just can be searched by system.The existence of semantic wide gap has greatly limited the application of CBIR in image retrieval.Automatic image annotation is in order to set up the contact between picture material and the keyword, automatically to give image to be marked with keyword assignment.Because the limitation of artificial mask method, increasing research steering utilizes the automatic image annotation technology to accomplish the task of image retrieval.
The purpose of image labeling is to let system automatically distribute the keyword that can describe its semantic content to image.The theory of existing image labeling technology general using machine learning goes out the relational model between feature space and the keyword space from marking the image focusing study, and utilizes this model to distribute the keyword of waiting to mark image as guidance.At present, the image labeling technology generally is divided three classes: based on the method for classification; Method based on probability model; Method based on internet hunt.Method based on classification is to regard image labeling as a kind of classification problem; A sorter all learnt out in each keyword from the image library that has marked; And utilize these different sorters to treat the mark image respectively and classify, the output result is that positive example shows that then the corresponding keyword of this sorter can be used for marking this width of cloth image.This method is one by one handled single keyword, does not have the inner link between the taking into account critical speech.Based on the method for probability model, be devoted to set up the characteristic of image and the joint ensemble between the keyword, only the characteristic of waiting to mark image need be provided, just can draw the probability that correspondent keyword occurs.The limitation of this method is that model extension property is relatively poor, the new keyword of coming in that adds, and all models need training again.Based on the method for internet hunt, accomplish image labeling through the extensive resource of excavating on the internet, improve the result of image labeling around literal such as utilizing around the internet epigraph.This method depends on the surrounding environment and the contextual information of internet epigraph, and the accuracy of annotation results is to be determined by the whether true of environmental information, and this has limited image labeling result's stability to a certain extent.
Summary of the invention
The objective of the invention is to deficiency to existing method; A kind of automated graphics semanteme marking method of propagating based on yardstick study and related label is proposed; Overcome the distance of only utilizing Euclidean distance etc. to define in advance in the prior art and weighed the defective of the Semantic Similarity between image; While has fully been excavated the correlativity between the keyword, makes that the result of mark is more accurate and effective.
For realizing above-mentioned purpose; The present invention at first extracts its content local and global characteristics fully describes image to every width of cloth image; And then with a kind of based on the distance scale-mahalanobis distance between the model learning image of structuring SVMs (Mahalanobis distance), then to the direct modeling of the inner link between the keyword of image, carry out label and propagate; And then the mahalanobis distance that will succeed in school is embedded into to be set up in the good label propagation model; Draw each keyword and belong to the probability of waiting to mark image, last setting threshold obtains the keyword of image.
The automated graphics semanteme marking method that the present invention is based on yardstick study and label propagation specifically comprises following step:
1. the color moment characteristic of image and Wavelet Texture are extracted the global characteristics descriptor of yardstick invariant features (scale-invariant feature transform is called for short SIFT) as image as its local feature description's in the extraction image library.
2. local feature and global characteristics are distinguished normalization, and be fused into a long proper vector as the characteristics of image descriptor.
3. proper vector and the label information with image library is integrated into the yardstick learning algorithm based on the structuring SVMs, draws the Ma Shi yardstick tolerance of similarity between image.
4. the relation between the image keyword is directly carried out modeling, fully excavate the correlativity between notion.
5. the scaling function of learning out is embedded in the keyword relationship modeling, finally draws the degree of confidence score of each keyword of waiting to mark image.
6. setting threshold is given image to be marked with the degree of confidence score greater than the keyword assignment of threshold value.
Particularly, according to an aspect of the present invention, a kind of automated graphics semanteme marking method of propagating based on yardstick study and related label is provided, comprises following concrete steps:
Step 1: the reading images storehouse, the feature description of extracting every width of cloth image is sub;
Step 2: draw similarity measurement between image through yardstick study;
Step 3: the inner link between the keyword of image is carried out relationship modeling;
Step 4: the similarity measurement of learning in the said step 2 is embedded in the process of keyword modeling;
Step 5: draw the degree of confidence score of the keyword of waiting to mark image, setting threshold is given image to be marked with the degree of confidence score greater than the keyword assignment of threshold value, accomplishes label and propagates.
Preferably, together, wherein, said global characteristics comprises that color moment and 62 dimensions are based on Gabor wave filter textural characteristics to feature description in the said step 1 with global characteristics and local Feature Fusion; Said local feature comprises the SIFT characteristic; With these characteristic normalization, be in turn connected into feature description of a long vector more respectively as image.
Preferably, the similarity measurement in the said step 2 realizes that through the yardstick study based on the structuring SVMs this implementation procedure may further comprise the steps:
Step 201: to the every width of cloth image x in the image library
i, according to its key word information generation ordering r
i, and design of graphics picture-sequence is right
As training set, utilize the modeling of structuring SVMs;
Step 202: the restrictive condition that provides optimization problem according to following principle: for x
i, the discriminant score of its correct sample ordering? M, ⌒ (x
i, r
i)/with the score of other any incorrect orderings
Satisfy inequality
Wherein
Be loss function, ⌒ is the mapping that the input space arrives output region,
Be about x in the image library
iOther any orderings; Under above-mentioned restrictive condition, find the solution feasible
The M and the σ that set up, wherein, σ is a slack variable, μ is an adjustable parameter, is used for controlling lax degree;
Step 203: the Optimization Model of utilizing the said step 202 of cutting planes method iterative to propose; This method circulation makes up working set
wherein;
is the set of all restrictive condition;
initial value is an empty set; Iteration begins: under current circulation; Algorithm is found the solution M and the σ under the work at present collection
, changes step 204 then over to;
Step 205: setting threshold; If the wavy amplitude of
then stops algorithm less than the threshold value that configures; Otherwise, return said step 203, the circulation of beginning next round.
Preferably, in the beginning of yardstick study, taken into full account the key word information of image set, for every width of cloth image generates the relevant ordering of image, and design of graphics picture-ordering is to gathering as training sample.
Preferably, keyword modeling in the said step 3 and label communication process are to accomplish through the mahalanobis distance yardstick that obtains in the said step 2 is embedded, and detailed process is following:
Step 301: the label communication process of taking into account critical speech mutual relationship is modeled as:
Wherein,
is that k keyword belongs to the degree of confidence score of waiting to mark image; δ (E) is an indicator function; And if only if, and incident E is that its value of true time is 1, otherwise is 0;
Step 302: for a keyword set Q; Provide its 0-1 vector representation form:
and if only if keyword i in set Q the time
value be 1, otherwise be 0;
Step 303: the δ indicator function is generalized to a recessed kernel function H; Obtaining a series of submodule function
is described below:
here,
is a kernel function;
Step 304: wait to mark image x for a pair
u, the degree of confidence score of keyword draws through following greedy algorithm:
Here,
It is a keyword set;
Step 305: obtain an x
uThe degree of confidence score vector of keyword
Step 306: threshold value θ is set
0, when
The time, k keyword passed to x
u
Preferably; After yardstick study finishes, only can preceding
width of cloth image of resulting image ordering be joined in keyword modeling and the label communication process.
Compare with existing method, superiority of the present invention is embodied in:
(1) not only extracts local feature but also extracted global characteristics, more fully portrayed the vision content of image, helped improving the accuracy of mark.
(2) similarity measurement between image is learnt to come out from image library, therefore is that sample is relevant, has overcome classic method and has used the distance (like Euclidean distance) that defines in advance to weigh the limitation of similarity between image.
(3) in the process of similarity yardstick tolerance study, added the key word information of image library, make that the yardstick of learning out can be better from the similarity between semantic level dimensioned plan picture.
(4) directly the inner link between the image keyword is carried out modeling, and be embedded into the scaling function of learning in this model, these two problems of the excavation that between study that has solved similarity measurement under the same framework simultaneously and keyword, concerns.
Description of drawings
Fig. 1 is the process flow diagram based on the automated graphics semantic tagger of yardstick study and the propagation of related label.
Fig. 2 by embodiment the example image of employing image library.
Fig. 3 by the present invention embodiment the annotation results on the employing image library.
Fig. 4 is the present invention and the mark comparing result of classical model on the part key speech.
Fig. 5 by the present invention embodiment the result for retrieval on the employing image library.
Embodiment
Below in conjunction with specific embodiment technical scheme of the present invention is done more detailed elaboration, the process flow diagram of operation is as shown in Figure 1.
Present embodiment is an example with the image labeling on the Core15k database.The example image of this database is as shown in Figure 2, comprises 374 keywords altogether, and every width of cloth image has been assigned with 1-5 keyword.
Below in conjunction with accompanying drawing embodiment of the present invention is done explanation more specifically, details are as follows:
Step 1, reading images storehouse, the feature description of extracting every width of cloth image is sub.
Image in the image library is extracted following three kinds of characteristics: (1) color moment characteristic; (2) 62 dimensions are based on Gabor wave filter textural characteristics; (3) SIFT characteristic.Preceding two kinds of characteristics are used for the global characteristics of picture engraving, a kind of local feature that comes picture engraving in back.With the normalization respectively of three kinds of characteristics, be fused into a long proper vector.
Step 2, based on the study of the yardstick of structuring SVMs.
Regard yardstick study the problem of category information retrieval as, come modeling and find the solution with the structuring SVMs.The image ordering that this yardstick of learning out based on the yardstick learning algorithm of structuring SVMs obtains is optimum.With X
InputThe presentation video storehouse, Y
OutputThe set of the image sequence of all images in the presentation video storehouse.For any two width of cloth image x in the image library
iAnd x
j,
Be mahalanobis distance (similarity size just) between the two, wherein M is a scaling function.For any sub-picture, r is its optimum image ordering,
Be Y
OutputIn other orderings arbitrarily, the order
Be the loss function between the different images ordering.
Yardstick learning algorithm step based on the structuring SVMs is specific as follows:
Step 201: to the every width of cloth image x in the image library
i,, generate its optimum image ordering r according to its key word information
i, and design of graphics picture-sequence is right
As training set, N is the number of total image.Utilize the structuring SVMs to the modeling of this problem;
Step 202: the restrictive condition that provides optimization problem according to following principle: for x
i, the discriminant score of its correct sample ordering? M, ⌒ (x
i, r
i)/with the score of other any incorrect orderings
Satisfy inequality
Wherein
Be loss function, ⌒ is the mapping that the input space arrives output region,
Be about x in the image library
iOther any orderings.Under above-mentioned restrictive condition, find the solution M and σ that feasible
sets up.Here, σ is a slack variable, and μ is an adjustable parameter, is used for controlling lax degree;
Step 203: the Optimization Model of utilizing the said step 202 of cutting planes method iterative to propose; This method circulation makes up working set
(set of all restrictive conditions;
initial value is an empty set); Iteration begins: under current circulation; Find the solution M and σ under the work at present collection
based on the structuring algorithm of support vector machine, change step 204 then over to;
Step 205: setting threshold; If the wavy amplitude of
then stops algorithm less than the threshold value that configures; Otherwise, return said step 203.
The scaling function of learning out through above-mentioned algorithm makes based on the image ordering of distance optimum; For pair image to be marked; Only get preceding
sub-picture of image sequence and participate in the following processes modeling; This has removed the redundant information in the image library to a certain extent, has saved the computing cost.
Step 3, the inner link between the keyword is carried out relationship modeling.
The process of modeling is followed such principle: one by one pass to the degree of confidence score of image to the element in the keyword set, must not be higher than the degree of confidence score that passes to certain sub-set of set image.Under the guidance of this principle, concrete modeling process is following:
Under the tolerance of scaling function M, the process model building that keyword is transmitted one by one is:
Wherein,
Be the degree of confidence score of this process,
L is the key word number, Q
iBe image x
iKeyword set; δ (E) is an indicator function, and value was 1 when and if only if incident E was genuine.The process that a certain subclass Q of keyword set is passed to image can be modeled as:
Wherein, q
u(Q) be the degree of confidence score of this process.Particularly,
Step 301: according to modeling principle, the label communication process of mutual relationship is modeled as between the taking into account critical speech:
The concrete derivation algorithm of this model is following:
Step 302: for a keyword set Q; Provide its 0-1 vector representation form:
and if only if keyword i in set Q the time
value be 1, otherwise be 0;
Step 303: the δ indicator function is generalized to a recessed kernel function H; Obtaining a series of submodule function
is described below:
here,
is a kernel function;
Step 304: wait to mark image x for a pair
u, the degree of confidence score of keyword draws through following greedy algorithm:
Through top five steps, obtained the highest keyword set of degree of confidence score of test pattern.Fig. 3 has provided the result of the present invention's automatic image annotation on the embodiment image library.Can find out that from this figure method of the present invention has been accomplished the task of image labeling effectively.Though some keyword that the present invention provides in an embodiment (providing with oblique font) is not included among the result of artificial mark, can find out that these keywords are meaningful and appropriate for the description of image, semantic content.Why the present invention can excavate abundanter keyword set, is because considered the mutual relationship between the keyword, and to direct modeling.
In order to verify validity of the present invention, on embodiment, calculated the size of accuracy rate (Precision) and precision ratio (Recall), and contrasted with several kinds of classical image labeling algorithms.These algorithms comprise Machine Translation Model (machine translation model, MT), stride the medium correlation model (cross-media relevance model, CMR), and mutually the symbiosis model (co-occurrence model, COM).Fig. 4 is the present invention and the comparing result of its excess-three kind model on the part key speech.Can find out that from this figure mark accuracy rate and the precision ratio of technology of the present invention on most keywords all is better than other three kinds of models.In addition, we have done quantitative test to the mark ability of different models, table 1 be the present invention on Embodiment C orel5K with the annotation results comparison diagram of above-mentioned different models.
Table 1
Wherein, N+ is the number with keyword of non-zero precision ratio.With respect to several classical models, the present invention has raising in various degree in accuracy rate and precision ratio.Fig. 5 is the instance that the present invention is applied in image retrieval.From top to bottom, term is: buildings, bear, flower, railway.For each term, return the higher image of the first five secondary score.
Claims (6)
1. an automated graphics semanteme marking method of propagating based on yardstick study and related label is characterized in that, comprises following concrete steps:
Step 1: the reading images storehouse, the feature description of extracting every width of cloth image is sub;
Step 2: draw similarity measurement between image through yardstick study;
Step 3: the inner link between the keyword of image is carried out relationship modeling;
Step 4: the similarity measurement of learning in the said step 2 is embedded in the process of keyword modeling;
Step 5: draw the degree of confidence score of the keyword of waiting to mark image, setting threshold is given image to be marked with the degree of confidence score greater than the keyword assignment of threshold value, accomplishes label and propagates.
2. the automated graphics semanteme marking method of propagating based on yardstick study and related label according to claim 1; It is characterized in that; Feature description in the said step 1 with global characteristics and local Feature Fusion together; Wherein, said global characteristics comprises that color moment and 62 dimensions are based on Gabor wave filter textural characteristics; Said local feature comprises the SIFT characteristic; With these characteristic normalization, be in turn connected into feature description of a long vector more respectively as image.
3. the automated graphics semanteme marking method of propagating based on yardstick study and related label according to claim 1 and 2; It is characterized in that; Similarity measurement in the said step 2 realizes that through the yardstick study based on the structuring SVMs this implementation procedure may further comprise the steps:
Step 201: to the every width of cloth image x in the image library
i, according to its key word information generation ordering r
i, and design of graphics picture-sequence is right
As training set, utilize the modeling of structuring SVMs;
Step 202: the restrictive condition that provides optimization problem according to following principle: for x
i, the discriminant score of its correct sample ordering<m, Ψ (x
i, r
i)>Score with other any incorrect orderings
Satisfy inequality
Wherein
Be loss function, Ψ is the mapping that the input space arrives output region,
Be about x in the image library
iOther any orderings; Under above-mentioned restrictive condition, find the solution and make mintr{ (M
TM)+and μ * σ } M and the σ that set up, wherein, σ is a slack variable, μ is an adjustable parameter, is used for controlling lax degree;
Step 203: the Optimization Model of utilizing the said step 202 of cutting planes method iterative to propose; The circulation of this method makes up working set Ω, and wherein, Ω is the set of all restrictive condition; The Ω initial value is an empty set; Iteration begins: under current circulation, algorithm is found the solution M and the σ under the work at present collection Ω, changes step 204 then over to;
Step 204: find out the maximum restrictive condition
of wavy amplitude and add into working set Ω to it:
Step 205: setting threshold; If the wavy amplitude of
then stops algorithm less than the threshold value that configures; Otherwise, return said step 203, the circulation of beginning next round.
4. according to each described automated graphics semanteme marking method of propagating based on yardstick study and related label in the claim 1 to 3; It is characterized in that; Beginning in yardstick study; Taken into full account the key word information of image set, for every width of cloth image generates the relevant ordering of image, and design of graphics picture-ordering is to gathering as training sample.
5. according to each described automated graphics semanteme marking method of propagating based on yardstick study and related label in the claim 1 to 4; It is characterized in that; Keyword modeling in the said step 3 and label communication process are to accomplish through the mahalanobis distance yardstick that obtains in the said step 2 is embedded, and detailed process is following:
Step 301: the label communication process of taking into account critical speech mutual relationship is modeled as:
Wherein,
is that k keyword belongs to the degree of confidence score of waiting to mark image; δ (E) is an indicator function; And if only if, and incident E is that its value of true time is 1, otherwise is 0;
Step 302: for a keyword set Q; Provide its 0-1 vector representation form:
and if only if keyword i in set Q the time
value be 1, otherwise be 0;
Step 303: the δ indicator function is generalized to a recessed kernel function H, obtains a series of submodule function T, be described below:
Here, H (x)=1-2
-α xIt is a kernel function;
Step 304: wait to mark image x for a pair
u, the degree of confidence score of keyword draws through following greedy algorithm:
Here,
It is a keyword set;
6. according to each described automated graphics semanteme marking method of propagating based on yardstick study and related label in the claim 1 to 5; It is characterized in that; After yardstick study finishes, only can preceding
width of cloth image of resulting image ordering be joined in keyword modeling and the label communication process.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2012100023165A CN102542067A (en) | 2012-01-06 | 2012-01-06 | Automatic image semantic annotation method based on scale learning and correlated label dissemination |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2012100023165A CN102542067A (en) | 2012-01-06 | 2012-01-06 | Automatic image semantic annotation method based on scale learning and correlated label dissemination |
Publications (1)
Publication Number | Publication Date |
---|---|
CN102542067A true CN102542067A (en) | 2012-07-04 |
Family
ID=46348946
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2012100023165A Pending CN102542067A (en) | 2012-01-06 | 2012-01-06 | Automatic image semantic annotation method based on scale learning and correlated label dissemination |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102542067A (en) |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103345481A (en) * | 2013-06-19 | 2013-10-09 | 新疆大学 | Method for labeling Uyghur image files |
CN103440352A (en) * | 2013-09-24 | 2013-12-11 | 中国科学院自动化研究所 | Method and device for analyzing correlation among objects based on deep learning |
CN103473275A (en) * | 2013-08-23 | 2013-12-25 | 中山大学 | Automatic image labeling method and automatic image labeling system by means of multi-feature fusion |
CN103631889A (en) * | 2013-11-15 | 2014-03-12 | 北京奇虎科技有限公司 | Image recognizing method and device |
CN103714178A (en) * | 2014-01-08 | 2014-04-09 | 北京京东尚科信息技术有限公司 | Automatic image marking method based on word correlation |
CN104036021A (en) * | 2014-06-26 | 2014-09-10 | 广西师范大学 | Method for semantically annotating images on basis of hybrid generative and discriminative learning models |
WO2015070678A1 (en) * | 2013-11-15 | 2015-05-21 | 北京奇虎科技有限公司 | Image recognition method, and method and device for mining main body information about image |
CN105808752A (en) * | 2016-03-10 | 2016-07-27 | 大连理工大学 | CCA and 2PKNN based automatic image annotation method |
CN106682060A (en) * | 2015-11-11 | 2017-05-17 | 奥多比公司 | Structured Knowledge Modeling, Extraction and Localization from Images |
CN107316042A (en) * | 2017-07-18 | 2017-11-03 | 盛世贞观(北京)科技有限公司 | A kind of pictorial image search method and device |
CN107423697A (en) * | 2017-07-13 | 2017-12-01 | 西安电子科技大学 | Activity recognition method based on non-linear fusion depth 3D convolution description |
CN107423318A (en) * | 2017-03-27 | 2017-12-01 | 北京珠穆朗玛移动通信有限公司 | A kind of method and mobile terminal of picture mark |
CN108182443A (en) * | 2016-12-08 | 2018-06-19 | 广东精点数据科技股份有限公司 | A kind of image automatic annotation method and device based on decision tree |
CN108304847A (en) * | 2017-11-30 | 2018-07-20 | 腾讯科技(深圳)有限公司 | Image classification method and device, personalized recommendation method and device |
CN108665055A (en) * | 2017-03-28 | 2018-10-16 | 上海荆虹电子科技有限公司 | A kind of figure says generation method and device |
CN109784359A (en) * | 2018-11-27 | 2019-05-21 | 北京邮电大学 | Image generating method, device, equipment and readable storage medium storing program for executing |
WO2020191706A1 (en) * | 2019-03-28 | 2020-10-01 | 香港纺织及成衣研发中心有限公司 | Active learning automatic image annotation system and method |
CN111753861A (en) * | 2019-03-28 | 2020-10-09 | 香港纺织及成衣研发中心有限公司 | Automatic image annotation system and method for active learning |
CN113139378A (en) * | 2021-03-18 | 2021-07-20 | 杭州电子科技大学 | Image description method based on visual embedding and condition normalization |
CN113408633A (en) * | 2021-06-29 | 2021-09-17 | 北京百度网讯科技有限公司 | Method, apparatus, device and storage medium for outputting information |
-
2012
- 2012-01-06 CN CN2012100023165A patent/CN102542067A/en active Pending
Non-Patent Citations (2)
Title |
---|
BIN WANG等: "INTEGRATING DISTANCE METRIC LEARNING INTO LABEL PROPAGATION MODEL FOR MULTI-LABEL IMAGE ANNOTATION", 《2011 18TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING》 * |
朱松豪等: "基于语义相似性的自动图片标注研究", 《和谐人机环境》 * |
Cited By (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103345481B (en) * | 2013-06-19 | 2016-08-24 | 新疆大学 | A kind of mask method of Uighur image file |
CN103345481A (en) * | 2013-06-19 | 2013-10-09 | 新疆大学 | Method for labeling Uyghur image files |
CN103473275A (en) * | 2013-08-23 | 2013-12-25 | 中山大学 | Automatic image labeling method and automatic image labeling system by means of multi-feature fusion |
CN103440352A (en) * | 2013-09-24 | 2013-12-11 | 中国科学院自动化研究所 | Method and device for analyzing correlation among objects based on deep learning |
CN103440352B (en) * | 2013-09-24 | 2017-04-19 | 中国科学院自动化研究所 | Method and device for analyzing correlation among objects based on deep learning |
WO2015070678A1 (en) * | 2013-11-15 | 2015-05-21 | 北京奇虎科技有限公司 | Image recognition method, and method and device for mining main body information about image |
CN103631889B (en) * | 2013-11-15 | 2017-04-12 | 北京奇虎科技有限公司 | Image recognizing method and device |
CN103631889A (en) * | 2013-11-15 | 2014-03-12 | 北京奇虎科技有限公司 | Image recognizing method and device |
CN103714178B (en) * | 2014-01-08 | 2017-01-25 | 北京京东尚科信息技术有限公司 | Automatic image marking method based on word correlation |
CN103714178A (en) * | 2014-01-08 | 2014-04-09 | 北京京东尚科信息技术有限公司 | Automatic image marking method based on word correlation |
CN104036021A (en) * | 2014-06-26 | 2014-09-10 | 广西师范大学 | Method for semantically annotating images on basis of hybrid generative and discriminative learning models |
CN106682060A (en) * | 2015-11-11 | 2017-05-17 | 奥多比公司 | Structured Knowledge Modeling, Extraction and Localization from Images |
CN105808752B (en) * | 2016-03-10 | 2018-04-10 | 大连理工大学 | A kind of automatic image marking method based on CCA and 2PKNN |
CN105808752A (en) * | 2016-03-10 | 2016-07-27 | 大连理工大学 | CCA and 2PKNN based automatic image annotation method |
CN108182443B (en) * | 2016-12-08 | 2020-08-07 | 广东精点数据科技股份有限公司 | Automatic image labeling method and device based on decision tree |
CN108182443A (en) * | 2016-12-08 | 2018-06-19 | 广东精点数据科技股份有限公司 | A kind of image automatic annotation method and device based on decision tree |
CN107423318A (en) * | 2017-03-27 | 2017-12-01 | 北京珠穆朗玛移动通信有限公司 | A kind of method and mobile terminal of picture mark |
CN108665055A (en) * | 2017-03-28 | 2018-10-16 | 上海荆虹电子科技有限公司 | A kind of figure says generation method and device |
CN108665055B (en) * | 2017-03-28 | 2020-10-23 | 深圳荆虹科技有限公司 | Method and device for generating graphic description |
CN107423697A (en) * | 2017-07-13 | 2017-12-01 | 西安电子科技大学 | Activity recognition method based on non-linear fusion depth 3D convolution description |
CN107423697B (en) * | 2017-07-13 | 2020-09-08 | 西安电子科技大学 | Behavior identification method based on nonlinear fusion depth 3D convolution descriptor |
CN107316042A (en) * | 2017-07-18 | 2017-11-03 | 盛世贞观(北京)科技有限公司 | A kind of pictorial image search method and device |
CN108304847B (en) * | 2017-11-30 | 2021-09-28 | 腾讯科技(深圳)有限公司 | Image classification method and device and personalized recommendation method and device |
CN108304847A (en) * | 2017-11-30 | 2018-07-20 | 腾讯科技(深圳)有限公司 | Image classification method and device, personalized recommendation method and device |
US11238315B2 (en) | 2017-11-30 | 2022-02-01 | Tencent Technology (Shenzhen) Company Limited | Image classification method, personalized recommendation method, computer device and storage medium |
CN109784359A (en) * | 2018-11-27 | 2019-05-21 | 北京邮电大学 | Image generating method, device, equipment and readable storage medium storing program for executing |
WO2020191706A1 (en) * | 2019-03-28 | 2020-10-01 | 香港纺织及成衣研发中心有限公司 | Active learning automatic image annotation system and method |
CN111753861A (en) * | 2019-03-28 | 2020-10-09 | 香港纺织及成衣研发中心有限公司 | Automatic image annotation system and method for active learning |
CN111753861B (en) * | 2019-03-28 | 2024-04-30 | 香港纺织及成衣研发中心有限公司 | Automatic image labeling system and method for active learning |
CN113139378A (en) * | 2021-03-18 | 2021-07-20 | 杭州电子科技大学 | Image description method based on visual embedding and condition normalization |
CN113408633A (en) * | 2021-06-29 | 2021-09-17 | 北京百度网讯科技有限公司 | Method, apparatus, device and storage medium for outputting information |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102542067A (en) | Automatic image semantic annotation method based on scale learning and correlated label dissemination | |
US7788265B2 (en) | Taxonomy-based object classification | |
CN110502621A (en) | Answering method, question and answer system, computer equipment and storage medium | |
CN111291210B (en) | Image material library generation method, image material recommendation method and related devices | |
Wilkinson et al. | Neural Ctrl-F: segmentation-free query-by-string word spotting in handwritten manuscript collections | |
Unar et al. | Detected text‐based image retrieval approach for textual images | |
US20090144277A1 (en) | Electronic table of contents entry classification and labeling scheme | |
CN108959559B (en) | Question and answer pair generation method and device | |
CN104376406A (en) | Enterprise innovation resource management and analysis system and method based on big data | |
CN104156433B (en) | Image retrieval method based on semantic mapping space construction | |
CN103425687A (en) | Retrieval method and system based on queries | |
US11861925B2 (en) | Methods and systems of field detection in a document | |
CN103488724A (en) | Book-oriented reading field knowledge map construction method | |
CN104517112A (en) | Table recognition method and system | |
CN1936892A (en) | Image content semanteme marking method | |
CN109272440B (en) | Thumbnail generation method and system combining text and image content | |
CN116450834A (en) | Archive knowledge graph construction method based on multi-mode semantic features | |
CN111737477A (en) | Intellectual property big data-based intelligence investigation method, system and storage medium | |
Lu et al. | Web multimedia object classification using cross-domain correlation knowledge | |
Pengcheng et al. | Fast Chinese calligraphic character recognition with large-scale data | |
CN115661846A (en) | Data processing method and device, electronic equipment and storage medium | |
Chang et al. | Enhancing POI search on maps via online address extraction and associated information segmentation | |
US20230138491A1 (en) | Continuous learning for document processing and analysis | |
US20230134218A1 (en) | Continuous learning for document processing and analysis | |
CN115186240A (en) | Social network user alignment method, device and medium based on relevance information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20120704 |