CN103714178A - Automatic image marking method based on word correlation - Google Patents
Automatic image marking method based on word correlation Download PDFInfo
- Publication number
- CN103714178A CN103714178A CN201410008553.1A CN201410008553A CN103714178A CN 103714178 A CN103714178 A CN 103714178A CN 201410008553 A CN201410008553 A CN 201410008553A CN 103714178 A CN103714178 A CN 103714178A
- Authority
- CN
- China
- Prior art keywords
- word
- mark
- image
- training set
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
- G06F18/2155—Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling
Abstract
The invention discloses an automatic image marking method based on work correlation. A training set T comprises l images, n marking words are marked on each image of the training set T, the training set T is provided with a corresponding vision lemma, and the image to be marked is I. The method includes the steps that a semantic vector of each marking word w is calculated according to a formula, the marking word w is represented by the vector form w=<v1, v2,......, vm>, ci is an associated word in a context, and the associated words in the context total m; semantic similarity of the marking words is calculated according to a formula, and vector module calculated is achieved as is shown in the specification; p(A) is calculated according to the formula, wherein A is a marking word group in w1, w2,......wn, and n is the number of the marking word groups; the conditional probability p(I/wi) is calculated according to a formula; the marking word group A of the image I to be marked is calculated according to the formula A=arg maxAp(I/A) p(A).
Description
Technical field
The present invention relates to image processing field, particularly a kind of image automatic annotation method based on correlativity between word.
Background technology
Along with the high speed development of multimedia and Internet technology, daily life, work more and more stronger to the dependence of the multimedia messagess such as image.Image retrieval based on semantic can not only accurately be expressed user's retrieval intention, is also convenient to user's use, so this retrieval mode not only becomes a kind of important form of image retrieval, and becomes the hot technology that researchist chases.
And automatic image annotation technology is an important and challenging job in Semantic Image Retrieval, the appearance of automatic image annotation technology is the semantic information comprising in automatic acquisition image vision content, it attempts between image bottom visual signature and high-level semantic, to build a bridge block, thereby in semantic level, semantic retrieval is made to support.Therefore, the automatic marking algorithm research based on image, semantic has become very active research branch and a gordian technique in field of image search, has good application prospect and researching value.
Automatic image annotation is exactly to allow computing machine automatically to the image without mark, add the semantic key words that can reflect picture material.It utilizes the image collection or other the obtainable information that have marked automatically to practise the relational model in semantic concept space and visual signature space, and with this model, marks the image of unknown semantics.By setting up a kind of mapping relations between the high-layer semantic information at image and low-level image feature, solve to a certain extent semantic gap problem.
The image automatic annotation method of associating media correlation model is current a kind of image labeling algorithm based on being most widely used in generation model image labeling method, has obtained scholar's broad research.The basic thought of this marking model is to utilize the method for probability statistics to set up the probability relativity in Image Visual Feature space and semantic concept space, the joint probability distribution existing between the two by statistical learning, find out that one group of semantic tagger word makes it and picture material between joint probability maximum, and the final mark using this group mark word as testing image.
But associating media correlation model belongs to a kind of of probability model, this class model has skewed popularity to the high mark word of occurrence frequency.Secondly in associating media correlation model automatic image annotation mark method, different candidates mark word and in mark process, are assumed to be it is separate, and the correlativity marking between word is not fully excavated.In fact with in piece image, between different labeled word, exist the multiple associations such as symbiosis, level or space.
The image that has comprised semantic objects such as " sun, sky, cloud, mountain, tree " such as a width, image vision content, can find out that " sun " and " sky " object exists certain spatial correlation, " sun " can not depart from " sky " this semantic object and independent existence; Equally, for " mountain " in picture material and " tree " two semantic objects, the existence of " tree " object is to take " mountain " semantic object as vision content background, the two has equally inseparable contact in image vision content, can not suppose utterly that these two mark words are separate marks.Therefore, associating media correlation model automatic image annotation algorithm thinks that in mark process different candidates mark way separate between word and have certain defect, may cause marking in annotation results semantic inconsistent phenomenon between word because ignoring between word correlativity.
Summary of the invention
In view of this, the invention provides a kind of image automatic annotation method based on correlativity between word, to overcome associating media correlation model automatic image annotation algorithm, in mark process, think that different candidates mark the defect that way separate between word exists, solve because ignoring between word correlativity and cause marking in annotation results semantic inconsistent problem between word.The technical scheme that the present invention proposes is:
An image automatic annotation method based on correlativity between word, training set T comprises l image, described l image construction image collection P=[p
1p
2p
l]; Each image labeling of described training set T has n mark word, and in training set T, all mark words form mark set of words W=[w
1w
2w
s]; Each image of training set T has corresponding vision lemma, and in training set T, all vision lemmas form the set B=[b of visual word unit
1b
2b
y], image to be marked is I, the method comprises:
A. according to formula
in calculation training set T, the semantic vector of each mark word w, is expressed as vector form w=< v by mark word w
1, v
2..., v
m>, wherein, c
ifor context relation word, total m context relation word, p (c
i) be context relation word c
ioverall distribution probability, p (c
i/ w) represent context relation word c
ithe ratio of the total degree that the co-occurrence number of times with mark word w in training set T and mark word w occur in training set T,
B. according to formula
calculate the semantic similarity between mark word, wherein || || for vectorial mould calculates;
C. according to formula
calculate p (A), wherein A is mark phrase { w
1, w
2... w
n, n is the number of mark phrase;
F. by formula A=argmax
ap (I/A) p (A) calculates the mark phrase A of image I to be marked.
In such scheme, step D further comprises:
P(w
i) be mark word w
ithe number of times occurring in training set T gathers with training the ratio that total degree appears in all mark words of T,
P(w
i, b
1..., b
n) computing method be:
Wherein, α
jwith β
jfor smoothing parameter, it is an empirical value;
# (w
i, J) represent mark word w
iin training image J, whether occur, if so, # (w
i, J)=1, otherwise # (w
i, J)=0;
# (w
i, T) represent mark word w
iin training set T, whether occur, if so, # (w
i, T)=1, otherwise # (w
i, T)=0;
# (b
k, J) represent vision lemma b
kin training image J, whether occur, if so, # (b
k, J)=1, otherwise # (b
k, J)=0;
| J| represents to mark in training image J total number of word and vision lemma; | T| represents to mark in training set T total number of word and vision lemma.
In such scheme, described context relation word is the mark word in training set T.
In sum, the technical scheme that the present invention proposes converts the joint probability calculation process of the mark word in associating media correlation model and image the probability of mark entry part hypograph appearance and two-part the solving of prior probability of mark phrase to, greatly reduce high frequency candidate and mark word for the impact of probability statistics model, make non-high frequency candidate mark the larger effect of word performance, recall ratio and precision ratio that non-high frequency candidate marks word have been improved, the similar language model of semanteme is incorporated in the middle of associating media correlation model simultaneously, by the similar language model of semanteme, remove to estimate the prior probability of one group of mark word, so more likely produce one group of mark word that semantic dependency is stronger.Thereby improve the integral body mark effect of image.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the embodiment of the present invention.
Embodiment
Clearer for what the object, technical solutions and advantages of the present invention were expressed, below in conjunction with drawings and the specific embodiments, the present invention is further described in more detail.
Technical scheme of the present invention is:
A. according to formula
in calculation training set T, the semantic vector of each mark word w, is expressed as vector form w=< v by mark word w
1, v
2..., v
m>, wherein, c
ifor context relation word, total m context relation word, p (c
i) be context relation word c
ioverall distribution probability, p (c
i/ w) represent context relation word c
ithe ratio of the total degree that the co-occurrence number of times with mark word w in training set T and mark word w occur in training set T,
B. according to formula
calculate the semantic similarity between mark word, wherein || || for vectorial mould calculates;
C. according to formula
calculate p (A), wherein A is mark phrase { w
1, w
2... w
n, n is the number of mark phrase;
E. basis
calculate p (I/A);
F. by formula A=arg max
ap (I/A) p (A) calculates the mark phrase A of image I to be marked.
Image labeling problem can be defined as at present: a given training set T, this training set T comprises image collection P and mark set of words W, and every width image p
iall completed mark word mark, the mark word of all images forms mark set of words W, does the one group of mark word A how choosing from described mark set of words W wherein mark a width new images I?
Image labeling method of the present invention adopts probability model, and its target is just to locate mark phrase A, its conditional probability p (A/I) maximum, that is:
A=arg?max
Ap(A/I) (3)
Wherein A is a mark phrase { w
1, w
2... w
n, one group of visual signature { b for image I
1, b
2, b
mrepresent, by image I is carried out pre-service (such as image cut apart, the operation such as feature extraction, characteristic value normalization) and image block region sort out computing and obtain.P (A/I) can be rewritten as following form:
Because the prior probability of piece image is considered to obey equally distributed conventionally, so p (I) can be regarded as a constant, and
p(A,I)=p(I/A)p(A) (5)
With formula (4), (5), formula (3) is simplified, is obtained:
A=arg?max
Ap(I/A)p(A) (6)
By p (I/A) and two probability of p (A) are combined, solve maximal value and find best mark phrase A.P (I/A) can obtain from original image marking model, and p (A) can obtain from language model.By giving different weights to two probability, represent original image model and the mark influential effect ability of language model to final acquisition:
It is carried out to following formal transformation:
A=arg?max
A(λ
1log?p(I/A)+λ
2log?p(A)) (8)
As long as calculate p (A) and p (I/A), just can obtain mark phrase A.Wherein, λ
1with λ
2be to determine in the machine learning of training plan image set and model process of establishing, carrying out in the automatic marking process of testing image is two constants.
The training set T that comprises l image of take below describes technical solution of the present invention as example, and image to be marked is I.L the image construction image collection P=[p of training set T
1p
2p
l]; Each image labeling of training set T has n mark word, and in training set T, all mark words form mark set of words W=[w
1w
2w
s]; Each image of training set T has corresponding vision lemma, and in training set T, all vision lemmas form the set B=[b of visual word unit
1b
2b
y].
Fig. 1 is the process flow diagram of the present embodiment, as shown in Figure 1, comprises the following steps:
Step 101: image I to be marked is carried out to image pre-service and segmented areas classification computing.
In this step, image I to be marked is carried out to image pre-service (image is cut apart, feature extraction, characteristic value normalization etc.), then carry out image block region and sort out computing, utilize clustering algorithm to sort out each image block region, and combine presentation video vision content: I={i with visual word unit
1i
2i
f.The preparation method of vision lemma is prior art, no longer describes in detail herein.
Step 102: calculate p (A) by the similar language model of semanteme.
In order to introduce correlation information between mark word in the similarity between mark word, the present invention has adopted each mark word w of semantic vector model representation: context relation set of words C=[c
1c
2c
m], each element c
irepresent a context relation word, total m context relation word, can choose all mark words of the mark set of words W in training set T as context relation word, i.e. C=W.Each the context relation word vector representation associated with it for w of mark word, i.e. w=< v
1, v
2..., v
m>, wherein each semantic component v
icalculating be defined as context relation word c
iconditional probability and context relation word c with respect to mark word w
ithe ratio of probability:
P (c wherein
i) expression context relation word c
ioverall distribution probability, for being uniformly distributed.Conditional probability p (c
i/ w) represent context relation word c
ithe ratio of the total degree that co-occurrence number of times during all image labelings of image collection P with mark word w in training set T occurs during at all image labelings of image collection P with mark word w:
P(c
i/ w) represent the intensity distributions of vocabulary w and context relation word co-occurrence, then be exactly in order to prevent semantic vector w=< v divided by the whole probability of each context relation word
1, v
2..., v
m> is dominated by the high context relation word of the frequency of occurrences, because high-frequency conjunctive word often also has very large conditional probability.As shown in table 1, wherein " sky ", " sun ", " clouds ", " town " represent a group context conjunctive word, and " tree ", " building ", " river " are one group of mark word, and the semantic vector of mark word represents as shown in table 1.
Table 1
? | sky | sun | clouds | town |
tree | 2.56 | 0.91 | 0.74 | 0.63 |
building | 5.01 | 0.57 | 2.41 | 21.19 |
river | 2.57 | 2.57 | 1.12 | 5.72 |
Then to calculate the semantic similarity between mark word.The calculating of similarity is as shown in Equation 11:
Wherein || || for vectorial mould calculates.
W
iw
jcalculating as shown in Equation 12:
C wherein
krepresent context relation word.Between mark word, semantic similarity is as shown in table 2.Similarity span is 0 to 1, and the similarity between two mark words of the higher expression of numerical value is higher, and the probability that they appear in same piece image is just larger.
Table 2
? | tree | road | | wood |
tree | ||||
1 | 0.1723 | 0.4311 | 0.2140 | |
road | 0.1723 | 1 | 0.1742 | 0.0021 |
sky | 0.4311 | 0.1742 | 1 | 0.0383 |
wood | 0.2140 | 0.0021 | 0.0383 | 1 |
Suppose in same mark, mark vocabulary is semantic relevant to context relation word, so one group of mark word A={w
1, w
2..., w
nprobability p (A) can obtain with similarity that other mark between word by calculating each mark word:
Formula 10,11,12 is updated in formula 13, can calculates the Probability p (A) of mark phrase:
Step 103: calculate p (I/A) by associating media correlation model.
In this step, first according to formula
design conditions Probability p (I/w
i).Wherein,
P(w
i) computing method be:
With mark word w
ithere is the ratio value representation vocabulary w of total degree in the number of times occurring and all mark words in training set T
iprior probability p (w
i):
P(w
i, b
1..., b
n) computing method be:
P (J) is illustrated in the probability of randomly drawing a width training image J in image collection P, is generally assumed to be and is uniformly distributed; p(w
i/ J) represent to occur vocabulary w in training image J
iposterior probability; And p (b
k/ J) represent to occur vision lemma b in training image J
kposterior probability.The probable value of each is estimated as follows:
Wherein, α
jwith β
jfor smoothing parameter, it is an empirical value; # (w
i, J) represent mark word w
iin training image J, whether occur, if so, # (w
i, J)=1, otherwise # (w
i, J)=0; # (w
i, T) represent mark word w
iin training set T, whether occur, if so, # (w
i, T)=1, otherwise # (w
i, T)=0; # (b
k, J) represent vision lemma b
kin training image J, whether occur, if so, # (b
k, J)=1, otherwise # (b
k, J)=0; | J| represents to mark in training image J total number of word and vision lemma; | T| represents to mark in training set T total number of word and vision lemma.
Step 104: the phrase to be marked that calculates image I to be marked.
Below p (A) and p (I/A) have been solved respectively, according to A=arg max
a(λ
1log p (I/A)+λ
2log p (A)) can be piece image I and calculate mark phrase A
The foregoing is only preferred embodiment of the present invention, not in order to limit the present invention, all any modifications of making within the spirit and principles in the present invention, be equal to replacement, improvement etc., within all should being included in the scope of protection of the invention.
Claims (3)
1. the image automatic annotation method based on correlativity between word, is characterized in that, training set T comprises l image, described l image construction image collection P=[p
1p
2p
l]; Each image labeling of described training set T has n mark word, and in training set T, all mark words form mark set of words W=[w
1w
2w
s]; Each image of training set T has corresponding vision lemma, and in training set T, all vision lemmas form the set B=[b of visual word unit
1b
2b
y], image to be marked is I, the method comprises:
A. according to formula
in calculation training set T, the semantic vector of each mark word w, is expressed as vector form w=< v by mark word w
1, v
2..., v
m>, wherein, c
ifor context relation word, total m context relation word, p (c
i) be context relation word c
ioverall distribution probability, p (c
i/ w) represent context relation word c
ithe ratio of the total degree that the co-occurrence number of times with mark word w in training set T and mark word w occur in training set T,
B. according to formula
calculate the semantic similarity between mark word, wherein || || for vectorial mould calculates;
C. according to formula
calculate p (A), wherein A is mark phrase { w
1, w
2... w
n, n is the number of mark phrase;
F. by formula A=argmax
ap (I/A) p (A) calculates the mark phrase A of image I to be marked.
2. method according to claim 1, is characterized in that, step D further comprises:
P(w
i) be mark word w
ithe number of times occurring in training set T gathers with training the ratio that total degree appears in all mark words of T,
P(w
i, b
1..., b
n) computing method be:
Wherein, α
jwith β
jfor smoothing parameter, it is an empirical value;
# (w
i, J) represent mark word w
iin training image J, whether occur, if so, # (w
i, J)=1, otherwise # (w
i, J)=0;
# (w
i, T) represent mark word w
iin training set T, whether occur, if so, # (w
i, T)=1, otherwise # (w
i, T)=0;
# (b
k, J) represent vision lemma b
kin training image J, whether occur, if so, # (b
k, J)=1, otherwise # (b
k, J)=0;
| J| represents to mark in training image J total number of word and vision lemma; | T| represents to mark in training set T total number of word and vision lemma.
3. method according to claim 1, is characterized in that, described context relation word is the mark word in training set T.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410008553.1A CN103714178B (en) | 2014-01-08 | 2014-01-08 | Automatic image marking method based on word correlation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410008553.1A CN103714178B (en) | 2014-01-08 | 2014-01-08 | Automatic image marking method based on word correlation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103714178A true CN103714178A (en) | 2014-04-09 |
CN103714178B CN103714178B (en) | 2017-01-25 |
Family
ID=50407153
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410008553.1A Active CN103714178B (en) | 2014-01-08 | 2014-01-08 | Automatic image marking method based on word correlation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103714178B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104794183A (en) * | 2015-04-10 | 2015-07-22 | 浙江大学 | Picture labeling method based on multiple views and multiple labels |
CN108268875A (en) * | 2016-12-30 | 2018-07-10 | 广东精点数据科技股份有限公司 | A kind of image meaning automatic marking method and device based on data smoothing |
CN108604902A (en) * | 2016-02-08 | 2018-09-28 | 皇家飞利浦有限公司 | Determine the device and method of cluster |
WO2020073952A1 (en) * | 2018-10-10 | 2020-04-16 | 腾讯科技(深圳)有限公司 | Method and apparatus for establishing image set for image recognition, network device, and storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1920820A (en) * | 2006-09-14 | 2007-02-28 | 浙江大学 | Image meaning automatic marking method based on marking significance sequence |
US20090289942A1 (en) * | 2008-05-20 | 2009-11-26 | Timothee Bailloeul | Image learning, automatic annotation, retrieval method, and device |
CN101620615A (en) * | 2009-08-04 | 2010-01-06 | 西南交通大学 | Automatic image annotation and translation method based on decision tree learning |
CN101685464A (en) * | 2009-06-18 | 2010-03-31 | 浙江大学 | Method for automatically labeling images based on community potential subject excavation |
CN102298606A (en) * | 2011-06-01 | 2011-12-28 | 清华大学 | Random walking image automatic annotation method and device based on label graph model |
CN102542067A (en) * | 2012-01-06 | 2012-07-04 | 上海交通大学 | Automatic image semantic annotation method based on scale learning and correlated label dissemination |
-
2014
- 2014-01-08 CN CN201410008553.1A patent/CN103714178B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1920820A (en) * | 2006-09-14 | 2007-02-28 | 浙江大学 | Image meaning automatic marking method based on marking significance sequence |
US20090289942A1 (en) * | 2008-05-20 | 2009-11-26 | Timothee Bailloeul | Image learning, automatic annotation, retrieval method, and device |
CN101685464A (en) * | 2009-06-18 | 2010-03-31 | 浙江大学 | Method for automatically labeling images based on community potential subject excavation |
CN101620615A (en) * | 2009-08-04 | 2010-01-06 | 西南交通大学 | Automatic image annotation and translation method based on decision tree learning |
CN102298606A (en) * | 2011-06-01 | 2011-12-28 | 清华大学 | Random walking image automatic annotation method and device based on label graph model |
CN102542067A (en) * | 2012-01-06 | 2012-07-04 | 上海交通大学 | Automatic image semantic annotation method based on scale learning and correlated label dissemination |
Non-Patent Citations (3)
Title |
---|
BING-KUN BAO ET AL.: ""Hidden-Concept Driven Multilabel Image Annotation and Label Ranking"", 《IEEE TRANSACTIONS ON MULTIMEDIA》 * |
MAHDIA BAKALEM ET AL.: ""Latent Semantic Analysis-based Image Auto Annotation"", 《IEEE CONF. ON MACHINE AND WEB INTELLIGENCE》 * |
刘咏梅等: ""词间相关性的CMRM图像标注方法"", 《智能系统学报》 * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104794183A (en) * | 2015-04-10 | 2015-07-22 | 浙江大学 | Picture labeling method based on multiple views and multiple labels |
CN108604902A (en) * | 2016-02-08 | 2018-09-28 | 皇家飞利浦有限公司 | Determine the device and method of cluster |
CN108268875A (en) * | 2016-12-30 | 2018-07-10 | 广东精点数据科技股份有限公司 | A kind of image meaning automatic marking method and device based on data smoothing |
CN108268875B (en) * | 2016-12-30 | 2020-12-08 | 广东精点数据科技股份有限公司 | Image semantic automatic labeling method and device based on data smoothing |
WO2020073952A1 (en) * | 2018-10-10 | 2020-04-16 | 腾讯科技(深圳)有限公司 | Method and apparatus for establishing image set for image recognition, network device, and storage medium |
US11853352B2 (en) | 2018-10-10 | 2023-12-26 | Tencent Technology (Shenzhen) Company Limited | Method and apparatus for establishing image set for image recognition, network device, and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN103714178B (en) | 2017-01-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Gao et al. | Database saliency for fast image retrieval | |
CN103617157B (en) | Based on semantic Text similarity computing method | |
CN102902821B (en) | The image high-level semantics mark of much-talked-about topic Network Based, search method and device | |
CN104391942B (en) | Short essay eigen extended method based on semantic collection of illustrative plates | |
CN103678670B (en) | Micro-blog hot word and hot topic mining system and method | |
CN104199857B (en) | A kind of tax document hierarchy classification method based on multi-tag classification | |
CN103544697B (en) | A kind of image partition method based on hypergraph analysis of spectrum | |
CN104881458B (en) | A kind of mask method and device of Web page subject | |
CN105844424A (en) | Product quality problem discovery and risk assessment method based on network comments | |
CN101950284A (en) | Chinese word segmentation method and system | |
CN100573557C (en) | A kind of SAR image partition method of short annealing based on MRF | |
CN103425757A (en) | Cross-medial personage news searching method and system capable of fusing multi-mode information | |
Tran et al. | Cluster-based similarity aggregation for ontology matching | |
CN104484380A (en) | Personalized search method and personalized search device | |
CN103714178A (en) | Automatic image marking method based on word correlation | |
CN102521368A (en) | Similarity matrix iteration based cross-media semantic digesting and optimizing method | |
WO2013118435A1 (en) | Semantic similarity level computation method, system and program | |
CN105653640A (en) | Collaborative filtering recommendation method based on trust mechanism | |
CN102637199B (en) | Image marking method based on semi-supervised subject modeling | |
CN110390022A (en) | A kind of professional knowledge map construction method of automation | |
Zhao et al. | Retrieving Social Flooding Images Based on Multimodal Information. | |
CN103927730A (en) | Image noise reduction method based on Primal Sketch correction and matrix filling | |
CN107301426A (en) | A kind of multi-tag clustering method of shoe sole print image | |
CN103064907A (en) | System and method for topic meta search based on unsupervised entity relation extraction | |
Ma et al. | PSVM: a preference-enhanced SVM model using preference data for classification |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |