CN107885854A - A kind of semi-supervised cross-media retrieval method of feature based selection and virtual data generation - Google Patents

A kind of semi-supervised cross-media retrieval method of feature based selection and virtual data generation Download PDF

Info

Publication number
CN107885854A
CN107885854A CN201711124618.9A CN201711124618A CN107885854A CN 107885854 A CN107885854 A CN 107885854A CN 201711124618 A CN201711124618 A CN 201711124618A CN 107885854 A CN107885854 A CN 107885854A
Authority
CN
China
Prior art keywords
msub
mrow
data
msup
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201711124618.9A
Other languages
Chinese (zh)
Other versions
CN107885854B (en
Inventor
孙建德
于恩
李静
张化祥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong Normal University
Original Assignee
Shandong Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong Normal University filed Critical Shandong Normal University
Priority to CN201711124618.9A priority Critical patent/CN107885854B/en
Publication of CN107885854A publication Critical patent/CN107885854A/en
Application granted granted Critical
Publication of CN107885854B publication Critical patent/CN107885854B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The present invention proposes a kind of semi-supervised cross-media retrieval method of feature based selection and virtual data generation.This method proposes to generate some virtual data points according to the feature of training data to expand training data, while uses l during two pairs of projection matrixes are learnt2,1Norm carries out feature selecting.Specifically, its class center is asked to every a kind of image and text first, then generates new data point at random around it, form new training data;Then, two pairs of projection matrixes are learnt using new data, at the same time, using l2,1Norm constraint, carry out feature selecting;Finally retrieval result is assessed with mAP values.This method not only generates some random number strong points to improve the diversity of training data, while some can be selected more to have the feature of distinction and abundant information when two pairs of projection matrixes are learnt.Experimental result on three different pieces of information collection also indicates that the superiority of the method.

Description

A kind of semi-supervised cross-media retrieval method of feature based selection and virtual data generation
Technical field
The present invention relates to cross-media retrieval method, is given birth to more specifically to a kind of selection of feature based and virtual data Into semi-supervised cross-media retrieval method.
Background technology
With the development of multimedia technology, increasing data can be expressed as different mode, and different modalities Data may have same semantic information.Therefore, how to explore these has identical semantic but shows as different modalities number Relation between becomes particularly important.Wherein, it is nearest for many years in, cross-media retrieval technology increasingly causes researcher's Concern.Cross-media retrieval just refers to retrieve other with identical semantic information by the use of a kind of data of mode as inquiry data The data of mode.By taking the retrieval of picture and text as an example, picture can be used to go text of the retrieval with corresponding semantic information, letter Claim:I2T;Or picture of the retrieval with corresponding semantic information is removed using text, referred to as:T2I.The present invention is between picture and text Retrieval exemplified by analyzed and tested, but the retrieval that the method can be expanded between other different modalities.
In cross-media retrieval technology, most important problem is:The data of different modalities have different character representations, this Different a bit to be characterized in different dimensional spaces, the similitude between such isomeric data cannot directly compare.Cause This, the problem of cross-media retrieval field is primarily upon, is how to cross over this semantic gap.A kind of popular solution method is just It is sub-space learning method.Sub-space learning method is intended to learn a potential semantic space, in this potential semantic space The similitude of isomeric data can be directly measured.Traditional sub-space learning method is a pair of projection matrixes of study, passes through this The data of different modalities are mapped in a potential semantic space by projection matrix can, such isomeric data it is similar Property can is measured.A kind of popular method is:Canonical correlation analysis (Canonical Correlation Analysis, CCA), CCA has learnt a pair of projection matrixes, is maximized when by the Feature Mapping of different modalities to semantic space Correlation between isomeric data.Based on CCA, semantic relevant matches (Semantic Correlation Match, SCM) use Logistic regression obtains semantic space.Another popular method is:Offset minimum binary (Partial Least Squares, PLS), PLS is intended to learn two potential semantic spaces by maximizing the correlation between isomeric data.In addition, Normalize multi-view analysis (generalized multi-view analysis, GMA) and GMLDA based on GMA and GMMFA obtains the feature of multi-angle by using label information, and achieves more preferable effect.
However, there is directionality, i.e. image retrieval text (I2T) and text retrieval image in common cross-media retrieval task (T2I), above method only learns a pair of projection matrixes, does not emphasize to inquire about the importance of data.Specifically, in I2T In task, picture is more decisive for study projection matrix;Similarly, in T2I tasks, then the importance of text is more emphasized. Therefore, the method for learning a pair of projection matrixes is extremely difficult to optimal effect.In order to emphasize that the weight of data is inquired about in different task The property wanted, cross-media retrieval (Modality-dependent Cross-media Retrieval, MDCR) method based on mode Two pairs of projection matrixes of study are proposed, i.e., a pair of projection matrixes is learnt respectively to I2T and T2I tasks, can thus taken into full account The importance of data is inquired about, therefore the precision retrieved also is greatly improved.
But above method all simply has the method for supervision, is only trained and have ignored using only markd data Unlabelled data, while cannot more expand intrinsic data set.Secondly, current method only from how to measure isomeric data it Between the angle of similitude set out, it is intended to learn more effective projection matrix, it is more accurate always to be obtained in semantic space Comparison, still, they all have ignored when learning projection matrix it is more rich to information content, more distinction feature choosing Select.Therefore, we have invented a kind of semi-supervised method that can generate virtual data point at random based on MDCR, while use l2,1 Norm carries out feature selecting.
The content of the invention
The invention provides the semi-supervised cross-media retrieval technology of a kind of selection of feature based and pseudo-random data generation.Pass The cross-media retrieval method of system, otherwise it is only to be used only to have what flag data was trained to have measure of supervision, otherwise it is to select one Part Unlabeled data adds the semi-supervised method of training.The present invention propose, generated in markd data basis some with Correlation pseudorandom virtual data point, so not only it is contemplated that unlabelled data, it is related that some can also be increased Virtual data point is to improve the precision of training.Meanwhile unlike traditional search method:Our method is for different The different projection matrix of tasking learning, l is used when learning projection matrix2,1Norm carries out feature selecting.On the whole, we Method simultaneously consider the diversity of training data and the selection of validity feature.
The concrete technical scheme of the present invention is as follows:
A kind of semi-supervised cross-media retrieval technology of feature based selection and virtual data generation, comprises the following steps:
Step 1:Data-oriented collectionN represents the sum of data pair, xiRepresentative picture feature, tiRepresent text Eigen, then, picture and text feature matrix can be expressed as:XG=[x1,x2,...,xn-1,xn] and TG=[t1, t2,t3,...,tn-1,tn];
Step 2:Pseudorandom virtual data point is generated, legacy data collection is expanded, specific method is:Calculate XGAnd TG Per a kind of class center, i.e., for every a kind of data, calculate average of such data per dimension, obtained each dimension average structure Into it is new vector as such class center;Then, using the average of every dimension as center, at it, generation n' is individual at random up and down Numerical value, the sets of random values in all dimensions is combined n' pseudorandom virtual data of generation, by the puppet Random virtual data point adds the data set G after legacy data collection is expandedall={ G, G'}, picture and text after expansion Eigenmatrix is expressed as:X=[x1,...xn,x1',x'2...x'n] and T=[t1,...,tn,t1',t'2,...,t'n];
Step 3:Build object function:
Objective function:
Wherein, U, V represent this method a pair of projection matrixes to be learnt, and C (U, V) is correlation analysis item so that multimode The data of state can be kept into potential semantic space to neighbor relationships;L (U, V) is from image or text modality feature Space to semantic space linear regression item, for keep with identical semantic different modalities data neighbor relationships;N(U, V it is) regular terms, the selection for feature;
According to formula (1), image inspection text I2T and the object function of text inspection image T2I retrieval tasks are respectively obtained, such as Under:
(1) I2T object function is:
Wherein, U1,V1It is the projection matrix that learn to obtain in I2T tasks, the U corresponded respectively in formula (1), V, β are Coefficient of balance and 0≤β≤1, Y are semantic matrixes;
(2) T2I object function is:
Wherein, U2,V2It is the projection matrix that learn to obtain in T2I tasks, the U, V corresponded respectively in formula (1);
Step 4:By iterative method, optimal projection matrix is obtained:
Because formula (2) and (3) are non-convex, therefore use the method for control variable to solve, i.e., local derviation asked to U and V respectively, And make it be equal to zero, projection matrix U and V value can be obtained;Then pass through continuous iteration, until convergence, try to achieve projection matrix U and V optimal value.
Especially, in step 3, N (U, V)=λ1||U||2,12||V||2,1, wherein λ1, λ2For balancing two canonicals , and all be positive number, this bound term is used to select when projection matrix learn more with distinction and abundant information Feature.
Brief description of the drawings
Fig. 1 is the inventive method flow chart.
Embodiment
1. data set is handled:
Wikipedia, altogether comprising 10 classes, 2866 picture-texts pair.We select 2173 picture-texts pair As initial training data, remainder is test data.Its picture feature be 4096 dimension CNN features, text feature 100 Tie up LDA features.
Pascal Sentence, 20 classes altogether, per 50 picture-texts pair of class.We select 30 figures in every class Picture-text is to as initial training data, remaining is test data.Its picture feature be 4096 dimension CNN features, text feature For 100 dimension LDA features.
INRIA-Websearch, altogether 353 class, 71478 image-texts pair.We randomly choose therein 70% and made For initial training data, remaining is test data.Its picture feature is the CNN features of 4096 dimensions, and text feature is 1000 dimension LDA Feature.
2. specific implementation step of the present invention:
Step 1:Data-oriented collectionN represents the sum of data pair, xiRepresentative picture feature, tiRepresent text Eigen, then, picture and text feature matrix can be expressed as:XG=[x1,x2,...,xn-1,xn] and TG=[t1, t2,t3,...,tn-1,tn]。
Step 2:Pseudorandom virtual data point is generated, legacy data collection is expanded, specific method is:Calculate XGAnd TG Per a kind of class center, i.e., for every a kind of data, calculate average of such data per dimension, obtained each dimension average structure Into it is new vector as such class center;Then, using the average of every dimension as center, at it, generation n' is individual at random up and down Numerical value, the sets of random values in all dimensions is combined n' pseudorandom virtual data of generationBy described in Pseudorandom virtual data point adds the data set G after legacy data collection is expandedall={ G, G'}, picture and text after expansion Eigen matrix is expressed as:X=[x1,...xn,x1',x'2...x'n] and T=[t1,...,tn,t′1,t′2,...,t ′n]。
Step 3:Build object function:
Objective function:
Wherein, U, V represent this method a pair of projection matrixes to be learnt, and C (U, V) is correlation analysis item so that multimode The data of state can be kept into potential semantic space to neighbor relationships;L (U, V) is from image or text modality feature Space to semantic space linear regression item, for keep with identical semantic different modalities data neighbor relationships;N(U, V it is) regular terms, the selection for feature;
According to formula (1), image inspection text I2T and the object function of text inspection image T2I retrieval tasks are respectively obtained, such as Under:
(1) I2T object function is:
Wherein, U1,V1It is the projection matrix that learn to obtain in I2T tasks, β is coefficient of balance and 0≤β≤1, Y are semantemes Matrix, and:N(U1,V1)=λ1||U1||2,12||V1||2,1, wherein λ1, λ2For balancing two regular terms, and all it is positive number;
(2) T2I object function is:
Wherein, U2,V2It is the projection matrix N (U that learn to obtain in T2I tasks2,V2)=λ1||U2||2,12||V2| |2,1
Step 4:By iterative method, optimal projection matrix is obtained:
Because formula (2) and (3) are non-convex, therefore use the method for control variable to solve, i.e., local derviation asked to U and V respectively, And make it be equal to zero, projection matrix U and V value can be obtained;Then pass through continuous iteration, until convergence, try to achieve projection matrix U and V optimal value.
Especially, for l2,1Norm can using mark come derivation, such as:Matrix U is defined, then:||U||2,1=Tr (UTRU), R is a diagonal matrix,uiU every a line is represented, ε is a minimum real number.
3. evaluation criteria (mAP)
We use mean accuracy average (mean average precision, mAP) evaluation criteria, last to evaluate Retrieval effectiveness.We define the average precision inquired about each time first:
Wherein, N represents the sum of sample in test data, has and corresponding category when in the sort result of ith retrieval Rel (i)=1 when signing identical, otherwise rel (i)=0.P (i) represents the precision of ith retrieval ordering result.So, all inquiries AP values be averagely exactly mAP.
4. algorithm is realized
(1)I2T:
Input:Picture feature matrix XGWith text feature matrix TG, sample labeling matrix Y, parameter lambda1, λ2, β
Generate virtual data:For every a kind of data, the average per dimension is calculated first, as in this kind of class The heart,
Then, using the average of every dimension as center, in its n' numerical value of generation at random up and down, in all dimensions with Machine value, which is combined, can be formed by the virtual data of n', finally, the virtual data of generation add input picture and Text feature matrix, obtain new training picture feature matrix X and text feature matrix T.
Initialization:Initial projection matrix U1, V1For unit matrix.
Solve optimal solution:According to the U tried to achieve1=(XXT1R11)-1[βXY+(1-β)XTTV1] and
V1=[(1- β) TTT2R12]-1(1-β)TXTU1, by constantly iteration, until result restrain to obtain it is optimal U1, V1
This process false code is as follows:
(2)T2I:
It is similar with I2T tasks, finally obtain optimal projection matrix U2, V2
5. results contrast
We are tested on three data sets respectively, and compared for other current popular 7 kinds of methods (PLS, CCA, SM, SCM, GMMFA, GMLDA, MDCR), following table shows that the inventive method all shows preferably to examine on different pieces of information collection Suo Xiaoguo.

Claims (2)

1. a kind of feature based selection and the semi-supervised cross-media retrieval technology of virtual data generation, comprise the following steps:
Step 1:Data-oriented collectionN represents the sum of data pair, xiRepresentative picture feature, tiRepresent text spy Sign, then, picture and text feature matrix can be expressed as:XG=[x1,x2,...,xn-1,xn] and TG=[t1,t2, t3,...,tn-1,tn];
Step 2:Pseudorandom virtual data point is generated, legacy data collection is expanded, specific method is:Calculate XGAnd TGIt is each The class center of class, i.e., for every a kind of data, average of such data per dimension is calculated, what obtained each dimension average was formed New vector is as such class center;Then, using the average of every dimension as center, n' number is generated at random up and down at it Value, the sets of random values in all dimensions is combined n' pseudorandom virtual data of generation By the puppet Random virtual data point adds the data set G after legacy data collection is expandedall={ G, G'}, picture and text after expansion Eigenmatrix is expressed as:X=[x1,...xn,x1',x'2...x'n] and T=[t1,...,tn,t1',t'2,...,t'n];
Step 3:Build object function:
Objective function:
Wherein, U, V represent this method a pair of projection matrixes to be learnt, and C (U, V) is correlation analysis item so that multi-modal Data can be kept into potential semantic space to neighbor relationships;L (U, V) is from image or text modality feature space To the linear regression item of semantic space, for keeping the neighbor relationships with identical semantic different modalities data;N (U, V) is Regular terms, the selection for feature;
According to formula (1), image inspection text I2T and the object function of text inspection image T2I retrieval tasks are respectively obtained, it is as follows:
(1) I2T object function is:
<mrow> <munder> <mrow> <mi>m</mi> <mi>i</mi> <mi>n</mi> </mrow> <mrow> <msub> <mi>U</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>V</mi> <mn>1</mn> </msub> </mrow> </munder> <mi>f</mi> <mrow> <mo>(</mo> <msub> <mi>U</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>V</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mi>&amp;beta;</mi> <mo>|</mo> <mo>|</mo> <msup> <mi>X</mi> <mi>T</mi> </msup> <msub> <mi>U</mi> <mn>1</mn> </msub> <mo>-</mo> <mi>Y</mi> <mo>|</mo> <msubsup> <mo>|</mo> <mi>F</mi> <mn>2</mn> </msubsup> <mo>+</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&amp;beta;</mi> <mo>)</mo> </mrow> <mo>|</mo> <mo>|</mo> <msup> <mi>X</mi> <mi>T</mi> </msup> <msub> <mi>U</mi> <mn>1</mn> </msub> <mo>-</mo> <msup> <mi>T</mi> <mi>T</mi> </msup> <msub> <mi>V</mi> <mn>1</mn> </msub> <mo>|</mo> <msubsup> <mo>|</mo> <mi>F</mi> <mn>2</mn> </msubsup> <mo>+</mo> <mi>N</mi> <mrow> <mo>(</mo> <msub> <mi>U</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>V</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow>
Wherein, U1,V1It is the projection matrix that learn to obtain in I2T tasks, the U corresponded respectively in formula (1), V, β are balances Coefficient and 0≤β≤1, Y are semantic matrixes;
(2) T2I object function is:
<mrow> <munder> <mrow> <mi>m</mi> <mi>i</mi> <mi>n</mi> </mrow> <mrow> <msub> <mi>U</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi>W</mi> <mn>2</mn> </msub> </mrow> </munder> <mi>f</mi> <mrow> <mo>(</mo> <msub> <mi>U</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi>V</mi> <mn>2</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mi>&amp;beta;</mi> <mo>|</mo> <mo>|</mo> <msup> <mi>T</mi> <mi>T</mi> </msup> <msub> <mi>V</mi> <mn>2</mn> </msub> <mo>-</mo> <mi>Y</mi> <mo>|</mo> <msubsup> <mo>|</mo> <mi>F</mi> <mn>2</mn> </msubsup> <mo>+</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&amp;beta;</mi> <mo>)</mo> </mrow> <mo>|</mo> <mo>|</mo> <msup> <mi>X</mi> <mi>T</mi> </msup> <msub> <mi>U</mi> <mn>2</mn> </msub> <mo>-</mo> <msup> <mi>T</mi> <mi>T</mi> </msup> <msub> <mi>V</mi> <mn>2</mn> </msub> <mo>|</mo> <msubsup> <mo>|</mo> <mi>F</mi> <mn>2</mn> </msubsup> <mo>+</mo> <mi>N</mi> <mrow> <mo>(</mo> <msub> <mi>U</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi>V</mi> <mn>2</mn> </msub> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow>
Wherein, U2,V2It is the projection matrix that learn to obtain in T2I tasks, the U, V corresponded respectively in formula (1);
Step 4:By iterative method, optimal projection matrix is obtained:
Because formula (2) and (3) are non-convex, therefore use the method for control variable to solve, i.e., local derviation is asked to U and V respectively, and make It is equal to zero, can obtain projection matrix U and V value;Then continuous iteration is passed through, until convergence, tries to achieve projection matrix U and V Optimal value.
2. feature based selection as claimed in claim 1 and the semi-supervised cross-media retrieval technology of virtual data generation, it is special Sign is:In step 3, N (U, V)=λ1||U||2,12||V||2,1, wherein λ1, λ2For balancing two regular terms, and all it is Positive number, this bound term are used for the feature that selection when projection matrix is learnt more has distinction and abundant information.
CN201711124618.9A 2017-11-14 2017-11-14 semi-supervised cross-media retrieval method based on feature selection and virtual data generation Active CN107885854B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711124618.9A CN107885854B (en) 2017-11-14 2017-11-14 semi-supervised cross-media retrieval method based on feature selection and virtual data generation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711124618.9A CN107885854B (en) 2017-11-14 2017-11-14 semi-supervised cross-media retrieval method based on feature selection and virtual data generation

Publications (2)

Publication Number Publication Date
CN107885854A true CN107885854A (en) 2018-04-06
CN107885854B CN107885854B (en) 2020-01-31

Family

ID=61776703

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711124618.9A Active CN107885854B (en) 2017-11-14 2017-11-14 semi-supervised cross-media retrieval method based on feature selection and virtual data generation

Country Status (1)

Country Link
CN (1) CN107885854B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109784405A (en) * 2019-01-16 2019-05-21 山东建筑大学 Cross-module state search method and system based on pseudo label study and semantic consistency
CN109857892A (en) * 2018-12-29 2019-06-07 西安电子科技大学 Semi-supervised cross-module state Hash search method based on category transmitting
CN112419324A (en) * 2020-11-24 2021-02-26 山西三友和智慧信息技术股份有限公司 Medical image data expansion method based on semi-supervised task driving

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105069173A (en) * 2015-09-10 2015-11-18 天津中科智能识别产业技术研究院有限公司 Rapid image retrieval method based on supervised topology keeping hash
US9332137B2 (en) * 2012-09-28 2016-05-03 Interactive Memories Inc. Method for form filling an address on a mobile computing device based on zip code lookup
CN106462642A (en) * 2014-06-24 2017-02-22 谷歌公司 Methods, Systems And Media For Performing Personalized Actions On Mobile Devices Associated With A Media Presentation Device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9332137B2 (en) * 2012-09-28 2016-05-03 Interactive Memories Inc. Method for form filling an address on a mobile computing device based on zip code lookup
CN106462642A (en) * 2014-06-24 2017-02-22 谷歌公司 Methods, Systems And Media For Performing Personalized Actions On Mobile Devices Associated With A Media Presentation Device
CN105069173A (en) * 2015-09-10 2015-11-18 天津中科智能识别产业技术研究院有限公司 Rapid image retrieval method based on supervised topology keeping hash

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YUNCHAO WEI等: "Modality-Dependent Cross-Media Retrieval", 《ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109857892A (en) * 2018-12-29 2019-06-07 西安电子科技大学 Semi-supervised cross-module state Hash search method based on category transmitting
CN109857892B (en) * 2018-12-29 2022-12-02 西安电子科技大学 Semi-supervised cross-modal Hash retrieval method based on class label transfer
CN109784405A (en) * 2019-01-16 2019-05-21 山东建筑大学 Cross-module state search method and system based on pseudo label study and semantic consistency
CN109784405B (en) * 2019-01-16 2020-09-08 山东建筑大学 Cross-modal retrieval method and system based on pseudo-tag learning and semantic consistency
CN112419324A (en) * 2020-11-24 2021-02-26 山西三友和智慧信息技术股份有限公司 Medical image data expansion method based on semi-supervised task driving
CN112419324B (en) * 2020-11-24 2022-04-19 山西三友和智慧信息技术股份有限公司 Medical image data expansion method based on semi-supervised task driving

Also Published As

Publication number Publication date
CN107885854B (en) 2020-01-31

Similar Documents

Publication Publication Date Title
Chen et al. Msdn: Mutually semantic distillation network for zero-shot learning
CN107402993B (en) The cross-module state search method for maximizing Hash is associated with based on identification
CN111639679B (en) Small sample learning method based on multi-scale metric learning
CN106649715B (en) A kind of cross-media retrieval method based on local sensitivity hash algorithm and neural network
CN106339756A (en) Training data generation method and device and searching method and device
CN104573669A (en) Image object detection method
CN110097095B (en) Zero sample classification method based on multi-view generation countermeasure network
CN109871454B (en) Robust discrete supervision cross-media hash retrieval method
CN111666427A (en) Entity relationship joint extraction method, device, equipment and medium
CN112418351B (en) Zero sample learning image classification method based on global and local context sensing
CN107885854A (en) A kind of semi-supervised cross-media retrieval method of feature based selection and virtual data generation
CN114186084B (en) Online multi-mode Hash retrieval method, system, storage medium and equipment
CN104424254A (en) Method and device for obtaining similar object set and providing similar object set
Nascimento et al. A robust indoor scene recognition method based on sparse representation
CN114117153A (en) Online cross-modal retrieval method and system based on similarity relearning
Chen et al. An Improved Deep Fusion CNN for Image Recognition.
Liu et al. Incorporating domain and sentiment supervision in representation learning for domain adaptation
CN105740917B (en) The semi-supervised multiple view feature selection approach of remote sensing images with label study
Ren et al. Improving cross-domain recommendation through probabilistic cluster-level latent factor model
CN105701225A (en) Cross-media search method based on unification association supergraph protocol
CN114780723B (en) Portrayal generation method, system and medium based on guide network text classification
Xu et al. DeMT: Deformable mixer transformer for multi-task learning of dense prediction
Dou et al. Function-words enhanced attention networks for few-shot inverse relation classification
CN104036021A (en) Method for semantically annotating images on basis of hybrid generative and discriminative learning models
CN106250918A (en) A kind of mixed Gauss model matching process based on the soil-shifting distance improved

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant