CN110163258A - A kind of zero sample learning method and system reassigning mechanism based on semantic attribute attention - Google Patents

A kind of zero sample learning method and system reassigning mechanism based on semantic attribute attention Download PDF

Info

Publication number
CN110163258A
CN110163258A CN201910335801.6A CN201910335801A CN110163258A CN 110163258 A CN110163258 A CN 110163258A CN 201910335801 A CN201910335801 A CN 201910335801A CN 110163258 A CN110163258 A CN 110163258A
Authority
CN
China
Prior art keywords
space
semantic
image
attention
hidden layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910335801.6A
Other languages
Chinese (zh)
Other versions
CN110163258B (en
Inventor
刘洋
蔡登�
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201910335801.6A priority Critical patent/CN110163258B/en
Publication of CN110163258A publication Critical patent/CN110163258A/en
Application granted granted Critical
Publication of CN110163258B publication Critical patent/CN110163258B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Abstract

The invention discloses a kind of zero sample learning method and system that mechanism is reassigned based on semantic attribute attention, wherein zero sample learning method includes: that (1) establishes the neural network model that mechanism is reassigned based on semantic attribute attention;(2) weight between semantic feature is redistributed using the attention in semantic attribute space;(3) using the image data set training neural network model with label;(4) similarity of the semantic prototype of semantic feature and unknown class after calculating image weighting, calculates the similarity of the hidden layer Feature prototype of hidden layer feature and unknown class, two kinds of similarities is added to obtain the similarity of test image Yu each unknown class;(5) according to it is all kinds of between similarity be ranked up, the class for choosing the class of maximum similarity as the image is predicted.The present invention can make zero sample learning in the training process, can closely contact semantic space and hidden layer space, so that the result in conjunction with the joint classification in two kinds of spaces is more robust.

Description

It is a kind of based on semantic attribute attention reassign mechanism zero sample learning method and System
Technical field
The present invention relates to zero sample learning categorizing system fields, are divided again more particularly, to one kind based on semantic attribute attention The zero sample learning method and system with mechanism.
Background technique
In recent years, target classification is come from industry by vast as an important branch in computer vision field With the concern of academia researcher.The fast development for benefiting from depth learning technology has the target classification task of supervision to achieve Very big progress, but simultaneously, there are some limitations for the training method under this surveillance requirements.In Supervised classification, each class Require enough training samples with label.In addition, the classifier of study can only be to the class for belonging to training data and being covered Example classify, lack handle previously invisible class ability.In practical applications, each class may be not enough Training sample, it is also possible to there is a situation where that unlapped class appears in test sample in training.The target of zero sample learning is Classify to the example for belonging to the class being not covered in training, has become a fast-developing side in machine learning field To in computer vision, natural language processing and general fit calculation aspect are had a wide range of applications.
Zero sample learning method of mainstream at present, the two stages for being mainly taken based on attribute derive the mark for carrying out forecast image Label.The process of the derivation method are as follows: one image of input, model predict each attribute of this image in the first stage, then Second stage possesses the class of most like property set by search to infer its class label.For example, Christoph H.Lampert etc. People is in 2013 in IEEE mode analysis and machine intelligence periodical (The IEEE Transactions on Pattern Analysis and Machine Intelligence) on article " the Attribute-based classification that delivers For zero-shot visual object categorization " propose DAP model pass through learning probability attributive classification Device estimates the posterior probability of each attribute of image, is then inferred by calculating posterior probability and the MAP estimation of class The class label of image.International computer visions in 2016 and pattern-recognition meeting (The Conference on Computer Vision and Pattern Recognition) on include " Recovering the missing link: Predicting class-attribute associations for unsupervised zero-shot learning》 In, propose to be first each attribute learning probability classifier, is then classified by the method for random forest, this classification side Method is capable of handling some insecure attributes.There is domain migration in this two stage method, although for example, goal task It is the label of the class of forecast image, the intermediate task of DAP is study classifier related with image attributes.
The latest developments of zero sample learning directly learn the mapping from image feature space to attribute semantemes space.Wherein, It is embodied within 2016 IEEE mode analysis and machine intelligence periodical (The IEEE Transactions on Pattern Analysis and Machine Intelligence) on article " Label-embedding for image Classification " propose ALE model, using it is a kind of based on ranking loss function study image and attribute space between Bilinearity compatibility function.International computer visions in 2017 and pattern-recognition meeting (The Conference on Computer Vision and Pattern Recognition) on " the Semantic autoencoder for that includes Zero-shot learning " propose that a kind of semantic autocoder pressure projects to characteristics of image can be by image reconstruction Semantic space.(the The International Conference on Computer of international computer vision conference in 2017 Vision on), an entitled " Predicting visual exemplars of unseen classes for zero- Shot learning " article proposition, by the way that class semantic expressiveness is projected in visual signature space and is carried out in these projections Arest neighbors classification.
Other than common semantic attribute space, some newest work in recent years are often related to from more zero sample learning A space carries out joint category inferences.For example, international computer vision conference (The International in 2017 Conference on Computer Vision) on an entitled " Learning discriminative latent Attributes for zero-shot classification " article propose LAD model obtained using dictionary learning Have any different but retain the hidden feature space of semantic information.International computer visions in 2018 and pattern-recognition meeting (The Conference on Computer Vision and Pattern Recognition) on, one is entitled The article of " Discriminative Learning of Latent Features for Zero-Shot Recognition " It proposes a kind of combination to maximize between class distance and minimize the new hidden feature space of inter- object distance, and in semantic space and hidden Joint deduction is carried out on feature space.In European Computer vision international conference (The European Conference in 2018 On Computer Vision) in, an entitled " Learning Class Prototypes via Structure Alignment for Zero-Shot Recognition " article propose CDL model in vision and semantic space simultaneously It is aligned class formation.However, these methods all think that every attribute is all of equal importance in assorting process in itself, and neglect Having omited them, there are different distributions, variance, comentropies etc. between inhomogeneity, and this processing mode is some challenging Image on, it is easy to cause wrong classification.
Summary of the invention
The present invention provides a kind of zero sample learning method and system that mechanism is reassigned based on semantic attribute attention, lead to It crosses and provides a kind of mode that attention reassigns to the attribute forecast of every image, measure again each when for the image classification The significance level of a attribute, and then achieve the effect that better zero sample learning.
Technical scheme is as follows:
A kind of zero sample learning method reassigning mechanism based on semantic attribute attention, comprising:
(1) neural network model for reassigning mechanism based on semantic attribute attention is established, is reflected comprising vision-semantic space Branch, vision-hidden layer space reflection branch and attention branch are penetrated, neural network model combines three branches to allow to carry out image When forward direction derives, semantic feature of the image in semantic attribute space, the hidden layer feature in hidden layer space are respectively obtained and in language The attention of adopted attribute space;
(2) weight between semantic feature is redistributed using the attention in semantic attribute space;
(3) using the image data set training neural network model with label;
(4) image to be tested is inputted, the semantic feature after calculating image weighting is similar to the semantic prototype of each unknown class Degree calculates the similarity of the hidden layer Feature prototype of hidden layer feature and unknown class, by two kinds of similarities be added to obtain test image with The similarity of each unknown class;
(5) according to it is all kinds of between similarity be ranked up, the class for choosing maximum similarity is pre- as the class of the image It surveys.
The zero sample learning method reassigned based on semantic attribute attention proposed by the invention, is a kind of modified version The algorithm that semantic attribute space and hidden layer spaces union are inferred.Semantic space and hidden layer compared to algorithm before, in this method Space combines even closer, provides the guidance of category information including 1) hidden layer space for semantic attribute space, produces neural network The raw attention to hold water;2) semantic attribute space provides initial methods for the prototype construction of unknown class in hidden layer space, from And reduce drawback brought by domain migration.Meanwhile models coupling reassigns the semantic attribute space and hidden layer space of attention Inferred, so that the stability of model prediction greatly improves.
In step (1), the vision-semantic space mapping branch and vision-hidden layer space reflection branch use VGG19 bone Frame network structure makees the Feature Mapping of different spaces as shared shallow-layer network using different full articulamentums respectively;
The attention branch uses the characteristic pattern of different layers in VGG19 back bone network the convolution kernel size of different parameters Make feature extraction for 3 single layer convolutional neural networks, and calculates the VGG19 characteristic pattern pair of different layers using the method for Fusion Features The attention in the semantic attribute space answered.
In step (1), obtain semantic feature of the image in semantic attribute space, the hidden layer feature in hidden layer space and Detailed process is as follows for the attention in semantic attribute space:
Image, which is extracted, using the depth convolutional neural networks of pre-training inputs xiDeep layer visual signature θi, use full connection The deep layer visual signature of image is mapped to semantic space and hidden layer space respectively by neural network, and the semantic space and hidden layer are empty Between calculation formula are as follows:
σi=FC2i)
Wherein,Indicate that image i is indicated in the vector of semantic space, σiIndicate that image i is indicated in the vector in hidden layer space, FC1Indicate the mapping function that semantic space is mapped to from visual space, FC2It indicates to be mapped to reflecting for hidden layer space from visual space Penetrate function;
The intermediate features figure about image i for choosing l layers in depth convolutional neural networks indicatesWith hidden layer space vector σi, calculation formula of the image i in the semantic attribute attention that space or depth perception is l Are as follows:
Wherein, WlAnd blIt is the parameter of single layer fully-connected network,It is vision when hidden layer vector indicates and depth is l Fusion Features expression, the calculation formula that the Fusion Features indicate are as follows:
Wherein, FsqIt is the transforming function transformation function of matrix, C × H × W size three-dimensional matrice is indicated to the two dimension for being converted to C × HW Matrix expression,Representing matrix is summed by channel,It is characteristic patternBy a series of convolution As a result, k indicate Fusion Features after port number, with semantic vector indicate and hidden layer vector expression length be consistent;Most After select to different number of plies l ∈ lBCarry out sum, calculation formula of the image i in the attention in semantic attribute space are as follows:
Wherein, pI, lThe semantic attribute attention for being l in space or depth perception for image i.
In step (2), the weight between semantic feature, calculation formula are redistributed using the attention in semantic attribute space Are as follows:
Wherein, diag (pi) be a k × k diagonal matrix, the value on diagonal line is pi
The detailed process of step (3) are as follows:
In (3-1) Data Preparation Process, former training dataset D is divided into the set of multiple triple compositions in advance
Wherein, for any triple WithIt is from same classDifferent images,Then be fromThe different class of imageImage;
In (3-2) training process, for each tripleNeural network model uses mixed damage Function L is lost to train neural network, the calculation formula of specific loss function L are as follows:
Wherein, LFIt is defined in the loss function in hidden layer space, LAIt is defined in the loss function in semantic attribute space;
Hidden layer space loss function comes to maximize between class distance simultaneously using triple loss function, minimize class it is interior away from From;The specific formula for calculation of hidden layer loss function are as follows:
The loss function in semantic attribute space using maximized based on cross entropy loss function semantic space classification Probability;The specific formula for calculation of semantic loss function are as follows:
Wherein, Y is the set of all trained classes,It is known class yiSemantic attribute prototype.
The specific steps of step (4) are as follows:
(4-1) is for input picture xi, using trained model prediction, its semantic vector is indicatedHidden layer vector table Show σiWith its semantic attribute attention pi
(4-2) is for any classification yu∈Yu, wherein YuIt indicates the unlapped classification of training, calculates separately image xiWith class Semantic prototypeWith class hidden layer prototypeCosine similarity in semantic attribute space and hidden layer space;Semantic attribute space Cosine similarity calculation formula are as follows:
The calculation formula of the cosine similarity in hidden layer space are as follows:
The cosine similarity of two spaces is summed to obtain image xiWith class yuBetween similarity, its calculation formula is:
The specific steps of step (5) are as follows:
Using nearest neighbor search algorithm, image x is calculatediIn class set YuOn class predictionIts calculation formula is:
The zero sample learning algorithm reassigned based on semantic attribute attention proposed by the invention, has zero sample learning All advantages, and can correctly distinguish it is some there are the difficult samples of semantic ambiguity, such as distinguished from pig and spot dog Out with spotted pig.In practice, zero more previous sample learning algorithm of proposed algorithm, the category on semantic space are found Property predicted value has much lower variance, and when this makes final searching classification, classification results are influenced by a variety of semantic attributes, Rather than by one or a small number of attribute forecast more outstanding is dominated.Joint of this algorithm in semantic space and hidden layer space is inferred On the basis of classification, so that the relationship of two spaces is even closer, avoids image and be correctly sorted in a feature space In another feature space the problem of mistake classification.
The present invention also provides a kind of zero sample learning systems that mechanism is reassigned based on semantic attribute attention, including meter It calculation machine memory, computer processor and is stored in the computer storage and can be held on the computer processor Capable computer program has with lower module in the computer storage:
Visual signature module captures the deep vision feature of input picture using depth convolutional neural networks;
Visual signature is mapped to semantic attribute space using full Connection Neural Network by vision-semantic mapping module;
Visual signature is mapped to hidden layer space using full Connection Neural Network by vision-hidden layer mapping block;
Semanteme pays attention to power module, utilizes the category information generative semantics space of the shallow-layer visual signature and hidden layer space of image Attribute attention;
Systematic searching module utilizes the semantic attribute space representation of image, the attention of hidden layer space representation and semantic space Power carries out image classification;
Classify generation module, for after category of model finishes, external output category result.
Compared with prior art, the invention has the following advantages:
1, semantic attribute attention reassignment algorithm proposed by the present invention, by using the mechanism of attention, so that each language There are competitive relations between adopted attribute forecast, so that classification results are determined by more semantic attributes, rather than are solely dependent upon portion Divide semantic attribute more outstanding, to avoid the mistake classification on some difficult samples with semantic ambiguity.
2, the domain migration problem when present invention can infer classification to avoid semantic space and hidden layer spaces union, therefore can be with Avoid the problem that classification results are intended to training covering class when general zero sample learning.
3, the present invention is by the way that experimental results demonstrate illustrate the model performance better than other reference line algorithms.It is demonstrate,proved from experiment The superiority of model is illustrated.
Detailed description of the invention
Fig. 1 is a kind of the whole of zero sample learning method that mechanism is reassigned based on semantic attribute attention of the embodiment of the present invention Body block schematic illustration;
Fig. 2 is the semantic operation schematic diagram for paying attention to power module of method in the embodiment of the present invention;
Fig. 3 is a kind of the whole of zero sample learning system that mechanism is reassigned based on semantic attribute attention of the embodiment of the present invention Body structural schematic diagram;
Fig. 4 is to be illustrated in the embodiment of the present invention using the prediction distribution of the different obtained semantic spaces of attention mechanism Figure.
Specific embodiment
The invention will be described in further detail with reference to the accompanying drawings and examples, it should be pointed out that reality as described below It applies example to be intended to convenient for the understanding of the present invention, and does not play any restriction effect to it.
As shown in Figure 1, master cast of the invention be divided into visual signature module and three branch modules to respectively correspond image defeated Three outputs entered, and these three branches are synchronized to the optimization process for being put into entire model.Specific step is as follows:
(a) visual signature module is study input picture x in zero sample learning training processiDeep vision feature θi, base Steps are as follows for this:
(a-1) network model parameter is initialized using the good large-scale neural network ResNet101 of pre-training.Input picture xi The center image x ' that is cut to 224 × 224 sizes firstiAs actually entering for network.
(a-2) take the feature vector of the last layer non-categorical layer of neural network as image x 'iDeep vision feature θi, the length of feature vector is denoted as V.
(b) vision-semantic mapping module provides the mapping in deep vision space to semantic space for zero sample learning process, Basic step is as follows:
(b-1) initialization model parameter.Initialize spatial mapping matrix W1∈Rk×VAnd b1∈Rk
(b-2) by visual signature θiIt is mapped to semantic space, obtains semantic space expressionSpecific formula for calculation are as follows:
(c) vision-hidden layer mapping block provides the mapping in deep vision space to hidden layer space for zero sample learning process, Basic step is as follows:
(c-1), similar with step (b-1), model is initialized, W is obtained2And b2
(c-2), similar with step (b-2), visual signature is mapped, specific formula for calculation are as follows:
σi=W2θi+b2
(d) semanteme-attention power module is zero sample using the part category information in hidden layer space and the shallow-layer visual information of image Weight is redistributed in the output of semantic space in this study, as shown in Fig. 2, basic step is as follows:
(d-1) the shallow-layer visual signature figure of certain layer l is chosenAnd hidden layer character representation σi∈Rk, As the input of semanteme-attention power module.Initialize convolutional neural networksParameter, the parameter W of fully-connected network FC3And b3
(d-2) shallow-layer visual signature figureBy a series of convolution transformsObtain characteristic patternWherein characteristic pattern size be H ' × W ', port number be k and with hidden layer feature σiLength keep one It causes.
(d-3) characteristic patternσ is indicated with the hidden layer of imagei∈RkBe added by channel, and will The matrix of generation is converted to column vector by channel, obtains hidden variable
(d-4) hidden variable obtains integrated images shallow-layer visual signature by full Connection Neural Network FCWith hidden layer feature σiThe space or depth perception in semantic space be l attention indicate pI, l, specific formula for calculation are as follows:
(d-5) four specific network depth l ∈ l are chosenB, repeat above-mentioned 1-4 step respectively, and by obtained attention Power adds up, and obtains image in total semantic attribute attention in semantic attribute space, its calculation formula is:
The training step of the zero sample learning method based on semantic attribute attention reassignment is as follows:
1. initializing training dataset D={ (xi, yi), wherein xiIndicate input picture, yi∈ Y indicates input picture Class label, Y indicates the set for the class that training set is covered, to every any sort ys∈ Y,It is such prototype in semantic space Vector.Data set is arranged to the set for being divided into multiple triplesWhereinWithCome From same classDifferent images,Then be fromThe different class of imageImage.
2. choosing a triple pairAs the input of network model, each image is obtained respectively in language The vector in adopted space, hidden layer space and semantic attribute attention indicates.
3. minimizing inter- object distance using based on triple loss function to maximize between class distance simultaneously.Hidden layer loses letter Several specific formula for calculation are as follows:
Use the probability maximized based on the loss function of cross entropy in semantic space classification.The tool of semantic loss function Body calculation formula are as follows:
4. repeating above-mentioned 2-3 step, the parameter of each module of training using gradient descent method.
5. using the average hidden layer space representation of training picture as the class hidden layer prototype of training covering class, calculation formula Are as follows:
Wherein NsIndicate class ysTraining samples number.S calculates the hidden layer vector expression that training does not cover class using ridge regressionIts specific formula for calculation are as follows:
Steps are as follows for the sample classification of the zero sample learning method based on semantic attribute attention reassignment:
1. for input picture xi, using trained model prediction, its semantic vector is indicatedHidden layer vector indicates σiWith its semantic attribute attention pi
2. for any classification yu∈Yu, wherein YuIt indicates the unlapped classification of training, calculates separately image xiWith class semanteme PrototypeWith class hidden layer prototypeIn the cosine similarity of semantic space and hidden layer space.The cosine similarity of semantic space Calculation formula are as follows:
The calculation formula of the cosine similarity in hidden layer space are as follows:
The cosine similarity of two spaces is summed to obtain image xiWith class yuBetween similarity, its calculation formula is:
3. using nearest neighbor search algorithm, image x is calculatediIn class set YuOn class predictionIts calculation formula is:
As shown in figure 3, a kind of zero sample classification system reassigned based on semantic attribute attention, is divided into six big moulds Block, is visual signature module respectively, vision-semantic mapping module, vision-hidden layer mapping block, semanteme-attention power module, point Class retrieval module, and classification generation module.
The above method is applied in the following example below, it is specific in embodiment to embody technical effect of the invention Step repeats no more.
The present embodiment is had and zero sample of other current forefronts in three large sizes public data collection AwA2, CUB and SUN Learning method compares.AwA2 is the medium scale data set an of coarseness, from 50 there are 85 users to define and belongs to 37322 images of the animal category of property.CUB is a fine-grained data set, by 11788 from 200 kinds of different birds A image composition, has 312 user-defined attributes.SUN is another fine-grained data set, including not from 717 With 14340 images of scene, 102 user-defined attributes are provided.Data set is divided into two parts: training set and test Collection, on different data sets by different divisions.On AwA2 data set, 40 animal classes are as training set, 10 animal classes As test set;Similar, training set of 150 classes as CUB, test set of 50 classes as CUB, 645 classes are as SUN Training set, test set of 72 classes as SUN.The judging quota of the present embodiment is that class is averaged recognition accuracy, is compared in total Zero specimen discerning algorithm of 5 current mainstreams, the results are shown in Table 1 for overall contrast.
Table 1
As it can be seen from table 1 the zero sample learning frame proposed by the present invention reassigned based on semantic attribute attention, Optimal effectiveness is obtained under major judging quota, sufficiently illustrates the superiority of inventive algorithm.
The algorithm proposed in order to further illustrate the present invention really by inhibiting part relatively prominent in semantic space Predicted value out, the present invention are compared simultaneously " without using attention mechanism ", " use the attention machine based on sigmoid function Experimental comparison of " the attention mechanism for using softmax function " that system " and the present invention use on CUB data set, experiment knot Fruit is as shown in table 2.
Table 2
Method Class Average Accuracy (%) Semantic forecast value variance × 10-3
w/o Attention 62.1 2.48
w/Sigmoid Attention 73.5 1.75
w/Softmax Attention 81.1 0.86
From on table 2 as can be seen that the attention mechanism based on softmax function has reached optimal experiment knot in experiment Fruit, and it was found that with the reduction of semantic space predicted value variance, the inhibitory effect of outburst prediction value is better, the effect of model It is better, absolutely prove validity of the attention mechanism in zero sample learning.
In addition to this, the kernel function estimation of semantic forecast value as shown in Figure 4 also reflect attention mechanism will be semantic empty Between prediction distribution be limited in a particular range, absolutely prove algorithm proposed by the invention by redistributing semanteme The attribute weight in space has reached the inhibitory effect to exception semantics predicted value.
Technical solution of the present invention and beneficial effect is described in detail in embodiment described above, it should be understood that Above is only a specific embodiment of the present invention, it is not intended to restrict the invention, it is all to be done in spirit of the invention Any modification, supplementary, and equivalent replacement, should all be included in the protection scope of the present invention.

Claims (8)

1. a kind of zero sample learning method for reassigning mechanism based on semantic attribute attention characterized by comprising
(1) neural network model that mechanism is reassigned based on semantic attribute attention is established, the neural network model includes view Feel-semantic attribute space reflection branch, vision-hidden layer space reflection branch and attention branch make image carry out network When forward direction derives, semantic feature of the image in semantic attribute space, the hidden layer feature in hidden layer space are respectively obtained and in language The attention of adopted attribute space;
(2) weight between semantic feature is redistributed using the attention in semantic attribute space;
(3) using the image data set training neural network model with label;
(4) image to be tested is inputted, the similarity of the semantic prototype of semantic feature and each unknown class after calculating image weighting, meter The similarity for calculating the hidden layer Feature prototype of hidden layer feature and unknown class, by two kinds of similarities be added to obtain test image with it is each unknown The similarity of class;
(5) according to it is all kinds of between similarity be ranked up, the class for choosing the class of maximum similarity as the image is predicted.
2. the zero sample learning method according to claim 1 for reassigning mechanism based on semantic attribute attention, feature It is, in step (1), the vision-semantic space mapping branch and vision-hidden layer space reflection branch use VGG19 skeleton Network structure makees the Feature Mapping of different spaces as shared shallow-layer network using different full articulamentums respectively;
The attention branch is 3 using the convolution kernel size of different parameters to the characteristic pattern of different layers in VGG19 back bone network Single layer convolutional neural networks make feature extraction, and it is corresponding using the VGG19 characteristic pattern that the method for Fusion Features calculates different layers Semantic attribute space attention.
3. the zero sample learning method according to claim 1 for reassigning mechanism based on semantic attribute attention, feature It is, in step (1), obtains semantic feature of the image in semantic attribute space, the hidden layer feature in hidden layer space and in language Detailed process is as follows for the attention of adopted attribute space:
Image, which is extracted, using the depth convolutional neural networks of pre-training inputs xiDeep layer visual signature θi, use full connection nerve The deep layer visual signature of image is mapped to semantic space and hidden layer space respectively by network, the semantic space and hidden layer space Calculation formula are as follows:
Wherein,Indicate that image i is indicated in the vector of semantic space, σiIndicate that image i is indicated in the vector in hidden layer space, FC1Table Show the mapping function that semantic space is mapped to from visual space, FC2Indicate the mapping letter that hidden layer space is mapped to from visual space Number;
The intermediate features figure about image i for choosing l layers in depth convolutional neural networks indicatesWith Hidden layer space vector σi, calculation formula of the image i in the semantic attribute attention that space or depth perception is l are as follows:
Wherein, WlAnd blIt is the parameter of single layer fully-connected network,It is visual signature when hidden layer vector indicates and depth is l Fusion expression, the calculation formula that the Fusion Features indicate are as follows:
Wherein, FsqIt is the transforming function transformation function of matrix, C × H × W size three-dimensional matrice is indicated to the two-dimensional matrix for being converted to C × HW It indicates,Representing matrix is summed by channel,It is characteristic patternBy a series of knot of convolution Fruit, k indicate the port number after Fusion Features, indicate with semantic vector and the length of hidden layer vector expression is consistent;Finally select It selects to different number of plies l ∈ lBCarry out sum, calculation formula of the image i in the attention in semantic attribute space are as follows:
Wherein, pI, lThe semantic attribute attention for being l in space or depth perception for image i.
4. the zero sample learning method according to claim 1 for reassigning mechanism based on semantic attribute attention, feature It is, in step (2), redistributes the weight between semantic feature, calculation formula using the attention in semantic attribute space are as follows:
Wherein, diag (pi) be a k × k diagonal matrix, the value on diagonal line is pi
5. the zero sample learning method according to claim 1 for reassigning mechanism based on semantic attribute attention, feature It is, the detailed process of step (3) are as follows:
In (3-1) Data Preparation Process, former training dataset D is divided into the set of multiple triple compositions in advance
Wherein, for any tripleWithIt is from same classDifferent images,Then Be fromThe different class of imageImage;
In (3-2) training process, for each tripleNeural network model uses mixed loss letter L is counted to train neural network, the calculation formula of specific loss function L are as follows:
Wherein, LFIt is defined in the loss function in hidden layer space, LAIt is defined in the loss function in semantic attribute space;
Hidden layer space loss function carrys out using triple loss function while maximizing between class distance, minimizes inter- object distance;It is hidden The specific formula for calculation of layer loss function are as follows:
The loss function in semantic attribute space uses the probability maximized based on cross entropy loss function in semantic space classification; The specific formula for calculation of semantic loss function are as follows:
Wherein, Y is the set of all trained classes,It is known class yiSemantic attribute prototype.
6. the zero sample learning method according to claim 1 for reassigning mechanism based on semantic attribute attention, feature It is, the specific steps of step (4) are as follows:
(4-1) is for input picture xi, using trained model prediction, its semantic vector is indicatedHidden layer vector indicates σi With its semantic attribute attention pi
(4-2) is for any classification yu∈Yu, wherein YuIt indicates the unlapped classification of training, calculates separately image xiIt is semantic former with class TypeWith class hidden layer prototypeCosine similarity in semantic attribute space and hidden layer space;The cosine in semantic attribute space The calculation formula of similarity are as follows:
The calculation formula of the cosine similarity in hidden layer space are as follows:
The cosine similarity of two spaces is summed to obtain image xiWith class yuBetween similarity, its calculation formula is:
7. the zero sample learning method according to claim 6 for reassigning mechanism based on semantic attribute attention, feature It is, the specific steps of step (5) are as follows:
Using nearest neighbor search algorithm, image x is calculatediIn class set YuOn class predictionIts calculation formula is:
8. a kind of zero sample learning system for reassigning mechanism based on semantic attribute attention, including computer storage, calculating Machine processor and it is stored in the computer program that can be executed in the computer storage and on the computer processor, It is characterized in that, having in the computer storage with lower module:
Visual signature module captures the deep vision feature of input picture using depth convolutional neural networks;
Visual signature is mapped to semantic attribute space using full Connection Neural Network by vision-semantic mapping module;
Visual signature is mapped to hidden layer space using full Connection Neural Network by vision-hidden layer mapping block;
Semanteme pays attention to power module, utilizes the attribute in the category information generative semantics space of the shallow-layer visual signature and hidden layer space of image Attention;
Systematic searching module, using the semantic attribute space representation of image, the attention of hidden layer space representation and semantic space into Row image classification;
Classify generation module, for after category of model finishes, external output category result.
CN201910335801.6A 2019-04-24 2019-04-24 Zero sample learning method and system based on semantic attribute attention redistribution mechanism Active CN110163258B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910335801.6A CN110163258B (en) 2019-04-24 2019-04-24 Zero sample learning method and system based on semantic attribute attention redistribution mechanism

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910335801.6A CN110163258B (en) 2019-04-24 2019-04-24 Zero sample learning method and system based on semantic attribute attention redistribution mechanism

Publications (2)

Publication Number Publication Date
CN110163258A true CN110163258A (en) 2019-08-23
CN110163258B CN110163258B (en) 2021-04-09

Family

ID=67639900

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910335801.6A Active CN110163258B (en) 2019-04-24 2019-04-24 Zero sample learning method and system based on semantic attribute attention redistribution mechanism

Country Status (1)

Country Link
CN (1) CN110163258B (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110866140A (en) * 2019-11-26 2020-03-06 腾讯科技(深圳)有限公司 Image feature extraction model training method, image searching method and computer equipment
CN111222471A (en) * 2020-01-09 2020-06-02 中国科学技术大学 Zero sample training and related classification method based on self-supervision domain perception network
CN111428733A (en) * 2020-03-12 2020-07-17 山东大学 Zero sample target detection method and system based on semantic feature space conversion
CN111461025A (en) * 2020-04-02 2020-07-28 同济大学 Signal identification method for self-evolving zero-sample learning
CN111738313A (en) * 2020-06-08 2020-10-02 大连理工大学 Zero sample learning algorithm based on multi-network cooperation
CN112100380A (en) * 2020-09-16 2020-12-18 浙江大学 Generation type zero sample prediction method based on knowledge graph
CN112257808A (en) * 2020-11-02 2021-01-22 郑州大学 Integrated collaborative training method and device for zero sample classification and terminal equipment
CN112633382A (en) * 2020-12-25 2021-04-09 浙江大学 Mutual-neighbor-based few-sample image classification method and system
CN112686318A (en) * 2020-12-31 2021-04-20 广东石油化工学院 Zero sample learning mechanism based on spherical embedding, spherical alignment and spherical calibration
CN113077427A (en) * 2021-03-29 2021-07-06 北京深睿博联科技有限责任公司 Generation method and device of category prediction model
CN113326892A (en) * 2021-06-22 2021-08-31 浙江大学 Relation network-based few-sample image classification method
CN113343941A (en) * 2021-07-20 2021-09-03 中国人民大学 Zero sample action identification method and system based on mutual information similarity
CN113435531A (en) * 2021-07-07 2021-09-24 中国人民解放军国防科技大学 Zero sample image classification method and system, electronic equipment and storage medium
CN113627470A (en) * 2021-07-01 2021-11-09 汕头大学 Zero-learning-based unknown event classification method for optical fiber early warning system
CN113642621A (en) * 2021-08-03 2021-11-12 南京邮电大学 Zero sample image classification method based on generation countermeasure network
CN114627312A (en) * 2022-05-17 2022-06-14 中国科学技术大学 Zero sample image classification method, system, equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679556A (en) * 2017-09-18 2018-02-09 天津大学 The zero sample image sorting technique based on variation autocoder
CN108846413A (en) * 2018-05-21 2018-11-20 复旦大学 A kind of zero sample learning method based on global semantic congruence network
CN109447115A (en) * 2018-09-25 2019-03-08 天津大学 Zero sample classification method of fine granularity based on multilayer semanteme supervised attention model

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679556A (en) * 2017-09-18 2018-02-09 天津大学 The zero sample image sorting technique based on variation autocoder
CN108846413A (en) * 2018-05-21 2018-11-20 复旦大学 A kind of zero sample learning method based on global semantic congruence network
CN109447115A (en) * 2018-09-25 2019-03-08 天津大学 Zero sample classification method of fine granularity based on multilayer semanteme supervised attention model

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110866140A (en) * 2019-11-26 2020-03-06 腾讯科技(深圳)有限公司 Image feature extraction model training method, image searching method and computer equipment
CN110866140B (en) * 2019-11-26 2024-02-02 腾讯科技(深圳)有限公司 Image feature extraction model training method, image searching method and computer equipment
CN111222471A (en) * 2020-01-09 2020-06-02 中国科学技术大学 Zero sample training and related classification method based on self-supervision domain perception network
CN111222471B (en) * 2020-01-09 2022-07-15 中国科学技术大学 Zero sample training and related classification method based on self-supervision domain perception network
CN111428733A (en) * 2020-03-12 2020-07-17 山东大学 Zero sample target detection method and system based on semantic feature space conversion
CN111428733B (en) * 2020-03-12 2023-05-23 山东大学 Zero sample target detection method and system based on semantic feature space conversion
CN111461025A (en) * 2020-04-02 2020-07-28 同济大学 Signal identification method for self-evolving zero-sample learning
CN111461025B (en) * 2020-04-02 2022-07-05 同济大学 Signal identification method for self-evolving zero-sample learning
CN111738313A (en) * 2020-06-08 2020-10-02 大连理工大学 Zero sample learning algorithm based on multi-network cooperation
CN112100380A (en) * 2020-09-16 2020-12-18 浙江大学 Generation type zero sample prediction method based on knowledge graph
CN112100380B (en) * 2020-09-16 2022-07-12 浙江大学 Generation type zero sample prediction method based on knowledge graph
CN112257808A (en) * 2020-11-02 2021-01-22 郑州大学 Integrated collaborative training method and device for zero sample classification and terminal equipment
CN112257808B (en) * 2020-11-02 2022-11-11 郑州大学 Integrated collaborative training method and device for zero sample classification and terminal equipment
CN112633382A (en) * 2020-12-25 2021-04-09 浙江大学 Mutual-neighbor-based few-sample image classification method and system
CN112633382B (en) * 2020-12-25 2024-02-13 浙江大学 Method and system for classifying few sample images based on mutual neighbor
CN112686318A (en) * 2020-12-31 2021-04-20 广东石油化工学院 Zero sample learning mechanism based on spherical embedding, spherical alignment and spherical calibration
CN112686318B (en) * 2020-12-31 2023-08-29 广东石油化工学院 Zero sample learning mechanism based on sphere embedding, sphere alignment and sphere calibration
CN113077427A (en) * 2021-03-29 2021-07-06 北京深睿博联科技有限责任公司 Generation method and device of category prediction model
CN113077427B (en) * 2021-03-29 2023-04-25 北京深睿博联科技有限责任公司 Method and device for generating class prediction model
CN113326892A (en) * 2021-06-22 2021-08-31 浙江大学 Relation network-based few-sample image classification method
CN113627470A (en) * 2021-07-01 2021-11-09 汕头大学 Zero-learning-based unknown event classification method for optical fiber early warning system
CN113627470B (en) * 2021-07-01 2023-09-05 汕头大学 Zero-order learning-based unknown event classification method for optical fiber early warning system
CN113435531A (en) * 2021-07-07 2021-09-24 中国人民解放军国防科技大学 Zero sample image classification method and system, electronic equipment and storage medium
CN113435531B (en) * 2021-07-07 2022-06-21 中国人民解放军国防科技大学 Zero sample image classification method and system, electronic equipment and storage medium
CN113343941A (en) * 2021-07-20 2021-09-03 中国人民大学 Zero sample action identification method and system based on mutual information similarity
CN113343941B (en) * 2021-07-20 2023-07-25 中国人民大学 Zero sample action recognition method and system based on mutual information similarity
CN113642621A (en) * 2021-08-03 2021-11-12 南京邮电大学 Zero sample image classification method based on generation countermeasure network
CN114627312B (en) * 2022-05-17 2022-09-06 中国科学技术大学 Zero sample image classification method, system, equipment and storage medium
CN114627312A (en) * 2022-05-17 2022-06-14 中国科学技术大学 Zero sample image classification method, system, equipment and storage medium

Also Published As

Publication number Publication date
CN110163258B (en) 2021-04-09

Similar Documents

Publication Publication Date Title
CN110163258A (en) A kind of zero sample learning method and system reassigning mechanism based on semantic attribute attention
He et al. An end-to-end steel surface defect detection approach via fusing multiple hierarchical features
Acharya et al. TallyQA: Answering complex counting questions
Das et al. Automatic clustering using an improved differential evolution algorithm
Xu et al. 3D attention-driven depth acquisition for object identification
CN107066559A (en) A kind of method for searching three-dimension model based on deep learning
Bespalov et al. Scale-space representation of 3d models and topological matching
Xiong et al. ASK: Adaptively selecting key local features for RGB-D scene recognition
CN112949740B (en) Small sample image classification method based on multilevel measurement
CN110084211B (en) Action recognition method
CN104751463B (en) A kind of threedimensional model optimal viewing angle choosing method based on sketch outline feature
Naqvi et al. Feature quality-based dynamic feature selection for improving salient object detection
CN106250918A (en) A kind of mixed Gauss model matching process based on the soil-shifting distance improved
CN110163130B (en) Feature pre-alignment random forest classification system and method for gesture recognition
Xia et al. Evaluation of saccadic scanpath prediction: Subjective assessment database and recurrent neural network based metric
Andreetto et al. Unsupervised learning of categorical segments in image collections
CN108805280A (en) A kind of method and apparatus of image retrieval
CN108596186B (en) Three-dimensional model retrieval method
CN113032613B (en) Three-dimensional model retrieval method based on interactive attention convolution neural network
Ngo et al. Similarity Shape Based on Skeleton Graph Matching.
CN114821632A (en) Method for re-identifying blocked pedestrians
Kusrini et al. Automatic Mango Leaf and Trunk Detection as Supporting Tool of Mango Pest Identifier (MPI)
Sun et al. Recursive templates segmentation and exemplars matching for human parsing
Hajihashemi et al. Human activity recognition in videos based on a Two Levels K-means and Hierarchical Codebooks
Bouksim et al. New approach for 3D Mesh Retrieval using artificial neural network and histogram of features

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant