CN105184298A - Image classification method through fast and locality-constrained low-rank coding process - Google Patents

Image classification method through fast and locality-constrained low-rank coding process Download PDF

Info

Publication number
CN105184298A
CN105184298A CN201510534124.2A CN201510534124A CN105184298A CN 105184298 A CN105184298 A CN 105184298A CN 201510534124 A CN201510534124 A CN 201510534124A CN 105184298 A CN105184298 A CN 105184298A
Authority
CN
China
Prior art keywords
vector
image
dictionary
dist
matrix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510534124.2A
Other languages
Chinese (zh)
Other versions
CN105184298B (en
Inventor
范敏
王芬
杜思远
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University
Original Assignee
Chongqing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University filed Critical Chongqing University
Priority to CN201510534124.2A priority Critical patent/CN105184298B/en
Publication of CN105184298A publication Critical patent/CN105184298A/en
Application granted granted Critical
Publication of CN105184298B publication Critical patent/CN105184298B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an image classification method through the fast and locality-constrained low-rank coding process. The method comprises a dictionary learning part, a training part and a testing part. During the dictionary learning part, a corresponding dictionary is obtained based on SIFT features extracted out of all data sets according to the k-means clustering method or other dictionary learning methods. During the training part, one part of images in each data set are selected as training images, and then the corresponding low-rank representation corresponding to each data set is learned based on the tags of the images. Through the SPM quantization process and the SVM classifier-based training process, category labels and a kernel function corresponding to a feature representation matrix are obtained. During the testing part, all remaining images in each data set are adopted as to-be-classified test images. For any unclassified image, the SIFT features of the image are extracted firstly, and then a low-rank coding matrix thereof in the dictionary is found out. Through the SPM quantization process, a final feature representation matrix is obtained. The final feature representation matrix is input into an SVM classifier obtained during the training part, so that the category of the image can be known.

Description

A kind of image classification method of quick local restriction low-rank coding
Technical field
The present invention relates to a kind of method utilizing quick local restriction low-rank to encode and carry out Images Classification, also precision is higher fast for the method, is a kind of sorting technique of data-driven of image content-based.
Background technology
Images Classification belongs to an important research direction of computer vision, and it comprises the pre-service of image, the expression of feature and the design of sorter, and wherein feature representation comprises the coding of the extraction of characteristics of image, the dimensionality reduction of feature and feature.Feature extraction mainly by attributes such as the space distributions of various content presentation color out in image, brightness, texture, shape and pixel to iamge description.Feature expression refers to carry out adding up (vector quantization), coding or additive method to form the last feature of piece image in the most basic feature base, can have better performance under normal circumstances than original essential characteristic; Good feature representation can promote the performance of Images Classification and identification.
The Measures compare represented for feature interpretation is in recent years many, there is visual word bag model (bagofwords, BOW), based on the spatial pyramid model (SparseCodingSpatialPyramidMatching of sparse coding, ScSPM), local restriction uniform enconding (Locality-constrainedLinearCoding, LLC), image classification method (the StructuredLow-rankRepresentation that structure low-rank represents, SLRR) and low-rank sparse decompose image classification method (Low-rankSparseCoding, LRSC), these methods overcome some difficult problems to a certain extent, but still all Shortcomings.
Visual word bag model (bagofwords, BOW), although be easy to build, have ignored similar image and has certain space structure similarity.Spatial pyramid model (SparseCodingSpatialPyramidMatching, ScSPM) based on sparse coding have ignored the correlativity between local description, and to the change of feature and noise-sensitive, and calculated amount is very large, consumption internal memory.Local restriction uniform enconding (Locality-constrainedLinearCoding, LLC), this locality result in openness to a certain extent, relative to sparse coding, the operand of algorithm is reduced greatly, speed improves, but LLC have ignored the spatial information of the one-piece construction feature between feature.Image classification method (the StructuredLow-rankRepresentation that structure low-rank represents, SLRR), there is good effect to recognition of face, be particularly useful for having the Images Classification of critical noisy (as block, lighting angle change etc.).Image classification method (Low-rankSparseCoding) (LRSC) nicety of grading that low-rank sparse is decomposed obtains certain raising, but have ignored local spatial information; Wherein sparse part is still very sensitive to coding.Based on the image classification method (FastLow-rankRepresentationbasedSpatialPyramidMatching that the quick low-rank of spatial pyramid represents, LrrSPM), the classification speed of the method is 5 ~ 16 times of ScSPM, but nicety of grading is lower than ScSPM and LLC method.
The coded system that current image classification method mostly adopts sparse coding and low-rank to represent, but in view of sparse coding to the change of image and noise more responsive, very consume internal memory; The coded system adopting low-rank to represent obtains a balance between precision and computation complexity.
Summary of the invention
In view of the defect of current sorting technique in classifying quality and classification speed, the present invention proposes a kind of image classification method of quick local restriction low-rank coding, and the method comprises dictionary learning part, training part and part of detecting three parts.Dictionary learning means suitable, to the SIFT feature of image zooming-out images all in image library, obtains dictionary through K-means cluster; Training part is first to the image zooming-out SIFT feature in training set, then try to achieve the feature coding of SIFT feature under dictionary to represent, adopt the low-rank coding method of local restriction, consider global structure consistance and the local spatial simlanty of image simultaneously, obtain feature coding matrix, behind spatial pyramid coupling core, pond, obtain final character representation matrix, adopt SVM method to train character representation matrix and class mark, set up disaggregated model.First part of detecting carries out feature extraction to any image, then encoder matrix under its dictionary is tried to achieve, after spatial pyramid coupling core, pond obtain character representation matrix, the SVM classifier obtained is trained in input, obtains the classification results belonging to this image.
The technical scheme adopted for realizing the object of the invention is such, and a kind of image classification method of quick local restriction low-rank coding, comprises dictionary learning part, training part and part of detecting.
Described dictionary learning part steps process is as follows:
1) utilize SIFT feature extracting method to carry out local shape factor to all images of image library, obtain input feature value matrix X=[x 1, x 2..., x n] ∈ R m × n, wherein x i∈ R nbe i-th component vector of eigenvectors matrix, wherein m is the dimension of vector, and n is vectorial number.
2) utilize K-means clustering or other dictionary learning methods, clustering processing is carried out to the eigenvectors matrix X of all image zooming-out, complete dictionary D=[d must be 1, d 2..., d k] ∈ R m × k, wherein m is the dimension of dictionary, and k is the number of dictionary base, and meet dictionary and cross completeness m < < k, k is generally 256,1024,4096 etc.
Described training department step by step process is as follows:
I) from image library, selected part image, as training image, utilizes SIFT feature extracting method to carry out local shape factor to all training images, obtains input feature value matrix X'=[x 1', x 2' ..., x n'] ∈ R m × n.Wherein x i' ∈ R nbe i-th component vector of eigenvectors matrix, wherein m is the dimension of vector, and n is vectorial number.
Ii) each proper vector x is calculated iand the Euclidean distance p between dictionary vector i, computing formula is as follows:
p i = exp ( d i s t ( x i &prime; , D ) &sigma; )
Dist (x i', D)=[dist (x i', d 1) ..., dist (x i', d k)] t, dist (x i', d i) be proper vector x i' d vectorial with each dictionary ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
Iii) calculate the character representation vector Z of local restriction low-rank coding, computing formula is:
Z = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i T p i ) - 1 D T X &prime;
Wherein λ 1, λ 2> 0 is weight; p ifor the Euclidean distance p that above-mentioned steps ii calculates i.
Iv) adopt spatial pyramid matching process to process to the character representation vector Z obtained, Z is divided into 2 l× 2 lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer icarry out maximum pond, i.e. z i=max{|z 1i|, | z 2i| ..., | z ji|, obtain the i-th vectorial z of final low-rank coding i, wherein z jii-th element in jth block, j=2 l× 2 l, obtain last character representation matrix Z' by this process.
V) character representation matrix Z' is inputted SVM classifier, according to every class image category label and its characteristic of correspondence representing matrix Z', study obtains the kernel function of SVM classifier.
Described part of detecting step process is as follows:
I) to all image zooming-out SIFT feature remaining in image library, input feature value matrix X*=[x is obtained 1*, x 1* ..., x n*] ∈ R m × n, wherein x i* ∈ R ni-th component vector of eigenvectors matrix.
II) each proper vector x is calculated i* the Euclidean distance p and between dictionary vector i, computing formula is as follows:
p i * = exp ( d i s t ( x i * , D ) &sigma; )
Dist (x i*, D)=[dist (x i*, d 1) ..., dist (x i*, d k)] t, dist (x i*, d i) be proper vector x i* with each dictionary vector d ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
III) calculate the character representation vector Z * of local restriction low-rank coding, computing formula is:
Z * = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i * T p i * ) - 1 D T X *
Wherein λ 1, λ 2> 0 is weight; p i* be Euclidean distance that above-mentioned steps II calculates.
IV) adopt spatial pyramid matching process, character representation vector Z * is divided into 2 l× 2 lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer i* carry out maximum pond, keep character representation vector to have certain robustness, i.e. z i*=max{|z 1i* |, | z 2i* | ..., | z ji* | }, obtain the i-th vectorial z of final low-rank coding i*, wherein z ji* be i-th element in jth block, j=2 l× 2 l, obtain last character representation matrix Z*' like this.
V) character representation matrix Z*' is inputted training department and divide the SVM classifier obtained, obtain the class label belonging to image to be classified.
VI) the statistical test classified image number percent of correctly classifying, obtains the nicety of grading that every class testing image is final.
Further, in the image classification method of described a kind of quick local restriction low-rank coding, the low-rank that in training part, in step I ii and part of detecting, Step II I calculates feature coding represents vector.
The coding method that existing low-rank represents is that the low-rank finding proper vector in the error range allowed represents, namely
m i n Z , E | | Z | | * + &lambda; 1 | | E | | p
s.t.X=DZ+E
Wherein μ 1, μ 2> 0 is error penalty term, λ 1> 0 is weights, and Z is the expression of feature vector, X under dictionary D.Wherein ‖ Z ‖ *nuclear norm, ‖ E ‖ pthe l of constraint sparse coding item 1norm or retrain a certain class interference l 2-1norm.
Existing low-rank optimized algorithm commonly adopts Lagrange multiplier to optimize AugmentedLagrangeMultiplealgorithm (ALM), and optimizing expression is as follows:
min Z , D | | Z | | * + &lambda; | | D - D 0 | | F + t r &lsqb; Y T ( D - D 0 - E 1 ) &rsqb; + t r &lsqb; Y T ( X - D Z - E 2 ) &rsqb; + &mu; 1 2 | | D - D 0 - E 1 | | F 2 + &mu; 2 2 | | X - D Z - E 2 | | F 2 - - - ( 1 )
Wherein ‖ ‖ ff norm (Frobeniusnorm), λ 1> 0 is weights, Y 1, Y 2lagrange multiplier, μ 1, μ 2> 0 is error penalty term.
Above formula 1 can adopt typical exactALM or inexactALM method, adopt alternative manner step by step respectively regeneration characteristics represent vector Z and initial dictionary D 0, the computation complexity of this method is O (mn 2), n is the number of proper vector.This calculating comes that is very large, and is a kind of method of iteration optimization of off-line, cannot obtain classification results (can not asking character representation if do not belonged to such other image) to the input picture of increment type.Meanwhile, existing low-rank constraint only considered the global structure consistance of characteristics of image, have ignored the local space similarity of feature, comprehensively can not represent the related information between characteristics of image.
The low-rank coding method step calculating local restriction in this method is as follows:
A) when dictionary error is zero, E 1=D-D 0=0, F norm ‖ ‖ fnuclear norm ‖ ‖ can be replaced *.Above-mentioned optimizing expression 1 becomes
m i n Z | | Z | | F + &lambda; 1 / 2 | | X - D Z | | F 2 - - - ( 2 )
B) combined coding mode is taked in low-rank constraint, on the conforming basis of characteristics of image global structure, makes up the local space similarity of feature, adds local restriction item above-mentioned expression formula 2 becomes:
Wherein λ 1, λ 2> 0 is unequal weight, and symbol ⊙ represents that the corresponding element of two vectors is multiplied, p ifor proper vector x iwith the Euclidean distance p of dictionary vector D i
p i = exp ( d i s t ( x i , D ) &sigma; )
Dist (x i, D) and=[dist (x i, d 1) ..., dist (x i, d k)] t, dist (x i, d i) be proper vector x iwith each dictionary vector d ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
C) differentiate is carried out to above-mentioned expression formula 3
Finally calculate character representation vector: Z = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i T p i ) - 1 D T X , Wherein I is the vector of unit length of k × k.
In differentiate process, take threshold value constraint item by character representation vector z iconstrain within 0.98.
Technique effect of the present invention is mathematical, the present invention adopts a kind of low-rank coding specification method of local restriction fast, consider global structure consistance and the local spatial simlanty of characteristics of image simultaneously, better can not only represent the related information between characteristics of image, obtain effective nicety of grading, and calculated amount reduces, classification speed is a bit larger tham LLC, is only 19.4 ~ 25% of ScSPM.
Accompanying drawing explanation
Fig. 1 is Images Classification main-process stream;
Fig. 2 is local restriction low-rank coding (FLCLR) process flow diagram;
Fig. 3 is the example images of classification in the middle part of table 1;
Fig. 4 is the example images of classification in the middle part of table 2;
Fig. 5 is the example images of classification in the middle part of table 4;
In figure: 1-feature extraction, 2-feature coding, 3-pond process.
Embodiment
Below in conjunction with embodiment, the invention will be further described, but should not be construed the above-mentioned subject area of the present invention and be only limitted to following embodiment.Without departing from the idea case in the present invention described above, according to ordinary skill knowledge and customary means, make various replacement and change, all should be included in protection scope of the present invention.
Below implement to choose Images Classification conventional authoritative image library Scene13 and Caltech101 as class test data set.Scene13 is the most frequently used scene classification data set, has 13 class images, comprises suburb (CALsuburb), seashore (MITcoast), and totally 4485 width such as forest (MITforest), every class comprises 200 to 400 pictures.Caltech101 image library comprises 101 classes (as animal, flower, face etc.) and a background classes image, totally 9144 width.Each class comprises 31 to 800 width images, and has difference in larger class, is the most frequently used database of Images Classification.
As shown in Figure 1, whole implementation step is divided into three parts, and Part I is dictionary learning part, extracts all SIFT feature according to data set, uses the dictionary learning methods such as k-menas clustering to obtain corresponding dictionary.Part II is training part, chooses the parts of images in the every class of each data set, and wherein Scene13 gets 100 at random as training image, and Caltech101 gets 30 at random as training image.Then the low-rank learning each classification corresponding for the label of these training images represents, quantizes to obtain obtaining class label and kernel function corresponding to character representation through svm classifier training through SPM.Part III is part of detecting, in each database, remaining all images are as test pattern to be sorted, the image of classification is not known for any one, first SIFT feature is extracted, then its low-rank coding vector under dictionary is tried to achieve, quantize to obtain final character representation matrix through SPM, inputted the SVM classifier that training part gained arrives, and then obtain the classification belonging to this image.
Pre-service illustrates: image size is adjusted to maximum 300 × 300 pixels, and the grid of the intensive Bian sample of feature extraction is 16 × 16 pixels, and its step-length is 4 pixels.
Dictionary learning part:
1) to image zooming-out SIFT local features all in image library, input feature value matrix X=[x is obtained 1, x 2..., x n] ∈ R m × n, wherein x i∈ R nbe i-th local feature vectors of image, wherein m is the dimension of vector, and n is vectorial number.
2) utilize K-means clustering or other dictionary learning methods, clustering processing is carried out to the proper vector of all image zooming-out, complete dictionary D=[d must be 1, d 2..., d k] ∈ R m × k, wherein k is the dimension of dictionary, and m is the number of dictionary base, meets dictionary and crosses completeness m < < k, and wherein k is that the dimension of dictionary gets 1024.
Training part:
I) SIFT feature is extracted to the training image chosen, obtain input feature value matrix X'=[x' 1, x' 2..., x' n] ∈ R m × n, wherein x i' ∈ R ni-th component vector of eigenvectors matrix.
Ii) each proper vector x in calculation training image i' and dictionary vector between Euclidean distance p i, computing formula is as follows:
p i = exp ( d i s t ( x i &prime; , D ) &sigma; )
Dist (x i', D)=[dist (x i', d 1) ..., dist (x i', d k)] t, dist (x i', d i) be proper vector x i' d vectorial with each dictionary ibetween Euclidean distance; σ is the weight of adjustment local restriction speed, gets 100.
Iii) calculate the character representation vector Z of local restriction low-rank coding, computing formula is:
Z = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i T p i ) - 1 D T X &prime;
Wherein I is unit matrix, and dimension is consistent with Z, λ 1, λ 2> 0 is weight, λ 1, λ 20.7 is respectively, 0.0001 according to the empirical value value of document Wang and Peng.P ifor the Euclidean distance calculated in above-mentioned steps 2.
Step I i and step I ii seeks to obtain character representation vector z, the i.e. X=DZ of eigenvectors matrix X under dictionary D ,wherein Z=[z 1, z 1..., z n] ∈ R k × n.Algorithm flow chart as shown in Figure 2.
Iv) adopt spatial pyramid matching process, Z is divided into 2 l× 2 lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer icarry out maximum pond, keep character representation vector to have certain robustness, i.e. z i=max{|z 1i|, | z 2i| ..., | z ji|, obtain the i-th vectorial z of final low-rank coding i, wherein z jii-th element in jth block, j=2 l× 2 l, obtain last character representation matrix Z' like this.
V) character representation matrix Z' is inputted SVM classifier, according to every class image category label and its characteristic of correspondence representing matrix Z', study obtains the kernel function of SVM classifier.
Part of detecting:
I) to all image zooming-out SIFT feature remaining in image library, input feature value matrix X*=[x is obtained 1*, x 1* ..., x n*] ∈ R m × n, wherein x i* ∈ R ni-th component vector of eigenvectors matrix.
II) each proper vector x is calculated i* the Euclidean distance p and between dictionary vector i, computing formula is as follows:
p i * = exp ( d i s t ( x i * , D ) &sigma; )
Dist (x i*, D)=[dist (x i*, d 1) ..., dist (x i*, d k)] t, dist (x i*, d i) be proper vector x i* with each dictionary vector d ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
III) calculate the character representation vector Z * of local restriction low-rank coding, computing formula is:
Z * = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i * T p i * ) - 1 D T X *
Wherein λ 1, λ 2> 0 is weight; p i* be Euclidean distance that above-mentioned steps II calculates.
IV) adopt spatial pyramid matching process, Z* is divided into 2 l× 2 lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer i* carry out maximum pond, keep character representation vector to have certain robustness, i.e. z i*=max{|z 1i* |, | z 2i* | ..., | z ji* | }, obtain the i-th vectorial z of final low-rank coding i*, wherein z ji* be i-th element in jth block, j=2 l× 2 l, obtain last character representation matrix Z*' like this.
V) character representation matrix Z*' is inputted training department and divide the SVM classifier obtained, obtain the class label belonging to image to be classified.
VI) the statistical test classified image number percent of correctly classifying, obtains the nicety of grading that every class testing image is final.
In order to contrast the validity of the inventive method, have chosen the corresponding programs of the most frequently used several associated picture sorting techniques in test, arranging of the experiment parameter of contrast test is identical, all obtains experiment comparison of classification result, as shown in list.
Table 1 is the part classifying precision (first group) of Scene13;
Table 2 is the part classifying precision (second group) of Scene13;
Table 3 is overall nicety of grading and the time of Scene13;
Table 4 is the part classifying precision (first group) of Caltech101;
Table 5 is overall nicety of grading and the time of Caltech101.
Wherein ScSPM is the spatial pyramid model image sorting technique based on sparse coding, LLC is local restriction uniform enconding image classification method, LrrSPM is the image classification method represented based on the quick low-rank of spatial pyramid, and FLCLR is the image classification method of the quick local restriction geocoding that this experiment proposes.
Table 1
Classes ScSPM LLC LrrSPM FLCLR
MITtallbuilding 0.7182 0.7727 0.8148 0.85714
bedroom 0.8787 0.8129 0.9074 0.9858
MITforest 0.8397 0.8431 0.8538 0.9414
MIThighway 0.6387 0.6774 0.7027 0.7818
MITcoast 0.83654 0.8846 0.8125 0.89375
kitchen 0.9118 0.9161 0.8992 0.9423
MITmountain 0.8980 0.9102 0.9025 0.93226
Mean 0.8174 0.8310 0.8420 0.9049
Table 2
Classes ScSPM LLC LrrSPM FLCLR
Livingroom 0.8550 0.8438 0.83125 0.8437
MITstreet 0.6638 0.6840 0.6293 0.6207
MITopencountry 0.9478 0.9391 0.8783 0.8922
MITinsidecity 0.8802 0.8654 0.8594 0.8608
CALsuburb 0.8323 0.8308 0.8154 0.8123
PARoffice 0.6725 0.6032 0.6032 0.6009
Mean 0.8086 0.7944 0.7695 0.7718
Table 3
Table 4
Classes ScSPM LLC LrrSPM FLCLR
car_side 0.7000 0.7318 0.7536 0.8364
chair 0.4615 0.6223 0.6485 0.72230
Faces 0.6470 0.6647 0.7059 0.7765
mayfly 0.7192 0.7192 0.7587 0.8462
pyramid 0.6667 0.7578 0.7783 0.8519
water_lilly 0.4286 0.6429 0.6943 0.7714
widsor_char 0.84615 0.9315 0.9534 0.9931
Table 5

Claims (1)

1. an image classification method for quick local restriction low-rank coding, is characterized in that: comprise dictionary learning part, training part and part of detecting;
Described dictionary learning part comprises the following steps process;
1) utilize SIFT feature extracting method to carry out local shape factor to all images of image library, obtain input feature value matrix X=[x 1, x 2..., x n] ∈ R m × n; Wherein x i∈ R nbe i-th component vector of eigenvectors matrix, m is the dimension of vector, and n is vectorial number;
2) utilize K-means clustering or other dictionary learning methods, clustering processing is carried out to the eigenvectors matrix X of all image zooming-out, complete dictionary D=[d must be 1, d 2..., d k] ∈ R m × k; Wherein m is the dimension of dictionary, and k is the quantity of dictionary base;
Described training department divides the process of comprising the following steps;
I) from image library, selected part image, as training image, utilizes SIFT feature extracting method to carry out local shape factor to all training images, obtains input feature value matrix X'=[x 1', x 2' ..., x n'] ∈ R m × n; Wherein x i' ∈ R nbe i-th component vector of eigenvectors matrix, wherein m is the dimension of vector, and n is vectorial number;
Ii) each proper vector x is calculated i' and dictionary vector D between Euclidean distance p i, computing formula is as follows:
p i = exp ( d i s t ( x i &prime; , D ) &sigma; )
Dist (x i', D)=[dist (x i', d 1) ..., dist (x i', d k)] t, dist (x i', d i) be proper vector x i' d vectorial with each dictionary ibetween Euclidean distance; σ is the weight of adjustment local restriction speed;
Iii) calculate the character representation vector Z of local restriction low-rank coding, computing formula is:
Z = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i T p i ) - 1 D T X &prime;
Wherein λ 1, λ 2> 0 is weight, and I is unit vector, Z=[z 1, z 1..., z n] ∈ R k × n; p ifor the Euclidean distance that above-mentioned steps ii calculates;
Iv) adopt spatial pyramid matching process to process to the character representation vector Z obtained, Z is divided into 2 l× 2 lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively; Then for the z of each block in every layer icarry out maximum pond, i.e. z i=max{|z 1i|, | z 2i| ..., | z ji|, obtain the i-th vectorial z of final low-rank coding i, wherein z jii-th element in jth block, j=2 l× 2 l, obtain last character representation matrix Z' by this process;
V) character representation matrix Z' is inputted SVM classifier; Each image is to there being a label, and according to every class image category label and its characteristic of correspondence representing matrix Z', study obtains the kernel function of SVM classifier;
Described part of detecting comprises the following steps process;
I) to remaining image after choosing training image in image library, extract SIFT feature, obtain input feature value matrix X*=[x 1 *, x 1 *..., x n *] ∈ R m × n, wherein x i *∈ R ni-th component vector of eigenvectors matrix;
II) each proper vector x is calculated i *and the Euclidean distance p between dictionary vector i, computing formula is as follows:
p i * = exp ( d i s t ( x i * , D ) &sigma; )
Dist (x i *, D) and=[dist (x i *, d 1) ..., dist (x i *, d k)] t, dist (x i *, d i) be proper vector x i *with each dictionary vector d ibetween Euclidean distance; σ is the weight of adjustment local restriction speed;
III) calculate the character representation vector Z * of local restriction low-rank coding, computing formula is:
Z * = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i * T p i * ) - 1 D T X *
Wherein λ 1, λ 2> 0 is weight, and wherein I is unit vector; p i *for the Euclidean distance that above-mentioned steps II calculates;
IV) adopt spatial pyramid matching process, character representation vector Z * is divided into 2 l× 2 lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively; Then for the z of each block in every layer i *carry out maximum pond, keep character representation vector to have certain robustness, i.e. z i *=max{|z 1i *|, | z 2i *| ..., | z ji *|, obtain the i-th vectorial z of final low-rank coding i *, wherein z ji *i-th element in jth block, j=2 l× 2 l, obtain last character representation matrix Z*' like this;
V) character representation matrix Z*' is inputted training department and divide the SVM classifier obtained, obtain the class label belonging to image to be classified;
VI) the statistical test classified image number percent of correctly classifying, obtains the nicety of grading that every class testing image is final;
Step II I in step I ii and part of detecting in described training part, calculate the character representation vector of local restriction low-rank coding, the low-rank coding method calculating local restriction comprises the following steps;
A) when dictionary error is zero, E 1=D-D 0=0, F norm || || fcan nuclear norm be replaced || || *; By existing expression formula min Z , D | | Z | | * + &lambda; | | D - D 0 | | F + t r &lsqb; Y T ( D - D 0 - E 1 ) &rsqb; + t r &lsqb; Y T ( X - D Z - E 2 ) &rsqb; + &mu; 1 2 | | D - D 0 - E 1 | | F 2 + &mu; 2 2 | | X - D Z - E 2 | | F 2 Optimization is turned into
min Z | | Z | | F + &lambda; 1 / 2 | | X - D Z | | F 2
B) combined coding mode is taked in low-rank constraint, on the conforming basis of characteristics of image global structure, makes up the local space similarity of feature, adds local restriction item expression formula after optimizing in above-mentioned steps a becomes:
Wherein λ 1, λ 2> 0 is unequal weight; Symbol ⊙ represents that the corresponding element of two vectors is multiplied; p ifor proper vector x iwith the Euclidean distance p of dictionary vector D i:
p i = exp ( d i s t ( x i , D ) &sigma; )
Dist (x i, D) and=[dist (x i, d 1) ..., dist (x i, d k)] t, dist (x i, d i) be proper vector x iwith each dictionary vector d ibetween Euclidean distance; σ is the weight of adjustment local restriction speed;
C) differentiate is carried out to the optimizing expression added in above-mentioned steps b after local restriction item
Finally calculate character representation vector: Z = ( &lambda; 1 I + DD T + &lambda; 1 &lambda; 2 &Sigma; i = 1 k p i T p i ) - 1 D T X , Wherein I is the vector of unit length of k × k.
CN201510534124.2A 2015-08-27 2015-08-27 A kind of image classification method of quick local restriction low-rank coding Active CN105184298B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510534124.2A CN105184298B (en) 2015-08-27 2015-08-27 A kind of image classification method of quick local restriction low-rank coding

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510534124.2A CN105184298B (en) 2015-08-27 2015-08-27 A kind of image classification method of quick local restriction low-rank coding

Publications (2)

Publication Number Publication Date
CN105184298A true CN105184298A (en) 2015-12-23
CN105184298B CN105184298B (en) 2018-10-19

Family

ID=54906364

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510534124.2A Active CN105184298B (en) 2015-08-27 2015-08-27 A kind of image classification method of quick local restriction low-rank coding

Country Status (1)

Country Link
CN (1) CN105184298B (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105718934A (en) * 2016-01-25 2016-06-29 无锡中科富农物联科技有限公司 Method for pest image feature learning and identification based on low-rank sparse coding technology
CN106056082A (en) * 2016-05-31 2016-10-26 杭州电子科技大学 Video action recognition method based on sparse low-rank coding
CN106407899A (en) * 2016-08-30 2017-02-15 电子科技大学 Crowd density spectrum estimation method based on local texture features
CN107092875A (en) * 2017-04-11 2017-08-25 福州大学 A kind of new scene recognition method
CN107392140A (en) * 2017-07-19 2017-11-24 苏州闻捷传感技术有限公司 Polarization SAR terrain classification method with incremental learning is decomposed based on low-rank sparse
CN107766790A (en) * 2017-08-31 2018-03-06 电子科技大学 A kind of Human bodys' response method based on local restriction low-rank coding
CN108052867A (en) * 2017-11-20 2018-05-18 河海大学 A kind of single sample face recognition method based on bag of words
CN109271949A (en) * 2018-09-28 2019-01-25 中国科学院长春光学精密机械与物理研究所 Multispectral image data extraction method, device, equipment and readable storage medium storing program for executing
CN109726730A (en) * 2017-10-27 2019-05-07 财团法人工业技术研究院 Automatic optics inspection image classification method, system and computer-readable medium
CN110580482A (en) * 2017-11-30 2019-12-17 腾讯科技(深圳)有限公司 Image classification model training, image classification and personalized recommendation method and device
CN110889345A (en) * 2019-11-15 2020-03-17 重庆邮电大学 Method for recovering shielding face by distinguishing low-rank matrix based on cooperative representation and classification
CN111091176A (en) * 2018-10-23 2020-05-01 三星电子株式会社 Data recognition apparatus and method, and training apparatus and method
CN111930934A (en) * 2020-06-05 2020-11-13 江苏理工学院 Clustering method based on dual local consistency constraint sparse concept decomposition
US11315231B2 (en) 2018-06-08 2022-04-26 Industrial Technology Research Institute Industrial image inspection method and system and computer readable recording medium
CN115082727A (en) * 2022-05-25 2022-09-20 江苏大学 Scene classification method and system based on multilayer local perception depth dictionary learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103413151A (en) * 2013-07-22 2013-11-27 西安电子科技大学 Hyperspectral image classification method based on image regular low-rank expression dimensionality reduction
US20140156575A1 (en) * 2012-11-30 2014-06-05 Nuance Communications, Inc. Method and Apparatus of Processing Data Using Deep Belief Networks Employing Low-Rank Matrix Factorization
CN104268556A (en) * 2014-09-12 2015-01-07 西安电子科技大学 Hyperspectral image classification method based on nuclear low-rank representing graph and spatial constraint

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140156575A1 (en) * 2012-11-30 2014-06-05 Nuance Communications, Inc. Method and Apparatus of Processing Data Using Deep Belief Networks Employing Low-Rank Matrix Factorization
CN103413151A (en) * 2013-07-22 2013-11-27 西安电子科技大学 Hyperspectral image classification method based on image regular low-rank expression dimensionality reduction
CN104268556A (en) * 2014-09-12 2015-01-07 西安电子科技大学 Hyperspectral image classification method based on nuclear low-rank representing graph and spatial constraint

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
DEVANSH ARPIT等: "Locality-constrained Low Rank Coding for Face Recognition", 《21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION》 *

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105718934A (en) * 2016-01-25 2016-06-29 无锡中科富农物联科技有限公司 Method for pest image feature learning and identification based on low-rank sparse coding technology
CN106056082A (en) * 2016-05-31 2016-10-26 杭州电子科技大学 Video action recognition method based on sparse low-rank coding
CN106056082B (en) * 2016-05-31 2019-03-08 杭州电子科技大学 A kind of video actions recognition methods based on sparse low-rank coding
CN106407899B (en) * 2016-08-30 2019-07-19 电子科技大学 Crowd density Power estimation method based on Local textural feature
CN106407899A (en) * 2016-08-30 2017-02-15 电子科技大学 Crowd density spectrum estimation method based on local texture features
CN107092875A (en) * 2017-04-11 2017-08-25 福州大学 A kind of new scene recognition method
CN107092875B (en) * 2017-04-11 2020-02-07 福州大学 Novel scene recognition method
CN107392140A (en) * 2017-07-19 2017-11-24 苏州闻捷传感技术有限公司 Polarization SAR terrain classification method with incremental learning is decomposed based on low-rank sparse
CN107392140B (en) * 2017-07-19 2021-01-08 苏州闻捷传感技术有限公司 Polarized SAR terrain classification method based on low-rank sparse decomposition and incremental learning
CN107766790A (en) * 2017-08-31 2018-03-06 电子科技大学 A kind of Human bodys' response method based on local restriction low-rank coding
CN107766790B (en) * 2017-08-31 2021-04-30 电子科技大学 Human behavior identification method based on local constraint low-rank coding
CN109726730A (en) * 2017-10-27 2019-05-07 财团法人工业技术研究院 Automatic optics inspection image classification method, system and computer-readable medium
CN108052867A (en) * 2017-11-20 2018-05-18 河海大学 A kind of single sample face recognition method based on bag of words
CN108052867B (en) * 2017-11-20 2021-11-23 河海大学 Single-sample face recognition method based on bag-of-words model
CN110580482B (en) * 2017-11-30 2022-04-08 腾讯科技(深圳)有限公司 Image classification model training, image classification and personalized recommendation method and device
CN110580482A (en) * 2017-11-30 2019-12-17 腾讯科技(深圳)有限公司 Image classification model training, image classification and personalized recommendation method and device
US11315231B2 (en) 2018-06-08 2022-04-26 Industrial Technology Research Institute Industrial image inspection method and system and computer readable recording medium
CN109271949A (en) * 2018-09-28 2019-01-25 中国科学院长春光学精密机械与物理研究所 Multispectral image data extraction method, device, equipment and readable storage medium storing program for executing
CN111091176A (en) * 2018-10-23 2020-05-01 三星电子株式会社 Data recognition apparatus and method, and training apparatus and method
CN110889345A (en) * 2019-11-15 2020-03-17 重庆邮电大学 Method for recovering shielding face by distinguishing low-rank matrix based on cooperative representation and classification
CN111930934A (en) * 2020-06-05 2020-11-13 江苏理工学院 Clustering method based on dual local consistency constraint sparse concept decomposition
CN111930934B (en) * 2020-06-05 2023-12-26 江苏理工学院 Clustering method based on constraint sparse concept decomposition of dual local agreement
CN115082727A (en) * 2022-05-25 2022-09-20 江苏大学 Scene classification method and system based on multilayer local perception depth dictionary learning

Also Published As

Publication number Publication date
CN105184298B (en) 2018-10-19

Similar Documents

Publication Publication Date Title
CN105184298A (en) Image classification method through fast and locality-constrained low-rank coding process
CN107609601B (en) Ship target identification method based on multilayer convolutional neural network
He et al. Learning and incorporating top-down cues in image segmentation
CN112528928B (en) Commodity identification method based on self-attention depth network
CN111753828B (en) Natural scene horizontal character detection method based on deep convolutional neural network
EP3029606A2 (en) Method and apparatus for image classification with joint feature adaptation and classifier learning
CN110619369A (en) Fine-grained image classification method based on feature pyramid and global average pooling
CN105574548A (en) Hyperspectral data dimensionality-reduction method based on sparse and low-rank representation graph
CN109063719B (en) Image classification method combining structure similarity and class information
CN103955702A (en) SAR image terrain classification method based on depth RBF network
CN104268593A (en) Multiple-sparse-representation face recognition method for solving small sample size problem
CN107943856A (en) A kind of file classification method and system based on expansion marker samples
CN103605972A (en) Non-restricted environment face verification method based on block depth neural network
CN103761531A (en) Sparse-coding license plate character recognition method based on shape and contour features
CN109657704B (en) Sparse fusion-based coring scene feature extraction method
CN104298977A (en) Low-order representing human body behavior identification method based on irrelevance constraint
CN114780767A (en) Large-scale image retrieval method and system based on deep convolutional neural network
CN107220656A (en) A kind of multiple labeling data classification method based on self-adaptive features dimensionality reduction
CN103942214B (en) Natural image classification method and device on basis of multi-modal matrix filling
Sen et al. Face recognition using deep convolutional network and one-shot learning
CN115995040A (en) SAR image small sample target recognition method based on multi-scale network
CN104318271A (en) Image classification method based on adaptability coding and geometrical smooth convergence
CN116796248A (en) Forest health environment assessment system and method thereof
CN109034213A (en) Hyperspectral image classification method and system based on joint entropy principle
CN111898704A (en) Method and device for clustering content samples

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant