CN105184298A - Image classification method through fast and locality-constrained low-rank coding process - Google Patents
Image classification method through fast and locality-constrained low-rank coding process Download PDFInfo
- Publication number
- CN105184298A CN105184298A CN201510534124.2A CN201510534124A CN105184298A CN 105184298 A CN105184298 A CN 105184298A CN 201510534124 A CN201510534124 A CN 201510534124A CN 105184298 A CN105184298 A CN 105184298A
- Authority
- CN
- China
- Prior art keywords
- vector
- image
- dictionary
- dist
- matrix
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2321—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
- G06F18/23213—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an image classification method through the fast and locality-constrained low-rank coding process. The method comprises a dictionary learning part, a training part and a testing part. During the dictionary learning part, a corresponding dictionary is obtained based on SIFT features extracted out of all data sets according to the k-means clustering method or other dictionary learning methods. During the training part, one part of images in each data set are selected as training images, and then the corresponding low-rank representation corresponding to each data set is learned based on the tags of the images. Through the SPM quantization process and the SVM classifier-based training process, category labels and a kernel function corresponding to a feature representation matrix are obtained. During the testing part, all remaining images in each data set are adopted as to-be-classified test images. For any unclassified image, the SIFT features of the image are extracted firstly, and then a low-rank coding matrix thereof in the dictionary is found out. Through the SPM quantization process, a final feature representation matrix is obtained. The final feature representation matrix is input into an SVM classifier obtained during the training part, so that the category of the image can be known.
Description
Technical field
The present invention relates to a kind of method utilizing quick local restriction low-rank to encode and carry out Images Classification, also precision is higher fast for the method, is a kind of sorting technique of data-driven of image content-based.
Background technology
Images Classification belongs to an important research direction of computer vision, and it comprises the pre-service of image, the expression of feature and the design of sorter, and wherein feature representation comprises the coding of the extraction of characteristics of image, the dimensionality reduction of feature and feature.Feature extraction mainly by attributes such as the space distributions of various content presentation color out in image, brightness, texture, shape and pixel to iamge description.Feature expression refers to carry out adding up (vector quantization), coding or additive method to form the last feature of piece image in the most basic feature base, can have better performance under normal circumstances than original essential characteristic; Good feature representation can promote the performance of Images Classification and identification.
The Measures compare represented for feature interpretation is in recent years many, there is visual word bag model (bagofwords, BOW), based on the spatial pyramid model (SparseCodingSpatialPyramidMatching of sparse coding, ScSPM), local restriction uniform enconding (Locality-constrainedLinearCoding, LLC), image classification method (the StructuredLow-rankRepresentation that structure low-rank represents, SLRR) and low-rank sparse decompose image classification method (Low-rankSparseCoding, LRSC), these methods overcome some difficult problems to a certain extent, but still all Shortcomings.
Visual word bag model (bagofwords, BOW), although be easy to build, have ignored similar image and has certain space structure similarity.Spatial pyramid model (SparseCodingSpatialPyramidMatching, ScSPM) based on sparse coding have ignored the correlativity between local description, and to the change of feature and noise-sensitive, and calculated amount is very large, consumption internal memory.Local restriction uniform enconding (Locality-constrainedLinearCoding, LLC), this locality result in openness to a certain extent, relative to sparse coding, the operand of algorithm is reduced greatly, speed improves, but LLC have ignored the spatial information of the one-piece construction feature between feature.Image classification method (the StructuredLow-rankRepresentation that structure low-rank represents, SLRR), there is good effect to recognition of face, be particularly useful for having the Images Classification of critical noisy (as block, lighting angle change etc.).Image classification method (Low-rankSparseCoding) (LRSC) nicety of grading that low-rank sparse is decomposed obtains certain raising, but have ignored local spatial information; Wherein sparse part is still very sensitive to coding.Based on the image classification method (FastLow-rankRepresentationbasedSpatialPyramidMatching that the quick low-rank of spatial pyramid represents, LrrSPM), the classification speed of the method is 5 ~ 16 times of ScSPM, but nicety of grading is lower than ScSPM and LLC method.
The coded system that current image classification method mostly adopts sparse coding and low-rank to represent, but in view of sparse coding to the change of image and noise more responsive, very consume internal memory; The coded system adopting low-rank to represent obtains a balance between precision and computation complexity.
Summary of the invention
In view of the defect of current sorting technique in classifying quality and classification speed, the present invention proposes a kind of image classification method of quick local restriction low-rank coding, and the method comprises dictionary learning part, training part and part of detecting three parts.Dictionary learning means suitable, to the SIFT feature of image zooming-out images all in image library, obtains dictionary through K-means cluster; Training part is first to the image zooming-out SIFT feature in training set, then try to achieve the feature coding of SIFT feature under dictionary to represent, adopt the low-rank coding method of local restriction, consider global structure consistance and the local spatial simlanty of image simultaneously, obtain feature coding matrix, behind spatial pyramid coupling core, pond, obtain final character representation matrix, adopt SVM method to train character representation matrix and class mark, set up disaggregated model.First part of detecting carries out feature extraction to any image, then encoder matrix under its dictionary is tried to achieve, after spatial pyramid coupling core, pond obtain character representation matrix, the SVM classifier obtained is trained in input, obtains the classification results belonging to this image.
The technical scheme adopted for realizing the object of the invention is such, and a kind of image classification method of quick local restriction low-rank coding, comprises dictionary learning part, training part and part of detecting.
Described dictionary learning part steps process is as follows:
1) utilize SIFT feature extracting method to carry out local shape factor to all images of image library, obtain input feature value matrix X=[x
1, x
2..., x
n] ∈ R
m × n, wherein x
i∈ R
nbe i-th component vector of eigenvectors matrix, wherein m is the dimension of vector, and n is vectorial number.
2) utilize K-means clustering or other dictionary learning methods, clustering processing is carried out to the eigenvectors matrix X of all image zooming-out, complete dictionary D=[d must be
1, d
2..., d
k] ∈ R
m × k, wherein m is the dimension of dictionary, and k is the number of dictionary base, and meet dictionary and cross completeness m < < k, k is generally 256,1024,4096 etc.
Described training department step by step process is as follows:
I) from image library, selected part image, as training image, utilizes SIFT feature extracting method to carry out local shape factor to all training images, obtains input feature value matrix X'=[x
1', x
2' ..., x
n'] ∈ R
m × n.Wherein x
i' ∈ R
nbe i-th component vector of eigenvectors matrix, wherein m is the dimension of vector, and n is vectorial number.
Ii) each proper vector x is calculated
iand the Euclidean distance p between dictionary vector
i, computing formula is as follows:
Dist (x
i', D)=[dist (x
i', d
1) ..., dist (x
i', d
k)]
t, dist (x
i', d
i) be proper vector x
i' d vectorial with each dictionary
ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
Iii) calculate the character representation vector Z of local restriction low-rank coding, computing formula is:
Wherein λ
1, λ
2> 0 is weight; p
ifor the Euclidean distance p that above-mentioned steps ii calculates
i.
Iv) adopt spatial pyramid matching process to process to the character representation vector Z obtained, Z is divided into 2
l× 2
lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer
icarry out maximum pond, i.e. z
i=max{|z
1i|, | z
2i| ..., | z
ji|, obtain the i-th vectorial z of final low-rank coding
i, wherein z
jii-th element in jth block, j=2
l× 2
l, obtain last character representation matrix Z' by this process.
V) character representation matrix Z' is inputted SVM classifier, according to every class image category label and its characteristic of correspondence representing matrix Z', study obtains the kernel function of SVM classifier.
Described part of detecting step process is as follows:
I) to all image zooming-out SIFT feature remaining in image library, input feature value matrix X*=[x is obtained
1*, x
1* ..., x
n*] ∈ R
m × n, wherein x
i* ∈ R
ni-th component vector of eigenvectors matrix.
II) each proper vector x is calculated
i* the Euclidean distance p and between dictionary vector
i, computing formula is as follows:
Dist (x
i*, D)=[dist (x
i*, d
1) ..., dist (x
i*, d
k)]
t, dist (x
i*, d
i) be proper vector x
i* with each dictionary vector d
ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
III) calculate the character representation vector Z * of local restriction low-rank coding, computing formula is:
Wherein λ
1, λ
2> 0 is weight; p
i* be Euclidean distance that above-mentioned steps II calculates.
IV) adopt spatial pyramid matching process, character representation vector Z * is divided into 2
l× 2
lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer
i* carry out maximum pond, keep character representation vector to have certain robustness, i.e. z
i*=max{|z
1i* |, | z
2i* | ..., | z
ji* | }, obtain the i-th vectorial z of final low-rank coding
i*, wherein z
ji* be i-th element in jth block, j=2
l× 2
l, obtain last character representation matrix Z*' like this.
V) character representation matrix Z*' is inputted training department and divide the SVM classifier obtained, obtain the class label belonging to image to be classified.
VI) the statistical test classified image number percent of correctly classifying, obtains the nicety of grading that every class testing image is final.
Further, in the image classification method of described a kind of quick local restriction low-rank coding, the low-rank that in training part, in step I ii and part of detecting, Step II I calculates feature coding represents vector.
The coding method that existing low-rank represents is that the low-rank finding proper vector in the error range allowed represents, namely
s.t.X=DZ+E
Wherein μ
1, μ
2> 0 is error penalty term, λ
1> 0 is weights, and Z is the expression of feature vector, X under dictionary D.Wherein ‖ Z ‖
*nuclear norm, ‖ E ‖
pthe l of constraint sparse coding item
1norm or retrain a certain class interference l
2-1norm.
Existing low-rank optimized algorithm commonly adopts Lagrange multiplier to optimize AugmentedLagrangeMultiplealgorithm (ALM), and optimizing expression is as follows:
Wherein ‖ ‖
ff norm (Frobeniusnorm), λ
1> 0 is weights, Y
1, Y
2lagrange multiplier, μ
1, μ
2> 0 is error penalty term.
Above formula 1 can adopt typical exactALM or inexactALM method, adopt alternative manner step by step respectively regeneration characteristics represent vector Z and initial dictionary D
0, the computation complexity of this method is O (mn
2), n is the number of proper vector.This calculating comes that is very large, and is a kind of method of iteration optimization of off-line, cannot obtain classification results (can not asking character representation if do not belonged to such other image) to the input picture of increment type.Meanwhile, existing low-rank constraint only considered the global structure consistance of characteristics of image, have ignored the local space similarity of feature, comprehensively can not represent the related information between characteristics of image.
The low-rank coding method step calculating local restriction in this method is as follows:
A) when dictionary error is zero, E
1=D-D
0=0, F norm ‖ ‖
fnuclear norm ‖ ‖ can be replaced
*.Above-mentioned optimizing expression 1 becomes
B) combined coding mode is taked in low-rank constraint, on the conforming basis of characteristics of image global structure, makes up the local space similarity of feature, adds local restriction item
above-mentioned expression formula 2 becomes:
Wherein λ
1, λ
2> 0 is unequal weight, and symbol ⊙ represents that the corresponding element of two vectors is multiplied, p
ifor proper vector x
iwith the Euclidean distance p of dictionary vector D
i
Dist (x
i, D) and=[dist (x
i, d
1) ..., dist (x
i, d
k)]
t, dist (x
i, d
i) be proper vector x
iwith each dictionary vector d
ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
C) differentiate is carried out to above-mentioned expression formula 3
Finally calculate character representation vector:
Wherein I is the vector of unit length of k × k.
In differentiate process, take threshold value constraint item by character representation vector z
iconstrain within 0.98.
Technique effect of the present invention is mathematical, the present invention adopts a kind of low-rank coding specification method of local restriction fast, consider global structure consistance and the local spatial simlanty of characteristics of image simultaneously, better can not only represent the related information between characteristics of image, obtain effective nicety of grading, and calculated amount reduces, classification speed is a bit larger tham LLC, is only 19.4 ~ 25% of ScSPM.
Accompanying drawing explanation
Fig. 1 is Images Classification main-process stream;
Fig. 2 is local restriction low-rank coding (FLCLR) process flow diagram;
Fig. 3 is the example images of classification in the middle part of table 1;
Fig. 4 is the example images of classification in the middle part of table 2;
Fig. 5 is the example images of classification in the middle part of table 4;
In figure: 1-feature extraction, 2-feature coding, 3-pond process.
Embodiment
Below in conjunction with embodiment, the invention will be further described, but should not be construed the above-mentioned subject area of the present invention and be only limitted to following embodiment.Without departing from the idea case in the present invention described above, according to ordinary skill knowledge and customary means, make various replacement and change, all should be included in protection scope of the present invention.
Below implement to choose Images Classification conventional authoritative image library Scene13 and Caltech101 as class test data set.Scene13 is the most frequently used scene classification data set, has 13 class images, comprises suburb (CALsuburb), seashore (MITcoast), and totally 4485 width such as forest (MITforest), every class comprises 200 to 400 pictures.Caltech101 image library comprises 101 classes (as animal, flower, face etc.) and a background classes image, totally 9144 width.Each class comprises 31 to 800 width images, and has difference in larger class, is the most frequently used database of Images Classification.
As shown in Figure 1, whole implementation step is divided into three parts, and Part I is dictionary learning part, extracts all SIFT feature according to data set, uses the dictionary learning methods such as k-menas clustering to obtain corresponding dictionary.Part II is training part, chooses the parts of images in the every class of each data set, and wherein Scene13 gets 100 at random as training image, and Caltech101 gets 30 at random as training image.Then the low-rank learning each classification corresponding for the label of these training images represents, quantizes to obtain obtaining class label and kernel function corresponding to character representation through svm classifier training through SPM.Part III is part of detecting, in each database, remaining all images are as test pattern to be sorted, the image of classification is not known for any one, first SIFT feature is extracted, then its low-rank coding vector under dictionary is tried to achieve, quantize to obtain final character representation matrix through SPM, inputted the SVM classifier that training part gained arrives, and then obtain the classification belonging to this image.
Pre-service illustrates: image size is adjusted to maximum 300 × 300 pixels, and the grid of the intensive Bian sample of feature extraction is 16 × 16 pixels, and its step-length is 4 pixels.
Dictionary learning part:
1) to image zooming-out SIFT local features all in image library, input feature value matrix X=[x is obtained
1, x
2..., x
n] ∈ R
m × n, wherein x
i∈ R
nbe i-th local feature vectors of image, wherein m is the dimension of vector, and n is vectorial number.
2) utilize K-means clustering or other dictionary learning methods, clustering processing is carried out to the proper vector of all image zooming-out, complete dictionary D=[d must be
1, d
2..., d
k] ∈ R
m × k, wherein k is the dimension of dictionary, and m is the number of dictionary base, meets dictionary and crosses completeness m < < k, and wherein k is that the dimension of dictionary gets 1024.
Training part:
I) SIFT feature is extracted to the training image chosen, obtain input feature value matrix X'=[x'
1, x'
2..., x'
n] ∈ R
m × n, wherein x
i' ∈ R
ni-th component vector of eigenvectors matrix.
Ii) each proper vector x in calculation training image
i' and dictionary vector between Euclidean distance p
i, computing formula is as follows:
Dist (x
i', D)=[dist (x
i', d
1) ..., dist (x
i', d
k)]
t, dist (x
i', d
i) be proper vector x
i' d vectorial with each dictionary
ibetween Euclidean distance; σ is the weight of adjustment local restriction speed, gets 100.
Iii) calculate the character representation vector Z of local restriction low-rank coding, computing formula is:
Wherein I is unit matrix, and dimension is consistent with Z, λ
1, λ
2> 0 is weight, λ
1, λ
20.7 is respectively, 0.0001 according to the empirical value value of document Wang and Peng.P
ifor the Euclidean distance calculated in above-mentioned steps 2.
Step I i and step I ii seeks to obtain character representation vector z, the i.e. X=DZ of eigenvectors matrix X under dictionary D
,wherein Z=[z
1, z
1..., z
n] ∈ R
k × n.Algorithm flow chart as shown in Figure 2.
Iv) adopt spatial pyramid matching process, Z is divided into 2
l× 2
lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer
icarry out maximum pond, keep character representation vector to have certain robustness, i.e. z
i=max{|z
1i|, | z
2i| ..., | z
ji|, obtain the i-th vectorial z of final low-rank coding
i, wherein z
jii-th element in jth block, j=2
l× 2
l, obtain last character representation matrix Z' like this.
V) character representation matrix Z' is inputted SVM classifier, according to every class image category label and its characteristic of correspondence representing matrix Z', study obtains the kernel function of SVM classifier.
Part of detecting:
I) to all image zooming-out SIFT feature remaining in image library, input feature value matrix X*=[x is obtained
1*, x
1* ..., x
n*] ∈ R
m × n, wherein x
i* ∈ R
ni-th component vector of eigenvectors matrix.
II) each proper vector x is calculated
i* the Euclidean distance p and between dictionary vector
i, computing formula is as follows:
Dist (x
i*, D)=[dist (x
i*, d
1) ..., dist (x
i*, d
k)]
t, dist (x
i*, d
i) be proper vector x
i* with each dictionary vector d
ibetween Euclidean distance, σ is the weight of adjustment local restriction speed.
III) calculate the character representation vector Z * of local restriction low-rank coding, computing formula is:
Wherein λ
1, λ
2> 0 is weight; p
i* be Euclidean distance that above-mentioned steps II calculates.
IV) adopt spatial pyramid matching process, Z* is divided into 2
l× 2
lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively.Then for the z of each block in every layer
i* carry out maximum pond, keep character representation vector to have certain robustness, i.e. z
i*=max{|z
1i* |, | z
2i* | ..., | z
ji* | }, obtain the i-th vectorial z of final low-rank coding
i*, wherein z
ji* be i-th element in jth block, j=2
l× 2
l, obtain last character representation matrix Z*' like this.
V) character representation matrix Z*' is inputted training department and divide the SVM classifier obtained, obtain the class label belonging to image to be classified.
VI) the statistical test classified image number percent of correctly classifying, obtains the nicety of grading that every class testing image is final.
In order to contrast the validity of the inventive method, have chosen the corresponding programs of the most frequently used several associated picture sorting techniques in test, arranging of the experiment parameter of contrast test is identical, all obtains experiment comparison of classification result, as shown in list.
Table 1 is the part classifying precision (first group) of Scene13;
Table 2 is the part classifying precision (second group) of Scene13;
Table 3 is overall nicety of grading and the time of Scene13;
Table 4 is the part classifying precision (first group) of Caltech101;
Table 5 is overall nicety of grading and the time of Caltech101.
Wherein ScSPM is the spatial pyramid model image sorting technique based on sparse coding, LLC is local restriction uniform enconding image classification method, LrrSPM is the image classification method represented based on the quick low-rank of spatial pyramid, and FLCLR is the image classification method of the quick local restriction geocoding that this experiment proposes.
Table 1
Classes | ScSPM | LLC | LrrSPM | FLCLR |
MITtallbuilding | 0.7182 | 0.7727 | 0.8148 | 0.85714 |
bedroom | 0.8787 | 0.8129 | 0.9074 | 0.9858 |
MITforest | 0.8397 | 0.8431 | 0.8538 | 0.9414 |
MIThighway | 0.6387 | 0.6774 | 0.7027 | 0.7818 |
MITcoast | 0.83654 | 0.8846 | 0.8125 | 0.89375 |
kitchen | 0.9118 | 0.9161 | 0.8992 | 0.9423 |
MITmountain | 0.8980 | 0.9102 | 0.9025 | 0.93226 |
Mean | 0.8174 | 0.8310 | 0.8420 | 0.9049 |
Table 2
Classes | ScSPM | LLC | LrrSPM | FLCLR |
Livingroom | 0.8550 | 0.8438 | 0.83125 | 0.8437 |
MITstreet | 0.6638 | 0.6840 | 0.6293 | 0.6207 |
MITopencountry | 0.9478 | 0.9391 | 0.8783 | 0.8922 |
MITinsidecity | 0.8802 | 0.8654 | 0.8594 | 0.8608 |
CALsuburb | 0.8323 | 0.8308 | 0.8154 | 0.8123 |
PARoffice | 0.6725 | 0.6032 | 0.6032 | 0.6009 |
Mean | 0.8086 | 0.7944 | 0.7695 | 0.7718 |
Table 3
Table 4
Classes | ScSPM | LLC | LrrSPM | FLCLR |
car_side | 0.7000 | 0.7318 | 0.7536 | 0.8364 |
chair | 0.4615 | 0.6223 | 0.6485 | 0.72230 |
Faces | 0.6470 | 0.6647 | 0.7059 | 0.7765 |
mayfly | 0.7192 | 0.7192 | 0.7587 | 0.8462 |
pyramid | 0.6667 | 0.7578 | 0.7783 | 0.8519 |
water_lilly | 0.4286 | 0.6429 | 0.6943 | 0.7714 |
widsor_char | 0.84615 | 0.9315 | 0.9534 | 0.9931 |
Table 5
Claims (1)
1. an image classification method for quick local restriction low-rank coding, is characterized in that: comprise dictionary learning part, training part and part of detecting;
Described dictionary learning part comprises the following steps process;
1) utilize SIFT feature extracting method to carry out local shape factor to all images of image library, obtain input feature value matrix X=[x
1, x
2..., x
n] ∈ R
m × n; Wherein x
i∈ R
nbe i-th component vector of eigenvectors matrix, m is the dimension of vector, and n is vectorial number;
2) utilize K-means clustering or other dictionary learning methods, clustering processing is carried out to the eigenvectors matrix X of all image zooming-out, complete dictionary D=[d must be
1, d
2..., d
k] ∈ R
m × k; Wherein m is the dimension of dictionary, and k is the quantity of dictionary base;
Described training department divides the process of comprising the following steps;
I) from image library, selected part image, as training image, utilizes SIFT feature extracting method to carry out local shape factor to all training images, obtains input feature value matrix X'=[x
1', x
2' ..., x
n'] ∈ R
m × n; Wherein x
i' ∈ R
nbe i-th component vector of eigenvectors matrix, wherein m is the dimension of vector, and n is vectorial number;
Ii) each proper vector x is calculated
i' and dictionary vector D between Euclidean distance p
i, computing formula is as follows:
Dist (x
i', D)=[dist (x
i', d
1) ..., dist (x
i', d
k)]
t, dist (x
i', d
i) be proper vector x
i' d vectorial with each dictionary
ibetween Euclidean distance; σ is the weight of adjustment local restriction speed;
Iii) calculate the character representation vector Z of local restriction low-rank coding, computing formula is:
Wherein λ
1, λ
2> 0 is weight, and I is unit vector, Z=[z
1, z
1..., z
n] ∈ R
k × n; p
ifor the Euclidean distance that above-mentioned steps ii calculates;
Iv) adopt spatial pyramid matching process to process to the character representation vector Z obtained, Z is divided into 2
l× 2
lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively; Then for the z of each block in every layer
icarry out maximum pond, i.e. z
i=max{|z
1i|, | z
2i| ..., | z
ji|, obtain the i-th vectorial z of final low-rank coding
i, wherein z
jii-th element in jth block, j=2
l× 2
l, obtain last character representation matrix Z' by this process;
V) character representation matrix Z' is inputted SVM classifier; Each image is to there being a label, and according to every class image category label and its characteristic of correspondence representing matrix Z', study obtains the kernel function of SVM classifier;
Described part of detecting comprises the following steps process;
I) to remaining image after choosing training image in image library, extract SIFT feature, obtain input feature value matrix X*=[x
1 *, x
1 *..., x
n *] ∈ R
m × n, wherein x
i *∈ R
ni-th component vector of eigenvectors matrix;
II) each proper vector x is calculated
i *and the Euclidean distance p between dictionary vector
i, computing formula is as follows:
Dist (x
i *, D) and=[dist (x
i *, d
1) ..., dist (x
i *, d
k)]
t, dist (x
i *, d
i) be proper vector x
i *with each dictionary vector d
ibetween Euclidean distance; σ is the weight of adjustment local restriction speed;
III) calculate the character representation vector Z * of local restriction low-rank coding, computing formula is:
Wherein λ
1, λ
2> 0 is weight, and wherein I is unit vector; p
i *for the Euclidean distance that above-mentioned steps II calculates;
IV) adopt spatial pyramid matching process, character representation vector Z * is divided into 2
l× 2
lblock, wherein l=1,2,4, represent the number of different three layers of corresponding spatial pyramid piecemeal respectively; Then for the z of each block in every layer
i *carry out maximum pond, keep character representation vector to have certain robustness, i.e. z
i *=max{|z
1i *|, | z
2i *| ..., | z
ji *|, obtain the i-th vectorial z of final low-rank coding
i *, wherein z
ji *i-th element in jth block, j=2
l× 2
l, obtain last character representation matrix Z*' like this;
V) character representation matrix Z*' is inputted training department and divide the SVM classifier obtained, obtain the class label belonging to image to be classified;
VI) the statistical test classified image number percent of correctly classifying, obtains the nicety of grading that every class testing image is final;
Step II I in step I ii and part of detecting in described training part, calculate the character representation vector of local restriction low-rank coding, the low-rank coding method calculating local restriction comprises the following steps;
A) when dictionary error is zero, E
1=D-D
0=0, F norm || ||
fcan nuclear norm be replaced || ||
*; By existing expression formula
Optimization is turned into
B) combined coding mode is taked in low-rank constraint, on the conforming basis of characteristics of image global structure, makes up the local space similarity of feature, adds local restriction item
expression formula after optimizing in above-mentioned steps a becomes:
Wherein λ
1, λ
2> 0 is unequal weight; Symbol ⊙ represents that the corresponding element of two vectors is multiplied; p
ifor proper vector x
iwith the Euclidean distance p of dictionary vector D
i:
Dist (x
i, D) and=[dist (x
i, d
1) ..., dist (x
i, d
k)]
t, dist (x
i, d
i) be proper vector x
iwith each dictionary vector d
ibetween Euclidean distance; σ is the weight of adjustment local restriction speed;
C) differentiate is carried out to the optimizing expression added in above-mentioned steps b after local restriction item
Finally calculate character representation vector:
Wherein I is the vector of unit length of k × k.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510534124.2A CN105184298B (en) | 2015-08-27 | 2015-08-27 | A kind of image classification method of quick local restriction low-rank coding |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510534124.2A CN105184298B (en) | 2015-08-27 | 2015-08-27 | A kind of image classification method of quick local restriction low-rank coding |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105184298A true CN105184298A (en) | 2015-12-23 |
CN105184298B CN105184298B (en) | 2018-10-19 |
Family
ID=54906364
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510534124.2A Active CN105184298B (en) | 2015-08-27 | 2015-08-27 | A kind of image classification method of quick local restriction low-rank coding |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105184298B (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105718934A (en) * | 2016-01-25 | 2016-06-29 | 无锡中科富农物联科技有限公司 | Method for pest image feature learning and identification based on low-rank sparse coding technology |
CN106056082A (en) * | 2016-05-31 | 2016-10-26 | 杭州电子科技大学 | Video action recognition method based on sparse low-rank coding |
CN106407899A (en) * | 2016-08-30 | 2017-02-15 | 电子科技大学 | Crowd density spectrum estimation method based on local texture features |
CN107092875A (en) * | 2017-04-11 | 2017-08-25 | 福州大学 | A kind of new scene recognition method |
CN107392140A (en) * | 2017-07-19 | 2017-11-24 | 苏州闻捷传感技术有限公司 | Polarization SAR terrain classification method with incremental learning is decomposed based on low-rank sparse |
CN107766790A (en) * | 2017-08-31 | 2018-03-06 | 电子科技大学 | A kind of Human bodys' response method based on local restriction low-rank coding |
CN108052867A (en) * | 2017-11-20 | 2018-05-18 | 河海大学 | A kind of single sample face recognition method based on bag of words |
CN109271949A (en) * | 2018-09-28 | 2019-01-25 | 中国科学院长春光学精密机械与物理研究所 | Multispectral image data extraction method, device, equipment and readable storage medium storing program for executing |
CN109726730A (en) * | 2017-10-27 | 2019-05-07 | 财团法人工业技术研究院 | Automatic optics inspection image classification method, system and computer-readable medium |
CN110580482A (en) * | 2017-11-30 | 2019-12-17 | 腾讯科技(深圳)有限公司 | Image classification model training, image classification and personalized recommendation method and device |
CN110889345A (en) * | 2019-11-15 | 2020-03-17 | 重庆邮电大学 | Method for recovering shielding face by distinguishing low-rank matrix based on cooperative representation and classification |
CN111091176A (en) * | 2018-10-23 | 2020-05-01 | 三星电子株式会社 | Data recognition apparatus and method, and training apparatus and method |
CN111930934A (en) * | 2020-06-05 | 2020-11-13 | 江苏理工学院 | Clustering method based on dual local consistency constraint sparse concept decomposition |
US11315231B2 (en) | 2018-06-08 | 2022-04-26 | Industrial Technology Research Institute | Industrial image inspection method and system and computer readable recording medium |
CN115082727A (en) * | 2022-05-25 | 2022-09-20 | 江苏大学 | Scene classification method and system based on multilayer local perception depth dictionary learning |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103413151A (en) * | 2013-07-22 | 2013-11-27 | 西安电子科技大学 | Hyperspectral image classification method based on image regular low-rank expression dimensionality reduction |
US20140156575A1 (en) * | 2012-11-30 | 2014-06-05 | Nuance Communications, Inc. | Method and Apparatus of Processing Data Using Deep Belief Networks Employing Low-Rank Matrix Factorization |
CN104268556A (en) * | 2014-09-12 | 2015-01-07 | 西安电子科技大学 | Hyperspectral image classification method based on nuclear low-rank representing graph and spatial constraint |
-
2015
- 2015-08-27 CN CN201510534124.2A patent/CN105184298B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140156575A1 (en) * | 2012-11-30 | 2014-06-05 | Nuance Communications, Inc. | Method and Apparatus of Processing Data Using Deep Belief Networks Employing Low-Rank Matrix Factorization |
CN103413151A (en) * | 2013-07-22 | 2013-11-27 | 西安电子科技大学 | Hyperspectral image classification method based on image regular low-rank expression dimensionality reduction |
CN104268556A (en) * | 2014-09-12 | 2015-01-07 | 西安电子科技大学 | Hyperspectral image classification method based on nuclear low-rank representing graph and spatial constraint |
Non-Patent Citations (1)
Title |
---|
DEVANSH ARPIT等: "Locality-constrained Low Rank Coding for Face Recognition", 《21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION》 * |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105718934A (en) * | 2016-01-25 | 2016-06-29 | 无锡中科富农物联科技有限公司 | Method for pest image feature learning and identification based on low-rank sparse coding technology |
CN106056082A (en) * | 2016-05-31 | 2016-10-26 | 杭州电子科技大学 | Video action recognition method based on sparse low-rank coding |
CN106056082B (en) * | 2016-05-31 | 2019-03-08 | 杭州电子科技大学 | A kind of video actions recognition methods based on sparse low-rank coding |
CN106407899B (en) * | 2016-08-30 | 2019-07-19 | 电子科技大学 | Crowd density Power estimation method based on Local textural feature |
CN106407899A (en) * | 2016-08-30 | 2017-02-15 | 电子科技大学 | Crowd density spectrum estimation method based on local texture features |
CN107092875A (en) * | 2017-04-11 | 2017-08-25 | 福州大学 | A kind of new scene recognition method |
CN107092875B (en) * | 2017-04-11 | 2020-02-07 | 福州大学 | Novel scene recognition method |
CN107392140A (en) * | 2017-07-19 | 2017-11-24 | 苏州闻捷传感技术有限公司 | Polarization SAR terrain classification method with incremental learning is decomposed based on low-rank sparse |
CN107392140B (en) * | 2017-07-19 | 2021-01-08 | 苏州闻捷传感技术有限公司 | Polarized SAR terrain classification method based on low-rank sparse decomposition and incremental learning |
CN107766790A (en) * | 2017-08-31 | 2018-03-06 | 电子科技大学 | A kind of Human bodys' response method based on local restriction low-rank coding |
CN107766790B (en) * | 2017-08-31 | 2021-04-30 | 电子科技大学 | Human behavior identification method based on local constraint low-rank coding |
CN109726730A (en) * | 2017-10-27 | 2019-05-07 | 财团法人工业技术研究院 | Automatic optics inspection image classification method, system and computer-readable medium |
CN108052867A (en) * | 2017-11-20 | 2018-05-18 | 河海大学 | A kind of single sample face recognition method based on bag of words |
CN108052867B (en) * | 2017-11-20 | 2021-11-23 | 河海大学 | Single-sample face recognition method based on bag-of-words model |
CN110580482B (en) * | 2017-11-30 | 2022-04-08 | 腾讯科技(深圳)有限公司 | Image classification model training, image classification and personalized recommendation method and device |
CN110580482A (en) * | 2017-11-30 | 2019-12-17 | 腾讯科技(深圳)有限公司 | Image classification model training, image classification and personalized recommendation method and device |
US11315231B2 (en) | 2018-06-08 | 2022-04-26 | Industrial Technology Research Institute | Industrial image inspection method and system and computer readable recording medium |
CN109271949A (en) * | 2018-09-28 | 2019-01-25 | 中国科学院长春光学精密机械与物理研究所 | Multispectral image data extraction method, device, equipment and readable storage medium storing program for executing |
CN111091176A (en) * | 2018-10-23 | 2020-05-01 | 三星电子株式会社 | Data recognition apparatus and method, and training apparatus and method |
CN110889345A (en) * | 2019-11-15 | 2020-03-17 | 重庆邮电大学 | Method for recovering shielding face by distinguishing low-rank matrix based on cooperative representation and classification |
CN111930934A (en) * | 2020-06-05 | 2020-11-13 | 江苏理工学院 | Clustering method based on dual local consistency constraint sparse concept decomposition |
CN111930934B (en) * | 2020-06-05 | 2023-12-26 | 江苏理工学院 | Clustering method based on constraint sparse concept decomposition of dual local agreement |
CN115082727A (en) * | 2022-05-25 | 2022-09-20 | 江苏大学 | Scene classification method and system based on multilayer local perception depth dictionary learning |
Also Published As
Publication number | Publication date |
---|---|
CN105184298B (en) | 2018-10-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105184298A (en) | Image classification method through fast and locality-constrained low-rank coding process | |
CN107609601B (en) | Ship target identification method based on multilayer convolutional neural network | |
He et al. | Learning and incorporating top-down cues in image segmentation | |
CN112528928B (en) | Commodity identification method based on self-attention depth network | |
CN111753828B (en) | Natural scene horizontal character detection method based on deep convolutional neural network | |
EP3029606A2 (en) | Method and apparatus for image classification with joint feature adaptation and classifier learning | |
CN110619369A (en) | Fine-grained image classification method based on feature pyramid and global average pooling | |
CN105574548A (en) | Hyperspectral data dimensionality-reduction method based on sparse and low-rank representation graph | |
CN109063719B (en) | Image classification method combining structure similarity and class information | |
CN103955702A (en) | SAR image terrain classification method based on depth RBF network | |
CN104268593A (en) | Multiple-sparse-representation face recognition method for solving small sample size problem | |
CN107943856A (en) | A kind of file classification method and system based on expansion marker samples | |
CN103605972A (en) | Non-restricted environment face verification method based on block depth neural network | |
CN103761531A (en) | Sparse-coding license plate character recognition method based on shape and contour features | |
CN109657704B (en) | Sparse fusion-based coring scene feature extraction method | |
CN104298977A (en) | Low-order representing human body behavior identification method based on irrelevance constraint | |
CN114780767A (en) | Large-scale image retrieval method and system based on deep convolutional neural network | |
CN107220656A (en) | A kind of multiple labeling data classification method based on self-adaptive features dimensionality reduction | |
CN103942214B (en) | Natural image classification method and device on basis of multi-modal matrix filling | |
Sen et al. | Face recognition using deep convolutional network and one-shot learning | |
CN115995040A (en) | SAR image small sample target recognition method based on multi-scale network | |
CN104318271A (en) | Image classification method based on adaptability coding and geometrical smooth convergence | |
CN116796248A (en) | Forest health environment assessment system and method thereof | |
CN109034213A (en) | Hyperspectral image classification method and system based on joint entropy principle | |
CN111898704A (en) | Method and device for clustering content samples |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |