CN106326935A - Image classification method based on sparse nonlinear subspace migration - Google Patents
Image classification method based on sparse nonlinear subspace migration Download PDFInfo
- Publication number
- CN106326935A CN106326935A CN201610748644.8A CN201610748644A CN106326935A CN 106326935 A CN106326935 A CN 106326935A CN 201610748644 A CN201610748644 A CN 201610748644A CN 106326935 A CN106326935 A CN 106326935A
- Authority
- CN
- China
- Prior art keywords
- phi
- data
- formula
- subspace
- vector
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an image classification method based on sparse nonlinear subspace migration. The method comprises the steps of mapping data from an original space to a regenerated kernel Hilbert space via a kernel method, mapping target domain training data XT to a preset subspace via predefined fundamental transformation P in the kernel Hilbert space to obtain target data PXT, mapping source domain training data XS to the preset subspace via the fundamental transformation P to obtain P[XS, XT], transforming the source domain data P[XS, XT] via a spare matrix Z, and distributing the P[XS, XT] and the PXT in the preset subspace in a sharing mode. The method has the advantages of improving the migration accuracy of image data in the preset subspace and being applicable to transformation of nonlinear data.
Description
Technical field
The invention belongs to a kind of image classification method, be specifically related to a kind of image migrated based on sparse nonlinear subspace
Sorting technique.
Background technology
In general, the data collected when sensor does not occur drift are referred to as source domain, in use for some time
The data that sensor obtains after drifting about are referred to as aiming field.At image domains, any one attitude of object can be referred to as
Source domain, will be changed the another kind of attitude that (such as angle, light intensity, backcolor change etc.) obtain and be referred to as by external condition
Aiming field, the basic task of image classification is exactly that the similar object that can these be under different conditions searches out.
In order to describe simplicity, present invention provide that:
" S " represents source domain, and " T " represents aiming field;Source domain training data is designated asAiming field training data is remembered
ForWherein D represents dimension, NSAnd NTIt it is the number of training sample in two territories;WithRepresent minority labelling in aiming field and a large amount of unlabelled data;Represent source data and target
The luv space of data is mapped to the basis conversion with dimension d of latent space;Z represents XsAnd XTBetween sparse reconstruction square
Battle array;1nRepresent the non-zero column vector of a length of n;I representation unit matrix;‖·‖0Represent the number of a vectorial nonzero element
Mesh;‖·‖1Represent L1-norm;‖·‖FThe Frobenius norm of representing matrix;[X]iRepresent i-th row of X.In the present invention,
Matrix uses capitalization black matrix, and vector is small letter black matrix, and variable is italic.
M.Shao,D.Kit,and Y.Fu,“Generalized Transfer Subspace Learning Through
Low-Rank Constraint, " Int.J.Comput.Vis., vol.109, pp.74-93,2014. (M.Shao, D.Kit and
Y.Fu was published in the paper of volume 109 the 74-93 page in " international computer vision periodical " " under low-rank constraints in 2014
Broad sense transfer sub-space learning "), it is provided that a kind of subspace moving method (being called for short LTSL) represented based on low-rank, LTSL
The data transfer principle figure of method is as shown in Figure 1:
Source domain training data XSWith aiming field training data XTOne conversion base W of trained acquisition, source domain data and aiming field
Data converted base W be mapped to subspace (subspace refer to by projection, it is achieved high dimensional feature to the mapping of lower dimensional space,
Ensureing not lose information, this is a kind of conventional dimensionality reduction thought, when this sub spaces training data, can effectively prevent as far as possible
Over-fitting), in subspace, source data WXSBy low-rank Z so that and aiming field data WXTDistribution is consistent, the most similar warp
Cross mapping closer to, thus realize classification.
(such as, principal component analysis (PCA), linear discriminant analysis (LDA), local keeps the sub-space learning method existed
Projection (LPP) etc.), it is used to Learning Subspaces projection W, in subspace, source data WXSRepresent that (LRR) is used for by low-rank
Rebuild target data WXT(because of target data WXTIt is to exist, WXSRepresented by low-rank and approach).
LTSL solves subspace branch problem, by minimizing object function as follows and realize (that is: find one group of W, Z,
The value of E so that the value of this formula minimizes):
s.t.WTXT=WTXSZ+E,WTU2W=I
In formula, the order of rank (.) representing matrix, ‖ E ‖2,1Represent l2,1-norm, and λ1And λ2It it is regularization coefficient.Express
Formula ‖ E ‖2,1Being used to encourage the error of E to be classified as 0, therefore noise or abnormal source domain can be deleted during adapting to.But
What the minima of rank (Z) was the most permissible finds a reconstructed coefficients matrix by inferior grade structure.F(W,Xs) it is a broad sense
Sub-space learning function, it can be write as Tr (WTU1W), U1And U2Being selected by sub-space learning pattern, learning model is such as
PCA or LDA etc..W is a kind of sub-space learning representation, and learning model has PCA or LDA;Z is that low-rank represents.
When sub-space learning pattern is selected, W fixes, use non-precision augmented vector approach solve above formula with
Obtain Z, E.Based on low-rank method for expressing knowledge adaptability should be provided a more complete theory and spatial analysis, but only
Under conditions of meeting subspace independence and having adequate data, the performance of LTSL is good.It practice, subspace is independent
It is the most inappeasable with data volume abundance, so the accuracy of the subspace moving method represented based on low-rank is relatively low.
It addition, LTSL method is only applicable to the conversion of linear data, but nonlinear data can not be processed, and this
Method invention is just being aimed at nonlinear data.
Summary of the invention
The technical problem to be solved is just to provide a kind of image migrated based on sparse nonlinear subspace and divides
Class method, it can improve the accuracy that view data migrates in subspace, and can be suitably used for the conversion of nonlinear data.
The technical problem to be solved is realized by such technical scheme, and it includes following steps:
Step 1, input source domain sampleWith aiming field sample Calculate initial dataXS 1Be source domain first sample,It is the N of source domainSIndividual sample
Originally, XT 1Be aiming field first sample,It is the N of aiming fieldTIndividual sample;
Step 2, calculates KTAnd K
By formulaWithSolve KT and K respectively;It is that initial data X projects to
The data of Hilbert space;
Step 3, obtains the Eigenvalues Decomposition of K
By formula K=VSVTSolve the Eigenvalues Decomposition of K;V is characterized the matrix of vector composition, element on the diagonal of S
It is characterized the eigenvalue corresponding to vector;
Step 4, initialization feature vector Φ
By formula Φ :=V (:, v) take the characteristic vector corresponding to front d eigenvalue of maximum of V, in formula v be the front d of V
Big characteristic vector corresponding to eigenvalue;
Step 5, updates Z
Φ is fixed, uses ADMM method, update with the Z in following formula (8):
In formula (8), Z is sparse matrix, λ1Represent balance parameter,Represent a length of NS+NTNon-zero column vector,
Represent a length of NTNon-zero column vector;
Step 6, updates Φ
Z is fixed, the method using Eigenvalues Decomposition, updates with the Φ in following formula (13):
s.t.ΦTK Φ=I
Step 7, checks convergence situation
If not restraining, then repeat step 5 and step 6;If convergence, then export the value of Φ and Z;
Step 8, byP must be mapped*, calculate source domain data XSIt is mapped to value M of default subspaceS,
Step 9, utilizes and maps P*, calculate the target training data X of labellingTlIt is mapped to value M of default subspaceTl,
Represent non-linear XS, project to Hilbert space and beThen by sparse square
Battle array Z-direction aiming field conversion;
Step 10, calculates not labeled target detection data XTuIt is mapped to value M of default subspaceTu,
Step 11, training data [M based on labellingS,MTl] and the labelling of they correspondences, utilize a two norm canonicals young waiter in a wineshop or an inn
Method is taken advantage of to train a grader W;
Step 12, by calculating decision functionNot labeled target detection data are carried out classification differentiation.
The solution have the advantages that: improve the accuracy that view data migrates in default subspace;Due to the present invention
First pass through kernel method and data are mapped to the core Hilbert space of regeneration from luv space, therefore can be suitably used for non-linear number
According to conversion.
Accompanying drawing explanation
The accompanying drawing of the present invention is described as follows:
Fig. 1 is the data transfer principle figure of the LTSL of background technology;
Fig. 2 is the data transfer principle figure of the present invention;
Fig. 3 is source domain and the three-dimensional distribution map of aiming field of data verification;
Fig. 4 is that the data of Fig. 3 two dimension after the present invention changes presets Subspace Distribution figure;
Fig. 5 is the sample image of an experimenter;
Fig. 6 is that the present invention is applied to the pose adjustment result figure in expression of smiling in Fig. 5.
Detailed description of the invention
The invention will be further described with embodiment below in conjunction with the accompanying drawings:
The present invention first passes through kernel method and data is mapped to the core Hilbert space (RKHS) of regeneration from luv space,
Again by conversion as shown in Figure 2, at core Hilbert space, by a predefined basis conversion P, aiming field is trained
Data XTIt is mapped to default subspace by basis conversion P and obtains target data PXT, by source domain training data XSTurned by basis
Change P to be mapped to default subspace and obtain PXS, owing to the present invention is by source domain and aiming field data P [XS,XT] come target data
Rebuild, therefore P [XS,XT] it is collectively referred to as again source domain data, then utilize source domain data P [XS,XT] pass through sparse matrix
Z changes, with PXTDistribution is shared in default subspace.
The principle of the present invention is:
LSDT (linearly) conversion and the difference changed of NLSDT (non-linear) are: due to can not be linear in lower dimensional space
The point set of segmentation, during by the point set that is converted in higher dimensional space, it is possible to become linear separability, but higher dimensional space again can
Cause dimension disaster, the method therefore using core, first initial data is projected on Hilbert space, reduce dimension and just can keep away
Exempt from dimension disaster, i.e. X projects to(X is initial data,It is the data after X projects to Hilbert space).
First nuclear matrix is definedAndWherein κ
It is kernel function, uses gaussian kernel function,xi、xjFor original number
According to,It is respectively initial data xi、xjProject to the data of Hilbert space.
Source domain data P [XS,XT] changed by sparse matrix Z, thus with aiming field data PXTIn default subspace
Interior sharing is distributed, then sparse nonlinear subspace migrates the mathematical model optimized and is
In formula (1), T0It it is sparse level;For the target data of default subspace,For in advance
If the source domain data of subspace;Z is to make to preset source domain and the sparse matrix of aiming field data sharing distribution in subspace.
Because based on L0The optimization of norm is non-convex problem, and model the most of the present invention employs the convex of this specification and approaches, i.e. L1
Norm.
In order to make the P of acquisition be able to ensure that, projection does not make data be distorted, and can keep a lot of useful information, that
There is following expression,
Formula (1) is combined with formula (2) formula, finally can be expressed as:
In formula (3), the row of P must be orthogonal, and necessary normalization so that PPT=I, I are unit matrixs, with this
Avoid degenerating.Furthermore, it is assumed that(Represent a length of NS+NTNon-zero column vector,Represent a length of NT's
Non-zero column vector) so that source domain data and aiming field data are in united affine default subspace, and non-linear default subspace
(affine default subspace and the linear difference presetting subspace are whether there is selected initial point.In affine default subspace
The status of any 2 is of equal value, and the initial point linearly presetting subspace is a special point), λ1And λ2Then represent balance parameter (i.e.
The regularization coefficient λ of background technology1And λ2, illustrate the weight of each).
To put it more simply, orderSo object function (3) can be written as:
In object function (3) or (4), the solution of P has a lot, i.e. solution is the most unique.A kind of method for solving of basis conversion P is such as
Under:
There is optimal solution P*, it can be expressed as intuitivelyIn the case of, initial data and number of targets
According to linear combination, representation is as follows:
In formula (5), φ is characterized vector, φT、Be respectively φ andTransposed matrix.
Convolution (4) and formula (5), object function can be expressed as:
OrderSo previously described formula (3) can be expressed as follows:
From formula (7) it can be seen that formula relates to two variable Z and Φ.In order to obtain the minimal solution of formula (7), can use
Fix a variable, and change another variable to solve this problem.Therefore, including following 2 key steps:
The first, sparse matrix (Z) is updated
In order to obtain Z, Φ can be fixed, then formula (7) can be expressed as:
This is typical sparse lasso trick (Lasso) optimization problem, and alternating direction multiplier method (ADMM) can be used to have
Effect ground solves.
Introducing auxiliary variable L, formula (8) can be rewritten as:
The enhancing Lagrangian of formula (9) can be expressed as:
In formula (10), YAAnd YBRepresent Lagrange multiplier, λ respectively1=μ1/2;λ1And μ1It is coefficient variation, Ke Yisui
Meaning selects, and takes λ1=μ1/ 2 are intended merely to convenience when that derivation obtaining (11) formula, and (two norm derivations just have 2 and 1/2 the most about
Fall);μ2Coefficient when being to solve with augmented vector approach.
For understanding the L of formula (9), now fix Z, and make the local derviation of formula (10)It it is 0, it is possible to obtain the expression formula of L
As follows:
Fixing L, then Z can be solved by following formula:
Therefore, using ADMM method, the step solving Z according to formula (8) is as follows:
Step 1), initiation parameter
Make Z, YA,YBIt is 0, μ2For λ1, ρ=1.1 and maxμ=106;Wherein YAAnd YBRepresent Lagrange multiplier respectively,
λ1And λ2Then represent balance parameter;
Step 2), update L
Z is fixed, by utilizing following formula that L is solved when not restraining when:
In formula, L is auxiliary variable, λ1=μ1/ 2, μ2Coefficient when being to solve with augmented vector approach;
Step 3), update Z
When not restraining when, fix L, by following formula, Z solved:
Step 4), update Lagrange multiplier
According to step 2) and step 3) more newly obtained L and Z, substitute into following formula,
YA=YA+μ2(L-Z),YB=YB+μ2(1TL-1T)
Multiplier after being updated;
In above-mentioned formula (10), YAAnd YBRepresent Lagrange multiplier respectively, owing to this is a typical sparse lasso trick optimization
Problem, therefore uses alternating direction multiplier method (ADMM), because formula (10) has Lagrange multiplier, therefore title is called again multiplication
Device;
Step 5), undated parameter μ2
μ2Take μ2ρ and maxμIn minima (μ2ρ and maxμBeing the value set, in step 1, initiation parameter part is the most supplementary
Their setting value);
Step 6), check convergence situation
If not restraining, then repeat step 2) to step 5);
If convergence, then export Z.
The second, Φ is updated
In order to solve Φ, by fixing Z, formula (7) can be written as:
The object function of formula (13) is expanded as follows:
In formula (14),
Make ΦTK Φ=I, then Φ ΦTKΦΦT=QKQT=Φ ΦT=Q, then object function can be reduced to:
Tr((λ1(K-1KT-Z)(K-1KT-Z)T-λ2I)KTQTK) (15)
According to formula K=VSVTEigenvalues Decomposition, wherein V is characterized vector, on S diagonal element be characterized vector right
The eigenvalue answered:
In formula,Characteristic function (15) dissolve for:
In formula (16),
And ΩTΩ=ΦTVSVTΦ=ΦTK Φ=I
Finally, formula (13) is rewritten as:
Optimal solution Ω*It is to be obtained by characteristic vector l of the minimal eigenvalue about Θ.
Ω*It is to pass throughAnd VVT=I is tried to achieve, Φ*Optimum can be represented as:
Therefore, the step of Φ is solved according to formula (13) as follows:
Step (1), seeks the Eigenvalues Decomposition of K
By K=VSVTObtain the Eigenvalues Decomposition of K;Wherein K=XTX, V are characterized vector, and on S diagonal, element is characterized
Vector characteristic of correspondence value;
Step (2), is solved Θ by following formula:
Step (3), obtains the Eigenvalues Decomposition of Θ
By formula Θ=U Σ UTSolving the Eigenvalues Decomposition of Θ, U is the characteristic vector of Θ, and Σ is a diagonal matrix, and it is right
Element on angle is characterized vector characteristic of correspondence value;
Step (4), solves Ω
Solve Ω by Ω=U (:, U), this formula be Θ is carried out feature decomposition after, take front d the minimal eigenvalue of U
Characteristic of correspondence vector, in formula, U is the characteristic vector corresponding to front d minimal eigenvalue of U;
The scope of d is all eigenvalue numbers of 1~U, circulates successively with eigenvalues all in U, and treating excess syndrome is tested effective
Those values;
Step (5), solves Φ
By formulaSolve Φ.
Summary content, obtains the image classification method migrated based on sparse nonlinear subspace of the present invention, including with
Lower step:
Step 1, input source domain sampleWith aiming field sample CalculateXS 1Be source domain first sample,It is the N of source domainSIndividual sample, XT 1It it is mesh
Mark territory first sample,It is the N of aiming fieldTIndividual sample;
Step 2, calculates KTAnd K
By formulaWithSolve KT and K respectively;It is that initial data X projects to
The data of Hilbert space;
Step 3, obtains the Eigenvalues Decomposition of K
By formula K=VSVTSolve the Eigenvalues Decomposition of K;V is characterized the matrix that vector is formed, element on S diagonal
It is characterized vector characteristic of correspondence value;
Step 4, initializes Φ
By formula Φ :=V (:, v) take the characteristic vector corresponding to front d eigenvalue of maximum of V, in formula v be the front d of V
Big characteristic vector corresponding to eigenvalue;
The scope of d is 1~(all eigenvalue numbers of V), circulates successively with eigenvalues all in V, and treating excess syndrome is tested effective
Those values;
Step 5, updates Z
Φ is fixed, uses ADMM method, update with the Z in following formula (8):
Step 6, updates Φ
Z is fixed, the method using Eigenvalues Decomposition, updates with the Φ in following formula (13):
s.t.ΦTK Φ=I
Step 7, checks convergence situation
If not restraining, then repeat step 5 and step 6;If convergence, then export the value of Φ and Z;
Step 8, byP must be mapped*, calculate source domain data XSIt is mapped to value M of default subspaceS,
Step 9, utilizes and maps P*, calculate the target training data X of labellingTlIt is mapped to value M of default subspaceTl,
Represent non-linear XS, project to Hilbert space and beBy sparse after P
Matrix Z-direction aiming field is changed;
Step 10, calculates not labeled target detection data XTuIt is mapped to value M of default subspaceTu,
The target training data X of labelling is setTlNot labeled target detection data XTu, it is because source domain to be passed through
Data XSAiming field data X with labellingTlTrain a grader W, then use not labeled aiming field data XTuEnter
Row test, it is judged that the quality of the grader W that step 11 trains, to check the quality of the present invention;
Step 11, training data [M based on labellingS,MTl] and the labelling of they correspondences, utilize a two norm canonicals young waiter in a wineshop or an inn
Method is taken advantage of to train a grader W;
Step 12, by calculating decision functionNot labeled target detection data are carried out classification differentiation.
Experimental verification
One, data verification
As it is shown on figure 3, for data produced by the array of default subspace, by the Gauss distribution of different modes and association side
Difference matrix, it is thus achieved that (target data is random producing for the target data of minority labelling and majority unlabelled target data distributed in three dimensions
The raw data meeting Gauss distribution), source domain has two class data, every class packet contains 50 samples, and easily in source domain
Find the decision boundary of two class data;In aiming field, also have two class data, every class have the sample of 5 labellings and 50 unmarked
Sample.From Fig. 3 it will be clear that: 1, of a sort data point has different distributions in source domain and aiming field;2、
The Optimal Separating Hyperplane of source domain is not suitable for the decision boundary (decision boundary refers to separate for inhomogeneity data line or plane) of aiming field.
The sparse Z subspace utilizing the present invention migrates, and after projection and rebuilding, obtains the source data of two-dimentional default subspace
And target data, as shown in Figure 4, the spatial offset rate between source data and target data decreases, and between this two class
Decision boundary clear, it is easy to finding a kind of grader to realize classification, so using present method invention to carry out data classification being
Feasible, effective.
Two, object identification
CMU Multi-PIE human face data collection is a public data collection, comprises the comprehensive face number of 337 experimenters altogether
According to collection, image therein passes through 15 postures, 20 kinds of illuminations, 6 kinds of expression, 4 realization captures of different stages.In order to be thought
The result wanted, in an experiment 60 experimenters before stage 1 and stage 2 have selected respectively, 7 kinds of expression figures of each experimenter
Picture.As shown in Figure 5: the stage 1 is the first row, face is with general expression;Stage 2 is the second row, and experimenter is also with smile.?
In this experiment, four kinds of Setup Experiments are as follows:
Arrange 1: the red rectangle front expression of each experimenter and the extreme attitude of blue rectangle (60 ° of angles) are respectively intended to
As source domain and the training data of aiming field, remaining expression is i.e. used as probe face.
Arrange 2: with reference to arranging 1, similarly configure.
1+2 is set: express one's feelings the two kinds of front expressions arranged in 1 and 2 and two kinds of extreme 60 ° of attitudes as training number
According to, remaining expression is used as probe face.
Arranged in a crossed manner: typically to express one's feelings arrange each experimenter in 1 as source domain, will arrange 2 each experimenters'
Expression of smiling, as aiming field, adapts to the change of expression with this.
Nonlinearity due to three-dimensional rotation face so that attitude tracking is extremely challenging.Fig. 6 is that the present invention is applied to
Stage 2 smile expression on pose adjustment result figure, it using in Fig. 5 front express one's feelings as source data, the extreme attitude in back
(60 ° of angles) is used as target data.We can see that from Fig. 6: owing to remaining the removal of (noise), target face alignment appearance
State is good.Using the invention of this nonlinear method to carry out recognition of face, its recognition accuracy compares with LTSL, and it the results are shown in Table 1.
Table 1. is for the recognition accuracy ratio of human face posture identification
Task | Source domain | Aiming field | LTSL-PCA | LTSL-LDA | The present invention |
Arrange 1 | Front | The angle of 60 degree | 55.7 | 56.0 | 63.7 |
Arrange 2 | Front | The angle of 60 degree | 58.7 | 60.7 | 70.7 |
1+2 is set | Front | The angle of 60 degree | 57.8 | 60.7 | 67.5 |
Arranged in a crossed manner | Arrange 1 | Arrange 2 | 96.7 | 96.7 | 99.4 |
This method is nonlinear method, owing to initial data is non-linear, is projected on Hilbert space carrying out again
Process, thus greatly improve recognition accuracy.By contrast linear method LTSL, as can be seen from Table 1: present method invention ratio
The recognition effect of linear method LTSL is good.
Claims (3)
1. the image classification method migrated based on sparse nonlinear subspace, is characterized in that, comprises the steps of
Step 1, input source domain sampleWith aiming field sample Calculate initial dataXS 1Be source domain first sample,It is the N of source domainSIndividual sample
Originally, XT 1Be aiming field first sample,It is the N of aiming fieldTIndividual sample;
Step 2, calculates KTAnd K
By formulaWithSolve KT and K respectively;It is that initial data X projects to Xi Er
The data in Bert space;
Step 3, obtains the Eigenvalues Decomposition of K
By formula K=VSVTSolve the Eigenvalues Decomposition of K;V is characterized the matrix of vector composition, and on the diagonal of S, element is characterized
Eigenvalue corresponding to vector;
Step 4, initialization feature vector Φ
By formulaTaking the characteristic vector corresponding to front d eigenvalue of maximum of V, in formula, v is front d the maximum feature of V
Characteristic vector corresponding to value;
Step 5, updates Z
Φ is fixed, uses ADMM method, update with the Z in following formula (8):
In formula (8), Z is sparse matrix, λ1Represent balance parameter,Represent a length of NS+NTNon-zero column vector,Represent
A length of NTNon-zero column vector;
Step 6, updates Φ
Z is fixed, the method using Eigenvalues Decomposition, updates with the Φ in following formula (13):
s.t.ΦTK Φ=I
Step 7, checks convergence situation
If not restraining, then repeat step 5 and step 6;If convergence, then export the value of Φ and Z;
Step 8, byP must be mapped*, calculate source domain data XSIt is mapped to value M of default subspaceS,
Step 9, utilizes and maps P*, calculate the target training data X of labellingTlIt is mapped to value M of default subspaceTl,
Represent non-linear XS, project to Hilbert space and beThen by sparse matrix Z
Change to aiming field;
Step 10, calculates not labeled target detection data XTuIt is mapped to value M of default subspaceTu,
Step 11, training data [M based on labellingS,MTl] and the labelling of they correspondences, utilize two norm canonical least squares
Method one grader W of training;
Step 12, by calculating decision functionNot labeled target detection data are carried out classification differentiation.
The image classification method migrated based on sparse nonlinear subspace the most according to claim 1, is characterized in that, in step
Rapid 5 steps updating Z include:
Step 1), initiation parameter
Make Z, YA,YBIt is 0, μ2For λ1, ρ=1.1 and maxμ=106;Wherein YAAnd YBRepresent Lagrange multiplier, λ respectively1And λ2
Then represent balance parameter;
Step 2), update L
Z is fixed, by utilizing following formula that L is solved when not restraining when:
In formula, L is auxiliary variable, λ1=μ1/ 2, μ2Coefficient when being to solve with augmented vector approach;
Step 3), update Z
When not restraining when, fix L, by following formula, Z solved:
Step 4), update Lagrange multiplier
According to step 2) and step 3) more newly obtained L and Z, substitute into following formula,
YA=YA+μ2(L-Z),YB=YB+μ2(1TL-1T)
Multiplier after being updated;
Step 5), undated parameter μ2
μ2Take μ2ρ and maxμIn minima;
Step 6), check convergence situation
If not restraining, then repeat step 2) to step 5);
If convergence, then export Z.
The image classification method migrated based on sparse nonlinear subspace the most according to claim 1 and 2, is characterized in that,
The step updating Φ in step 6 includes:
Step (1), seeks the Eigenvalues Decomposition of K
By K=VSVTObtain the Eigenvalues Decomposition of K;Wherein K=XTX, V are characterized vector, and on S diagonal, element is characterized vector
Characteristic of correspondence value;
Step (2), is solved Θ by following formula:
Step (3), obtains the Eigenvalues Decomposition of Θ
By formula Θ=U Σ UTSolving the Eigenvalues Decomposition of Θ, U is the characteristic vector of Θ, and Σ is a diagonal matrix, on its diagonal angle
Element be characterized vector characteristic of correspondence value;
Step (4), solves Ω
Solve Ω by Ω=U (:, U), this formula be Θ is carried out feature decomposition after, front d the minimal eigenvalue taking U is corresponding
Characteristic vector, in formula, U is the characteristic vector corresponding to front d minimal eigenvalue of U;
Step (5), solves Φ
By formulaSolve Φ.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610748644.8A CN106326935A (en) | 2016-08-29 | 2016-08-29 | Image classification method based on sparse nonlinear subspace migration |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610748644.8A CN106326935A (en) | 2016-08-29 | 2016-08-29 | Image classification method based on sparse nonlinear subspace migration |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106326935A true CN106326935A (en) | 2017-01-11 |
Family
ID=57788252
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610748644.8A Pending CN106326935A (en) | 2016-08-29 | 2016-08-29 | Image classification method based on sparse nonlinear subspace migration |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106326935A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107729914A (en) * | 2017-09-06 | 2018-02-23 | 鲁小杰 | A kind of detection method of pathological data |
CN108798641A (en) * | 2018-06-19 | 2018-11-13 | 东北大学 | A kind of Diagnosing The Faults of Sucker Rod Pumping System method based on subspace transfer learning |
CN108875928A (en) * | 2017-05-15 | 2018-11-23 | 广东石油化工学院 | Multi-output regression network and learning method |
CN110146655A (en) * | 2019-05-31 | 2019-08-20 | 重庆大学 | A kind of anti-electronic nose bleach-out process based on adaptive sub-space learning normal form |
CN111461191A (en) * | 2020-03-25 | 2020-07-28 | 杭州跨视科技有限公司 | Method and device for determining image sample set for model training and electronic equipment |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101079103A (en) * | 2007-06-14 | 2007-11-28 | 上海交通大学 | Human face posture identification method based on sparse Bayesian regression |
CN104392243A (en) * | 2014-11-18 | 2015-03-04 | 西北工业大学 | Nonlinear un-mixing method of hyperspectral images based on kernel sparse nonnegative matrix decomposition |
-
2016
- 2016-08-29 CN CN201610748644.8A patent/CN106326935A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101079103A (en) * | 2007-06-14 | 2007-11-28 | 上海交通大学 | Human face posture identification method based on sparse Bayesian regression |
CN104392243A (en) * | 2014-11-18 | 2015-03-04 | 西北工业大学 | Nonlinear un-mixing method of hyperspectral images based on kernel sparse nonnegative matrix decomposition |
Non-Patent Citations (1)
Title |
---|
LEI ZHANG等: "LSDT: Latent Sparse Domain Transfer Learning for Visual Adaptation", 《IEEE TRANSACTIONS ON I《IEEE TRANSACTIONS ON IMAGE PROCESSING》MAGE PROCESSING》 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108875928A (en) * | 2017-05-15 | 2018-11-23 | 广东石油化工学院 | Multi-output regression network and learning method |
CN108875928B (en) * | 2017-05-15 | 2021-02-26 | 广东石油化工学院 | Multi-output regression network and learning method |
CN107729914A (en) * | 2017-09-06 | 2018-02-23 | 鲁小杰 | A kind of detection method of pathological data |
CN107729914B (en) * | 2017-09-06 | 2021-06-25 | 鲁小杰 | Pathological data detection method |
CN108798641A (en) * | 2018-06-19 | 2018-11-13 | 东北大学 | A kind of Diagnosing The Faults of Sucker Rod Pumping System method based on subspace transfer learning |
CN110146655A (en) * | 2019-05-31 | 2019-08-20 | 重庆大学 | A kind of anti-electronic nose bleach-out process based on adaptive sub-space learning normal form |
CN110146655B (en) * | 2019-05-31 | 2021-12-21 | 重庆大学 | Electronic nose drift resistance method based on self-adaptive subspace learning paradigm |
CN111461191A (en) * | 2020-03-25 | 2020-07-28 | 杭州跨视科技有限公司 | Method and device for determining image sample set for model training and electronic equipment |
CN111461191B (en) * | 2020-03-25 | 2024-01-23 | 杭州跨视科技有限公司 | Method and device for determining image sample set for model training and electronic equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106326935A (en) | Image classification method based on sparse nonlinear subspace migration | |
CN103632138B (en) | A kind of face identification method of low-rank piecemeal rarefaction representation | |
Zhang et al. | Robust low-rank kernel multi-view subspace clustering based on the schatten p-norm and correntropy | |
Yang et al. | Kernel ICA: An alternative formulation and its application to face recognition | |
Zhou et al. | Bayesian low-tubal-rank robust tensor factorization with multi-rank determination | |
Allen et al. | Capturing inter-subject variability with group independent component analysis of fMRI data: a simulation study | |
Małek et al. | The VIMOS Public Extragalactic Redshift Survey (VIPERS)-A support vector machine classification of galaxies, stars, and AGNs | |
CN111275192B (en) | Auxiliary training method for improving accuracy and robustness of neural network simultaneously | |
CN105405098A (en) | Image super-resolution reconstruction method based on sparse representation and adaptive filtering | |
CN104809471B (en) | A kind of high spectrum image residual error integrated classification method based on spatial spectral information | |
Sohl-Dickstein et al. | Hamiltonian annealed importance sampling for partition function estimation | |
CN104751176A (en) | Method for selecting high-spectrum remote-sensing image wave band | |
CN105929113A (en) | Electronic nose signal error adaptive-learning method with subspace projection | |
Galvin et al. | Radio galaxy zoo: Knowledge transfer using rotationally invariant self-organizing maps | |
Boroson et al. | Exploring the spectral space of low redshift QSOs | |
CN103745233A (en) | Hyper-spectral image classifying method based on spatial information transfer | |
Finnveden et al. | Understanding when spatial transformer networks do not support invariance, and what to do about it | |
Sun et al. | Mixed noise removal for hyperspectral images based on global tensor low-rankness and nonlocal SVD-aided group sparsity | |
CN104050489B (en) | SAR ATR method based on multicore optimization | |
Kalaganis et al. | Revisiting Riemannian geometry-based EEG decoding through approximate joint diagonalization | |
CN104050482B (en) | A kind of manifold learning generalization algorithm based on local linear smoothing | |
CN106326936A (en) | Image classification method based on sparse preset subspace migration | |
CN106960225A (en) | A kind of sparse image classification method supervised based on low-rank | |
CN109815889A (en) | A kind of across resolution ratio face identification method based on character representation collection | |
Qi et al. | A novel joint dictionary framework for sparse hyperspectral unmixing incorporating spectral library |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20170111 |
|
WD01 | Invention patent application deemed withdrawn after publication |