CN106326935A - Image classification method based on sparse nonlinear subspace migration - Google Patents

Image classification method based on sparse nonlinear subspace migration Download PDF

Info

Publication number
CN106326935A
CN106326935A CN201610748644.8A CN201610748644A CN106326935A CN 106326935 A CN106326935 A CN 106326935A CN 201610748644 A CN201610748644 A CN 201610748644A CN 106326935 A CN106326935 A CN 106326935A
Authority
CN
China
Prior art keywords
phi
data
formula
subspace
vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610748644.8A
Other languages
Chinese (zh)
Inventor
张磊
邓平聆
段青言
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University
Original Assignee
Chongqing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University filed Critical Chongqing University
Priority to CN201610748644.8A priority Critical patent/CN106326935A/en
Publication of CN106326935A publication Critical patent/CN106326935A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an image classification method based on sparse nonlinear subspace migration. The method comprises the steps of mapping data from an original space to a regenerated kernel Hilbert space via a kernel method, mapping target domain training data XT to a preset subspace via predefined fundamental transformation P in the kernel Hilbert space to obtain target data PXT, mapping source domain training data XS to the preset subspace via the fundamental transformation P to obtain P[XS, XT], transforming the source domain data P[XS, XT] via a spare matrix Z, and distributing the P[XS, XT] and the PXT in the preset subspace in a sharing mode. The method has the advantages of improving the migration accuracy of image data in the preset subspace and being applicable to transformation of nonlinear data.

Description

A kind of image classification method migrated based on sparse nonlinear subspace
Technical field
The invention belongs to a kind of image classification method, be specifically related to a kind of image migrated based on sparse nonlinear subspace Sorting technique.
Background technology
In general, the data collected when sensor does not occur drift are referred to as source domain, in use for some time The data that sensor obtains after drifting about are referred to as aiming field.At image domains, any one attitude of object can be referred to as Source domain, will be changed the another kind of attitude that (such as angle, light intensity, backcolor change etc.) obtain and be referred to as by external condition Aiming field, the basic task of image classification is exactly that the similar object that can these be under different conditions searches out.
In order to describe simplicity, present invention provide that:
" S " represents source domain, and " T " represents aiming field;Source domain training data is designated asAiming field training data is remembered ForWherein D represents dimension, NSAnd NTIt it is the number of training sample in two territories;WithRepresent minority labelling in aiming field and a large amount of unlabelled data;Represent source data and target The luv space of data is mapped to the basis conversion with dimension d of latent space;Z represents XsAnd XTBetween sparse reconstruction square Battle array;1nRepresent the non-zero column vector of a length of n;I representation unit matrix;‖·‖0Represent the number of a vectorial nonzero element Mesh;‖·‖1Represent L1-norm;‖·‖FThe Frobenius norm of representing matrix;[X]iRepresent i-th row of X.In the present invention, Matrix uses capitalization black matrix, and vector is small letter black matrix, and variable is italic.
M.Shao,D.Kit,and Y.Fu,“Generalized Transfer Subspace Learning Through Low-Rank Constraint, " Int.J.Comput.Vis., vol.109, pp.74-93,2014. (M.Shao, D.Kit and Y.Fu was published in the paper of volume 109 the 74-93 page in " international computer vision periodical " " under low-rank constraints in 2014 Broad sense transfer sub-space learning "), it is provided that a kind of subspace moving method (being called for short LTSL) represented based on low-rank, LTSL The data transfer principle figure of method is as shown in Figure 1:
Source domain training data XSWith aiming field training data XTOne conversion base W of trained acquisition, source domain data and aiming field Data converted base W be mapped to subspace (subspace refer to by projection, it is achieved high dimensional feature to the mapping of lower dimensional space, Ensureing not lose information, this is a kind of conventional dimensionality reduction thought, when this sub spaces training data, can effectively prevent as far as possible Over-fitting), in subspace, source data WXSBy low-rank Z so that and aiming field data WXTDistribution is consistent, the most similar warp Cross mapping closer to, thus realize classification.
(such as, principal component analysis (PCA), linear discriminant analysis (LDA), local keeps the sub-space learning method existed Projection (LPP) etc.), it is used to Learning Subspaces projection W, in subspace, source data WXSRepresent that (LRR) is used for by low-rank Rebuild target data WXT(because of target data WXTIt is to exist, WXSRepresented by low-rank and approach).
LTSL solves subspace branch problem, by minimizing object function as follows and realize (that is: find one group of W, Z, The value of E so that the value of this formula minimizes):
s.t.WTXT=WTXSZ+E,WTU2W=I
In formula, the order of rank (.) representing matrix, ‖ E ‖2,1Represent l2,1-norm, and λ1And λ2It it is regularization coefficient.Express Formula ‖ E ‖2,1Being used to encourage the error of E to be classified as 0, therefore noise or abnormal source domain can be deleted during adapting to.But What the minima of rank (Z) was the most permissible finds a reconstructed coefficients matrix by inferior grade structure.F(W,Xs) it is a broad sense Sub-space learning function, it can be write as Tr (WTU1W), U1And U2Being selected by sub-space learning pattern, learning model is such as PCA or LDA etc..W is a kind of sub-space learning representation, and learning model has PCA or LDA;Z is that low-rank represents.
When sub-space learning pattern is selected, W fixes, use non-precision augmented vector approach solve above formula with Obtain Z, E.Based on low-rank method for expressing knowledge adaptability should be provided a more complete theory and spatial analysis, but only Under conditions of meeting subspace independence and having adequate data, the performance of LTSL is good.It practice, subspace is independent It is the most inappeasable with data volume abundance, so the accuracy of the subspace moving method represented based on low-rank is relatively low.
It addition, LTSL method is only applicable to the conversion of linear data, but nonlinear data can not be processed, and this Method invention is just being aimed at nonlinear data.
Summary of the invention
The technical problem to be solved is just to provide a kind of image migrated based on sparse nonlinear subspace and divides Class method, it can improve the accuracy that view data migrates in subspace, and can be suitably used for the conversion of nonlinear data.
The technical problem to be solved is realized by such technical scheme, and it includes following steps:
Step 1, input source domain sampleWith aiming field sample Calculate initial dataXS 1Be source domain first sample,It is the N of source domainSIndividual sample Originally, XT 1Be aiming field first sample,It is the N of aiming fieldTIndividual sample;
Step 2, calculates KTAnd K
By formulaWithSolve KT and K respectively;It is that initial data X projects to The data of Hilbert space;
Step 3, obtains the Eigenvalues Decomposition of K
By formula K=VSVTSolve the Eigenvalues Decomposition of K;V is characterized the matrix of vector composition, element on the diagonal of S It is characterized the eigenvalue corresponding to vector;
Step 4, initialization feature vector Φ
By formula Φ :=V (:, v) take the characteristic vector corresponding to front d eigenvalue of maximum of V, in formula v be the front d of V Big characteristic vector corresponding to eigenvalue;
Step 5, updates Z
Φ is fixed, uses ADMM method, update with the Z in following formula (8):
min Z | | Z | | 1 + λ 1 | | Φ T K T - Φ T K Z | | F 2 - - - ( 8 )
s . t .1 N S + N T T Z = 1 N T T
In formula (8), Z is sparse matrix, λ1Represent balance parameter,Represent a length of NS+NTNon-zero column vector, Represent a length of NTNon-zero column vector;
Step 6, updates Φ
Z is fixed, the method using Eigenvalues Decomposition, updates with the Φ in following formula (13):
min Φ λ 1 | | Φ T K T - Φ T K Z | | F 2 + λ 2 T r ( ( I - ΦΦ T K ) T K ( I - ΦΦ T K ) ) - - - ( 13 )
s.t.ΦTK Φ=I
Step 7, checks convergence situation
If not restraining, then repeat step 5 and step 6;If convergence, then export the value of Φ and Z;
Step 8, byP must be mapped*, calculate source domain data XSIt is mapped to value M of default subspaceS,
Step 9, utilizes and maps P*, calculate the target training data X of labellingTlIt is mapped to value M of default subspaceTl,
Represent non-linear XS, project to Hilbert space and beThen by sparse square Battle array Z-direction aiming field conversion;
Step 10, calculates not labeled target detection data XTuIt is mapped to value M of default subspaceTu,
Step 11, training data [M based on labellingS,MTl] and the labelling of they correspondences, utilize a two norm canonicals young waiter in a wineshop or an inn Method is taken advantage of to train a grader W;
Step 12, by calculating decision functionNot labeled target detection data are carried out classification differentiation.
The solution have the advantages that: improve the accuracy that view data migrates in default subspace;Due to the present invention First pass through kernel method and data are mapped to the core Hilbert space of regeneration from luv space, therefore can be suitably used for non-linear number According to conversion.
Accompanying drawing explanation
The accompanying drawing of the present invention is described as follows:
Fig. 1 is the data transfer principle figure of the LTSL of background technology;
Fig. 2 is the data transfer principle figure of the present invention;
Fig. 3 is source domain and the three-dimensional distribution map of aiming field of data verification;
Fig. 4 is that the data of Fig. 3 two dimension after the present invention changes presets Subspace Distribution figure;
Fig. 5 is the sample image of an experimenter;
Fig. 6 is that the present invention is applied to the pose adjustment result figure in expression of smiling in Fig. 5.
Detailed description of the invention
The invention will be further described with embodiment below in conjunction with the accompanying drawings:
The present invention first passes through kernel method and data is mapped to the core Hilbert space (RKHS) of regeneration from luv space, Again by conversion as shown in Figure 2, at core Hilbert space, by a predefined basis conversion P, aiming field is trained Data XTIt is mapped to default subspace by basis conversion P and obtains target data PXT, by source domain training data XSTurned by basis Change P to be mapped to default subspace and obtain PXS, owing to the present invention is by source domain and aiming field data P [XS,XT] come target data Rebuild, therefore P [XS,XT] it is collectively referred to as again source domain data, then utilize source domain data P [XS,XT] pass through sparse matrix Z changes, with PXTDistribution is shared in default subspace.
The principle of the present invention is:
LSDT (linearly) conversion and the difference changed of NLSDT (non-linear) are: due to can not be linear in lower dimensional space The point set of segmentation, during by the point set that is converted in higher dimensional space, it is possible to become linear separability, but higher dimensional space again can Cause dimension disaster, the method therefore using core, first initial data is projected on Hilbert space, reduce dimension and just can keep away Exempt from dimension disaster, i.e. X projects to(X is initial data,It is the data after X projects to Hilbert space).
First nuclear matrix is definedAndWherein κ It is kernel function, uses gaussian kernel function,xi、xjFor original number According to,It is respectively initial data xi、xjProject to the data of Hilbert space.
Source domain data P [XS,XT] changed by sparse matrix Z, thus with aiming field data PXTIn default subspace Interior sharing is distributed, then sparse nonlinear subspace migrates the mathematical model optimized and is
In formula (1), T0It it is sparse level;For the target data of default subspace,For in advance If the source domain data of subspace;Z is to make to preset source domain and the sparse matrix of aiming field data sharing distribution in subspace.
Because based on L0The optimization of norm is non-convex problem, and model the most of the present invention employs the convex of this specification and approaches, i.e. L1 Norm.
In order to make the P of acquisition be able to ensure that, projection does not make data be distorted, and can keep a lot of useful information, that There is following expression,
Formula (1) is combined with formula (2) formula, finally can be expressed as:
In formula (3), the row of P must be orthogonal, and necessary normalization so that PPT=I, I are unit matrixs, with this Avoid degenerating.Furthermore, it is assumed that(Represent a length of NS+NTNon-zero column vector,Represent a length of NT's Non-zero column vector) so that source domain data and aiming field data are in united affine default subspace, and non-linear default subspace (affine default subspace and the linear difference presetting subspace are whether there is selected initial point.In affine default subspace The status of any 2 is of equal value, and the initial point linearly presetting subspace is a special point), λ1And λ2Then represent balance parameter (i.e. The regularization coefficient λ of background technology1And λ2, illustrate the weight of each).
To put it more simply, orderSo object function (3) can be written as:
In object function (3) or (4), the solution of P has a lot, i.e. solution is the most unique.A kind of method for solving of basis conversion P is such as Under:
There is optimal solution P*, it can be expressed as intuitivelyIn the case of, initial data and number of targets According to linear combination, representation is as follows:
In formula (5), φ is characterized vector, φTBe respectively φ andTransposed matrix.
Convolution (4) and formula (5), object function can be expressed as:
OrderSo previously described formula (3) can be expressed as follows:
min Z , Φ | | z | | 1 + λ 1 | | Φ T K T - Φ T K Z | | F 2 + λ 2 T r ( ( I - ΦΦ T K ) T K ( I - ΦΦ T K ) ) - - - ( 7 )
s . t . Φ T K Φ = I , 1 N S + N T T Z = 1 N T T
From formula (7) it can be seen that formula relates to two variable Z and Φ.In order to obtain the minimal solution of formula (7), can use Fix a variable, and change another variable to solve this problem.Therefore, including following 2 key steps:
The first, sparse matrix (Z) is updated
In order to obtain Z, Φ can be fixed, then formula (7) can be expressed as:
min Z | | Z | | 1 + λ 1 | | Φ T K T - Φ T K Z | | F 2 s . t .1 N S + N T T Z = 1 N T T - - - ( 8 )
This is typical sparse lasso trick (Lasso) optimization problem, and alternating direction multiplier method (ADMM) can be used to have Effect ground solves.
Introducing auxiliary variable L, formula (8) can be rewritten as:
min Z , L | | Z | | 1 + λ 1 | | Φ T K T - Φ T K L | | F 2 s . t . L = Z , 1 N S + N T T L = 1 N T T - - - ( 9 )
The enhancing Lagrangian of formula (9) can be expressed as:
J ( Z , L ) = | | Z | | 1 + &mu; 1 2 | | &Phi; T K T - &Phi; T K L | | F 2 + < Y A , L - Z > + < Y B , 1 N S + N T T L - 1 N T T > + &mu; 1 2 ( | | L - Z | | F 2 + | | 1 N S + N T T L - 1 N T T | | 2 2 ) - - - ( 10 )
In formula (10), YAAnd YBRepresent Lagrange multiplier, λ respectively11/2;λ1And μ1It is coefficient variation, Ke Yisui Meaning selects, and takes λ11/ 2 are intended merely to convenience when that derivation obtaining (11) formula, and (two norm derivations just have 2 and 1/2 the most about Fall);μ2Coefficient when being to solve with augmented vector approach.
For understanding the L of formula (9), now fix Z, and make the local derviation of formula (10)It it is 0, it is possible to obtain the expression formula of L As follows:
L = ( &mu; 1 K T &Phi;&Phi; T K + &mu; 2 I + &mu; 2 1 N S + N T 1 N S + N T T ) - 1 ( &mu; 1 K T &Phi;&Phi; T K T - Y A - 1 N S + N T Y B + &mu; 2 Z + &mu; 2 1 N S + N T 1 N T T ) - - - ( 11 )
Fixing L, then Z can be solved by following formula:
min Z | | Z | | 1 + &mu; 2 2 | | L - Z | | F 2 + < Y A , L - Z > &Proportional; min Z | | Z | | 1 + &mu; 2 2 ( | | L - Z | | F 2 + &mu; 2 2 < Y A , L - Z > ) &Proportional; min Z | | Z | | 1 + &mu; 2 2 | | Z - ( L + Y A &mu; 2 ) | | F 2 - - - ( 12 )
Therefore, using ADMM method, the step solving Z according to formula (8) is as follows:
Step 1), initiation parameter
Make Z, YA,YBIt is 0, μ2For λ1, ρ=1.1 and maxμ=106;Wherein YAAnd YBRepresent Lagrange multiplier respectively, λ1And λ2Then represent balance parameter;
Step 2), update L
Z is fixed, by utilizing following formula that L is solved when not restraining when:
L = ( &mu;K T &Phi;&Phi; T K + &mu; 2 I + &mu; 2 1 N S + N T 1 N S + N T T ) - 1 ( &mu;K T &Phi;&Phi; T K T - Y A - 1 N S + N T Y B + &mu; 2 Z + &mu; 2 1 N S + N T 1 N T T )
In formula, L is auxiliary variable, λ11/ 2, μ2Coefficient when being to solve with augmented vector approach;
Step 3), update Z
When not restraining when, fix L, by following formula, Z solved:
m i n Z | | Z | | 1 + &mu; 2 2 | | Z - ( L + Y A &mu; 2 ) | | F 2
Step 4), update Lagrange multiplier
According to step 2) and step 3) more newly obtained L and Z, substitute into following formula,
YA=YA2(L-Z),YB=YB2(1TL-1T)
Multiplier after being updated;
In above-mentioned formula (10), YAAnd YBRepresent Lagrange multiplier respectively, owing to this is a typical sparse lasso trick optimization Problem, therefore uses alternating direction multiplier method (ADMM), because formula (10) has Lagrange multiplier, therefore title is called again multiplication Device;
Step 5), undated parameter μ2
μ2Take μ2ρ and maxμIn minima (μ2ρ and maxμBeing the value set, in step 1, initiation parameter part is the most supplementary Their setting value);
Step 6), check convergence situation
If not restraining, then repeat step 2) to step 5);
If convergence, then export Z.
The second, Φ is updated
In order to solve Φ, by fixing Z, formula (7) can be written as:
The object function of formula (13) is expanded as follows:
In formula (14),
Make ΦTK Φ=I, then Φ ΦTKΦΦT=QKQT=Φ ΦT=Q, then object function can be reduced to:
Tr((λ1(K-1KT-Z)(K-1KT-Z)T2I)KTQTK) (15)
According to formula K=VSVTEigenvalues Decomposition, wherein V is characterized vector, on S diagonal element be characterized vector right The eigenvalue answered:
K T Q T K = VS 1 2 &Omega;&Omega; T S 1 2 V T
In formula,Characteristic function (15) dissolve for:
T r ( &Omega; T S 1 2 V T ( &lambda; 1 ( K - 1 K T - Z ) ( K - 1 K T - Z ) T - &lambda; 2 I ) VS 1 2 &Omega; ) = T r ( &Omega; T &Theta; &Omega; ) - - - ( 16 )
In formula (16),
And ΩTΩ=ΦTVSVTΦ=ΦTK Φ=I
Finally, formula (13) is rewritten as:
&Omega; * = arg min &Omega; T r ( &Omega; T &Theta; &Omega; ) , s . t . &Omega; T &Omega; = I - - - ( 17 )
Optimal solution Ω*It is to be obtained by characteristic vector l of the minimal eigenvalue about Θ.
Ω*It is to pass throughAnd VVT=I is tried to achieve, Φ*Optimum can be represented as:
&Phi; * = VS - 1 2 &Omega; * - - - ( 18 )
Therefore, the step of Φ is solved according to formula (13) as follows:
Step (1), seeks the Eigenvalues Decomposition of K
By K=VSVTObtain the Eigenvalues Decomposition of K;Wherein K=XTX, V are characterized vector, and on S diagonal, element is characterized Vector characteristic of correspondence value;
Step (2), is solved Θ by following formula:
&Theta; = S 1 2 V T ( &lambda; 1 ( K - 1 K T - Z ) ( K - 1 K T - Z ) T - &lambda; 2 I ) VS 1 2
Step (3), obtains the Eigenvalues Decomposition of Θ
By formula Θ=U Σ UTSolving the Eigenvalues Decomposition of Θ, U is the characteristic vector of Θ, and Σ is a diagonal matrix, and it is right Element on angle is characterized vector characteristic of correspondence value;
Step (4), solves Ω
Solve Ω by Ω=U (:, U), this formula be Θ is carried out feature decomposition after, take front d the minimal eigenvalue of U Characteristic of correspondence vector, in formula, U is the characteristic vector corresponding to front d minimal eigenvalue of U;
The scope of d is all eigenvalue numbers of 1~U, circulates successively with eigenvalues all in U, and treating excess syndrome is tested effective Those values;
Step (5), solves Φ
By formulaSolve Φ.
Summary content, obtains the image classification method migrated based on sparse nonlinear subspace of the present invention, including with Lower step:
Step 1, input source domain sampleWith aiming field sample CalculateXS 1Be source domain first sample,It is the N of source domainSIndividual sample, XT 1It it is mesh Mark territory first sample,It is the N of aiming fieldTIndividual sample;
Step 2, calculates KTAnd K
By formulaWithSolve KT and K respectively;It is that initial data X projects to The data of Hilbert space;
Step 3, obtains the Eigenvalues Decomposition of K
By formula K=VSVTSolve the Eigenvalues Decomposition of K;V is characterized the matrix that vector is formed, element on S diagonal It is characterized vector characteristic of correspondence value;
Step 4, initializes Φ
By formula Φ :=V (:, v) take the characteristic vector corresponding to front d eigenvalue of maximum of V, in formula v be the front d of V Big characteristic vector corresponding to eigenvalue;
The scope of d is 1~(all eigenvalue numbers of V), circulates successively with eigenvalues all in V, and treating excess syndrome is tested effective Those values;
Step 5, updates Z
Φ is fixed, uses ADMM method, update with the Z in following formula (8):
m i n Z | | Z | | 1 + &lambda; 1 | | &Phi; T K T - &Phi; T K Z | | F 2 - - - ( 8 )
s . t .1 N S + N T T Z = 1 N T T
Step 6, updates Φ
Z is fixed, the method using Eigenvalues Decomposition, updates with the Φ in following formula (13):
min &Phi; &lambda; 1 | | &Phi; T K T - &Phi; T K Z | | F 2 + &lambda; 2 T r ( ( I - &Phi;&Phi; T K ) T K ( I - &Phi;&Phi; T K ) ) - - - ( 13 )
s.t.ΦTK Φ=I
Step 7, checks convergence situation
If not restraining, then repeat step 5 and step 6;If convergence, then export the value of Φ and Z;
Step 8, byP must be mapped*, calculate source domain data XSIt is mapped to value M of default subspaceS,
Step 9, utilizes and maps P*, calculate the target training data X of labellingTlIt is mapped to value M of default subspaceTl,
Represent non-linear XS, project to Hilbert space and beBy sparse after P Matrix Z-direction aiming field is changed;
Step 10, calculates not labeled target detection data XTuIt is mapped to value M of default subspaceTu,
The target training data X of labelling is setTlNot labeled target detection data XTu, it is because source domain to be passed through Data XSAiming field data X with labellingTlTrain a grader W, then use not labeled aiming field data XTuEnter Row test, it is judged that the quality of the grader W that step 11 trains, to check the quality of the present invention;
Step 11, training data [M based on labellingS,MTl] and the labelling of they correspondences, utilize a two norm canonicals young waiter in a wineshop or an inn Method is taken advantage of to train a grader W;
Step 12, by calculating decision functionNot labeled target detection data are carried out classification differentiation.
Experimental verification
One, data verification
As it is shown on figure 3, for data produced by the array of default subspace, by the Gauss distribution of different modes and association side Difference matrix, it is thus achieved that (target data is random producing for the target data of minority labelling and majority unlabelled target data distributed in three dimensions The raw data meeting Gauss distribution), source domain has two class data, every class packet contains 50 samples, and easily in source domain Find the decision boundary of two class data;In aiming field, also have two class data, every class have the sample of 5 labellings and 50 unmarked Sample.From Fig. 3 it will be clear that: 1, of a sort data point has different distributions in source domain and aiming field;2、 The Optimal Separating Hyperplane of source domain is not suitable for the decision boundary (decision boundary refers to separate for inhomogeneity data line or plane) of aiming field.
The sparse Z subspace utilizing the present invention migrates, and after projection and rebuilding, obtains the source data of two-dimentional default subspace And target data, as shown in Figure 4, the spatial offset rate between source data and target data decreases, and between this two class Decision boundary clear, it is easy to finding a kind of grader to realize classification, so using present method invention to carry out data classification being Feasible, effective.
Two, object identification
CMU Multi-PIE human face data collection is a public data collection, comprises the comprehensive face number of 337 experimenters altogether According to collection, image therein passes through 15 postures, 20 kinds of illuminations, 6 kinds of expression, 4 realization captures of different stages.In order to be thought The result wanted, in an experiment 60 experimenters before stage 1 and stage 2 have selected respectively, 7 kinds of expression figures of each experimenter Picture.As shown in Figure 5: the stage 1 is the first row, face is with general expression;Stage 2 is the second row, and experimenter is also with smile.? In this experiment, four kinds of Setup Experiments are as follows:
Arrange 1: the red rectangle front expression of each experimenter and the extreme attitude of blue rectangle (60 ° of angles) are respectively intended to As source domain and the training data of aiming field, remaining expression is i.e. used as probe face.
Arrange 2: with reference to arranging 1, similarly configure.
1+2 is set: express one's feelings the two kinds of front expressions arranged in 1 and 2 and two kinds of extreme 60 ° of attitudes as training number According to, remaining expression is used as probe face.
Arranged in a crossed manner: typically to express one's feelings arrange each experimenter in 1 as source domain, will arrange 2 each experimenters' Expression of smiling, as aiming field, adapts to the change of expression with this.
Nonlinearity due to three-dimensional rotation face so that attitude tracking is extremely challenging.Fig. 6 is that the present invention is applied to Stage 2 smile expression on pose adjustment result figure, it using in Fig. 5 front express one's feelings as source data, the extreme attitude in back (60 ° of angles) is used as target data.We can see that from Fig. 6: owing to remaining the removal of (noise), target face alignment appearance State is good.Using the invention of this nonlinear method to carry out recognition of face, its recognition accuracy compares with LTSL, and it the results are shown in Table 1.
Table 1. is for the recognition accuracy ratio of human face posture identification
Task Source domain Aiming field LTSL-PCA LTSL-LDA The present invention
Arrange 1 Front The angle of 60 degree 55.7 56.0 63.7
Arrange 2 Front The angle of 60 degree 58.7 60.7 70.7
1+2 is set Front The angle of 60 degree 57.8 60.7 67.5
Arranged in a crossed manner Arrange 1 Arrange 2 96.7 96.7 99.4
This method is nonlinear method, owing to initial data is non-linear, is projected on Hilbert space carrying out again Process, thus greatly improve recognition accuracy.By contrast linear method LTSL, as can be seen from Table 1: present method invention ratio The recognition effect of linear method LTSL is good.

Claims (3)

1. the image classification method migrated based on sparse nonlinear subspace, is characterized in that, comprises the steps of
Step 1, input source domain sampleWith aiming field sample Calculate initial dataXS 1Be source domain first sample,It is the N of source domainSIndividual sample Originally, XT 1Be aiming field first sample,It is the N of aiming fieldTIndividual sample;
Step 2, calculates KTAnd K
By formulaWithSolve KT and K respectively;It is that initial data X projects to Xi Er The data in Bert space;
Step 3, obtains the Eigenvalues Decomposition of K
By formula K=VSVTSolve the Eigenvalues Decomposition of K;V is characterized the matrix of vector composition, and on the diagonal of S, element is characterized Eigenvalue corresponding to vector;
Step 4, initialization feature vector Φ
By formulaTaking the characteristic vector corresponding to front d eigenvalue of maximum of V, in formula, v is front d the maximum feature of V Characteristic vector corresponding to value;
Step 5, updates Z
Φ is fixed, uses ADMM method, update with the Z in following formula (8):
min Z | | Z | | 1 + &lambda; 1 | | &Phi; T K T - &Phi; T K Z | | F 2 - - - ( 8 )
s . t .1 N S + N T T Z = 1 N T T
In formula (8), Z is sparse matrix, λ1Represent balance parameter,Represent a length of NS+NTNon-zero column vector,Represent A length of NTNon-zero column vector;
Step 6, updates Φ
Z is fixed, the method using Eigenvalues Decomposition, updates with the Φ in following formula (13):
min &Phi; &lambda; 1 | | &Phi; T K T - &Phi; T K Z | | F 2 + &lambda; 2 T r ( ( I - &Phi;&Phi; T K ) T K ( I - &Phi;&Phi; T K ) ) - - - ( 13 )
s.t.ΦTK Φ=I
Step 7, checks convergence situation
If not restraining, then repeat step 5 and step 6;If convergence, then export the value of Φ and Z;
Step 8, byP must be mapped*, calculate source domain data XSIt is mapped to value M of default subspaceS,
Step 9, utilizes and maps P*, calculate the target training data X of labellingTlIt is mapped to value M of default subspaceTl,
Represent non-linear XS, project to Hilbert space and beThen by sparse matrix Z Change to aiming field;
Step 10, calculates not labeled target detection data XTuIt is mapped to value M of default subspaceTu,
Step 11, training data [M based on labellingS,MTl] and the labelling of they correspondences, utilize two norm canonical least squares Method one grader W of training;
Step 12, by calculating decision functionNot labeled target detection data are carried out classification differentiation.
The image classification method migrated based on sparse nonlinear subspace the most according to claim 1, is characterized in that, in step Rapid 5 steps updating Z include:
Step 1), initiation parameter
Make Z, YA,YBIt is 0, μ2For λ1, ρ=1.1 and maxμ=106;Wherein YAAnd YBRepresent Lagrange multiplier, λ respectively1And λ2 Then represent balance parameter;
Step 2), update L
Z is fixed, by utilizing following formula that L is solved when not restraining when:
L = ( &mu; 1 K T &Phi;&Phi; T K + &mu; 2 I + &mu; 2 1 N S + N T 1 N S + N T T ) - 1 ( &mu; 1 K T &Phi;&Phi; T K T - Y A - 1 N S + N T Y B + &mu; 2 Z + &mu; 2 1 N S + N T 1 N T T )
In formula, L is auxiliary variable, λ11/ 2, μ2Coefficient when being to solve with augmented vector approach;
Step 3), update Z
When not restraining when, fix L, by following formula, Z solved:
m i n Z | | Z | | 1 + &mu; 2 2 | | Z - ( L + Y A &mu; 2 ) | | F 2
Step 4), update Lagrange multiplier
According to step 2) and step 3) more newly obtained L and Z, substitute into following formula,
YA=YA2(L-Z),YB=YB2(1TL-1T)
Multiplier after being updated;
Step 5), undated parameter μ2
μ2Take μ2ρ and maxμIn minima;
Step 6), check convergence situation
If not restraining, then repeat step 2) to step 5);
If convergence, then export Z.
The image classification method migrated based on sparse nonlinear subspace the most according to claim 1 and 2, is characterized in that, The step updating Φ in step 6 includes:
Step (1), seeks the Eigenvalues Decomposition of K
By K=VSVTObtain the Eigenvalues Decomposition of K;Wherein K=XTX, V are characterized vector, and on S diagonal, element is characterized vector Characteristic of correspondence value;
Step (2), is solved Θ by following formula:
&Theta; = S 1 2 V T ( &lambda; 1 ( K - 1 K T - Z ) ( K - 1 K T - Z ) T - &lambda; 2 I ) VS 1 2
Step (3), obtains the Eigenvalues Decomposition of Θ
By formula Θ=U Σ UTSolving the Eigenvalues Decomposition of Θ, U is the characteristic vector of Θ, and Σ is a diagonal matrix, on its diagonal angle Element be characterized vector characteristic of correspondence value;
Step (4), solves Ω
Solve Ω by Ω=U (:, U), this formula be Θ is carried out feature decomposition after, front d the minimal eigenvalue taking U is corresponding Characteristic vector, in formula, U is the characteristic vector corresponding to front d minimal eigenvalue of U;
Step (5), solves Φ
By formulaSolve Φ.
CN201610748644.8A 2016-08-29 2016-08-29 Image classification method based on sparse nonlinear subspace migration Pending CN106326935A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610748644.8A CN106326935A (en) 2016-08-29 2016-08-29 Image classification method based on sparse nonlinear subspace migration

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610748644.8A CN106326935A (en) 2016-08-29 2016-08-29 Image classification method based on sparse nonlinear subspace migration

Publications (1)

Publication Number Publication Date
CN106326935A true CN106326935A (en) 2017-01-11

Family

ID=57788252

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610748644.8A Pending CN106326935A (en) 2016-08-29 2016-08-29 Image classification method based on sparse nonlinear subspace migration

Country Status (1)

Country Link
CN (1) CN106326935A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107729914A (en) * 2017-09-06 2018-02-23 鲁小杰 A kind of detection method of pathological data
CN108798641A (en) * 2018-06-19 2018-11-13 东北大学 A kind of Diagnosing The Faults of Sucker Rod Pumping System method based on subspace transfer learning
CN108875928A (en) * 2017-05-15 2018-11-23 广东石油化工学院 Multi-output regression network and learning method
CN110146655A (en) * 2019-05-31 2019-08-20 重庆大学 A kind of anti-electronic nose bleach-out process based on adaptive sub-space learning normal form
CN111461191A (en) * 2020-03-25 2020-07-28 杭州跨视科技有限公司 Method and device for determining image sample set for model training and electronic equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101079103A (en) * 2007-06-14 2007-11-28 上海交通大学 Human face posture identification method based on sparse Bayesian regression
CN104392243A (en) * 2014-11-18 2015-03-04 西北工业大学 Nonlinear un-mixing method of hyperspectral images based on kernel sparse nonnegative matrix decomposition

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101079103A (en) * 2007-06-14 2007-11-28 上海交通大学 Human face posture identification method based on sparse Bayesian regression
CN104392243A (en) * 2014-11-18 2015-03-04 西北工业大学 Nonlinear un-mixing method of hyperspectral images based on kernel sparse nonnegative matrix decomposition

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LEI ZHANG等: "LSDT: Latent Sparse Domain Transfer Learning for Visual Adaptation", 《IEEE TRANSACTIONS ON I《IEEE TRANSACTIONS ON IMAGE PROCESSING》MAGE PROCESSING》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108875928A (en) * 2017-05-15 2018-11-23 广东石油化工学院 Multi-output regression network and learning method
CN108875928B (en) * 2017-05-15 2021-02-26 广东石油化工学院 Multi-output regression network and learning method
CN107729914A (en) * 2017-09-06 2018-02-23 鲁小杰 A kind of detection method of pathological data
CN107729914B (en) * 2017-09-06 2021-06-25 鲁小杰 Pathological data detection method
CN108798641A (en) * 2018-06-19 2018-11-13 东北大学 A kind of Diagnosing The Faults of Sucker Rod Pumping System method based on subspace transfer learning
CN110146655A (en) * 2019-05-31 2019-08-20 重庆大学 A kind of anti-electronic nose bleach-out process based on adaptive sub-space learning normal form
CN110146655B (en) * 2019-05-31 2021-12-21 重庆大学 Electronic nose drift resistance method based on self-adaptive subspace learning paradigm
CN111461191A (en) * 2020-03-25 2020-07-28 杭州跨视科技有限公司 Method and device for determining image sample set for model training and electronic equipment
CN111461191B (en) * 2020-03-25 2024-01-23 杭州跨视科技有限公司 Method and device for determining image sample set for model training and electronic equipment

Similar Documents

Publication Publication Date Title
CN106326935A (en) Image classification method based on sparse nonlinear subspace migration
CN103632138B (en) A kind of face identification method of low-rank piecemeal rarefaction representation
Zhang et al. Robust low-rank kernel multi-view subspace clustering based on the schatten p-norm and correntropy
Yang et al. Kernel ICA: An alternative formulation and its application to face recognition
Zhou et al. Bayesian low-tubal-rank robust tensor factorization with multi-rank determination
Allen et al. Capturing inter-subject variability with group independent component analysis of fMRI data: a simulation study
Małek et al. The VIMOS Public Extragalactic Redshift Survey (VIPERS)-A support vector machine classification of galaxies, stars, and AGNs
CN111275192B (en) Auxiliary training method for improving accuracy and robustness of neural network simultaneously
CN105405098A (en) Image super-resolution reconstruction method based on sparse representation and adaptive filtering
CN104809471B (en) A kind of high spectrum image residual error integrated classification method based on spatial spectral information
Sohl-Dickstein et al. Hamiltonian annealed importance sampling for partition function estimation
CN104751176A (en) Method for selecting high-spectrum remote-sensing image wave band
CN105929113A (en) Electronic nose signal error adaptive-learning method with subspace projection
Galvin et al. Radio galaxy zoo: Knowledge transfer using rotationally invariant self-organizing maps
Boroson et al. Exploring the spectral space of low redshift QSOs
CN103745233A (en) Hyper-spectral image classifying method based on spatial information transfer
Finnveden et al. Understanding when spatial transformer networks do not support invariance, and what to do about it
Sun et al. Mixed noise removal for hyperspectral images based on global tensor low-rankness and nonlocal SVD-aided group sparsity
CN104050489B (en) SAR ATR method based on multicore optimization
Kalaganis et al. Revisiting Riemannian geometry-based EEG decoding through approximate joint diagonalization
CN104050482B (en) A kind of manifold learning generalization algorithm based on local linear smoothing
CN106326936A (en) Image classification method based on sparse preset subspace migration
CN106960225A (en) A kind of sparse image classification method supervised based on low-rank
CN109815889A (en) A kind of across resolution ratio face identification method based on character representation collection
Qi et al. A novel joint dictionary framework for sparse hyperspectral unmixing incorporating spectral library

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20170111

WD01 Invention patent application deemed withdrawn after publication