CN107491419A - A kind of linear discriminant analysis method with bilinearity low-rank subspace - Google Patents

A kind of linear discriminant analysis method with bilinearity low-rank subspace Download PDF

Info

Publication number
CN107491419A
CN107491419A CN201710793137.0A CN201710793137A CN107491419A CN 107491419 A CN107491419 A CN 107491419A CN 201710793137 A CN201710793137 A CN 201710793137A CN 107491419 A CN107491419 A CN 107491419A
Authority
CN
China
Prior art keywords
matrix
discriminant analysis
low
loss function
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710793137.0A
Other languages
Chinese (zh)
Other versions
CN107491419B (en
Inventor
何小海
苏婕
卿粼波
周文
周文一
王正勇
吴晓红
熊淑华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan University
Original Assignee
Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan University filed Critical Sichuan University
Priority to CN201710793137.0A priority Critical patent/CN107491419B/en
Publication of CN107491419A publication Critical patent/CN107491419A/en
Application granted granted Critical
Publication of CN107491419B publication Critical patent/CN107491419B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Computational Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computing Systems (AREA)
  • Algebra (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Monitoring And Testing Of Nuclear Reactors (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a kind of linear discriminant analysis method with bilinearity low-rank subspace.Comprise the following steps:For matrix data sample set, retain the form of script matrix as input data, build the discriminant analysis model based on matrix, loss function is built using the least square method based on matrix, introduce nuclear norm regular terms and low-rank constraint is carried out to mapping matrix stack, loss function is solved using alternating direction algorithm (ADMM) framework of multiplier, to seek the bilinearity low-rank subspace of matrix samples collection, so as to carry out dimensionality reduction to matrix samples collection.Linear discriminant analysis method of the present invention with bilinearity low-rank subspace is obviously improved than other method in matrix samples collection dimensionality reduction effect, and evaluation index general performance is fine.

Description

A kind of linear discriminant analysis method with bilinearity low-rank subspace
Technical field
The present invention devises a kind of linear discriminant analysis method with bilinearity low-rank subspace, is related to machine learning, Data Dimensionality Reduction, data mining technology field.
Background technology
With the fast development of big data, the correlation technique such as data mining, machine learning and artificial intelligence obtains widely Using.Requirement of the people to data becomes to become increasingly complex, because people are preserving the knot of data message and heuristic data behind There are more demands in terms of structure correlation, this causes the complexity of model structure and the inefficiencies of model training.Therefore, obtain Correlation between data is very necessary to reduce data redudancy and computation complexity.
A kind of intuitively method for reducing data redudancy is that the structural dependence between data is obtained by dimensionality reduction.PCA With LDA be two kinds of common dimension reduction methods, the feature extraction being widely used in area of pattern recognition, such as Eigenfaces and Signal structure correlation analysis in Fisherfaces, and signal transacting, PCA and LDA have obtained very big in dimension-reduction treatment Success.
PCA and LDA is by carrying out Eigenvalues Decomposition to covariance matrix or hash matrix, therefore both can be equivalent to Generalized eigenvalue decomposition problem.Input sample of data is expressed as vector or scalar by classical PCA and LDA, based on these vectors It is modeled, goes out to try to achieve mapping matrix.However, the data such as image, electroencephalogram (EEG), itself are matrix forms, if by its Represent to turn into vector, data dimension can be made very high.In this case, will during classical PCA and LDA progress Eigenvalues Decomposition Become abnormal difficult while cause huge amount of calculation.For LDA, or even lack sampling problem occurs, and lead to not Solve characteristic value;In addition, for this kind of matrix data, it is expressed as vector form, its original internal ranks knot can be destroyed Structure information.
In order to overcome drawbacks described above, existing researcher proposes several dimension reduction methods based on data matrix.Such as IMPCA, 2DPCA, (2D) 2PCA, 2DLDA, 2D-LDA and BDCA.However, these component analyzing methods still need carry out feature It is worth decomposition computation, calculation cost is very big.Because under certain condition, linear discriminant analysis can be equivalent to linear regression, therefore The loss function of least square method structure linear discriminant analysis can be used, so as to avoid meter huge caused by Eigenvalues Decomposition Calculation amount.
The content of the invention
The present invention provides a kind of more efficiently linear discriminant with bilinearity low-rank subspace to solve the above problems Analysis method (Linear Discriminant Analysis with Bilinear Low-Rank Subspace, BLR- LDA).The present invention be directed to matrix samples data set, the discriminant analysis model based on matrix is built, using with particular indicators The least square method structure loss function of matrix, and low-rank is carried out about to mapping matrix stack by introducing nuclear norm regular terms Beam, to obtain the structural dependence of input sample internal matrix, more preferable discriminant analysis effect can be obtained.
The present invention is achieved through the following technical solutions above-mentioned purpose:
A kind of linear discriminant analysis method with bilinearity low-rank subspace, comprises the following steps:
(1) matrix data sample set is directed to, regard it as input data in the form of script matrix.
(2) the discriminant analysis model based on matrix is built, using the least square method structure loss letter pair based on matrix Matrix samples collection enters row matrix discriminant analysis, to seek the two-wire subspace of matrix samples collection.
(3) low-rank constraint is carried out to mapping matrix stack by adding nuclear norm regular terms in loss function, so as to obtain The ranks correlation of matrix data sample interior.
(4) assume that each mapping matrix that mapping matrix is concentrated is separate, former loss function is converted into k son of solution Lose letter.
(5) mapping matrix is solved using the alternating direction algorithm (ADMM) of multiplier, is mapped to matrix data sample set Low-rank subspace.
Brief description of the drawings
Fig. 1 is classical LDA block schematic illustrations
Fig. 2 is BLR-LDA block schematic illustrations of the present invention
Embodiment
The invention will be further described below in conjunction with the accompanying drawings:
Fig. 1 is classical LDA block schematic illustrations, is comprised the following steps:
(1) the n set of data samples characterized by matrix form are inputted, are vectorial shape by each data sample matrixing Formula:
Wherein,Represent i-th of input sample of data matrix, xiRepresent its corresponding vector form.Then n number Data matrix is constituted according to vector
(2) mapping matrix is usedLow-dimensional will be mapped to by the primary data sample collection of vector representation Space, pass through XTW is operated, the data vector after being mapped
Fig. 2 is BLR-LDA block schematic illustrations of the present invention.Comprise the following steps:
(1) the n data sample X characterized by matrix form are inputtediTo retain its structural information, correspondingly, in classical LDA Mapping matrix W each row correspond to a mapping matrix:D=p × q, j=1 ..., k.Then in classical LDA Mapping matrix W by one group of new mapping matrix collection WjRepresent.
(2) under certain condition, classical LDA can be equivalent to linear regression, and this allows classical LDA directly to convert For least square problem, huge amount of calculation as caused by Eigenvalues Decomposition is avoided.Classical LDA can represent as follows:
Wherein,It is the indicator matrix of centralization, it is defined as follows:
Wherein, n is total number of samples amount, njIt is jth class sample size.
(3) above equation relation, mapping matrix are based onIt is calculated by equation below:
Wherein,It is that i-th of data sample is vectorial, wjIt is the jth row of mapping matrix W in classical LDA.
(4) for the structural information of retention data matrix, we are vectorial by data sampleBe converted to data sample MatrixCorrespondingly, the jth of projection matrix arranges in classical LDABe converted to matrix formCause This, in above formulaRepresent as follows:
(5) in order to obtain the structural information of data sample matrix, we are to each WjNuclear norm regularization is used alone. Loss function based on mapping matrix optimization problem is specified as follows:
Wherein, τ is the hyper parameter obtained by cross validation.
(6) in BLR-LDA frameworks, it will be assumed that each WjIt is separate.Therefore, above-mentioned optimization problem can To be expressed as k sub- loss functions:
We solve above-mentioned optimization problem using alternating direction (ADMM) algorithm of multiplier, fissioned on WjFission and auxiliary variable SjTwo subproblems.Above formula can be equivalently written as:
s.t.Wj-Sj=0
Wherein,G(Sj)=τ | | Sj||*
(7) solved the above problems using augmented vector approach:
Wherein, ΛjIt is Lagrange multiplier matrix, ρ > 0 are hyper parameters.
First, on auxiliary variable SjThe solution method of subproblem is as follows:
Optimal solution Sj (t)Obtained by the t times iteration of following analytical expression:
Wherein, ρ Wjj=U Σ VTIt is singular value decomposition form.For arbitrary τ > 0,It is unusual It is worth threshold operator, Sτ(Σ)ii=max (Σii-τ,0)。
Then we solve second subproblem Wj, expression formula is as follows:
WjIt is a clear and definite convex optimization problem, we take gradient to decline to solve it.Changed using following form In generation, updates Wj
Wherein, α > 0 are learning rate, are a hyper parameters.WjPart derivation formula it is as follows:
In addition, LaGrange parameter ΛjUpdated with following single gradient step:
(8) ADMM algorithms are used, continuous alternating iteration updates Wj、Sj、ΛjUntil convergence, you can try to achieve one group and optimal reflect Penetrate matrix stack, Wj, so as to which matrix samples collection is mapped into bilinearity low-rank subspace.

Claims (4)

  1. A kind of 1. linear discriminant analysis method with bilinearity low-rank subspace, it is characterised in that comprise the following steps:
    Step 1:For n matrix data sample set { Xi|Xi∈Rp×q, p × q=d, 1≤i≤n } (such as:View data, brain Electromyographic data), it regard each data sample as input data in the form of script matrix;
    Step 2:The discriminant analysis model based on matrix is built, loss function L is built using the least square method based on matrix (W(1,2,…k)), by L (W(1,2,…k)) solution is iterated, so as to obtain mapping matrix collection Wj∈Rp×q(j=1,2 ... k), To seek the two-wire subspace of matrix samples collection, so as to carry out dimensionality reduction to matrix samples collection;
    Step 3:By in loss function L (W(1,2,…k)) in add nuclear norm canonical | | Wj||*Item is to each mapping matrix WjEnter Row low-rank constrains, so as to obtain the ranks correlation of matrix data sample interior;
    Step 4:Assuming that k WjBetween independently of each other, will solve loss function L (W(1,2,…k)) the problem of to be converted into solution k sub Loss function L (Wj) the problem of, so as to k L (Wj) solve respectively;
    Step 5:Sub- loss function L (Wj) non-smooth, using multiplier alternating direction algorithm (ADMM) framework to L (Wj) asked Solution, and former optimization problem is changed into two sub- optimization problems and solved, mapping matrix collection is finally given after iteration, so as to Original matrix sample set is mapped to low-rank subspace, to reach the purpose of Data Dimensionality Reduction.
  2. 2. the matrix samples collection of claim 1 all input matrix discriminant analysis models in the matrix form so that primary data sample Ranks relational structural information retained.
  3. 3. when in claim 2 to the loss function structure of matrix discriminant analysis model, the least square based on matrix is employed Method, high calculation cost caused by Eigenvalues Decomposition in conventional linear discriminant analysis is avoided, while it also avoid traditional wire Lack sampling problem in property discriminant analysis method.
  4. 4. by loss function L (W in claim 3(1,2,…k)) in each WjIntroduce the constraint of nuclear norm low-rank so that model Original matrix data sample ranks correlation information can be obtained.
CN201710793137.0A 2017-09-06 2017-09-06 Linear discriminant analysis method with bilinear low-rank subspace Active CN107491419B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710793137.0A CN107491419B (en) 2017-09-06 2017-09-06 Linear discriminant analysis method with bilinear low-rank subspace

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710793137.0A CN107491419B (en) 2017-09-06 2017-09-06 Linear discriminant analysis method with bilinear low-rank subspace

Publications (2)

Publication Number Publication Date
CN107491419A true CN107491419A (en) 2017-12-19
CN107491419B CN107491419B (en) 2020-06-23

Family

ID=60652213

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710793137.0A Active CN107491419B (en) 2017-09-06 2017-09-06 Linear discriminant analysis method with bilinear low-rank subspace

Country Status (1)

Country Link
CN (1) CN107491419B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120254077A1 (en) * 2011-03-31 2012-10-04 Fatih Porikli Data Driven Frequency Mapping for Kernels Used in Support Vector Machines
CN104463246A (en) * 2014-12-08 2015-03-25 天津大学 Manifold-based linear regression learning method
CN105389343A (en) * 2015-10-23 2016-03-09 北京工业大学 Vectorized dimension reduction method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120254077A1 (en) * 2011-03-31 2012-10-04 Fatih Porikli Data Driven Frequency Mapping for Kernels Used in Support Vector Machines
CN104463246A (en) * 2014-12-08 2015-03-25 天津大学 Manifold-based linear regression learning method
CN105389343A (en) * 2015-10-23 2016-03-09 北京工业大学 Vectorized dimension reduction method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
杨晓敏 等: "基于投影中心距离的支持向量预选取", 《四川大学学报(自然科学版)》 *

Also Published As

Publication number Publication date
CN107491419B (en) 2020-06-23

Similar Documents

Publication Publication Date Title
Wang et al. Deepvid: Deep visual interpretation and diagnosis for image classifiers via knowledge distillation
WO2022041678A1 (en) Remote sensing image feature extraction method employing tensor collaborative graph-based discriminant analysis
CN104268593B (en) The face identification method of many rarefaction representations under a kind of Small Sample Size
Boughida et al. A novel approach for facial expression recognition based on Gabor filters and genetic algorithm
CN105631479B (en) Depth convolutional network image labeling method and device based on non-equilibrium study
CN108304357B (en) Chinese character library automatic generation method based on font manifold
CN107203787B (en) Unsupervised regularization matrix decomposition feature selection method
CN107122809A (en) Neural network characteristics learning method based on image own coding
CN106203356B (en) A kind of face identification method based on convolutional network feature extraction
CN103400143A (en) Data subspace clustering method based on multiple view angles
CN114398961A (en) Visual question-answering method based on multi-mode depth feature fusion and model thereof
CN106295694A (en) Face recognition method for iterative re-constrained group sparse representation classification
CN109376787B (en) Manifold learning network and computer vision image set classification method based on manifold learning network
CN106682606A (en) Face recognizing method and safety verification apparatus
CN105868796A (en) Design method for linear discrimination of sparse representation classifier based on nuclear space
CN108491863A (en) Color image processing method based on Non-negative Matrix Factorization and convolutional neural networks
CN115050064A (en) Face living body detection method, device, equipment and medium
US20230101539A1 (en) Physiological electric signal classification processing method and apparatus, computer device and storage medium
CN117690178B (en) Face image recognition method and system based on computer vision
CN115658886A (en) Intelligent liver cancer staging method, system and medium based on semantic text
Wang et al. CWC-transformer: a visual transformer approach for compressed whole slide image classification
CN106803105A (en) A kind of image classification method based on rarefaction representation dictionary learning
CN116758379B (en) Image processing method, device, equipment and storage medium
CN109447147A (en) The image clustering method decomposed based on the sparse matrix of depths of digraph
Li et al. Locally-enriched cross-reconstruction for few-shot fine-grained image classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant