CN110348287A - A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram - Google Patents

A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram Download PDF

Info

Publication number
CN110348287A
CN110348287A CN201910441208.XA CN201910441208A CN110348287A CN 110348287 A CN110348287 A CN 110348287A CN 201910441208 A CN201910441208 A CN 201910441208A CN 110348287 A CN110348287 A CN 110348287A
Authority
CN
China
Prior art keywords
dictionary
matrix
objective function
sample
space
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910441208.XA
Other languages
Chinese (zh)
Inventor
唐厂
万诚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Geosciences
Original Assignee
China University of Geosciences
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Geosciences filed Critical China University of Geosciences
Priority to CN201910441208.XA priority Critical patent/CN110348287A/en
Publication of CN110348287A publication Critical patent/CN110348287A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/192Recognition using electronic means using simultaneous comparisons or correlations of the image signals with a plurality of references
    • G06V30/194References adjustable by an adaptive method, e.g. learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The present invention is a kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram, disclose a kind of unsupervised feature selection approach and system based on dictionary and the study of sample similar diagram, in this method and system, due to the new dictionary base space of one by being given to raw data matrix X, in the new dictionary base space W of generation, l is used2,1The row sparsity that norm applies W is used to the importance of measures characteristic with this.Compared with the rudimentary representation method in the past in original feature space, it is based on dictionary and sample similar diagram learning model disclosed in the present invention, captures higher level and more abstract expression, has broad application prospects.

Description

A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram
Technical field
The present invention relates to signal processings, data analysis field, and in particular to one kind is learnt based on dictionary and sample similar diagram Unsupervised feature selection approach and device.
Background technique
With the fast development of sensor and the Internet media, higher-dimension analysis and big data have become one and have challenge Property and inevitable problem.In numerous applications, such as computer vision, machine learning, pattern-recognition and medical analysis all A large amount of high dimensional data is used.Although data are typically denoted as high dimensional feature vector, for learning tasks, only Have a very little but unknown character subset be very important with it is distinguishing.Directly handling these high dimensional datas can not only show Write and increase demand to computing resource, and due to dimension disaster (curse of dimensionality) can reduce it is many existing There is the performance of algorithm.Feature selecting, which is typically considered, reduces uncorrelated or redundancy feature quantity to improve learning algorithm performance A significant process.
According to whether the label information of sample data classification is utilized, feature selection approach generally can be divided into three classes: have prison Superintend and direct feature selecting (Supervised feature selection), unsupervised feature selecting (Unsupervised feature ) and semi-supervised feature selecting (Semi-supervised feature selection) selection.In general, unsupervised feature Selection method may be summarized to be three kinds, that is, filtering type (Filter), packaged type (Wrapper) and embedded (Embedded).Phase Than in other two methods, embedded methods construct a learning model usually to export the subset of feature, it is in many aspects Excellent performance can be obtained, and is rapidly developed.Data representation based on embedded methods has obtained widely answering With, and propose various models.These methods assume that initial data can be by executing reconstruct letter on selected character subset Then number adds a special regularization matrix to indicate on the characterization coefficient matrix of feature selecting.Wherein, what is used is most Existing algorithm model be from indicate algorithm, it assumes that each feature may be expressed as the linear combination of its correlated characteristic.This Outside, for the partial structurtes of data, usually figure Laplace operator is embedded into embedded UFS method.However, traditional Embedded methods at least there are problems that two.First, existing most of algorithms are that feature is carried out on original feature space Selection frequently can lead to result inaccuracy due to the noise and redundancy feature of original feature space, and can not utilize higher level More abstract expression.Second, traditional UFS method usually constructs similar matrix respectively and carries out feature selecting, therefore, rear In continuous process, the similar matrix obtained in the initial data is remained unchanged, but the noise sample for including in real data and spy Sign inevitably makes similar matrix become unreliable.This insecure similar matrix may destroy the part of initial data Popular structure, to make result that can not optimize.
Summary of the invention
The technical problem to be solved in the present invention is that in view of the drawbacks of the prior art, providing a kind of based on dictionary and sample The unsupervised feature selection approach and system-DGL-UFS of similar diagram study, and other unsupervised feature selecting algorithms, such as: LS, Baseline, SCUFS and RJGSC etc. are compared, and DGL-UFS method is on dictionary base space rather than in essential characteristic sky Between it is upper select feature, and one similar matrix of study automatically, to keep the local manifolds structure of data.
The technical solution adopted by the present invention to solve the technical problems is: constructing a kind of based on dictionary graphics similar with sample The unsupervised feature selection approach practised, comprising the following steps:
S1, a raw data matrix X={ x is given1,x2,…xn}={ f1;f2…;fd}∈Rd×n;Wherein:
N is number of samples, and d is the quantity of feature vector;I ∈ n, xi∈Rd×1What is indicated is i-th of sample;J ∈ d, fj∈ Rd×1It is j-th of feature vector;
S2, study have the dictionary D ∈ R of m base vectord×m, utilize the raw data matrix that dictionary D is given to step S1 X is reconstructed, and obtains a new dictionary base space W;Wherein, l is utilized2,1Norm applies row sparsity to W, so that from dictionary D In select corresponding base vector to indicate each sample;
S3, Laplce's restricted model based on graph model, definition constraint regularization term are utilized;
S4, the constraint regularization term defined in conjunction with the obtained dictionary base space W of step S2 and step S3, in dictionary base In space W, feature selecting is carried out, specifically: construction DGL-UFS model objective function solves and obtains W;Wherein, excellent using iteration Change strategy, by carrying out constraint to each other to multiple variables in objective function, W is finally calculated.
Further, raw data matrix X is reconstructed, resulting reconstruction coefficients matrix W ∈ Rm×nAs newly generated Dictionary space;Wherein, the objective function about base vector and new dictionary base space learning can indicate are as follows:
Wherein,Expression takes square of the F norm of matrix " ";β is one and takes positive balance parameters;The i-th of dictionary D List is shown as di, wherein constraint condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;||W||2,1 It indicates to utilize l2,1Norm applies row sparsity to dictionary base space W;M is the positive integer greater than 1.
Further, regularization term expression formula is constrained defined in step S3 are as follows:
Wherein, SijIt is data point xiSimilarity measure between xj;L=A-S is figure Laplacian Matrix, and A is one right Angular moment battle array, whereinS is similarity matrix;wiOr wjIt is arranged for the i-th or j of W.
Further, the DGL-UFS model objective function of definition are as follows:
s.t.||di||2≤1,ST1=1, S >=0, i=1,2 ... m
Wherein,Square of the F norm of representing matrix " ";α be taken for one of balance chart regular terms it is positive normal Number, β are one and take positive balance parameters;L=A-S is figure Laplacian Matrix, and A is a diagonal matrix, and S is similar matrix;| |W||2,1It indicates to utilize l2,1Norm applies row sparsity to dictionary base space W;The i-th list of dictionary D is shown as di, wherein constraint Condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;M is the positive integer greater than 1.
Further, in step S4, the specific steps of objective function are solved by alternating iteration optimisation strategy are as follows:
S51, initialization dictionary base D and similar matrix S;
S52, preset parameter D and S, objective function is converted are as follows:
Wherein, ginseng is further updated by solving above-mentioned objective function using an iteration weight weighted least square algorithm Number W;
S53, preset parameter W and S, objective function is converted are as follows:
Wherein, using ADMM-Alternating Direction method of Multipliers algorithm, by drawing Enter a matrix of variables into objective function, further undated parameter D;
S54, preset parameter W and D, objective function is converted are as follows:
Wherein, due to the similarity vector of each sample be it is independent, for i-th of sample, current goal function representation are as follows:
By introducing Lagrange's multiplier, the similar matrix s of current i-th of sample is obtainedi;Further, for each Sample successively updates the similar matrix of each sample using above-mentioned identical method, realizes the update to parameter S;
S55, repeat step S52-S54, until objective function convergence when, export convergent under parameter S, D and W。
A kind of unsupervised feature selecting device learnt based on dictionary and sample similar diagram proposed by the present invention, the device packet It includes with lower module:
Data setup module, for giving a raw data matrix X={ x1,x2,…xn}={ f1;f2…;fd}∈Rd ×n;Wherein:
N is number of samples, and d is the quantity of feature vector;I ∈ n, xi∈Rd×1What is indicated is i-th of sample;J ∈ d, fj∈ Rd×1It is j-th of feature vector;
Space Reconstruction module, for learning the dictionary D ∈ R with m base vectord×m, mould is arranged to data using dictionary D The given raw data matrix X of block is reconstructed, and obtains a new dictionary base space W;Wherein, l is utilized2,1Norm applies W Row sparsity, so that selecting corresponding base vector from dictionary D to indicate each sample;
Regularization module is constrained, for utilizing Laplce's restricted model based on graph model, definition constraint regularization term;
Feature selection module, the dictionary base space W for being obtained in conjunction with Space Reconstruction module, and constraint regularization module The constraint regularization term of definition carries out feature selecting in dictionary base space W, specifically: construction DGL-UFS simulated target letter Number solves and obtains W;Wherein, using iteration optimization strategy, by carrying out pact to each other to multiple variables in objective function W is finally calculated in beam.
Further, raw data matrix X is reconstructed, resulting reconstruction coefficients matrix W ∈ Rm×nAs newly generated Dictionary space;Wherein, the objective function about base vector and new dictionary base space learning can indicate are as follows:
Wherein,Expression takes square of the F norm of matrix " ";β is one and takes positive balance parameters;The i-th of dictionary D List is shown as di, wherein constraint condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;||W||2,1 It indicates to utilize l2,1Norm applies row sparsity to dictionary base space W;M is the positive integer greater than 1.
Further, constraint regularization term expression formula defined in regularization module is constrained are as follows:
Wherein, SijIt is data point xiAnd xjBetween similarity measure;L=A-S is figure Laplacian Matrix, and A is one right Angular moment battle array, whereinS is similarity matrix;wiOr wjIt is arranged for the i-th or j of W.
Further, the DGL-UFS model objective function of definition are as follows:
s.t.||di||2≤1,ST1=1, S >=0, i=1,2 ... m;
Wherein,Square of the F norm of representing matrix " ";α be taken for one of balance chart regular terms it is positive normal Number, β are one and take positive balance parameters;L=A-S is figure Laplacian Matrix, and A is a diagonal matrix, and S is similar matrix;| |W||2,1It indicates to utilize l2,1Norm applies row sparsity to dictionary base space W;The i-th list of dictionary D is shown as di, wherein constraint Condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;M is the positive integer greater than 1.
Further, in feature selection module, the specific steps of objective function are solved by alternating iteration optimisation strategy are as follows:
S51, initialization dictionary base D and similar matrix S;
S52, preset parameter D and S, objective function is converted are as follows:
Wherein, ginseng is further updated by solving above-mentioned objective function using an iteration weight weighted least square algorithm Number W;
S53, preset parameter W and S, objective function is converted are as follows:
Wherein, using ADMM-Alternating Direction method of Multipliers algorithm, by drawing Enter a matrix of variables into objective function, further undated parameter D;
S54, preset parameter W and D, objective function is converted are as follows:
Wherein, due to the similarity vector of each sample be it is independent, for i-th of sample, current goal function representation are as follows:
By introducing Lagrange's multiplier, the similar matrix s of current i-th of sample is obtainedi;Further, for each Sample successively updates the similar matrix of each sample using above-mentioned identical method, realizes the update to parameter S;
S55, repeat step S52-S54, until objective function convergence when, export convergent under parameter S, D and W。
In a kind of unsupervised feature selection approach and system based on dictionary and the study of sample similar diagram of the present invention In, due to W={ w1,w2,…,wnIt is that a new space of X uses l in new dictionary base space2,1Norm applies W Row sparsity can be used to the importance of measures characteristic.Compared with the rudimentary representation method in the past in original feature space, DGL-UFS captures higher level and more abstract expression.
Detailed description of the invention
Present invention will be further explained below with reference to the attached drawings and examples, in attached drawing:
Fig. 1 is the method flow diagram that the present invention realizes unsupervised feature selecting;
Fig. 2 is unsupervised feature selecting apparatus structure schematic diagram disclosed by the invention;
Fig. 3 is the iterative algorithm figure disclosed by the invention for solving dictionary base space W;
Fig. 4 is the algorithm pattern disclosed by the invention that objective function is solved using iteration optimization algorithms;
Fig. 5 is the statistics of eight databases;
Fig. 6 is different cluster result figure (ACC% ± std%) of the feature selection approach on each database;
Fig. 7 is different cluster result figure (NMI% ± std%) of the feature selection approach on each database;
Fig. 8 is on different data sets, and distinct methods correspond to the ACC value of the selected feature of different number;
Fig. 9 is on different data sets, and distinct methods correspond to the NMI value of the selected feature of different number;
Figure 10 is ACC value of the present invention in the case where keeping parameter alpha=1 to change the value of β;
Figure 11 is NMI value of the present invention in the case where keeping parameter alpha=1 to change the value of β;
Figure 12 is ACC value of the present invention in the case where keeping parameter beta=1 to change the value of α;
Figure 13 is NMI value of the present invention in the case where keeping parameter beta=1 to change the value of α;
Figure 14 is the convergence curve of algorithm 1 and algorithm 2 on Yale data set;
Figure 15 is the convergence curve of algorithm 1 and algorithm 2 on COIL20 data set;
Figure 16 is the convergence curve of algorithm 1 and algorithm 2 on Isolet data set;
Figure 17 is the convergence curve of algorithm 1 and algorithm 2 on Prostate_GE data set;
Figure 18 is the convergence curve of algorithm 1 and algorithm 2 on Lung data set;
Figure 19 is the convergence curve of algorithm 1 and algorithm 2 on SRBCT data set.
Specific embodiment
For a clearer understanding of the technical characteristics, objects and effects of the present invention, now control attached drawing is described in detail A specific embodiment of the invention.
Referring to FIG. 1, it realizes the method flow diagram of unsupervised feature selecting, a kind of base disclosed by the invention for the present invention In the unsupervised feature selection approach that dictionary and sample similar diagram learn, comprising the following steps:
S1, a raw data matrix X={ x is given1,x2,…xn}={ f1;f2…;fd}∈Rd×n;Wherein:
N is number of samples, and i ∈ n, d are feature quantity, j ∈ d;xi∈Rd×1That indicate is i-th of sample sample, fj∈Rd ×1It is j-th of feature vector;
S2, study have the dictionary D ∈ R of m base vectord×m, utilize the raw data matrix X that dictionary D is given to step S1 It is reconstructed, obtains a new dictionary base space W;Wherein, l is utilized2,1Norm applies row sparsity to W, so that from dictionary D In select corresponding base vector to indicate each sample;
In fact, Yang et al. has confirmed that the base of different number will affect the quality of W, it is disclosed by the invention in order to make Algorithm model can export feature importance, in the present embodiment, only be provided with the radix equal with the dimension of feature.About base It can be indicated with the objective function of space learning are as follows:
Wherein,Square of the F norm of representing matrix " ";β is one and takes positive balance parameters;The i-th column of dictionary D It is expressed as di, wherein constraint condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;||W||2,1Table Show and utilizes l2,1Norm applies row sparsity to dictionary base space W;M is the positive integer greater than 1.
In the present embodiment, l is used2,1Norm and do not have to the l in traditional dictionary learning model1Norm, main cause have two It is a: first, l1Norm results in the sparsity of element, not can reflect the importance of different characteristic, on the contrary, l2,1Norm can be with Row sparsity is applied to W, distance can be measured in characteristic dimension by regularization.Second, row sparsity is forced on W, Closely related base can be selected to indicate each sample from D, also the local attribute of available initial data.
S3, using Laplce's restricted model based on graph model, the local geometric retained between raw data matrix X closes System, and define constraint regularization term:
Wherein, SijIt is data point xiAnd xjBetween similarity measure;L=A-S is figure Laplacian Matrix, and A is one right Angular moment battle array, whereinS is similarity matrix;wiOr wjIt is arranged for the i-th or j of W.
Previous in research, calculated between data sample usually using a certain distance measurement (such as gaussian kernel function) Similitude.However, in the follow-up process, the similar matrix obtained in the initial data is pre-defined and fixed, and is showed The noise sample and feature for including in real data inevitably make similar matrix become unreliable.We attempt in optimization process In learn similar matrix automatically, rather than use precalculated similar matrix.
S4, the constraint regularization term defined in conjunction with the obtained dictionary base space W of step S2 and step S3, in dictionary base In space W, feature selecting is carried out, specifically: construction DGL-UFS model objective function solves and obtains W;Wherein, excellent using iteration Change strategy, by carrying out constraint to each other to multiple variables in objective function, W is finally calculated.
The DGL-UFS model objective function of definition are as follows:
s.t.||di||2≤1,ST1=1, S >=0, i=1,2 ... m;
Wherein, α is to take positive constant for one of balance chart regular terms.
Due to W={ w1,w2,…,wnIt is that a new space of X uses l in new dictionary base space2,1Norm is to W The row sparsity of application can be used to the importance of measures characteristic.With the rudimentary representation method phase in the past in original feature space Than DGL-UFS captures higher level and more abstract expression.
Referring to FIG. 2, it is unsupervised feature selecting apparatus structure schematic diagram disclosed by the invention, which includes data Setup module L1, Space Reconstruction module L2, constraint regularization module L3 and feature selection module L4, wherein;
Data setup module L1 is for giving a raw data matrix X={ x1,x2,…xn}={ f1;f2;fd}∈Rd×n; Wherein:
N is number of samples, and i ∈ n, d are feature quantity, j ∈ d;xi∈Rd×1That indicate is i-th of sample sample, fj∈Rd ×1It is j-th of feature vector;
Space Reconstruction module L2 is used to learn the dictionary D ∈ R with m base vectord×m, mould is arranged to data using dictionary D The given raw data matrix X of block is reconstructed, and obtains a new dictionary base space W;Wherein, l is utilized2,1Norm applies W Row sparsity, so that selecting corresponding base vector from dictionary D to indicate each sample;
It constrains regularization module L3 to be used to utilize Laplce's restricted model based on graph model, retains raw data matrix Local geometric relationship between X, and define constraint regularization term;
The dictionary base space W that feature selection module L4 is used to that Space Reconstruction module to be combined to obtain, and constraint regularization mould The constraint regularization term that block defines carries out feature selecting in dictionary base space W, specifically: construction DGL-UFS simulated target Function solves and obtains W;Wherein, using iteration optimization strategy, by carrying out pact to each other to multiple variables in objective function W is finally calculated in beam.
Fig. 3-Fig. 4 is please referred to, is changed for the iterative algorithm figure disclosed by the invention for solving dictionary base space W, and use The algorithm pattern of objective function is solved for optimization algorithm.
The specific steps of objective function are solved by alternating iteration optimisation strategy are as follows:
(1) fixed D and S, updates W:
Fixed D and S, solves W, then objective function is converted are as follows:
Due to the l of W2,1Norm regularization, above formula are convex, but and it is rough, added again in the present embodiment using an iteration Least-squares algorithm is weighed to solve it.Assuming that updating the current of W in the subproblem of W is estimated as Wt, define a diagonal weight Matrix Gt, i-th of diagonal element elementThen W is updated by solving following weighted least-squares problemt+1:
To above formula about W derivation, and the result that derivation is arranged is zero, is obtained:
(DTD+βG)W+αWLS=DTX
Above-mentioned equation is Silvester (Sylvester) equation, due to DTD+ β G is stringent orthotropicity, therefore on Formula has stable solution.
(2) fixed W and S, updates D:
Fixed W and S, solves D, then objective function is converted are as follows:
In order to solve above formula, in the present embodiment, ADMM (Alternating Direction method of is used Multipliers the optimal solution of D) is obtained by introducing a matrix of variables, is obtained:
Pass through the available optimal D of following iterative step:
Y(t+1)=Y(t)+D(t+1)-H(t+1)
Wherein, Y is Lagrange's multiplier, and κ is parameter, and t indicates the number of iterations.
(3) fixed W and D, updates S:
Fixed W and D, solves S, then objective function is converted are as follows:
Due to the similarity vector of each sample be it is independent, for i-th of sample, have:
s.t.1Tsi=1,0≤sij≤1
For the ease of indicating, a matrix is indicated with M in the present embodimentVector mi∈Rn×1, then on Formula can be write as:
By introducing Lagrange's multiplier, available si
In concrete implementation, D is initialized with random matrix, is obtained with traditional gaussian kernel function initialization S:
Wherein, Nk(xi) indicate xiK arest neighbors set, σ be a width parameter (in initialization step, k=5, σ =1).
In order to sufficiently verify the validity of DGL-UFS method of the present invention, first on six common master databases (Yale, Isolet, COIL20, Prostate_GE, Lung, SRBCT) tests the performance of DGL-UFS method, while with following seven Kind unsupervised feature selecting algorithm popular at present is compared:
(1) Baseline: all primitive characters are all used.
(2) LS: Laplce's score feature selecting, this method choose the feature for being best suitable for Laplacian matrix.
(3) MCFS: multiple cluster feature selection, this method use l1Feature selection process specification is turned to spectrum letter by norm Cease regression problem.
(4) SOGFS: while the unsupervised feature selection approach of feature selecting and local Structure learning is carried out, so as to Adaptively determine similar matrix.
(5) CDL-FS: a kind of coupling analysis-synthesis dictionary learning method of new unsupervised feature selecting uses synthesis Dictionary reconstructed sample, and analyze dictionary and parsing coding is carried out to sample and is sample allocation probability.
(6) SCUFS: the unsupervised feature selecting of subspace cluster guidance learns instruction by the subspace clustering based on expression Practice the cluster labels of sample, selection can retain the feature of cluster labels well.
(7) RJGSC: a kind of unsupervised feature selecting of new joint figure sparse coding model, the model is first with Some dictionary learning methods extract the base of training data, then extract initial data and generate new expression in base space.
In experiment, to DGL-UFS method and other seven kinds of unsupervised feature selection approach on database disclosed in six Compare test.Six databases include a face image database (Yale), a voice signal data library (Isolet), an object image data library (COIL20) and three biometric databases (Prostate_GE, Lung and SRBCT). The statistics of these databases is as shown in Figure 5.
Similar to previous unsupervised feature selection approach, K-means collection also is executed using the feature selected in the present invention Group, the evaluation criterion being widely used using two kinds, that is, the accuracy rate clustered (ACC) and normalized mutual information (NMI).ACC and The value of NMI is bigger, then it represents that method performance is better.Assuming that qiIt is cluster result, piIt is true tag, then the definition of ACC is such as Under:
Wherein, if as x=y, δ (x, y)=1, otherwise δ (x, y)=0.map(qi) it is a best mapping letter Number, its function are the true tag progress of experiment is obtained by Kuhn-Munkres algorithm cluster labels and sample Match.
Given two variables P and Q, NMI is defined as:
Wherein, H (P) and H (Q) respectively indicates the entropy of P and Q, and I (P, Q) indicates the mutual information of P and Q between the two.P is defeated Enter the cluster result of sample, Q is their true tag.NMI reflects the consistent degree between cluster result and true tag.
DGL-UFS algorithm and the parameter of other control methods will be configured in an experiment, for LS, MCFS, The size k=5 of neighbour's parameter of all databases is arranged in SOGFS, SCUFS and RJGSC.It is public in order to be carried out to different methods Flat comparison, in the present embodiment, using " grid search " strategy from { 10-3,10-2,10-1,1,10,102,103In adjustment it is all The parameter of method.Due to the optimal number of the feature of selection be it is unknown, for all databases we use " grid search " plan Slightly from the quantity of the different features selected of { 20,30 ..., 90,100 } setting.
It is special to the low-dimensional selected by them using K-means algorithm after different characteristic selection algorithm completes feature selecting Sign is clustered.In view of the performance of K-means cluster will receive the influence of initialization, repeat 20 times it is different it is random just Beginningization experiment, finally records their average value.
Interpretation of result:
Fig. 6 and Fig. 7 gives ACC and NMI value of the distinct methods on six databases.As can be seen that proposed The performance of DGL-UFS is better than other state-of-the-art methods always, this shows that method proposed by the present invention can be captured than original The rudimentary higher level of expression and more abstract expression.Specifically, with RJGSC (RJGSC independently execute dictionary basic learning and Feature selecting) it compares, it has more than 5% to most of data sets (such as Yale, Isolet, Prostate_GE and SRBCT) It improves.
In order to verify influence of the feature selecting to cluster result, the present embodiment illustrates all methods in Fig. 8, Fig. 9 and exists On disparate databases, by the asynchronous performance of selection characteristic.ACC value and NMI value and different data concentrate the number of selected feature It measures related.It can be seen that the performance of this method, which is stablized, is better than other methods in certain characteristic range.It is worth noting that, When characteristic is smaller, compared to Baseline method, the ACC value of DGL-UFS method is higher, this has turned out our feature The advantage of selection method.
Parameter Sensitivity Analysis:
In the present invention containing there are two balance parameters (α and β), in order to study the present invention to the susceptibility of parameter, we are to it In a parameter fixed, a remaining parameter is changed.Specific operating process are as follows: firstly, fixed α= 1, change the value of β, ACC the and NMI value on disparate databases is as shown in Fig. 9, Figure 11.Secondly, fixing Beta=1, changes the value of α, ACC and NMI value on disparate databases is as shown in Figure 12 and Figure 13.
As can be seen that in most cases, DGL-UFS is for different α and β when the quantity of selected feature is fixed Performance is stablized.For data set Prostate_GE, NMI has some sensitivity to β, available best as β=1 NMI value;It is as a result stable as β > 1 for data set SRBCT;DGL-UFS is unwise other than to data set SRBCT Sense, when α > 0.01 and selected characteristic are greater than 80, the NMI on data set SRBCT keeps stablizing.
The time complexity of DGL-UFS algorithm is analyzed:
Through iteration optimization algorithms during solving objective function, the main time spends in three parts: solving W, asks It solves D and solves S.For updating the part W, the classic algorithm of Sylvester equation is Bartels Stewart algorithm, complicated Degree is o (m3), if T1For the number of iterations of algorithm one, then the complexity of iteration time each time in algorithm two is o (T1m3); For updating the part D, if T2For the number of iterations of ADMM algorithm, due to m=d, the iteration time each time in algorithm two is multiple Miscellaneous degree is o (T2(d2n+d3));For updating S portion, prime cost is to calculate distance matrix M, and time complexity is o (n2)。
The analysis of DGL-UFS convergence:
When for solving W, using iteration weight weighted least square algorithm in the present embodiment, in iterative process each time In, an approximate solution is obtained by Silvester (Sylvester) equation;For solving D, ADMM algorithm can guarantee to optimize D converges on optimal solution;For solve S portion, be based on Lagrangian and KKT condition, the optimization of S be also converge on it is optimal Solution.
Convergence curve of the objective function on different data sets is as shown in Figure 14~Figure 19;It can be seen that algorithm 1 is restrained Speed is fast, almost restrains in 5 iteration, and the target value of algorithm 2 can also be substantially reduced with preceding 10 iteration.
The embodiment of the present invention is described with above attached drawing, but the invention is not limited to above-mentioned specific Embodiment, the above mentioned embodiment is only schematical, rather than restrictive, those skilled in the art Under the inspiration of the present invention, without breaking away from the scope protected by the purposes and claims of the present invention, it can also make very much Form, all of these belong to the protection of the present invention.

Claims (10)

1. a kind of unsupervised feature selection approach learnt based on dictionary and sample similar diagram, which is characterized in that including following step It is rapid:
S1, a raw data matrix X={ x is given1,x2,…xn}={ f1;f2…;fd}∈Rd×n;Wherein:
N is number of samples, and d is the quantity of feature vector;I ∈ n, xi∈Rd×1What is indicated is i-th of sample;J ∈ d, fj∈Rd×1 It is j-th of feature vector;
S2, study have the dictionary D ∈ R of m base vectord×m, the raw data matrix X that step S1 gives is carried out using dictionary D Reconstruct, obtains a new dictionary base space W;Wherein, l is utilized2,1Norm applies row sparsity to W, so that selecting from dictionary D Selecting corresponding base vector indicates each sample;
S3, Laplce's restricted model based on graph model, definition constraint regularization term are utilized;
S4, the constraint regularization term defined in conjunction with the obtained dictionary base space W of step S2 and step S3, in dictionary base space W On, feature selecting is carried out, specifically: construction DGL-UFS model objective function solves and obtains W;Wherein, iteration optimization plan is utilized Slightly, by carrying out constraint to each other to multiple variables in objective function, W is finally calculated.
2. unsupervised feature selection approach according to claim 1, which is characterized in that carry out weight to raw data matrix X Structure, resulting reconstruction coefficients matrix W ∈ Rm×nAs newly generated dictionary space;Wherein, empty about base vector and new dictionary base Between the objective function that learns can indicate are as follows:
Wherein,Expression takes square of the F norm of matrix " ";β is one and takes positive balance parameters;The i-th list of dictionary D It is shown as di, wherein constraint condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;||W||2,1It indicates Utilize l2,1Norm applies row sparsity to dictionary base space W;M is the positive integer greater than 1.
3. unsupervised feature selection approach according to claim 1, which is characterized in that constrain canonical defined in step S3 Change item expression formula are as follows:
Wherein, SijIt is data point xiAnd xjBetween similarity measure;L=A-S is figure Laplacian Matrix, and A is one to angular moment Battle array, whereinS is similarity matrix;wiOr wjIt is arranged for the i-th or j of W.
4. unsupervised feature selection approach according to claim 1,2 or 3, which is characterized in that the DGL-UFS model of definition Objective function are as follows:
s.t.||di||2≤1,ST1=1, S >=0, i=1,2 ... m;
Wherein,Square of the F norm of representing matrix " ";α is to take positive constant for one of balance chart regular terms, and β is One takes positive balance parameters;L=A-S is figure Laplacian Matrix, and A is a diagonal matrix, and S is similar matrix;||W||2,1 It indicates to utilize l2,1Norm applies row sparsity to dictionary base space W;The i-th list of dictionary D is shown as di, wherein constraint condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;M is the positive integer greater than 1.
5. unsupervised feature selection approach according to claim 4, which is characterized in that in step S4, pass through alternating iteration The specific steps of optimisation strategy solution objective function are as follows:
S51, initialization dictionary base D and similar matrix S;
S52, preset parameter D and S, objective function is converted are as follows:
Wherein, using an iteration weight weighted least square algorithm, by solving above-mentioned objective function, further undated parameter W;
S53, preset parameter W and S, objective function is converted are as follows:
Wherein, using ADMM-Alternating Direction method of Multipliers algorithm, by introducing one A matrix of variables is into objective function, further undated parameter D;
S54, preset parameter W and D, objective function is converted are as follows:
Wherein, due to the similarity vector of each sample be it is independent, for i-th of sample, current goal function representation are as follows:
By introducing Lagrange's multiplier, the similar matrix s of current i-th of sample is obtainedi;Further, needle for each sample, Using above-mentioned identical method, the similar matrix of each sample is successively updated, realizes the update to parameter S;
S55, step S52-S54 is repeated, when objective function convergence, exports parameter S, D and W under convergent.
6. a kind of unsupervised feature selecting device learnt based on dictionary and sample similar diagram, which is characterized in that the device includes With lower module:
Data setup module, for giving a raw data matrix X={ x1,x2,…xn}={ f1;f2…;fd}∈Rd×n;Its In:
N is number of samples, and d is the quantity of feature vector;I ∈ n, xi∈Rd×1What is indicated is i-th of sample;J ∈ d, fj∈Rd×1 It is j-th of feature vector;
Space Reconstruction module, for learning the dictionary D ∈ R with m base vectord×m, data setup module is given using dictionary D Fixed raw data matrix X is reconstructed, and obtains a new dictionary base space W;Wherein, l is utilized2,1It is dilute that norm applies row to W Property is dredged, so that selecting corresponding base vector from dictionary D to indicate each sample;
Regularization module is constrained, for utilizing Laplce's restricted model based on graph model, definition constraint regularization term;
Feature selection module, the dictionary base space W for being obtained in conjunction with Space Reconstruction module, and constraint regularization module definition Constraint regularization term, in dictionary base space W, carry out feature selecting, specifically: construction DGL-UFS model objective function, ask Solution obtains W;Wherein, using iteration optimization strategy, by carrying out constraint to each other to multiple variables in objective function, finally W is calculated.
7. unsupervised feature selecting device according to claim 6, which is characterized in that carry out weight to raw data matrix X Structure, resulting reconstruction coefficients matrix W ∈ Rm×nAs newly generated dictionary space;Wherein, empty about base vector and new dictionary base Between the objective function that learns can indicate are as follows:
Wherein,Expression takes square of the F norm of matrix " ";β is one and takes positive balance parameters;The i-th list of dictionary D It is shown as di, wherein constraint condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;||W||2,1It indicates Utilize l2,1Norm applies row sparsity to dictionary base space W;M is the positive integer greater than 1.
8. unsupervised feature selecting device according to claim 6, which is characterized in that defined in constraint regularization module Constrain regularization term expression formula are as follows:
Wherein, SijIt is data point xiAnd xjBetween similarity measure;L=A-S is figure Laplacian Matrix, and A is one to angular moment Battle array, whereinS is similarity matrix;wiOr wjIt is arranged for the i-th or j of W.
9. the unsupervised feature selecting device according to claim 6,7 or 8, which is characterized in that the DGL-UFS model of definition Objective function are as follows:
s.t.||di||2≤1,ST1=1, S >=0, i=1,2 ... m;
Wherein,Square of the F norm of representing matrix " ";α is to take positive constant for one of balance chart regular terms, and β is One takes positive balance parameters;L=A-S is figure Laplacian Matrix, and A is a diagonal matrix, and S is similar matrix;||W||2,1 It indicates to utilize l2,1Norm applies row sparsity to dictionary base space W;The i-th list of dictionary D is shown as di, wherein constraint condition | | di||2≤ 1 makes the atom in the dictionary space learnt become compact;M is the positive integer greater than 1.
10. unsupervised feature selection approach according to claim 9, which is characterized in that in feature selection module, pass through friendship The specific steps of objective function are solved for iteration optimization strategy are as follows:
S51, initialization dictionary base D and similar matrix S;
S52, preset parameter D and S, objective function is converted are as follows:
Wherein, using an iteration weight weighted least square algorithm, by solving above-mentioned objective function, further undated parameter W;
S53, preset parameter W and S, objective function is converted are as follows:
Wherein, using ADMM-Alternating Direction method of Multipliers algorithm, by introducing one A matrix of variables is into objective function, further undated parameter D;
S54, preset parameter W and D, objective function is converted are as follows:
Wherein, due to the similarity vector of each sample be it is independent, for i-th of sample, current goal function representation are as follows:
By introducing Lagrange's multiplier, the similar matrix s of current i-th of sample is obtainedi;Further, needle for each sample, Using above-mentioned identical method, the similar matrix of each sample is successively updated, realizes the update to parameter S;
S55, step S52-S54 is repeated, when objective function convergence, exports parameter S, D and W under convergent.
CN201910441208.XA 2019-05-24 2019-05-24 A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram Pending CN110348287A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910441208.XA CN110348287A (en) 2019-05-24 2019-05-24 A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910441208.XA CN110348287A (en) 2019-05-24 2019-05-24 A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram

Publications (1)

Publication Number Publication Date
CN110348287A true CN110348287A (en) 2019-10-18

Family

ID=68174629

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910441208.XA Pending CN110348287A (en) 2019-05-24 2019-05-24 A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram

Country Status (1)

Country Link
CN (1) CN110348287A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111027636A (en) * 2019-12-18 2020-04-17 山东师范大学 Unsupervised feature selection method and system based on multi-label learning
CN111813982A (en) * 2020-07-23 2020-10-23 中原工学院 Data processing method and device based on subspace clustering algorithm of spectral clustering
CN113486980A (en) * 2021-07-29 2021-10-08 中南大学 Aluminum electrolysis cell condition identification method, system and equipment based on self-expression dictionary pair
WO2022166362A1 (en) * 2021-02-03 2022-08-11 浙江师范大学 Unsupervised feature selection method based on latent space learning and manifold constraints

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111027636A (en) * 2019-12-18 2020-04-17 山东师范大学 Unsupervised feature selection method and system based on multi-label learning
CN111813982A (en) * 2020-07-23 2020-10-23 中原工学院 Data processing method and device based on subspace clustering algorithm of spectral clustering
CN111813982B (en) * 2020-07-23 2021-09-03 中原工学院 Data processing method and device based on subspace clustering algorithm of spectral clustering
WO2022166362A1 (en) * 2021-02-03 2022-08-11 浙江师范大学 Unsupervised feature selection method based on latent space learning and manifold constraints
CN113486980A (en) * 2021-07-29 2021-10-08 中南大学 Aluminum electrolysis cell condition identification method, system and equipment based on self-expression dictionary pair
CN113486980B (en) * 2021-07-29 2023-02-24 中南大学 Aluminum electrolysis cell condition identification method, system and equipment based on self-expression dictionary pair

Similar Documents

Publication Publication Date Title
CN110348287A (en) A kind of unsupervised feature selection approach and device based on dictionary and sample similar diagram
Sidi et al. Unsupervised co-segmentation of a set of shapes via descriptor-space spectral clustering
CN108427921A (en) A kind of face identification method based on convolutional neural networks
CN108171122A (en) The sorting technique of high-spectrum remote sensing based on full convolutional network
Tang et al. Tensor multi-elastic kernel self-paced learning for time series clustering
CN108021930B (en) Self-adaptive multi-view image classification method and system
JP2002513979A (en) System, method, and computer program product for representing proximity data in multidimensional space
CN107203787A (en) A kind of unsupervised regularization matrix characteristics of decomposition system of selection
CN112949740B (en) Small sample image classification method based on multilevel measurement
Dai et al. Metric imitation by manifold transfer for efficient vision applications
CN109325513A (en) A kind of image classification network training method based on magnanimity list class single image
Bhute et al. Content based image indexing and retrieval
CN111695455B (en) Low-resolution face recognition method based on coupling discrimination manifold alignment
CN111444802B (en) Face recognition method and device and intelligent terminal
CN110096991A (en) A kind of sign Language Recognition Method based on convolutional neural networks
CN112232374A (en) Irrelevant label filtering method based on depth feature clustering and semantic measurement
Xu et al. A novel image feature extraction algorithm based on the fusion AutoEncoder and CNN
Zhao et al. Hyperspectral target detection method based on nonlocal self-similarity and rank-1 tensor
Zhang et al. SaltISCG: Interactive salt segmentation method based on CNN and graph cut
CN114898167A (en) Multi-view subspace clustering method and system based on inter-view difference detection
Afzal et al. Discriminative feature abstraction by deep L2 hypersphere embedding for 3D mesh CNNs
CN109993208A (en) A kind of clustering processing method having noise image
CN116229179A (en) Dual-relaxation image classification method based on width learning system
Ji et al. Hyperspectral image classification based on unsupervised regularization
Wang et al. Conscience online learning: an efficient approach for robust kernel-based clustering

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20191018

RJ01 Rejection of invention patent application after publication