CN108520279A - A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part - Google Patents

A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part Download PDF

Info

Publication number
CN108520279A
CN108520279A CN201810326062.XA CN201810326062A CN108520279A CN 108520279 A CN108520279 A CN 108520279A CN 201810326062 A CN201810326062 A CN 201810326062A CN 108520279 A CN108520279 A CN 108520279A
Authority
CN
China
Prior art keywords
matrix
sample
semi
supervised
sparse
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810326062.XA
Other languages
Chinese (zh)
Inventor
黄冬梅
张明华
张晓桐
邹亚晴
李永兰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Maritime University
Shanghai Ocean University
Original Assignee
Shanghai Maritime University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Maritime University filed Critical Shanghai Maritime University
Priority to CN201810326062.XA priority Critical patent/CN108520279A/en
Publication of CN108520279A publication Critical patent/CN108520279A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/211Selection of the most significant subset of features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24147Distances to closest patterns, e.g. nearest neighbour classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/30Noise filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/513Sparse representations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/58Extraction of image or video features relating to hyperspectral data

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)

Abstract

The present invention relates to a kind of semi-supervised dimension reduction methods of high spectrum image of the sparse insertion in part, the described method comprises the following steps:Step S1. sets higher dimensional space RDIn there are data set X={ x1, x2..., xl, xl+1..., xl+u, l+u=N, wherein preceding l sample XlIt is N to have category sample, class label c, Different categories of samples numberi, i=(1,2 ..., c), rear u sample XuIt is no category sample;Step S2. builds sparse coefficient matrix S by rarefaction representation;Step S3. is based on the sparse embedded projection algorithm construction projection matrix W in semi-supervised part;Step S4. finds out lower-dimensional subspace Y=W according to projection matrix WTX={ y1, y2..., yN}.Its advantage is shown:By carrying out the semi-supervised dimensionality reduction of the sparse insertion in part to high spectrum image, the category information of data was both utilized, has also maintained data local characteristics and reduces the noise information of image, to improve the nicety of grading of image.

Description

A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part
Technical field
The present invention relates to the dimensionality reduction technology fields of high spectrum image, specifically, being a kind of bloom of the sparse insertion in part Spectrogram is as semi-supervised dimension reduction method.
Background technology
The spatial resolution of high-spectrum remote sensing is high, and object spectrum curve is continuous, has stronger terrain classification and knowledge Other ability, in fields such as geologic prospect, medical detection, life science, forensic identification, military surveillance, environmental monitoring, precision agricultures It is widely used.High-spectrum seems the remote sensing images obtained by imaging spectrometer, with imaging spectrometer space and The raising of resolution ratio index between spectrum, while having the high-spectrum remote sensing data volume of two-dimensional space and one-dimensional spectral information in finger Number increases.The drastically expansion of data volume not only brings huge difficulty to data storage with transmission, while also exacerbating data The complexity of processing procedure reduces data-handling efficiency, or even will appear dimension disaster problem.Dimensionality reduction can be high dimensional data A low-dimensional representation method for including sample important information is found, is the pretreated important link of classification hyperspectral imagery.Dimensionality reduction The useful information of data can be not only preserved, and data volume can be greatly reduced, and it is possible to prevente effectively from dimension disaster, makes Data indicate simpler clear, are conducive to subsequent classification.
Whether considered according to sample data or use classes label, traditional dimension reduction method can be divided into three classes:Supervision drop Dimension, unsupervised dimensionality reduction and semi-supervised dimensionality reduction.Unsupervised dimensionality reduction using the spatial relationship of data come keep in data manifold knot Structure, representative algorithm include that Principal Component Analysis (principle component analysis, PCA) and part retain throwing Shadow (locality preserving projections, LPP) carries out unsupervised dimensionality reduction to high spectrum image, it cannot be guaranteed that from Distance is constant between sample data after higher-dimension to low dimension projective.Having supervision dimensionality reduction to refer to one has the projection of distinctive to be dropped Dimension, representative algorithm include local sensitivity discriminant analysis (Locality Sensitive Discriminant Analysis, LSDA) and neighborhood keeps embedded (Neighborhood Preserving Embedding, NPE), has to high spectrum image Dimensionality reduction is supervised, the category of image needs a large amount of manpower, material resources, and there are category data to be difficult to obtain.Semi-supervised dimensionality reduction being capable of profit With the distributed intelligence of data, the structural information of data and a small amount of label information come the performance of boosting algorithm, representative drop Dimension algorithm include semi-supervised locality preserving projections (Semi-Supervised Locality Preserving Projection, SSLPP) and based on semi-supervised feature extraction (the Semi-Supervised Dimensionality Reduction constrained in pairs Based on pairwise constraint propagation, SSDR-PCP).
Although above algorithm achieves preferable effect in some applications, all there is a disadvantage in them:It is right Noise not robust.
High spectrum image field, due to high spectrum image band class information more than and redundancy characteristic, need introduce category letter It ceasing to improve nicety of grading, the cost of category information is big and is not easy to obtain, though and semi-supervised dimension-reduction algorithm traditional at present Semi-supervised thought is so utilized, but only considered the part of properties of data, not can guarantee the local characteristics of data and does not examine Consider the influence of the noise information of image data, therefore, the semi-supervised dimension reduction method based on the sparse insertion in part is worth research.
Chinese patent literature CN201310565426.7, the applying date 20131115, patent name is:One kind is dilute based on part The face emotion identification method for dredging presentation class device, discloses a kind of face emotion recognition based on local rarefaction representation grader Method, it is characterised in that including step:Acquire Facial Expression Image;The feature of Gabor wavelet transition structure Facial Expression Image Vector;Feature is selected using feature selecting algorithm MFCS;Emotional category is identified using the grader of local rarefaction representation.
There is a kind of face emotion identification method based on local rarefaction representation grader of above patent document emotion to know Other accuracy rate is high, and emotion recognition is fast, it is insensitive to face the advantages that.But about a kind of category information not only utilizing sample, but also The dimension reduction method of the sparse insertion of sample data local message is considered, the distance between similar sample can be minimized, is maximized not Distance between similar sample, and the extraction of the local message of data can be kept and reduce the dry of the noise information of image It disturbs, is then disclosed without corresponding to improve the technical solution of nicety of grading of sample.
In conclusion in the higher-dimension inhomogeneous data have partially overlap or lean on it is close, although LPP has local reservation Characteristic, but LPP does not utilize the category information of sample, LPP different classes can be projected to together, can obtain undesirable point Class precision.Class label information is utilized in semi-supervised thought, there is no neighborhood information is considered, cannot reflect data set well Partial structurtes.Dimensionality reduction is carried out to high spectrum image at present, the feature of extraction can be influenced by noise information, seriously affect point Class effect.And studies have shown that the local message of data for dimensionality reduction and classification be all beneficial, traditional LPP dimension-reduction algorithms only It is single from the local message angle of image data, does not filter the noise information of image.One kind is needed both to utilize sample This category information can minimize between similar sample it is contemplated that the dimension reduction method of the sparse insertion of sample data local message Distance, maximize inhomogeneity sample between distance, and can keep the local message of data extraction and reduce image Noise information interference, to improve the semi-supervised dimensionality reduction of high spectrum image of the sparse insertion in the part of nicety of grading of sample Method, and the semi-supervised dimension reduction method of high spectrum image about the sparse insertion in this part yet there are no report.
Invention content
The purpose of the present invention is being directed to deficiency in the prior art, provide it is a kind of not only utilizing the category information of sample, but also examine The dimension reduction method of the sparse insertion of sample data local message is considered, the distance between similar sample can be minimized, has maximized different Distance between class sample, and the extraction of the local message of data can be kept and reduce the interference of the noise information of image, The semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part to improve the nicety of grading of sample.
Another object of the present invention is:A kind of semi-supervised dimension reduction method of high spectrum image using the sparse insertion in part is provided Technology path
To achieve the above object, the technical solution adopted by the present invention is that:
A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part, the described method comprises the following steps:
Step S1. sets higher dimensional space RDIn there are data set X={ x1, x2..., xl, xl+1..., xl+u, l+u=N, wherein Preceding l sample XlIt is N to have category sample, class label c, Different categories of samples numberi, i=(1,2 ..., c), rear u sample XuIt is Without category sample;
Step S2. builds sparse coefficient matrix S by rarefaction representation;
Step S3. is based on the sparse embedded projection algorithm construction projection matrix W in semi-supervised part;
Step S4. finds out lower-dimensional subspace Y=W according to projection matrix WTX={ y1,y2,…,yN}。
As a kind of perferred technical scheme, following steps are specifically included in step S2:
Step S21. is for any one data xiLinear combination is carried out, obtaining linear combination is:
xi=si,1x1+…+si,i-1xi-1+si,i+1xi+1+…+si,nxn
Wherein:si=[sI, 1,…sI, i-1,0,sI, i+1,…,si,n]TFor coefficient matrix, si,jIndicate sample xiReconstruct system Number;
Step S22. builds mathematical model, and the mathematical model built is as follows:
s.t.xi=Xsi
||si||0Indicate siL0Norm, for weighing siSparsity;
Step S23. is by l0Norm minimum problem is converted into l1Norm minimum problem:
s.t.xi=Xsi1=1Tsi
Wherein, 1 complete 1 vector is indicated, | | si||1Indicate siL1Norm;
Step S24. obtains sparse coefficient matrix s=[s by the calculation formula in step S231,s2,…sn];
As a kind of perferred technical scheme, following steps are specifically included in step S3:
Step S31. improves semi-supervised similar weight matrix;
Wherein, qijFor the element of improved semi-supervised similar weight matrix Q, aij=exp (- | | xi-xj||2/ σ), σ is institute There are square of Euclidean distance average value between sample pair, J (xi) it is xiK neighbours domain, k be neighbour's parameter;aijFor partial weight value, Enable (1+aij) it is that weights, (1-a are differentiated in classij) weights are differentiated between class;
Step S32. calculates the diagonal matrix D of similar weight matrix, according to sparse by improving semi-supervised similar weight matrix Coefficient matrix S, to calculate Laplacian Matrix L*, utilize Laplacian Matrix L*Construct the sparse embedded projection in semi-supervised part The target function type of algorithm, formula are as follows:
Wherein, L*=D-QST-SQ+SDST, S=[s1,s2,…,sn];Matrix D is the diagonal matrix of Q,
Diagonal matrix element is Dii=∑jqij
Step S33. introduces constraints WTXXTW=I, wherein I are unit matrix, introduce method of Lagrange multipliers by step Formula overall goals function is converted into as follows in S32:
F (W)=WTXL*XTW-λ(WTXXTW-I)
Derivation, such as following formula are carried out to the W of the formula:
It enablesIt obtains as follows:
XL*XTW=λ XXTW
Wherein λ is generalized eigenvalue, passes through the corresponding feature vector of a maximum eigenvalue before solution, composition projection matrix W =[w1,w2,...,wa]。
To realize above-mentioned second purpose, the technical solution adopted by the present invention is that:
A kind of semi-supervised dimensionality reduction of high spectrum image using the sparse insertion in part described at least one above-described embodiment Technology path, it is specific as follows:
The first step, existing high spectrum image, the interpretation figure marked according to expert refers to legend, from this width Different classes of sample data { x is extracted in image1, x2, ..., xN, the input number as high-spectrum image dimensionality reduction of the present invention According to;
The initial data obtained in the first step is carried out simple data prediction by second step;
Third walks, and sparse coefficient matrix S is built by rarefaction representation;
4th step solves sample data the Euclidean distance dist=between them | | xi-xj||;
5th step, by the 4th step solve Euclidean distance, according in second step to sample add category information the case where, It is divided into 5 kinds of situations to solve, finally obtains the similar weight matrix Q between sample point;
6th step, the similar weight matrix Q acquired by the 5th step according to similar weight matrix Q and acquire diagonal matrix D;
7th step, the value acquired by third step and the 6th step bring L into*=D-QST-SQ+SDST
8th step brings the value that the 7th step acquires into following formula:
Wherein, L*=D-QST-SQ+SDST, S=[s1,s2,…,sn], matrix D is the diagonal matrix of Q, diagonal matrix element For Dii=∑jqij
Sample data dimension is arranged and according to formula XL in 9th step*XTW=λ XXTW finds out transition matrix W=(w1, w2..., wd);
Tenth step, calculation formula yi=WT×xi, find out sample data Y={ y in lower dimensional space1,y2,…,yN};
11st step is added category information as training sample data to a few sample after dimensionality reduction, does not add in conjunction with most The test sample data of category classify to the sample after dimensionality reduction using nearest neighbor classifier, by classification results and known master The Comparative result of componential analysis PCA, local retaining projection LPP, calculate nicety of grading;
As a kind of perferred technical scheme, select number of tags for 50labels, when d=15, k=24, nicety of grading is most It is high.
As a kind of perferred technical scheme, the calculation formula of nicety of grading is as follows:NErr=sum (class~= classLabel);Rate=1-nErr/length (class);Class be by nearest neighbor algorithm find out come test sample Category, classLabel is the category for the test sample known before dimensionality reduction, and nErr is error rate, and rate is identification Rate.
As a kind of perferred technical scheme, pretreatment includes:Sample data adds category X={ x1, x2..., xl, xl+1..., xl+u, data normalization processing.
The invention has the advantages that:
1. once following technique effect may be implemented input application in the present invention:This programme is by carrying out high spectrum image The semi-supervised dimensionality reduction of the sparse insertion in part, had both been utilized the category information of data, and had also maintained data local characteristics and reduce The noise information of image, to improve the nicety of grading of image.
2. algorithm proposed by the present invention makes full use of the local retention performance of LPP algorithms, in the similar weight square of LPP algorithms Introduced in battle array has category sample data and largely without category sample data on a small quantity, and in the algorithm believes neighbour between sample data The difference combination situation of breath and category information considers respectively, including the 1) similitude between similar label neighbour sample number strong point; 2) similitude between inhomogeneity label neighbour sample number strong point;3) there is category and without the phase between category neighbour's sample number strong point Like property;4) without category and without the similitude between category neighbour's sample number strong point;
3. a kind of semi-supervised dimension-reduction algorithm of high spectrum image of the sparse insertion in part of the present invention, being suitable for processing has wave The high spectrum image that section is more and redundancy, Spectral correlation are strong, data volume is huge, dimension is high, dimensionality reduction greatly reduce answering for calculating Miscellaneous degree reduces the differentiation error caused by redundancy, improves the classification performance of image, and traditional high spectrum image Not only store and transmit it is of high cost, in practical applications handle higher-dimension data also bring along many problems;
4. matching noise is removed using sparse matrix, to effective filter out noise information during study, to noise Robust;
5. the present invention proposes a kind of category information both utilizing sample, it is contemplated that sample data local message is sparse embedding The dimension reduction method entered can minimize the distance between similar sample, maximize the distance between inhomogeneity sample, and can keep The extraction of the local message of data and the interference of the noise information of reduction image, to improve the nicety of grading of sample;
6. by rarefaction representation, base signal matrix that eshaustibility may be less indicates initial data and at the same time subtracting as far as possible Few reconstruction error;
7. improved semi-supervised similar weight matrix has the following advantages:(1)0≤aij≤ 1,1≤(1+aij)≤2,0 ≤(1-aij)≤1, differentiate between class in weights and class differentiate weights by noise be compressed in it is certain within the scope of;(2) by aij= exp(-||xi-xj||2/ σ) (σ is square of Euclidean distance average value between all data) it is found that with Euclidean distance reduction, Partial weight value increases, and differentiates that weights reduce between class, differentiates that weights increase in class so that the similar weights of similar exemplar are got over Greatly rather than the weights of similar exemplar are accordingly reduced.
Description of the drawings
Attached drawing 1 is rarefaction representation schematic diagram.
2 dimensionality reduction technology route schematic diagram of attached drawing.
Attached drawing 3 is that number of tags is 20labels and the result of PCA, LPP carry out contrast schematic diagram.
Attached drawing 4 is that number of tags is 30labels and the result of PCA, LPP carry out contrast schematic diagram.
Attached drawing 5 is that number of tags is 40labels and the result of PCA, LPP carry out contrast schematic diagram.
Attached drawing 6 is that number of tags is 50labels and the result of PCA, LPP carry out contrast schematic diagram.
Specific implementation mode
It elaborates below in conjunction with the accompanying drawings to specific implementation mode provided by the invention.
The present invention proposes a modified hydrothermal process:The semi-supervised dimensionality reduction of high spectrum image based on the sparse insertion in part is calculated The key of method, the algorithm is that the partial structurtes using LPP keep and introduce category information and reduce noise information, constructs One new projection target function, to obtain the projection matrix based on ISWSSFE rarefaction representations.This method can be more fully It extracts local discriminant information and effectively filters noise information so that classification hyperspectral imagery precision is obviously improved.
If higher dimensional space RDIn there are data set X={ x1, x2..., xl, xl+1..., xl+u, l+u=N, wherein preceding l sample This XlIt is N to have category sample, class label c, Different categories of samples numberi, i=(1,2 ..., c), rear u sample XuIt is no category Sample.
1, rarefaction representation
The target of rarefaction representation indicates initial data and at the same time reducing as far as possible with base signal matrix as few as possible Reconstruction error.For any one data xi, it can be by other samples (in addition to itself) linear combination:
xi=si,1x1+…+si,i-1xi-1+si,i+1xi+1+…+si,nxn (1)
Wherein:si=[sI, 1,…sI, i-1,0,sI, i+1,…,si,n]TFor coefficient matrix, si,jIndicate sample xiReconstruct system Number.
Mathematical model is as follows:
s.t.xi=Xsi
||si||0Indicate siL0Norm, for weighing siSparsity, but minimize l0Norm problem is a NP- Hard problems solve difficult.It therefore, can be by l0Norm minimum problem is converted into l1Norm minimum problem:
s.t.xi=Xsi1=1Tsi
Wherein, 1 complete 1 vector is indicated, | | si||1Indicate siL1Norm.By solving formula (3), it can be deduced that sparse coefficient Matrix s=[s1,s2,…sn]。
2, the sparse embedded projection algorithm (ISWSSFE) in semi-supervised part
LPP is unsupervised locality preserving projections dimension-reduction algorithm, and the category information of sample is not added in similar weight matrix, Therefore very sensitive to noise.The present invention utilizes the category information of sample, it is proposed that improved semi-supervised similar weight matrix:
Wherein, qijFor the element of improved semi-supervised similar weight matrix Q, aij=exp (- | | xi-xj||2/ σ), σ is institute There is square of Euclidean distance average value between sample pair.J(xi) it is xiK neighbours domain, k be neighbour's parameter.aijFor partial weight value, Enable (1+aij) it is that weights, (1-a are differentiated in classij) differentiate that weights, improved semi-supervised similar weighted value consider sample between class Class label information, which reflects the local neighborhood structure and category information of sample data.
The advantages of improved semi-supervised similar weight matrix:(1)0≤aij≤ 1,1≤(1+aij)≤2,0≤(1-aij)≤ 1, differentiate between class in weights and class differentiate weights by noise be compressed in it is certain within the scope of;(2) by aij=exp (- | | xi-xj| |2/ σ) (σ is square of Euclidean distance average value between all data) it is found that with Euclidean distance reduction, partial weight value increase Greatly, differentiate that weights reduce between class, differentiate that weights increase in class so that the similar weights of similar exemplar are bigger rather than similar The weights of exemplar are accordingly reduced.
Based on improved semi-supervised similar weight matrix, improved semi-supervised localized target function is provided, the throwing of LPP is made Shadow matrix W can keep the classification information of sample in the case where ensureing that part retains and be made an uproar to image using rarefaction representation Acoustic intelligence is filtered, and formula is as follows:
Wherein, L*=D-QST-SQ+SDST, S=[s1,s2,...,sn].Matrix D is the diagonal matrix of Q, opposite angle matrix element Element is Dii=∑jqij
Introduce constraints WTXXTW=I, wherein I are unit matrix.
Introducing method of Lagrange multipliers converts formula (5) overall goals function to as follows:
F (W)=WTXL*XTW-λ(WTXXTW-I) (6)
Derivation, such as following formula are carried out to the W of formula (6):
It enablesFormula (7) is converted into formula (8), as follows:
XL*XTW=λ XXTW (8)
Projection matrix W is actually to solve generalized eigenvalue and feature vector problem, and wherein λ is generalized eigenvalue, is led to Cross the corresponding feature vector of the preceding a maximum eigenvalue of solution, composition projection matrix W=[w1,w2,...,wa], low-dimensional is empty after projection Between data be:
Y=WTX
Input:Sample data setParameter is k
Output:Dimensionality reduction matrix W=[w1,w2,...,wa] and lower-dimensional subspace Y=WTX={ y1,y2,…,yN}
Steps are as follows for the realization of algorithm:
(1) according to aij=exp (- | | xi-xj||2/ σ) the improved semi-supervised similar weight matrix Q of construction;
(2) by improving semi-supervised similar weight matrix, the diagonal matrix D of similar weight matrix is calculated;
(3) sparse coefficient matrix S is constructed according to formula (3), to calculate Laplacian Matrix L*
(4) Laplacian Matrix L is utilized*Construct the target function type (5) of the sparse embedded projection algorithm in semi-supervised part;
(5) Generalized Characteristic Equation is solved according to formula (8), the corresponding feature vector of a maximum eigenvalue before obtaining is constituted and thrown Shadow matrix W;
(6) lower-dimensional subspace Y=W is found out according to formula (9)TX={ y1,y2,…,yN}。
The present invention proposes the semi-supervised dimension-reduction algorithm (ISWSSFE) of the high spectrum image based on the sparse insertion in part, right Pavia University high spectrum images are handled, take respectively number of tags be 20labels, 30labels, 40labels, 50labels is tested, and is compared with the result of PCA, LPP, and d is dimension, k be neighbour's parameter (k 24, d be 5~ 50), as a result as seen in figures 3-6, from the line chart of 20labels, 30labels, 40labels and 50labels it can be found that When number of tags is 50labels, when d=15, k=24, the discrimination of sample is up to 93.56%.
Fig. 2 substantially illustrates the technology path of the dimensionality reduction of the high spectrum image based on the technology of the present invention.
The first step, existing high spectrum image, the interpretation figure marked according to expert refers to legend, from this width Different classes of sample data { x is extracted in image1, x2, ..., xN, the input number as high-spectrum image dimensionality reduction of the present invention According to.
The initial data (103 * 1800 samples of dimension) obtained in the first step is carried out simple data and located in advance by second step Reason.Pretreatment includes:Sample data adds category X={ x1, x2..., xl, xl+1..., xl+u, data normalization processing.
Third walks, and the sparse coefficient matrix S (1800*1800) of training sample is acquired according to formula (3).
4th step solves sample data Euclidean distance (1800*1800) dist=between them | | xi-xj||。
5th step, by the 4th step solve Euclidean distance, according in second step to sample add category information the case where, It is divided into 5 kinds of situations to solve, finally obtains the similar weight matrix Q (1800*1800) between sample point as shown in formula (4).
6th step, the similar weight matrix Q acquired by the 5th step according to similar weight matrix Q and acquire diagonal matrix D。
7th step, the value acquired by third step and the 6th step bring L into*=D-QST-SQ+SDST, acquire
8th step brings the value that the 7th step acquires into formula (5).
9th step, setting sample data dimension d (value range is 5~50), and according to formula (8), find out transition matrix W =(w1, w2..., wd) (103*d dimensions).
Tenth step, calculation formula yi=WT×xi, find out sample data Y={ y in lower dimensional space1,y2,…,yN(d ties up * 1800)。
11st step is added category information as training sample data to a few sample after dimensionality reduction, does not add in conjunction with most The test sample data of category classify to the sample after dimensionality reduction using nearest neighbor classifier, by classification results with it is known The Comparative result of (PCA, LPP) utilizes formula nErr=sum (class~=classLabel);Rate=1-nErr/ Length (class) (class be by nearest neighbor algorithm find out come test sample category, classLabel is in dimensionality reduction The category for the test sample known before, nErr are error rates, and rate is discrimination) calculate nicety of grading.
12nd step, it can be found that working as from the line chart of 20labels, 30labels, 40labels and 50labels Number of tags is 50labels, and when d=15, k=24, the discrimination of sample is up to 93.56%.
The present invention once puts into application, and following technique effect may be implemented:This programme passes through to high spectrum image carry out office The category information of data had both been utilized in the semi-supervised dimensionality reduction of the sparse insertion in portion, also maintains data local characteristics and reduces figure The noise information of picture, to improve the nicety of grading of image.Algorithm proposed by the present invention makes full use of the part of LPP algorithms to protect Characteristic is held, introduced in the similar weight matrix of LPP algorithms has category sample data and largely without category sample data on a small quantity, and The difference combination situation of neighbor information between sample data and category information is considered respectively in the algorithm, including 1) similar label Similitude between neighbour's sample number strong point;2) similitude between inhomogeneity label neighbour sample number strong point;3) have category and Without the similitude between category neighbour's sample number strong point;4) without category and without the similitude between category neighbour's sample number strong point; A kind of semi-supervised dimension-reduction algorithm of high spectrum image of the sparse insertion in part of the present invention, being suitable for processing has wave band more and superfluous High spectrum image remaining, Spectral correlation is strong, data volume is huge, dimension is high, dimensionality reduction greatly reduce the complexity of calculating, subtract Lack the differentiation error caused by redundancy, improves the classification performance of image, and traditional high spectrum image not only stores With transmission cost height, the data for handling higher-dimension in practical applications also bring along many problems;It goes to match using sparse matrix Noise, to effective filter out noise information during study, to noise robustness;The present invention proposes one kind and both utilizing sample This category information can minimize between similar sample it is contemplated that the dimension reduction method of the sparse insertion of sample data local message Distance, maximize inhomogeneity sample between distance, and can keep the local message of data extraction and reduce image Noise information interference, to improve the nicety of grading of sample;By rarefaction representation, base signal square that eshaustibility may be less Battle array indicates initial data and at the same time reducing reconstruction error as far as possible;Having for improved semi-supervised similar weight matrix is following Advantage:(1)0≤aij≤ 1,1≤(1+aij)≤2,0≤(1-aij)≤1 differentiates between class in weights and class and differentiates weights by noise Within the scope of being compressed in centainly;(2) by aij=exp (- | | xi-xj||2/ σ) (σ is Euclidean distance average value between all data Square) it is found that with Euclidean distance reduction, partial weight value increases, differentiates that weights reduce between class, differentiate that weights increase in class Greatly so that the similar weights of similar exemplar are bigger rather than the weights of similar exemplar are accordingly reduced.
The above is only a preferred embodiment of the present invention, it is noted that for the ordinary skill people of the art Member, under the premise of not departing from the method for the present invention, can also make several improvement and supplement, these are improved and supplement also should be regarded as Protection scope of the present invention.

Claims (7)

1. a kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part, which is characterized in that the method includes following Step:
Step S1. sets higher dimensional space RDIn there are data set X={ x1, x2..., xl, xl+1..., xl+u, l+u=N, wherein first l Sample XlIt is N to have category sample, class label c, Different categories of samples numberi, i=(1,2 ..., c), rear u sample XuIt is no class Standard specimen sheet;
Step S2. builds sparse coefficient matrix S by rarefaction representation;
Step S3. is based on the sparse embedded projection algorithm construction projection matrix W in semi-supervised part;
Step S4. finds out lower-dimensional subspace Y=W according to projection matrix WTX={ y1,y2,…,yN}。
2. the semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part according to claim 1, which is characterized in that step Following steps are specifically included in rapid S2:
Step S21. is for any one data xiLinear combination is carried out, obtaining linear combination is:
xi=si,1x1+…+si,i-1xi-1+si,i+1xi+1+…+si,nxn
Wherein:si=[sI, 1,…sI, i-1,0,sI, i+1,…,si,n]TFor coefficient matrix, si,jIndicate sample xiReconstruction coefficients;
Step S22. builds mathematical model, and the mathematical model built is as follows:
s.t.xi=Xsi
||si||0Indicate siL0Norm, for weighing siSparsity;
Step S23. is by l0Norm minimum problem is converted into l1Norm minimum problem:
s.t.xi=Xsi1=1Tsi
Wherein, 1 complete 1 vector is indicated, | | si||1Indicate siL1Norm;
Step S24. obtains sparse coefficient matrix s=[s by the calculation formula in step S231,s2,…sn]。
3. the semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part according to claim 1, which is characterized in that step Following steps are specifically included in rapid S3:
Step S31. improves semi-supervised similar weight matrix;
Wherein, qijFor the element of improved semi-supervised similar weight matrix Q, aij=exp (- | | xi-xj||2/ σ), σ is all samples Square of Euclidean distance average value between this pair, J (xi) it is xiK neighbours domain, k be neighbour's parameter;aijFor partial weight value, (1 is enabled +aij) it is that weights, (1-a are differentiated in classij) weights are differentiated between class;
Step S32. calculates the diagonal matrix D of similar weight matrix, according to sparse coefficient by improving semi-supervised similar weight matrix Matrix S, to calculate Laplacian Matrix L*, utilize Laplacian Matrix L*Construct the sparse embedded projection algorithm in semi-supervised part Target function type, formula is as follows:
Wherein, L*=D-QST-SQ+SDST, S=[s1,s2,...,sn];Matrix D is the diagonal matrix of Q,
Diagonal matrix element is Dii=∑jqij
Step S33. introduces constraints WTXXTW=I, wherein I are unit matrix, introduce method of Lagrange multipliers by step S32 Middle formula overall goals function is converted into as follows:
F (W)=WTXL*XTW-λ(WTXXTW-I)
Derivation, such as following formula are carried out to the W of the formula:
It enablesIt obtains as follows:
XL*XTW=λ XXTW
Wherein λ is generalized eigenvalue, passes through the corresponding feature vector of a maximum eigenvalue before solution, composition projection matrix W= [w1,w2,…,wa]。
4. a kind of utilizing the local semi-supervised dimension reduction method of high spectrum image of sparse insertion of claim 1-3 any one of them Technology path, it is specific as follows:
The first step, existing high spectrum image, the interpretation figure marked according to expert extracts difference from diagram picture Sample data { the x of classification1, x2, ..., xN, the input data as high-spectrum image dimensionality reduction;
The initial data obtained in the first step is carried out simple data prediction by second step;
Third walks, and sparse coefficient matrix S is built by rarefaction representation;
4th step solves sample data the Euclidean distance dist=between them | | xi-xj||;
5th step, by the 4th step solve Euclidean distance, according in second step to sample add category information the case where, be divided into 5 kinds of situations solve, and finally obtain the similar weight matrix Q between sample point;
6th step, the similar weight matrix Q acquired by the 5th step according to similar weight matrix Q and acquire diagonal matrix D;
7th step, the value acquired by third step and the 6th step bring L into*=D-QST-SQ+SDST
8th step brings the value that the 7th step acquires into following formula:
Wherein, L*=D-QST-SQ+SDST, S=[s1,s2,…,sn], matrix D is the diagonal matrix of Q, and diagonal matrix element is Dii =∑jqij
Sample data dimension is arranged and according to formula XL in 9th step*XTW=λ XXTW finds out transition matrix W=(w1, w2..., wd);
Tenth step, calculation formula yi=WT×xi, find out sample data Y={ y in lower dimensional space1,y2,…,yN};
11st step is added category information as training sample data to a few sample after dimensionality reduction, does not add category in conjunction with most Test sample data, classified to the sample after dimensionality reduction using nearest neighbor classifier, by classification results and known principal component The Comparative result of analytic approach PCA, local retaining projection LPP, calculate nicety of grading.
5. a kind of technology path of the semi-supervised dimensionality reduction of high spectrum image of the sparse insertion in part according to claim 4, It is characterized in that, selects number of tags for 50labels, when d=15, k=24, as the calculating parameter for calculating nicety of grading.
6. a kind of technology path of the semi-supervised dimensionality reduction of high spectrum image of the sparse insertion in part according to claim 4, It is characterized in that, the calculation formula of nicety of grading is as follows:
NErr=sum (class~=classLabel);Rate=1-nErr/length (class);Class is by nearest Adjacent algorithm finds out the category for the test sample come, and classLabel is the category for the test sample known before dimensionality reduction, NErr is error rate, and rate is discrimination.
7. a kind of technology path of the semi-supervised dimensionality reduction of high spectrum image of the sparse insertion in part according to claim 4, It is characterized in that, pretreatment includes:Sample data adds category X={ x1, x2..., xl, xl+1..., xl+u, data normalization processing.
CN201810326062.XA 2018-04-12 2018-04-12 A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part Pending CN108520279A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810326062.XA CN108520279A (en) 2018-04-12 2018-04-12 A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810326062.XA CN108520279A (en) 2018-04-12 2018-04-12 A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part

Publications (1)

Publication Number Publication Date
CN108520279A true CN108520279A (en) 2018-09-11

Family

ID=63432420

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810326062.XA Pending CN108520279A (en) 2018-04-12 2018-04-12 A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part

Country Status (1)

Country Link
CN (1) CN108520279A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110689064A (en) * 2019-09-23 2020-01-14 云南电网有限责任公司电力科学研究院 Image semi-supervised classification method and device and computer readable storage medium
CN111611963A (en) * 2020-05-29 2020-09-01 扬州大学 Face recognition method based on neighbor preserving canonical correlation analysis
CN111783615A (en) * 2020-06-28 2020-10-16 南京工程学院 Non-constrained face recognition method based on weighted block tensor sparse graph mapping
CN112101381A (en) * 2020-08-30 2020-12-18 西南电子技术研究所(中国电子科技集团公司第十研究所) Tensor collaborative drawing discriminant analysis remote sensing image feature extraction method
CN112836671A (en) * 2021-02-26 2021-05-25 西北工业大学 Data dimension reduction method based on maximization ratio and linear discriminant analysis
CN112944104A (en) * 2021-03-03 2021-06-11 杭州申昊科技股份有限公司 Pipeline robot for detecting defects and control method and control system thereof

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103593676A (en) * 2013-11-29 2014-02-19 重庆大学 High-spectral remote-sensing image classification method based on semi-supervision sparse discriminant embedding

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103593676A (en) * 2013-11-29 2014-02-19 重庆大学 High-spectral remote-sensing image classification method based on semi-supervision sparse discriminant embedding

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张晓涛: "基于半监督PCA-LPP流形学习算法的故障降维辨识", 《中南大学学报》 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110689064A (en) * 2019-09-23 2020-01-14 云南电网有限责任公司电力科学研究院 Image semi-supervised classification method and device and computer readable storage medium
CN110689064B (en) * 2019-09-23 2023-06-09 云南电网有限责任公司电力科学研究院 Image semi-supervised classification method, device and computer readable storage medium
CN111611963A (en) * 2020-05-29 2020-09-01 扬州大学 Face recognition method based on neighbor preserving canonical correlation analysis
CN111611963B (en) * 2020-05-29 2023-12-26 扬州大学 Face recognition method based on neighbor preservation canonical correlation analysis
CN111783615A (en) * 2020-06-28 2020-10-16 南京工程学院 Non-constrained face recognition method based on weighted block tensor sparse graph mapping
CN111783615B (en) * 2020-06-28 2023-08-22 南京工程学院 Unconstrained face recognition method based on weighting block tensor sparse map mapping
CN112101381A (en) * 2020-08-30 2020-12-18 西南电子技术研究所(中国电子科技集团公司第十研究所) Tensor collaborative drawing discriminant analysis remote sensing image feature extraction method
CN112836671A (en) * 2021-02-26 2021-05-25 西北工业大学 Data dimension reduction method based on maximization ratio and linear discriminant analysis
WO2022178978A1 (en) * 2021-02-26 2022-09-01 西北工业大学 Data dimensionality reduction method based on maximum ratio and linear discriminant analysis
CN112836671B (en) * 2021-02-26 2024-03-08 西北工业大学 Data dimension reduction method based on maximized ratio and linear discriminant analysis
CN112944104A (en) * 2021-03-03 2021-06-11 杭州申昊科技股份有限公司 Pipeline robot for detecting defects and control method and control system thereof

Similar Documents

Publication Publication Date Title
CN108520279A (en) A kind of semi-supervised dimension reduction method of high spectrum image of the sparse insertion in part
Dong et al. Target detection based on random forest metric learning
Xing et al. Pan-sharpening via deep metric learning
Ma et al. Local-manifold-learning-based graph construction for semisupervised hyperspectral image classification
Chen et al. Hyperspectral image classification using dictionary-based sparse representation
CN109543602B (en) Pedestrian re-identification method based on multi-view image feature decomposition
CN108520281A (en) A kind of semi-supervised dimension reduction method of high spectrum image kept based on overall situation and partial situation
CN110084159A (en) Hyperspectral image classification method based on the multistage empty spectrum information CNN of joint
CN109766858A (en) Three-dimensional convolution neural network hyperspectral image classification method combined with bilateral filtering
CN104392251B (en) Hyperspectral image classification method based on semi-supervised dictionary learning
CN105760900B (en) Hyperspectral image classification method based on neighbour's propagation clustering and sparse Multiple Kernel Learning
CN106096506A (en) Based on the SAR target identification method differentiating doubledictionary between subclass class
Zhang et al. Locality-constrained sparse representation for hyperspectral image classification
CN104268556A (en) Hyperspectral image classification method based on nuclear low-rank representing graph and spatial constraint
CN104281835B (en) Face recognition method based on local sensitive kernel sparse representation
CN111680579B (en) Remote sensing image classification method for self-adaptive weight multi-view measurement learning
Luo et al. Fusion of graph embedding and sparse representation for feature extraction and classification of hyperspectral imagery
CN105869161B (en) Hyperspectral image band selection method based on image quality evaluation
Sarker et al. A multidimensional pixel-wise convolutional neural network for hyperspectral image classification
Zhu et al. EDLAD: An encoder-decoder long short-term memory network-based anomaly detector for hyperspectral images
CN107203779A (en) The EO-1 hyperion dimension reduction method kept based on empty spectrum information
Yang et al. Hyperspectral image classification based on spatial and spectral features and sparse representation
CN117115675A (en) Cross-time-phase light-weight spatial spectrum feature fusion hyperspectral change detection method, system, equipment and medium
Feng et al. Unsupervised feature extraction in hyperspectral image based on improved neighborhood preserving embedding
McCleary et al. Sign language recognition using micro-Doppler and explainable deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20180911

RJ01 Rejection of invention patent application after publication