CN104008383A - Hyperspectral image characteristic extraction algorithm based on manifold learning linearization - Google Patents

Hyperspectral image characteristic extraction algorithm based on manifold learning linearization Download PDF

Info

Publication number
CN104008383A
CN104008383A CN201410286545.3A CN201410286545A CN104008383A CN 104008383 A CN104008383 A CN 104008383A CN 201410286545 A CN201410286545 A CN 201410286545A CN 104008383 A CN104008383 A CN 104008383A
Authority
CN
China
Prior art keywords
matrix
manifold learning
algorithm
linearization
hyperspectral image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410286545.3A
Other languages
Chinese (zh)
Other versions
CN104008383B (en
Inventor
张淼
赖镇洲
刘攀
沈毅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Institute of Technology Shenzhen
Original Assignee
Harbin Institute of Technology Shenzhen
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Institute of Technology Shenzhen filed Critical Harbin Institute of Technology Shenzhen
Priority to CN201410286545.3A priority Critical patent/CN104008383B/en
Publication of CN104008383A publication Critical patent/CN104008383A/en
Application granted granted Critical
Publication of CN104008383B publication Critical patent/CN104008383B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

一种基于流形学习线性化的高光谱图像特征提取算法,属于高光谱图像数据处理与应用技术领域。本发明针对流形学习算法无泛化能力的不足,提出了一种改进的流形学习线性化算法。所述方法包括如下步骤:一、计算初步降维结果和拉普拉斯矩阵;二、构建矩阵方程组常数项矩阵和系数矩阵;三、计算特征转换矩阵;四、通过特征转换矩阵计算最终降维结果。本发明针对LPP、NPE和LLTSA线性化流形学习算法中全局线性映射的假设在很多时候不成立的不足,在原有的代价函数中加入了偏离原流形学习算法结果的惩罚项,并且舍去了原目标函数中的约束项,将最优特征转换矩阵的求解转换为一个矩阵方程组的求解问题。该方法适用于高光谱图像的特征提取。A hyperspectral image feature extraction algorithm based on manifold learning linearization belongs to the technical field of hyperspectral image data processing and application. The invention proposes an improved manifold learning linearization algorithm aiming at the lack of generalization ability of the manifold learning algorithm. The method comprises the following steps: 1. Calculating the preliminary dimensionality reduction result and the Laplacian matrix; 2. Constructing a constant term matrix and a coefficient matrix of the matrix equation system; 3. Calculating the feature transformation matrix; dimension results. The present invention aims at the deficiency that the assumption of global linear mapping in LPP, NPE and LLTSA linearized manifold learning algorithms is often untenable, and adds a penalty item that deviates from the result of the original manifold learning algorithm to the original cost function, and discards the The constraint item in the original objective function converts the solution of the optimal feature transformation matrix into a solution problem of a matrix equation system. This method is suitable for feature extraction of hyperspectral images.

Description

基于流形学习线性化的高光谱图像特征提取算法Hyperspectral Image Feature Extraction Algorithm Based on Manifold Learning Linearization

技术领域technical field

本发明属于高光谱图像数据处理与应用技术领域,涉及一种高光谱图像特征提取算法,具体涉及一种基于流形学习线性化的高光谱图像特征提取算法。The invention belongs to the technical field of hyperspectral image data processing and application, and relates to a hyperspectral image feature extraction algorithm, in particular to a hyperspectral image feature extraction algorithm based on manifold learning linearization.

背景技术Background technique

高光谱图像是信息量巨大的数据立方体,每个像元都对应一条包含上百个波段的谱线,这为人们研究物质与光谱曲线之间的关系提供了可能。但是高光谱数据存在数据冗余和维数灾难问题,人们有迫切的需求去消除高光谱数据的这种信息的冗余性。高光谱数据的这种冗余性主要是高光谱数据波段间的相关性造成的,降维是一种重要的预处理方法,尽管像PCA(Principal Component Analysis)和LDA(LinearDiscriminantAnalysis)这样的线性降维算法实现简单,但是高光谱图像具有非线性特性,流形学习算法可以更好地挖掘高光谱数据的非线性结构,提高数据分析能力。经典的流形学习算法有LE(LaplacianEgenmap)、LLE(Locally Linear Embedding)和LTSA(Local TangentSpace Alignment)算法,可用于高光谱图像的特征提取算法。Hyperspectral images are data cubes with a huge amount of information, and each pixel corresponds to a spectral line containing hundreds of bands, which provides the possibility for people to study the relationship between substances and spectral curves. However, there are data redundancy and dimensionality disaster problems in hyperspectral data, and there is an urgent need to eliminate the information redundancy of hyperspectral data. This redundancy of hyperspectral data is mainly caused by the correlation between the bands of hyperspectral data. Dimensionality reduction is an important preprocessing method, although linear reduction methods such as PCA (Principal Component Analysis) and LDA (Linear Discriminant Analysis) Dimensional algorithms are simple to implement, but hyperspectral images have nonlinear characteristics. Manifold learning algorithms can better mine the nonlinear structure of hyperspectral data and improve data analysis capabilities. The classic manifold learning algorithms include LE (Laplacian Egenmap), LLE (Locally Linear Embedding) and LTSA (Local TangentSpace Alignment) algorithms, which can be used for feature extraction algorithms of hyperspectral images.

但是经典流形学习算法如LE无泛化能力,当有新的高光谱样本数据出现时,只有把新的样本数据和原有的样本数据结合在一起再一次地进行整体的学习才能够得到新样本数据的降维结果。当新的样本数据的个数相比于原有的样本数据的个数显得很少的时候,显然这样重复大规模的计算会大大增加算法的时间复杂度。从高光谱分类需求来说,流形学习算法必须具备泛化能力,因为很多时候无法将训练数据和待分类的测试数据放在一起进行学习,对于这种情况,待分类的测试数据必须要通过泛化算法映射到低维的特征空间,否则将无法在低维的特征空间对新的高光谱数据进行分类。However, classic manifold learning algorithms such as LE have no generalization ability. When new hyperspectral sample data appears, only by combining the new sample data with the original sample data and performing overall learning again can new Dimensionality reduction results for sample data. When the number of new sample data is very small compared with the number of original sample data, it is obvious that repeated large-scale calculations will greatly increase the time complexity of the algorithm. From the perspective of hyperspectral classification requirements, the manifold learning algorithm must have generalization ability, because in many cases it is impossible to put the training data and the test data to be classified together for learning. In this case, the test data to be classified must pass The generalization algorithm is mapped to a low-dimensional feature space, otherwise it will not be possible to classify new hyperspectral data in a low-dimensional feature space.

线性降维方法如PCA等虽然在处理具有非线性的数据效果不佳,但是它们却能够获得一个全局的映射函数,因而具备泛化能力。为此很多学者对流形学习线性化来解决流形学习无泛化能力的问题。其中非常典型的就是对LE、LLE、LTSA的线性化,分别得到LPP(Locality preserving projections)、NPE(Neighborhood PreservingEmbedding)、LLTSA(Linear Local Tangent Space Alignment)。Although linear dimensionality reduction methods such as PCA are not effective in dealing with nonlinear data, they can obtain a global mapping function and thus have generalization capabilities. For this reason, many scholars linearize manifold learning to solve the problem that manifold learning has no generalization ability. The very typical ones are the linearization of LE, LLE, and LTSA, and LPP (Locality preserving projections), NPE (Neighborhood Preserving Embedding), and LLTSA (Linear Local Tangent Space Alignment) are obtained respectively.

流形学习算法具备统一的框架,对于如LE、LLE和LTSA等基于谱分解的流形学习算法而言,最优的降维结果Y*都可以通过下面的最优化问题来求解:The manifold learning algorithm has a unified framework. For manifold learning algorithms based on spectral decomposition such as LE, LLE, and LTSA, the optimal dimensionality reduction result Y * can be solved by the following optimization problem:

YY ** == minmin YY trtr (( YLYYLY TT )) sthe s .. tt YBYYBY YY == II -- -- -- (( 11 )) ,,

其中tr(·)是求矩阵的迹运算符,L为拉普拉斯矩阵,B是约束矩阵。LPP、NPE和LLTSA都假设降维过程中存在一个全局的线性映射:Among them, tr(·) is the trace operator for finding the matrix, L is the Laplacian matrix, and B is the constraint matrix. LPP, NPE, and LLTSA all assume that there is a global linear map in the dimensionality reduction process:

Y=VTX  (2),Y = V T X (2),

使得式(1)中的优化问题转变为:The optimization problem in formula (1) is transformed into:

VV ** == minmin VV trtr (( VV TT XLXXLX TT VV )) sthe s .. tt VV TT XVXXVX TT VV == II -- -- -- (( 33 )) ..

获得了最优的线性映射矩阵V*之后,不论是新的样本还是旧的样本,都可以通过式(2)映射到降维空间。但是式(2)这个假设在很多时候是不成立的。为了获得全局的线性映射,这些算法其实是对原本算法的局部保持特性做了一定的妥协。After obtaining the optimal linear mapping matrix V * , whether it is a new sample or an old sample, it can be mapped to the dimensionality reduction space by formula (2). But the assumption of formula (2) is not true in many cases. In order to obtain a global linear map, these algorithms actually compromise the local preservation characteristics of the original algorithm.

发明内容Contents of the invention

本发明针对流形学习算法无泛化能力的不足,提出了一种改进的流形学习线性化算法,并将其用于高光谱图像的特征提取算法。Aiming at the lack of generalization ability of the manifold learning algorithm, the invention proposes an improved manifold learning linearization algorithm, and uses it in the feature extraction algorithm of the hyperspectral image.

本发明的目的是通过以下技术方案实现的:The purpose of the present invention is achieved through the following technical solutions:

为了使得线性化的过程尽可能不要牺牲原流形学习算法保持局部特性的能力,本发明在原有的代价函数中加入了偏离原流形学习算法结果的惩罚项,使得式(3)中的优化问题转变为:In order to avoid sacrificing the ability of the original manifold learning algorithm to maintain local characteristics during the linearization process, the present invention adds a penalty term that deviates from the result of the original manifold learning algorithm in the original cost function, so that the optimization in formula (3) The problem turns into:

minmin trtr (( VV TT XLXXLX TT VV )) ++ αα || || YY LL -- VV TT Xx || || Ff 22 -- -- -- (( 44 )) ,,

其中YL是已经学习了的降维结果,α是惩罚系数,可以通过LE、LLE、LTSA或其他任何一种效果良好但是不具备泛化能力的流形学习算法获得。相比于式(3),式(4)不再需要添加约束项VTXBXTV=I,这使得优化问题的搜索域得到了拓宽,可以获得比传统降维算法更好的降维效果。这种泛化算法是基于YL已知去求取一个线性化映射矩阵V的过程,是一种全局的线性回归算法。原有的流形学习算法都是通过特征分解来获得最优的降维结果,本发明提出的改进方法通过求解一个矩阵方程来获得最优的降维结果。Among them, Y L is the dimension reduction result that has been learned, and α is the penalty coefficient, which can be obtained by LE, LLE, LTSA or any other manifold learning algorithm with good effect but no generalization ability. Compared with formula (3), formula (4) no longer needs to add the constraint item V T XBX T V = I, which broadens the search domain of the optimization problem and can obtain better dimension reduction effect than traditional dimension reduction algorithms . This generalization algorithm is a process of finding a linearized mapping matrix V based on the known Y L , which is a global linear regression algorithm. The original manifold learning algorithm obtains the optimal dimension reduction result through feature decomposition, and the improved method proposed by the present invention obtains the optimal dimension reduction result by solving a matrix equation.

这里给出对(4)中的优化问题的求解过程:Here is the solution process of the optimization problem in (4):

gg (( VV )) == trtr (( VV TT XLXXLX TT VV )) ++ αα || || YY LL -- VV TT Xx || || Ff 22 == trtr (( VV TT XLXXLX TT VV )) ++ αtrαtr {{ [[ YY LL -- VV TT Xx ]] [[ YY LL -- VV TT Xx ]] TT }} == trtr (( VV TT Xx (( αIα I ++ LL )) Xx TT VV )) ++ αtrαtr (( YY LL YY LL TT )) -- αtrαtr (( YY LL Xx TT VV )) -- αtrαtr (( VV TT XTXT LL TT )) -- -- -- (( 55 ))

代价函数g(V)对V求导得:The cost function g(V) is derived from V:

∂∂ gg (( VV )) ∂∂ VV == 22 [[ XAXXAX TT ]] VV -- 22 αα XYX Y LL TT == 00 -- -- -- (( 66 )) ,,

其中A=αI+L。where A=αI+L.

于是式(4)中的最优化问题变为一个求解矩阵方程的问题:So the optimization problem in formula (4) becomes a problem of solving the matrix equation:

[XAXT]V=aXYL T  (7)。[XAX T ] V = aXY L T (7).

等价于下面的线性方程组问题:Equivalent to the following system of linear equations problem:

vecvec (( VV )) == (( XAXXAX TT ⊗⊗ II dd )) -- 11 vecvec (( αα XYX Y LL TT )) == [[ (( XAXXAX TT )) -- 11 ⊗⊗ II dd ]] vecvec (( αα XYX Y LL TT )) -- -- -- (( 88 )) ..

所以有:F:

VV == avecavec {{ [[ (( XAXXAX TT )) -- 11 ⊗⊗ II dd ]] vecvec (( BB )) ,, dd }} -- -- -- (( 99 )) ..

式(9)是一个函数表达式,其中vec(·)是矩阵向量化运算符,对于vec(B),即将D×d维的矩阵转换为D×d维列向量。avec(·)是向量矩阵化运算符。对于D×d维的矩阵,vec(·)的运算规则为将矩阵每一行的元素转换为一个d维的列向量,并且将每行转换后的列向量依行号排成一个D×d维列向量,avec(·)则是其逆运算,即将D×d维的列向量转换为D×d维矩阵。Equation (9) is a functional expression, where vec( ) is a matrix vectorization operator, and for vec(B), it is to convert a D×d-dimensional matrix into a D×d-dimensional column vector. avec( ) is the vector matrixization operator. For a D×d-dimensional matrix, the operation rule of vec(·) is to convert the elements of each row of the matrix into a d-dimensional column vector, and arrange the converted column vectors of each row into a D×d-dimensional one according to the row number column vector, and avec(·) is its inverse operation, which converts a D×d-dimensional column vector into a D×d-dimensional matrix.

对(9)进一步化简得:(9) is further simplified to get:

V=(XAXT)-1B  (10)。V = (XAX T ) -1 B (10).

本发明提供的基于流形学习线性化的高光谱图像特征提取算法,具体步骤如下:The hyperspectral image feature extraction algorithm based on manifold learning linearization provided by the present invention, the specific steps are as follows:

步骤一:计算初步降维结果和拉普拉斯矩阵。Step 1: Calculate the preliminary dimensionality reduction results and Laplacian matrix.

给定高光谱数据集X,X是D×N维矩阵,D是数据维数,N是样本个数,d是降维维数,执行如LE、LLE以及LTSA算法中的某一种流形学习算法,得到这种算法的初步降维结果YL和拉普拉斯矩阵L,其中YL是d×N维矩阵,L是N×N维矩阵。Given a hyperspectral data set X, where X is a D×N dimensional matrix, D is the data dimension, N is the number of samples, and d is the dimensionality reduction dimension, execute a certain manifold such as LE, LLE, and LTSA algorithms Learning algorithm, get the preliminary dimension reduction result Y L and Laplacian matrix L of this algorithm, where Y L is a d×N dimensional matrix, and L is an N×N dimensional matrix.

步骤二:构建矩阵方程组常数项矩阵和系数矩阵。Step 2: Construct the constant term matrix and coefficient matrix of the matrix equation system.

LPP、NPE和LLTSA等线性化的流形学习通常都要通过对N×N维的拉普拉斯矩阵L特征分解来获取最优的特征转换矩阵,计算量大,本步骤的创新点在于将最优特征转换矩阵的求解转换为一个矩阵方程组的求解问题。Linear manifold learning such as LPP, NPE, and LLTSA usually obtains the optimal feature transformation matrix by decomposing the N×N-dimensional Laplacian matrix L eigenvalues, which requires a large amount of calculation. The innovation of this step is to The solution of the optimal eigentransition matrix is transformed into a solution problem of a system of matrix equations.

矩阵方程组如式(7)所示,这里给出常数项矩阵和系数矩阵的构建公式:The matrix equations are shown in formula (7), here are the construction formulas of the constant term matrix and the coefficient matrix:

1)构建D×d维的矩阵方程组常数项矩阵B:1) Construct a D×d-dimensional matrix equation system constant item matrix B:

B=αXYL TB=αXY L T ,

其中α是一个为正数的惩罚系数,默认可设置为1。Where α is a positive penalty coefficient, which can be set to 1 by default.

2)构建D×D维的矩阵方程组系数矩阵C:2) Construct D×D-dimensional matrix equation system coefficient matrix C:

C=X(αI+L)XTC=X(αI+L)X T ,

其中,I是N×N维的单位矩阵。Wherein, I is an N×N-dimensional identity matrix.

步骤三:计算特征转换矩阵。Step 3: Calculate the feature transformation matrix.

矩阵方程组的求解过程中,通常要对矩阵进行向量化以及对向量进行矩阵化运算,而且还涉及到矩阵的Kronecker积,这些运算使得矩阵方程组的求解时间大消耗的内存空间也大。本步骤的创新点在于将矩阵方程组求解的问题化简为两个矩阵相乘来实现,实现方便,算法复杂度小。In the process of solving the matrix equations, it is usually necessary to vectorize the matrix and perform matrix operations on the vectors, and also involves the Kronecker product of the matrix. These operations make the solution of the matrix equations take a long time and consume a lot of memory space. The innovation of this step is to simplify the problem of solving matrix equations into two matrix multiplications, which is easy to implement and has low algorithm complexity.

1)对方程组系数矩阵C进行求逆得到矩阵H:1) Invert the coefficient matrix C of the system of equations to obtain the matrix H:

H=C-1H=C -1 ;

2)通过矩阵H和方程组常数项矩阵B相乘得特征转换矩阵V:2) The feature transformation matrix V is obtained by multiplying the matrix H and the constant item matrix B of the equation system:

V=HB。V=HB.

步骤四:通过特征转换矩阵计算最终降维结果:Step 4: Calculate the final dimension reduction result through the feature transformation matrix:

Y=VTX,Y = V T X,

其中,Y是最终的降维结果。Among them, Y is the final dimensionality reduction result.

本发明的有益效果在于:The beneficial effects of the present invention are:

1、这种算法是一种全局的线性回归算法,可以使得任何一种不具备泛化能力的流形学习算法通过该方法进行线性化之后,获得泛化能力,并能可以获得比传统降维算法更好的特征提取效果。1. This algorithm is a global linear regression algorithm, which can enable any manifold learning algorithm that does not have generalization ability to obtain generalization ability after linearization through this method, and can obtain better than traditional dimensionality reduction. The algorithm has better feature extraction effect.

2、针对LPP、NPE和LLTSA线性化流形学习算法中全局线性映射的假设在很多时候不成立的不足,在原有的代价函数中加入了偏离原流形学习算法结果的惩罚项,并且舍去了原目标函数中的约束项,这使得优化问题的搜索域得到了拓展,可以获得比传统降维算法更好的降维效果。2. In view of the fact that the assumption of global linear mapping in the LPP, NPE and LLTSA linearized manifold learning algorithms is not valid in many cases, a penalty item for deviating from the original manifold learning algorithm result is added to the original cost function, and the The constraints in the original objective function expand the search domain of the optimization problem and achieve better dimensionality reduction results than traditional dimensionality reduction algorithms.

附图说明Description of drawings

图1为本发明步骤流程图;Fig. 1 is a flow chart of steps of the present invention;

图2为本发明中的高光谱图像数据的特征散点图;Fig. 2 is a characteristic scatter diagram of the hyperspectral image data in the present invention;

图3为基于本发明的特征提取之后的高光谱图像的分类精度。Fig. 3 is the classification accuracy of the hyperspectral image after feature extraction based on the present invention.

具体实施方式Detailed ways

下面结合附图对本发明的技术方案作进一步的说明,但并不局限如此,凡是对本发明技术方案进行修改或者等同替换,而不脱离本发明技术方案的精神和范围,均应涵盖在本发明的保护范围中。The technical solution of the present invention will be further described below in conjunction with the accompanying drawings, but it is not limited to this. Any modification or equivalent replacement of the technical solution of the present invention without departing from the spirit and scope of the technical solution of the present invention should be covered by the technical solution of the present invention. in the scope of protection.

本发明在第一步骤中需要使用一种现有的流形学习算法获得拉普拉斯矩阵和初步的降维结果,这里以LLE算法为例,将其作为第一步骤中的流形学习算法,然后使用本发明所提出的算法对高光谱图像进行特征提取。实验用高光谱图像数据选用IND PINE高光谱图像,该高光谱图像是由美国的肯尼迪航天中心在对美国印第安纳州的一个农田进行拍照而得,这幅图像中一共有16种不同的农作物,图像的空间分辨率是20×20m2,每个像素点有224个波段,覆盖光谱范围是0.2~2.4μm的波谱范围,谱分辨率为10nm。In the first step, the present invention needs to use an existing manifold learning algorithm to obtain the Laplacian matrix and preliminary dimensionality reduction results. Here, the LLE algorithm is used as an example of the manifold learning algorithm in the first step. , and then use the algorithm proposed by the present invention to extract features from the hyperspectral image. The hyperspectral image data used in the experiment is the IND PINE hyperspectral image. The hyperspectral image was taken by the Kennedy Space Center of the United States in a farmland in Indiana, USA. There are 16 different crops in this image. The spatial resolution is 20×20m 2 , each pixel has 224 bands, the spectral range is 0.2-2.4μm, and the spectral resolution is 10nm.

随机地从IND PINE高光谱图像中选择1500个像素点作为训练样本,随机地从剩余的样本中选择1500个像素点作为测试样本。本实施例中,使用本发明对训练样本进行学习,获取特征转换矩阵,然后使用特征转换矩阵对测试样本进行特征提取,并且使用KNN分类器对测试样本进行分类。为了验证本发明特征提取地有效性,对同样的训练数据、测试数据和分类器,同时使用LLE经典的线性化算法NPE作为对比算法。Randomly select 1500 pixels from the IND PINE hyperspectral image as training samples, and randomly select 1500 pixels from the remaining samples as test samples. In this embodiment, the present invention is used to study the training samples, obtain the feature transformation matrix, then use the feature transformation matrix to extract the features of the test samples, and use the KNN classifier to classify the test samples. In order to verify the effectiveness of feature extraction in the present invention, for the same training data, test data and classifier, the classic linearization algorithm NPE of LLE is used as a comparison algorithm.

如图1所示,采用本发明进行特征提取的具体步骤如下:As shown in Figure 1, the specific steps of feature extraction using the present invention are as follows:

步骤一:通过LLE流形学习算法获得降维结果YL和拉普拉斯矩阵L。Step 1: Obtain the dimensionality reduction result Y L and the Laplacian matrix L through the LLE manifold learning algorithm.

1)输入训练集其中每个样本数据维数为224,测试样本个数为1500;设置近邻域的个数为20;设置降维结果维数30。1) Input training set The data dimension of each sample is 224, the number of test samples is 1500; the number of nearest neighbors is set to 20; the dimensionality of the dimension reduction result is set to 30.

2)寻找邻域集2) Find the neighborhood set

对训练集中的每个样本xi,其中i表示xi在训练样本集X中的位置索引号,i=1,2,...,1500,通过样本间欧式距离排序,搜索训练集中与xi最近的20个样本,构成样本xi的邻域集 For each sample x i in the training set, where i represents the position index number of x i in the training sample set X, i=1, 2, ..., 1500, sort by the Euclidean distance between samples, search the training set and x The nearest 20 samples of i constitute the neighborhood set of sample x i

3)构建重构系数矩阵W:3) Construct the reconstruction coefficient matrix W:

WW ijij == TT (( xx jj ,, Xx ii )) // SS (( xx jj ,, Xx ii )) xx jj ∈∈ Xx ii 00 othersothers ii ,, jj == 1,21,2 ,, .. .. .. ,, 15001500 ,,

其中:Wij为xj对xi的重构系数,xj为训练样本集X中位置索引号为j的测试样本,其中,T(xj,Xi)通过下面公式计算:Among them: W ij is the reconstruction coefficient of x j to x i , x j is the test sample whose position index number is j in the training sample set X, where T(x j ,X i ) is calculated by the following formula:

TT (( xx jj ,, Xx ii )) == ΣΣ ll == 11 2020 PP ll ,, indexindex (( xx jj ,, Xx ii )) ..

其中:index(xj,Xi)表示xj在Xi中位置索引,T计算的是矩阵P中index(xj,Xi)列的所有元素的和;矩阵P为邻域集Xi的局部协方差矩阵的逆,可以通过下面公式计算:Among them: index(x j , Xi ) indicates the position index of x j in Xi , and T calculates the sum of all elements in the column index(x j , Xi ) in matrix P; matrix P is the neighborhood set Xi The inverse of the local covariance matrix of can be calculated by the following formula:

P=[(Xi-xi)T(Xi-xi)]-1P=[(X i -xi ) T (X i -xi )] -1 ,

其中:S为矩阵P的所有元素的和,可以通过下面公式计算:Among them: S is the sum of all elements of the matrix P, which can be calculated by the following formula:

SS (( xx jj ,, Xx ii )) == ΣΣ ll == 11 2020 ΣΣ mm == 11 2020 PP ll ,, mm ..

4)构建拉普拉斯矩阵:4) Construct the Laplacian matrix:

L=(I-W)T(I-W),L=(IW) T (IW),

其中:I为N×N维单位矩阵。Where: I is an N×N-dimensional identity matrix.

5)进行特征分解:5) Perform feature decomposition:

Lz=λz。Lz=λz.

6)获取初步降维结果:6) Obtain preliminary dimensionality reduction results:

其中:zii是L的第ii小的特征值对应的特征向量,ii=2,...,d+1。Where: z ii is the eigenvector corresponding to the ii- th smallest eigenvalue of L, ii =2,...,d+1.

步骤二:构建矩阵方程组常数项矩阵B和系数矩阵C。Step 2: Construct the constant term matrix B and the coefficient matrix C of the matrix equation system.

1)设置正数惩罚系数α,这里设为α=1。1) Set a positive penalty coefficient α, where α=1.

2)矩阵方程组常数项矩阵B通过下面公式计算:2) The constant item matrix B of the matrix equation system is calculated by the following formula:

B=αXYL TB = αXYLT .

3)矩阵方程组系数矩阵C通过下面公式计算:3) The coefficient matrix C of the matrix equation system is calculated by the following formula:

C=X(αI+L)XTC=X(αI+L) XT .

步骤三:计算特征转换矩阵V。Step 3: Calculate the feature transformation matrix V.

1)对方程组系数矩阵C进行求逆得到矩阵H:1) Invert the coefficient matrix C of the system of equations to obtain the matrix H:

H=C-1H=C -1 .

2)通过矩阵H和方程组常数项矩阵B相乘得特征转换矩阵V:2) The feature transformation matrix V is obtained by multiplying the matrix H and the constant item matrix B of the equation system:

V=HB。V=HB.

步骤四:通过特征转换矩阵计算最终降维结果。Step 4: Calculate the final dimension reduction result through the feature transformation matrix.

1)通过下面公式计算训练集的最终降维结果Y:1) Calculate the final dimensionality reduction result Y of the training set by the following formula:

Y=(V)TX。Y = (V) T X.

2)通过下面公式计算测试集的最终降维结果2) Calculate the final dimensionality reduction result of the test set by the following formula :

YY ~~ == (( VV )) TT Xx ~~ ,,

其中:为测试样本集。in: is the test sample set.

测试集和训练集的最终降维结果如图2所示。从图中可以看到新样本的泛化结果和测试样本的降维结果基本一致,可知本发明所提特征提取算法的泛化效果较好。The final dimensionality reduction results of the test set and training set are shown in Figure 2. It can be seen from the figure that the generalization result of the new sample is basically consistent with the dimensionality reduction result of the test sample, and it can be seen that the generalization effect of the feature extraction algorithm proposed in the present invention is better.

步骤五:使用KNN分类算法对测试样本进行分类测试。Step 5: Use the KNN classification algorithm to classify the test samples.

1)计算测试样本集的最终降维结果中的每个测试样本与所有训练样本集的最终降维结果Y的欧式距离。1) Calculate the final dimensionality reduction result of the test sample set The Euclidean distance between each test sample in and the final dimensionality reduction result Y of all training sample sets.

2)取距离最近的5个训练样本作为测试样本的近邻。2) Take the 5 closest training samples as the neighbors of the test samples.

3)根据这5个近邻归属的主要类别,来对测试样本分类。3) Classify the test samples according to the main categories that these five neighbors belong to.

4)计算分类的正确率。4) Calculate the correct rate of classification.

分类正确率如图3所示。从图3中可以看到通过本发明所提特征提取算法进行高光谱图像的泛化之后通过KNN分类具有较高的整体分类精度,LLE-GLR算法的分类精度明显要高于LLE、NPE算法的分类精度,这说明本发明中提出的特征提取算法有利于提高高光谱图像的分类精度。The classification accuracy rate is shown in Figure 3. As can be seen from Fig. 3, after the generalization of the hyperspectral image by the proposed feature extraction algorithm of the present invention, the KNN classification has a higher overall classification accuracy, and the classification accuracy of the LLE-GLR algorithm is obviously higher than that of the LLE and NPE algorithms. Classification accuracy, which shows that the feature extraction algorithm proposed in the present invention is beneficial to improve the classification accuracy of hyperspectral images.

综合上述分析,可知本发明提出的基于流形学习线性化的高光谱图像特征提取算法,对已有的流形学习线性化算法的确进行了改良。本发明中提出的特征提取算法可以有效的解决现有的流形学习对新样本无法学习的不足,同时还可以提高高光谱图像的分类精度,具有很大的工程实际价值。Based on the above analysis, it can be seen that the hyperspectral image feature extraction algorithm based on manifold learning linearization proposed by the present invention has indeed improved the existing manifold learning linearization algorithm. The feature extraction algorithm proposed in the present invention can effectively solve the problem that the existing manifold learning cannot learn new samples, and can also improve the classification accuracy of hyperspectral images, which has great engineering practical value.

Claims (1)

1.一种基于流形学习线性化的高光谱图像特征提取算法,其特征在于所述高光谱图像特征提取算法步骤如下:1. a hyperspectral image feature extraction algorithm based on manifold learning linearization, characterized in that the hyperspectral image feature extraction algorithm step is as follows: 一、给定高光谱数据集X,通过流形学习算法获得初步降维结果YL和拉普拉斯矩阵L,其中X是D×N维矩阵,D是数据维数,N是样本个数,YL是d×N维矩阵,L是N×N维矩阵,d是降维维数;1. Given a hyperspectral data set X, obtain the preliminary dimensionality reduction result Y L and the Laplacian matrix L through the manifold learning algorithm, where X is a D×N dimensional matrix, D is the data dimension, and N is the number of samples , Y L is a d×N dimensional matrix, L is an N×N dimensional matrix, and d is the dimensionality reduction; 二、构建矩阵方程组常数项矩阵B和系数矩阵C:2. Construct the constant term matrix B and the coefficient matrix C of the matrix equation system: 1)构建D×d维的矩阵方程组常数项矩阵B:1) Construct a D×d-dimensional matrix equation system constant item matrix B: B=αXYL TB=αXY L T , 其中,α是一个为正数的惩罚系数;Among them, α is a penalty coefficient that is a positive number; 2)构建D×D维的矩阵方程组系数矩阵C:2) Construct D×D-dimensional matrix equation system coefficient matrix C: C=X(αI+L)XTC=X(αI+L)X T , 其中,I是N×N维的单位矩阵;Wherein, I is an identity matrix of N×N dimensions; 三、计算特征转换矩阵:3. Calculate the feature transformation matrix: 1)对矩阵方程组系数矩阵C进行求逆得到矩阵H:1) Invert the coefficient matrix C of the matrix equation system to obtain the matrix H: H=C-1H=C -1 ; 2)通过矩阵H和矩阵方程组常数项矩阵B相乘得特征转换矩阵V:2) By multiplying the matrix H and the constant item matrix B of the matrix equation system, the characteristic transformation matrix V is obtained: V=HB;V=HB; 四、通过特征转换矩阵计算最终降维结果:4. Calculate the final dimension reduction result through the feature transformation matrix: Y=VTX,Y = V T X, 其中,Y是最终的降维结果。Among them, Y is the final dimensionality reduction result.
CN201410286545.3A 2014-06-24 2014-06-24 Based on manifold learning linearizing high spectrum image feature extracting method Active CN104008383B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410286545.3A CN104008383B (en) 2014-06-24 2014-06-24 Based on manifold learning linearizing high spectrum image feature extracting method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410286545.3A CN104008383B (en) 2014-06-24 2014-06-24 Based on manifold learning linearizing high spectrum image feature extracting method

Publications (2)

Publication Number Publication Date
CN104008383A true CN104008383A (en) 2014-08-27
CN104008383B CN104008383B (en) 2017-03-08

Family

ID=51369032

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410286545.3A Active CN104008383B (en) 2014-06-24 2014-06-24 Based on manifold learning linearizing high spectrum image feature extracting method

Country Status (1)

Country Link
CN (1) CN104008383B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107203779A (en) * 2017-05-11 2017-09-26 中国科学院西安光学精密机械研究所 Hyperspectral dimensionality reduction method based on spatial-spectral information maintenance
CN110222631A (en) * 2019-06-04 2019-09-10 电子科技大学 Based on deblocking it is parallel cut space arrangement SAR image target recognition method
CN110781974A (en) * 2019-10-31 2020-02-11 上海融军科技有限公司 Dimension reduction method and system for hyperspectral image
CN112016366A (en) * 2019-05-31 2020-12-01 北京车和家信息技术有限公司 Obstacle positioning method and device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060165284A1 (en) * 2005-01-25 2006-07-27 Shmuel Aharon Manifold learning for discriminating pixels in multi-channel images, with application to image/volume/video segmentation and clustering
CN103136736A (en) * 2013-03-19 2013-06-05 哈尔滨工业大学 Hyperspectral remote sensing data non-linear dimension descending method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060165284A1 (en) * 2005-01-25 2006-07-27 Shmuel Aharon Manifold learning for discriminating pixels in multi-channel images, with application to image/volume/video segmentation and clustering
CN103136736A (en) * 2013-03-19 2013-06-05 哈尔滨工业大学 Hyperspectral remote sensing data non-linear dimension descending method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZHI HE ET AL.: "Hyperspectral image classification with multivariate empirical mode decomposition-based features", 《2014 IEEE INTERNATIONAL INSTRUMENTATION AND MEASUREMENT TECHNOLOGY CONFERENCE》 *
普晗晔 等: "基于流形学习的新高光谱图像降维算法", 《红外与激光工程》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107203779A (en) * 2017-05-11 2017-09-26 中国科学院西安光学精密机械研究所 Hyperspectral dimensionality reduction method based on spatial-spectral information maintenance
CN112016366A (en) * 2019-05-31 2020-12-01 北京车和家信息技术有限公司 Obstacle positioning method and device
CN110222631A (en) * 2019-06-04 2019-09-10 电子科技大学 Based on deblocking it is parallel cut space arrangement SAR image target recognition method
CN110222631B (en) * 2019-06-04 2022-03-15 电子科技大学 Target recognition method of tangential space arrangement SAR image based on data block and parallel
CN110781974A (en) * 2019-10-31 2020-02-11 上海融军科技有限公司 Dimension reduction method and system for hyperspectral image

Also Published As

Publication number Publication date
CN104008383B (en) 2017-03-08

Similar Documents

Publication Publication Date Title
Zhao et al. Incomplete multi-view clustering via deep semantic mapping
WO2022041678A1 (en) Remote sensing image feature extraction method employing tensor collaborative graph-based discriminant analysis
CN104281855B (en) Hyperspectral image classification method based on multi-task low rank
CN106446936B (en) Hyperspectral data classification method based on convolutional neural network combined spatial spectrum data to waveform map
CN107590515B (en) Hyperspectral image classification method of self-encoder based on entropy rate superpixel segmentation
CN112381144B (en) Heterogeneous deep network method for non-European and Euclidean domain space spectrum feature learning
CN107341786A (en) The infrared and visible light image fusion method that wavelet transformation represents with joint sparse
CN104732546B (en) The non-rigid SAR image registration method of region similitude and local space constraint
CN105574534A (en) Significant object detection method based on sparse subspace clustering and low-order expression
CN104318243B (en) High-spectral data dimension reduction method based on rarefaction representation and empty spectrum Laplce's figure
CN103413151A (en) Hyperspectral image classification method based on image regular low-rank expression dimensionality reduction
CN107451565A (en) A kind of semi-supervised small sample deep learning image model classifying identification method
CN107609573A (en) High spectrum image time varying characteristic extracting method based on low-rank decomposition and empty spectrum constraint
CN108197650A (en) The high spectrum image extreme learning machine clustering method that local similarity is kept
CN104008383B (en) Based on manifold learning linearizing high spectrum image feature extracting method
CN106023221A (en) Remote sensing image segmentation method based on nonnegative low-rank sparse correlated drawing
CN102800113B (en) Digital image analysis method based on fractal dimension
CN105550641A (en) Age estimation method and system based on multi-scale linear differential textural features
CN104463219A (en) Polarimetric SAR image classification method based on eigenvector measurement spectral clustering
CN101877065B (en) Extraction and identification method of non-linear authentication characteristic of facial image under small sample condition
Zhang et al. Discriminative tensor sparse coding for image classification.
CN104463246A (en) Manifold-based linear regression learning method
CN102902984B (en) Remote-sensing image semi-supervised projection dimension reducing method based on local consistency
CN104050482B (en) A kind of manifold learning generalization algorithm based on local linear smoothing
CN106446965A (en) Spacecraft visible light image classification method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant