CN114373080A - Hyperspectral classification method of lightweight hybrid convolution model based on global reasoning - Google Patents
Hyperspectral classification method of lightweight hybrid convolution model based on global reasoning Download PDFInfo
- Publication number
- CN114373080A CN114373080A CN202210279207.1A CN202210279207A CN114373080A CN 114373080 A CN114373080 A CN 114373080A CN 202210279207 A CN202210279207 A CN 202210279207A CN 114373080 A CN114373080 A CN 114373080A
- Authority
- CN
- China
- Prior art keywords
- global
- hyperspectral
- classification
- hyperspectral classification
- model based
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 35
- 239000000284 extract Substances 0.000 claims abstract description 20
- 238000000605 extraction Methods 0.000 claims abstract description 14
- 238000013527 convolutional neural network Methods 0.000 claims description 64
- 230000003595 spectral effect Effects 0.000 claims description 43
- 238000000513 principal component analysis Methods 0.000 claims description 14
- 238000012545 processing Methods 0.000 claims description 7
- 230000003993 interaction Effects 0.000 claims description 5
- 238000013528 artificial neural network Methods 0.000 claims description 2
- 238000010276 construction Methods 0.000 claims description 2
- 238000012549 training Methods 0.000 abstract description 23
- 230000010365 information processing Effects 0.000 abstract description 2
- 239000010410 layer Substances 0.000 description 21
- 230000000694 effects Effects 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 3
- 238000011156 evaluation Methods 0.000 description 3
- 238000002474 experimental method Methods 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 238000010521 absorption reaction Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 230000007812 deficiency Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000012880 independent component analysis Methods 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 235000008331 Pinus X rigitaeda Nutrition 0.000 description 1
- 235000011613 Pinus brutia Nutrition 0.000 description 1
- 241000018646 Pinus brutia Species 0.000 description 1
- 241001466077 Salina Species 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000007637 random forest analysis Methods 0.000 description 1
- 239000002356 single layer Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
- 238000010396 two-hybrid screening Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
本发明公开基于全局推理的轻量化混合卷积模型的高光谱分类方法,属于信息处理技术领域,主要用于小样本数据的高光谱图像分类。所述基于全局推理的轻量化混合卷积模型包括一层二维卷积、一层三维卷积和一个全局推理模块;加入全局推理模块,通过对不同区域间的上下文关系推理有效提取高光谱图像的全局特征和深层特征信息,以代替深层三维卷积的特征提取,大大降低了模型的复杂度和计算成本。公开数据集中的测试结果表明,本发明的分类性能优于当前最好的分类方法,仅需少量训练样本,就可有效提取高光谱图像的空‑谱联合特征,解决了仅使用二维卷积造成的信道关系信息缺失的问题以及采用深层三维卷积造成的模型复杂度和计算成本大大增加的问题。
The invention discloses a hyperspectral classification method of a lightweight hybrid convolution model based on global reasoning, belongs to the technical field of information processing, and is mainly used for hyperspectral image classification of small sample data. The lightweight hybrid convolution model based on global inference includes a layer of two-dimensional convolution, a layer of three-dimensional convolution and a global inference module; adding a global inference module can effectively extract hyperspectral images by inferring contextual relationships between different regions To replace the feature extraction of deep 3D convolution, the complexity and computational cost of the model are greatly reduced. The test results in the public data set show that the classification performance of the present invention is better than the current best classification method, and only a small number of training samples can be used to effectively extract the space-spectral joint feature of the hyperspectral image, which solves the problem of using only two-dimensional convolution. The problem of missing channel relationship information and the problem of greatly increased model complexity and computational cost caused by the use of deep three-dimensional convolution.
Description
技术领域technical field
本发明公开基于全局推理的轻量化混合卷积模型的高光谱分类方法,属于信息处理技术领域。The invention discloses a hyperspectral classification method of a lightweight hybrid convolution model based on global reasoning, and belongs to the technical field of information processing.
背景技术Background technique
高光谱图像(HSI)因具有众多的光谱波段给地物信息的提取带来了极大的帮助,但大量的光谱数据也造成了地物信息的冗余,同时因光谱分辨率的提高使得各波段间的相关性加强,这给高光谱图像的分类带来了不小的挑战。尤其是当训练样本较少时,会给HSI的分类带来更大的困难。Hyperspectral image (HSI) has brought great help to the extraction of ground object information due to its numerous spectral bands, but a large amount of spectral data also caused the redundancy of ground object information. The correlation between the bands is strengthened, which brings a lot of challenges to the classification of hyperspectral images. Especially when there are few training samples, it will bring greater difficulties to the classification of HSI.
目前,HSI分类的方法主要有两类:基于光谱信息的传统方法和基于深度学习的方法。传统方法主要是利用不同地物的光谱特征信息进行分类,其中支持向量机、k近邻、随机森林等最具有代表性。为了获取更好的分类性能,主成分分析(PCA),独立成分分析(ICA)和线性判别分析(LDA)等方法被应用以有效的进行数据降维或特征提取。尽管传统的HSI分类方法取得了良好的分类性能,但由于它们处理时很大程度上依赖的是手工制作的特征或提取的浅层特征,特征表达能力有限,无法更好地适应复杂的HSI分类任务。At present, there are mainly two types of HSI classification methods: traditional methods based on spectral information and methods based on deep learning. The traditional method mainly uses the spectral feature information of different objects to classify, among which support vector machine, k-nearest neighbor and random forest are the most representative. In order to obtain better classification performance, methods such as Principal Component Analysis (PCA), Independent Component Analysis (ICA) and Linear Discriminant Analysis (LDA) are applied to effectively perform data dimensionality reduction or feature extraction. Although traditional HSI classification methods have achieved good classification performance, they cannot better adapt to complex HSI classification because they largely rely on hand-crafted features or extracted shallow features. Task.
近年来,深度学习因其强大的特征表达能力已被成功应用在众多的计算机视觉任务中,如图像分类,图像分割,目标检测等。巨大的应用潜力使得深度学习方法在HSI分类领域也取得了显著的效果,其中基于卷积神经网络(CNN)的HSI分类方法最为流行。W.Hu等人首次采用深度CNN直接在光谱域中对高光谱图像进行分类,获得了良好的分类效果。KonstantinosMakantasis等人使用2D-CNN对高光谱像元的邻域进行空间特征提取,同时采用PCA方法有效降维,降低了模型的计算成本。但该模型仅能提取空间特征信息,无法获得光谱信道间的关系信息。YushiChen等人将3D-CNN方法用于HSI任务,通过3D-CNN同时提取光谱-空间的联合特征,有效提升了分类精度。在此基础上,M.He等人提出了多尺度的深度3D-CNN模型,该模型可以端到端地从HSI数据中共同学习多尺度空间特征和光谱特征,进一步提升了HSI的分类性能。In recent years, deep learning has been successfully applied in numerous computer vision tasks, such as image classification, image segmentation, object detection, etc., due to its powerful feature representation ability. The huge application potential makes deep learning methods also achieve remarkable results in the field of HSI classification, among which HSI classification methods based on convolutional neural networks (CNN) are the most popular. W. Hu et al. used deep CNN for the first time to classify hyperspectral images directly in the spectral domain, and obtained good classification results. Konstantinos Makantasis et al. used 2D-CNN to extract spatial features from the neighborhood of hyperspectral pixels, and used PCA method to effectively reduce dimensionality and reduce the computational cost of the model. However, this model can only extract spatial feature information, and cannot obtain the relationship information between spectral channels. YushiChen et al. applied the 3D-CNN method to the HSI task, and simultaneously extracted the spectral-spatial joint features through 3D-CNN, which effectively improved the classification accuracy. On this basis, M. He et al. proposed a multi-scale deep 3D-CNN model, which can jointly learn multi-scale spatial features and spectral features from HSI data end-to-end, further improving the classification performance of HSI.
然而,利用3D-CNN模型进行HSI分类,将需要深度的卷积层才能有效提取空-谱联合特征,从而导致模型复杂度和训练样本数量的大大增加。基于此,有研究人员已经开始将2D-CNN与3D-CNN结合用于HSI分类。2020年,SwalpaKumarRoy等人提出一种基于HybridSN模型的HSI分类方法,该方法将3D-CNN和2D-CNN组合,充分提取空间特征信息和光谱特征信息,相比于单独使用3D-CNN进行HSI分类,在计算成本和分类性能方面均取得了较大的提升。2021年,SaeedGhaderizadeh等人提出了一种3D-2DCNN模型。该模型通过引入三维深度可分卷积块和快速卷积块,使模型的鲁棒性和效率更高。但值得注意的是,为使模型具有更强的空间特征提取能力,模型中往往会包含多个甚至深度的3D卷积层,在文献中,都采用了多至三层的3D卷积进行特征的提取。这种通过多层3D卷积的堆叠提取深层特征的方法,一方面会使得模型的复杂度和计算成本增加,另一方面也需要大量的样本进行训练,而对于HSI来说,大量样本的标注也是一个不小的挑战。However, utilizing the 3D-CNN model for HSI classification will require deep convolutional layers to effectively extract the joint spatial-spectral features, resulting in a large increase in model complexity and the number of training samples. Based on this, some researchers have begun to combine 2D-CNN with 3D-CNN for HSI classification. In 2020, SwalpaKumarRoy et al. proposed an HSI classification method based on the HybridSN model, which combines 3D-CNN and 2D-CNN to fully extract spatial feature information and spectral feature information, compared to using 3D-CNN alone for HSI classification. , which has achieved great improvements in computational cost and classification performance. In 2021, SaeedGhaderizadeh et al. proposed a 3D-2DCNN model. This model makes the model more robust and efficient by introducing 3D depthwise separable convolution blocks and fast convolution blocks. But it is worth noting that in order to make the model have stronger spatial feature extraction ability, the model often contains multiple or even deep 3D convolution layers. In the literature, up to three layers of 3D convolution are used for feature extraction. extraction. This method of extracting deep features through the stacking of multi-layer 3D convolutions increases the complexity and computational cost of the model on the one hand, and requires a large number of samples for training on the other hand. For HSI, the annotation of a large number of samples It's also quite a challenge.
发明内容SUMMARY OF THE INVENTION
本发明的目的在于提供基于全局推理的轻量化混合卷积模型的高光谱分类方法,以解决现有技术中,高光谱分类方法包括多个三维卷积层而造成模型复杂、计算成本高、样本需求大的问题。The purpose of the present invention is to provide a hyperspectral classification method based on a lightweight hybrid convolutional model based on global reasoning, so as to solve the problem that in the prior art, the hyperspectral classification method includes multiple three-dimensional convolutional layers, resulting in complex models, high computational costs, and high sample size. high demand problem.
基于全局推理的轻量化混合卷积模型的高光谱分类方法,包括:A hyperspectral classification method based on a lightweight hybrid convolutional model based on global inference, including:
S1.设原始高光谱数据表示为I∈RM×N×D,I是原始高光谱分类输入,M和N是空间维的宽和高,D是光谱波段数;S1. Let the original hyperspectral data be represented as I∈RM×N×D, where I is the original hyperspectral classification input, M and N are the width and height of the spatial dimension, and D is the number of spectral bands;
S2.采用主成分分析对高光谱分类光谱信息降维,经过主成分分析处理后的高光谱数据表示为X∈RM×N×B,其空间维数保持不变,光谱波段数由D变为B;S2. Principal component analysis is used to reduce the dimension of spectral information for hyperspectral classification. The hyperspectral data processed by principal component analysis is expressed as X∈RM×N×B, and its spatial dimension remains unchanged, and the number of spectral bands changes from D to B;
S3.对高光谱分类的邻域块提取进行预处理,对高光谱分类的每一个像元,都以像元为中心点提取一个色斑,色斑的标签即为中心像元的标签,表示为Z∈RS×S×B,色斑的宽和高为S,光谱波段数为B;S3. Preprocess the neighborhood block extraction of hyperspectral classification. For each pixel of hyperspectral classification, extract a color spot with the pixel as the center point. The label of the color spot is the label of the center pixel, indicating is Z∈RS×S×B, the width and height of the color spot are S, and the number of spectral bands is B;
S4.将高光谱分类数据送入基于全局推理的轻量化混合卷积模型中进行处理。S4. The hyperspectral classification data is sent to a lightweight hybrid convolution model based on global inference for processing.
优选地,所述基于全局推理的轻量化混合卷积模型包括二维卷积神经网络、三维卷积神经网络和全局推理模块。Preferably, the lightweight hybrid convolutional model based on global inference includes a two-dimensional convolutional neural network, a three-dimensional convolutional neural network and a global inference module.
优选地,所述三维卷积神经网络采用8个3×3×7的卷积核,空间维度为3×3,光谱维度为7。Preferably, the three-dimensional convolutional neural network adopts 8 convolution kernels of 3×3×7, the spatial dimension is 3×3, and the spectral dimension is 7.
优选地,所述二维卷积神经网络采用16个3×3的卷积核,空间维度为3×3。Preferably, the two-dimensional convolutional neural network adopts 16 3×3 convolution kernels, and the spatial dimension is 3×3.
优选地,所述全局推理模块是二维的,其构建方法为,将输入数据的坐标空间上不相交区域聚合的特征投影到一个潜在的交互空间中,用一个单独的特征即一个节点来表示,应用图卷积的方式来建模和推理每对节点之间的上下文关系,再对其进行反向投影,将具有关系信息的特征变换回原始坐标空间,获得相关的全局特征信息。Preferably, the global reasoning module is two-dimensional, and its construction method is to project the aggregated features of disjoint regions in the coordinate space of the input data into a potential interaction space, and use a single feature, that is, a node to represent , using graph convolution to model and reason the context relationship between each pair of nodes, and then back-project it to transform the features with relationship information back to the original coordinate space to obtain relevant global feature information.
优选地,步骤S4包括:Preferably, step S4 includes:
S4.1.采用三维卷积神经网络对预处理后的高光谱分类数据进行空间特征和光谱特征的联合提取;S4.1. Use a three-dimensional convolutional neural network to jointly extract the spatial and spectral features of the preprocessed hyperspectral classification data;
S4.2.采用二维卷积神经网络对S4.1处理后的高光谱分类数据进行空间特征和光谱特征的联合提取;S4.2. Use a two-dimensional convolutional neural network to jointly extract spatial features and spectral features from the hyperspectral classification data processed in S4.1;
S4.3.采用全局推理模块对S4.2处理后的高光谱分类数据进行全局特征提取。S4.3. Use the global inference module to extract global features from the hyperspectral classification data processed in S4.2.
与现有技术相比,本发明仅使用一层3D-CNN和一层2D-CNN的轻量级混合卷积网络模型,降低网络模型复杂度的同时,能够有效提取HSI的空-谱联合特征;在设计的混合卷积网络模型中加入了轻量化全局推理模块,通过对不同区域间的上下文关系进行推理来有效提取HSI的全局特征,HSI的分类性能得到显著提升;提出的模型不仅降低了现有混合网络模型的复杂度,而且在仅需很少的样本训练时,就能获得非常好的分类性能。Compared with the prior art, the present invention only uses a lightweight hybrid convolutional network model of one layer of 3D-CNN and one layer of 2D-CNN, which can effectively extract the space-spectral joint feature of HSI while reducing the complexity of the network model. ; In the designed hybrid convolutional network model, a lightweight global reasoning module is added to effectively extract the global features of HSI by inferring the contextual relationship between different regions, and the classification performance of HSI is significantly improved; the proposed model not only reduces the The complexity of existing hybrid network models, and very good classification performance can be obtained when training only a few samples.
附图说明Description of drawings
图1是本发明的基于全局推理的轻量化混合卷积模型的结构图;Fig. 1 is the structure diagram of the lightweight hybrid convolution model based on global reasoning of the present invention;
图2是全局推理模块的原理框架图。Figure 2 is a schematic diagram of the global reasoning module.
具体实施方式Detailed ways
下面结合具体实施方式对本发明作进一步详细说明:Below in conjunction with specific embodiment, the present invention is described in further detail:
实施例首先介绍基于全局推理的轻量化混合卷积模型模型的总体结构,然后介绍3D-CNN和2D-CNN混合卷积模型的结构和参数设计,最后给出轻量化全局推理模块的核心思想。The embodiment first introduces the overall structure of the lightweight hybrid convolution model based on global reasoning, then introduces the structure and parameter design of the 3D-CNN and 2D-CNN hybrid convolution models, and finally gives the core idea of the lightweight global inference module.
深度卷积神经网络模型:单一的深度二维卷积或三维卷积网络作为高光谱分类的特征提取模型,会导致缺失光谱信道关系或模型复杂度高。现有的用于高光谱分类的混合卷积网络模型中3D-CNN的层数也较多,一定程度增加了模型的复杂度、计算成本以及样本的标注成本。本发明仅需一层3D-CNN和一层2D-CNN组合,再加上全局推理模块,即可在少量样本的情况下获得高性能的高光谱分类效果。Deep convolutional neural network model: A single deep 2D convolutional or 3D convolutional network as a feature extraction model for hyperspectral classification can lead to missing spectral channel relationships or high model complexity. The existing hybrid convolutional network model for hyperspectral classification also has a large number of layers of 3D-CNN, which increases the complexity of the model, the computational cost and the cost of labeling the sample to a certain extent. The present invention only needs a combination of a layer of 3D-CNN and a layer of 2D-CNN, plus a global inference module, to obtain high-performance hyperspectral classification effects with a small number of samples.
设原始高光谱数据表示为I∈RM×N×D,其中I是原始高光谱分类输入,M和N空间维的宽和高,D是光谱波段数,一般有几十甚至几百个波段。高光谱分类众多的波段信息不仅会给后续处理任务带来较大的计算成本,而且当训练样本有限时,高光谱分类的总体分类精度会随着特征维数的增加而降低,即“维数灾难”问题。因此,降低高光谱分类的光谱维度、减少信息的冗余势在必行。主成分分析(PCA)作为常见的降维方式,核心思想是对各个不同数据之间的特征相似度进行计算,提取主要特征。本发明采用PCA的方法对高光谱分类光谱信息降维,以有效改善因信息冗余造成计算成本大和分类精度低的问题。经过PCA处理后表示为X∈RM×N×B,空间维数保持不变,光谱波段数由D变为B。Let the original hyperspectral data be represented as I∈R M×N×D , where I is the original hyperspectral classification input, M and N spatial dimensions of width and height, D is the number of spectral bands, generally tens or even hundreds of bands . The large number of band information in hyperspectral classification will not only bring a large computational cost to subsequent processing tasks, but also when the training samples are limited, the overall classification accuracy of hyperspectral classification will decrease with the increase of feature dimension, that is, “dimension disaster" issue. Therefore, it is imperative to reduce the spectral dimension of hyperspectral classification and reduce the redundancy of information. Principal component analysis (PCA) is a common dimensionality reduction method, and the core idea is to calculate the feature similarity between different data and extract the main features. The invention adopts the PCA method to reduce the dimension of the hyperspectral classification spectral information, so as to effectively improve the problems of high computational cost and low classification accuracy caused by information redundancy. After PCA processing, it is expressed as X∈R M×N×B , the spatial dimension remains unchanged, and the number of spectral bands changes from D to B.
为了更好的利用后续卷积网络进行分类处理,在PCA降维后,对高光谱分类的邻域块提取进行预处理。即对高光谱分类的每一个像元,都以其为中心点提取一个patch。每个patch的标签即为中心像元的标签,表示为Z∈RS×S×B,patch的宽和高为S,光谱波段数为B。In order to better utilize the subsequent convolutional network for classification processing, after PCA dimensionality reduction, the neighborhood block extraction of hyperspectral classification is preprocessed. That is, for each pixel of hyperspectral classification, a patch is extracted from it as the center point. The label of each patch is the label of the central pixel, expressed as Z∈R S×S×B , the width and height of the patch are S, and the number of spectral bands is B.
高光谱分类数据经过PCA以及邻域块提取的预处理后,送入基于全局推理的轻量化混合卷积模型模型中进行处理。单层3D-CNN和2D-CNN的混合完成对光谱和空间特征的联合提取并降低模型复杂度,全局推理模块通过对不同区域间的上下文关系进行推理来提取全局特征,解决了为提取表达能力更强的特征需采用计算成本大的多层3D-CNN的问题。The hyperspectral classification data is preprocessed by PCA and neighborhood block extraction, and then sent to a lightweight hybrid convolution model based on global inference for processing. The mixture of single-layer 3D-CNN and 2D-CNN completes the joint extraction of spectral and spatial features and reduces the complexity of the model. The global inference module extracts global features by inferring the contextual relationship between different regions, which solves the problem of extracting expressive capabilities. Stronger features require computationally expensive multi-layer 3D-CNN problems.
3D-2D CNN:高光谱分类是一个包括几十甚至几百个波段的体数据,包含空间信息和光谱信息。对于传统的2D-CNN,仅能提取空间特征信息;而3D-CNN却能同时提取空间特征和光谱特征。所以,近年来基于3D-CNN的高光谱分类成为主流,但由于深度3D-CNN模型复杂,导致模型的计算成本大且需要大量的训练样本。现有的几种用于高光谱分类的混合卷积模型具有良好的分类性能,能在一定程度上降低训练样本的数量。但多层的3D-CNN堆叠带来计算成本提高的问题仍不可忽视,也没能彻底解决训练样本数量较大的问题。为了确保高光谱分类数据的空间信息和光谱信息的完整性,首先采用3D-CNN对预处理后的高光谱分类数据进行空间特征和光谱特征的联合提取。在3D-CNN层,采用了8个3×3×7的卷积核,即空间维度为3×3,光谱维度为7。在图1中体现为K1 1=3,K2 1=3,K3 1=7。为充分提取高光谱分类数据的空间特征信息,在采用3D-CNN进行特征提取后,再用一层2D-CNN提取,以提升高光谱分类的分类效果。在2D-CNN中,采用了16个3×3的卷积核,即空间维度为3×3。在图1中,体现为K1 3=3,K2 3=3。3D-2D CNN: Hyperspectral classification is a volume data that includes dozens or even hundreds of bands, including spatial information and spectral information. For traditional 2D-CNN, only spatial feature information can be extracted; while 3D-CNN can extract spatial features and spectral features at the same time. Therefore, in recent years, hyperspectral classification based on 3D-CNN has become the mainstream, but due to the complexity of the deep 3D-CNN model, the computational cost of the model is large and a large number of training samples are required. Several existing hybrid convolution models for hyperspectral classification have good classification performance and can reduce the number of training samples to a certain extent. However, the problem of increased computational cost brought about by multi-layer 3D-CNN stacking cannot be ignored, and the problem of a large number of training samples cannot be completely solved. In order to ensure the integrity of the spatial and spectral information of the hyperspectral classification data, 3D-CNN is firstly used to jointly extract the spatial and spectral features of the preprocessed hyperspectral classification data. In the 3D-CNN layer, 8 convolution kernels of 3×3×7 are used, that is, the spatial dimension is 3×3 and the spectral dimension is 7. In FIG. 1 , it is shown that K 1 1 =3, K 2 1 =3, and K 3 1 =7. In order to fully extract the spatial feature information of hyperspectral classification data, after using 3D-CNN for feature extraction, a layer of 2D-CNN extraction is used to improve the classification effect of hyperspectral classification. In 2D-CNN, 16 3×3 convolution kernels are used, that is, the spatial dimension is 3×3. In Figure 1, it is shown that K 1 3 =3, K 2 3 =3.
单个卷积层能够较好的捕获卷积核覆盖的局部关系,但捕捉表达能力更强的全局区域之间的关系则需要堆叠多个甚至深度的卷积层,这无疑增加了CNN全局推理的难度和成本。现有的几种用于高光谱分类的混合卷积模型均是采用这种低效的多层3D-CNN堆叠的设计方式提取全局特征。本发明提出的混合卷积模型中添加的全局推理模块就是为了克服多层卷积运算对全局关系建模的固有不足。A single convolutional layer can better capture the local relationship covered by the convolution kernel, but capturing the relationship between global regions with stronger expressiveness requires stacking multiple or even deep convolutional layers, which undoubtedly increases the global inference ability of CNN. difficulty and cost. Several existing hybrid convolutional models for hyperspectral classification all use this inefficient multi-layer 3D-CNN stacking design to extract global features. The global reasoning module added in the hybrid convolution model proposed by the present invention is to overcome the inherent deficiency of the multi-layer convolution operation for modeling the global relationship.
全局推理模块是一个轻量化全局推理模块,通过对区域之间进行全局关系建模和推理,提取出相关的全局、深层次的特征,其原理框架如图2所示,表示了交互空间和坐标空间之间的投影和反向投影,首先,将输入数据的坐标空间上不相交区域聚合的特征投影到一个潜在的交互空间中,用一个单独的特征即一个节点来表示。通过这种方式,关系推理被简化为在图2顶部显示的较小的图上对节点之间进行交互建模。然后,应用图卷积的方式来建模和推理每对节点之间的上下文关系。最后,再对其进行反向投影,将具有关系信息的特征变换回原始坐标空间,即可获得相关的全局特征信息。The global reasoning module is a lightweight global reasoning module. It extracts relevant global and deep-level features by modeling and reasoning the global relationship between regions. The principle framework is shown in Figure 2, which represents the interaction space and coordinates. Projection and back-projection between spaces. First, the features aggregated from disjoint regions in the coordinate space of the input data are projected into a potential interaction space, represented by a single feature, a node. In this way, relational reasoning is reduced to modeling interactions between nodes on the smaller graph shown at the top of Figure 2. Then, a graph convolution approach is applied to model and reason about the contextual relationship between each pair of nodes. Finally, it is back-projected to transform the features with relational information back to the original coordinate space, and then the relevant global feature information can be obtained.
高光谱分类数据经过3D-CNN层提取空-谱联合特征后,有着较为完整的空间信息和光谱信息,这有利于对高光谱分类进行全局关系推理。因此,在3D-CNN层后加入模块,以提取高光谱分类的全局特征。因3D层次的模块复杂度较高,本发明采用2D层次的模块来聚合高光谱分类数据的全局特征,这恰好与后续2D-CNN层提取空间特征信息匹配。充分的结构组合实验验证了“3D-CNN++2D-CNN”组合的模型分类性能最好。针对IP数据集所提出模型的结构参数中,可训练的权重参数的总数为2,275,504,如表1。The hyperspectral classification data has relatively complete spatial information and spectral information after the 3D-CNN layer extracts the spatial-spectral joint features, which is conducive to the global relational reasoning for the hyperspectral classification. Therefore, a module is added after the 3D-CNN layer to extract global features for hyperspectral classification. Due to the high complexity of the modules at the 3D level, the present invention adopts the modules at the 2D level to aggregate the global features of the hyperspectral classification data, which exactly matches the spatial feature information extracted by the subsequent 2D-CNN layer. The sufficient structure combination experiment verifies that the "3D-CNN++2D-CNN" combination has the best model classification performance. Among the structural parameters of the proposed model for the IP dataset, the total number of trainable weight parameters is 2,275,504, as shown in Table 1.
表1 IP数据集的本发明所用模型概述Table 1 Overview of the models used in the present invention for the IP dataset
实施例使用了三个高光谱分类的公共数据集作为实验数据集,即Salinas Scene(SA)、University of Pavia (UP)和Indian Pines (IP)。SA数据集有16个类别,空间维为512×217,在360-2500nm波长范围内有224个光谱波段。去除20个吸水谱带后保留204个光谱波段。UP数据集有9个类别,空间维为610×340,在430-860nm波长范围内有103个光谱波段。IP数据集有16个类别,空间维为145×145,在400-2500nm波长范围内,有224个光谱波段。去除24个吸水谱带后保留200个光谱波段。The examples used three public datasets for hyperspectral classification as experimental datasets, namely Salinas Scene (SA), University of Pavia (UP) and Indian Pines (IP). The SA dataset has 16 categories with a spatial dimension of 512 × 217 and 224 spectral bands in the wavelength range of 360-2500 nm. After removing 20 water absorption bands, 204 spectral bands remained. The UP dataset has 9 categories with a spatial dimension of 610 × 340 and 103 spectral bands in the wavelength range of 430-860 nm. The IP dataset has 16 categories with a spatial dimension of 145 × 145 and 224 spectral bands in the 400-2500 nm wavelength range. 200 spectral bands remained after removing 24 water absorption bands.
所有的实验都是在一台联想-Y7000P电脑上进行的。该电脑配备了GTX 1660ti图形处理器(GPU)和16 GB RAM。使用SGD优化器,学习率设置为0.01。并采用momentum和L2正则化,参数分别设为0.8和0.0005。在模型训练中使用大小为32的mini-batches。另外,为了加快训练过程和提升泛化性能,又对模型进行了batch normalization处理。将网络模型训练的迭代次数设置为50。为了与其他方法公平比较,对于不同的数据集,从PCA降维处理后的数据中提取相同的邻域块,大小为27×27×20。All experiments are performed on a Lenovo-Y7000P computer. The computer is equipped with a GTX 1660ti Graphics Processing Unit (GPU) and 16 GB of RAM. Using the SGD optimizer, the learning rate is set to 0.01. Momentum and L2 regularization are used, and the parameters are set to 0.8 and 0.0005, respectively. Use mini-batches of size 32 in model training. In addition, in order to speed up the training process and improve the generalization performance, batch normalization is performed on the model. Set the number of iterations for network model training to 50. For a fair comparison with other methods, for different datasets, the same neighborhood blocks of size 27 × 27 × 20 are extracted from the PCA-reduced data.
使用Kappa系数(Kappa)、总体精度(OA)和平均精度(AA)作为评价指标来判断高光谱分类性能。Kappa反映了信息之间的一致性信息;OA是指正确分类的个数所占比例;AA是指每一种类别正确分类比例的平均值。The hyperspectral classification performance was judged using Kappa coefficient (Kappa), overall accuracy (OA) and average accuracy (AA) as evaluation metrics. Kappa reflects the consistent information between information; OA refers to the proportion of the correct classification; AA refers to the average of the correct classification proportion of each category.
不同的Window Size和Spectral Dimension对LH-CNN模型性能的影响中,窗口尺寸和光谱维数为27×27×20的邻域块时,效果最好,如表2和表3所示。Among the effects of different Window Size and Spectral Dimension on the performance of the LH-CNN model, the window size and spectral dimension of 27 × 27 × 20 neighborhood blocks have the best effect, as shown in Table 2 and Table 3.
表2 空间窗口大小对模型性能的影响Table 2 Influence of spatial window size on model performance
表3 光谱维数对模型性能的影响Table 3 Influence of spectral dimension on model performance
为验证所设计的模型在少量训练样本条件下的分类性能,与高光谱分类常用的CNN方法和目前最好的混合卷积方法,即2-D-CNN,3-D-CNN,multi-scale 3-D deepconvolutional neural network(M3D)-CNN,HybridSN and 3D-2D CNN进行对比。各种方法的训练样本均从数据集中随机选取,样本比例为1%。在只有1%训练数据情况下,本发明提出的模型相比于其他方法,性能更好,如表4所示。In order to verify the classification performance of the designed model under the condition of a small number of training samples, it is compared with the commonly used CNN methods for hyperspectral classification and the current best hybrid convolution methods, namely 2-D-CNN, 3-D-CNN, multi-scale 3-D deepconvolutional neural network (M3D)-CNN, HybridSN and 3D-2D CNN for comparison. The training samples of various methods are randomly selected from the data set, and the sample ratio is 1%. In the case of only 1% training data, the model proposed by the present invention has better performance than other methods, as shown in Table 4.
表4 IP、UP和SA数据集的分类准确率(百分比)(1%样本用于培训)Table 4 Classification accuracy (percentage) of IP, UP and SA datasets (1% of samples are used for training)
特别是在样本量较少的IP数据集上的实验,从Kappa、OA和AA三种评价指标来看,提出的模型性能远超过最好的混合CNN模型。混合CNN模型的复杂度以及在不同数据集中的计算开销,本发明提出的模型无论在模型复杂度方面还是计算开销方面都是最优的,全局推理模块对于高光谱分类性能的影响显著。Especially in the experiments on the IP dataset with a small sample size, from the three evaluation indicators of Kappa, OA and AA, the performance of the proposed model far exceeds the best hybrid CNN model. The complexity of the hybrid CNN model and the computational cost in different data sets, the model proposed by the present invention is optimal in terms of model complexity and computational cost, and the global inference module has a significant impact on the performance of hyperspectral classification.
为了进一步验证所提模型在少量训练样本情况下的分类性能,减少训练样本数量,在UP和SA数据集中随机选取0.5%的样本数据作为训练样本(这里IP数据集因部分类别样本数量较少,无法选取0.5%的样本数据),设计的模型性能最好,且远超过最好的两种混合CNN的分类结果,尤其是在UP数据集上的分类结果,各类评价指标几乎都超出6个百分点,如表5,充分证明了所设计模型在少量训练样本情况下的优越性。In order to further verify the classification performance of the proposed model in the case of a small number of training samples and reduce the number of training samples, 0.5% of the sample data is randomly selected as training samples in the UP and SA data sets (here, the IP data set has a small number of samples in some categories. 0.5% of the sample data cannot be selected), the designed model has the best performance, and far exceeds the classification results of the best two hybrid CNNs, especially the classification results on the UP data set, all kinds of evaluation indicators almost exceed 6 Percentage points, as shown in Table 5, fully demonstrate the superiority of the designed model in the case of a small number of training samples.
表5 UP和SA数据集的分类准确率(百分比)(0.5%样本用于训练)Table 5 Classification accuracy (percentage) of UP and SA datasets (0.5% samples are used for training)
值得注意的是,在0.5%的训练数据下,3D-2D CNN的性能稍低于HybridSN,这与在1%的训练数据下的结果不同。分析原因,或许是由于3D-2D CNN模型更复杂、参数更多,需要更多的样本数据才能训练出分类性能好的模型,这也是复杂网络模型本身的不足。It is worth noting that under 0.5% training data, the performance of 3D-2D CNN is slightly lower than HybridSN, which is different from the results under 1% training data. The reason for the analysis may be that the 3D-2D CNN model is more complex and has more parameters, requiring more sample data to train a model with good classification performance, which is also the deficiency of the complex network model itself.
当然,上述说明并非是对本发明的限制,本发明也并不仅限于上述举例,本技术领域的技术人员在本发明的实质范围内所做出的变化、改型、添加或替换,也应属于本发明的保护范围。Of course, the above description is not intended to limit the present invention, and the present invention is not limited to the above examples. Changes, modifications, additions or substitutions made by those skilled in the art within the essential scope of the present invention should also belong to the present invention. the scope of protection of the invention.
Claims (6)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210279207.1A CN114373080B (en) | 2022-03-22 | 2022-03-22 | Hyperspectral classification method based on lightweight hybrid convolution model based on global inference |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210279207.1A CN114373080B (en) | 2022-03-22 | 2022-03-22 | Hyperspectral classification method based on lightweight hybrid convolution model based on global inference |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114373080A true CN114373080A (en) | 2022-04-19 |
CN114373080B CN114373080B (en) | 2022-07-29 |
Family
ID=81146006
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210279207.1A Active CN114373080B (en) | 2022-03-22 | 2022-03-22 | Hyperspectral classification method based on lightweight hybrid convolution model based on global inference |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114373080B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114511786A (en) * | 2022-04-20 | 2022-05-17 | 中国石油大学(华东) | A cloud removal method for remote sensing images by fusing multi-temporal information and sub-channel dense convolution |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110210313A (en) * | 2019-05-06 | 2019-09-06 | 河海大学 | United Hyperspectral Remote Sensing Imagery Classification method is composed based on multiple dimensioned PCA-3D-CNN sky |
US20200151546A1 (en) * | 2018-11-08 | 2020-05-14 | Netflix, Inc. | Identifying image aesthetics using region composition graphs |
CN111192270A (en) * | 2020-01-03 | 2020-05-22 | 中山大学 | Point cloud semantic segmentation method based on point global context reasoning |
CN111461181A (en) * | 2020-03-16 | 2020-07-28 | 北京邮电大学 | A kind of vehicle fine-grained classification method and device |
CN111695469A (en) * | 2020-06-01 | 2020-09-22 | 西安电子科技大学 | Hyperspectral image classification method of lightweight depth separable convolution feature fusion network |
CN111967297A (en) * | 2020-06-28 | 2020-11-20 | 北京百度网讯科技有限公司 | Semantic segmentation method and device for image, electronic equipment and medium |
US20210081780A1 (en) * | 2019-09-13 | 2021-03-18 | Honda Motor Co., Ltd. | System and method for providing object-level driver attention reasoning with a graph convolution network |
CN112820377A (en) * | 2021-02-02 | 2021-05-18 | 中国科学技术大学 | Radiotherapy plan automatic generation method based on deep learning |
CN113837314A (en) * | 2021-10-09 | 2021-12-24 | 陕西科技大学 | A Hyperspectral Image Classification Method Based on Hybrid Convolutional Neural Networks |
CN113920442A (en) * | 2021-09-29 | 2022-01-11 | 中国人民解放军火箭军工程大学 | Hyperspectral classification method combining graph structure and convolutional neural network |
CN114092963A (en) * | 2021-10-14 | 2022-02-25 | 北京百度网讯科技有限公司 | Key point detection and model training method, device, equipment and storage medium |
-
2022
- 2022-03-22 CN CN202210279207.1A patent/CN114373080B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200151546A1 (en) * | 2018-11-08 | 2020-05-14 | Netflix, Inc. | Identifying image aesthetics using region composition graphs |
CN110210313A (en) * | 2019-05-06 | 2019-09-06 | 河海大学 | United Hyperspectral Remote Sensing Imagery Classification method is composed based on multiple dimensioned PCA-3D-CNN sky |
US20210081780A1 (en) * | 2019-09-13 | 2021-03-18 | Honda Motor Co., Ltd. | System and method for providing object-level driver attention reasoning with a graph convolution network |
CN111192270A (en) * | 2020-01-03 | 2020-05-22 | 中山大学 | Point cloud semantic segmentation method based on point global context reasoning |
CN111461181A (en) * | 2020-03-16 | 2020-07-28 | 北京邮电大学 | A kind of vehicle fine-grained classification method and device |
CN111695469A (en) * | 2020-06-01 | 2020-09-22 | 西安电子科技大学 | Hyperspectral image classification method of lightweight depth separable convolution feature fusion network |
CN111967297A (en) * | 2020-06-28 | 2020-11-20 | 北京百度网讯科技有限公司 | Semantic segmentation method and device for image, electronic equipment and medium |
CN112820377A (en) * | 2021-02-02 | 2021-05-18 | 中国科学技术大学 | Radiotherapy plan automatic generation method based on deep learning |
CN113920442A (en) * | 2021-09-29 | 2022-01-11 | 中国人民解放军火箭军工程大学 | Hyperspectral classification method combining graph structure and convolutional neural network |
CN113837314A (en) * | 2021-10-09 | 2021-12-24 | 陕西科技大学 | A Hyperspectral Image Classification Method Based on Hybrid Convolutional Neural Networks |
CN114092963A (en) * | 2021-10-14 | 2022-02-25 | 北京百度网讯科技有限公司 | Key point detection and model training method, device, equipment and storage medium |
Non-Patent Citations (4)
Title |
---|
TIANYI WU 等: "GINet: Graph Interaction Network for Scene Parsing", 《SPRINGER NATURE SWITZERLAND AG 2020》 * |
TIANYI WU 等: "GINet: Graph Interaction Network for Scene Parsing", 《SPRINGER NATURE SWITZERLAND AG 2020》, 31 December 2020 (2020-12-31), pages 34 - 51 * |
刘漳辉 等: "基于语义传播与前/背景感知的图像语义分割网络", 《模式识别与人工智能》, 31 January 2022 (2022-01-31), pages 71 - 81 * |
廖金雷 等: "融合植被指数的3D-2D-CNN高光谱图像植被分类方法", 《科学技术与工程》, 31 December 2021 (2021-12-31), pages 11656 - 11662 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114511786A (en) * | 2022-04-20 | 2022-05-17 | 中国石油大学(华东) | A cloud removal method for remote sensing images by fusing multi-temporal information and sub-channel dense convolution |
CN114511786B (en) * | 2022-04-20 | 2022-07-19 | 中国石油大学(华东) | A cloud removal method for remote sensing images combining multi-temporal information and sub-channel dense convolution |
Also Published As
Publication number | Publication date |
---|---|
CN114373080B (en) | 2022-07-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112926396B (en) | Action identification method based on double-current convolution attention | |
CN106372648B (en) | Plankton image classification method based on multi-feature fusion convolutional neural network | |
CN108427924B (en) | A Text Regression Detection Method Based on Rotation Sensitive Features | |
CN109766858A (en) | Three-dimensional convolution neural network hyperspectral image classification method combined with bilateral filtering | |
CN109711283B (en) | Occlusion expression recognition method combining double dictionaries and error matrix | |
CN107437099A (en) | A kind of specific dress ornament image recognition and detection method based on machine learning | |
CN110533024B (en) | Double-quadratic pooling fine-grained image classification method based on multi-scale ROI (region of interest) features | |
Wang et al. | Object instance detection with pruned Alexnet and extended training data | |
CN106504064A (en) | Clothes classification based on depth convolutional neural networks recommends method and system with collocation | |
CN111625667A (en) | Three-dimensional model cross-domain retrieval method and system based on complex background image | |
CN104636732B (en) | A kind of pedestrian recognition method based on the deep belief network of sequence | |
CN106203363A (en) | Human skeleton motion sequence Activity recognition method | |
CN104036255A (en) | Facial expression recognition method | |
CN109948593A (en) | Crowd Counting Method Based on MCNN Combined with Global Density Features | |
CN112070158A (en) | A Facial Defect Detection Method Based on Convolutional Neural Network and Bilateral Filtering | |
CN105205449A (en) | Sign language recognition method based on deep learning | |
Khalid et al. | DFGNN: An interpretable and generalized graph neural network for deepfakes detection | |
CN110674685A (en) | A Human Analytical Segmentation Model and Method Based on Edge Information Enhancement | |
Yuan et al. | Few-shot scene classification with multi-attention deepemd network in remote sensing | |
CN113486712B (en) | A deep learning-based multi-face recognition method, system and medium | |
CN114612802A (en) | System and method for classifying fine granularity of ship target based on MBCNN | |
CN113505856A (en) | Hyperspectral image unsupervised self-adaptive classification method | |
CN114373080B (en) | Hyperspectral classification method based on lightweight hybrid convolution model based on global inference | |
CN115661754B (en) | Pedestrian re-recognition method based on dimension fusion attention | |
CN114898157A (en) | Global learning device and method for hyperspectral image classification |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |