CN112580670A - Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning - Google Patents
Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning Download PDFInfo
- Publication number
- CN112580670A CN112580670A CN202011633323.6A CN202011633323A CN112580670A CN 112580670 A CN112580670 A CN 112580670A CN 202011633323 A CN202011633323 A CN 202011633323A CN 112580670 A CN112580670 A CN 112580670A
- Authority
- CN
- China
- Prior art keywords
- hyperspectral
- spatial
- data
- network
- layer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/243—Classification techniques relating to the number of classes
- G06F18/2431—Multiple classes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a hyperspectral and spatial spectrum combined feature extraction method based on transfer learning, and belongs to the field of deep learning remote sensing. The method for extracting the space-spectrum combined characteristic of the hyperspectral data comprises the steps of firstly designing 1D CNN and 2D CNN to respectively extract the spectrum and the space characteristic of the hyperspectral data, and then fusing the two parts of characteristics. In order to overcome the contradiction that a deep neural network needs a large amount of training data and hyperspectral data lacks of marked samples, the invention adopts a method that a model ResNet-18 pre-trained on an RGB image data set ImageNet is transferred to a hyperspectral image target domain, thereby realizing network parameter sharing and reducing the calculation cost of the training model. And training a SoftMax layer to realize a hyperspectral target classification task based on the extracted combined features. And finally, the model after migration is more suitable for the hyperspectral data through a fine-tuned migration learning strategy, and the classification precision is improved. The invention has clear structure, easy realization and deep theoretical basis and practical significance.
Description
Technical Field
The invention mainly relates to the field of hyperspectral image classification, in particular to a hyperspectral data space spectrum combined feature extraction method based on transfer learning, which is used for hyperspectral classification.
Background
With the development of spectral imaging technology, hyperspectral imaging has attracted extensive attention in the field of remote sensing due to its broadband detection from visible light to near-infrared wavelength range. The applications of target detection and image classification for hyperspectral data are becoming more and more mature. The early hyperspectral classification method relies on artificial expertise to a great extent, and classification basis is mostly based on shallow features, so that the classification precision is low and the cost is high. With the development of deep learning in recent years, the deep learning method has better feature extraction and expression capability by simulating more complex functions and extracting features hierarchically.
The deep learning method is applied to the hyperspectral classification problem, so that the algorithm performance can be improved, the cost of a large number of manual labels is reduced, and deeper features are excavated. Some representative methods are a Stacked Autoencoder (SAE), a limited boltzmann machine (RBM), and a Convolutional Neural Network (CNN). The CNN is a deep learning network most widely applied in the field of hyperspectral classification, and commonly adopted structures include 1D CNN, 2D CNN, 3D CNN, mixed structure CNN networks and the like.
However, there are two contradictions to the hyperspectral classification problem based on deep learning. Firstly, the spectrum characteristics of different hyperspectral categories have low difference, and the error rate of classification by only utilizing the spectrum characteristics is high. Secondly, along with the increase of the number of the wave bands participating in the operation, the classification precision is increased firstly and then decreased, and the deep learning model is easy to generate the phenomenon of overfitting. For the problem of intra-class difference, a corresponding method is usually adopted to embed the spatial domain information into the spectral features, i.e., the spatial spectral features of the hyperspectrum are comprehensively considered and then classified. Some representative methods for the hough (Hughes) phenomenon are Transfer Learning (TL), data enhancement (DA) and unsupervised/semi-supervised learning. Transfer learning refers to a method of transferring network parameters from a source domain to a target domain, thereby reducing the random initialization process and achieving better performance. The advantages are that: (1) a large number of training samples are not required; (2) training time is saved by sharing the weights of the pre-trained model.
In many methods for applying transfer learning to hyperspectral image problems, a source domain and a target domain which are transferred generally belong to homologous or heterologous hyperspectral data. For example, Yang et al propose that a dual-branch CNN structure based on transfer learning is used for hyperspectral classification by extracting spatio-spectral union features, and although this method saves a random initialization process through transfer learning, a large amount of hyperspectral labeled data is still required to pre-train a model. For example, in a three-dimensional residual network and a transfer learning hyperspectral image classification method provided in patent CN109754017A, a deeper small sample hyperspectral image classification of a network model is realized by transferring hyperspectral Data acquired by different sensors, but the invention aims at transfer learning among heterogeneous hyperspectrum, but does not solve the problem that the feature dimension and the sample number are not matched due to the hous phenomenon Hughes in the hyperspectral classification problem, the feature extracted by a 3D network structure is deeper, the feature dimension is higher, and the Data Shift (Data Shift) problem among heterogeneous hyperspectrum cannot be effectively improved.
However, in recent years, training test data sets of various deep neural network models (such as ResNet, VGG-Net, Google-Net, etc.) which perform better in various image classification tasks are all mass-collected RGB images. In order to fully utilize such excellent model structures and hopefully achieve the effect of improving the hyperspectral classification by transfer learning, and solve the problem of few hyperspectral labeled training data, the problem of dimension mismatch in the process of transferring the RGB image to the hyperspectral image must be solved at first.
To apply the model trained on the 2D RGB image dataset directly to the 3D hyperspectral data, a method of morphing the RGB data or the model may be used. For example, Zhang et al expand RGB images into 3D cube data and input the 3D cube data into a 3D CNN model for pre-training, and then input hyperspectral cube data into the model to realize transfer learning. The method has the advantages that the existing large number of RGB image data resources can be fully utilized, the random initialization process is saved, and the characteristic expression capability of the model is improved. However, due to the physical difference between the hyperspectral data and the RGB data, the classification precision of the migrated model on the hyperspectral data is not high enough, and the model which is trained by expanding and pre-training the RGB image has no way to fully utilize the hyperspectral abundant spectral features, weakens the expression of the spectral features, and does not fundamentally solve the problem that the characteristic dimension caused by the Hughes phenomenon in the hyperspectral classification problem is not matched with the number of samples.
Generally, the existing hyperspectral image classification method has poor classification effect or takes too long time, is subject to high cost of manual labeling of hyperspectral data, and has less data. The factors seriously restrict the application of artificial intelligence in the field of hyperspectral image classification. Aiming at the defects of the existing hyperspectral image classification method based on deep learning, the invention provides an improved hyperspectral image classification method based on the past hyperspectral spectral feature or spatial feature classification. On the basis of fully utilizing hyperspectral data information, a double-branched 1D CNN structure and a double-branched 2D CNN structure are respectively used for extracting hyperspectral spectral information and spatial information, and classification prediction is realized according to fusion of two characteristics. Meanwhile, by adopting a transfer learning method in the hyperspectral spatial feature extraction process, the structural superiority of ResNet-18 is fully utilized, a random initialization process of network parameters is omitted by virtue of a large number of RGB image data set pre-training models, the aim of reducing training time is fulfilled, and finally, a fine-tuning strategy is utilized to act on the whole target domain hyperspectral data set, so that the effect of obviously improving classification precision is achieved.
Disclosure of Invention
The invention mainly solves the technical problems that: aiming at the problems of high marking cost, few data set sources, long training time consumption and the like of the traditional hyperspectral image classification data, the invention provides the hyperspectral classification method which is clear in structure, easy to realize, good in classification effect and short in time consumption.
In the invention, a network (ResNet-18) pre-trained on a 2D RGB image data set (ImageNet) is adopted to extract spatial features of hyperspectral data, meanwhile, spectral features are extracted through 1D CNN and embedded into the whole model, the information of the hyperspectral data set is fully utilized, and finally, the model after migration is more suitable for the data set with a hyperspectral target domain through a fine-tuning strategy.
In the hyperspectral image classification problem, different from the action range of the traditional transfer learning strategy, millions of 2D RGB images are regarded as a source domain (source domain) of transfer learning, and a hyperspectral data set is regarded as a target domain of the transfer learning. The invention designs a double-branch feature extraction network model which is respectively used for extracting the spectral features and the spatial features of the hyperspectral image. The 1D CNN branch is formed by connecting three convolution layers, three maximum pooling layers and a full-link layer in a crossed manner and aims to extract the spectral characteristics of hyperspectrum.
The depth of the network is an important factor for realizing good feature expression capability of the deep neural network, however, under the condition that the depth of the network is continuously deepened, gradient explosion becomes a great obstacle for training the deep network, and the phenomenon can cause that the network cannot be converged or starts to degenerate even if the network convergence accuracy rate is close to saturation. The network degradation can be prevented by superposing a layer of identity mapping on the basis of a shallow network. This is because the multi-layer nonlinear network cannot approximate the fitted identity mapping network. ResNet is a special case in short-cut connections (shortcuts connections) based identity mapping, and does not introduce additional parameters and increase the complexity of calculation. The ResNet structure has been the first of the classification tasks of the ImageNet game because of its excellent performance, and is also widely used in the fields of image detection, segmentation, recognition, and the like. The ResNet network is simple in structure and suitable for transfer learning in an image classification task. To extract deeper spatial features while avoiding excessive superparameters, the present invention selects the ResNet-18 network as the network structure in the spatial branch.
The technical scheme of the invention is as follows:
a hyperspectral and spatial spectrum combined feature extraction method based on transfer learning comprises the following steps:
s1: representing the original hyperspectral data as I e RM×N×LWhere I represents each sample in a data cube captured by a hyperspectral image sensor, M is the height, N is the width, and L represents the spectrum; before training a neural network model, preprocessing an original hyperspectral image, and performing Principal Component Analysis (PCA) operation along a spectral channel to reduce the dimension of hyperspectral data so as to reduce the calculation cost; reducing the number of spectral channels from L to k by PCA and keeping the space size unchanged; data after PCA pre-processing is expressed as I epsilon RM×N×kWhere k is the number of principal components, where the value of k is set to 3 corresponding to the number of RGB channels.
After RCA operation is performed, the hyperspectral data is expressed as I e RM×N×3Representing each pixel of the hyperspectral image as xij=[x1,x2,...,xk]Wherein x isij∈XLExtracting r pixels around each pixel to form an image, and representing each sample asThe following:
when x isijWhen the image data is positioned at the edge of the image, zero filling operation is carried out on the part exceeding the edge of the image, so that the image data input into the spatial feature extraction network is preprocessed;
after pre-processing, each pixel is converted from the original 1D vector to a picture of size (2r +1) × (2r +1) × 3; after data preprocessing, making I be belonged to R(2r+1)×(2r+1)×3As input to the spatial branch, let I ∈ R3×3×LAs input to the spectral branch;
s2, extracting spatial features based on a ResNet-18 network model structure, and directly transferring the pre-trained network parameters on a two-dimensional image data set ImageNet into a network for extracting the spatial features of hyperspectral data by using a transfer learning method, wherein the ResNet-18 network structure consists of 17 convolution layers and 1 full-connection layer, ResNet-18 is represented by stacked residual blocks, and the residual blocks of different convolution structures are connected through downsampling with stride of 2; the ResNet-18 network adopts the sizes of a convolution kernel and a pooling layer kernel, specifically, a convolution layer is formed by a filter with the size of 3 multiplied by 3, a pooling layer is formed by a filter with the size of 2 multiplied by 2, a full-link layer is formed after the average pooling operation after convolution, the FC layer of the last layer of the ResNet-18 structure comprises 1000 channels, and the last part of the model is used for classification by a SoftMax classifier; the migration model for spatial feature extraction deletes the FC layer of the ResNet-18 structure, newly adds an FC layer with the number of neurons being 1024, and keeps the rest unchanged to serve as a spatial branch network; a schematic diagram of a spatial feature extraction framework based on ResNet-18 transfer learning is shown in fig. 2.
S3, applying the 1D CNN network to spectral feature extraction of hyperspectrum, wherein the 1D CNN network is formed by connecting three convolution layers with convolution kernel size of 1 x 1 and filter size of 512, three maximum pooling layers with size of 2 x 2 and a full-link layer containing 1024 neurons; after extraction of spectral and spatial features by 1D CNN and ResNet-18, respectively, the last two fully connected layers from the spatial and spectral branches are connected together as follows:
where F denotes a cascaded FC layer of the spatial spectrum joint feature,the FC layer representing the spatial branch is,the last FC layer representing a spectral branch; the method comprises the steps of fusing two full-link layers to achieve space-spectrum combined feature extraction, and finally adding a SoftMax classifier to output to predict probability; the schematic diagram of the network structure is shown in fig. 1.
S4, adopting a fine-tuning strategy method to take the process of the network parameters of the ResNet-18 model which is transferred on the RGB image and is pre-trained as a step for randomly initializing the alternative space branch model, then training the whole model with the spectrum branches, and updating and optimizing each parameter in the new training process, so that the adaptability of the transfer model to the hyperspectral data is higher, and the classification effect is better.
Compared with the prior art, the invention has the advantages that:
1. in the invention, a spatial feature extraction model based on transfer learning is provided by applying a transfer learning method to a hyperspectral image classification problem. This is the first attempt to apply a pre-trained ResNet-18 based network on an RGB image dataset (ImageNet) directly to a hyperspectral classification target domain, enabling better classification performance than transfer learning between two homologous or heterologous hyperspectral datasets.
2. In order to fully utilize spectral information and spatial information provided by a hyperspectral imaging technology, a double-branch feature extractor structure is designed to combine spatial and spectral features. By training the whole double-branch model in combination with the transfer learning, the classification effect of the combined space spectrum feature given to the transfer learning is expected to be better than that of the spectrum feature or the space feature alone.
3. In order to make the migrated model more suitable for a new hyperspectral data set, the parameters of the whole model are optimized and updated by adopting a fine-tuning strategy, so that the classification precision is further improved, the model can express hyperspectral features more accurately on the premise of not generating overfitting, and the problem of model incompatibility caused by trans-physical imaging migration learning is relieved to a certain extent.
Drawings
FIG. 1 is a schematic diagram of a network framework structure for performing space-spectrum joint feature extraction according to the present invention;
FIG. 2 is a schematic diagram of a network structure of a transfer learning ResNet-18 proposed by the present invention;
FIG. 3 is a schematic representation of a hyperspectral data set employed in an example embodiment of the invention;
FIG. 4 is a diagram illustrating the pre-processing operation of spatially branched input data in an exemplary embodiment of the present invention;
fig. 5 is an overall framework diagram of the present invention.
Detailed Description
The invention will be described in further detail below with reference to the drawings and specific examples.
A hyperspectral and spatial spectrum combined feature extraction method based on transfer learning comprises the following steps:
s1: representing the original hyperspectral data as I e RM×N×LWhere I represents each sample in a data cube captured by a hyperspectral image sensor, M is the height, N is the width, and L represents the spectrum; before training a neural network model, preprocessing an original hyperspectral image, and performing Principal Component Analysis (PCA) operation along a spectral channel to reduce the dimension of hyperspectral data so as to reduce the calculation cost; reducing the number of spectral channels from L to k by PCA and keeping the space size unchanged; data after PCA pre-processing is expressed as I epsilon RM×N×kWhere k is the number of principal components, whereThe value of k is set to 3 corresponding to the number of RGB channels.
After performing the PCA operation, the hyperspectral data is represented as I ∈ RM×N×3Representing each pixel of the hyperspectral image as xij=[x1,x2,...,xk]Wherein x isij∈XLExtracting r pixels surrounding each pixel to compose an image, and representing each sample as follows:
when x isijWhen the image data is positioned at the edge of the image, zero filling operation is carried out on the part exceeding the edge of the image, so that the image data input into the spatial feature extraction network is preprocessed;
after pre-processing, each pixel is converted from the original 1D vector to a picture of size (2r +1) × (2r +1) × 3; after data preprocessing, making I be belonged to R(2r+1)×(2r+1)×3As input to the spatial branch, let I ∈ R3×3×LAs input to the spectral branch;
s2: extracting spatial features based on a ResNet-18 network model structure, and directly transferring network parameters pre-trained on a two-dimensional image data set ImageNet into a network for extracting the spatial features of hyperspectral data by using a transfer learning method, wherein the ResNet-18 network structure consists of 17 convolutional layers and 1 full-connection layer, ResNet-18 is represented by stacked residual blocks, and the residual blocks of different convolutional structures are connected through downsampling with stride of 2; the ResNet-18 network adopts the sizes of a convolution kernel and a pooling layer kernel, specifically, a convolution layer is formed by a filter with the size of 3 multiplied by 3, a pooling layer is formed by a filter with the size of 2 multiplied by 2, a full-link layer is formed after the average pooling operation after convolution, the FC layer of the last layer of the ResNet-18 structure comprises 1000 channels, and the last part of the model is used for classification by a SoftMax classifier; the migration model for spatial feature extraction deletes the FC layer of the ResNet-18 structure, newly adds an FC layer with the number of neurons being 1024, and keeps the rest unchanged to serve as a spatial branch network; a schematic diagram of a spatial feature extraction framework based on ResNet-18 transfer learning is shown in fig. 2.
S3: the 1D CNN network is used for spectral feature extraction of hyperspectrum, wherein the 1D CNN network is formed by connecting three convolution layers with convolution kernel size of 1 x 1 and filter size of 512, three maximum pooling layers with size of 2 x 2 and a full-connection layer containing 1024 neurons; after extraction of spectral and spatial features by 1D CNN and ResNet-18, respectively, the last two fully connected layers from the spatial and spectral branches are connected together as follows:
where F denotes a cascaded FC layer of the spatial spectrum joint feature,the FC layer representing the spatial branch is,the last FC layer representing a spectral branch; the method comprises the steps of fusing two full-link layers to achieve space-spectrum combined feature extraction, and finally adding a SoftMax classifier to output to predict probability; the schematic diagram of the network structure is shown in FIG. 1
S4: and a fine-tuning strategy method is adopted to only serve the process of the network parameters of the ResNet-18 model which is transferred on the RGB image and is pre-trained as a step for randomly initializing the alternative space branch model, then the whole model with the spectrum branches is trained, and each parameter is updated and optimized in a new training process, so that the adaptability of the transfer model to the hyperspectral data is higher, and the classification effect is better.
As shown in fig. 3, in order to more vividly describe the method provided by the present invention, the present invention adopts a hyperspectral image scene of the University of pavian (Pavia University) as a target hyperspectral dataset, and realizes the classification prediction of Pavia University data by transferring a model pre-trained on an RGB dataset ImageNet to a hyperspectral dataset Pavia University. The Pavia University dataset was acquired by the ROSIS sensor at a pixel size of 610 × 340 with a spectral band number of 103. The ground truth of the university of parkia is divided into 9 categories.
S1: as shown in fig. 4, after PCA preprocessing is performed on the Pavia University hyperspectral dataset, the first three principal components are extracted, and a picture similar to an RGB image composed of three channels is formed. Then, the field r of each pixel is expanded, and zero padding is performed on the edge pixels, so that each pixel is converted into a picture with the size of (2r +1) × (2r +1) × 3 from the original one-dimensional spectral vector, and the picture is used as the input of the spatial branch in the invention. According to the input picture size requirement of the ResNet-18 model, the value of the field r is larger than 32, and therefore 33 is selected as the size of r. After preprocessing, each pixel is converted into a picture with the size of 67 × 67 × 3 from the original one-dimensional spectral vector as the input of a spatial feature extraction branch, each 3 × 3 pixel is extracted by a spectral branch, and a rectangular data block with the spectral dimension of 103 is used as the input of the spectral branch.
S2, as shown in FIG. 2, the model of migration of the spatial-branched network structure core block is composed of classical ResNet-18, which removes the last full-link layer and adds a full-link layer composed of 1024 neurons. The network parameters are obtained by direct migration of parameters pre-trained on the RGB image dataset ImageNet.
And S3, as shown in the figure 1, performing feature extraction and feature fusion on the hyperspectral data set Pavia University through spectrum branching and space branching to obtain a hyperspectral classification model based on transfer learning and space spectrum combined feature extraction.
S4, performing optimization, updating and retraining on the parameters of the whole network by a fine adjustment method, and finally performing prediction and classification on 9 categories corresponding to the SoftMax layer to obtain prediction results of different pixels in the hyperspectral image, wherein a flow diagram of the whole method is shown in FIG. 5.
The above is only a preferred embodiment of the present invention, and the protection scope of the present invention is not limited to the above-mentioned embodiments, and all technical solutions belonging to the idea of the present invention belong to the protection scope of the present invention. It should be noted that modifications and embellishments within the scope of the invention may be made by those skilled in the art without departing from the principle of the invention.
Claims (1)
1. The hyperspectral and spatial spectrum combined feature extraction method based on transfer learning is characterized by comprising the following steps:
s1, representing the original hyperspectral data as I e RM×N×LWhere I represents each sample in a data cube captured by a hyperspectral image sensor, M is the height, N is the width, and L represents the spectrum; before training a neural network model, preprocessing an original hyperspectral image, and performing Principal Component Analysis (PCA) operation along a spectral channel to reduce the dimension of hyperspectral data so as to reduce the calculation cost; reducing the number of spectral channels from L to k by PCA and keeping the space size unchanged; data after PCA pre-processing is expressed as I epsilon RM×N×kWhere k is the number of principal components, where the value of k is set to 3 corresponding to the number of RGB channels;
after performing the PCA operation, the hyperspectral data is represented as I ∈ RM×N×3Representing each pixel of the hyperspectral image as xij=[x1,x2,…,xk]Wherein x isij∈XLExtracting r pixels surrounding each pixel to compose an image, and representing each sample as follows:
when x isijWhen the image data is positioned at the edge of the image, zero filling operation is carried out on the part exceeding the edge of the image, so that the image data input into the spatial feature extraction network is preprocessed;
after pre-processing, each pixel is converted from the original 1D vector to a picture of size (2r +1) × (2r +1) × 3; after data preprocessing, making I be belonged to R(2r+1)×(2r+1)×3As input to the spatial branch, let I eR3×3×LAs input to the spectral branch;
s2, extracting spatial features based on a ResNet-18 network model structure, and directly transferring the pre-trained network parameters on a two-dimensional image data set ImageNet into a network for extracting the spatial features of hyperspectral data by using a transfer learning method, wherein the ResNet-18 network structure consists of 17 convolution layers and 1 full-connection layer, ResNet-18 is represented by stacked residual blocks, and the residual blocks of different convolution structures are connected through downsampling with stride of 2; the ResNet-18 network adopts the sizes of a convolution kernel and a pooling layer kernel, specifically, a convolution layer is formed by a filter with the size of 3 multiplied by 3, a pooling layer is formed by a filter with the size of 2 multiplied by 2, a full-link layer is formed after the average pooling operation after convolution, the FC layer of the last layer of the ResNet-18 structure comprises 1000 channels, and the last part of the model is used for classification by a SoftMax classifier; the migration model for spatial feature extraction deletes the FC layer of the ResNet-18 structure, newly adds an FC layer with the number of neurons being 1024, and keeps the rest unchanged to serve as a spatial branch network;
s3, applying the 1D CNN network to spectral feature extraction of hyperspectrum, wherein the 1D CNN network is formed by connecting three convolution layers with convolution kernel size of 1 x 1 and filter size of 512, three maximum pooling layers with size of 2 x 2 and a full-link layer containing 1024 neurons; after extraction of spectral and spatial features by 1D CNN and ResNet-18, respectively, the last two fully connected layers from the spatial and spectral branches are connected together as follows:
where F denotes a cascaded FC layer of the spatial spectrum joint feature,the FC layer representing the spatial branch is,the last FC layer representing a spectral branch; the method comprises the steps of fusing two full-link layers to achieve space-spectrum combined feature extraction, and finally adding a SoftMax classifier to output to predict probability;
s4, adopting a fine-tuning strategy method to take the process of the network parameters of the ResNet-18 model which is transferred on the RGB image and is pre-trained as a step for randomly initializing the alternative space branch model, then training the whole model with the spectrum branches, and updating and optimizing each parameter in the new training process, so that the adaptability of the transfer model to the hyperspectral data is higher, and the classification effect is better.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011633323.6A CN112580670B (en) | 2020-12-31 | 2020-12-31 | Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011633323.6A CN112580670B (en) | 2020-12-31 | 2020-12-31 | Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112580670A true CN112580670A (en) | 2021-03-30 |
CN112580670B CN112580670B (en) | 2022-04-19 |
Family
ID=75144539
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011633323.6A Active CN112580670B (en) | 2020-12-31 | 2020-12-31 | Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112580670B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113159189A (en) * | 2021-04-23 | 2021-07-23 | 南京理工大学 | Hyperspectral image classification method and system based on double-branch multi-attention convolutional neural network |
CN113708863A (en) * | 2021-09-10 | 2021-11-26 | 中国人民解放军63891部队 | Method and device for constructing spectrum sensing training data set |
CN114037671A (en) * | 2021-11-01 | 2022-02-11 | 大连医科大学附属第二医院 | Microscopic hyperspectral leukocyte detection method based on improved fast RCNN |
CN114202690A (en) * | 2021-12-09 | 2022-03-18 | 东北林业大学 | Multi-scale network analysis method based on mixed multilayer perceptron |
CN115375951A (en) * | 2022-09-20 | 2022-11-22 | 中国矿业大学 | Small sample hyperspectral image classification method based on primitive migration network |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2187339A1 (en) * | 2008-11-12 | 2010-05-19 | Fundación Robotiker | Method for integrating spectral and spatial features for classifying materials |
CN107451616A (en) * | 2017-08-01 | 2017-12-08 | 西安电子科技大学 | Multi-spectral remote sensing image terrain classification method based on the semi-supervised transfer learning of depth |
CN109766858A (en) * | 2019-01-16 | 2019-05-17 | 中国人民解放军国防科技大学 | Three-dimensional convolution neural network hyperspectral image classification method combined with bilateral filtering |
CN109886870A (en) * | 2018-12-29 | 2019-06-14 | 西北大学 | Remote sensing image fusion method based on binary channels neural network |
CN109961096A (en) * | 2019-03-19 | 2019-07-02 | 大连理工大学 | A kind of multimode high spectrum image migration classification method |
CN111191736A (en) * | 2020-01-05 | 2020-05-22 | 西安电子科技大学 | Hyperspectral image classification method based on depth feature cross fusion |
CN111723731A (en) * | 2020-06-18 | 2020-09-29 | 西安电子科技大学 | Hyperspectral image classification method based on spatial spectrum convolution kernel, storage medium and device |
EP3716136A1 (en) * | 2019-03-26 | 2020-09-30 | Koninklijke Philips N.V. | Tumor boundary reconstruction using hyperspectral imaging |
CN111914696A (en) * | 2020-07-16 | 2020-11-10 | 河海大学 | Hyperspectral remote sensing image classification method based on transfer learning |
CN111985543A (en) * | 2020-08-06 | 2020-11-24 | 西北大学 | Construction method, classification method and system of hyperspectral image classification model |
-
2020
- 2020-12-31 CN CN202011633323.6A patent/CN112580670B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2187339A1 (en) * | 2008-11-12 | 2010-05-19 | Fundación Robotiker | Method for integrating spectral and spatial features for classifying materials |
CN107451616A (en) * | 2017-08-01 | 2017-12-08 | 西安电子科技大学 | Multi-spectral remote sensing image terrain classification method based on the semi-supervised transfer learning of depth |
CN109886870A (en) * | 2018-12-29 | 2019-06-14 | 西北大学 | Remote sensing image fusion method based on binary channels neural network |
CN109766858A (en) * | 2019-01-16 | 2019-05-17 | 中国人民解放军国防科技大学 | Three-dimensional convolution neural network hyperspectral image classification method combined with bilateral filtering |
CN109961096A (en) * | 2019-03-19 | 2019-07-02 | 大连理工大学 | A kind of multimode high spectrum image migration classification method |
EP3716136A1 (en) * | 2019-03-26 | 2020-09-30 | Koninklijke Philips N.V. | Tumor boundary reconstruction using hyperspectral imaging |
CN111191736A (en) * | 2020-01-05 | 2020-05-22 | 西安电子科技大学 | Hyperspectral image classification method based on depth feature cross fusion |
CN111723731A (en) * | 2020-06-18 | 2020-09-29 | 西安电子科技大学 | Hyperspectral image classification method based on spatial spectrum convolution kernel, storage medium and device |
CN111914696A (en) * | 2020-07-16 | 2020-11-10 | 河海大学 | Hyperspectral remote sensing image classification method based on transfer learning |
CN111985543A (en) * | 2020-08-06 | 2020-11-24 | 西北大学 | Construction method, classification method and system of hyperspectral image classification model |
Non-Patent Citations (2)
Title |
---|
CHENG DENG 等: "Active Transfer Learning Network: A Unified Deep Joint Spectral–Spatial Feature Learning Model for Hyperspectral Image Classification", 《IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING》 * |
DEEP RESIDUAL PROTOTYPE LEARNING NETWORK FOR HYPERSPECTRAL IMAGE: "Liu, Yu 等", 《PROCEEDINGS OF THE SPIE》 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113159189A (en) * | 2021-04-23 | 2021-07-23 | 南京理工大学 | Hyperspectral image classification method and system based on double-branch multi-attention convolutional neural network |
CN113708863A (en) * | 2021-09-10 | 2021-11-26 | 中国人民解放军63891部队 | Method and device for constructing spectrum sensing training data set |
CN113708863B (en) * | 2021-09-10 | 2023-08-01 | 中国人民解放军63891部队 | Method and device for constructing spectrum sensing training data set |
CN114037671A (en) * | 2021-11-01 | 2022-02-11 | 大连医科大学附属第二医院 | Microscopic hyperspectral leukocyte detection method based on improved fast RCNN |
CN114202690A (en) * | 2021-12-09 | 2022-03-18 | 东北林业大学 | Multi-scale network analysis method based on mixed multilayer perceptron |
CN114202690B (en) * | 2021-12-09 | 2024-04-12 | 东北林业大学 | Multi-scale network analysis method based on hybrid multi-layer perceptron |
CN115375951A (en) * | 2022-09-20 | 2022-11-22 | 中国矿业大学 | Small sample hyperspectral image classification method based on primitive migration network |
Also Published As
Publication number | Publication date |
---|---|
CN112580670B (en) | 2022-04-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112580670B (en) | Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning | |
Guo et al. | Scene-driven multitask parallel attention network for building extraction in high-resolution remote sensing images | |
CN111915530B (en) | End-to-end-based haze concentration self-adaptive neural network image defogging method | |
CN111259828B (en) | High-resolution remote sensing image multi-feature-based identification method | |
CN110136075B (en) | Remote sensing image defogging method for generating countermeasure network based on edge sharpening cycle | |
CN111339862B (en) | Remote sensing scene classification method and device based on channel attention mechanism | |
CN110751271B (en) | Image traceability feature characterization method based on deep neural network | |
CN113538457A (en) | Video semantic segmentation method utilizing multi-frequency dynamic hole convolution | |
He et al. | LKAT-GAN: a GAN for thermal infrared image colorization based on large kernel and attentionunet-transformer | |
Jiang et al. | An Improved Semantic Segmentation Method for Remote Sensing Images Based on Neural Network. | |
CN113763417B (en) | Target tracking method based on twin network and residual error structure | |
CN113139431A (en) | Image saliency target detection method based on deep supervised learning | |
CN117115563A (en) | Remote sensing land coverage classification method and system based on regional semantic perception | |
Luo et al. | A fast denoising fusion network using internal and external priors | |
CN117115632A (en) | Underwater target detection method, device, equipment and medium | |
Özyurt et al. | A new method for classification of images using convolutional neural network based on Dwt-Svd perceptual hash function | |
CN115965968A (en) | Small sample target detection and identification method based on knowledge guidance | |
Yang et al. | Multicue contrastive self-supervised learning for change detection in remote sensing | |
CN111860668B (en) | Point cloud identification method for depth convolution network of original 3D point cloud processing | |
CN112560824B (en) | Facial expression recognition method based on multi-feature adaptive fusion | |
Zhang et al. | Single image haze removal for aqueous vapour regions based on optimal correction of dark channel | |
Khoond et al. | Image enhancement using nonlocal prior and gradient residual minimization for improved visualization of deep underwater image | |
Zeng et al. | Analysis of hyperspectral image classification technology and application based on convolutional neural networks | |
Anand et al. | A Survey on Semantic Segmentation Models for Underwater Images | |
Futschik | Colorization of black-and-white images using deep neural networks |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |