CN111368776A - High-resolution remote sensing image classification method based on deep ensemble learning - Google Patents
High-resolution remote sensing image classification method based on deep ensemble learning Download PDFInfo
- Publication number
- CN111368776A CN111368776A CN202010173481.1A CN202010173481A CN111368776A CN 111368776 A CN111368776 A CN 111368776A CN 202010173481 A CN202010173481 A CN 202010173481A CN 111368776 A CN111368776 A CN 111368776A
- Authority
- CN
- China
- Prior art keywords
- classification
- remote sensing
- layer
- network
- training set
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 54
- 238000012549 training Methods 0.000 claims abstract description 67
- 230000011218 segmentation Effects 0.000 claims abstract description 20
- 230000005484 gravity Effects 0.000 claims abstract description 13
- 238000012360 testing method Methods 0.000 claims description 49
- 238000013527 convolutional neural network Methods 0.000 claims description 27
- 230000006870 function Effects 0.000 claims description 21
- 238000011176 pooling Methods 0.000 claims description 14
- 230000004913 activation Effects 0.000 claims description 11
- 230000008569 process Effects 0.000 claims description 10
- 238000010606 normalization Methods 0.000 claims description 8
- 238000007781 pre-processing Methods 0.000 claims description 6
- 238000005192 partition Methods 0.000 claims description 5
- 238000005070 sampling Methods 0.000 claims description 5
- 238000012545 processing Methods 0.000 claims description 4
- 238000011478 gradient descent method Methods 0.000 claims description 3
- 238000005457 optimization Methods 0.000 claims description 3
- 238000012952 Resampling Methods 0.000 claims description 2
- 238000013145 classification model Methods 0.000 claims description 2
- 230000004927 fusion Effects 0.000 claims description 2
- 238000005096 rolling process Methods 0.000 claims description 2
- 238000012546 transfer Methods 0.000 claims description 2
- 238000001228 spectrum Methods 0.000 abstract description 3
- 238000013528 artificial neural network Methods 0.000 abstract description 2
- 238000007636 ensemble learning method Methods 0.000 abstract 1
- 238000010586 diagram Methods 0.000 description 9
- 238000000605 extraction Methods 0.000 description 6
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000003595 spectral effect Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 238000003709 image segmentation Methods 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000001668 ameliorated effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/13—Satellite images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/254—Fusion techniques of classification results, e.g. of results related to same input data
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Astronomy & Astrophysics (AREA)
- Remote Sensing (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a high-resolution remote sensing image classification method based on deep ensemble learning, which has the following thought: firstly, using pixel brightness values as classification features to perform a full-connection network classification experiment; secondly, performing object-oriented segmentation, and extracting a convolution block by taking the center of gravity as the center to perform convolution neural network classification; cutting all the original images into image blocks, and performing one vs all multi-element classification by using a U-Net complete convolution network; and finally, training a fully-connected network on the classification results of the first three deep network-based classifiers to perform probability combination, thereby realizing better classification performance. By using the deep ensemble learning method, the classification accuracy can be improved by effectively combining information such as spectrum, space and the like, the advantages of the base classifier are integrated, and the classification accuracy better than that of a single classifier can be obtained on the single classification and the overall classification accuracy of the high-resolution remote sensing image classification.
Description
Technical Field
The invention belongs to the technical field of remote sensing image analysis, and particularly relates to a high-resolution remote sensing image classification method based on deep ensemble learning.
Background
The remote sensing image classification goes through the process from visual interpretation to computer automatic classification, the traditional computer classification mainly comprises supervised classification and unsupervised classification based on pixels, the characteristic of the classification is mainly spectral characteristic, and the classification is carried out according to the difference of the pixel brightness values; with the improvement of the spatial resolution of the remote sensing image, the difference of the pixel brightness values of the same ground features is large, and the traditional pixel brightness value classification-based method cannot meet the classification precision required by the high-resolution remote sensing image.
Since the convolutional neural network designed by Alex Krizhevsky in 2012 plucks the canopy of the ILSVRC visual competition at one stroke, the hot tide of deep learning is raised, the application range of the convolutional neural network is wider and wider, and the accuracy is higher and higher along with the deepening of the network depth. The convolutional neural network is applied to high-spatial-resolution remote sensing image classification, automatic feature extraction is carried out by applying a convolutional layer and a pooling layer, and a weight is continuously corrected according to training loss, so that effective features suitable for a specific scene are extracted, and classification is further carried out by a classifier. Compared with the traditional computer classification, the method focuses more on the feature extraction in the early stage, and the classifier parameter optimization is not carried out to improve the classification precision.
However, for a high-resolution remote sensing image, a single classifier always has a certain problem, spectral information and spatial information of the image cannot be combined, and classification accuracy needs to be further improved.
Disclosure of Invention
Aiming at the defects of the traditional high-resolution remote sensing image classification method, the invention provides a high-resolution remote sensing image classification method based on deep ensemble learning. The method adopts three methods of pixel brightness value classification based on a full-connection network, object-oriented classification based on a convolutional neural network and remote sensing image classification based on a complete convolutional network for deep ensemble learning, effectively combines the spectrum and space information of the high-resolution remote sensing image, and improves the classification accuracy; the classification precision of single classification and overall classification of high-resolution remote sensing image classification is improved, and better classification performance is realized.
In order to achieve the technical purpose, the invention is realized by adopting the following technical scheme.
A high-resolution remote sensing image classification method based on deep ensemble learning comprises the following steps:
the samples in the training set and the test set are high-resolution remote sensing images and DSM data thereof respectively, and the DSM is a digital surface model;
and 4, classifying the samples in the test set by adopting a deep ensemble learning network, and outputting corresponding prediction class labels.
Compared with the prior art, the invention has the following advantages:
(1) the invention can construct different base classifiers aiming at specific application, obtain good accuracy on the base classifiers, and further integrate the base classifiers through a proper combination strategy to obtain better classification accuracy than a single classifier.
(2) The method adopts three methods of pixel brightness value classification based on a full-connection network, object-oriented classification based on a convolutional neural network and remote sensing image classification based on a complete convolutional network for deep ensemble learning, effectively combines the spectrum and space information of the high-resolution remote sensing image, and improves the classification accuracy; the classification precision of single classification and overall classification of high-resolution remote sensing image classification is improved, and better classification performance is realized.
Drawings
The present invention will be described in further detail with reference to the accompanying drawings and specific embodiments.
FIG. 1 is a schematic diagram of a deep ensemble learning network structure according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a fully-connected network model based on pixel brightness value classification according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a convolutional neural network model based on object-oriented classification according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a complete convolution network model based on remote sensing image classification according to an embodiment of the invention;
FIG. 5 is a schematic diagram of a segment and a center of gravity according to an embodiment of the present invention;
fig. 6 is a schematic diagram of a result of classification using only a fully-connected network based on pixel brightness value classification according to an embodiment of the present invention, where (a) is a label graph, (b) is a corresponding predicted classification result graph, and (c) is a median-filtered classification result graph;
FIG. 7 is a diagram illustrating the results of classification using only a convolutional neural network based on object-oriented classification according to an embodiment of the present invention, where (a) is a label graph and (b) is a corresponding predicted classification result graph;
FIG. 8 is a diagram illustrating the results of classification using only a complete convolutional network based on remote sensing image classification in accordance with an embodiment of the present invention; wherein, (a) is a label graph, and (b) is a corresponding prediction classification result graph;
FIG. 9 is a schematic diagram illustrating the results of remote sensing image classification by deep ensemble learning using a deep ensemble learning network according to an embodiment of the present invention, in which (a) is a label graph and (b) is a corresponding predicted classification result graph;
Detailed Description
Referring to fig. 1, a high-resolution remote sensing image classification method based on deep ensemble learning includes the following steps:
specifically, the obtained high-resolution remote sensing image and DSM data corresponding to the high-resolution remote sensing image are used as samples to form a training set and a test set respectively, and the number of samples of the training set is not less than that of the test set.
The full-connection network model based on pixel brightness value classification is composed of N full-connection layers, wherein the first N-1 layers respectively adopt a ReLU function as an activation function, and the last layer adopts a Softmax function as an activation function; the Dropout layer is used for regularization before the last fully connected layer to prevent overfitting. In the embodiment of the invention, a 7-layer fully-connected neural network is built, and the network structure is shown in figure 2.
The convolutional neural network model based on object-oriented classification comprises a plurality of convolutional layers, a maximum pooling layer, full-link layers and output layers, wherein one maximum pooling layer is arranged behind each convolutional layer, the last pooling layer is connected with the plurality of full-link layers, and the last full-link layer is arranged behind the output layers; wherein, the activation function of each convolution layer and the full connection layer is a ReLU function, and the activation function of the output layer is a Softmax function; the dryout layer is used for regularization before the last fully connected layer.
In the embodiment of the invention, a convolutional network comprising four convolutional layers, a pooling layer and three full-connection layers is built, the number of filter templates of the four convolutional layers is 128, 96, 64 and 64 respectively, the size of the filter template is 5 x 5, the filter templates are initialized by using Henormal, activation functions are ReLU functions, each convolutional layer is followed by a maximum pooling layer, the size of the filter template of the pooling layer is 3 x 3, the step size is 1, the number of neurons of the first full-connection layer and the second full-connection layer is 128, a Dropout layer is used for regularization after the second full-connection layer, an output layer comprising 6 neurons is followed, and the output layer adopts a Softmax function as the activation function. The network structure is shown in fig. 3.
The complete convolution network model based on remote sensing image classification is composed of a plurality of binary classifiers, wherein each binary classifier comprises a convolution layer, a Dropout regularization layer, a pooling layer, and a convolution layer, a fusion layer and an upper sampling layer which are symmetrical to the convolution layer, the Dropout regularization layer and the pooling layer.
The invention uses U-Net network to train, cut 15 images and DSM and label graph into 512 x 512 size images, and then expand data, apply the architecture as shown in figure 4 to train 5 groups of binary classifiers, and then use one vs all method to get the result of multi-element classification.
2.1, preprocessing the training set to obtain a preprocessed training set;
wherein, the preprocessing comprises unbalanced sample processing, data expansion and normalization;
specifically, the unbalanced sample treatment is: in the remote sensing image classification process, if the number of samples of each class in the training data set is extremely unbalanced, the number is an important factor influencing the classification result. In actual processing, the effects of unbalanced samples can be reduced or eliminated, both in terms of data and algorithms. In the data layer, the data resampling is carried out on the small category, namely, on the basis of the original samples of the small category, one of the value spaces of each attribute feature of all the samples under the category is randomly selected to form a new sample, so that the sample data of the small category is increased. Or data down-sampling large classes, i.e. randomly discarding a fraction of samples to keep the class number balanced, but this approach affects the generalization ability of the model. In the aspect of algorithm, the number of samples of each category is counted firstly, then corresponding weight coefficients are set, different sampling proportions are tried for each category, and therefore penalty cost of misclassification of the samples of the small categories is reduced.
The data expansion is as follows: horizontally turning (mirroring) or randomly rotating the original images in the training set; wherein, the original data volume can be doubled by horizontal turning; the random rotation is to designate an angle, so that the original image randomly generates the number of angles of the required expansion multiple within the given angle, and rotates.
Normalization: for color images of three wave bands of near infrared, red and green of QB (Quickbird) remote sensing satellite data, the gray values are all between 0 and 255, so the normalization is to simply scale the gray values from 0 to 255 to 0 to 1; for aviation images, color images comprising three bands of near infrared, red and green and DSMs with one band are adopted, and the normalization method of the three bands of near infrared, red and green is as above; for DSM data, firstly counting all values, finding out a minimum value and a value interval, subtracting the minimum value from each pixel value, and dividing the value number of the value interval; finally, all pixel values are normalized to be between 0 and 1. For example: by counting the values, the values are all found to be between 200 and 300, so that the minimum value 200 is subtracted from all the pixel values, and then the pixel values are divided by the value interval 100, thereby classifying all the pixel values to be between 0 and 1.
2.2, training the full-connection network model based on pixel brightness value classification by adopting the training set after preprocessing, wherein the specific process is as follows:
taking each pixel in each sample in the preprocessed training set as an input unit, and forming an input feature vector by the brightness values of each pixel after the normalization of four wave bands, namely near infrared, red, green and DSM; and updating the optimized network parameters by using a random gradient descent method until all samples in the preprocessed training set are input, and obtaining the base classifier corresponding to the fully-connected network.
The method comprises the following steps of testing a base classifier corresponding to the full-connection network by adopting a test set, and specifically: taking each original image in the test set as a test sample, forming an input feature vector by pixel brightness values of four wave bands of near infrared, red, green and DSM of each pixel of one test sample, and correspondingly outputting a corresponding class probability value;
in the method using pixels as the classification unit in this embodiment, because of intra-class heterogeneity, many isolated pixels may appear as a result, so that the classification result is relatively broken, and therefore, the classification result is smoothed by using median filtering, and the smoothing window size is 9 × 9.
Illustratively, the training data is the feature row vector of the first 10 images of the 15 images, and the test data is the feature row vector of the last 5 images.
Model training and testing: in the training phase, the Batch _ size is set to 256, the epoch is 50, the gradient descent is performed using the stochastic gradient descent method, and the learning rate is 10-4Learning rate decay is set to 10-6Finally, training was performed on 10 images and testing was performed on another image. The results are shown in FIG. 6 and Table 1;
the network classification precision evaluation of the invention mainly compares the classification result with the reference label graph to quantitatively measure the accuracy of the classification result. The commonly used evaluation indexes mainly comprise overall precision, accuracy, F1 score and the like; the method takes the F1 score and the accuracy as indexes to evaluate the classification precision.
The F1 score is the harmonic mean of precision and recall, defined as:
wherein Precision represents the Precision rate,recall represents the Recall rate of the call,TP is true positive, i.e., the prediction is positive, and the label is also positive; FP was false positive, i.e. prediction was positive and signature was negative; FN is false negative, i.e. predicted negative, and the tag is positive.
ACC denotes the accuracy (accuracuracy), which is defined as:
wherein TP is true positive, i.e., the prediction is positive, and the label is also positive; FP was false positive, i.e. prediction was positive and signature was negative; TN is true negative, i.e. predicted negative, the label is also negative; FN is false negative, i.e. predicted negative, and the tag is positive.
TABLE 1 Pixel luminance based classification results for fully connected networks
As can be seen from fig. 6 and table 1, the fully-connected network model can accurately identify the outline of the building, but the chimney or skylight on the roof of the building can be mistakenly classified into the automobile category due to the large intra-category difference; because the number of the automobile types is basically balanced with large types such as buildings, impervious surfaces and the like after the automobile types are expanded, and because of the diversity of the automobile colors, a plurality of impervious surfaces are wrongly classified into automobiles (the accuracy of the automobiles in the corresponding table 1 is low); the confusion between the low vegetation and the tree is serious, because the spectral characteristics of the low vegetation and the tree are similar, DSM assistance is required to distinguish the trees from each other, but because of the difference of the land surface height in a large range, the DSM value of the tree in the overall range is not always larger than that of the low vegetation; the three types of buildings, impervious surfaces and automobiles are basically not mixed with the two types of low vegetation and trees.
2.2, training the convolutional neural network model based on object-oriented classification by adopting the preprocessed training set, wherein the specific process is as follows:
2.2.1, performing multi-scale segmentation on the label graph of the preprocessed sample in the training set to correspondingly obtain a plurality of label graph segmentation blocks; wherein, the gray value in each partition block is the same; according to the embodiment of the invention, eCoginization software is adopted to carry out multi-scale segmentation on the tag graph, the segmentation result is derived into a tif grid graph, the segmentation scale parameter is 20, the shape index is 0.4, and the compactness is 0.5.
2.2.2, extracting the gravity center of each label graph segmentation block; wherein, the abscissa of the gravity center is the sum of the abscissas of all pixels in the segmentation block, and then the sum is divided by the number of the pixels in the segmentation block, and the result is rounded downwards; the vertical coordinate of the gravity center is the sum of the vertical coordinates of all pixels in the segmentation block, and then the sum is divided by the number of the pixels in the segmentation block, and the result is rounded downwards. The results of extracting the center of gravity of the different-sized blocks are shown in fig. 5, in which each hatched portion is a block and black squares are the extracted center of gravity. As can be seen from the figure, the barycentric coordinates of the different sized segments are different.
2.2.3, taking an image block with the size of a × b by taking the gravity center of each label map partition block as the center to obtain a plurality of image blocks of a × b × k, namely extracting a plurality of rolling blocks;
where k is the number of channels in the convolutional layer.
2.2.4, training a convolutional neural network by taking each image block as input data, and performing gradient descent by using a RMSProp (root mean square transfer) optimization method to optimize parameters of the convolutional neural network.
The method comprises the following steps of testing a base classifier corresponding to the convolutional neural network by adopting a test set, and specifically comprises the following steps: carrying out multi-scale segmentation on the original test image in the test set to obtain a plurality of test image segmentation blocks; performing gravity center extraction and convolution block extraction on the test data to obtain input data of the network test; inputting the input data of the network test into the trained base classifier, and outputting the probability value of the corresponding class.
Illustratively, in the embodiment of the invention, training is carried out on a 10-image segmentation block, and testing is carried out on another image, the size of the volume block is 32 × 32 × 4, the training stage is that the Batch _ size is set to be 12, the epoch is 100, gradient descent is carried out by using a RMSProp optimizer method, and the learning rate is 10-3Learning rate decay is set to 10-5. The test results are shown in fig. 7 and table 2;
TABLE 2 classification results based on convolutional neural networks for object-oriented classification
As can be seen from fig. 7 and table 2, in the convolutional neural network based on object-oriented classification, the confusion of the skylight of the roof of the building and the car class is greatly reduced, because the spatial information is applied to the classification characteristics, but like the building in the lower left part of the middle of the image, the skylight of the roof is still mistakenly classified into the car class due to the large area of the skylight; in the method, because the number of the expanded automobile types is less compared with the large types, the condition that the building is wrongly divided into the automobiles is greatly reduced, and meanwhile, the divided automobiles are also reduced; confusion between low vegetation and trees is not improved in this method; but compared with the method of taking blocks by a sliding window, the method reduces the confusion condition of class boundaries and reduces the calculation redundancy by the method of object-oriented segmentation.
2.3, training the complete convolution network model based on remote sensing image classification by adopting the training set after preprocessing, wherein the specific process is as follows:
and 2.3.1, cutting the preprocessed images in the training set according to a certain size to obtain image blocks, wherein for example, the training images are completely cut into 512 × 512 image blocks, and different partition block sizes can be designed according to different video memory capacities.
Step 2.3.2, setting a corresponding number of binary classification models according to the classes of the label graphs corresponding to the samples in the preprocessed training set, wherein one label graph class corresponds to one binary classifier; converting each label graph into a corresponding binary label graph; and training a corresponding binary classifier for each binary label graph to obtain corresponding network parameters which respectively correspond to 5 categories, and training 5U-Net networks to obtain classification parameters for different categories.
The binary label graph only comprises one target class, and the rest classes are backgrounds; the binary classifier is a U-Net network;
the method comprises the following steps of testing a base classifier corresponding to the complete convolution network by adopting a test set, specifically: cutting an original test image in the test set to obtain a plurality of image blocks, namely input data of a base classifier corresponding to the complete convolution network; and inputting the input data into the base classifier corresponding to the trained complete convolution network, and outputting a corresponding class probability value.
Illustratively, the original label graph has 5 categories: the method comprises the steps of converting an original label graph into 5 binary label graphs which respectively only comprise one class and other classes become backgrounds, respectively corresponding to the 5 classes, training 5U-Net networks to obtain classification parameters aiming at different classes, wherein the 5 binary label graphs are used for training the 5U-Net networks. The network design is shown in FIG. 3; on the basis of an original network architecture, a regularization layer is added into a down sampling path to improve the generalization capability of the model. The test results are shown in fig. 8 and table 3.
Table 3 shows the classification results of the complete convolution network based on remote sensing image classification
As can be seen from fig. 8 and table 3, in the classification method based on the complete convolutional network, the classification accuracy of each category is improved except for the vehicle category, because a binary classification mode is adopted in the classification process, and a classifier is respectively constructed for each category, the classification method is more pertinent; however, for the automobile category, data expansion is not considered, and the number of the negative samples is much larger than that of the positive samples in the binary classification process, so that the automobile category has almost no correctly classified pixels in the classification result.
specifically, as shown in fig. 1, a fully-connected network is established at the output ends of three basis classifiers, and the class probability output by the three basis classifiers and discriminated for each pixel is used as a training feature, so that a fully-connected network is trained on a test image by using a prediction probability and a label graph.
And 4, classifying the samples in the test set by adopting a deep ensemble learning network, and outputting corresponding prediction class labels.
And (3) testing the test images in the test set by adopting the deep ensemble learning network obtained by training in the step (3) to obtain a prediction class label. Therefore, the advantages of each method are integrated to obtain the final classification result. The integrated prediction results of the present invention are shown in fig. 9 and table 4.
Table 4 shows the classification results of the remote sensing images based on the deep ensemble learning network
As can be seen from fig. 9 and table 4, for both low vegetation and high vegetation, the base class classifier before the extraction has a good extraction effect, and the classification accuracy after the integration is not greatly improved. However, the method integrates the classification advantages of two methods based on pixels and object-oriented methods for vehicles and the classification advantages of a complete convolution network for buildings, has no obvious short board on the classification precision of each small category, and improves the overall classification precision to a certain extent. The problem of confusion between building rooftops and vehicles, which arises in both pixel-based and object-oriented approaches, is substantially eliminated, while the problem of confusion between portions of the building and the impervious surface is ameliorated.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present invention without departing from the spirit and scope of the invention; thus, if such modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.
Claims (9)
1. A high-resolution remote sensing image classification method based on deep ensemble learning is characterized by comprising the following steps:
step 1, acquiring a training set and a test set; respectively constructing a full-connection network model based on pixel brightness value classification, a convolutional neural network model based on object-oriented classification and a full convolutional network model based on remote sensing image classification;
the samples in the training set and the test set are high-resolution remote sensing images and DSM data thereof respectively, and the DSM is a digital surface model;
step 2, training a full-connection network model based on pixel brightness value classification, a convolutional neural network model based on object-oriented classification and a full convolutional network model based on remote sensing image classification by adopting a training set respectively to obtain corresponding base classifiers correspondingly; respectively testing the three base classifiers by adopting a test set, and correspondingly outputting corresponding class probability values;
step 3, setting a fully-connected network at the output ends of the three base classifiers obtained in the step 2, taking the class probability values output by the three base classifiers as input feature data, taking the label graph as a target class, and training the fully-connected network to obtain a deep integrated learning network;
and 4, classifying the samples in the test set by adopting a deep ensemble learning network, and outputting corresponding prediction class labels.
2. The deep ensemble learning-based high-resolution remote sensing image classification method according to claim 1, wherein the fully-connected network model based on pixel brightness value classification is composed of N fully-connected layers, wherein the first N-1 layers respectively adopt a ReLU function as an activation function, and the last layer adopts a Softmax function as an activation function; the Dropout layer is used for regularization before the last fully connected layer to prevent overfitting.
3. The method for classifying high-resolution remote sensing images based on deep ensemble learning according to claim 1, wherein the convolutional neural network model based on object-oriented classification comprises a plurality of convolutional layers, a maximum pooling layer, fully-connected layers and output layers, wherein one maximum pooling layer is arranged behind each convolutional layer, the plurality of fully-connected layers are connected behind the last pooling layer, and the output layer is arranged behind the last fully-connected layer; the activation function of each convolution layer and the activation function of the full connection layer are respectively a ReLU function, and the activation function of the output layer is a Softmax function; the dryout layer is used for regularization before the last fully connected layer.
4. The method for classifying high-resolution remote sensing images based on deep ensemble learning according to claim 1, wherein the complete convolution network model based on remote sensing image classification is composed of a plurality of binary classifiers, and each binary classifier comprises a convolution layer, a Dropout regularization layer, a pooling layer, and a convolution layer, a fusion layer and an upsampling layer which are symmetrical with the convolution layer, the Dropout regularization layer and the pooling layer.
5. The method for classifying high-resolution remote sensing images based on deep ensemble learning according to claim 1, wherein the step 2 comprises the following substeps:
2.1, preprocessing the training set to obtain a preprocessed training set;
wherein the preprocessing comprises unbalanced sample processing, data expansion and normalization;
2.2, training a full-connection network model based on pixel brightness value classification by adopting the preprocessed training set to obtain a base classifier corresponding to the full-connection network; testing the base classifier corresponding to the full-connection network by adopting a test set, and outputting corresponding class probability;
2.3, training the convolutional neural network model based on the object-oriented classification by adopting the preprocessed training set to obtain a base classifier corresponding to the convolutional neural network; testing a base classifier corresponding to the convolutional neural network by adopting a test set, and outputting a corresponding class probability value;
2.4, training a complete convolution network model based on remote sensing image classification by adopting the preprocessed training set to obtain a base classifier corresponding to the complete convolution network; and testing the base classifier corresponding to the complete convolution network by adopting the test set, and outputting a corresponding class probability value.
6. The high-resolution remote sensing image classification method based on deep ensemble learning according to claim 5, wherein the unbalanced sample processing specifically comprises: resampling data of the small class samples, downsampling the large class data or adopting different sampling proportions for different classes;
the data expansion is as follows: horizontally turning or randomly rotating the original images in the training set;
the normalization is as follows: for the color images of the near infrared band, the red band and the green band of the remote sensing satellite data, the gray value is scaled from 0-255 to 0-1;
for aviation images, the aviation images comprise color images of three wave bands of near infrared, red and green and DSM data of one wave band, wherein the color images of the three wave bands of the near infrared, the red and the green are scaled from a gray value of 0-255 to a gray value of 0-1; for DSM data, firstly counting all values, finding out a minimum value and a value interval, subtracting the minimum value from each pixel value, and dividing the value number of the value interval; finally, all pixel values are normalized to be between 0 and 1.
7. The method for classifying high-resolution remote sensing images based on deep ensemble learning according to claim 5, wherein the fully-connected network model based on pixel brightness value classification is trained by adopting a preprocessed training set in the specific process;
taking each pixel in each sample in the preprocessed training set as an input unit, and forming an input feature vector by the brightness values of each pixel after the normalization of four wave bands, namely near infrared, red, green and DSM; and updating the optimized network parameters by using a random gradient descent method until all samples in the preprocessed training set are input, and obtaining the base classifier corresponding to the fully-connected network.
8. The deep ensemble learning-based high-resolution remote sensing image classification method according to claim 5, wherein the training of the convolutional neural network model based on object-oriented classification is performed by using a preprocessed training set, and the specific process is as follows:
2.2.1, performing multi-scale segmentation on the label graph of the preprocessed sample in the training set to correspondingly obtain a plurality of label graph segmentation blocks; wherein, the gray value of the pixel in each partition block is the same;
2.2.2, extracting the gravity center of each label graph segmentation block; wherein, the abscissa of the gravity center is the sum of the abscissas of all pixels in the segmentation block, and then the sum is divided by the number of the pixels in the segmentation block, and the obtained result is rounded downwards; the vertical coordinate of the gravity center is the sum of the vertical coordinates of all pixels in the segmentation block, and then the sum is divided by the number of the pixels in the segmentation block, and the obtained result is rounded downwards;
2.2.3, taking an image block with the size of a × b by taking the gravity center of each label map partition block as the center to obtain a plurality of image blocks of a × b × k, namely extracting a plurality of rolling blocks;
wherein k is the number of channels of the convolutional layer;
2.2.4, training a convolutional neural network by taking each image block as input data, and performing gradient descent by using a root-mean-square transfer optimization method to optimize parameters of the convolutional neural network.
9. The deep ensemble learning-based high-resolution remote sensing image classification method according to claim 5, wherein the training of the complete convolution network model based on remote sensing image classification is performed by using a preprocessed training set, and the specific process is as follows:
step 2.3.1, cutting the preprocessed images in the training set according to a preset size to obtain image blocks;
step 2.3.2, setting a corresponding number of binary classification models according to the classes of the label graphs corresponding to the samples in the preprocessed training set, wherein one label graph class corresponds to one binary classifier; converting each label graph into a corresponding binary label graph;
the binary label graph only comprises one target class, and the rest classes are backgrounds; the binary classifier is a U-Net network.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010173481.1A CN111368776B (en) | 2020-03-13 | 2020-03-13 | High-resolution remote sensing image classification method based on deep ensemble learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010173481.1A CN111368776B (en) | 2020-03-13 | 2020-03-13 | High-resolution remote sensing image classification method based on deep ensemble learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111368776A true CN111368776A (en) | 2020-07-03 |
CN111368776B CN111368776B (en) | 2024-03-22 |
Family
ID=71211853
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010173481.1A Active CN111368776B (en) | 2020-03-13 | 2020-03-13 | High-resolution remote sensing image classification method based on deep ensemble learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111368776B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113449666A (en) * | 2021-07-07 | 2021-09-28 | 中南大学 | Remote sensing image multi-scale target detection method based on data fusion and feature selection |
CN113537509A (en) * | 2021-06-28 | 2021-10-22 | 南方科技大学 | Collaborative model training method and device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108764316A (en) * | 2018-05-18 | 2018-11-06 | 河海大学 | Remote sensing images scene classification method based on depth convolutional neural networks and Multiple Kernel Learning |
CN108830296A (en) * | 2018-05-18 | 2018-11-16 | 河海大学 | A kind of improved high score Remote Image Classification based on deep learning |
WO2018214195A1 (en) * | 2017-05-25 | 2018-11-29 | 中国矿业大学 | Remote sensing imaging bridge detection method based on convolutional neural network |
US20200026953A1 (en) * | 2018-07-23 | 2020-01-23 | Wuhan University | Method and system of extraction of impervious surface of remote sensing image |
-
2020
- 2020-03-13 CN CN202010173481.1A patent/CN111368776B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018214195A1 (en) * | 2017-05-25 | 2018-11-29 | 中国矿业大学 | Remote sensing imaging bridge detection method based on convolutional neural network |
CN108764316A (en) * | 2018-05-18 | 2018-11-06 | 河海大学 | Remote sensing images scene classification method based on depth convolutional neural networks and Multiple Kernel Learning |
CN108830296A (en) * | 2018-05-18 | 2018-11-16 | 河海大学 | A kind of improved high score Remote Image Classification based on deep learning |
US20200026953A1 (en) * | 2018-07-23 | 2020-01-23 | Wuhan University | Method and system of extraction of impervious surface of remote sensing image |
Non-Patent Citations (2)
Title |
---|
周敏;史振威;丁火平;: "遥感图像飞机目标分类的卷积神经网络方法", 中国图象图形学报, no. 05 * |
王鑫;李可;宁晨;黄凤辰;: "基于深度卷积神经网络和多核学习的遥感图像分类方法", 电子与信息学报, no. 05 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113537509A (en) * | 2021-06-28 | 2021-10-22 | 南方科技大学 | Collaborative model training method and device |
CN113449666A (en) * | 2021-07-07 | 2021-09-28 | 中南大学 | Remote sensing image multi-scale target detection method based on data fusion and feature selection |
Also Published As
Publication number | Publication date |
---|---|
CN111368776B (en) | 2024-03-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108961235B (en) | Defective insulator identification method based on YOLOv3 network and particle filter algorithm | |
CN109684922B (en) | Multi-model finished dish identification method based on convolutional neural network | |
CN105809121A (en) | Multi-characteristic synergic traffic sign detection and identification method | |
CN107358260B (en) | Multispectral image classification method based on surface wave CNN | |
CN113421269A (en) | Real-time semantic segmentation method based on double-branch deep convolutional neural network | |
CN108573276A (en) | A kind of change detecting method based on high-resolution remote sensing image | |
CN110363215A (en) | The method that SAR image based on production confrontation network is converted into optical imagery | |
CN110084108A (en) | Pedestrian re-identification system and method based on GAN neural network | |
CN109344891A (en) | A kind of high-spectrum remote sensing data classification method based on deep neural network | |
CN106503739A (en) | The target in hyperspectral remotely sensed image svm classifier method and system of combined spectral and textural characteristics | |
CN110728197B (en) | Single-tree-level tree species identification method based on deep learning | |
CN112131946B (en) | Automatic extraction method for vegetation and water information of optical remote sensing image | |
CN110287798B (en) | Vector network pedestrian detection method based on feature modularization and context fusion | |
CN109801305B (en) | SAR image change detection method based on deep capsule network | |
CN111898621A (en) | Outline shape recognition method | |
CN110516648B (en) | Ramie plant number identification method based on unmanned aerial vehicle remote sensing and pattern identification | |
CN113673556B (en) | Hyperspectral image classification method based on multiscale dense convolution network | |
CN111368776B (en) | High-resolution remote sensing image classification method based on deep ensemble learning | |
CN112115795B (en) | Hyperspectral image classification method based on Triple GAN | |
CN109165698A (en) | A kind of image classification recognition methods and its storage medium towards wisdom traffic | |
CN112733736A (en) | Class imbalance hyperspectral image classification method based on enhanced oversampling | |
CN111666813A (en) | Subcutaneous sweat gland extraction method based on three-dimensional convolutional neural network of non-local information | |
CN112883852B (en) | Hyperspectral image classification system and method | |
CN117475327A (en) | Multi-target detection positioning method and system based on remote sensing image in city | |
Li et al. | Incremental learning of infrared vehicle detection method based on SSD |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |