CN110211138B - Remote sensing image segmentation method based on confidence points - Google Patents
Remote sensing image segmentation method based on confidence points Download PDFInfo
- Publication number
- CN110211138B CN110211138B CN201910494015.0A CN201910494015A CN110211138B CN 110211138 B CN110211138 B CN 110211138B CN 201910494015 A CN201910494015 A CN 201910494015A CN 110211138 B CN110211138 B CN 110211138B
- Authority
- CN
- China
- Prior art keywords
- image
- green
- pixel
- red
- test set
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10032—Satellite or aerial image; Remote sensing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30181—Earth observation
Abstract
The invention discloses a remote sensing image segmentation method based on confidence points, which mainly solves the defect of low segmentation precision of high-resolution remote sensing images in the prior art. The method comprises the following specific steps: (1) constructing a convolutional neural network; (2) generating two training sets; (3) generating two test sets; (4) predicting the ground feature class labels of the test set; (5) labeling each confidence point pixel; (6) correcting the building category; (7) updating the plant category; and (8) obtaining a final ground object type label. The method has the advantages of improving the precision of the vegetation and the building category, having better segmentation effect on the low-resolution remote sensing image and having better segmentation effect on the high-resolution remote sensing image.
Description
Technical Field
The invention belongs to the technical field of image processing, and further relates to a remote sensing image segmentation method based on confidence points in the technical field of remote sensing image processing. The method can be used for segmenting the high-resolution multiband remote sensing image acquired by the satellite to obtain the segmentation image with the ground object class label.
Background
Remote sensing image segmentation is a technique and process for dividing a remote sensing image into several specific regions with unique properties. At present, deep neural network learning technologies such as U-Net, PSPNet, deep Lab and the like are more adopted in engineering practice. The method for segmenting the remote sensing image based on the deep learning neural network is characterized in that the neural network is used for extracting the characteristics of the remote sensing image, and the category of each pixel is predicted through a training network, so that the segmentation image with the category label is finally obtained. The method in the prior art has a good segmentation effect on a public data set with low resolution, but the effect is not ideal for a remote sensing image with high resolution, such as a WorldView-3 satellite remote sensing image.
The patent document of Changan university in its application "a region-based multi-feature fusion high-resolution remote sensing image segmentation method" (patent application No. 201610643629.7, publication No. 106296680A) discloses a region-based multi-feature fusion high-resolution remote sensing image segmentation method. The method comprises the steps of firstly segmenting an initial remote sensing image, secondly calculating the textural feature distance, the spectral feature distance and the shape feature distance of any neighborhood in each segmented region in the initial segmented image, and finally applying a RAG method and an NNG method to carry out relevant region combination. Although the method has high segmentation precision and high execution efficiency on the high-resolution second-order remote sensing image with the resolution of 1 meter, the method still has the defects that the remote sensing image is segmented by adopting the spectral characteristic distance, the textural characteristic distance and the shape characteristic distance, and information of other wave bands of the remote sensing image is not utilized, so that the vegetation type segmentation precision is not high. In addition, due to the existence of building shadows and occlusion problems, the segmentation precision of specific building categories is not high.
The patent document of Haihe university discloses a multiband high-resolution remote sensing image segmentation method based on gray level co-occurrence matrix (patent application number: 20131056019.2, publication number: 103578110A). The method comprises the steps of independently segmenting each wave band image by adopting precipitation watershed transformation, and then overlapping segmentation results of each wave band. And finally, combining fragment areas in the segmentation result by using an area combination strategy based on multiband spectral information, thereby realizing image segmentation. Although the method can overcome the phenomena of over-segmentation and under-segmentation, and has good segmentation precision and stability, the method still has the defects that the method is not high in universality of the remote sensing image with high resolution because the method is suitable for the three-band panchromatic-multispectral fusion remote sensing image of the Shanghai region in China with the resolution of 2.5 meters, and in addition, the method adopts a gray level co-occurrence matrix method to segment the remote sensing image, the problems of existence and shielding of building shadows are not considered, and the segmentation precision of specific building categories is not high. .
Disclosure of Invention
The invention aims to provide a remote sensing image segmentation method based on confidence points aiming at the defects of the prior art. The method has high segmentation efficiency, and can realize a good segmentation result for the remote sensing image with high resolution, particularly vegetation and building categories.
The idea for achieving the purpose of the invention is to construct two training sets and two testing sets, construct a LinkNet network, set parameters of each layer, obtain two trained networks after training, send the two testing sets into the two trained networks respectively, predict ground object type labels of the testing sets through the networks, and finally obtain the ground object type labels after remote sensing image segmentation through marking confidence point pixels, updating vegetation types and correcting building types.
In order to achieve the purpose, the method comprises the following specific steps:
(1) Constructing a LinkNet neural network:
(1a) Building a 14-layer LinkNet neural network;
(1b) Setting parameters of each layer in a LinkNet neural network;
(2) Two training sets were generated:
(2a) Selecting at least 20000 images from a remote sensing image data set to form a basic training set, wherein 10000 red, green and blue RGB images are selected, and the rest 10000 images are multispectral images; the ground feature type label of each pixel of each image in the basic training set is one of five ground feature type labels of a road, a viaduct, a building, vegetation and the ground, each red, green and blue RGB image corresponds to one satellite shooting area, each multispectral image comprises eight-waveband information, and the satellite shooting area of each multispectral image is the same as the satellite shooting area of each red, green and blue RGB image;
(2b) Synthesizing all red, green and blue (RGB) images in the basic training set into an enhanced red, green and blue (RGB) image by using a synthesis method, wherein the enhanced red, green and blue (RGB) image is used as a first training set;
(2c) Synthesizing all multispectral images in the basic training set into a multiband pseudo-color image by using a synthesis method to serve as a second training set;
(3) Training a neural network:
(3a) Inputting the first training set into a LinkNet network for training to obtain a trained first segmentation network;
(3b) Inputting the second training set into a LinkNet network for training to obtain a trained second segmentation network;
(4) Two test sets were generated:
(4a) Selecting at least 6000 images from the remote sensing image data set to form a basic test set, wherein 3000 red, green and blue (RGB) images are selected, and the rest 3000 images are multispectral images; the ground feature type label of each pixel of each image in the basic test set is one of five ground feature type labels of a road, an overpass, a building, vegetation and the ground, each red, green and blue RGB image corresponds to one satellite shooting area, each multispectral image comprises eight-waveband information, and the satellite shooting area of each multispectral image is the same as the satellite shooting area of each red, green and blue RGB image;
(4b) Synthesizing the red, green and blue RGB images in the basic test set into an enhanced red, green and blue RGB image by using a synthesis method to serve as a first test set;
(4c) Synthesizing the multispectral images in the basic test set into a multiband pseudo-color image serving as a second test set by using a synthesis method;
(5) And (3) performing ground object class label prediction on the test set:
(5a) Sequentially inputting each image of the first test set into a first segmentation network, and outputting a ground object class label of each pixel in each image of the first test set and a ground object class label probability value of each pixel in each image;
(5b) Sequentially inputting each image of the second test set into a second segmentation network, and outputting the ground object class label of each pixel in each image in the second test set and the ground object class label probability value of each pixel in each image;
(6) Labeling each confidence point pixel:
traversing each pixel in each image in the first test set; marking all pixels with the surface feature label probability value of the pixel being more than 0.9 as confidence point pixels;
(7) Correcting the building type:
traversing all the ground feature type labels in the first test set to be each pixel of the building, changing the ground feature type labels of all the pixels with confidence point pixels in the range of 5 multiplied by 5 pixel points adjacent to each pixel into the building, and changing the ground feature type labels of the rest pixels into the ground; obtaining the ground object class label of each pixel in each image in the updated first test set
(8) Updating the plant category:
(8a) Multiplying the average value of pixel values of coastline waveband Coastal corresponding to all the pixels with the vegetation in the second test set by 0.8 to obtain a threshold value of coastline waveband Coastal, multiplying the average value of pixel values of Yellow waveband Yellow corresponding to all the pixels with the vegetation in the second test set by 0.8 to obtain a threshold value of Yellow waveband Yellow, and multiplying the average value of pixel values of near infrared secondary waveband NIR2 corresponding to all the pixels with the vegetation in the second test set by 0.2 to obtain a threshold value of near infrared secondary waveband NIR 2;
(8b) Traversing each pixel in each image in the second test set, updating the ground object class labels of all pixels of which the probability value of the ground object class label is less than or equal to 0.5, the pixel value corresponding to the coastline wave band Coastal is less than the threshold value of the coastline wave band Coastal, the pixel value of the Yellow wave band Yellow is less than the threshold value of the Yellow wave band Yellow, and the pixel value of the near infrared secondary wave band NIR2 is greater than the threshold value of the near infrared secondary wave band NIR2 into vegetation, and keeping the original ground object class labels of the rest pixels unchanged; obtaining a ground object type label of each pixel in each image in the updated second test set;
(9) Obtaining a final ground object class label:
(9a) Replacing the vegetation type label of each pixel in each image in the updated first test set by using the vegetation type label of each pixel in each image in the updated second test set to obtain the surface feature type label of each pixel in each image in the first test set after vegetation type replacement;
(9b) And the feature class label of each pixel in each image in the first test set after vegetation class replacement is used as the feature class label after the remote sensing image is segmented.
Compared with the prior art, the invention has the following advantages:
firstly, the ground object type labels of the vegetation output by the segmentation network are updated by setting three thresholds of coastline wave bands coast, yellow wave bands Yellow and near infrared two-wave band NIR2, so that the defect of low vegetation type segmentation precision caused by the fact that a remote sensing image is segmented by adopting a spectral characteristic distance, a texture characteristic distance and a shape characteristic distance in the prior art is overcome, even scattered vegetation can be segmented, and the segmentation precision of the vegetation type is improved.
Secondly, because the invention inputs the enhanced red, green and blue RGB image and the multiband pseudo-color image to the LinkNet neural network for training, the invention overcomes the defects that the prior art has better segmentation effect on the public data set with lower resolution and has low universality on the remote sensing image with high resolution, so that the invention not only has better segmentation effect on the remote sensing image with low resolution, but also has better segmentation effect on the remote sensing image with high resolution.
Thirdly, the invention utilizes the confidence point pixel to correct the ground feature type label of the building output by the segmentation network, thereby overcoming the defect of low segmentation precision of the concrete building type caused by the existence of the shadow and the shielding problem of the building in the prior art and improving the segmentation precision of the building type.
Drawings
FIG. 1 is an overall flow chart of the present invention.
The specific implementation mode is as follows:
step 1, constructing a neural network.
Building a 14-layer LinkNet neural network;
the structure of the LinkNet neural network with 14 layers is as follows in sequence: input layer → 1 st convolution layer → 1 st largest pooling layer → 1 st encoder → 2 nd encoder → 3 rd encoder → 4 th decoder → 3 rd decoder → 2 nd decoder → 1 st full convolution layer → 2 nd full convolution layer; wherein, the output of the 1 st coder is connected with the input of the 1 st decoder, the output of the 2 nd coder is connected with the input of the 2 nd decoder, and the output of the 3 rd coder is connected with the input of the 3 rd decoder.
The structure of the encoder is as follows in sequence: input layer → 1 st encoding convolution layer → 2 nd encoding convolution layer → 3 rd encoding convolution layer → 4 th encoding convolution layer; wherein, the input layer is connected with the output of the 2 nd coding convolution layer, and the input of the 3 rd coding convolution layer is connected with the output of the 4 th coding convolution layer; the decoder structure is as follows: input layer → 1 st decoded convolutional layer → 1 st decoded full convolutional layer → 2 nd decoded convolutional layer.
Setting parameters of each layer in a LinkNet neural network;
the parameters of each layer in the linkNet neural network are set as follows: setting the convolution kernel size of the 1 st convolution layer to 7 × 7 and the step size to 2; setting the size of the pooling window of the 1 st largest pooling layer to 3 x 3, and setting the step size to 2; setting the convolution kernel sizes of the 1 st and 2 nd full convolution layers as 3 x 3 and 2 x 2 in sequence, and setting the step sizes as 2; the convolution kernel size for the 2 nd convolution layer is set to 3 x 3 with the step size set to 1.
Wherein the parameter settings of the encoder are as follows: setting the sizes of convolution kernels of the 1 st, 2 nd, 3 th and 4 th coding convolution layers as 3 x 3, and sequentially setting the step sizes as 2, 1 and 1; the parameters of the decoder are as follows: setting the sizes of convolution kernels of the 1 st decoding convolution layer and the 2 nd decoding convolution layer as 1 x 1, and setting the step size as 1; the convolution kernel size of the 1 st decoded full convolution layer is set to 3 x 3 and the step size is set to 2.
And 2, generating two training sets.
Selecting at least 20000 images from a remote sensing image data set to form a basic training set, wherein 10000 red, green and blue RGB images are selected, and the rest 10000 images are multispectral images; the ground feature type label of each pixel of each image in the basic training set is one of five ground feature type labels of a road, a viaduct, a building, vegetation and the ground, each red, green and blue RGB image corresponds to one satellite shooting area, each multispectral image comprises eight-waveband information, and the satellite shooting area of each multispectral image is the same as the satellite shooting area of each red, green and blue RGB image;
the eight-band information means: coastline wave band Coastal, blue wave band Blue, green wave band Green, yellow wave band Yellow, red wave band Red edge, near infrared first wave band NIR1 and near infrared second wave band NIR2.
Synthesizing all red, green and blue RGB images in the basic training set into an enhanced red, green and blue RGB image by using a synthesis method to serve as a first training set;
the enhanced red, green and blue RGB image combines the original red, green and blue RGB image and the information of the near infrared two NIR2 wave bands, enhances the information of the remote sensing image such as contrast, detail texture and the like, and has good detection effect on the ground, buildings, elevated roads and other categories.
And synthesizing all multispectral images in the basic training set into a multiband pseudo-color image as a second training set by using a synthesis method.
The multi-band pseudo-color image combines a coastline Coastal waveband, a Yellow waveband and a near infrared two NIR2 waveband in the multi-spectral image, and highlights the category information of vegetation.
The synthesis method comprises the following steps:
step 1, synthesizing a green channel in each red, green and blue RGB image in a basic training set or a basic test set and corresponding near infrared two NIR2 waveband information into a green channel of an enhanced red, green and blue RGB image of the red, green and blue RGB image according to the following formula:
N=G*w+R*(1-w)
wherein, N represents the green channel of the enhanced RGB image, G represents the green channel of the RGB image of the basic training set or the basic testing set, x represents the multiplication operation, w represents the weighting of the green channel of the enhanced RGB image and the green channel of the RGB image of the basic training set or the basic testing set, and is used for controlling the degree of the enhanced RGB image, the weighting value is 0.8, R represents the near-infrared two 2 wave bands in the NIR image corresponding to the RGB image of the basic training set or the basic testing set.
And 2, replacing the green channel of each enhanced red, green and blue RGB image in the corresponding red, green and blue RGB image in the basic training set or the basic test set by the green channel of each enhanced red, green and blue RGB image to obtain the enhanced red, green and blue RGB image after the red, green and blue RGB image is updated.
And 3, forming all the enhanced red, green and blue RGB images into a first training set or a first testing set.
And 4, respectively normalizing the values of three wave bands, namely coastline wave band coast, yellow wave band Yellow and near infrared two wave band NIR2, in each multispectral image, respectively multiplying the three normalized values by 255, respectively replacing a red channel R, a green channel G and a blue channel B of a red, green and blue RGB image in a basic training set or a basic test set corresponding to the multispectral image by the product of the three normalized values, and forming the multiband pseudo-color image by the three updated channels of the red channel R, the green channel G and the blue channel B.
Step 5, forming all multiband pseudo-color images into a second training set or a second testing set;
and 3, training a neural network.
Inputting the first training set into a LinkNet network for training to obtain a trained first segmentation network;
and inputting the second training set into a LinkNet network for training to obtain a trained second segmentation network.
LinkNet is a network structure proposed in 2018 NIPS paper LinkNet, which focuses mainly on the speed problem of Semantic Segmentation and obtains a better Segmentation effect.
And 4, generating two test sets.
Selecting at least 6000 images from the remote sensing image data set to form a basic test set, wherein 3000 red, green and blue (RGB) images are selected, and the rest 3000 images are multispectral images; the ground feature type label of each pixel of each image in the basic test set is one of five ground feature type labels of a road, an overpass, a building, vegetation and the ground, each red, green and blue RGB image corresponds to one satellite shooting area, each multispectral image comprises eight-waveband information, and the satellite shooting area of each multispectral image is the same as the satellite shooting area of each red, green and blue RGB image;
synthesizing the red, green and blue RGB images in the basic test set into an enhanced red, green and blue RGB image by using a synthesis method to serve as a first test set;
synthesizing the multispectral image in the basic test set into a multiband pseudo-color image as a second test set by using a synthesis method;
the synthesis was as described in step 2.
And 5, predicting the ground feature class labels of the test set.
Sequentially inputting each image of the first test set into a first segmentation network, and outputting a ground object class label of each pixel in each image in the first test set and a ground object class label probability value of each pixel in each image;
sequentially inputting each image of the second test set into a second segmentation network, and outputting the ground object class label of each pixel in each image in the second test set and the ground object class label probability value of each pixel in each image;
the feature type label probability value is that each image is input to a segmentation network, the segmentation network outputs five output values of each image, wherein each pixel of each image corresponds to a road, an overhead bridge, a building, vegetation and the ground, normalization operation is carried out on the five output values, the maximum value of the five normalized output values is taken, and the maximum value is used as the feature type label probability value of the pixel.
And 6, marking each confidence point pixel.
Traversing each pixel in each image in the first test set; all pixels in which the surface feature class label probability value of the pixel is greater than 0.9 are marked as confidence point pixels.
And 7, correcting the building type.
Traversing all the ground feature type labels in the first test set to be each pixel of the building, changing the ground feature type labels of all the pixels with confidence point pixels in the range of 5 multiplied by 5 pixel points adjacent to each pixel into the building, and changing the ground feature type labels of the rest pixels into the ground; obtaining a ground object type label of each pixel in each image in the updated first test set;
because the sun altitude and the shooting angle are different during satellite shooting, shadow and shading problems can occur around the building, the segmentation results of the difficult areas are not accurate enough, and therefore, the segmentation results output by the segmentation network can form scattered segmentation results in the areas, and the category of the building is corrected through the confidence point pixels.
And 8, updating the plant category.
Multiplying the average value of pixel values of all the pixels with the vegetation type in the second test set, which correspond to a coastline waveband Coastal, by 0.8 to obtain a coastline waveband Coastal threshold value, multiplying the average value of pixel values of all the pixels with the vegetation type in the second test set, which correspond to a Yellow waveband Yellow by 0.8 to obtain a Yellow waveband Yellow threshold value, and multiplying the average value of pixel values of all the pixels with the vegetation type in the second test set, which correspond to a near infrared second waveband NIR2 by 0.2 to obtain a near infrared second waveband NIR2 threshold value;
traversing each pixel in each image in the second test set, updating the ground object class labels of all pixels of which the probability value of the ground object class label is less than or equal to 0.5, the pixel value corresponding to the coastline wave band Coastal is less than the threshold value of the coastline wave band Coastal, the pixel value of the Yellow wave band Yellow is less than the threshold value of the Yellow wave band Yellow, and the pixel value of the near infrared secondary wave band NIR2 is greater than the threshold value of the near infrared secondary wave band NIR2 into vegetation, and keeping the original ground object class labels of the rest pixels unchanged; obtaining a ground object type label of each pixel in each image in the updated second test set;
since the division of the vegetation region by using only three band thresholds may be affected by noise, such as objects with similar colors to vegetation, etc., the introduction of the ground object class label probability value avoids this problem.
And 9, obtaining a final ground object type label.
Replacing the vegetation type label of each pixel in each image in the updated first test set by using the vegetation type label of each pixel in each image in the updated second test set to obtain a ground feature type label of each pixel in each image in the first test set after vegetation type replacement;
and the ground feature class label of each pixel in each image in the first test set after vegetation class replacement is used as the ground feature class label after remote sensing image segmentation.
Because the vegetation class segmentation precision of each pixel in each image in the second test set is higher, and the segmentation effect of each pixel in each image in the first test set on other classes, such as the ground, buildings, viaducts and the like is better, the two segmentation results are fused to obtain the final accurate ground object class label after the remote sensing image is segmented.
Claims (6)
1. A remote sensing image segmentation method based on confidence points is characterized in that a neural network is used for predicting ground object class labels, the class labels obtained according to a multi-view and double-data model are fused, and the confidence points are used for correcting building classes in a segmentation map, and the method comprises the following steps:
(1) Constructing a LinkNet neural network:
(1a) Building a 14-layer LinkNet neural network;
(1b) Setting parameters of each layer in a LinkNet neural network;
(2) Two training sets were generated:
(2a) Selecting at least 20000 images from a remote sensing image data set to form a basic training set, wherein 10000 red, green and blue RGB images are selected, and the rest 10000 images are multispectral images; the ground feature type label of each pixel of each image in the basic training set is one of five ground feature type labels of a road, a viaduct, a building, vegetation and the ground, each red, green and blue RGB image corresponds to one satellite shooting area, each multispectral image comprises eight-waveband information, and the satellite shooting area of each multispectral image is the same as the satellite shooting area of each red, green and blue RGB image;
(2b) Synthesizing all red, green and blue (RGB) images in the basic training set into an enhanced red, green and blue (RGB) image by using a synthesis method, wherein the enhanced red, green and blue (RGB) image is used as a first training set;
(2c) Synthesizing all multispectral images in the basic training set into a multiband pseudo-color image by using a synthesis method to serve as a second training set;
(3) Training a neural network:
(3a) Inputting the first training set into a LinkNet network for training to obtain a trained first segmentation network;
(3b) Inputting the second training set into a LinkNet network for training to obtain a trained second segmentation network;
(4) Two test sets were generated:
(4a) Selecting at least 6000 images from the remote sensing image data set to form a basic test set, wherein 3000 red, green and blue (RGB) images are selected, and the rest 3000 images are multispectral images; the ground feature type label of each pixel of each image in the basic test set is one of five ground feature type labels of a road, an overpass, a building, vegetation and the ground, each red, green and blue RGB image corresponds to one satellite shooting area, each multispectral image comprises eight-waveband information, and the satellite shooting area of each multispectral image is the same as the satellite shooting area of each red, green and blue RGB image;
(4b) Synthesizing the red, green and blue RGB images in the basic test set into an enhanced red, green and blue RGB image by using a synthesis method to serve as a first test set;
(4c) Synthesizing the multispectral images in the basic test set into a multiband pseudo-color image serving as a second test set by using a synthesis method;
(5) And (3) predicting the ground object class labels of the test set:
(5a) Sequentially inputting each image of the first test set into a first segmentation network, and outputting a ground object class label of each pixel in each image of the first test set and a ground object class label probability value of each pixel in each image;
(5b) Sequentially inputting each image of the second test set into a second segmentation network, and outputting the ground object class label of each pixel in each image in the second test set and the ground object class label probability value of each pixel in each image;
(6) Labeling each confidence point pixel:
traversing each pixel in each image in the first test set; marking all pixels with the surface feature label probability value of the pixel being more than 0.9 as confidence point pixels;
(7) Modified building type:
traversing all the ground feature type labels in the first test set to be each pixel of the building, changing the ground feature type labels of all the pixels with confidence point pixels in the range of 5 multiplied by 5 pixel points adjacent to each pixel into the building, and changing the ground feature type labels of the rest pixels into the ground; obtaining a ground object type label of each pixel in each image in the updated first test set;
(8) Updating the plant category:
(8a) Multiplying the average value of pixel values of coastline waveband Coastal corresponding to all the pixels with the vegetation in the second test set by 0.8 to obtain a threshold value of coastline waveband Coastal, multiplying the average value of pixel values of Yellow waveband Yellow corresponding to all the pixels with the vegetation in the second test set by 0.8 to obtain a threshold value of Yellow waveband Yellow, and multiplying the average value of pixel values of near infrared secondary waveband NIR2 corresponding to all the pixels with the vegetation in the second test set by 0.2 to obtain a threshold value of near infrared secondary waveband NIR 2;
(8b) Traversing each pixel in each image in the second test set, updating the ground object class labels of all pixels of which the probability value of the ground object class label is less than or equal to 0.5, the pixel value corresponding to the coastline wave band Coastal is less than the threshold value of the coastline wave band Coastal, the pixel value of the Yellow wave band Yellow is less than the threshold value of the Yellow wave band Yellow, and the pixel value of the near infrared secondary wave band NIR2 is greater than the threshold value of the near infrared secondary wave band NIR2 into vegetation, and keeping the original ground object class labels of the rest pixels unchanged; obtaining a ground object type label of each pixel in each image in the updated second test set;
(9) Obtaining a final ground object type label:
(9a) Replacing the vegetation type label of each pixel in each image in the updated first test set by using the vegetation type label of each pixel in each image in the updated second test set to obtain a ground feature type label of each pixel in each image in the first test set after vegetation type replacement;
(9b) And the ground feature class label of each pixel in each image in the first test set after vegetation class replacement is used as the ground feature class label after remote sensing image segmentation.
2. The remote sensing image segmentation method based on the confidence points as claimed in claim 1, wherein the structure of the LinkNet neural network with 14 layers in step (1 a) is as follows in sequence: input layer → 1 st convolutional layer → 1 st maximally pooling layer → 1 st encoder → 2 nd encoder → 3 rd encoder → 4 th decoder → 3 rd decoder → 2 nd decoder → 1 st full convolutional layer → 2 nd full convolutional layer; wherein, the output of the 1 st coder is connected with the input of the 1 st decoder, the output of the 2 nd coder is connected with the input of the 2 nd decoder, and the output of the 3 rd coder is connected with the input of the 3 rd decoder;
the structure of the encoder is as follows in sequence: input layer → 1 st coding convolution layer → 2 nd coding convolution layer → 3 rd coding convolution layer → 4 th coding convolution layer; wherein, the input layer is connected with the output of the 2 nd coding convolution layer, and the input of the 3 rd coding convolution layer is connected with the output of the 4 th coding convolution layer; the decoder structure is: input layer → 1 st decoded convolutional layer → 1 st decoded full convolutional layer → 2 nd decoded convolutional layer.
3. The remote sensing image segmentation method based on the confidence point as claimed in claim 1, wherein the setting of the parameters of each layer in the LinkNet neural network in the step (1 b) is as follows: setting the convolution kernel size of the 1 st convolution layer to 7 × 7 and setting the step size to 2; setting the size of the pooling window of the 1 st largest pooling layer to 3 x 3, and setting the step size to 2; the convolution kernels of the 1 st and 2 nd full convolution layers are sequentially set to be 3 x 3 and 2 x 2, and the step length is set to be 2; setting the convolution kernel size of the 2 nd convolution layer as 3 x 3 and setting the step size as 1;
wherein the parameter settings of the encoder are as follows: setting the sizes of convolution kernels of the 1 st, 2 nd, 3 rd and 4 th coding convolution layers as 3 x 3, and sequentially setting the step sizes as 2, 1 and 1; the parameters of the decoder are as follows: setting the sizes of convolution kernels of the 1 st decoding convolution layer and the 2 nd decoding convolution layer as 1 x 1, and setting the step size as 1; the convolution kernel size of the 1 st decoded full convolution layer is set to 3 x 3 and the step size is set to 2.
4. The remote sensing image segmentation method based on the confidence point as claimed in claim 1, wherein the eight-band information in the steps (2 a) and (4 a) is: coastline wave band Coastal, blue wave band Blue, green wave band Green, yellow wave band Yellow, red wave band Red, near infrared first wave band NIR1, near infrared second wave band NIR2.
5. The remote sensing image segmentation method based on the confidence point as set forth in claim 1, wherein the synthesis method in the steps (2 b), (2 c), (4 b) and (4 c) comprises the following steps:
firstly, synthesizing a green channel in each red, green and blue RGB image in a basic training set or a basic test set and corresponding near infrared two NIR2 waveband information according to the following formula:
N=G*w+R*(1-w)
n represents a green channel of the enhanced red, green and blue RGB image, G represents a green channel of the red, green and blue RGB image of the basic training set or the basic test set, x represents multiplication operation, w represents a weight of the green channel of the enhanced red, green and blue RGB image and the green channel of the red, green and blue RGB image in the basic training set or the basic test set, the weight is 0.8, R represents a near-infrared two NIR2 wave band in the multispectral image corresponding to the red, green and blue RGB image in the basic training set or the basic test set;
replacing the green channel of each enhanced red, green and blue RGB image in the corresponding red, green and blue RGB image in the basic training set or the basic test set by the green channel of each enhanced red, green and blue RGB image to obtain the enhanced red, green and blue RGB image after the red, green and blue RGB image is updated;
thirdly, forming a first training set or a first testing set by all the enhanced red, green and blue RGB images;
fourthly, respectively normalizing the values of three wave bands, namely coastline wave band coast, yellow wave band Yellow and near infrared two wave band NIR2, in each multispectral image, respectively multiplying the three normalized values by 255, respectively replacing a red channel R, a green channel G and a blue channel B of a red, green and blue RGB image in a basic training set or a basic test set corresponding to the multispectral image by the product of the three normalized values, and forming the multiband pseudo-color image by the three updated channels of the red channel R, the green channel G and the blue channel B;
and a fifth step of combining all the multi-band pseudo-color images into a second training set or a second test set.
6. The remote sensing image segmentation method based on the confidence point according to the claim 1, wherein the feature label probability value in the steps (5 a) and (5 b) is that each image is input to the segmentation network, the segmentation network outputs five output values of each image, each pixel of which corresponds to a road, an overhead bridge, a building, vegetation and the ground, the five output values are normalized, the maximum value of the normalized five values is taken, and the maximum value is taken as the feature label probability value of the pixel.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910494015.0A CN110211138B (en) | 2019-06-08 | 2019-06-08 | Remote sensing image segmentation method based on confidence points |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910494015.0A CN110211138B (en) | 2019-06-08 | 2019-06-08 | Remote sensing image segmentation method based on confidence points |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110211138A CN110211138A (en) | 2019-09-06 |
CN110211138B true CN110211138B (en) | 2022-12-02 |
Family
ID=67791544
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910494015.0A Active CN110211138B (en) | 2019-06-08 | 2019-06-08 | Remote sensing image segmentation method based on confidence points |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110211138B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111401275B (en) * | 2020-03-20 | 2022-11-25 | 内蒙古工业大学 | Information processing method and device for identifying grassland edge |
CN112800915B (en) * | 2021-01-20 | 2023-06-27 | 北京百度网讯科技有限公司 | Building change detection method, device, electronic equipment and storage medium |
CN113139550B (en) * | 2021-03-29 | 2022-07-12 | 山东科技大学 | Remote sensing image coastline extraction method based on deep semantic segmentation network |
CN113284171B (en) * | 2021-06-18 | 2023-04-07 | 成都天巡微小卫星科技有限责任公司 | Vegetation height analysis method and system based on satellite remote sensing stereo imaging |
CN113538280B (en) * | 2021-07-20 | 2022-03-04 | 江苏天汇空间信息研究院有限公司 | Remote sensing image splicing method for removing color lines and incomplete images based on matrix binarization |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103971115A (en) * | 2014-05-09 | 2014-08-06 | 中国科学院遥感与数字地球研究所 | Automatic extraction method for newly-increased construction land image spots in high-resolution remote sensing images based on NDVI and PanTex index |
CN105427309A (en) * | 2015-11-23 | 2016-03-23 | 中国地质大学(北京) | Multiscale hierarchical processing method for extracting object-oriented high-spatial resolution remote sensing information |
CN106683112A (en) * | 2016-10-10 | 2017-05-17 | 中国交通通信信息中心 | High-resolution image-based road region building change extraction method |
CN106709948A (en) * | 2016-12-21 | 2017-05-24 | 浙江大学 | Quick binocular stereo matching method based on superpixel segmentation |
CN107392925A (en) * | 2017-08-01 | 2017-11-24 | 西安电子科技大学 | Remote sensing image terrain classification method based on super-pixel coding and convolutional neural networks |
CN107392130A (en) * | 2017-07-13 | 2017-11-24 | 西安电子科技大学 | Classification of Multispectral Images method based on threshold adaptive and convolutional neural networks |
CN108573276A (en) * | 2018-03-12 | 2018-09-25 | 浙江大学 | A kind of change detecting method based on high-resolution remote sensing image |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9704068B2 (en) * | 2012-06-22 | 2017-07-11 | Google Inc. | System and method for labelling aerial images |
-
2019
- 2019-06-08 CN CN201910494015.0A patent/CN110211138B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103971115A (en) * | 2014-05-09 | 2014-08-06 | 中国科学院遥感与数字地球研究所 | Automatic extraction method for newly-increased construction land image spots in high-resolution remote sensing images based on NDVI and PanTex index |
CN105427309A (en) * | 2015-11-23 | 2016-03-23 | 中国地质大学(北京) | Multiscale hierarchical processing method for extracting object-oriented high-spatial resolution remote sensing information |
CN106683112A (en) * | 2016-10-10 | 2017-05-17 | 中国交通通信信息中心 | High-resolution image-based road region building change extraction method |
CN106709948A (en) * | 2016-12-21 | 2017-05-24 | 浙江大学 | Quick binocular stereo matching method based on superpixel segmentation |
CN107392130A (en) * | 2017-07-13 | 2017-11-24 | 西安电子科技大学 | Classification of Multispectral Images method based on threshold adaptive and convolutional neural networks |
CN107392925A (en) * | 2017-08-01 | 2017-11-24 | 西安电子科技大学 | Remote sensing image terrain classification method based on super-pixel coding and convolutional neural networks |
CN108573276A (en) * | 2018-03-12 | 2018-09-25 | 浙江大学 | A kind of change detecting method based on high-resolution remote sensing image |
Also Published As
Publication number | Publication date |
---|---|
CN110211138A (en) | 2019-09-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110211138B (en) | Remote sensing image segmentation method based on confidence points | |
CN109558806B (en) | Method for detecting high-resolution remote sensing image change | |
CN110969088B (en) | Remote sensing image change detection method based on significance detection and deep twin neural network | |
CN111274865A (en) | Remote sensing image cloud detection method and device based on full convolution neural network | |
CN108229425A (en) | A kind of identifying water boy method based on high-resolution remote sensing image | |
CN108416784B (en) | Method and device for rapidly extracting boundary of urban built-up area and terminal equipment | |
CN107564017B (en) | Method for detecting and segmenting urban high-resolution remote sensing image shadow | |
CN111832518B (en) | Space-time fusion-based TSA remote sensing image land utilization method | |
CN110751727A (en) | Synthetic image construction method based on Landsat long-time sequence | |
CN110717921B (en) | Full convolution neural network semantic segmentation method of improved coding and decoding structure | |
CN114494821B (en) | Remote sensing image cloud detection method based on feature multi-scale perception and self-adaptive aggregation | |
CN114005042A (en) | Remote sensing image urban building extraction method based on shadow compensation and U-net | |
CN109961105B (en) | High-resolution remote sensing image classification method based on multitask deep learning | |
CN116343053B (en) | Automatic solid waste extraction method based on fusion of optical remote sensing image and SAR remote sensing image | |
CN102063700A (en) | Satellite remote sensing image generating method and system | |
CN102855627A (en) | City remote sensing image shadow detection method based on spectral characteristic and topological relation | |
Van de Voorde et al. | Extraction of land use/land cover related information from very high resolution data in urban and suburban areas | |
CN111539965A (en) | Urban surface water domestic high-resolution remote sensing self-adaptive extraction method cooperating with multi-source information | |
CN115205713A (en) | Method for recovering details of scenery color and texture in shadow area of remote sensing image of unmanned aerial vehicle | |
CN107507151B (en) | Multispectral remote sensing image real color restoration method and system | |
CN102222235A (en) | Object-oriented hyperspectral classification processing method based on object integration height information | |
CN107564024B (en) | SAR image aggregation region extraction method based on single-side aggregation line segment | |
CN109060813A (en) | Earth and stone material gradation automatic testing method based on image optimization classification | |
CN107705330A (en) | Visibility intelligence estimating and measuring method based on road camera | |
CN110751699B (en) | Color reconstruction method of optical remote sensing image based on convolutional neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |