CN111291818A - Non-uniform class sample equalization method for cloud mask - Google Patents

Non-uniform class sample equalization method for cloud mask Download PDF

Info

Publication number
CN111291818A
CN111291818A CN202010099382.3A CN202010099382A CN111291818A CN 111291818 A CN111291818 A CN 111291818A CN 202010099382 A CN202010099382 A CN 202010099382A CN 111291818 A CN111291818 A CN 111291818A
Authority
CN
China
Prior art keywords
cloud
samples
mask
training
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010099382.3A
Other languages
Chinese (zh)
Other versions
CN111291818B (en
Inventor
吴炜
高星宇
范菁
沈瑛
夏列钢
葛炜炜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University of Technology ZJUT
Original Assignee
Zhejiang University of Technology ZJUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University of Technology ZJUT filed Critical Zhejiang University of Technology ZJUT
Priority to CN202010099382.3A priority Critical patent/CN111291818B/en
Publication of CN111291818A publication Critical patent/CN111291818A/en
Application granted granted Critical
Publication of CN111291818B publication Critical patent/CN111291818B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

A cloud mask-oriented heterogeneous class sample equalization method comprises the following steps: step 1: data preprocessing and sample acquisition; step 2: grouping samples; and step 3: training a cloud mask model; and 4, step 4: masking and evaluating a classifier; and 5: performing iterative training; step 6: a cloud mask; and 7: and (5) performing mask data post-processing. The sample equalization method effectively solves the problem of sample unbalance caused by uneven cloud types in the remote sensing image, and realizes effective identification and segmentation of various types of clouds in the image, so that the cloud mask precision is improved; the samples are selectively input, so that the influence of the missing or wrongly detected samples can be highlighted, the small-class samples are enhanced, the features extracted by the deep learning model are effectively adjusted, and the problem of missing detection or wrong detection of the small-class clouds is solved.

Description

Non-uniform class sample equalization method for cloud mask
Technical Field
The invention relates to a non-uniform class sample equalization method, in particular to a non-uniform class sample equalization method in a cloud mask training process based on deep learning.
Background
With the development of remote sensing technology, remote sensing images are widely applied to various fields such as crop mapping, forest monitoring and the like. However, due to the influence of the cloud in the imaging process of the remote sensing image, data loss or distortion exists in the corresponding area of the obtained image, and therefore the accuracy and the application significance of the information extraction result are influenced. The problem of data loss or distortion caused by cloud is particularly prominent in cloudy and rainy areas such as southern China and the like, and the data acquirability can be improved and the guarantee capability can be enhanced by using images partially covered by the cloud. In the application of partially missing images, the cloud masking process of marking pixels affected by clouds on the image pixel by pixel is the basis of various applications.
The cloud mask method based on supervised learning can learn cloud characteristics according to marked samples and adjust parameters of a classifier, has the potential of obtaining high cloud mask precision, and characteristic design and classifier selection are important factors for determining classification precision. Common features (e.g., spectral, shape, and texture features, etc.), machine learning methods (e.g., support vector machines and shallow neural networks, etc.), and combinations thereof have all been used for cloud masks (picrorhiza, chengchun, lianda. landau. Landsat image cloud detection with joint cloud volume automatic assessment and weighted support vector machines [ J ] mapping bulletin, 2014,43(8): 848-854.). However, these features are artificially designed, and since the different kinds of clouds have different features, it is difficult to accurately describe and distinguish all clouds by using a single or multiple features, thereby limiting the applicability of the method. Therefore, the samples need to be reselected and the classifier model needs to be adjusted each time the cloud mask is generated, resulting in higher time consumption and financial cost, and reducing the practicability of the algorithm.
In recent years, deep learning is widely applied and has great success in the field of computer vision because the feature extraction and classifier optimization are coupled together to realize collaborative optimization, thereby providing a new technical idea and scheme for cloud masks. One exemplary method is as follows: similar pixels can be aggregated into super pixels through a linear iterative clustering algorithm, and then the super pixels are used as input to carry out cloud detection. Meanwhile, the convolution operation can not only utilize the spectral characteristics of a single pixel, but also extract the spatial characteristics formed by the convolution window pixels, and the cloud masking by directly utilizing the convolution characteristics is favorable for improving the masking precision (Xie F, ShiM, Shi Z, et al. multilevel closed detection in Remote Sensing images based on horizontal masking J. IEEE Journal of Selected Topics in Applied easy alignment and Remote Sensing,2017,10(8): 3631-). The method for improving the masking precision of the image is simple and convenient to operate and has a wide application prospect. From the above discussion it can be seen that: the cloud mask method based on deep learning has realized a higher-precision mask effect.
However, as the cloud is a composite category, the cloud can be divided into 3 groups of low cloud, medium cloud and high cloud 10 according to the height of the cloud and by combining information such as whether precipitation, temperature and reverse radiation are generated, wherein the height of the low cloud is not more than 4km, and the low cloud comprises types such as layer cloud, rain layer cloud, rain cloud and the like; the height of the middle clouds is 4-6km, including high-volume clouds and high-layer clouds; the height of the high clouds is more than 6km, including the rolling clouds, the rolling clouds and the rolling layer clouds. Fig. 1 gives a schematic of the various clouds, and it can be seen that: the different types of clouds exhibit morphologies and color characteristics that are different, even very different; even the same type of cloud varies in its characteristics as its thickness, height, etc. varies. Meanwhile, the number of clouds is also very different, and the number of clouds varies with time and area, i.e. the unbalanced categories (King Shuaihui, Han Shi just, Yao Shi just, etc.. macroscopic feature analysis of various clouds in China and surrounding areas based on CloudSat data [ J ]. meteorology report, 2011,69(5): 883-.
The deep learning model needs to accurately recognize the features of different types of clouds to realize accurate recognition of the different types of clouds, but because: (1) the proportion of different types of clouds is different greatly and changes along with the influence of factors such as different latitudes and different seasons, and samples with balanced proportion are difficult to obtain; (2) the separability of the cloud is determined not only by the cloud itself, but also by the differentiability of the cloud from the underlying surface (such as city, forest, water body, desert, etc.), and the cloud and the underlying surface form a complex combination relationship. Due to the factors, the deep learning model cannot well learn the small-class cloud characteristics, and cloud missing of the corresponding type is caused.
In the field of computer vision and the like, one method for processing the imbalance category is to select samples by category to ensure the relative balance of various sample numbers. However, the cloud type determination method requires other auxiliary information such as height, and only relies on a visual determination method, which is also a challenging task for professional technicians. Thus, choosing samples category by category is difficult to implement. In order to solve the problems, the invention provides a sample equalization method based on packet training, which is characterized in that the mask accuracy of a classifier on various types of clouds is judged by using the current classification result, and missing detection and false detection samples are selectively input, so that the sensitivity of the classifier on the missing detection or false detection samples is enhanced.
Disclosure of Invention
Aiming at the problem of sample imbalance caused by non-uniform classes, the invention provides a sample equalization method, so that a classifier can be ensured to accurately identify different types of clouds.
The method has the principle that samples are grouped, a relatively stable deep learning model is trained through a previous group of samples, and then cloud masking and result evaluation are carried out on a next group of samples; and neglecting the samples which can be accurately identified by the current classifier, and inputting the false detection samples and the missed detection samples into the classifier for iterative training, so that the proportion of large classes in the total samples is reduced, the proportion of small classes in the total samples is increased, and the sample equalization is realized.
According to the principle, on the basis of obtaining an image to be masked, the cloud mask-oriented non-uniform type sample equalization method disclosed by the invention comprises the following implementation steps:
step 1: data preprocessing and sample acquisition;
the vegetation on the standard false color image is red, the distinguishability of various ground objects and clouds is large, and the standard false color mode is used for wave band synthesis.
Because the original images adopt different quantization bits, the same gray value has different meanings, and the gray value is compressed to an interval [1,255] by a percentage truncation mode, so that the images obtained by different sensors have similar colors.
Since the image is distributed in the form of scenes, the size of the image is much larger than the processing capacity of the computer, the image is sliced, and image blocks with the size h x w are output, and the number of the image blocks is M.
On the basis, a polygon of an area where the cloud is located on the image block is marked, and a remote sensing image cloud mask sample set G is manufactured, wherein the remote sensing image cloud mask sample set G comprises N image blocks.
Step 2: grouping samples;
sample G was divided into K groups of samples:
G={G1,G2,...,GK} (1)
wherein the sample subset G1As a pre-training classifier to obtain a relatively robust deep learning model; sample subset { G }2,...,Gk-1Testing the group of samples and selecting the samples in an iteration mode; sample subset GKFor detecting the accuracy.
Here, according to the client requirement, whether an intersection exists between the sample subsets can be ensured.
And step 3: training a cloud mask model;
after setting the training hyper-parameters, the sample subset G is used1Training a deep learning model by passing through a subset of samples G1The model obtained by training is C1
And 4, step 4: masking and evaluating a classifier;
use of C1For G2The samples in (1) are predicted to obtain a mask result, and the mask result is then compared with G2The tag data in (1) are compared, thereby evaluating the classification result.
The cloud mask results can be classified into three types of correct T, error F and omission M. Wherein, the correct representation sample is marked as cloud, and the classification result is also cloud; the error indicates that the sample is non-cloud and the classifier is extracted as cloud; missing indicates that the sample is marked as a cloud, but the classifier is not extracted. On the basis, three evaluation indexes, namely average detection rate (AR), average accuracy rate (AP) and Overall Accuracy (OA), are selected to evaluate the mask accuracy:
Figure BDA0002386415890000031
Figure BDA0002386415890000032
Figure BDA0002386415890000033
where, # denotes the number of pixels. It is worth noting that: since the number of background pixels formed by the clean ground object is much larger than that of the cloud pixels, so that the overall accuracy is difficult to reflect the real detection effect, the partial pixels are ignored when calculating the OA, and the OA-to-intersection ratio (IoU) is the same.
At the same time, grouping G samplesKThe classification result in (1) is subjected to cloud mask and precision evaluation thereof, and the precision is recorded as OA (G)K,C1),
And 5: performing iterative training;
for a block of cloud, those with both AR and AP greater than 0.5 are taken as correct samples, and others are taken as erroneous samples. Neglecting correct samples, adding error and missing samples, returning to step 3, and training to obtain new classifier C2(ii) a Then according to step 4, the sample G is processed3Evaluation was carried out.
The above process is repeated until all sample sets are exhausted or the classifier has stabilized, where classifier stabilization refers to the classification accuracy OA (G) of the ith timeK,Ci) Classification accuracy OA (G) with order i +1K,Ci+1) Is less than or equal to the tolerance e:
|OA(GK,Ci)-OA(GK,Ci+1)|≤e (5)
where i represents the number of iterations of the classifier.
The stable classifier obtained in the above step is marked as CK-1
Step 6: a cloud mask;
using classifier model CK-1Test set G to be processedKAnd carrying out cloud masking and evaluating the cloud masking precision.
And 7: post-processing mask data;
performing post-processing on the mask data, and removing clouds in the small communication area smaller than T; filling clean connected pixels smaller than T in the large cloud.
Preferably, step 1, standard false color synthesis is performed on the near-infrared band corresponding to red, the red band corresponding to green and the green band corresponding to blue to obtain a standard false color image; directly reducing the image from 16 bits to 8 bits by using percentage truncation, firstly counting an image histogram, extracting a gray value a when the gray value reaches 1% and a gray value b when the gray value reaches 99%, and then linearly stretching the [ a, b ] to [1,255 ];
cutting the remote sensing image into image blocks with the size of 512 × 512 pixels, generating a data set required by training, and obtaining 16000 image blocks in total through the mode;
in order to make a cloud mask label of the data set, the ArcMap is used for labeling the cloud on the image block, making a mask label with clear cloud boundary, and then converting the mask label into a grid format.
Preferably, the grouping of samples of step 2 is specifically: the first group of samples is used as a pre-training data set and is used for training a relatively stable deep learning model; taking the second group to the last but one group of samples as precision test samples, carrying out precision evaluation, and adding the samples for incremental training; the last group of samples do not participate in training and are independently used for detecting the precision;
carrying out image slicing according to the step 1, obtaining 16000 image blocks in total, firstly using 10000 samples for training, dividing 6000 samples into 6 groups for image testing, and 4000 samples for testing; the samples are divided into 8 groups, and the number of the samples in each group is (6000, 1000, 1000, 1000, 1000, 1000, 4000).
Preferably, the cloud mask model training in step 3 specifically comprises: in terms of initializing weights, a pre-training model of ResNet101 on a COCO data set is used for initializing Mask R-CNN network weights; in the aspect of super-parameter setting, the iteration number is set to 160, the batch _ size is set to 2, the pool _ size is set to 7, the mask _ pool _ size is set to 14, the initial learning rate is 0.001, and the learning rate update coefficient is 0.9;
inputting the first group of data in the grouping of the step 2 into a Mask R-CNN model, training according to the hyper-parameters, and obtaining a stable cloud Mask model C through training1
The invention has the advantages that: (1) the sample equalization method effectively solves the problem of sample unbalance caused by uneven cloud types in the remote sensing image, realizes effective identification and segmentation of various types of clouds in the image, and improves the accuracy of the cloud mask. (2) The samples are selectively input, so that the influence of the missing or wrongly detected samples can be highlighted, the small-class samples are enhanced, the features extracted by the deep learning model are effectively adjusted, and the problem of missing detection or wrong detection of the small-class clouds is solved.
Drawings
Fig. 1 is a cloud classification schematic.
FIG. 2 is a schematic flow chart of the cloud mask-oriented non-uniform class sample equalization method of the present invention.
FIG. 3 is a sample illustration of the invention. FIG. 3(a) is an original image; fig. 3(b) is artificial cloud mask data.
FIG. 4 is an exemplary diagram of the prediction results of the original Mask R-CNN model in the embodiment of the present invention: fig. 4(a) is an original image and fig. 4(b) is cloud mask data; FIG. 4(c) is the original mask result; fig. 4(d) shows mask result evaluation.
FIG. 5 is a graph of results of different iteration numbers of iterative training in an embodiment of the invention: FIG. 5(a) is an original remote sensing image; FIG. 5(b) shows the mask result, i.e., true value; fig. 5(c) - (i) use iterative masking results for samples of different input sets.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described below with reference to the accompanying drawings and embodiments. It should be understood that the detailed description and specific examples, while indicating the preferred embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention. Therefore, all other embodiments that can be obtained by a person skilled in the art without making any inventive step based on the embodiments of the present invention belong to the protection scope of the present invention.
Assuming that an image set to be masked is obtained, a partial region on the image is shielded by the cloud to cause data loss, and the cloud in the region needs to be masked. In the embodiment, 24 scenes of Landsat 5TM images, 20 scenes of Landsat8OLI images and 17 scenes of Sentinel 2Multi Spectral Imager images are selected in total, and because the invention is mainly based on the visual characteristics of the images for segmentation, various image types are not distinguished, and the images are processed as the same data.
Fig. 2 is a flowchart of the present invention, and a specific implementation method of the cloud mask-oriented non-uniform class sample equalization method of the present invention is as follows:
step 1: data preprocessing and sample acquisition;
because most images have near infrared, red and green wave bands, the near infrared wave band is corresponding to red, the red wave band is corresponding to green, and the green wave band is corresponding to blue to carry out standard false color synthesis to obtain a standard false color image. The vegetation on the image is red, and various ground objects are greatly different from clouds, so that the vegetation is high in separability.
Because the Sentinel 2MSI and Landsat8OLI images are quantized by 16 bits and the meanings of the same gray value are different, the invention does not perform complicated operations such as radiation correction and the like because the cloud mask is mainly performed on the basis of the visual characteristics of the images, and the images are directly reduced from 16 bits to 8 bits by using percentage truncation. The implementation is that firstly, the image histogram is counted, the gray value a when the gray value reaches 1% and the gray value b when the gray value reaches 99% are extracted, and then [ a, b ] is linearly stretched to [1,255 ].
Because the remote sensing image is generally distributed in a scene mode, the length and the width of a single scene image are large, and the single scene image is difficult to be directly input into a deep learning model for processing, the remote sensing image is cut into image blocks with the size of 512 pixels by 512 pixels to generate a data set required by training, and 16000 image blocks are obtained in the mode.
In order to make a cloud mask label of the data set, the ArcMap is used for labeling the cloud on the image block, making a mask label with clear cloud boundary, and then converting the mask label into a grid format. The sample results of this project selection are shown in fig. 3. Fig. 3(a) is an original image, and fig. 3(b) is an example of a mask according to the present invention.
Step 2: grouping samples;
the invention provides a packet training strategy to solve the problem of unbalanced samples. The grouping principle is as follows: the first group of samples is used as a pre-training data set and is used for training a relatively stable deep learning model; taking the second group to the last but one group of samples as precision test samples, carrying out precision evaluation, and adding the samples for incremental training; the last group of samples does not participate in training and is used for detecting the precision independently.
After the image slicing is performed according to the step 1, a total of 16000 image blocks is obtained. The invention firstly uses 10000 samples for training, 6000 samples are divided into 6 groups for image test, and 4000 samples for test. The samples are divided into 8 groups, and the number of the samples in each group is (6000, 1000, 1000, 1000, 1000, 1000, 4000).
The grouping of samples may be in a repeatable or non-repeatable manner, and the number or proportion of samples in each group may also be adjusted as required.
And step 3: training a cloud mask model;
the Mask R-CNN and other example segmentation methods have good effects in various applications, so the implementation process of the patent is described by taking the Mask R-CNN as an example, and certainly, more advanced classifiers can be selected for cloud masking according to requirements.
In initializing weights, Mask R-CNN network weights are initialized using a pre-trained model of ResNet101 on the COCO dataset. In the super-parameter setting aspect, the iteration number is set to 160, batch _ size is set to 2, pool _ size is set to 7, mask _ pool _ size is set to 14, the initial learning rate is set to 0.001, and the learning rate update coefficient is set to 0.9.
Inputting the first group of data in the grouping into a Mask R-CNN model, training according to the hyper-parameters of the steps, and training to obtain a stable cloud Mask model C1
And 4, step 4: grouping training and evaluation;
model C trained using step 31For sample subset G2And carrying out cloud mask to obtain a cloud mask result.
Then, for the packet GKPerforming cloud mask to obtain mask result, and calculating its overall accuracy OA (G)K,C1)。
And 5: performing iterative training;
the results of the mask are compared with the labels and can be divided into three types, namely correct T, error F and missing M,
on the basis, evaluating AR and AP of each cloud object, adding the cloud objects smaller than 0.5 into the training set, and returning to the step 3 to perform retraining by using the new training set; and then carrying out step 4 to carry out cloud mask on the sample.
The above process is repeated until all the sample subsets have been used or until the test set G has been completedKThe classification precision of the method is stable.
The present embodiment uses the OA change of less than 0.1% before and after the two times as the indicator of the classifier stability.
FIG. 4 is an example of the masking results herein. FIG. 4(a) is an original input image of the text method; FIG. 4(b) is a masking result of the method herein.
FIG. 5 shows the iterative masking results of the method herein, where FIG. 5(a) is the original image with two different types of clouds, thin and thick; FIG. 5(b) is a cloud mask result; fig. 5(c) - (i) show the effect of masking the clouds by sequentially using the sample subsets, and since there are fewer samples of thin clouds, there are more missed detections in the early masking results of fig. 5(c), and after the samples are added, the thin clouds are gradually detected, and the classification accuracy is improved.
Step 6: a cloud mask;
using classifier model CK-1Test set G to be processedKAnd carrying out cloud masking to obtain a pixel-level cloud masking result.
And 7: performing mask result post-processing;
and on the basis of obtaining the training model, carrying out cloud masking on the remote sensing image sample for testing to obtain a cloud masking result. Removing small communication clouds with less than 10 pixels generated in the cloud mask result; or a clean area of less than 10 pixels in the bulk communication area.
The embodiments described in this specification are merely illustrative of implementations of the inventive concept and the scope of the present invention should not be considered limited to the specific forms set forth in the embodiments but rather by the equivalents thereof as may occur to those skilled in the art upon consideration of the present inventive concept.

Claims (4)

1. A cloud mask-oriented non-uniform class sample equalization method comprises the following implementation steps:
step 1: data preprocessing and sample acquisition;
the vegetation on the standard false color image is red, the distinguishability of various ground objects and clouds is large, and the standard false color mode is used for wave band synthesis;
because the original images adopt different quantization bits, the same gray value has different meanings, and the gray value is compressed to an interval [1,255] by a percentage truncation mode, so that the images obtained by different sensors have similar colors;
because the image is distributed in a scene mode, the size of the image is far larger than the processing capacity of a computer, the image is sliced, and image blocks with the size h x w are output, and the number of the image blocks is M;
on the basis, marking a polygon of an area where the cloud is located on the image block, and manufacturing a remote sensing image cloud mask sample set G which comprises N image blocks;
step 2: grouping samples;
sample G was divided into K groups of samples:
G={G1,G2,...,GK} (1)
wherein the sample subset G1As a pre-training classifier to obtain a relatively robust deep learning model; sample subset { G }2,...,Gk-1Testing the group of samples and selecting the samples in an iteration mode; sample subset GKThe detection precision is used;
whether intersections exist among all sample subsets can be guaranteed according to the needs of clients;
and step 3: training a cloud mask model;
after setting the training hyper-parameters, the sample subset G is used1Training a deep learning model by passing through a subset of samples G1The model obtained by training is C1
And 4, step 4: masking and evaluating a classifier;
use of C1For G2The samples in (1) are predicted to obtain a mask result, and the mask result is then compared with G2The tag data in (1) are compared, thereby evaluating the classification result;
cloud mask results can be divided into three types of correct T, error F and omission M; wherein, the correct representation sample is marked as cloud, and the classification result is also cloud; the error indicates that the sample is non-cloud and the classifier is extracted as cloud; omission indicates that the sample is marked as a cloud, but the classifier is not extracted; on the basis, three evaluation indexes, namely average detection rate (AR), average accuracy rate (AP) and Overall Accuracy (OA), are selected to evaluate the mask accuracy:
Figure FDA0002386415880000021
Figure FDA0002386415880000022
Figure FDA0002386415880000023
wherein, # denotes the number of pixels; it is worth noting that: since the number of background pixels formed by the clean ground objects is far larger than that of cloud pixels, the whole precision is difficult to reflect the real detection effect, and therefore, when the OA is calculated, the part of pixels are ignored, and the OA and the cross-over ratio (IoU) are the same;
at the same time, grouping G samplesKThe classification result in (1) is subjected to cloud mask and precision evaluation thereof, and the precision is recorded as OA (G)K,C1),
And 5: performing iterative training;
for a cloud, taking the samples with both AR and AP larger than 0.5 as correct samples, and taking the other samples as error samples; neglecting correct samples, adding error and missing samples, returning to step 3, and training to obtain new classifier C2(ii) a Then according to step 4, the sample G is processed3Carrying out evaluation;
the above process is repeated until all sample sets are exhausted or the classifier has stabilized, where classifier stabilization refers to the classification accuracy OA (G) of the ith timeK,Ci) Classification accuracy OA (G) with order i +1K,Ci+1) Is less than or equal to the tolerance e:
|OA(GK,Ci)-OA(GK,Ci+1)|≤e (5)
wherein i represents the number of iterations of the classifier;
the stable classifier obtained in the above step is marked as CK-1
Step 6: a cloud mask;
using classifier model CK-1Carrying out cloud mask on the test set to be processed, and evaluating the accuracy of the cloud mask;
and 7: post-processing mask data;
performing post-processing on the mask data, and removing clouds in the small communication area smaller than T; filling clean connected pixels smaller than T in the large cloud.
2. The cloud mask-oriented non-uniform class sample equalization method as recited in claim 1, wherein: step 1, performing standard false color synthesis on a near infrared band corresponding to red, a red band corresponding to green and a green band corresponding to blue to obtain a standard false color image; directly reducing the image from 16 bits to 8 bits by using percentage truncation, firstly counting an image histogram, extracting a gray value a when the gray value reaches 1% and a gray value b when the gray value reaches 99%, and then linearly stretching the [ a, b ] to [1,255 ];
cutting the remote sensing image into image blocks with the size of 512 × 512 pixels, generating a data set required by training, and obtaining 16000 image blocks in total through the mode;
in order to make a cloud mask label of the data set, the ArcMap is used for labeling the cloud on the image block, making a mask label with clear cloud boundary, and then converting the mask label into a grid format.
3. The cloud mask-oriented non-uniform class sample equalization method as recited in claim 1, wherein: the sample grouping of step 2 is specifically: the first group of samples is used as a pre-training data set and is used for training a relatively stable deep learning model; taking the second group to the last but one group of samples as precision test samples, carrying out precision evaluation, and adding the samples for incremental training; the last group of samples do not participate in training and are independently used for detecting the precision;
carrying out image slicing according to the step 1, obtaining 16000 image blocks in total, firstly using 10000 samples for training, dividing 6000 samples into 6 groups for image testing, and 4000 samples for testing; the samples are divided into 8 groups, and the number of the samples in each group is (6000, 1000, 1000, 1000, 1000, 1000, 4000).
4. The cloud mask-oriented non-uniform class sample equalization method as recited in claim 1, wherein: the specific method for training the cloud mask model in the step 3 is as follows: in terms of initializing weights, a pre-training model of ResNet101 on a COCO data set is used for initializing Mask R-CNN network weights; in the aspect of super-parameter setting, the iteration number is set to 160, the batch _ size is set to 2, the pool _ size is set to 7, the mask _ pool _ size is set to 14, the initial learning rate is 0.001, and the learning rate update coefficient is 0.9;
inputting the first group of data in the grouping of the step 2 into a Mask R-CNN model, training according to the hyper-parameters, and obtaining a stable cloud Mask model C through training1
CN202010099382.3A 2020-02-18 2020-02-18 Non-uniform class sample equalization method for cloud mask Active CN111291818B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010099382.3A CN111291818B (en) 2020-02-18 2020-02-18 Non-uniform class sample equalization method for cloud mask

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010099382.3A CN111291818B (en) 2020-02-18 2020-02-18 Non-uniform class sample equalization method for cloud mask

Publications (2)

Publication Number Publication Date
CN111291818A true CN111291818A (en) 2020-06-16
CN111291818B CN111291818B (en) 2022-03-18

Family

ID=71024534

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010099382.3A Active CN111291818B (en) 2020-02-18 2020-02-18 Non-uniform class sample equalization method for cloud mask

Country Status (1)

Country Link
CN (1) CN111291818B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111898503A (en) * 2020-07-20 2020-11-06 中国农业科学院农业资源与农业区划研究所 Crop identification method and system based on cloud coverage remote sensing image and deep learning
CN113159146A (en) * 2021-04-08 2021-07-23 浙江天行健智能科技有限公司 Sample generation method, target detection model training method, target detection method and device
US11429070B2 (en) 2020-03-13 2022-08-30 Guangdong University Of Technology Inhomogeneous sample equalization method and system for product assembly process

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101980202A (en) * 2010-11-04 2011-02-23 西安电子科技大学 Semi-supervised classification method of unbalance data
CN102521656A (en) * 2011-12-29 2012-06-27 北京工商大学 Integrated transfer learning method for classification of unbalance samples
CN104951809A (en) * 2015-07-14 2015-09-30 西安电子科技大学 Unbalanced data classification method based on unbalanced classification indexes and integrated learning
CN106228130A (en) * 2016-07-19 2016-12-14 武汉大学 Remote sensing image cloud detection method of optic based on fuzzy autoencoder network
CN107610114A (en) * 2017-09-15 2018-01-19 武汉大学 Optical satellite remote sensing image cloud snow mist detection method based on SVMs
CN108460421A (en) * 2018-03-13 2018-08-28 中南大学 The sorting technique of unbalanced data
CN109934154A (en) * 2019-03-08 2019-06-25 北京科技大学 A kind of remote sensing image variation detection method and detection device
CN110378872A (en) * 2019-06-10 2019-10-25 河海大学 A kind of multi-source adaptive equalization transfer learning method towards crack image detection

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101980202A (en) * 2010-11-04 2011-02-23 西安电子科技大学 Semi-supervised classification method of unbalance data
CN102521656A (en) * 2011-12-29 2012-06-27 北京工商大学 Integrated transfer learning method for classification of unbalance samples
CN104951809A (en) * 2015-07-14 2015-09-30 西安电子科技大学 Unbalanced data classification method based on unbalanced classification indexes and integrated learning
CN106228130A (en) * 2016-07-19 2016-12-14 武汉大学 Remote sensing image cloud detection method of optic based on fuzzy autoencoder network
CN107610114A (en) * 2017-09-15 2018-01-19 武汉大学 Optical satellite remote sensing image cloud snow mist detection method based on SVMs
CN108460421A (en) * 2018-03-13 2018-08-28 中南大学 The sorting technique of unbalanced data
CN109934154A (en) * 2019-03-08 2019-06-25 北京科技大学 A kind of remote sensing image variation detection method and detection device
CN110378872A (en) * 2019-06-10 2019-10-25 河海大学 A kind of multi-source adaptive equalization transfer learning method towards crack image detection

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11429070B2 (en) 2020-03-13 2022-08-30 Guangdong University Of Technology Inhomogeneous sample equalization method and system for product assembly process
CN111898503A (en) * 2020-07-20 2020-11-06 中国农业科学院农业资源与农业区划研究所 Crop identification method and system based on cloud coverage remote sensing image and deep learning
CN111898503B (en) * 2020-07-20 2021-02-26 中国农业科学院农业资源与农业区划研究所 Crop identification method and system based on cloud coverage remote sensing image and deep learning
CN113159146A (en) * 2021-04-08 2021-07-23 浙江天行健智能科技有限公司 Sample generation method, target detection model training method, target detection method and device

Also Published As

Publication number Publication date
CN111291818B (en) 2022-03-18

Similar Documents

Publication Publication Date Title
CN110443143B (en) Multi-branch convolutional neural network fused remote sensing image scene classification method
CN109800736B (en) Road extraction method based on remote sensing image and deep learning
CN107016405B (en) A kind of pest image classification method based on classification prediction convolutional neural networks
CN111291818B (en) Non-uniform class sample equalization method for cloud mask
CN111274865A (en) Remote sensing image cloud detection method and device based on full convolution neural network
CN111310756B (en) Damaged corn particle detection and classification method based on deep learning
CN111611874B (en) Face mask wearing detection method based on ResNet and Canny
CN108257151B (en) PCANet image change detection method based on significance analysis
CN108960404B (en) Image-based crowd counting method and device
CN111639587B (en) Hyperspectral image classification method based on multi-scale spectrum space convolution neural network
CN109740483A (en) A kind of rice growing season detection method based on deep-neural-network
CN106897681A (en) A kind of remote sensing images comparative analysis method and system
CN104182985A (en) Remote sensing image change detection method
CN108171119B (en) SAR image change detection method based on residual error network
CN112200121A (en) Hyperspectral unknown target detection method based on EVM and deep learning
CN113312993B (en) Remote sensing data land cover classification method based on PSPNet
CN109886146B (en) Flood information remote sensing intelligent acquisition method and device based on machine vision detection
CN113887472A (en) Remote sensing image cloud detection method based on cascade color and texture feature attention
CN112258525A (en) Image abundance statistics and population recognition algorithm based on bird high frame frequency sequence
CN110458019B (en) Water surface target detection method for eliminating reflection interference under scarce cognitive sample condition
CN114266932A (en) Self-learning-based semi-supervised labeling method for remote sensing multispectral data
CN104732246B (en) A kind of semi-supervised coorinated training hyperspectral image classification method
CN115830514B (en) Whole river reach surface flow velocity calculation method and system suitable for curved river channel
CN111882573A (en) Cultivated land plot extraction method and system based on high-resolution image data
CN116563205A (en) Wheat spike counting detection method based on small target detection and improved YOLOv5

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant