WO2022083202A1 - Fine water body extraction method based on u-net neural network - Google Patents

Fine water body extraction method based on u-net neural network Download PDF

Info

Publication number
WO2022083202A1
WO2022083202A1 PCT/CN2021/108847 CN2021108847W WO2022083202A1 WO 2022083202 A1 WO2022083202 A1 WO 2022083202A1 CN 2021108847 W CN2021108847 W CN 2021108847W WO 2022083202 A1 WO2022083202 A1 WO 2022083202A1
Authority
WO
WIPO (PCT)
Prior art keywords
water body
neural network
training
net neural
data
Prior art date
Application number
PCT/CN2021/108847
Other languages
French (fr)
Chinese (zh)
Inventor
蔡玉林
秦鹏
孙孟昊
王雪丽
柳佳
范璞然
李龙坤
徐慧宇
Original Assignee
山东科技大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 山东科技大学 filed Critical 山东科技大学
Publication of WO2022083202A1 publication Critical patent/WO2022083202A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/182Network patterns, e.g. roads or rivers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30181Earth observation
    • G06T2207/30184Infrastructure

Definitions

  • the invention discloses a fine water body extraction method based on U-net neural network, and belongs to the technical field of convolution neural network and water body extraction.
  • the invention discloses a fine water body extraction method based on U-net neural network, so as to solve the problem that the recognition effect of small water bodies in remote sensing images is not good in the prior art.
  • Fine water extraction method based on U-net neural network including:
  • each combination will synthesize a multi-band remote sensing image, and input the multi-band remote sensing images into the U-net network for testing to determine the multi-band remote sensing that can obtain the best water body classification effect.
  • image called the optimal remote sensing image
  • the original image format is .tiff
  • the parameter used in the principal component analysis is a covariance matrix.
  • output bands that are uncorrelated with each other are generated.
  • ENVI is output in the following order: the variance of the first principal component is the largest, the variance of the second principal component is the second largest, and so on, the final principal component has the smallest variance.
  • the label data sets the water body value to 1 and the non-water body value to 0.
  • the verification set does not participate in the training, and is used to calculate the accuracy of the U-net neural network after the training is completed.
  • the training data of the optimal remote sensing image and the training data corresponding to the label data are divided, and a 64*64 image is output as the training data in step S5.
  • the U-net neural network is deepened, and the deepening includes a 3 ⁇ 3 convolutional layer, a 2 ⁇ 2 downsampling layer, a 2 ⁇ 2 upsampling layer and a skip connection layer , after the convolution operation is performed, the Relu activation function is used.
  • the setting of the U-net neural network uses the Adam optimizer to dynamically set the learning rate; each layer of the U-net neural network uses a random deactivation method to prevent overfitting, use Batch normalization prevents gradients from disappearing.
  • the generations used for U-net neural network training are divided into several batches, and the training accuracy and the verification accuracy are respectively output after the training of one batch is completed. If the accuracy of a certain number of generations is continuously maintained at 90% The above and no major changes occur, then stop the training of the U-net neural network.
  • the test data of the optimal remote sensing image is divided into 64*64 images; in the threshold segmentation, the value greater than 0.1 is assigned a value of 1, indicating a water body, and the rest is assigned a value of 0, indicating that non-water bodies.
  • the present invention has the following beneficial effects: the depth of the U-net neural network is deepened, so that it can adapt to hyperspectral data with more bands; the method of principal component analysis is used to reduce the dimension of the input data and ensure that the information is as far as possible No loss; about 90% of the water body recognition rate is achieved on the 10-meter-resolution remote sensing data, some small water bodies are identified, and there is no misclassification with roads, buildings, vegetation, etc.
  • FIG. 2 is a comparison diagram of the actual water body distribution and the water body extracted by the method of the present invention.
  • Fine water extraction method based on U-net neural network including:
  • the data of this embodiment are selected from the Zhubai No. 1 hyperspectral data in the area near Taihu Lake, Suzhou City, Jiangsu province. Since all 32 bands are input into neural network training, it is easy to cause memory overflow and cannot complete the training.
  • the format of the 32 bands is The original image of .tiff is imported into ENVI, and the principal component analysis function is selected to perform principal component analysis; the parameter is the covariance matrix, and the parameter setting is shown in Figure 1; the band with strong correlation is converted to the band with weak correlation, and the correlation is weak
  • the wavebands are used as principal components; the process of principal component analysis is as follows: establish a coordinate system with the origin as the mean value of the data, maximize the variance of the data through the rotation of the coordinate axis, and generate uncorrelated output bands; after forming the principal components, ENVI follows the Output in the following order: the first principal component has the largest variance, the second principal component has the second largest variance, and so on, the last principal component has the smallest variance, and the smaller the variance, the less information the principal component contains.
  • the processing preserves the original multi-band information and has a good control over the data dimension, reducing the amount of computation and shortening the computation time.
  • the first few principal components basically contain all the information of the original data, respectively input the first few principal components into the U-net network for testing, and perform various combinations of different principal components, each combination A multi-band remote sensing image is synthesized, and the multi-band remote sensing image is input into the U-net network for testing, and the multi-band remote sensing image that can obtain the best water classification effect is determined, which is called the optimal remote sensing image; the comparison results are shown in Table 1. It is shown that the classification effect achieved in the multi-band image input network composed of the first, second, third, fourth, fifth and sixth principal components is better, and the accuracy of the increase of the number of bands is smaller and the computational cost is larger.
  • step S4 Divide the optimal remote sensing image into training data and test data. About 3/4 of the upper left corner is the training data, and the remaining 1/4 is the test data; 15% of the data in the training data is used as the validation set; The optimal remote sensing image is divided in the same way; the verification set does not participate in the training, and is used to calculate the accuracy of the U-net neural network after the training is completed; due to the large size of the original image, in order to reduce and prevent memory overflow, before the execution of step S5, The training data of the optimal remote sensing image and the training data corresponding to the label data are divided, and 64*64 images are output for inputting into the network for training, and finally 1640 pieces of training data are obtained as the training data in step S5.
  • the training data is input into the U-net neural network for training once for one generation, with a total of 300 generations. In order to prevent memory overflow, every 8 generations is a batch. The accuracy of the neural network training is judged. The highest accuracy is 100%. If the accuracy of 10 generations is maintained above 90% continuously without major changes, the training of the U-net neural network will be stopped.
  • the test data of the optimal remote sensing image is divided into 64*64 images, and a total of 620 pieces of test data are obtained; the test data of the optimal remote sensing image is input into the trained U-net neural network to obtain the output image; the output image is thresholded Segmentation, splicing, and restoration to the original size; in the threshold segmentation, the value greater than 0.1 is assigned a value of 1, indicating a water body, and the rest is assigned a value of 0, indicating a non-water body.

Abstract

Disclosed is a fine water body extraction method based on a U-net neural network, which method relates to the technical field of convolutional neural networks and water body extraction, and particularly relates to water body extraction by means of hyperspectral data. The method comprises: importing original images of all wavebands into ENVI for principal component analysis; forming a variety of combinations of different principal components; forming label data; dividing an optimal remote-sensing image into training data and test data; inputting all the training data into a U-net neural network for training; inputting the test data of the optimal remote-sensing image into the trained U-net neural network, so as to obtain an output image; performing threshold value segmentation and splicing on the output image, and restoring the output image to an original size; and comparing the output image which has been restored to the original size with the test data in the label data, so as to evaluate the precision of fine water body extraction.

Description

基于U-net神经网络的精细水体提取方法A Fine Water Body Extraction Method Based on U-net Neural Network 技术领域technical field
本发明公开了基于U-net神经网络的精细水体提取方法,属于卷积神经网络和水体提取技术领域。The invention discloses a fine water body extraction method based on U-net neural network, and belongs to the technical field of convolution neural network and water body extraction.
背景技术Background technique
随着遥感技术的不断发展,使用遥感手段自动提取地表水体信息已成为全球信息提取研究方面的热点。由于水体在近红外波段附近光谱反射特征的不同,通过设置阈值可识别水体,但这种采用单一波段的方法很难将背景与水体完全区分开;使用多个波段的组合进行比值运算及指数提取方法能突出水体信息,但只适用于大范围水体提取,对一些细小水体的识别效果并不理想。With the continuous development of remote sensing technology, the use of remote sensing methods to automatically extract surface water information has become a hot spot in global information extraction research. Due to the different spectral reflection characteristics of water bodies near the near-infrared band, the water body can be identified by setting a threshold, but this method using a single band is difficult to completely distinguish the background from the water body; the combination of multiple bands is used for ratio calculation and index extraction. The method can highlight the water body information, but it is only suitable for large-scale water body extraction, and the recognition effect for some small water bodies is not ideal.
发明内容SUMMARY OF THE INVENTION
本发明公开了基于U-net神经网络的精细水体提取方法,以解决现有技术中对遥感影像的细小水体识别效果不好的问题。The invention discloses a fine water body extraction method based on U-net neural network, so as to solve the problem that the recognition effect of small water bodies in remote sensing images is not good in the prior art.
基于U-net神经网络的精细水体提取方法,包括:Fine water extraction method based on U-net neural network, including:
S1.将全部波段的原始影像导入ENVI,进行主成分分析,使相关性强的波段转换为相关性弱的波段,相关性弱的波段作为主成分;S1. Import the original images of all the bands into ENVI, and perform principal component analysis, so that the bands with strong correlation are converted into bands with weak correlation, and the bands with weak correlation are used as principal components;
S2.将不同的主成分进行多种组合,每种组合均合成一幅多波段遥感影像,把多波段遥感影像分别输入U-net网络进行试验,确定能获得最好水体分类效果的多波段遥感影像,称为最优遥感影像;S2. Perform multiple combinations of different principal components, each combination will synthesize a multi-band remote sensing image, and input the multi-band remote sensing images into the U-net network for testing to determine the multi-band remote sensing that can obtain the best water body classification effect. image, called the optimal remote sensing image;
S3.对最优遥感影像进行水体目视解译、使用ENVI勾画感兴趣区域、标注水体,形成标签数据;S3. Visually interpret the water body of the optimal remote sensing image, use ENVI to delineate the area of interest, label the water body, and form label data;
S4.将最优遥感影像划分为训练数据和测试数据,训练数据中一部分数据作为验证集;对标签数据作出与S4中最优遥感影像相同方式的划分;S4. Divide the optimal remote sensing image into training data and test data, and use a part of the data in the training data as a validation set; divide the label data in the same way as the optimal remote sensing image in S4;
S5.将所有训练数据输入U-net神经网络进行训练,经过卷积操作和最大池化的下采样操作,获得第一次水体特征;经过卷积操作和上采样操作,获得第二次水体特征;将上述两次水体特征进行连接,提取最终的水体特征;将最终水体特征与真实水体特征进行对比,不断优化网络参数,使U-net神经网络输出的水体和真实水体不断接近,完成对U-net神经网络的训练;S5. Input all the training data into the U-net neural network for training. After convolution operation and maximum pooling downsampling operation, the first water body feature is obtained; after convolution operation and upsampling operation, the second water body feature is obtained. ; Connect the above two water body features to extract the final water body features; compare the final water body features with the real water body features, and continuously optimize the network parameters, so that the water body output by the U-net neural network is constantly close to the real water body, and the U-net neural network output is continuously close to the real water body. - training of net neural network;
S6.将最优遥感影像的测试数据输入训练后的U-net神经网络,得到输出影像;将输出影像进行阈值分割、拼接,还原至原始大小;S6. Input the test data of the optimal remote sensing image into the U-net neural network after training to obtain the output image; perform threshold segmentation and splicing, and restore the output image to the original size;
S7.把还原至原始大小的输出影像与标签数据中的测试数据进行对比,评价精细水体提取的精度。S7. Compare the output image restored to the original size with the test data in the label data to evaluate the precision of fine water body extraction.
优选地,所述步骤S1中原始影像格式为.tiff,进行主成分分析时使用的参数为协方差矩阵,过程为:建立一个原点为数据均值的坐标系统,通过坐标轴的旋转使数据的方差达到最大,生成互不相关的输出波段。Preferably, in the step S1, the original image format is .tiff, and the parameter used in the principal component analysis is a covariance matrix. At the maximum, output bands that are uncorrelated with each other are generated.
优选地,所述步骤S1中,形成主成分后ENVI按照以下顺序输出:第一主成分方差最大,第二主成分方差第二大,以此类推,最后的主成分方差最小。Preferably, in the step S1, after the principal components are formed, ENVI is output in the following order: the variance of the first principal component is the largest, the variance of the second principal component is the second largest, and so on, the final principal component has the smallest variance.
优选地,所述步骤S3中,标签数据设置水体值为1,非水体值为0。Preferably, in the step S3, the label data sets the water body value to 1 and the non-water body value to 0.
优选地,所述步骤S4中,验证集不参与训练,用于计算训练完成后U-net神经网络的精度。Preferably, in the step S4, the verification set does not participate in the training, and is used to calculate the accuracy of the U-net neural network after the training is completed.
优选地,所述步骤S5执行前,将最优遥感影像的训练数据、对应标签数据的训练数据进行分割,输出64*64的影像,作为步骤S5中的训练数据。Preferably, before the step S5 is performed, the training data of the optimal remote sensing image and the training data corresponding to the label data are divided, and a 64*64 image is output as the training data in step S5.
优选地,所述步骤S5中,对U-net神经网络进行加深处理,加深后包含了3×3的卷积层、2×2的下采样层、2×2的上采样层和跳跃连接层,卷积操作执行后,使用Relu激活函数。Preferably, in the step S5, the U-net neural network is deepened, and the deepening includes a 3×3 convolutional layer, a 2×2 downsampling layer, a 2×2 upsampling layer and a skip connection layer , after the convolution operation is performed, the Relu activation function is used.
优选地,所述步骤S5中,U-net神经网络的设置使用Adam优化器,对学习率进行动态设置;U-net神经网络的每一层都使用随机失活的方法防止过拟合、使用批标准化防止梯度消失。Preferably, in the step S5, the setting of the U-net neural network uses the Adam optimizer to dynamically set the learning rate; each layer of the U-net neural network uses a random deactivation method to prevent overfitting, use Batch normalization prevents gradients from disappearing.
优选地,所述步骤S5中,将U-net神经网络训练所用的世代分为若干批次,一个批次训练完成后分别输出训练精度、验证精度,若有一定量世代的精度连续维持在90%以上且不发生大的改变,则停止U-net神经网络的训练。Preferably, in the step S5, the generations used for U-net neural network training are divided into several batches, and the training accuracy and the verification accuracy are respectively output after the training of one batch is completed. If the accuracy of a certain number of generations is continuously maintained at 90% The above and no major changes occur, then stop the training of the U-net neural network.
优选地,所述步骤S6执行前,将最优遥感影像的测试数据分割成64*64的影像;所述阈值分割中,值大于0.1的赋值为1,表示水体,其余的赋值为0,表示非水体。Preferably, before the step S6 is executed, the test data of the optimal remote sensing image is divided into 64*64 images; in the threshold segmentation, the value greater than 0.1 is assigned a value of 1, indicating a water body, and the rest is assigned a value of 0, indicating that non-water bodies.
优选地,所述步骤S7中,以回召率和准确率作为精度评价标准:回召率=(神经网络正确识别出水体的像素个数)/(实际水体像素个数),准确率=(神经网络正确识别出水体的像素个数)/(神经网络识别出水体的像素总个数)。Preferably, in the step S7, the recall rate and the accuracy rate are used as the precision evaluation criteria: the recall rate=(the number of pixels of the water body correctly identified by the neural network)/(the actual number of pixels of the water body), the accuracy rate=( The number of pixels that the neural network correctly identifies the water body)/(the total number of pixels that the neural network recognizes the water body).
与现有技术相比,本发明的有益效果是:加深了U-net神经网络深度,使之适应波段较多的高光谱数据;利用主成分分析的方法来减少输入数据的维度并保证信息尽量不受损失;在十米分辨率的遥感数据上实现了约90%的水体识别率,一些细小的水体都被识别出来,且没有与道路、建筑、植被等错分。Compared with the prior art, the present invention has the following beneficial effects: the depth of the U-net neural network is deepened, so that it can adapt to hyperspectral data with more bands; the method of principal component analysis is used to reduce the dimension of the input data and ensure that the information is as far as possible No loss; about 90% of the water body recognition rate is achieved on the 10-meter-resolution remote sensing data, some small water bodies are identified, and there is no misclassification with roads, buildings, vegetation, etc.
附图说明Description of drawings
图1为主成分分析参数设置图;Figure 1. Principal component analysis parameter setting diagram;
图2为实际水体分布与本发明所用方法提取水体的对比图。FIG. 2 is a comparison diagram of the actual water body distribution and the water body extracted by the method of the present invention.
具体实施方式Detailed ways
下面结合具体实施方式对本发明作进一步详细说明:Below in conjunction with specific embodiment, the present invention is described in further detail:
基于U-net神经网络的精细水体提取方法,包括:Fine water extraction method based on U-net neural network, including:
S1.本实施例数据选取的是江苏省苏州市太湖附近地区的珠海一号高光谱数据,由于将全部32个波段输入神经网络训练容易导致内存溢出无法完成训练,因此将32个波段的格式为.tiff的原始影像导入ENVI,选择主成分分析功能,进行主成分分析;参数为协方差矩阵,参数设置如图1所示;使相关性强的波段转换为相关性弱的波段,相关性弱的波段作为主成分;进行主成分分析时过程为:建立一个原点为数据均值的坐标系统,通过坐标轴的旋转使数据的方差达到最大,生成互不相关的输出波段;形成主成分后ENVI按照以下顺序输出:第一主成分方差最大,第二主成分方差第二大,以此类推,最后的主成分方差最小,方差越小的主成分所含信息相对也较少,通过主成分分析的处理既保留原始的多波段信息又能很好的控制数据维度,减少计算量并缩短计算时间。S1. The data of this embodiment are selected from the Zhuhai No. 1 hyperspectral data in the area near Taihu Lake, Suzhou City, Jiangsu Province. Since all 32 bands are input into neural network training, it is easy to cause memory overflow and cannot complete the training. Therefore, the format of the 32 bands is The original image of .tiff is imported into ENVI, and the principal component analysis function is selected to perform principal component analysis; the parameter is the covariance matrix, and the parameter setting is shown in Figure 1; the band with strong correlation is converted to the band with weak correlation, and the correlation is weak The wavebands are used as principal components; the process of principal component analysis is as follows: establish a coordinate system with the origin as the mean value of the data, maximize the variance of the data through the rotation of the coordinate axis, and generate uncorrelated output bands; after forming the principal components, ENVI follows the Output in the following order: the first principal component has the largest variance, the second principal component has the second largest variance, and so on, the last principal component has the smallest variance, and the smaller the variance, the less information the principal component contains. The processing preserves the original multi-band information and has a good control over the data dimension, reducing the amount of computation and shortening the computation time.
S2.进行主成分分析后,一般前几个主成分基本包含原始数据全部的信息,分别将前几个主成分输入U-net网络进行试验,将不同的主成分进行多种组合,每种组合均合成一幅多波段遥感影像,把多波段遥感影像分别输入U-net网络进行试验,确定能获得最好水体分类效果的多波段遥感影像,称为最优遥感影像;对比结果如表1所示,综合对比选择第一、二、三、四、五、六主成分组成的多波段图像输入网络中实现的分类效果较好,继续增加波段数量精度增加较小且花费计算成本较大。S2. After performing principal component analysis, generally the first few principal components basically contain all the information of the original data, respectively input the first few principal components into the U-net network for testing, and perform various combinations of different principal components, each combination A multi-band remote sensing image is synthesized, and the multi-band remote sensing image is input into the U-net network for testing, and the multi-band remote sensing image that can obtain the best water classification effect is determined, which is called the optimal remote sensing image; the comparison results are shown in Table 1. It is shown that the classification effect achieved in the multi-band image input network composed of the first, second, third, fourth, fifth and sixth principal components is better, and the accuracy of the increase of the number of bands is smaller and the computational cost is larger.
表1 前几个主成分的分类精度Table 1 Classification accuracy of the first few principal components
主成分main ingredient 1-3主成分1-3 principal components 1-4主成分1-4 Principal Components 1-5主成分1-5 principal components 1-6主成分1-6 Principal Components 1-7主成分1-7 Principal Components 1-8主成分1-8 principal components
精度precision 86.41%86.41% 87.23%87.23% 89.45%89.45% 92.32%92.32% 92.65%92.65% 92.89%92.89%
S3.对最优遥感影像进行水体目视解译、使用ENVI勾画感兴趣区域、标注水体,形成标签数据;标签数据设置水体值为1,非水体值为0。S3. Visually interpret the optimal remote sensing image for water bodies, use ENVI to delineate areas of interest, and label water bodies to form label data; the label data is set to 1 for water bodies and 0 for non-water bodies.
S4.将最优遥感影像划分为训练数据和测试数据,左上角约3/4为训练数据,其余1/4为测试数据;训练数据中15%的数据作为验证集;对标签数据作出与S4中最优遥感影像相同方式的划分;验证集不参与训练,用于计算训练完成后U-net神经网络的精度;由于原图像尺寸较大,为了降低防止内存溢出,所述步骤S5执行前,将最优遥感影像的训练数据、对应标签数据的训练数据进行分割,输出64*64的影像,以便输入网络进行训练,最终得到1640张训练数据,作为步骤S5中的训练数据。S4. Divide the optimal remote sensing image into training data and test data. About 3/4 of the upper left corner is the training data, and the remaining 1/4 is the test data; 15% of the data in the training data is used as the validation set; The optimal remote sensing image is divided in the same way; the verification set does not participate in the training, and is used to calculate the accuracy of the U-net neural network after the training is completed; due to the large size of the original image, in order to reduce and prevent memory overflow, before the execution of step S5, The training data of the optimal remote sensing image and the training data corresponding to the label data are divided, and 64*64 images are output for inputting into the network for training, and finally 1640 pieces of training data are obtained as the training data in step S5.
S5.为了更充分的利用珠海一号高光谱数据的优势,提取各波段的有用信息,对U-net神经网络进行加深处理,加深后包含了22个3×3的卷积层、5个2×2的下采样层、5个2×2的上采样层和5个跳跃连接层,卷积操作执行后,使用Relu激活函数;U-net神经网络的设置使用Adam优化器,对学习率进行动态设置;U-net神经网络的每一层都使用随机失活的方法防止过拟合、使用批标准化防止梯度消失。S5. In order to make full use of the advantages of Zhuhai No. 1 hyperspectral data, extract useful information of each band, and deepen the U-net neural network. After deepening, it contains 22 3×3 convolutional layers, 5 2 ×2 downsampling layer, 5 2×2 upsampling layers and 5 skip connection layers, after the convolution operation is performed, the Relu activation function is used; the setting of the U-net neural network uses the Adam optimizer, and the learning rate is adjusted. Dynamic settings; each layer of the U-net neural network uses random deactivation to prevent overfitting and batch normalization to prevent vanishing gradients.
将训练数据输入U-net神经网络训练一次为一个世代,共300个世代,为了防止内存溢出,每8个世代为一个批次,每个批次训练完成后分别输出训练精度、验证精度,以便进行神经网络训练时精度的判断,精度最高为100%,若有10个世代的精度连续维持在90%以上且不发生大的改变,则停止U-net神经网络的训练。The training data is input into the U-net neural network for training once for one generation, with a total of 300 generations. In order to prevent memory overflow, every 8 generations is a batch. The accuracy of the neural network training is judged. The highest accuracy is 100%. If the accuracy of 10 generations is maintained above 90% continuously without major changes, the training of the U-net neural network will be stopped.
将所有训练数据输入U-net神经网络进行训练,经过11次卷积操作和5次最大池化的下采样操作,获得第一次水体特征;经过过11次卷积操作和5次上采样操作,获得第二次水体特征;将上述两次水体特征进行连接,提取最终的水体特征;将最终水体特征与真实水体特征进行对比,不断优化网络参数,使U-net神经网络输出的水体和真实水体不断接近,完成对U-net神经网络的训练;训练结束后,最终得到的训练精度为94.21%,验证精度为92.32%。Input all training data into U-net neural network for training, after 11 convolution operations and 5 maximum pooling downsampling operations, the first water feature is obtained; after 11 convolution operations and 5 upsampling operations , obtain the second water body feature; connect the above two water body features to extract the final water body feature; compare the final water body feature with the real water body feature, and continuously optimize the network parameters to make the water body output by the U-net neural network match the real water body. The water body is constantly approaching, and the training of the U-net neural network is completed; after the training, the final training accuracy is 94.21%, and the verification accuracy is 92.32%.
将最优遥感影像的测试数据分割成64*64的影像,共得到620张测试数据;将最优遥感影像的测试数据输入训练后的U-net神经网络,得到输出影像;将输出影像进行阈值分割、拼接,还原至原始大小;所述阈值分割中,值大于0.1的赋值为1,表示水体,其余的赋值为0,表示非水体。The test data of the optimal remote sensing image is divided into 64*64 images, and a total of 620 pieces of test data are obtained; the test data of the optimal remote sensing image is input into the trained U-net neural network to obtain the output image; the output image is thresholded Segmentation, splicing, and restoration to the original size; in the threshold segmentation, the value greater than 0.1 is assigned a value of 1, indicating a water body, and the rest is assigned a value of 0, indicating a non-water body.
S7.把还原至原始大小的输出影像与标签数据中的测试数据进行对比,对比结果如图2所示;评价精细水体提取的精度,以回召率和准确率作为精度评价标准:召回率显示了水体识别的完整程度,回召率=(神经网络正确识别出水体的像素个数)/(实际水体像素个数);准确率显示了神经网络识别出的水体的正确率,准确率=(神经网络正确识别出水体的像素个数)/(神经网络识别出水体的像素总个数);最终结果显示U-net神经网络的召回率为89.03%,准确率为89.50%。S7. Compare the output image restored to the original size with the test data in the label data, and the comparison result is shown in Figure 2; to evaluate the precision of fine water extraction, the recall rate and accuracy rate are used as the precision evaluation criteria: the recall rate display The completeness of the water body identification, the recall rate = (the number of pixels that the neural network correctly identifies the water body)/(the actual number of water body pixels); the accuracy rate shows the correct rate of the water body identified by the neural network, and the accuracy rate = ( The number of pixels that the neural network correctly identifies the water body)/(the total number of pixels that the neural network recognizes the water body); the final result shows that the recall rate of the U-net neural network is 89.03%, and the accuracy rate is 89.50%.
当然,上述说明并非是对本发明的限制,本发明也并不仅限于上述举例,本技术领域的技术人员在本发明的实质范围内所做出的变化、改型、添加或替换,也应属于本发明的保护范围。Of course, the above description is not intended to limit the present invention, and the present invention is not limited to the above examples. Changes, modifications, additions or substitutions made by those skilled in the art within the essential scope of the present invention should also belong to the present invention. the scope of protection of the invention.

Claims (10)

  1. 基于U-net神经网络的精细水体提取方法,其特征在于,包括:The fine water body extraction method based on U-net neural network is characterized in that, it includes:
    S1.将全部波段的原始影像导入ENVI,进行主成分分析,使相关性强的波段转换为相关性弱的波段,相关性弱的波段作为主成分;S1. Import the original images of all the bands into ENVI, and perform principal component analysis, so that the bands with strong correlation are converted into bands with weak correlation, and the bands with weak correlation are used as principal components;
    S2.将不同的主成分进行多种组合,每种组合均合成一幅多波段遥感影像,把多波段遥感影像分别输入U-net网络进行试验,确定能获得最好水体分类效果的多波段遥感影像,称为最优遥感影像;S2. Perform multiple combinations of different principal components, each combination will synthesize a multi-band remote sensing image, and input the multi-band remote sensing images into the U-net network for testing to determine the multi-band remote sensing that can obtain the best water body classification effect. image, called the optimal remote sensing image;
    S3.对最优遥感影像进行水体目视解译、使用ENVI勾画感兴趣区域、标注水体,形成标签数据;S3. Visually interpret the water body of the optimal remote sensing image, use ENVI to delineate the area of interest, label the water body, and form label data;
    S4.将最优遥感影像划分为训练数据和测试数据,训练数据中一部分数据作为验证集;对标签数据作出与S4中最优遥感影像相同方式的划分;S4. Divide the optimal remote sensing image into training data and test data, and use a part of the data in the training data as a validation set; divide the label data in the same way as the optimal remote sensing image in S4;
    S5.将所有训练数据输入U-net神经网络进行训练,经过卷积操作和最大池化的下采样操作,获得第一次水体特征;经过卷积操作和上采样操作,获得第二次水体特征;将上述两次水体特征进行连接,提取最终的水体特征;将最终水体特征与真实水体特征进行对比,不断优化网络参数,使U-net神经网络输出的水体和真实水体不断接近,完成对U-net神经网络的训练;S5. Input all the training data into the U-net neural network for training. After convolution operation and maximum pooling downsampling operation, the first water body feature is obtained; after convolution operation and upsampling operation, the second water body feature is obtained. ; Connect the above two water body features to extract the final water body features; compare the final water body features with the real water body features, and continuously optimize the network parameters, so that the water body output by the U-net neural network is constantly close to the real water body, and the U-net neural network output is continuously close to the real water body. - training of net neural network;
    S6.将最优遥感影像的测试数据输入训练后的U-net神经网络,得到输出影像;将输出影像进行阈值分割、拼接,还原至原始大小;S6. Input the test data of the optimal remote sensing image into the U-net neural network after training to obtain the output image; perform threshold segmentation and splicing, and restore the output image to the original size;
    S7.把还原至原始大小的输出影像与标签数据中的测试数据进行对比,评价精细水体提取的精度。S7. Compare the output image restored to the original size with the test data in the label data to evaluate the precision of fine water body extraction.
  2. 根据权利要求1所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S1中原始影像格式为.tiff,进行主成分分析时使用的参数为协方差矩阵,过程为:建立一个原点为数据均值的坐标系统,通过坐标轴的旋转使数据的方差达到最大,生成互不相关的输出波段。The method for fine water body extraction based on U-net neural network according to claim 1, characterized in that, in the step S1, the original image format is .tiff, and the parameter used in the principal component analysis is a covariance matrix, and the process is: : Establish a coordinate system with the origin as the mean value of the data, and maximize the variance of the data through the rotation of the coordinate axis to generate uncorrelated output bands.
  3. 根据权利要求2所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S1中,形成主成分后ENVI按照以下顺序输出:第一主成分方差最大,第二主成分方差第二大,以此类推,最后的主成分方差最小。The method for fine water body extraction based on U-net neural network according to claim 2, characterized in that, in the step S1, after the principal components are formed, ENVI is output in the following order: the first principal component has the largest variance, and the second principal component has the largest variance. The variance is the second largest, and so on, and the final principal component has the smallest variance.
  4. 根据权利要求1所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S3中,标签数据设置水体值为1,非水体值为0。The method for fine water body extraction based on U-net neural network according to claim 1, characterized in that, in the step S3, the label data sets the water body value to 1 and the non-water body value to 0.
  5. 根据权利要求1所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S4中,验证集不参与训练,用于计算训练完成后U-net神经网络的精度。The method for extracting fine water bodies based on U-net neural network according to claim 1, characterized in that, in the step S4, the verification set does not participate in the training, and is used to calculate the accuracy of the U-net neural network after the training is completed.
  6. 根据权利要求1所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述 步骤S5执行前,将最优遥感影像的训练数据、对应标签数据的训练数据进行分割,输出64*64的影像,作为步骤S5中的训练数据。The method for fine water body extraction based on U-net neural network according to claim 1, characterized in that, before the step S5 is executed, the training data of the optimal remote sensing image and the training data corresponding to the label data are divided, and the output 64 *64 images are used as training data in step S5.
  7. 根据权利要求6所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S5中,对U-net神经网络进行加深处理,加深后包含了3×3的卷积层、2×2的下采样层、2×2的上采样层和跳跃连接层,卷积操作执行后,使用Relu激活函数。The method for fine water body extraction based on U-net neural network according to claim 6, characterized in that, in the step S5, the U-net neural network is deepened, and after the deepening, a 3×3 convolutional layer is included , 2×2 downsampling layer, 2×2 upsampling layer and skip connection layer, after the convolution operation is performed, the Relu activation function is used.
  8. 根据权利要求7所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S5中,U-net神经网络的设置使用Adam优化器,对学习率进行动态设置;U-net神经网络的每一层都使用随机失活的方法防止过拟合、使用批标准化防止梯度消失。The method for fine water body extraction based on U-net neural network according to claim 7, characterized in that, in the step S5, the setting of the U-net neural network uses an Adam optimizer to dynamically set the learning rate; Each layer of the net neural network uses random deactivation to prevent overfitting and batch normalization to prevent vanishing gradients.
  9. 根据权利要求8所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S5中,将U-net神经网络训练所用的世代分为若干批次,一个批次训练完成后分别输出训练精度、验证精度,若有一定量世代的精度连续维持在90%以上且不发生大的改变,则停止U-net神经网络的训练。The method for fine water body extraction based on U-net neural network according to claim 8, wherein in the step S5, the generations used for U-net neural network training are divided into several batches, and the training is completed in one batch Then output the training accuracy and verification accuracy respectively. If the accuracy of a certain number of generations is continuously maintained above 90% without major changes, the training of the U-net neural network is stopped.
  10. 根据权利要求1所述的基于U-net神经网络的精细水体提取方法,其特征在于,所述步骤S6执行前,将最优遥感影像的测试数据分割成64*64的影像;所述阈值分割中,值大于0.1的赋值为1,表示水体,其余的赋值为0,表示非水体;步骤S7中,以回召率和准确率作为精度评价标准:回召率=(神经网络正确识别出水体的像素个数)/(实际水体像素个数),准确率=(神经网络正确识别出水体的像素个数)/(神经网络识别出水体的像素总个数)。The method for fine water body extraction based on U-net neural network according to claim 1, wherein, before the step S6 is executed, the test data of the optimal remote sensing image is divided into 64*64 images; the threshold segmentation In step S7, the recall rate and the accuracy rate are used as the precision evaluation criteria: the recall rate = (the neural network correctly identifies the water body The number of pixels of the actual water body)/(the actual number of pixels of the water body), the accuracy rate = (the number of pixels that the neural network correctly identifies the water body)/(the total number of pixels that the neural network identifies the water body).
PCT/CN2021/108847 2020-10-20 2021-07-28 Fine water body extraction method based on u-net neural network WO2022083202A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011121566.1A CN112232229B (en) 2020-10-20 2020-10-20 Fine water body extraction method based on U-net neural network
CN202011121566.1 2020-10-20

Publications (1)

Publication Number Publication Date
WO2022083202A1 true WO2022083202A1 (en) 2022-04-28

Family

ID=74118697

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/108847 WO2022083202A1 (en) 2020-10-20 2021-07-28 Fine water body extraction method based on u-net neural network

Country Status (2)

Country Link
CN (1) CN112232229B (en)
WO (1) WO2022083202A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115170947A (en) * 2022-05-12 2022-10-11 广东省科学院广州地理研究所 Estuary turbid zone and water body classification method, device and equipment based on remote sensing image
CN116310604A (en) * 2023-05-23 2023-06-23 北京大学第三医院(北京大学第三临床医学院) Placenta implantation parting assessment tool and method
CN116704350A (en) * 2023-06-16 2023-09-05 浙江时空智子大数据有限公司 Water area change monitoring method and system based on high-resolution remote sensing image and electronic equipment
CN116879192A (en) * 2023-09-07 2023-10-13 航天宏图信息技术股份有限公司 Water bloom prediction method, device, equipment and medium based on satellite remote sensing data
CN117612017A (en) * 2024-01-23 2024-02-27 江西啄木蜂科技有限公司 Environment-adaptive remote sensing image change detection method

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112232229B (en) * 2020-10-20 2022-04-01 山东科技大学 Fine water body extraction method based on U-net neural network
CN116071373A (en) * 2023-03-01 2023-05-05 南通大学 Automatic U-net model tongue segmentation method based on fusion PCA

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103366189A (en) * 2013-07-10 2013-10-23 河海大学 Intelligent classification method for high-spectrum remote sensing image
WO2017041842A1 (en) * 2015-09-09 2017-03-16 Planetek Hellas E.P.E Methods of compressing and decompressing spectral data
CN111104889A (en) * 2019-12-04 2020-05-05 山东科技大学 Water body remote sensing identification method based on U-net
CN112232229A (en) * 2020-10-20 2021-01-15 山东科技大学 Fine water body extraction method based on U-net neural network

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107066995A (en) * 2017-05-25 2017-08-18 中国矿业大学 A kind of remote sensing images Bridges Detection based on convolutional neural networks
CN108537192B (en) * 2018-04-17 2021-11-30 福州大学 Remote sensing image earth surface coverage classification method based on full convolution network
CN110427836B (en) * 2019-07-11 2020-12-01 重庆市地理信息和遥感应用中心(重庆市测绘产品质量检验测试中心) High-resolution remote sensing image water body extraction method based on multi-scale optimization

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103366189A (en) * 2013-07-10 2013-10-23 河海大学 Intelligent classification method for high-spectrum remote sensing image
WO2017041842A1 (en) * 2015-09-09 2017-03-16 Planetek Hellas E.P.E Methods of compressing and decompressing spectral data
CN111104889A (en) * 2019-12-04 2020-05-05 山东科技大学 Water body remote sensing identification method based on U-net
CN112232229A (en) * 2020-10-20 2021-01-15 山东科技大学 Fine water body extraction method based on U-net neural network

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115170947A (en) * 2022-05-12 2022-10-11 广东省科学院广州地理研究所 Estuary turbid zone and water body classification method, device and equipment based on remote sensing image
CN115170947B (en) * 2022-05-12 2023-04-18 广东省科学院广州地理研究所 Estuary turbid zone and water body classification method, device and equipment based on remote sensing image
CN116310604A (en) * 2023-05-23 2023-06-23 北京大学第三医院(北京大学第三临床医学院) Placenta implantation parting assessment tool and method
CN116310604B (en) * 2023-05-23 2024-02-13 北京大学第三医院(北京大学第三临床医学院) Placenta implantation parting assessment tool and method
CN116704350A (en) * 2023-06-16 2023-09-05 浙江时空智子大数据有限公司 Water area change monitoring method and system based on high-resolution remote sensing image and electronic equipment
CN116704350B (en) * 2023-06-16 2024-01-30 浙江时空智子大数据有限公司 Water area change monitoring method and system based on high-resolution remote sensing image and electronic equipment
CN116879192A (en) * 2023-09-07 2023-10-13 航天宏图信息技术股份有限公司 Water bloom prediction method, device, equipment and medium based on satellite remote sensing data
CN116879192B (en) * 2023-09-07 2023-12-12 航天宏图信息技术股份有限公司 Water bloom prediction method, device, equipment and medium based on satellite remote sensing data
CN117612017A (en) * 2024-01-23 2024-02-27 江西啄木蜂科技有限公司 Environment-adaptive remote sensing image change detection method

Also Published As

Publication number Publication date
CN112232229A (en) 2021-01-15
CN112232229B (en) 2022-04-01

Similar Documents

Publication Publication Date Title
WO2022083202A1 (en) Fine water body extraction method based on u-net neural network
CN109446992B (en) Remote sensing image building extraction method and system based on deep learning, storage medium and electronic equipment
CN108615010B (en) Facial expression recognition method based on parallel convolution neural network feature map fusion
CN111815601B (en) Texture image surface defect detection method based on depth convolution self-encoder
EP3091479B1 (en) Method and apparatus for fingerprint identification
CN108256482B (en) Face age estimation method for distributed learning based on convolutional neural network
CN101620669B (en) Method for synchronously recognizing identities and expressions of human faces
CN104573731A (en) Rapid target detection method based on convolutional neural network
CN101520894A (en) Method for extracting significant object based on region significance
CN109872305B (en) No-reference stereo image quality evaluation method based on quality map generation network
CN112837234B (en) Human face image restoration method based on multi-column gating convolution network
CN112766279B (en) Image feature extraction method based on combined attention mechanism
CN111967464B (en) Weak supervision target positioning method based on deep learning
CN110909615B (en) Target detection method based on multi-scale input mixed perception neural network
CN112785526B (en) Three-dimensional point cloud restoration method for graphic processing
CN111368825A (en) Pointer positioning method based on semantic segmentation
Shawon et al. Silicon wafer map defect classification using deep convolutional neural network with data augmentation
CN115526864A (en) Steel rail surface defect detection method based on improved characteristic pyramid network and metric learning
Zhao et al. Fingerprint pre-processing and feature engineering to enhance agricultural products categorization
CN113705466B (en) Face five sense organ shielding detection method for shielding scene, especially under high imitation shielding
CN114372926A (en) Traditional Chinese medicine tongue tenderness identification method based on image restoration and convolutional neural network
CN110443790B (en) Cartilage identification method and system in medical image
CN106384364A (en) LPP-ELM based objective stereoscopic image quality evaluation method
CN115588136A (en) Neural network hyperspectral image classification method based on multi-level spatial spectrum fusion
CN102214292B (en) Illumination processing method for human face images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21881622

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21881622

Country of ref document: EP

Kind code of ref document: A1