CN117612031A - Remote sensing identification method for abandoned land based on semantic segmentation - Google Patents
Remote sensing identification method for abandoned land based on semantic segmentation Download PDFInfo
- Publication number
- CN117612031A CN117612031A CN202410086382.8A CN202410086382A CN117612031A CN 117612031 A CN117612031 A CN 117612031A CN 202410086382 A CN202410086382 A CN 202410086382A CN 117612031 A CN117612031 A CN 117612031A
- Authority
- CN
- China
- Prior art keywords
- abandoned
- semantic segmentation
- remote sensing
- barren
- land
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 61
- 238000000034 method Methods 0.000 title claims abstract description 25
- 238000012549 training Methods 0.000 claims abstract description 32
- 238000012360 testing method Methods 0.000 claims abstract description 19
- 238000012795 verification Methods 0.000 claims abstract description 15
- 238000005520 cutting process Methods 0.000 claims abstract description 7
- 238000010586 diagram Methods 0.000 claims description 15
- 230000006870 function Effects 0.000 claims description 10
- 230000008569 process Effects 0.000 claims description 8
- 230000004913 activation Effects 0.000 claims description 7
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000012545 processing Methods 0.000 claims description 5
- 238000013528 artificial neural network Methods 0.000 claims description 4
- 238000000605 extraction Methods 0.000 claims description 4
- 241000282326 Felis catus Species 0.000 claims description 3
- 230000003213 activating effect Effects 0.000 claims description 3
- 238000006243 chemical reaction Methods 0.000 claims description 3
- 238000013507 mapping Methods 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 238000000926 separation method Methods 0.000 claims description 3
- 238000004458 analytical method Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 3
- 238000012300 Sequence Analysis Methods 0.000 description 2
- 238000012271 agricultural production Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000007405 data analysis Methods 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000003595 spectral effect Effects 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 241000196324 Embryophyta Species 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013136 deep learning model Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013508 migration Methods 0.000 description 1
- 230000005012 migration Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/13—Satellite images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/52—Scale-space analysis, e.g. wavelet analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/188—Vegetation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Databases & Information Systems (AREA)
- Molecular Biology (AREA)
- Mathematical Physics (AREA)
- Medical Informatics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Astronomy & Astrophysics (AREA)
- Remote Sensing (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a remote sensing identification method of a abandoned land based on semantic segmentation, which comprises the following steps of; a1, acquiring an original remote sensing image of a region, and acquiring a permanent basic farmland image according to the permanent basic farmland division condition of the region; a2, acquiring a abandoned barren map spot, keeping the same scale and coordinate information of the abandoned barren map spot and a permanent basic farmland image, and cutting out an area covered by the abandoned barren map spot on the permanent basic farmland image to serve as a abandoned barren land sample, and taking an uncut area as a non-abandoned land sample; a3, dividing the abandoned barren land sample and the non-abandoned barren land sample into a training set, a verification set and a test set; a4, constructing a semantic segmentation model, and simultaneously inputting a training set and a verification set into the semantic segmentation model; training and fine-tuning semantic segmentation model parameters; and A5, acquiring the weight of the semantic segmentation model, and acquiring a binarized image of the test set after putting the weight into the test set. The model constructed in this way has significantly improved recognition capability and recognition accuracy for the abandoned land.
Description
Technical Field
The invention belongs to the technical fields of computational vision and remote sensing image segmentation, and particularly relates to a remote sensing identification method of abandoned land based on semantic segmentation.
Background
With the advent of the new century, grain demand is in a gradual rising state. If the wasted land can be effectively utilized, precious resources are provided for solving the grain safety problem. Thus, there is a need to find ways to identify and utilize such abandoned lands.
And the land identification can accurately detect the idle and uncultivated land through satellite remote sensing and other means. Through the technology, the utilization condition of the land can be known, and a solution can be formulated in a targeted manner so as to exert the potential of the land to the greatest extent. The current abandoned land identification has the following technologies:
the first is remote sensing data analysis and time sequence analysis; in the patent application with the publication number of CN103914678A, a combination scheme of remote sensing data analysis and time sequence analysis is adopted, and characteristic information among data is fused by utilizing multi-sensor, multi-resolution, multi-spectrum and multi-temporal remote sensing data, so as to construct a combination solution of exploring texture characteristics and vegetation indexes. In the patent publication No. CN116129284A, a long-time sequence remote sensing image is adopted, an exponential relation among various vegetation is constructed, and a cultivated land and abandoned land sample library is established. Solutions based on remote sensing image analysis and timing analysis typically require the combination of multispectral data and imagery across multiple points in time. The effectiveness of this approach is that the multispectral data can provide rich spectral information for analysis of the abandoned land area at the current stage. Analysis of these spectral features helps to determine the utilization of the land and thus helps to identify the abandoned land. The image collection in different periods, especially the images covering different time points in the same area, can greatly enhance the generalization capability of the multispectral analysis. This means that it is possible to have its academic recognition of the general characteristics of the wasteland, so that the results can be successfully applied to the wasteland detection tasks at different points in time. The migration learning mode can keep accuracy in different time periods, so that the change of the abandoned land can be effectively monitored. However, this approach presents some challenges. Firstly, acquiring multispectral data and images across multiple time points requires high requirements on image acquisition equipment, requiring advanced remote sensing techniques and equipment. Second, collecting time-series image data over many years requires considerable time and resources. This may include image acquisition, processing and sorting at different points in time. This time-consuming and labor-consuming process may have an impact on the implementation of the solution.
The second is machine learning; in the patent application with the publication number of CN112749628A, classification of vegetation and non-vegetation is realized by constructing a plurality of models, and characteristic classification of cloud and non-cloud is realized on the basis of the classification, so that a classification result of the abandoned land is obtained; the patent publication number CN116343111A combines contrast learning and semantic segmentation to realize identification of the abandoned land, divides the regional image into a plurality of categories by contrast learning to distinguish the regional image difference, and then adopts a scheme of semantic segmentation to extract a specific abandoned land region in the image. The feature extraction method based on machine learning generally adopts a scheme based on semantic segmentation when solving the problem of abandoned land detection. The semantic segmentation can distribute different parts of pixels in the image to different semantic categories, so that the ground object can be accurately segmented. However, relying solely on semantic segmentation methods may not adequately learn the characteristics of the abandoned land because the characteristics of the abandoned land may have similarities to other features, resulting in misclassification. In order to overcome the problem, the existing method adopts a feature analysis strategy, and introduces feature engineering into a machine learning process to guide a semantic segmentation model to reduce misjudgment. The goal of this strategy is to allow the model to better understand the unique characteristics of the abandoned land, thereby improving the accuracy of the test. The multi-model combined strategy is excellent in terms of improving accuracy, but has some inconveniences in terms of later service deployment.
Disclosure of Invention
The invention aims to provide a remote sensing identification method of a abandoned land based on semantic segmentation, which aims to solve the following technical problems in the background art:
the existing scheme has high requirements on the spectrum characteristics of the remote sensing image, has insufficient identification performance of the abandoned land due to more interference in the image, is complicated in later deployment of the model, and is not beneficial to updating and deployment of the model.
In order to solve the technical problems, the invention adopts the following technical scheme:
a remote sensing identification method of abandoned land based on semantic segmentation comprises the following steps;
a1, acquiring an original remote sensing image of a region, and acquiring a permanent basic farmland image according to the permanent basic farmland division condition of the region;
a2, acquiring a abandoned barren map spot, keeping the same scale and coordinate information of the abandoned barren map spot and a permanent basic farmland image, and cutting out an area covered by the abandoned barren map spot on the permanent basic farmland image to serve as a abandoned barren land sample, and taking an uncut area as a non-abandoned land sample;
a3, dividing the abandoned barren land sample and the non-abandoned barren land sample into a training set, a verification set and a test set;
a4, constructing a semantic segmentation model, and simultaneously inputting a training set and a verification set into the semantic segmentation model; introducing the heavy parameter convolution into a semantic segmentation model, and merging multiple branches of the heavy parameter into one branch in an inference stage;
data is processedFeature extraction input to semantic separation model constructionIn the network, feature sets F with different scales are obtained i ;
Wherein I represents an input sample picture, R represents a real number domain, C represents the number of channels of the input picture, and H and W represent the height and width of the input sample picture;
training phase characteristics F i The acquisition formula of (1) is:
inference stage feature F i The acquisition formula of (1) is:
wherein Conv (I) represents a feature map obtained by performing convolution operation on the input feature map I; BN (I) represents a characteristic diagram obtained by carrying out batch normalization operation on the input characteristic diagram I; f (Conv (I)) represents the processing of activating functions on the feature map obtained by the convolution operation;
training and fine-tuning semantic segmentation model parameters;
and A5, acquiring the weight of the semantic segmentation model, and acquiring a binarized image of the test set after putting the weight into the test set.
Further, remote sensing images corresponding to the abandoned map spots are segmented by using remote sensing image processing software.
Further, the training set accounts for 60% -80% of the total data, the verification set accounts for 10% -20% of the total data, and the test set accounts for 10% -20% of the total data.
Further, the semantic segmentation model adopts FCNHead with coordinate convolution, the coordinate convolution is introduced into a fully connected network, and the difference between local features and global features is learned;
scaling the dimensions to unity first, F i The conversion formula of (2) is as follows:
wherein F is m Represents F i Unified ruler of (2)Degree feature map, deConv () represents deconvolution operation;
feature F to be of uniform scale m Inputting into a neural network layer Coordconv to obtain a feature F with coordinate information c :
Wherein F is c Representing the output characteristic diagram, C x Representative and input feature map F m Feature map of corresponding x-coordinate information, C y Representative and input feature map F m Characteristic diagram of corresponding y coordinate information, cat represents splicing operation, F m 、C x And C y The three feature graphs are spliced according to the channel dimension;
will F c The calculation formula of the abandoned map spots M and M obtained by the convolution mapping mode is as follows:
wherein Sigmoid represents a Sigmoid activation function.
Further, in the Conv reasoning process, the BN layer parameters are fused with the Conv parameters, and the specific formula is as follows:
wherein, reLU represents ReLU activation function, and F represents final output characteristic diagram.
Further, inputting the multilayer features extracted by the FCNHead into a fully-connected network with coordinate convolution to obtain a segmentation result of the abandoned map spots.
Further, the trained semantic segmentation model is converted into a TensorRT format.
Compared with the prior art, the invention has the following beneficial effects:
according to the invention, the original remote sensing image of the region is obtained, and the permanent basic farmland image is obtained according to the permanent basic farmland division condition of the region, so that the training interference of the background information on the abandoned land recognition model can be reduced. Because permanent basic farmlands are clearly delimited by governments, farmlands that have to be agrochemically or used for other non-agricultural production activities have a clear association with the abandoned land. This may allow the model to focus more on various features of the abandoned land that are permanently relevant to the underlying farmland.
The invention can accurately acquire the abandoned map spot by keeping the same scale and coordinate information of the abandoned map spot and the permanent basic farmland image and cutting out the area covered by the abandoned map spot on the permanent basic farmland image as the abandoned land sample. Thus, the recognition capability and recognition accuracy of the model on the abandoned land can be improved.
Compared with the prior art, the invention only needs common optical images, can reduce partial noise interference by preprocessing the images, and can improve the identification precision of the abandoned land by utilizing a semantic segmentation model, thereby effectively reducing the deployment cost.
The invention combines multiple branches of heavy parameters into one branch in the reasoning stage, which has the functions of reducing the calculation complexity of the model and improving the reasoning speed. In the training stage, multiple branches can help the model to learn characteristics better, but in the reasoning stage, in order to reduce the calculation amount and memory occupation of the model, multiple branches are fused into one branch, so that the model is lighter, and later updating and deployment of the model are facilitated.
Detailed Description
The following description of the technical solutions in the embodiments of the present invention will be clear and complete, and it is obvious that the described embodiments are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Embodiment one.
A remote sensing identification method of abandoned land based on semantic segmentation comprises the following steps;
a1, acquiring an original remote sensing image of a region, and acquiring a permanent basic farmland image according to the permanent basic farmland division condition of the region;
specifically, the original remote sensing image of the region adopts satellite images, coordinate verification is needed to be carried out on the images and the permanent basic farmland pattern files after the original remote sensing image of the region is obtained, the coordinate information of the images is corrected, the sizes of the images are adjusted to be consistent with the pattern, and finally, all farmland pattern spots are screened out by the permanent basic farmland pattern, so that the original remote sensing image of the region is used as a training data sample.
So as to improve the accuracy and stability of subsequent treatment. It should be noted that, the permanent basic farmland is defined as farmland which is approved to be non-agricultural or used for other non-agricultural production activities in order to maintain farmland production, to ensure grain safety and ecological balance, within a range clearly defined by the government of the country, region or place. This demarcation is based on important considerations in ensuring agricultural sustainable development, maintaining agricultural resources and ecological environment. The abandoned land is from a permanent basic farmland. In view of the fact that the abandoned land is derived from the permanent basic farmland, the permanent basic farmland pattern obtained through the regional permanent basic farmland division condition can reduce the training interference of the rest background information on the abandoned land recognition model, namely the noise reduction is realized, so that the model focuses more on various characteristics of the abandoned land related to the permanent basic farmland.
A2, acquiring a abandoned barren map spot, keeping the same scale and coordinate information of the abandoned barren map spot and a permanent basic farmland image, and cutting out an area covered by the abandoned barren map spot on the permanent basic farmland image to serve as a abandoned barren land sample, and taking an uncut area as a non-abandoned land sample;
specifically, the abandoned map spots can be obtained by dividing the remote sensing image. Registering the abandoned map spots with the permanent basic farmland image to enable the abandoned map spots to have the same scale and coordinate information. Therefore, the corresponding relation between the image spots and the images can be ensured to be correct in the subsequent cutting and analyzing process. Cutting the image spots and the images according to the positions and boundary information of the image spots to obtain the abandoned image spots and the corresponding image areas. The cut pattern spots and the image should have the same scale and coordinate information. Dividing the cut image area into a abandoned land sample and a non-abandoned land sample according to the coverage condition of the image spots. Regarding the area covered by the pattern spots, the pattern spots are considered as a abandoned land sample; for uncovered areas, a sample of the unprecedented land is considered. The land samples that are not being abandoned may be marked 1 and the land samples that are not being abandoned may be marked 0.
A3, dividing the abandoned barren land sample and the non-abandoned barren land sample into a training set, a verification set and a test set;
a4, constructing a semantic segmentation model, and simultaneously inputting a training set and a verification set into the semantic segmentation model; introducing the heavy parameter convolution into a semantic segmentation model, and merging multiple branches of the heavy parameter into one branch in an inference stage;
data is processedInputting into a feature extraction network constructed by a semantic separation model to obtain feature sets F with different scales i ;
Wherein I represents an input sample picture, R represents a real number domain, C represents the number of channels of the input picture, and H and W represent the height and width of the input sample picture; the input feature map I may be understood as a three-dimensional tensor, in which each element corresponds to a pixel point in the feature map, and the number of channels c represents different feature information contained in the feature map. In deep learning, the input feature map I is an important input in the neural network, and is subjected to a series of operations such as convolution, pooling, activation and the like, so as to finally obtain the output of the network.
Training phase characteristics F i The acquisition formula of (1) is:
inference stage feature F i The acquisition formula of (1) is:
wherein Conv (I) represents a feature map obtained by performing convolution operation on the input feature map I; BN (I) represents a characteristic diagram obtained by carrying out batch normalization operation on the input characteristic diagram I; f (Conv (I)) represents the processing of activating functions on the feature map obtained by the convolution operation;
then, training and fine tuning semantic segmentation model parameters are needed;
specifically, the semantic segmentation model is a deep learning model, and is used for assigning each pixel in an image to a different class, so as to realize the recognition and segmentation of the pixel level of different objects in the image. In the scheme, a semantic segmentation model can be constructed, and a training set and a verification set are simultaneously input into the model for training. The training set is used to train parameters of the model and the validation set is used to evaluate the performance of the model on unseen data. Fine-tuning semantic segmentation model parameters refers to continuing training on already trained models to accommodate new tasks or datasets. In the scheme, a trained semantic segmentation model can be used as a basic model, and fine adjustment is performed on the data set of the abandoned map spots so as to improve the recognition capability of the model on the abandoned map spots. When the semantic segmentation model is constructed, the training set and the verification set are simultaneously input into the model to verify the model in the training process so as to monitor the performance of the model and timely adjust the super parameters to improve the generalization capability of the model. Therefore, the model can be ensured not to be fitted with training data in the training process, and meanwhile, the performance of the model on the verification set can be better known.
Introducing the heavy parameter convolution into the semantic segmentation model can enlarge the receptive field of the convolution operation, so that the model can better capture the global information of the input data. The heavy parameter convolution realizes the expansion of the receptive field by introducing intervals between convolution kernels, which is helpful for improving the understanding capability of a model to input data, and particularly for semantic segmentation tasks, semantic information of different scales in an image can be better captured.
The effect of merging multiple branches of heavy parameters into one branch in the reasoning stage is to reduce the calculation complexity of the model and improve the reasoning speed. In the training stage, multiple branches can help the model to learn characteristics better, but in the reasoning stage, in order to reduce the calculation amount and memory occupation of the model, multiple branches are fused into one branch, so that the model is lighter and more suitable for being deployed in practical application.
Preferably, the divided data sets are preprocessed, input formats required by the model are obtained, and data enhancement schemes such as rotation, scaling, clipping and the like are performed on the data.
Further preferably, the preprocessed data is input into FCNHead with a heavy parameter convolution, and the multi-scale features are obtained by multi-layer convolution.
Inputting the multi-scale features into the deconvolution, unifying the feature scales, and facilitating the subsequent segmentation; and finally, transmitting the features with uniform dimensions to a CoordFCN capable of learning coordinate information to obtain model predicted abandoned map spots.
Because the phenomenon of the abandoned land exists in the permanent basic farmland and occupies only a few acres, the pattern spots containing the abandoned land after the data are cut are far less than the pattern spots without the abandoned land, and the loss weights are proportionally divided for balancing the categories of model learning, so that the loss weights of the pattern spots without the abandoned land are restrained.
And A5, acquiring the weight of the semantic segmentation model, and acquiring a binarized image of the test set after putting the weight into the test set.
Specifically, a corresponding deep learning framework, such as TensorFlow, pyTorch, is used for loading trained model weights; and inputting the images of the test set into the model for prediction. The test set should contain similar images to the training set so that the model can make accurate predictions on the new data. And inputting the images of the test set into the weighted model for prediction. The prediction result will be a classification result at one pixel level, and each pixel point is classified into different categories. And according to the prediction result, dividing each pixel point into two categories of the abandoned land and other ground features. By setting a threshold, a binary image can be obtained by setting a certain category of the prediction result to 1 and other categories to 0, wherein the abandoned land is marked as 1 and the other land features are marked as 0.
In a preferred embodiment, remote sensing image processing software is utilized to divide the remote sensing image corresponding to the abandoned map spots to obtain the abandoned map spots.
In a preferred embodiment, the training set is 60% -80% of the total data, the verification set is 10% -20% of the total data, and the test set is 10% -20% of the total data.
In a preferred embodiment, the semantic segmentation model employs FCNHead with a coordinate convolution, which is introduced into a fully connected network, learning the differences between local and global features;
scaling the dimensions to unity first, F i The conversion formula of (2) is as follows:
wherein F is m Represents F i DeConv () represents a deconvolution operation;
feature F to be of uniform scale m Inputting into a neural network layer Coordconv to obtain a feature F with coordinate information c :
The semantic segmentation model adopts a segmentation head module of a CoordFCN. Expanding the input multi-scale features to the same size in a deconvolution mode, and compressing the stacked features in a convolution mode; and simultaneously, coordconv is introduced into a fully connected network, and the relation between local characteristics and global characteristics is learned, so that the characteristics between crops and weeds and shrubs are more easily distinguished.
Wherein F is c Representing the output characteristic diagram, C x Representative and input feature map F m Feature map of corresponding x-coordinate information, C y Representative and input feature map F m Characteristic diagram of corresponding y coordinate information, cat represents splicing operation, F m 、C x And C y The three feature graphs are spliced according to the channel dimension;
will F c The calculation formula of the abandoned map spots M and M obtained by the convolution mapping mode is as follows:
wherein Sigmoid represents a Sigmoid activation function.
In a preferred embodiment, in the Conv reasoning process, the BN layer parameters are fused with the Conv parameters, and the specific formula is expressed as follows:
wherein, reLU represents ReLU activation function, and F represents final output characteristic diagram.
In a preferred embodiment, the FCNHead extracted multi-layer features are input into a fully connected network with coordinate convolution to obtain the segmentation result of the abandoned map spots. The specific flow is as follows:
the FCNHead was used to extract the multi-layer features of the image.
And splicing the multilayer features with corresponding coordinate information to obtain a feature map with position information.
The feature map with the position information is input into a fully connected network with the coordinate convolution for processing.
The output result is the dividing result of the abandoned map spots.
The flow can fully utilize the position information and combine the expression capability of the fully connected network, thereby improving the accuracy and the robustness of the segmentation result.
In a preferred embodiment, the trained semantic segmentation model is converted to a TensorRT format. Thereby realizing end-to-end acceleration reasoning on the GPU server.
Although embodiments of the present invention have been shown and described, it will be understood by those skilled in the art that various changes, modifications, substitutions and alterations can be made therein without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.
Claims (7)
1. A remote sensing identification method of abandoned land based on semantic segmentation is characterized in that: comprises the following steps of;
a1, acquiring an original remote sensing image of a region, and acquiring a permanent basic farmland image according to the permanent basic farmland division condition of the region;
a2, acquiring a abandoned barren map spot, keeping the same scale and coordinate information of the abandoned barren map spot and a permanent basic farmland image, and cutting out an area covered by the abandoned barren map spot on the permanent basic farmland image to serve as a abandoned barren land sample, and taking an uncut area as a non-abandoned land sample;
a3, dividing the abandoned barren land sample and the non-abandoned barren land sample into a training set, a verification set and a test set;
a4, constructing a semantic segmentation model, and simultaneously inputting a training set and a verification set into the semantic segmentation model; introducing the heavy parameter convolution into a semantic segmentation model, and merging multiple branches of the heavy parameter into one branch in an inference stage;
data is processedInputting into a feature extraction network constructed by a semantic separation model to obtain feature sets F with different scales i ;
Wherein I represents an input sample picture, R represents a real number domain, C represents the number of channels of the input picture, and H and W represent the height and width of the input sample picture;
training phase characteristics F i The acquisition formula of (1) is:
inference stage feature F i The acquisition formula of (1) is:
wherein Conv (I) represents a feature map obtained by performing convolution operation on the input feature map I; BN (I) represents a characteristic diagram obtained by carrying out batch normalization operation on the input characteristic diagram I; f (Conv (I)) represents the processing of activating functions on the feature map obtained by the convolution operation;
training and fine-tuning semantic segmentation model parameters;
and A5, acquiring the weight of the semantic segmentation model, and acquiring a binarized image of the test set after putting the weight into the test set.
2. A method of remote sensing identification of a abandoned land based on semantic segmentation as defined in claim 1, wherein: and dividing the remote sensing image corresponding to the abandoned map spots by using remote sensing image processing software.
3. A method of remote sensing identification of a abandoned land based on semantic segmentation as defined in claim 1, wherein: the training set accounts for 60% -80% of the total data, the verification set accounts for 10% -20% of the total data, and the test set accounts for 10% -20% of the total data.
4. A method of remote sensing identification of a abandoned land based on semantic segmentation as defined in claim 1, wherein: the semantic segmentation model adopts FCNHead with coordinate convolution, the coordinate convolution is introduced into a fully connected network, and the difference between local features and global features is learned;
scaling the dimensions to unity first, F i The conversion formula of (2) is as follows:
wherein F is m Represents F i DeConv () represents a deconvolution operation;
feature F to be of uniform scale m Inputting into a neural network layer Coordconv to obtain a feature F with coordinate information c :
Wherein F is c Representing the output characteristic diagram, C x Representative and input feature mapF m Feature map of corresponding x-coordinate information, C y Representative and input feature map F m Characteristic diagram of corresponding y coordinate information, cat represents splicing operation, F m 、C x And C y The three feature graphs are spliced according to the channel dimension;
will F c The calculation formula of the abandoned map spots M and M obtained by the convolution mapping mode is as follows:
wherein Sigmoid represents a Sigmoid activation function.
5. A method of remote sensing identification of a abandoned land based on semantic segmentation as defined in claim 4, wherein: in the Conv reasoning process, the BN layer parameters are fused with the Conv parameters, and the specific formula is as follows:
wherein, reLU represents ReLU activation function, and F represents final output characteristic diagram.
6. A method of remote sensing identification of a abandoned land based on semantic segmentation as defined in claim 4, wherein: inputting the multilayer features extracted by FCNHead into a fully-connected network with coordinate convolution to obtain a segmentation result of the abandoned map spots.
7. A method of remote sensing identification of a abandoned land based on semantic segmentation as defined in claim 1, wherein: and converting the trained semantic segmentation model into a TensorRT format.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410086382.8A CN117612031A (en) | 2024-01-22 | 2024-01-22 | Remote sensing identification method for abandoned land based on semantic segmentation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410086382.8A CN117612031A (en) | 2024-01-22 | 2024-01-22 | Remote sensing identification method for abandoned land based on semantic segmentation |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117612031A true CN117612031A (en) | 2024-02-27 |
Family
ID=89960131
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202410086382.8A Pending CN117612031A (en) | 2024-01-22 | 2024-01-22 | Remote sensing identification method for abandoned land based on semantic segmentation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117612031A (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111985376A (en) * | 2020-08-13 | 2020-11-24 | 湖北富瑞尔科技有限公司 | Remote sensing image ship contour extraction method based on deep learning |
CN116958827A (en) * | 2023-06-16 | 2023-10-27 | 广东省国土资源测绘院 | Deep learning-based abandoned land area extraction method |
-
2024
- 2024-01-22 CN CN202410086382.8A patent/CN117612031A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111985376A (en) * | 2020-08-13 | 2020-11-24 | 湖北富瑞尔科技有限公司 | Remote sensing image ship contour extraction method based on deep learning |
CN116958827A (en) * | 2023-06-16 | 2023-10-27 | 广东省国土资源测绘院 | Deep learning-based abandoned land area extraction method |
Non-Patent Citations (6)
Title |
---|
WU Y等: "A Remote Sensing Method for Crop Mapping Based on Multiscale Neighborhood Feature Extraction", 《REMOTE SENSING》, vol. 15, no. 1, 22 December 2022 (2022-12-22), pages 1 - 23 * |
YAO X等: "Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features", 《SENSORS》, vol. 19, no. 12, 21 June 2019 (2019-06-21), pages 1 - 16 * |
吴强强等: "空间信息感知语义分割模型的高分辨率遥感影像道路提取", 《遥感学报》, vol. 26, no. 9, 25 September 2022 (2022-09-25), pages 1872 - 1885 * |
张宁: "基于深度学习的无监督遥感图像超分辨率重构技术研究", 《中国博士学位论文全文数据库 (工程科技Ⅱ辑)》, no. 4, 15 April 2023 (2023-04-15), pages 028 - 3 * |
李子茂等: "基于融合注意力机制的小样本遥感场景分类方法", 《国外电子测量技术》, vol. 42, no. 7, 15 July 2023 (2023-07-15), pages 59 - 67 * |
王恩德;齐凯;李学鹏;彭良玉;: "基于神经网络的遥感图像语义分割方法", 《光学学报》, vol. 39, no. 12, 27 August 2019 (2019-08-27), pages 93 - 104 * |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wen et al. | Directionally constrained fully convolutional neural network for airborne LiDAR point cloud classification | |
Pandey et al. | An intelligent system for crop identification and classification from UAV images using conjugated dense convolutional neural network | |
CN111126287B (en) | Remote sensing image dense target deep learning detection method | |
Paris et al. | A novel approach to the unsupervised update of land-cover maps by classification of time series of multispectral images | |
CN108052966A (en) | Remote sensing images scene based on convolutional neural networks automatically extracts and sorting technique | |
Cai et al. | Residual-capsule networks with threshold convolution for segmentation of wheat plantation rows in UAV images | |
CN109063660B (en) | Crop identification method based on multispectral satellite image | |
Xu et al. | Instance segmentation method for weed detection using UAV imagery in soybean fields | |
Baeta et al. | Learning deep features on multiple scales for coffee crop recognition | |
CN115965812B (en) | Evaluation method for classification of unmanned aerial vehicle images on wetland vegetation species and land features | |
CN113435254A (en) | Sentinel second image-based farmland deep learning extraction method | |
CN117409339A (en) | Unmanned aerial vehicle crop state visual identification method for air-ground coordination | |
Bhujade et al. | Role of digital, hyper spectral, and SAR images in detection of plant disease with deep learning network | |
CN112380917A (en) | A unmanned aerial vehicle for crops plant diseases and insect pests detect | |
Bhuyar et al. | Crop classification with multi-temporal satellite image data | |
CN113989652B (en) | Method and system for detecting farmland change under layered multiple judgment rules | |
Vaidhehi et al. | RETRACTED ARTICLE: An unique model for weed and paddy detection using regional convolutional neural networks | |
Granwehr et al. | Analysis on digital image processing for plant health monitoring | |
Jadhav et al. | Comprehensive review on machine learning for plant disease identification and classification with image processing | |
CN117612031A (en) | Remote sensing identification method for abandoned land based on semantic segmentation | |
Ouchra et al. | Comparison of Machine Learning Methods for Satellite Image Classification: A Case Study of Casablanca Using Landsat Imagery and Google Earth Engine | |
CN114463642A (en) | Cultivated land plot extraction method based on deep learning | |
Suneetha et al. | Object based Classification of Multispectral Remote Sensing Images for Forestry Applications | |
CN113762128A (en) | Hyperspectral image classification method based on unsupervised learning | |
CN117456287B (en) | Method for observing population number of wild animals by using remote sensing image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |