WO2024087574A1 - Panoptic segmentation-based optical remote-sensing image raft mariculture area classification method - Google Patents

Panoptic segmentation-based optical remote-sensing image raft mariculture area classification method Download PDF

Info

Publication number
WO2024087574A1
WO2024087574A1 PCT/CN2023/092747 CN2023092747W WO2024087574A1 WO 2024087574 A1 WO2024087574 A1 WO 2024087574A1 CN 2023092747 W CN2023092747 W CN 2023092747W WO 2024087574 A1 WO2024087574 A1 WO 2024087574A1
Authority
WO
WIPO (PCT)
Prior art keywords
segmentation
panoramic
raft
image
instance
Prior art date
Application number
PCT/CN2023/092747
Other languages
French (fr)
Chinese (zh)
Inventor
汪承义
郭艳君
陈建胜
杜云艳
王雷
汪祖家
Original Assignee
中国科学院空天信息创新研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国科学院空天信息创新研究院 filed Critical 中国科学院空天信息创新研究院
Publication of WO2024087574A1 publication Critical patent/WO2024087574A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A40/00Adaptation technologies in agriculture, forestry, livestock or agroalimentary production
    • Y02A40/80Adaptation technologies in agriculture, forestry, livestock or agroalimentary production in fisheries management
    • Y02A40/81Aquaculture, e.g. of fish

Definitions

  • the present application relates to the field of marine remote sensing and image processing technology, and in particular to a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images.
  • Marine raft aquaculture is an important part of marine aquaculture. Compared with ponds and tidal flats near the coast, raft aquaculture has a wide range and is dispersed in the region.
  • the traditional on-site measurement method is not only time-consuming and labor-intensive, but also difficult to obtain accurate results for large areas.
  • the development of remote sensing technology has greatly made up for the shortcomings of traditional ground measurement, such as small coverage and low data acquisition efficiency.
  • the use of deep learning methods to realize intelligent information extraction of remote sensing images can quickly and accurately obtain the distribution and aquaculture type information of marine aquaculture areas, which is a reliable and advanced technical means for dynamic monitoring of marine raft aquaculture.
  • Synthetic aperture radar has the characteristics of all-day and all-weather, and has been widely used in the field of remote sensing.
  • SAR images have disadvantages such as low resolution, susceptibility to noise interference, severe geometric distortion, and fewer available features.
  • Optical remote sensing has a wide range of applications. Optical remote sensing clearly describes the boundary information of ground objects and contains rich spectral information, which is conducive to the extraction of raft aquaculture boundary and aquaculture type information. However, some optical images are interfered by clouds, fog and light. These interference factors restrict the extraction of feature information of optical remote sensing images and increase the difficulty of target recognition and segmentation in optical remote sensing images.
  • the existing convolutional neural network-based marine aquaculture area extraction is mainly divided into semantic segmentation and instance segmentation, for example, improved SOLO, D-ResUnet, HCHNet and other segmentation algorithms.
  • Segmentation and instance segmentation are both pixel-level classification.
  • the predicted value of each pixel is mapped to a probability value of [0,1] through the Softmax function, and then the error between the predicted value and the true label value is judged by the cross entropy loss function.
  • the model is continuously trained by the gradient descent method to minimize the error between the two.
  • the more types of semantic segmentation targets that is, the more data set labels, the more interference items are encountered when identifying and segmenting each target.
  • the present application provides a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation to solve the above problems.
  • the present application provides a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images, including:
  • the pre-trained panoramic segmentation model includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module;
  • the semantic segmentation branch network uses the semantic segmentation branch network to perform semantic segmentation on the image to be segmented to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area;
  • the instance segmentation branch network is used to perform instance segmentation on the image to be segmented to obtain An initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories;
  • the panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
  • the semantic segmentation branch network is an improved U 2 -Net network
  • the improved U 2 -Net network includes at least 6 U-shaped secondary encoders and 5 U-shaped secondary decoders, the 6 U-shaped secondary encoders are sequentially 4 first secondary encoders and 2 second secondary encoders, and the 5 U-shaped secondary decoders are sequentially 4 first secondary decoders and 1 second secondary decoder;
  • the first secondary encoder and the first secondary decoder are both composed of a first convolution block, an LSFE module, a plurality of down-sampling modules, a DPC module, a second convolution block, a first convolution block, and a plurality of up-sampling modules in sequence;
  • the LSFE module is used to extract the features of the breeding area within a large field of view, and includes a separable convolution and an output filter;
  • the DPC module is used to capture long-range context information, which includes separable convolution and output channels.
  • the instance segmentation branch network includes an improved SOTR network, and the improved SOTR network includes at least a Transformer module; wherein the Transformer module includes a separable convolution and an iABN synchronization layer; and the Transformer module is used to predict the category of each instance.
  • the instance segmentation branch network also includes a feature extraction module, and the feature extraction module includes a mobile reverse bottleneck unit and a bidirectional feature pyramid network;
  • instance segmentation is performed on the image to be segmented through the improved SOTR network to obtain an initial instance segmentation result.
  • the pre-trained panoramic segmentation model is trained in the following manner:
  • the labels include semantic labels of raft aquaculture areas and non-raft aquaculture areas and instance labels of various aquaculture area categories;
  • panoramic fusion module to adaptively fuse the training semantic segmentation result and the training instance segmentation result to obtain a training multi-classification result
  • a total loss is obtained according to the first loss and the second loss, and the panoramic segmentation model is trained based on the training multi-classification result and the total loss until the panoramic segmentation model converges to obtain a trained panoramic segmentation model.
  • the method further includes:
  • inputting the training data set into the semantic segmentation branch network includes:
  • the step of inputting the training data set into the instance segmentation branch network comprises:
  • the shared synthetic dataset is input into the instance segmentation branch network.
  • the training data set includes a label data set and an adversarial sample set.
  • the label data set is a data set with corresponding labels after annotation, and the adversarial sample set is obtained by performing adversarial training on the segmentation results of the training instances.
  • the label data set is obtained in the following manner:
  • the multiple aquaculture area categories include at least fish, algae and shellfish.
  • the present application also provides a device for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images, comprising:
  • An image acquisition module is used to acquire an image to be segmented, wherein the image to be segmented is an optical remote sensing image of a marine aquaculture area;
  • An image segmentation module is used to input the image to be segmented into a pre-trained panoramic segmentation model to predict a multi-classification segmentation result, wherein the multi-classification segmentation result includes raft aquaculture area, non-raft aquaculture area and multiple aquaculture area categories;
  • the pre-trained panoramic segmentation model includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module;
  • the semantic segmentation branch network uses the semantic segmentation branch network to perform semantic segmentation on the image to be segmented to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area;
  • instance segmentation branch network uses the instance segmentation branch network to perform instance segmentation on the image to be segmented to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories;
  • the panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
  • the method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation uses a semantic segmentation branch network and an instance segmentation branch network to parallelly segment the segmented image, and fuses the outputs of the two branch networks through a parameter-free panoramic fusion module, and finally obtains a multi-classification segmentation result to achieve multi-task classification.
  • the adaptive fusion method of the panoramic fusion module can more completely utilize the logical outputs of the semantic segmentation head and the instance segmentation head to improve the accuracy of multi-classification tasks.
  • FIG1 is a schematic diagram of a flow chart of a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application;
  • FIG2 is a second flow chart of a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application;
  • FIG3 is a schematic diagram of the structure of an existing U 2 -Net network
  • FIG4 is a structural comparison diagram of the En_1 substructure in the existing U 2 -Net network and the improved En_1 substructure of the present application;
  • FIG5 is a schematic diagram of the training process of the panoptic segmentation model provided in an embodiment of the present application.
  • FIG6 is a schematic diagram of the structure of a device for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application.
  • this application unifies the semantic segmentation and instance segmentation prediction subnetworks, and fuses the outputs to form an overall panoramic segmentation network model; constructs a panoramic segmentation label dataset for the multi-classification task of marine raft aquaculture areas, thereby realizing multi-classification tasks, and making the classification of marine raft aquaculture areas more refined, and improving the segmentation accuracy of the model.
  • the following is a specific description of the optical remote sensing image marine raft aquaculture area classification method based on panoramic segmentation proposed in this application in conjunction with the accompanying drawings.
  • Figure 1 is one of the flow charts of the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application
  • Figure 2 is the second flow chart of the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application.
  • the optical remote sensing image based on panoramic segmentation of the marine raft aquaculture area Classification methods include:
  • the multi-classification segmentation results include raft aquaculture areas, non-raft aquaculture areas and multiple aquaculture area categories, and the multiple aquaculture area categories include fish, algae, shellfish and others.
  • the pre-trained panoramic segmentation model (High precision panoptic segmentation, HPPS) includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module.
  • the image to be segmented is semantically segmented using the semantic segmentation branch network to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area.
  • the image to be segmented is subjected to instance segmentation using the instance segmentation branch network to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories.
  • the panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
  • the panoramic fusion module is a parameter-free panoramic fusion module, which selectively attenuates or amplifies the fused logical output score based on the pixel-based head prediction adaptability to adaptively fuse the initial semantic segmentation results and the initial instance segmentation results.
  • the entire panoramic segmentation network is jointly optimized to obtain the final multi-classification and high-precision panoramic segmentation output results of the marine raft aquaculture area, realizing the multi-classification task of the optical remote sensing image marine raft aquaculture area.
  • the image to be segmented is input into the pre-trained panoramic segmentation model
  • the image to be segmented is standardized and pre-processed, and then slidingly cropped into an image of 2048*2048 size, and the cropped images are sequentially input into the trained HPPS model (i.e., the panoramic segmentation model).
  • the output result of the HPPS model is the multi-classification result of the offshore aquaculture area, and all the images are spliced to obtain the overall panoramic segmentation result map of the image to be segmented.
  • the optical remote sensing image marine raft aquaculture area classification method based on panoramic segmentation uses a semantic segmentation branch network and an instance segmentation branch network to parallelly segment the image to be segmented, and fuses the outputs of the two branch networks through a parameter-free panoramic fusion module to finally obtain a multi-classification segmentation result and realize multi-task classification.
  • This method can more completely utilize the logical outputs of the semantic segmentation head and the instance segmentation head to improve the accuracy of multi-classification tasks.
  • the semantic segmentation branch network is an improved U 2 -Net network
  • the semantic segmentation branch network is an improved U 2 -Net network
  • the improved U 2 -Net network includes at least 6 U-shaped secondary encoders and 5 U-shaped secondary decoders, the 6 U-shaped secondary encoders are sequentially 4 first secondary encoders and 2 second secondary encoders, the 5 U-shaped secondary decoders are sequentially 4 first secondary decoders and 1 second secondary decoder;
  • the first secondary encoder and the first secondary decoder are both composed of a first convolution block, an LSFE module, multiple down-sampling modules, a DPC module, a second convolution block, a first convolution block and multiple up-sampling modules in sequence.
  • the LSFE module is used to extract the features of the breeding area within a large field of view, and includes separable convolution and output filters, specifically including two 3 ⁇ 3 separable convolutions and 128 output filters.
  • the DPC module is used to capture remote context information, which includes separable convolution and output channels. Specifically, it includes a 3 ⁇ 3 separable convolution and 256 output channels, and is extended to five parallel branches. Then the outputs of all parallel branches are connected to generate a tensor with 1280 channels, which is finally input to a 1 ⁇ 1 convolution with 256 output channels. The output of the 1 ⁇ 1 convolution is the output of the DPC module.
  • the U 2 -Net network is a saliency detection model, and its specific network structure diagram is shown in FIG3 . It is a two-level nested U-shaped structure.
  • the overall U-shaped structure is referred to as the primary structure, and each small U-shaped structure contained in the primary structure is referred to as a secondary structure. This application does not make improvements on the primary structure, but makes specific improvements on the secondary structure.
  • the improved U 2 -Net network model proposed in this application not only realizes the macro detection of marine aquaculture areas, but also needs to identify, extract and classify each small piece of aquaculture raft.
  • this application improves the secondary U-shaped structure, specifically using a large scale feature extractor (LSFE) module and a dense prediction cell (DPC) module in the secondary structure.
  • LSFE large scale feature extractor
  • DPC dense prediction cell
  • the improved U 2 -Net network provided in the present application is also shown in FIG3 in its structure at this level, which specifically includes 6 U-shaped secondary encoders (i.e., En_1 to En_6) and 5 U-shaped secondary decoders (i.e., De_1 to De_5), wherein, in structure, En_1 corresponds to De_1, En_2 corresponds to De_2, En_3 corresponds to De_3, En_4 corresponds to De_4, and En_5 corresponds to De_5.
  • 6 U-shaped secondary encoders i.e., En_1 to En_6
  • 5 U-shaped secondary decoders i.e., De_1 to De_5
  • En_1 corresponds to De_1
  • En_2 corresponds to De_2
  • En_3 corresponds to De_3
  • En_4 corresponds to De_4
  • En_5 corresponds to De_5.
  • the present application improves the first four substructures of the U 2 -Net network, namely, the structures of En_1, De_1, En_2, De_2, En_3, De_3, En_4, and De_4, while En_5, En_6, and De_5 are not improved and still use the structures in the prior art, which will not be described in detail in the present application.
  • the improved En_1 network structure is composed of 1 first convolution block, 1 LSFE module (i.e., 2 in Figure 4), 4 downsampling modules, 1 DPC module (i.e., 4 in Figure 4), 1 second convolution block, 1 first convolution block, and 5 upsampling modules. That is, the second first convolution block in the prior art is replaced by an LSFE module, the last downsampling module is replaced by a DPC module, and the rest are the same as before. Since the structural difference between En_2, En_3, and En_4 in the prior art and the aforementioned En_1 is the decrease in the number of downsampling modules and the number of upsampling modules.
  • the improved En_2 network structure is composed of 1 first convolution block, 1 LSFE module, 3 downsampling modules, 1 DPC module, 1 second convolution block, 1 first convolution block, and 4 upsampling modules.
  • the improved En_3 network structure is composed of 1 first convolution block, 1 LSFE module, 2 downsampling modules, 1 DPC module, 1 second convolution block, 1 first convolution block and 3 upsampling modules.
  • the improved En_4 network structure is composed of 1 first convolution block, 1 LSFE module, 1 downsampling module, 1 DPC module, 1 second convolution block, 1 first convolution block and 2 upsampling modules.
  • De_1, De_2, De_3, and De_4 correspond one-to-one to the aforementioned improved En_1, En_2, En_3, and En_4, and will not be described in detail here.
  • This application adopts binary classification semantic segmentation, and its corresponding data set labels only include label 0 (non-raft aquaculture area) and label 1 (raft aquaculture area).
  • the loss function involved in the semantic segmentation branch network adopts the binary cross entropy loss function.
  • the embodiment of the present application provides a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images.
  • the LSFE module is used to extract the features of aquaculture areas within a large field of view to achieve macroscopic detection of marine aquaculture areas.
  • the DPC module is used to capture remote context information to identify, extract and classify each small aquaculture raft, thereby improving the segmentation accuracy of the semantic segmentation branch network.
  • the instance segmentation branch network includes an improved SOTR network
  • the improved SOTR network includes at least a Transformer module; wherein the Transformer module includes a separable convolution and an iABN synchronization layer; and the Transformer module is used to predict each instance category.
  • SOTR uses Transformer to simplify the segmentation process, using two parallel subtasks: 1) predicting each instance category through Transformer; 2) dynamically generating segmentation masks using a multi-level upsampling module.
  • the encoder-decoder Transformer model unifies the instance segmentation tasks through a series of learnable mask embeddings.
  • This application extends the Transformer with separable convolution and iABN (inplace activated batch normalization) synchronization layers, which improves the segmentation accuracy and training convergence to a certain extent.
  • the method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation improves the segmentation accuracy and training convergence of the model through the expansion of separable convolution and iABN synchronization layers.
  • the instance segmentation branch network also includes a feature extraction module, which includes a moving reverse bottleneck unit and a bidirectional feature pyramid network (Feature pyramid network, FPN).
  • a feature extraction module which includes a moving reverse bottleneck unit and a bidirectional feature pyramid network (Feature pyramid network, FPN).
  • FPN feature pyramid network
  • the feature extraction module is used to extract features from the image to be segmented to obtain multi-scale features.
  • instance segmentation is performed on the image to be segmented through the improved SOTR network to obtain an initial instance segmentation result.
  • the method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation realizes the extraction of multi-scale features by moving the reverse bottleneck unit and the bidirectional feature pyramid network, and can obtain more large-scale features, small-scale features, shallow information, and deep information to improve the accuracy of instance segmentation.
  • FIG5 is a schematic diagram of the training process of the panoramic segmentation model provided in an embodiment of the present application; as shown in FIG5 , the pre-trained panoramic segmentation model is trained in the following manner:
  • the labels include semantic labels of raft aquaculture areas and non-raft aquaculture areas and instance labels of various aquaculture area categories.
  • the training data set needs to be annotated.
  • the specific annotations are divided into “stuff” and “thing” classes.
  • the "stuff” class is annotated with a semantic mask
  • 0 represents a non-raft aquaculture area
  • 1 represents a raft aquaculture area
  • the "thing” class is annotated with an instance mask, including four instances of fish, algae, shellfish, and others.
  • semantic labels of the background (i.e., non-raft aquaculture area) and foreground (i.e., raft aquaculture area) are created for the training set and the test set, and on this basis, instance labels of instance 1 (fish), instance 2 (algae), instance 3 (shellfish), and instance 4 (others) are created.
  • the training data set is input into the semantic segmentation branch network, a training semantic segmentation result is predicted, and a loss between the training semantic segmentation result and the semantic label is calculated to obtain a first loss.
  • the training data set is input into the instance segmentation branch network, the training instance segmentation result is predicted, and the loss between the training instance segmentation result and the instance label is calculated to obtain a second loss.
  • the panoramic fusion module is used to adaptively fuse the training semantic segmentation result and the training instance segmentation result to obtain a training multi-classification result.
  • a total loss is obtained according to the first loss and the second loss, and the panoramic segmentation model is trained based on the training multi-classification result and the total loss until the panoramic segmentation model converges to obtain a trained panoramic segmentation model.
  • the total loss is obtained by adaptively weighting the first loss and the second loss according to the logical output score of the attenuated or amplified fusion.
  • this application trains the panoramic segmentation model by sharing the synthetic data set, and after the panoramic segmentation model reaches convergence accuracy, the panoramic segmentation model is tested by the test set to obtain qualitative evaluation results.
  • the model is trained again by adjusting hyperparameters, supplementing training samples, etc. until it meets the qualitative evaluation requirements.
  • the model is then tested with the test set and the panoramic segmentation accuracy PQ of the model is evaluated to finally obtain a trained panoramic segmentation model.
  • the embodiment of the present application provides a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation.
  • the dataset labels mentioned above not only classify and label the raft aquaculture areas and background areas as a whole, but also further classify and label the various aquaculture categories in the raft aquaculture areas in a refined manner to achieve multi-classification tasks.
  • the method further includes:
  • the normalized vegetation index feature (Normalized Difference Vegetation Index, NDVI) and the normalized water index feature (Normalized Difference Water Index, NDWI) of the training data set are constructed respectively.
  • the normalized vegetation index feature and the normalized water body index feature are fused with the training data set to obtain a shared synthetic data set.
  • inputting the training data set into the semantic segmentation branch network includes:
  • the shared synthetic dataset is input into the semantic segmentation branch network.
  • the step of inputting the training data set into the instance segmentation branch network comprises:
  • the shared synthetic dataset is input into the instance segmentation branch network.
  • the normalized vegetation index feature and the normalized water index feature can also be fused with the image to be segmented after the image to be segmented is obtained, so as to segment the fused image using the panoramic segmentation model.
  • the method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation obtains a shared synthetic data set by fusing two custom features, NDVI and NDWI, into a training data set, thereby being able to more fully and effectively utilize the rich variety of information in optical remote sensing images to improve the segmentation accuracy of the panoramic segmentation model.
  • the training data set includes a labeled data set and an adversarial sample set
  • the labeled data set is a data set with corresponding labels after being annotated
  • the adversarial sample set is obtained by performing adversarial training on the training instance segmentation results.
  • the multi-classification data label set of the marine aquaculture area is used to perform adversarial training on the instance segmentation branch to improve the anti-interference ability of multi-target multi-classification.
  • the adversarial samples generated during the adversarial training are added to the training dataset and together with the label dataset constitute the training dataset.
  • adversarial training can be achieved by adding a discriminator or generating new samples based on gradient feedback.
  • the adversarial training method is a conventional method and this application does not limit it.
  • the embodiment of the present application provides a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation.
  • An adversarial sample set is obtained through adversarial training, and a panoramic segmentation model is trained based on the adversarial sample set and a label data set, thereby improving the anti-interference ability of the panoramic segmentation model.
  • the label data set is obtained by:
  • An optical remote sensing image of a training marine aquaculture area is obtained, and the optical remote sensing image is subjected to at least storage format unification, cloud removal, normalization, and cropping.
  • the marine raft aquaculture area classification device based on panoramic segmentation of optical remote sensing images provided in the present application.
  • the marine raft aquaculture area classification device based on panoramic segmentation of optical remote sensing images described below and the marine raft aquaculture area classification method based on panoramic segmentation of optical remote sensing images described above can be referenced to each other.
  • Figure 6 is a structural schematic diagram of the optical remote sensing image marine raft aquaculture area classification device based on panoramic segmentation provided in an embodiment of the present application; as shown in Figure 6, the optical remote sensing image marine raft aquaculture area classification device based on panoramic segmentation includes an image acquisition module 601 and an image segmentation module 602.
  • the image acquisition module 601 is used to acquire the image to be segmented, where the image to be segmented is an optical remote sensing image of a marine aquaculture area.
  • the image segmentation module 602 is used to input the image to be segmented into a pre-trained panoramic image.
  • segmentation model multi-classification segmentation results are predicted.
  • the multi-classification segmentation results include raft aquaculture areas, non-raft aquaculture areas and multiple aquaculture area categories, and the multiple aquaculture area categories include fish, algae, shellfish and others.
  • the pre-trained panoramic segmentation model (High precision panoptic segmentation, HPPS) includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module.
  • the image to be segmented is semantically segmented using the semantic segmentation branch network to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area.
  • the image to be segmented is subjected to instance segmentation using the instance segmentation branch network to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories.
  • the panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
  • the panoramic fusion module is a parameter-free panoramic fusion module, which selectively attenuates or amplifies the fused logical output score based on the pixel-based head prediction adaptability to adaptively fuse the initial semantic segmentation results and the initial instance segmentation results.
  • the entire panoramic segmentation network is jointly optimized to obtain the final multi-classification and high-precision panoramic segmentation output results of the marine raft aquaculture area, realizing the multi-classification task of the optical remote sensing image marine raft aquaculture area.
  • the image to be segmented is input into the pre-trained panoramic segmentation model
  • the image to be segmented is standardized and pre-processed, and then slidingly cropped into an image of 2048*2048 size, and the cropped images are sequentially input into the trained HPPS model (i.e., the panoramic segmentation model).
  • the output result of the HPPS model is the multi-classification result of the offshore aquaculture area, and all the images are spliced to obtain the overall panoramic segmentation result map of the image to be segmented.
  • the optical remote sensing image marine raft aquaculture area classification device based on panoramic segmentation uses a semantic segmentation branch network and an instance segmentation branch network to parallelly segment the segmented image, and fuses the outputs of the two branch networks through a parameter-free panoramic fusion module, and finally obtains a multi-classification segmentation result to achieve multi-task classification.
  • the adaptive fusion method of the panoramic fusion module can more completely utilize the logical outputs of the semantic segmentation head and the instance segmentation head to improve the accuracy of multi-classification tasks.
  • the above-described device embodiments are merely illustrative, and the components described as separate components are The units may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of this embodiment. Those of ordinary skill in the art may understand and implement it without creative work.
  • each implementation method can be implemented by means of software plus a necessary general hardware platform, and of course, it can also be implemented by hardware.
  • the above technical solution is essentially or the part that contributes to the prior art can be embodied in the form of a software product, and the computer software product can be stored in a computer-readable storage medium, such as ROM/RAM, a disk, an optical disk, etc., including a number of instructions for a computer device (which can be a personal computer, a server, or a network device, etc.) to execute the methods of each embodiment or some parts of the embodiment.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

Provided in the present application is a panoptic segmentation-based optical remote-sensing image raft mariculture area classification method, comprising: acquiring an image to be segmented, said image being an optical remote-sensing image of a mariculture area (S101); and inputting said image into a pre-trained panoptic segmentation model to predict a multi-classification segmentation result (S102), in which semantic segmentation is performed on said image by using a semantic segmentation branch network so as to obtain an initial semantic segmentation result, instance segmentation is performed on said image by using an instance segmentation branch network so as to obtain an initial instance segmentation result, and the initial semantic segmentation result and the initial instance segmentation result are fused by using a panoptic fusion module so as to obtain the multi-classification segmentation result. By effectively using a variety of rich information in remote-sensing images, the present application can implement high-precision multi-classification recognition tasks of raft mariculture areas, and improve the segmentation precision of panoptic segmentation models.

Description

基于全景分割的光学遥感影像海洋筏式养殖区分类方法Classification method of marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本申请要求于2022年10月27日提交的申请号为202211328346.5,发明名称为“基于全景分割的光学遥感影像海洋筏式养殖区分类方法”的中国专利申请的优先权,其通过引用方式全部并入本文。This application claims priority to Chinese patent application No. 202211328346.5, filed on October 27, 2022, and entitled “Classification method for marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images”, which is incorporated herein by reference in its entirety.
技术领域Technical Field
本申请涉及海洋遥感和图像处理技术领域,尤其涉及一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法。The present application relates to the field of marine remote sensing and image processing technology, and in particular to a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images.
背景技术Background technique
海洋筏式养殖是海洋养殖的重要组成部分,相对于靠近海岸的池塘及滩涂养殖,筏式养殖范围广且区域分散,采用传统的现场测量方法,不仅费时费力而且难以得到大区域准确的结果。遥感技术的发展则极大地弥补了传统地面测量覆盖范围小、数据获取效率低等缺点。同时,利用深度学习方法实现遥感影像的智能信息提取可以快速、准确地获取海洋养殖区的分布和养殖类型信息,是开展海洋筏式养殖动态监测的可靠和先进的技术手段。面向海洋资源开发利用和支持监管的实施,需要在不降低分割精度的前提下对海洋养殖区的精细分类,例如细分为鱼类(网箱)、藻类(延绳)、贝类(浮筏)和其它等。Marine raft aquaculture is an important part of marine aquaculture. Compared with ponds and tidal flats near the coast, raft aquaculture has a wide range and is dispersed in the region. The traditional on-site measurement method is not only time-consuming and labor-intensive, but also difficult to obtain accurate results for large areas. The development of remote sensing technology has greatly made up for the shortcomings of traditional ground measurement, such as small coverage and low data acquisition efficiency. At the same time, the use of deep learning methods to realize intelligent information extraction of remote sensing images can quickly and accurately obtain the distribution and aquaculture type information of marine aquaculture areas, which is a reliable and advanced technical means for dynamic monitoring of marine raft aquaculture. In order to develop and utilize marine resources and support the implementation of supervision, it is necessary to finely classify marine aquaculture areas without reducing the segmentation accuracy, such as subdividing them into fish (cages), algae (longlines), shellfish (rafts) and others.
合成孔径雷达(SAR)具有全天时、全天候的特点,在遥感领域得到较多的应用,但是SAR图像具有分辨率较低、容易受到噪声的干扰、几何畸变严重、可用特征较少等缺点。光学遥感应用范围广,光学遥感对地物的边界信息描述清晰,同时包含丰富的光谱信息,有利于筏式养殖边界及养殖类型信息提取。但是部分光学影像存在一定云、雾和光照的干扰,这些干扰因素制约了光学遥感图像特征信息的提取,加大了光学遥感图像中目标识别和分割的难度。Synthetic aperture radar (SAR) has the characteristics of all-day and all-weather, and has been widely used in the field of remote sensing. However, SAR images have disadvantages such as low resolution, susceptibility to noise interference, severe geometric distortion, and fewer available features. Optical remote sensing has a wide range of applications. Optical remote sensing clearly describes the boundary information of ground objects and contains rich spectral information, which is conducive to the extraction of raft aquaculture boundary and aquaculture type information. However, some optical images are interfered by clouds, fog and light. These interference factors restrict the extraction of feature information of optical remote sensing images and increase the difficulty of target recognition and segmentation in optical remote sensing images.
现有的基于卷积神经网络的海水养殖区域的提取主要分为语义分割和实例分割,例如,改进的SOLO、D-ResUnet、HCHNet等分割算法。语义 分割和实例分割都属于像素级分类,其中,在语义分割模型训练过程中对每个像素点的预测值通过Softmax函数映射到[0,1]的概率值,然后通过交叉熵损失函数判断预测值与真实标签值的误差,通过梯度下降法不断的训练模型使两者之间的误差达到极小值。语义分割的目标种类越多,即数据集标签越多,在对每个目标进行识别和分割时受到的干扰项越多,其反映到数学模型上时,每个样本的预测值的概率分布越分散,概率分布的方差越大,此时增大了样本预测概率分布聚焦到一个标签值上的难度,使得损失函数的收敛速度降低,同时降低了分割和识别的精准度。由以上分析可得,语义分割任务中多种类识别和高精度分割存在相互制约的关系,因此,简单的目标检测、识别和分割无法完成对海洋筏式养殖区的精细化分类。The existing convolutional neural network-based marine aquaculture area extraction is mainly divided into semantic segmentation and instance segmentation, for example, improved SOLO, D-ResUnet, HCHNet and other segmentation algorithms. Segmentation and instance segmentation are both pixel-level classification. In the process of semantic segmentation model training, the predicted value of each pixel is mapped to a probability value of [0,1] through the Softmax function, and then the error between the predicted value and the true label value is judged by the cross entropy loss function. The model is continuously trained by the gradient descent method to minimize the error between the two. The more types of semantic segmentation targets, that is, the more data set labels, the more interference items are encountered when identifying and segmenting each target. When reflected in the mathematical model, the more dispersed the probability distribution of the predicted value of each sample is, the greater the variance of the probability distribution is. At this time, the difficulty of focusing the sample prediction probability distribution on a label value is increased, which reduces the convergence speed of the loss function and reduces the accuracy of segmentation and recognition. From the above analysis, it can be seen that there is a mutual constraint between multi-class recognition and high-precision segmentation in the semantic segmentation task. Therefore, simple target detection, recognition and segmentation cannot complete the refined classification of marine raft aquaculture areas.
现有的用于筏式养殖区提取的卷积神经网络模型的数据标签大多只适合于语义或者实例分割的单任务模型训练,缺少用于区分不同养殖类型的多分类的海洋养殖区全景分割的数据标签。Most of the data labels of existing convolutional neural network models used for raft aquaculture area extraction are only suitable for single-task model training of semantic or instance segmentation, and lack data labels for multi-classification panoramic segmentation of marine aquaculture areas to distinguish different aquaculture types.
综上所述,亟需一种能够实现多分类的海洋养殖区全景分割方法。In summary, there is an urgent need for a panoramic segmentation method of marine aquaculture areas that can achieve multi-classification.
发明内容Summary of the invention
本申请提供一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,用以解决上述问题。The present application provides a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation to solve the above problems.
本申请提供一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,包括:The present application provides a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images, including:
获取待分割图像,所述待分割图像为海洋养殖区的光学遥感影像;Acquire an image to be segmented, where the image to be segmented is an optical remote sensing image of a marine aquaculture area;
将所述待分割图像输入至预先训练好的全景分割模型中,预测得到多分类分割结果,所述多分类分割结果包括筏式养殖区、非筏式养殖区以及多种养殖区类别;Inputting the image to be segmented into a pre-trained panoramic segmentation model to predict a multi-classification segmentation result, wherein the multi-classification segmentation result includes raft aquaculture area, non-raft aquaculture area and multiple aquaculture area categories;
其中,所述预先训练好的全景分割模型包括语义分割分支网络、实例分割分支网络以及全景融合模块;The pre-trained panoramic segmentation model includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module;
利用所述语义分割分支网络对所述待分割图像进行语义分割,以获得初始语义分割结果,所述初始语义分割结果包括初始筏式养殖区与初始非筏式养殖区;Using the semantic segmentation branch network to perform semantic segmentation on the image to be segmented to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area;
利用所述实例分割分支网络对所述待分割图像进行实例分割,以获得 初始实例分割结果,所述初始实例分割结果包括多种初始养殖区类别;The instance segmentation branch network is used to perform instance segmentation on the image to be segmented to obtain An initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories;
利用所述全景融合模块对所述初始语义分割结果以及所述初始实例分割结果进行融合,以获得多分类分割结果。The panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,所述语义分割分支网络为改进的U2-Net网络,所述改进的U2-Net网络至少包括6个U型结构的次级编码器以及5个U型结构的次级解码器,所述6个U型结构的次级编码器依次为4个第一次级编码器以及2个第二次级编码器,所述5个U型结构的次级解码器依次为4个第一次级解码器以及1个第二次级解码器;所述第一次级编码器以及所述第一次级解码器均由第一卷积块、LSFE模块、多个下采样模块、DPC模块、第二卷积块、第一卷积块以及多个上采样模块依次构成;According to a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided by the present application, the semantic segmentation branch network is an improved U 2 -Net network, and the improved U 2 -Net network includes at least 6 U-shaped secondary encoders and 5 U-shaped secondary decoders, the 6 U-shaped secondary encoders are sequentially 4 first secondary encoders and 2 second secondary encoders, and the 5 U-shaped secondary decoders are sequentially 4 first secondary decoders and 1 second secondary decoder; the first secondary encoder and the first secondary decoder are both composed of a first convolution block, an LSFE module, a plurality of down-sampling modules, a DPC module, a second convolution block, a first convolution block, and a plurality of up-sampling modules in sequence;
其中,所述LSFE模块用于提取大视野范围内养殖区的特征,其包括可分离卷积以及输出滤波器;The LSFE module is used to extract the features of the breeding area within a large field of view, and includes a separable convolution and an output filter;
所述DPC模块用于捕获远程上下文信息,其包括可分离卷积以及输出通道。The DPC module is used to capture long-range context information, which includes separable convolution and output channels.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,所述实例分割分支网络包括改进的SOTR网络,所述改进的SOTR网络至少包括Transformer模块;其中,所述Transformer模块包括可分离卷积以及iABN同步层;所述Transformer模块用于预测每个实例类别。According to a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided in the present application, the instance segmentation branch network includes an improved SOTR network, and the improved SOTR network includes at least a Transformer module; wherein the Transformer module includes a separable convolution and an iABN synchronization layer; and the Transformer module is used to predict the category of each instance.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,所述实例分割分支网络还包括特征提取模块,所述特征提取模块包括移动反向瓶颈单元以及双向特征金字塔网络;According to a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided by the present application, the instance segmentation branch network also includes a feature extraction module, and the feature extraction module includes a mobile reverse bottleneck unit and a bidirectional feature pyramid network;
利用所述特征提取模块对所述待分割图像进行特征提取,以获得多尺度特征;Using the feature extraction module to extract features from the image to be segmented to obtain multi-scale features;
基于所述多尺度特征,通过所述改进的SOTR网络对所述待分割图像进行实例分割,以获得初始实例分割结果。Based on the multi-scale features, instance segmentation is performed on the image to be segmented through the improved SOTR network to obtain an initial instance segmentation result.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,所述预先训练好的全景分割模型通过如下方式训练得到:According to a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided in the present application, the pre-trained panoramic segmentation model is trained in the following manner:
获取训练数据集以及其对应的标签,并构建全景分割模型;其中,所 述标签包括筏式养殖区与非筏式养殖区的语义标签以及多种养殖区类别的实例标签;Obtain the training data set and its corresponding labels, and build a panoramic segmentation model; The labels include semantic labels of raft aquaculture areas and non-raft aquaculture areas and instance labels of various aquaculture area categories;
将所述训练数据集输入所述语义分割分支网络,预测得到训练用语义分割结果,计算训练用语义分割结果与所述语义标签之间的损失,获得第一损失;Inputting the training data set into the semantic segmentation branch network, predicting a training semantic segmentation result, calculating the loss between the training semantic segmentation result and the semantic label, and obtaining a first loss;
将所述训练数据集输入所述实例分割分支网络,预测得到训练用实例分割结果,计算训练用实例分割结果与所述实例标签之间的损失,获得第二损失;Inputting the training data set into the instance segmentation branch network, predicting a training instance segmentation result, calculating the loss between the training instance segmentation result and the instance label, and obtaining a second loss;
利用所述全景融合模块对所述训练用语义分割结果与所述训练用实例分割结果进行自适应融合,获得训练用多分类结果;Using the panoramic fusion module to adaptively fuse the training semantic segmentation result and the training instance segmentation result to obtain a training multi-classification result;
根据所述第一损失以及所述第二损失获取总损失,基于所述训练用多分类结果以及所述总损失对所述全景分割模型进行训练,直至所述全景分割模型收敛,以获得训练好的全景分割模型。A total loss is obtained according to the first loss and the second loss, and the panoramic segmentation model is trained based on the training multi-classification result and the total loss until the panoramic segmentation model converges to obtain a trained panoramic segmentation model.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,在所述获取训练数据集以及其对应的标签之后,该方法还包括:According to a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided by the present application, after obtaining the training data set and its corresponding labels, the method further includes:
分别构建所述训练数据集的归一化植被指数特征和归一化水体指数特征;Respectively constructing the normalized vegetation index feature and the normalized water index feature of the training data set;
将所述归一化植被指数特征和归一化水体指数特征与所述训练数据集进行融合,以获得共享合成数据集;fusing the normalized vegetation index feature and the normalized water index feature with the training data set to obtain a shared synthetic data set;
相应地,所述将所述训练数据集输入所述语义分割分支网络,包括:Accordingly, inputting the training data set into the semantic segmentation branch network includes:
将所述共享合成数据集输入所述语义分割分支网络;Inputting the shared synthetic dataset into the semantic segmentation branch network;
所述将所述训练数据集输入所述实例分割分支网络,包括:The step of inputting the training data set into the instance segmentation branch network comprises:
将所述共享合成数据集输入所述实例分割分支网络。The shared synthetic dataset is input into the instance segmentation branch network.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,所述训练数据集包括标签数据集以及对抗样本集,所述标签数据集为标注后具有对应标签的数据集,所述对抗样本集通过对所述训练用实例分割结果进行对抗训练得到。According to a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided in the present application, the training data set includes a label data set and an adversarial sample set. The label data set is a data set with corresponding labels after annotation, and the adversarial sample set is obtained by performing adversarial training on the segmentation results of the training instances.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,所述标签数据集通过如下方式获得:According to a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in this application, the label data set is obtained in the following manner:
获取训练用海洋养殖区的光学遥感影像,对所述光学遥感影像至少执 行存储格式统一、去云雾处理、归一化处理以及裁剪处理。Obtain an optical remote sensing image of a marine aquaculture area for training, and perform at least The row storage format is unified, and the cloud removal, normalization and cropping processes are performed.
根据本申请提供的一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,所述多种养殖区类别至少包括鱼类、藻类以及贝类。According to a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in the present application, the multiple aquaculture area categories include at least fish, algae and shellfish.
本申请还提供一种基于全景分割的光学遥感影像海洋筏式养殖区分类装置,包括:The present application also provides a device for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images, comprising:
图像获取模块,用于获取待分割图像,所述待分割图像为海洋养殖区的光学遥感影像;An image acquisition module is used to acquire an image to be segmented, wherein the image to be segmented is an optical remote sensing image of a marine aquaculture area;
图像分割模块,用于将所述待分割图像输入至预先训练好的全景分割模型中,预测得到多分类分割结果,所述多分类分割结果包括筏式养殖区、非筏式养殖区以及多种养殖区类别;An image segmentation module is used to input the image to be segmented into a pre-trained panoramic segmentation model to predict a multi-classification segmentation result, wherein the multi-classification segmentation result includes raft aquaculture area, non-raft aquaculture area and multiple aquaculture area categories;
其中,所述预先训练好的全景分割模型包括语义分割分支网络、实例分割分支网络以及全景融合模块;The pre-trained panoramic segmentation model includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module;
利用所述语义分割分支网络对所述待分割图像进行语义分割,以获得初始语义分割结果,所述初始语义分割结果包括初始筏式养殖区与初始非筏式养殖区;Using the semantic segmentation branch network to perform semantic segmentation on the image to be segmented to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area;
利用所述实例分割分支网络对所述待分割图像进行实例分割,以获得初始实例分割结果,所述初始实例分割结果包括多种初始养殖区类别;Using the instance segmentation branch network to perform instance segmentation on the image to be segmented to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories;
利用所述全景融合模块对所述初始语义分割结果以及所述初始实例分割结果进行融合,以获得多分类分割结果。The panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
本申请提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,通过语义分割分支网络与实例分割分支网络对待分割图像并行分割,并将两个分支网络的输出通过无参数的全景融合模块进行融合,最后获得多分类分割结果,实现多任务分类。且全景融合模块这种自适应融合方式能够更加完整利用语义分割头和实例分割头的逻辑输出,提升多分类任务的准确性。The method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided in this application uses a semantic segmentation branch network and an instance segmentation branch network to parallelly segment the segmented image, and fuses the outputs of the two branch networks through a parameter-free panoramic fusion module, and finally obtains a multi-classification segmentation result to achieve multi-task classification. Moreover, the adaptive fusion method of the panoramic fusion module can more completely utilize the logical outputs of the semantic segmentation head and the instance segmentation head to improve the accuracy of multi-classification tasks.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚地说明本申请或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本申请的一些实施例,对于本领域普通技术人员来讲,在 不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the present application or the prior art, the following briefly introduces the drawings required for use in the embodiments or the description of the prior art. Obviously, the drawings described below are some embodiments of the present application. For ordinary technicians in this field, Other drawings can be obtained based on these drawings without any creative work.
图1是本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法的流程示意图之一;FIG1 is a schematic diagram of a flow chart of a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application;
图2是本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法的流程示意图之二;FIG2 is a second flow chart of a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application;
图3是现有的U2-Net网络的结构示意图;FIG3 is a schematic diagram of the structure of an existing U 2 -Net network;
图4是现有的U2-Net网络中En_1次级结构与本申请改进后En_1次级结构的结构对比图;FIG4 is a structural comparison diagram of the En_1 substructure in the existing U 2 -Net network and the improved En_1 substructure of the present application;
图5是本申请实施例提供的全景分割模型的训练过程示意图;FIG5 is a schematic diagram of the training process of the panoptic segmentation model provided in an embodiment of the present application;
图6是本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类装置的结构示意图。FIG6 is a schematic diagram of the structure of a device for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application.
具体实施方式Detailed ways
为使本申请的目的、技术方案和优点更加清楚,下面将结合本申请中的附图,对本申请中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the purpose, technical solutions and advantages of this application clearer, the technical solutions in this application will be clearly and completely described below in conjunction with the drawings in this application. Obviously, the described embodiments are part of the embodiments of this application, not all of them. Based on the embodiments in this application, all other embodiments obtained by ordinary technicians in this field without creative work are within the scope of protection of this application.
基于上述提到的现有的卷积神经网络模型只能实现语义或者实例分割的单任务,且面对复杂多变的海洋环境,需要更加充分、有效地利用光学遥感影像中丰富的各种信息,因此,本申请通过统一语义分割和实例分割预测子网络,并将输出融合形成整体全景分割网络模型;构建海洋筏式养殖区多分类任务的全景分割标签数据集,从而实现多分类任务,并使海洋筏式养殖区的分类更加精细化,提升模型的分割精度。下面结合附图对本申请提出的基于全景分割的光学遥感影像海洋筏式养殖区分类方法进行具体说明。Based on the existing convolutional neural network models mentioned above, they can only realize single tasks of semantic or instance segmentation, and in the face of complex and changeable marine environments, it is necessary to make more full and effective use of the rich information in optical remote sensing images. Therefore, this application unifies the semantic segmentation and instance segmentation prediction subnetworks, and fuses the outputs to form an overall panoramic segmentation network model; constructs a panoramic segmentation label dataset for the multi-classification task of marine raft aquaculture areas, thereby realizing multi-classification tasks, and making the classification of marine raft aquaculture areas more refined, and improving the segmentation accuracy of the model. The following is a specific description of the optical remote sensing image marine raft aquaculture area classification method based on panoramic segmentation proposed in this application in conjunction with the accompanying drawings.
图1是本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法的流程示意图之一;图2是本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法的流程示意图之二。Figure 1 is one of the flow charts of the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application; Figure 2 is the second flow chart of the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images provided in an embodiment of the present application.
如图1以及2所示,该基于全景分割的光学遥感影像海洋筏式养殖区 分类方法,包括:As shown in Figures 1 and 2, the optical remote sensing image based on panoramic segmentation of the marine raft aquaculture area Classification methods include:
S101,获取待分割图像,所述待分割图像为海洋养殖区的光学遥感影像。S101, obtaining an image to be segmented, wherein the image to be segmented is an optical remote sensing image of a marine aquaculture area.
S102,将所述待分割图像输入至预先训练好的全景分割模型中,预测得到多分类分割结果。S102, inputting the image to be segmented into a pre-trained panoramic segmentation model to predict a multi-classification segmentation result.
其中,所述多分类分割结果包括筏式养殖区、非筏式养殖区以及多种养殖区类别,多种养殖区类别又包括鱼类、藻类、贝类以及其他。The multi-classification segmentation results include raft aquaculture areas, non-raft aquaculture areas and multiple aquaculture area categories, and the multiple aquaculture area categories include fish, algae, shellfish and others.
所述预先训练好的全景分割模型(High precision panoptic segmentation,HPPS)包括语义分割分支网络、实例分割分支网络以及全景融合模块。The pre-trained panoramic segmentation model (High precision panoptic segmentation, HPPS) includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module.
利用所述语义分割分支网络对所述待分割图像进行语义分割,以获得初始语义分割结果,所述初始语义分割结果包括初始筏式养殖区与初始非筏式养殖区。The image to be segmented is semantically segmented using the semantic segmentation branch network to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area.
利用所述实例分割分支网络对所述待分割图像进行实例分割,以获得初始实例分割结果,所述初始实例分割结果包括多种初始养殖区类别。The image to be segmented is subjected to instance segmentation using the instance segmentation branch network to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories.
利用所述全景融合模块对所述初始语义分割结果以及所述初始实例分割结果进行融合,以获得多分类分割结果。The panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
其中,全景融合模块为无参数全景融合模块,其基于像素的头部预测适应性,有选择地衰减或放大融合的逻辑输出分数来自适应融合初始语义分割结果与初始实例分割结果。通过这种端到端方式联合优化整个全景分割网络,从而获得最终的海洋筏式养殖区的多分类高精度的全景分割输出结果,实现了光学遥感影像海洋筏式养殖区多分类任务。Among them, the panoramic fusion module is a parameter-free panoramic fusion module, which selectively attenuates or amplifies the fused logical output score based on the pixel-based head prediction adaptability to adaptively fuse the initial semantic segmentation results and the initial instance segmentation results. Through this end-to-end approach, the entire panoramic segmentation network is jointly optimized to obtain the final multi-classification and high-precision panoramic segmentation output results of the marine raft aquaculture area, realizing the multi-classification task of the optical remote sensing image marine raft aquaculture area.
另外,在将所述待分割图像输入至预先训练好的全景分割模型中之前,对待分割图像进行标准化预处理,然后滑动裁剪成2048*2048大小的图像,将裁剪后的图像依次输入训练好的HPPS模型(即全景分割模型),该HPPS模型的输出结果即为海上养殖区域的多分类结果,对所有的图像进行拼接获得待分割图像的整体的全景分割结果图。In addition, before the image to be segmented is input into the pre-trained panoramic segmentation model, the image to be segmented is standardized and pre-processed, and then slidingly cropped into an image of 2048*2048 size, and the cropped images are sequentially input into the trained HPPS model (i.e., the panoramic segmentation model). The output result of the HPPS model is the multi-classification result of the offshore aquaculture area, and all the images are spliced to obtain the overall panoramic segmentation result map of the image to be segmented.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,通过语义分割分支网络与实例分割分支网络对待分割图像并行分割,并将两个分支网络的输出通过无参数的全景融合模块进行融合,最后获得多分类分割结果,实现多任务分类。且全景融合模块这种自适应融合 方式能够更加完整利用语义分割头和实例分割头的逻辑输出,提升多分类任务的准确性。The optical remote sensing image marine raft aquaculture area classification method based on panoramic segmentation provided in the embodiment of the present application uses a semantic segmentation branch network and an instance segmentation branch network to parallelly segment the image to be segmented, and fuses the outputs of the two branch networks through a parameter-free panoramic fusion module to finally obtain a multi-classification segmentation result and realize multi-task classification. This method can more completely utilize the logical outputs of the semantic segmentation head and the instance segmentation head to improve the accuracy of multi-classification tasks.
进一步地,所述语义分割分支网络为改进的U2-Net网络,所述语义分割分支网络为改进的U2-Net网络,所述改进的U2-Net网络至少包括6个U型结构的次级编码器以及5个U型结构的次级解码器,所述6个U型结构的次级编码器依次为4个第一次级编码器以及2个第二次级编码器,所述5个U型结构的次级解码器依次为4个第一次级解码器以及1个第二次级解码器;所述第一次级编码器以及所述第一次级解码器均由第一卷积块、LSFE模块、多个下采样模块、DPC模块、第二卷积块、第一卷积块以及多个上采样模块依次构成。Furthermore, the semantic segmentation branch network is an improved U 2 -Net network, the semantic segmentation branch network is an improved U 2 -Net network, the improved U 2 -Net network includes at least 6 U-shaped secondary encoders and 5 U-shaped secondary decoders, the 6 U-shaped secondary encoders are sequentially 4 first secondary encoders and 2 second secondary encoders, the 5 U-shaped secondary decoders are sequentially 4 first secondary decoders and 1 second secondary decoder; the first secondary encoder and the first secondary decoder are both composed of a first convolution block, an LSFE module, multiple down-sampling modules, a DPC module, a second convolution block, a first convolution block and multiple up-sampling modules in sequence.
其中,所述LSFE模块用于提取大视野范围内养殖区的特征,其包括可分离卷积以及输出滤波器。具体包括两个3×3可分离的卷积以及128个输出滤波器。The LSFE module is used to extract the features of the breeding area within a large field of view, and includes separable convolution and output filters, specifically including two 3×3 separable convolutions and 128 output filters.
所述DPC模块用于捕获远程上下文信息,其包括可分离卷积以及输出通道。具体包括一个3×3可分离卷积和256个输出通道,并扩展到五个并行分支,然后将所有并行分支的输出连接起来,以产生具有1280个通道的张量,该张量最终输入至具有256个输出通道的1×1卷积,该1×1卷积的输出即为DPC模块的输出。The DPC module is used to capture remote context information, which includes separable convolution and output channels. Specifically, it includes a 3×3 separable convolution and 256 output channels, and is extended to five parallel branches. Then the outputs of all parallel branches are connected to generate a tensor with 1280 channels, which is finally input to a 1×1 convolution with 256 output channels. The output of the 1×1 convolution is the output of the DPC module.
现有技术中,U2-Net网络是一种显著性检测模型,其具体网络结构图如图3所示,它是两级嵌套的U型结构,本申请将整体的U型结构称为本级结构,将本级结构内包含的每一个小的U型结构称为次级结构,本申请未在本级结构上进行改进,而是在次级结构上进行了具体的改进。In the prior art, the U 2 -Net network is a saliency detection model, and its specific network structure diagram is shown in FIG3 . It is a two-level nested U-shaped structure. In this application, the overall U-shaped structure is referred to as the primary structure, and each small U-shaped structure contained in the primary structure is referred to as a secondary structure. This application does not make improvements on the primary structure, but makes specific improvements on the secondary structure.
其中,改进的具体构思为:由于遥感影像是从卫星平台对海洋海岸大范围区域对地观测,视角宽广,数据庞大,它宏观综合地反映海洋养殖区的分布状况。因此,本申请提出的改进U2-Net网络模型不仅要实现对海洋养殖区的宏观检测,而且需要对每一小块养殖浮筏进行识别、提取和分类。为了实现这一目标,本申请对次级U型结构进行了改进,具体是在次级结构中使用了大型特征提取器(Large Scale Feature Extractor,LSFE)模块以及密集预测单元(Dense Prediction Cells,DPC)模块。 The specific concept of the improvement is: since remote sensing images are observations of a large area of the ocean coast from a satellite platform, with a wide viewing angle and huge data, they reflect the distribution of marine aquaculture areas in a macro and comprehensive manner. Therefore, the improved U 2 -Net network model proposed in this application not only realizes the macro detection of marine aquaculture areas, but also needs to identify, extract and classify each small piece of aquaculture raft. In order to achieve this goal, this application improves the secondary U-shaped structure, specifically using a large scale feature extractor (LSFE) module and a dense prediction cell (DPC) module in the secondary structure.
下面从整体的网络结构出发,对改进的U2-Net网络进行具体描述。The following describes the improved U 2 -Net network in detail based on the overall network structure.
首先,本申请提供的改进的U2-Net网络其在本级结构上也如图3所示,其具体包括6个U型结构的次级编码器(即En_1~En_6)以及5个U型结构的次级解码器(即De_1~De_5),其中,在结构上En_1与De_1对应,En_2与De_2对应,En_3与De_3对应,En_4与De_4对应,En_5与De_5对应。First, the improved U 2 -Net network provided in the present application is also shown in FIG3 in its structure at this level, which specifically includes 6 U-shaped secondary encoders (i.e., En_1 to En_6) and 5 U-shaped secondary decoders (i.e., De_1 to De_5), wherein, in structure, En_1 corresponds to De_1, En_2 corresponds to De_2, En_3 corresponds to De_3, En_4 corresponds to De_4, and En_5 corresponds to De_5.
本申请对U2-Net网络的前四个次级结构进行改进,也即对En_1、De_1、En_2、De_2、En_3、De_3、En_4、De_4的结构进行改进,而对En_5、En_6以及De_5未进行改进,仍沿用现有技术中的结构,本申请在此不做详细描述。The present application improves the first four substructures of the U 2 -Net network, namely, the structures of En_1, De_1, En_2, De_2, En_3, De_3, En_4, and De_4, while En_5, En_6, and De_5 are not improved and still use the structures in the prior art, which will not be described in detail in the present application.
以En_1为例,现有的En_1网络结构依次为2个第一卷积块(即图4中的①Conv+BN+RELU)、5个下采样模块(即图4中的③Downsample×1/2Conv+BN+RELU)、1个第二卷积块(即图4中的⑤Conv+BN+RELU dilation=4)、1个第一卷积块以及5个上采样模块(即图4中的⑧Upsample×2Conv+BN+RELU)。Taking En_1 as an example, the existing network structure of En_1 consists of 2 first convolution blocks (i.e. ①Conv+BN+RELU in Figure 4), 5 downsampling modules (i.e. ③Downsample×1/2Conv+BN+RELU in Figure 4), 1 second convolution block (i.e. ⑤Conv+BN+RELU dilation=4 in Figure 4), 1 first convolution block and 5 upsampling modules (i.e. ⑧Upsample×2Conv+BN+RELU in Figure 4).
改进后的En_1网络结构依次为1个第一卷积块、1个LSFE模块(即图4中的②)、4个下采样模块、1个DPC模块(即图4中的④)、1个第二卷积块、1个第一卷积块以及5个上采样模块。也即,现有技术中的位置第二的第一卷积块替换为LSFE模块,最后一个下采样模块替换为DPC模块,其余均沿用之前的。由于现有技术中En_2、En_3、En_4与前述En_1相比,在结构上的区别是下采样模块数量以及上采样模块数量上的递减。因此,同理,改进后的En_2网络结构依次为1个第一卷积块、1个LSFE模块、3个下采样模块、1个DPC模块、1个第二卷积块、1个第一卷积块以及4个上采样模块。改进后的En_3网络结构依次为1个第一卷积块、1个LSFE模块、2个下采样模块、1个DPC模块、1个第二卷积块、1个第一卷积块以及3个上采样模块。改进后的En_4网络结构依次为1个第一卷积块、1个LSFE模块、1个下采样模块、1个DPC模块、1个第二卷积块、1个第一卷积块以及2个上采样模块。The improved En_1 network structure is composed of 1 first convolution block, 1 LSFE module (i.e., ② in Figure 4), 4 downsampling modules, 1 DPC module (i.e., ④ in Figure 4), 1 second convolution block, 1 first convolution block, and 5 upsampling modules. That is, the second first convolution block in the prior art is replaced by an LSFE module, the last downsampling module is replaced by a DPC module, and the rest are the same as before. Since the structural difference between En_2, En_3, and En_4 in the prior art and the aforementioned En_1 is the decrease in the number of downsampling modules and the number of upsampling modules. Therefore, in the same way, the improved En_2 network structure is composed of 1 first convolution block, 1 LSFE module, 3 downsampling modules, 1 DPC module, 1 second convolution block, 1 first convolution block, and 4 upsampling modules. The improved En_3 network structure is composed of 1 first convolution block, 1 LSFE module, 2 downsampling modules, 1 DPC module, 1 second convolution block, 1 first convolution block and 3 upsampling modules. The improved En_4 network structure is composed of 1 first convolution block, 1 LSFE module, 1 downsampling module, 1 DPC module, 1 second convolution block, 1 first convolution block and 2 upsampling modules.
另外,改进后的De_1、De_2、De_3、De_4与前述的改进后的En_1、En_2、En_3、En_4一一对应,在此不再详述。 In addition, the improved De_1, De_2, De_3, and De_4 correspond one-to-one to the aforementioned improved En_1, En_2, En_3, and En_4, and will not be described in detail here.
本申请采用的是二分类语义分割,其对应的数据集标签只包含标签0(非筏式养殖区)和标签1(筏式养殖区)。在训练过程中语义分割分支网络所涉及到的损失函数采用二元交叉熵损失函数。This application adopts binary classification semantic segmentation, and its corresponding data set labels only include label 0 (non-raft aquaculture area) and label 1 (raft aquaculture area). During the training process, the loss function involved in the semantic segmentation branch network adopts the binary cross entropy loss function.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,通过LSFE模块来提取大视野范围内养殖区的特征,以实现对海洋养殖区的宏观检测;通过DPC模块捕获远程上下文信息,以实现对每一小块养殖浮筏进行识别、提取和分类,从而提升语义分割分支网络的分割精度。The embodiment of the present application provides a method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images. The LSFE module is used to extract the features of aquaculture areas within a large field of view to achieve macroscopic detection of marine aquaculture areas. The DPC module is used to capture remote context information to identify, extract and classify each small aquaculture raft, thereby improving the segmentation accuracy of the semantic segmentation branch network.
进一步地,所述实例分割分支网络包括改进的SOTR网络,所述改进的SOTR网络至少包括Transformer模块;其中,所述Transformer模块包括可分离卷积以及iABN同步层;所述Transformer模块用于预测每个实例类别。Furthermore, the instance segmentation branch network includes an improved SOTR network, and the improved SOTR network includes at least a Transformer module; wherein the Transformer module includes a separable convolution and an iABN synchronization layer; and the Transformer module is used to predict each instance category.
现有技术中,SOTR利用Transformer简化了分割流程,使用两个并行子任务:1)通过Transformer预测每个实例类别;2)利用多级上采样模块动态生成分割掩码。其中,编码器-解码器Transformer模型通过一系列可学习的掩码嵌入将实例分割任务统一。本申请对Transformer使用可分离卷积和iABN(inplace activated batch normalization)同步层进行扩展,在一定程度上提高了分割精度和训练收敛性。In the prior art, SOTR uses Transformer to simplify the segmentation process, using two parallel subtasks: 1) predicting each instance category through Transformer; 2) dynamically generating segmentation masks using a multi-level upsampling module. Among them, the encoder-decoder Transformer model unifies the instance segmentation tasks through a series of learnable mask embeddings. This application extends the Transformer with separable convolution and iABN (inplace activated batch normalization) synchronization layers, which improves the segmentation accuracy and training convergence to a certain extent.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,通过可分离卷积和iABN同步层的扩展,提升了模型的分割精度和训练收敛性。The method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided in the embodiment of the present application improves the segmentation accuracy and training convergence of the model through the expansion of separable convolution and iABN synchronization layers.
进一步地,所述实例分割分支网络还包括特征提取模块,所述特征提取模块包括移动反向瓶颈单元以及双向特征金字塔网络(Feature pyramid network,FPN)。Furthermore, the instance segmentation branch network also includes a feature extraction module, which includes a moving reverse bottleneck unit and a bidirectional feature pyramid network (Feature pyramid network, FPN).
利用所述特征提取模块对所述待分割图像进行特征提取,以获得多尺度特征。The feature extraction module is used to extract features from the image to be segmented to obtain multi-scale features.
基于所述多尺度特征,通过所述改进的SOTR网络对所述待分割图像进行实例分割,以获得初始实例分割结果。 Based on the multi-scale features, instance segmentation is performed on the image to be segmented through the improved SOTR network to obtain an initial instance segmentation result.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,通过移动反向瓶颈单元以及双向特征金字塔网络实现多尺度特征的提取,能够获取和更多的大尺度特征、小尺度特征、浅层信息、深层信息,用以提升实例分割的准确性。The method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided in the embodiment of the present application realizes the extraction of multi-scale features by moving the reverse bottleneck unit and the bidirectional feature pyramid network, and can obtain more large-scale features, small-scale features, shallow information, and deep information to improve the accuracy of instance segmentation.
图5是本申请实施例提供的全景分割模型的训练过程示意图;如图5所示,所述预先训练好的全景分割模型通过如下方式训练得到:FIG5 is a schematic diagram of the training process of the panoramic segmentation model provided in an embodiment of the present application; as shown in FIG5 , the pre-trained panoramic segmentation model is trained in the following manner:
获取训练数据集以及其对应的标签,并构建全景分割模型。Get the training dataset and its corresponding labels and build a panoptic segmentation model.
其中,所述标签包括筏式养殖区与非筏式养殖区的语义标签以及多种养殖区类别的实例标签。The labels include semantic labels of raft aquaculture areas and non-raft aquaculture areas and instance labels of various aquaculture area categories.
需要说明的是,在获取训练数据集之后需要对训练数据集进行标注,具体标注分为“stuff”类和“thing”类。其中,“stuff”类用语义掩膜标注,0表示非筏式养殖区,1表示筏式养殖区;“thing”类用实例掩膜标注,包含鱼类、藻类、贝类和其它等4个实例。根据上述标注类别,对训练集和测试集创建背景(即非筏式养殖区)以及前景(即筏式养殖区)的语义标签,在此基础上创建实例1(鱼类)、实例2(藻类)、实例3(贝类)、实例4(其它)的实例标签。It should be noted that after obtaining the training data set, the training data set needs to be annotated. The specific annotations are divided into "stuff" and "thing" classes. Among them, the "stuff" class is annotated with a semantic mask, 0 represents a non-raft aquaculture area, and 1 represents a raft aquaculture area; the "thing" class is annotated with an instance mask, including four instances of fish, algae, shellfish, and others. According to the above annotation categories, semantic labels of the background (i.e., non-raft aquaculture area) and foreground (i.e., raft aquaculture area) are created for the training set and the test set, and on this basis, instance labels of instance 1 (fish), instance 2 (algae), instance 3 (shellfish), and instance 4 (others) are created.
将所述训练数据集输入所述语义分割分支网络,预测得到训练用语义分割结果,计算训练用语义分割结果与所述语义标签之间的损失,获得第一损失。The training data set is input into the semantic segmentation branch network, a training semantic segmentation result is predicted, and a loss between the training semantic segmentation result and the semantic label is calculated to obtain a first loss.
将所述训练数据集输入所述实例分割分支网络,预测得到训练用实例分割结果,计算训练用实例分割结果与所述实例标签之间的损失,获得第二损失。The training data set is input into the instance segmentation branch network, the training instance segmentation result is predicted, and the loss between the training instance segmentation result and the instance label is calculated to obtain a second loss.
利用所述全景融合模块对所述训练用语义分割结果与所述训练用实例分割结果进行自适应融合,获得训练用多分类结果。The panoramic fusion module is used to adaptively fuse the training semantic segmentation result and the training instance segmentation result to obtain a training multi-classification result.
根据所述第一损失以及所述第二损失获取总损失,基于所述训练用多分类结果以及所述总损失对所述全景分割模型进行训练,直至所述全景分割模型收敛,以获得训练好的全景分割模型。A total loss is obtained according to the first loss and the second loss, and the panoramic segmentation model is trained based on the training multi-classification result and the total loss until the panoramic segmentation model converges to obtain a trained panoramic segmentation model.
其中,总损失根据衰减或放大融合的逻辑输出分数对第一损失与第二损失自适应加权求和得到。 The total loss is obtained by adaptively weighting the first loss and the second loss according to the logical output score of the attenuated or amplified fusion.
需要说明的是,本申请通过共享合成数据集对全景分割模型进行训练,在全景分割模型达到收敛精度后。再由测试集对全景分割模型进行测试,得出定性评估结果。It should be noted that this application trains the panoramic segmentation model by sharing the synthetic data set, and after the panoramic segmentation model reaches convergence accuracy, the panoramic segmentation model is tested by the test set to obtain qualitative evaluation results.
若测试不满足精度要求,则通过调整超参数、补充训练样本等手段再次训练模型,直至满足定性评估要求后,应用检测集对模型进行检测并评估模型的全景分割精度PQ,最终才获得训练好的全景分割模型。If the test does not meet the accuracy requirements, the model is trained again by adjusting hyperparameters, supplementing training samples, etc. until it meets the qualitative evaluation requirements. The model is then tested with the test set and the panoramic segmentation accuracy PQ of the model is evaluated to finally obtain a trained panoramic segmentation model.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,上述提到的数据集标签不仅对筏式养殖区和背景区域进行整体分类标注,同时进一步地对筏式养殖区的多种养殖类别进行了精细化分类标注,以实现多分类任务。The embodiment of the present application provides a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation. The dataset labels mentioned above not only classify and label the raft aquaculture areas and background areas as a whole, but also further classify and label the various aquaculture categories in the raft aquaculture areas in a refined manner to achieve multi-classification tasks.
进一步地,在所述获取训练数据集以及其对应的标签之后,该方法还包括:Furthermore, after obtaining the training data set and its corresponding labels, the method further includes:
分别构建所述训练数据集的归一化植被指数特征(Normalized Difference Vegetation Index,NDVI)和归一化水体指数特征(Normalized Difference Water Index,NDWI)。The normalized vegetation index feature (Normalized Difference Vegetation Index, NDVI) and the normalized water index feature (Normalized Difference Water Index, NDWI) of the training data set are constructed respectively.
将所述归一化植被指数特征和归一化水体指数特征与所述训练数据集进行融合,以获得共享合成数据集。The normalized vegetation index feature and the normalized water body index feature are fused with the training data set to obtain a shared synthetic data set.
相应地,所述将所述训练数据集输入所述语义分割分支网络,包括:Accordingly, inputting the training data set into the semantic segmentation branch network includes:
将所述共享合成数据集输入所述语义分割分支网络。The shared synthetic dataset is input into the semantic segmentation branch network.
所述将所述训练数据集输入所述实例分割分支网络,包括:The step of inputting the training data set into the instance segmentation branch network comprises:
将所述共享合成数据集输入所述实例分割分支网络。The shared synthetic dataset is input into the instance segmentation branch network.
需要说明的是,归一化植被指数特征和归一化水体指数特征也可以在获取待分割图像之后与待分割图像进行融合,从而利用全景分割模型对融合得到的图像进行分割。It should be noted that the normalized vegetation index feature and the normalized water index feature can also be fused with the image to be segmented after the image to be segmented is obtained, so as to segment the fused image using the panoramic segmentation model.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,通过在训练数据集中融合NDVI和NDWI两种自定义特征,获得共享合成数据集,从而能够更加充分、有效地利用光学遥感影像中丰富的各种信息,以提升全景分割模型的分割精度。 The method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation provided in the embodiment of the present application obtains a shared synthetic data set by fusing two custom features, NDVI and NDWI, into a training data set, thereby being able to more fully and effectively utilize the rich variety of information in optical remote sensing images to improve the segmentation accuracy of the panoramic segmentation model.
进一步地,所述训练数据集包括标签数据集以及对抗样本集,所述标签数据集为标注后具有对应标签的数据集,所述对抗样本集通过对所述训练用实例分割结果进行对抗训练得到。Furthermore, the training data set includes a labeled data set and an adversarial sample set, the labeled data set is a data set with corresponding labels after being annotated, and the adversarial sample set is obtained by performing adversarial training on the training instance segmentation results.
具体地,应用海洋养殖区多分类数据标签集对实例分割分支进行对抗训练,以提高多目标多分类的抗干扰能力。对抗训练过程中生成的对抗样本添加到训练数据集中,与标签数据集共同构成训练数据集。Specifically, the multi-classification data label set of the marine aquaculture area is used to perform adversarial training on the instance segmentation branch to improve the anti-interference ability of multi-target multi-classification. The adversarial samples generated during the adversarial training are added to the training dataset and together with the label dataset constitute the training dataset.
需要说明的是,对抗训练可以是通过添加鉴别器,亦或是根据梯度回传生成新样本,对抗训练方法为常规方法,本申请对此不做限定。It should be noted that adversarial training can be achieved by adding a discriminator or generating new samples based on gradient feedback. The adversarial training method is a conventional method and this application does not limit it.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,通过对抗训练获得对抗样本集,并基于对抗样本集以及标签数据集对全景分割模型实现训练,从而能够提升全景分割模型的抗干扰能力。The embodiment of the present application provides a method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation. An adversarial sample set is obtained through adversarial training, and a panoramic segmentation model is trained based on the adversarial sample set and a label data set, thereby improving the anti-interference ability of the panoramic segmentation model.
进一步地,所述标签数据集通过如下方式获得:Furthermore, the label data set is obtained by:
获取训练用海洋养殖区的光学遥感影像,对所述光学遥感影像至少执行存储格式统一、去云雾处理、归一化处理以及裁剪处理。An optical remote sensing image of a training marine aquaculture area is obtained, and the optical remote sensing image is subjected to at least storage format unification, cloud removal, normalization, and cropping.
具体地,先获取覆盖我国海岸线30km范围内的Sentinel-2、GF-1(PMS/WFV)以及Landsat的中分辨率光学遥感影像。然后,将光学遥感影像进行存储格式统一、去云雾处理和归一化处理,之后再进行滑动裁剪成2048*2048大小的标准图像,从而形成标准的标签数据集。Specifically, we first obtain medium-resolution optical remote sensing images from Sentinel-2, GF-1 (PMS/WFV) and Landsat covering 30km of my country's coastline. Then, we unify the storage format, remove the cloud and fog, and normalize the optical remote sensing images, and then slide and crop them into standard images of 2048*2048 size, thus forming a standard label dataset.
下面对本申请提供的基于全景分割的光学遥感影像海洋筏式养殖区分类装置进行描述,下文描述的基于全景分割的光学遥感影像海洋筏式养殖区分类装置与上文描述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法可相互对应参照。The following is a description of the marine raft aquaculture area classification device based on panoramic segmentation of optical remote sensing images provided in the present application. The marine raft aquaculture area classification device based on panoramic segmentation of optical remote sensing images described below and the marine raft aquaculture area classification method based on panoramic segmentation of optical remote sensing images described above can be referenced to each other.
图6是本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类装置的结构示意图;如图6所示,基于全景分割的光学遥感影像海洋筏式养殖区分类装置包括图像获取模块601以及图像分割模块602。Figure 6 is a structural schematic diagram of the optical remote sensing image marine raft aquaculture area classification device based on panoramic segmentation provided in an embodiment of the present application; as shown in Figure 6, the optical remote sensing image marine raft aquaculture area classification device based on panoramic segmentation includes an image acquisition module 601 and an image segmentation module 602.
图像获取模块601,用于获取待分割图像,所述待分割图像为海洋养殖区的光学遥感影像。The image acquisition module 601 is used to acquire the image to be segmented, where the image to be segmented is an optical remote sensing image of a marine aquaculture area.
图像分割模块602,用于将所述待分割图像输入至预先训练好的全景 分割模型中,预测得到多分类分割结果。The image segmentation module 602 is used to input the image to be segmented into a pre-trained panoramic image. In the segmentation model, multi-classification segmentation results are predicted.
其中,所述多分类分割结果包括筏式养殖区、非筏式养殖区以及多种养殖区类别,多种养殖区类别又包括鱼类、藻类、贝类以及其他。The multi-classification segmentation results include raft aquaculture areas, non-raft aquaculture areas and multiple aquaculture area categories, and the multiple aquaculture area categories include fish, algae, shellfish and others.
所述预先训练好的全景分割模型(High precision panoptic segmentation,HPPS)包括语义分割分支网络、实例分割分支网络以及全景融合模块。The pre-trained panoramic segmentation model (High precision panoptic segmentation, HPPS) includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module.
利用所述语义分割分支网络对所述待分割图像进行语义分割,以获得初始语义分割结果,所述初始语义分割结果包括初始筏式养殖区与初始非筏式养殖区。The image to be segmented is semantically segmented using the semantic segmentation branch network to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area.
利用所述实例分割分支网络对所述待分割图像进行实例分割,以获得初始实例分割结果,所述初始实例分割结果包括多种初始养殖区类别。The image to be segmented is subjected to instance segmentation using the instance segmentation branch network to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories.
利用所述全景融合模块对所述初始语义分割结果以及所述初始实例分割结果进行融合,以获得多分类分割结果。The panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
其中,全景融合模块为无参数全景融合模块,其基于像素的头部预测适应性,有选择地衰减或放大融合的逻辑输出分数来自适应融合初始语义分割结果与初始实例分割结果。通过这种端到端方式联合优化整个全景分割网络,从而获得最终的海洋筏式养殖区的多分类高精度的全景分割输出结果,实现了光学遥感影像海洋筏式养殖区多分类任务。Among them, the panoramic fusion module is a parameter-free panoramic fusion module, which selectively attenuates or amplifies the fused logical output score based on the pixel-based head prediction adaptability to adaptively fuse the initial semantic segmentation results and the initial instance segmentation results. Through this end-to-end approach, the entire panoramic segmentation network is jointly optimized to obtain the final multi-classification and high-precision panoramic segmentation output results of the marine raft aquaculture area, realizing the multi-classification task of the optical remote sensing image marine raft aquaculture area.
另外,在将所述待分割图像输入至预先训练好的全景分割模型中之前,对待分割图像进行标准化预处理,然后滑动裁剪成2048*2048大小的图像,将裁剪后的图像依次输入训练好的HPPS模型(即全景分割模型),该HPPS模型的输出结果即为海上养殖区域的多分类结果,对所有的图像进行拼接获得待分割图像的整体的全景分割结果图。In addition, before the image to be segmented is input into the pre-trained panoramic segmentation model, the image to be segmented is standardized and pre-processed, and then slidingly cropped into an image of 2048*2048 size, and the cropped images are sequentially input into the trained HPPS model (i.e., the panoramic segmentation model). The output result of the HPPS model is the multi-classification result of the offshore aquaculture area, and all the images are spliced to obtain the overall panoramic segmentation result map of the image to be segmented.
本申请实施例提供的基于全景分割的光学遥感影像海洋筏式养殖区分类装置,通过语义分割分支网络与实例分割分支网络对待分割图像并行分割,并将两个分支网络的输出通过无参数的全景融合模块进行融合,最后获得多分类分割结果,实现多任务分类。且全景融合模块这种自适应融合方式能够更加完整利用语义分割头和实例分割头的逻辑输出,提升多分类任务的准确性。The optical remote sensing image marine raft aquaculture area classification device based on panoramic segmentation provided in the embodiment of the present application uses a semantic segmentation branch network and an instance segmentation branch network to parallelly segment the segmented image, and fuses the outputs of the two branch networks through a parameter-free panoramic fusion module, and finally obtains a multi-classification segmentation result to achieve multi-task classification. Moreover, the adaptive fusion method of the panoramic fusion module can more completely utilize the logical outputs of the semantic segmentation head and the instance segmentation head to improve the accuracy of multi-classification tasks.
以上所描述的装置实施例仅仅是示意性的,其中作为分离部件说明的 单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。本领域普通技术人员在不付出创造性的劳动的情况下,即可以理解并实施。The above-described device embodiments are merely illustrative, and the components described as separate components are The units may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of this embodiment. Those of ordinary skill in the art may understand and implement it without creative work.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到各实施方式可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件。基于这样的理解,上述技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品可以存储在计算机可读存储介质中,如ROM/RAM、磁碟、光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行各个实施例或者实施例的某些部分的方法。Through the description of the above implementation methods, those skilled in the art can clearly understand that each implementation method can be implemented by means of software plus a necessary general hardware platform, and of course, it can also be implemented by hardware. Based on this understanding, the above technical solution is essentially or the part that contributes to the prior art can be embodied in the form of a software product, and the computer software product can be stored in a computer-readable storage medium, such as ROM/RAM, a disk, an optical disk, etc., including a number of instructions for a computer device (which can be a personal computer, a server, or a network device, etc.) to execute the methods of each embodiment or some parts of the embodiment.
最后应说明的是:以上实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围。 Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present application, rather than to limit it. Although the present application has been described in detail with reference to the aforementioned embodiments, those skilled in the art should understand that they can still modify the technical solutions described in the aforementioned embodiments, or replace some of the technical features therein with equivalents. However, these modifications or replacements do not deviate the essence of the corresponding technical solutions from the spirit and scope of the technical solutions of the embodiments of the present application.

Claims (10)

  1. 一种基于全景分割的光学遥感影像海洋筏式养殖区分类方法,包括:A method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images, comprising:
    获取待分割图像,所述待分割图像为海洋养殖区的光学遥感影像;Acquire an image to be segmented, where the image to be segmented is an optical remote sensing image of a marine aquaculture area;
    将所述待分割图像输入至预先训练好的全景分割模型中,预测得到多分类分割结果,所述多分类分割结果包括筏式养殖区、非筏式养殖区以及多种养殖区类别;Inputting the image to be segmented into a pre-trained panoramic segmentation model to predict a multi-classification segmentation result, wherein the multi-classification segmentation result includes a raft aquaculture area, a non-raft aquaculture area, and multiple aquaculture area categories;
    其中,所述预先训练好的全景分割模型包括语义分割分支网络、实例分割分支网络以及全景融合模块;The pre-trained panoramic segmentation model includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module;
    利用所述语义分割分支网络对所述待分割图像进行语义分割,以获得初始语义分割结果,所述初始语义分割结果包括初始筏式养殖区与初始非筏式养殖区;Using the semantic segmentation branch network to perform semantic segmentation on the image to be segmented to obtain an initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area;
    利用所述实例分割分支网络对所述待分割图像进行实例分割,以获得初始实例分割结果,所述初始实例分割结果包括多种初始养殖区类别;Using the instance segmentation branch network to perform instance segmentation on the image to be segmented to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories;
    利用所述全景融合模块对所述初始语义分割结果以及所述初始实例分割结果进行融合,以获得多分类分割结果。The panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
  2. 根据权利要求1所述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,其中,所述语义分割分支网络为改进的U2-Net网络,所述改进的U2-Net网络至少包括6个U型结构的次级编码器以及5个U型结构的次级解码器,所述6个U型结构的次级编码器依次为4个第一次级编码器以及2个第二次级编码器,所述5个U型结构的次级解码器依次为4个第一次级解码器以及1个第二次级解码器;所述第一次级编码器以及所述第一次级解码器均由第一卷积块、LSFE模块、多个下采样模块、DPC模块、第二卷积块、第一卷积块以及多个上采样模块依次构成;According to the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images in claim 1, wherein the semantic segmentation branch network is an improved U2 -Net network, and the improved U2 -Net network includes at least 6 U-shaped secondary encoders and 5 U-shaped secondary decoders, the 6 U-shaped secondary encoders are sequentially 4 first secondary encoders and 2 second secondary encoders, and the 5 U-shaped secondary decoders are sequentially 4 first secondary decoders and 1 second secondary decoder; the first secondary encoder and the first secondary decoder are both composed of a first convolution block, an LSFE module, a plurality of down-sampling modules, a DPC module, a second convolution block, a first convolution block, and a plurality of up-sampling modules in sequence;
    其中,所述LSFE模块用于提取大视野范围内养殖区的特征,其包括可分离卷积以及输出滤波器;The LSFE module is used to extract the features of the breeding area within a large field of view, and includes a separable convolution and an output filter;
    所述DPC模块用于捕获远程上下文信息,其包括可分离卷积以及输出通道。The DPC module is used to capture long-range context information, which includes separable convolution and output channels.
  3. 根据权利要求1所述的基于全景分割的光学遥感影像海洋筏式养 殖区分类方法,其中,所述实例分割分支网络包括改进的SOTR网络,所述改进的SOTR网络至少包括Transformer模块;其中,所述Transformer模块包括可分离卷积以及iABN同步层;所述Transformer模块用于预测每个实例类别。The optical remote sensing image marine raft culture method based on panoramic segmentation according to claim 1 A method for classifying regions, wherein the instance segmentation branch network includes an improved SOTR network, and the improved SOTR network includes at least a Transformer module; wherein the Transformer module includes a separable convolution and an iABN synchronization layer; and the Transformer module is used to predict the category of each instance.
  4. 根据权利要求3所述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,其中,所述实例分割分支网络还包括特征提取模块,所述特征提取模块包括移动反向瓶颈单元以及双向特征金字塔网络;According to the method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation according to claim 3, wherein the instance segmentation branch network further includes a feature extraction module, and the feature extraction module includes a moving reverse bottleneck unit and a bidirectional feature pyramid network;
    利用所述特征提取模块对所述待分割图像进行特征提取,以获得多尺度特征;Using the feature extraction module to extract features from the image to be segmented to obtain multi-scale features;
    基于所述多尺度特征,通过所述改进的SOTR网络对所述待分割图像进行实例分割,以获得初始实例分割结果。Based on the multi-scale features, instance segmentation is performed on the image to be segmented through the improved SOTR network to obtain an initial instance segmentation result.
  5. 根据权利要求1-4任一所述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,其中,所述预先训练好的全景分割模型通过如下方式训练得到:According to any one of claims 1 to 4, the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images, wherein the pre-trained panoramic segmentation model is trained in the following manner:
    获取训练数据集以及其对应的标签,并构建全景分割模型;其中,所述标签包括筏式养殖区与非筏式养殖区的语义标签以及多种养殖区类别的实例标签;Obtain a training data set and its corresponding labels, and construct a panoramic segmentation model; wherein the labels include semantic labels of raft aquaculture areas and non-raft aquaculture areas and instance labels of multiple aquaculture area categories;
    将所述训练数据集输入所述语义分割分支网络,预测得到训练用语义分割结果,计算训练用语义分割结果与所述语义标签之间的损失,获得第一损失;Inputting the training data set into the semantic segmentation branch network, predicting a training semantic segmentation result, calculating the loss between the training semantic segmentation result and the semantic label, and obtaining a first loss;
    将所述训练数据集输入所述实例分割分支网络,预测得到训练用实例分割结果,计算训练用实例分割结果与所述实例标签之间的损失,获得第二损失;Inputting the training data set into the instance segmentation branch network, predicting a training instance segmentation result, calculating the loss between the training instance segmentation result and the instance label, and obtaining a second loss;
    利用所述全景融合模块对所述训练用语义分割结果与所述训练用实例分割结果进行自适应融合,获得训练用多分类结果;Using the panoramic fusion module, the training semantic segmentation result and the training instance segmentation result are adaptively fused to obtain a training multi-classification result;
    根据所述第一损失以及所述第二损失获取总损失,基于所述训练用多分类结果以及所述总损失对所述全景分割模型进行训练,直至所述全景分割模型收敛,以获得训练好的全景分割模型。A total loss is obtained according to the first loss and the second loss, and the panoramic segmentation model is trained based on the training multi-classification result and the total loss until the panoramic segmentation model converges to obtain a trained panoramic segmentation model.
  6. 根据权利要求5所述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,其中,在所述获取训练数据集以及其对应的标签之后,该 方法还包括:According to the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images according to claim 5, after obtaining the training data set and its corresponding labels, the The method also includes:
    分别构建所述训练数据集的归一化植被指数特征和归一化水体指数特征;Respectively constructing the normalized vegetation index feature and the normalized water index feature of the training data set;
    将所述归一化植被指数特征和归一化水体指数特征与所述训练数据集进行融合,以获得共享合成数据集;fusing the normalized vegetation index feature and the normalized water index feature with the training data set to obtain a shared synthetic data set;
    相应地,所述将所述训练数据集输入所述语义分割分支网络,包括:Accordingly, inputting the training data set into the semantic segmentation branch network includes:
    将所述共享合成数据集输入所述语义分割分支网络;Inputting the shared synthetic dataset into the semantic segmentation branch network;
    所述将所述训练数据集输入所述实例分割分支网络,包括:The step of inputting the training data set into the instance segmentation branch network comprises:
    将所述共享合成数据集输入所述实例分割分支网络。The shared synthetic dataset is input into the instance segmentation branch network.
  7. 根据权利要求5所述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,其中,所述训练数据集包括标签数据集以及对抗样本集,所述标签数据集为标注后具有对应标签的数据集,所述对抗样本集通过对所述训练用实例分割结果进行对抗训练得到。According to the method for classifying marine raft aquaculture areas in optical remote sensing images based on panoramic segmentation in claim 5, the training data set includes a labeled data set and an adversarial sample set, the labeled data set is a data set with corresponding labels after annotation, and the adversarial sample set is obtained by adversarial training on the segmentation results of the training instances.
  8. 根据权利要求7所述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,其中,所述标签数据集通过如下方式获得:According to the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images according to claim 7, wherein the label data set is obtained by:
    获取训练用海洋养殖区的光学遥感影像,对所述光学遥感影像至少执行存储格式统一、去云雾处理、归一化处理以及裁剪处理。An optical remote sensing image of a training marine aquaculture area is obtained, and the optical remote sensing image is subjected to at least storage format unification, cloud removal, normalization, and cropping.
  9. 根据权利要求7所述的基于全景分割的光学遥感影像海洋筏式养殖区分类方法,其中,所述多种养殖区类别至少包括鱼类、藻类以及贝类。According to the method for classifying marine raft aquaculture areas based on panoramic segmentation of optical remote sensing images in claim 7, the multiple aquaculture area categories include at least fish, algae and shellfish.
  10. 一种基于全景分割的光学遥感影像海洋筏式养殖区分类装置,包括:An optical remote sensing image marine raft aquaculture area classification device based on panoramic segmentation, comprising:
    图像获取模块,用于获取待分割图像,所述待分割图像为海洋养殖区的光学遥感影像;An image acquisition module is used to acquire an image to be segmented, wherein the image to be segmented is an optical remote sensing image of a marine aquaculture area;
    图像分割模块,用于将所述待分割图像输入至预先训练好的全景分割模型中,预测得到多分类分割结果,所述多分类分割结果包括筏式养殖区、非筏式养殖区以及多种养殖区类别;An image segmentation module is used to input the image to be segmented into a pre-trained panoramic segmentation model to predict a multi-classification segmentation result, wherein the multi-classification segmentation result includes raft aquaculture areas, non-raft aquaculture areas, and multiple aquaculture area categories;
    其中,所述预先训练好的全景分割模型包括语义分割分支网络、实例分割分支网络以及全景融合模块;The pre-trained panoramic segmentation model includes a semantic segmentation branch network, an instance segmentation branch network and a panoramic fusion module;
    利用所述语义分割分支网络对所述待分割图像进行语义分割,以获得 初始语义分割结果,所述初始语义分割结果包括初始筏式养殖区与初始非筏式养殖区;The semantic segmentation branch network is used to perform semantic segmentation on the image to be segmented to obtain An initial semantic segmentation result, wherein the initial semantic segmentation result includes an initial raft aquaculture area and an initial non-raft aquaculture area;
    利用所述实例分割分支网络对所述待分割图像进行实例分割,以获得初始实例分割结果,所述初始实例分割结果包括多种初始养殖区类别;Using the instance segmentation branch network to perform instance segmentation on the image to be segmented to obtain an initial instance segmentation result, wherein the initial instance segmentation result includes a plurality of initial breeding area categories;
    利用所述全景融合模块对所述初始语义分割结果以及所述初始实例分割结果进行融合,以获得多分类分割结果。 The panoramic fusion module is used to fuse the initial semantic segmentation result and the initial instance segmentation result to obtain a multi-classification segmentation result.
PCT/CN2023/092747 2022-10-27 2023-05-08 Panoptic segmentation-based optical remote-sensing image raft mariculture area classification method WO2024087574A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211328346.5A CN115908894A (en) 2022-10-27 2022-10-27 Optical remote sensing image ocean raft type culture area classification method based on panoramic segmentation
CN202211328346.5 2022-10-27

Publications (1)

Publication Number Publication Date
WO2024087574A1 true WO2024087574A1 (en) 2024-05-02

Family

ID=86475256

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/092747 WO2024087574A1 (en) 2022-10-27 2023-05-08 Panoptic segmentation-based optical remote-sensing image raft mariculture area classification method

Country Status (2)

Country Link
CN (1) CN115908894A (en)
WO (1) WO2024087574A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115908894A (en) * 2022-10-27 2023-04-04 中国科学院空天信息创新研究院 Optical remote sensing image ocean raft type culture area classification method based on panoramic segmentation
CN116452901B (en) * 2023-06-19 2023-09-15 中国科学院海洋研究所 Automatic extraction method for ocean culture area of remote sensing image based on deep learning
CN117036982B (en) * 2023-10-07 2024-01-09 山东省国土空间数据和遥感技术研究院(山东省海域动态监视监测中心) Method and device for processing optical satellite image of mariculture area, equipment and medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108460411A (en) * 2018-02-09 2018-08-28 北京市商汤科技开发有限公司 Example dividing method and device, electronic equipment, program and medium
CN111292331A (en) * 2020-02-23 2020-06-16 华为技术有限公司 Image processing method and device
CN114842215A (en) * 2022-04-20 2022-08-02 大连海洋大学 Fish visual identification method based on multi-task fusion
WO2022160753A1 (en) * 2021-01-27 2022-08-04 上海商汤智能科技有限公司 Image processing method and apparatus, and electronic device and storage medium
CN115100652A (en) * 2022-08-02 2022-09-23 北京卫星信息工程研究所 Electronic map automatic generation method based on high-resolution remote sensing image
CN115908894A (en) * 2022-10-27 2023-04-04 中国科学院空天信息创新研究院 Optical remote sensing image ocean raft type culture area classification method based on panoramic segmentation

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108460411A (en) * 2018-02-09 2018-08-28 北京市商汤科技开发有限公司 Example dividing method and device, electronic equipment, program and medium
CN111292331A (en) * 2020-02-23 2020-06-16 华为技术有限公司 Image processing method and device
WO2022160753A1 (en) * 2021-01-27 2022-08-04 上海商汤智能科技有限公司 Image processing method and apparatus, and electronic device and storage medium
CN114842215A (en) * 2022-04-20 2022-08-02 大连海洋大学 Fish visual identification method based on multi-task fusion
CN115100652A (en) * 2022-08-02 2022-09-23 北京卫星信息工程研究所 Electronic map automatic generation method based on high-resolution remote sensing image
CN115908894A (en) * 2022-10-27 2023-04-04 中国科学院空天信息创新研究院 Optical remote sensing image ocean raft type culture area classification method based on panoramic segmentation

Also Published As

Publication number Publication date
CN115908894A (en) 2023-04-04

Similar Documents

Publication Publication Date Title
WO2024087574A1 (en) Panoptic segmentation-based optical remote-sensing image raft mariculture area classification method
Feng et al. Convolutional neural networks for intra-hour solar forecasting based on sky image sequences
CN112418117A (en) Small target detection method based on unmanned aerial vehicle image
CN115223063B (en) Deep learning-based unmanned aerial vehicle remote sensing wheat new variety lodging area extraction method and system
CN110555841A (en) SAR image change detection method based on self-attention image fusion and DEC
Nurkarim et al. Building footprint extraction and counting on very high-resolution satellite imagery using object detection deep learning framework
CN112434723A (en) Day/night image classification and object detection method based on attention network
CN114926469A (en) Semantic segmentation model training method, semantic segmentation method, storage medium and terminal
CN111259733A (en) Point cloud image-based ship identification method and device
CN113627440A (en) Large-scale point cloud semantic segmentation method based on lightweight neural network
CN114943902A (en) Urban vegetation unmanned aerial vehicle remote sensing classification method based on multi-scale feature perception network
Jiang et al. Forest-CD: Forest change detection network based on VHR images
CN116740516A (en) Target detection method and system based on multi-scale fusion feature extraction
Sun et al. Image target detection algorithm compression and pruning based on neural network
Zhao et al. Ocean ship detection and recognition algorithm based on aerial image
CN116524255A (en) Wheat scab spore identification method based on Yolov5-ECA-ASFF
Wang et al. Simultaneous extracting area and quantity of agricultural greenhouses in large scale with deep learning method and high-resolution remote sensing images
CN116503602A (en) Unstructured environment three-dimensional point cloud semantic segmentation method based on multi-level edge enhancement
Zhao et al. Recognition and Classification of Concrete Cracks under Strong Interference Based on Convolutional Neural Network.
CN116543191A (en) Multi-source remote sensing data self-adaptive fusion ground object classification method
Dong et al. Accurate and real-time visual detection algorithm for environmental perception of USVS under all-weather conditions
Yin et al. M2F2-RCNN: Multi-functional faster RCNN based on multi-scale feature fusion for region search in remote sensing images
CN113780117A (en) Method for rapidly identifying and extracting relevant parameters of estuary plume profile
Xia et al. Application of yolov7 and transformer structures to small object (license plate) detection in complex traffic scenes
Greenwell et al. Implicit land use mapping using social media imagery