CN112330562A - Heterogeneous remote sensing image transformation method and system - Google Patents

Heterogeneous remote sensing image transformation method and system Download PDF

Info

Publication number
CN112330562A
CN112330562A CN202011236018.3A CN202011236018A CN112330562A CN 112330562 A CN112330562 A CN 112330562A CN 202011236018 A CN202011236018 A CN 202011236018A CN 112330562 A CN112330562 A CN 112330562A
Authority
CN
China
Prior art keywords
image
visible light
generator
sar
noise reduction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011236018.3A
Other languages
Chinese (zh)
Other versions
CN112330562B (en
Inventor
刘瑜
谭大宁
李刚
丁自然
孙顺
姜乔文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Naval Aeronautical University
Original Assignee
Tsinghua University
Naval Aeronautical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University, Naval Aeronautical University filed Critical Tsinghua University
Priority to CN202011236018.3A priority Critical patent/CN112330562B/en
Publication of CN112330562A publication Critical patent/CN112330562A/en
Application granted granted Critical
Publication of CN112330562B publication Critical patent/CN112330562B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/29Graphical models, e.g. Bayesian networks
    • G06F18/295Markov models or related models, e.g. semi-Markov models; Markov random fields; Networks embedding Markov models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/40Analysis of texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a heterogeneous remote sensing image transformation method and a system, wherein a visible light gray image is set as a middle state from an SAR image to a visible light color image, the SAR image is generated into the visible light gray image through an SAR image noise reduction generator and is identified through a first identifier, the visible light gray image is generated into the visible light color image through the visible light image generator and is identified through a second identifier, the distortion rate from the SAR image to the visible light color image is reduced, and the quality and the efficiency of remote sensing image transformation are improved.

Description

Heterogeneous remote sensing image transformation method and system
Technical Field
The invention relates to the technical field of image transformation, in particular to a heterogeneous remote sensing image transformation method and system.
Background
With the continuous development of information perception technology, the processing and application of remote sensing images gradually change from single mode to multi-mode. The reason is that the image acquired by a single sensor cannot meet the requirements of certain specific scenes, and under the condition that the performance of the sensor cannot be effectively improved temporarily, the heterogeneous images acquired by multiple types of sensors are combined into a feasible way.
Taking optical and Synthetic Aperture Radar (SAR) images as an example, optical images are sensitive to weather, especially cloud layers, which severely limits their use for observation and monitoring of ground targets. A possible alternative is to correlate images acquired by different modality sensors at the same time, using a weather insensitive SAR image as a compensation. In this respect, deep learning based approaches provide a good solution. Under heterogeneous image-driven coupling time series, deep learning based methods can learn the mapping from SAR images to visible light images and use it for the transformation of known SAR images to unknown available light images. The obtained result has the spectrum information of the visible light image and can also obtain the all-weather observation advantages of the SAR image all day long.
Heterogeneous transformations of remote sensing images can be classified as an image style transformation problem, and can be further classified into feature-based methods and methods based on generation of a countermeasure network GAN in terms of image style transformation. The method based on the characteristics adopts the idea that the style and semantic content of an original image are respectively extracted by utilizing the characteristic extraction capability of an artificial neural network, the local affine transformation from input to output is limited in a color space by considering the distortion of certain style and content, the constraint is expressed as a self-defined fully differentiable energy item, and then the combined reconstruction is carried out to generate a new style image. Another idea of style transformation is based on a method of generating a countermeasure network GAN. The idea of the method is to adopt a conditional generation countermeasure network cGAN to perform conditional generation and transformation on the image, so that the generated image has the semantic information of an original image domain and the style characteristics of a target image domain.
For the two existing methods, although the characteristic-based method has good effect on artistic image style transformation, when a new style photo is generated, the loss function needs to be derived for many times, so that the parameter calculation amount is large when the method is used for heterogeneous transformation of remote sensing images, and the processing instantaneity is poor. Compared with the former method, the method for generating the anti-network cGAN based on the condition has the advantages of low calculation cost, higher flexibility, ideal conversion result between general images with clear structures and symmetrical information, and serious texture loss and distortion when the method is used for heterogeneous conversion of remote sensing images, which influences final decision and judgment.
At present, the heterogeneous remote sensing image transformation for generating the countermeasure network based on the conditions has the following technical difficulties: firstly, due to the principle difference of the SAR image and the visible light image acquisition sensor, the SAR image and the visible light image acquisition sensor have the characteristic of information asymmetry, and experiments are carried out by utilizing the existing image transformation models such as pix2pix, Cycle-GAN and the like, so that the RGB visible light image generated by the SAR image has serious distortion phenomena such as characteristic loss and the like, so that the interpretability and the reliability of the final transformation result are low, and the subsequent research and decision are influenced. Therefore, the influence of multiplicative gamma noise pollution on the SAR image on image cross-modal transformation needs to be reduced, the retention and recovery capability of the model on the main characteristics of the SAR image is improved, and the finally transformed visible light image has the semantic information of the original SAR image and certain spectral information, so that more reliable image information is provided for geological disaster monitoring, site investigation, sea area monitoring and the like; and secondly, the information of the visible light image exists in a plurality of channels, the SAR image is a gray level image, the information of the visible light image exists in one channel, channel difference and semantic gap exist between the visible light image and the SAR image, and the problems of inaccurate mapping of spectral information and semantic information and the like exist when a generator and an identifier are directly adopted to generate an anti-network for heterogeneous transformation.
In terms of generating heterogeneous image transformation of a countermeasure network based on conditions, no transformation model aiming at SAR image to visible light image (RGB) can solve the problems well at present. In the aspects of accuracy and real-time performance of heterogeneous transformation, the existing methods cannot meet the requirements of complication and diversification of remote sensing images, so that the problem needs to be solved urgently.
Disclosure of Invention
Based on the above, the invention aims to provide a heterogeneous remote sensing image transformation method and system, which improve the quality and efficiency of remote sensing image transformation.
In order to achieve the purpose, the invention provides the following scheme:
a heterogeneous remote sensing image transformation method, comprising:
denoising the plurality of SAR images to obtain a visible light gray level image of each SAR image;
obtaining a visible light color image of each SAR image;
the SAR image is used as the input of an SAR image noise reduction generator, the visible light gray level image is used as the output of the SAR image noise reduction generator, a first gradient discriminator is used for discriminating the visible light gray level image output by the SAR image noise reduction generator, and the SAR image noise reduction generator and the first gradient discriminator are trained;
training the visible light image generator and a second gradient discriminator by taking the visible light gray scale image as the input of the visible light image generator, taking the visible light color image as the output of the visible light image generator and discriminating the visible light color image output by the visible light image generator by using the second gradient discriminator;
inputting the SAR image to be transformed into the trained SAR image noise reduction generator to obtain a visible light gray image to be transformed, and inputting the visible light gray image to be transformed into the trained visible light image generator to obtain a transformed visible light color image.
Optionally, the SAR image noise reduction generator is of an improved U-Net network structure, an input layer and an output layer of the SAR image noise reduction generator are both single channels, and each adjacent single-layer convolutional layer in the SAR image noise reduction generator is connected by adopting a residual block and a jump connection.
Optionally, the visible light image generator is of a modified U-Net network structure, and the number of output layer channels of the visible light image generator is 3.
Optionally, the first gradient discriminator and the second gradient discriminator are both markov discriminators.
Optionally, the first training loss function for training the SAR image noise reduction generator and the first gradient discriminator is:
Figure BDA0002766670030000031
wherein,
Figure BDA0002766670030000032
representing conditional generation-countering network loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000033
representing a reconstruction loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image,
Figure BDA0002766670030000034
representing a texture retention loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000035
representing the SAR image noise reduction generator, x(i)Representing the set of SAR images, y(i)Representing a set of visible light grayscale images;
the second training loss function training the visible light image generator and the second gradient discriminator is:
Figure BDA0002766670030000041
wherein,
Figure BDA0002766670030000042
representing conditional generation-countering network loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA0002766670030000043
representing a reconstruction loss of the visible light image generator from inputting the visible light gray scale image to outputting a visible light color image;
Figure BDA0002766670030000044
representing a texture-preserving loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA0002766670030000045
representing a visible light image generator, z(i)Representing a set of visible color images;
α1denotes a first weight parameter, β1Representing a second weight parameter, α2Denotes a third weight parameter, β2Representing a fourth weight parameter;
stopping training the SAR image noise reduction generator, the first gradient discriminator, the visible light image generator and the second gradient discriminator when the sum of the first training loss function and the second training loss function is less than a set threshold.
The invention also discloses a heterogeneous remote sensing image transformation system, which comprises:
and the visible light gray level image acquisition module is used for denoising the SAR images to obtain the visible light gray level image of each SAR image.
The visible light color image acquisition module is used for acquiring a visible light color image of each SAR image;
the first training module is used for taking the SAR image as the input of the SAR image noise reduction generator, taking the visible light gray image as the output of the SAR image noise reduction generator, identifying the visible light gray image output by the SAR image noise reduction generator by using a first gradient identifier, and training the SAR image noise reduction generator and the first gradient identifier;
a second training module, which takes the visible light gray scale image as the input of a visible light image generator, takes the visible light color image as the output of the visible light image generator, and uses a second gradient discriminator to discriminate the visible light color image output by the visible light image generator, so as to train the visible light image generator and the second gradient discriminator;
and the SAR image conversion module is used for inputting the SAR image to be converted into the trained SAR image noise reduction generator to obtain the visible light gray image to be converted, and inputting the visible light gray image to be converted into the trained visible light image generator to obtain the converted visible light color image.
Optionally, the SAR image noise reduction generator is of an improved U-Net network structure, an input layer and an output layer of the SAR image noise reduction generator are both single channels, and each adjacent single-layer convolutional layer in the SAR image noise reduction generator is connected by adopting a residual block and a jump connection.
Optionally, the visible light image generator is of a modified U-Net network structure, and the number of output layer channels of the visible light image generator is 3.
Optionally, the first gradient discriminator and the second gradient discriminator are both markov discriminators.
Optionally, the first training loss function for training the SAR image noise reduction generator and the first gradient discriminator is:
Figure BDA0002766670030000051
wherein,
Figure BDA0002766670030000052
representing conditional generation-countering network loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000053
representing a reconstruction loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image,
Figure BDA0002766670030000054
representing a texture retention loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000055
representing the SAR image noise reduction generator, x(i)Representing the set of SAR images, y(i)A set of visible light grayscale images is represented,
the second training loss function training the visible light image generator and the second gradient discriminator is:
Figure BDA0002766670030000056
wherein,
Figure BDA0002766670030000057
representing conditional generation-countering network loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA0002766670030000058
representing a reconstruction loss of the visible light image generator from inputting the visible light gray scale image to outputting a visible light color image;
Figure BDA0002766670030000059
representing a texture-preserving loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA00027666700300000510
representing a visible light image generator, z(i)Representing a set of visible color images;
α1denotes a first weight parameter, β1Representing a second weight parameter, α2Denotes a third weight parameter, β2Representing a fourth weight parameter;
stopping training the SAR image noise reduction generator, the first gradient discriminator, the visible light image generator and the second gradient discriminator when the sum of the first training loss function and the second training loss function is less than a set threshold.
According to the specific embodiment provided by the invention, the invention discloses the following technical effects:
the invention discloses a heterogeneous remote sensing image transformation method and a system, wherein a visible light gray image is set as a middle state from an SAR image to a visible light color image, the SAR image is generated into the visible light gray image through an SAR image noise reduction generator, the visible light gray image is identified through a first identifier, the visible light gray image is generated into the visible light color image through the visible light image generator, the visible light color image is identified through a second identifier, the distortion rate from the SAR image to the visible light color image is reduced, and the quality and the efficiency of remote sensing image transformation are improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without inventive exercise.
FIG. 1 is a schematic flow chart of a heterogeneous remote sensing image transformation method of the invention;
FIG. 2 is a schematic diagram of the heterogeneous image transformation from SAR images to visible color images of the present invention;
FIG. 3 is a schematic diagram of a network structure of an SAR image noise reduction generator according to the present invention;
FIG. 4 is a schematic diagram of a network structure of a visible light image generator according to the present invention;
FIG. 5 is a diagram of a first authenticator network architecture in accordance with the present invention;
FIG. 6 is a diagram of a second discriminator network according to the invention;
FIG. 7 is a schematic structural diagram of a heterogeneous remote sensing image transformation system according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The invention aims to provide a heterogeneous remote sensing image transformation method and a heterogeneous remote sensing image transformation system, which improve the quality and efficiency of remote sensing image transformation.
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
Fig. 1 is a schematic flow chart of a heterogeneous remote sensing image transformation method of the present invention, and as shown in fig. 1, the heterogeneous remote sensing image transformation method includes:
step 101: and denoising the plurality of SAR images to obtain a visible light gray level image of each SAR image.
Step 102: and obtaining a visible light color image of each SAR image.
Step 103: and taking the SAR image as the input of an SAR image noise reduction generator, taking the visible light gray image as the output of the SAR image noise reduction generator, identifying the visible light gray image output by the SAR image noise reduction generator by using a first gradient identifier, and training the SAR image noise reduction generator and the first gradient identifier.
Step 104: and training the visible light image generator and a second gradient discriminator by taking the visible light gray scale image as the input of the visible light image generator, taking the visible light color image as the output of the visible light image generator, and discriminating the visible light color image output by the visible light image generator by using the second gradient discriminator.
Step 105: inputting the SAR image to be transformed into the trained SAR image noise reduction generator to obtain a visible light gray image to be transformed, and inputting the visible light gray image to be transformed into the trained visible light image generator to obtain a transformed visible light color image.
The SAR image noise reduction generator is of an improved U-Net network structure, an input layer and an output layer of the SAR image noise reduction generator are single channels, and each adjacent single-layer convolution layer in the SAR image noise reduction generator is connected through a residual block and a jump connection.
The visible light image generator is of an improved U-Net network structure, and the number of output layer channels of the visible light image generator is 3.
The first gradient discriminator and the second gradient discriminator are both Markov discriminators.
The first training loss function for training the SAR image noise reduction generator and the first gradient discriminator is as follows:
Figure BDA0002766670030000071
wherein,
Figure BDA0002766670030000072
representing conditional generation-countering network loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000073
representing a reconstruction loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image,
Figure BDA0002766670030000074
representing a texture retention loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000075
representing the SAR image noise reduction generator, x(i)Representing the set of SAR images, y(i)Representing a set of visible light grayscale images.
The second training loss function training the visible light image generator and the second gradient discriminator is:
Figure BDA0002766670030000081
wherein,
Figure BDA0002766670030000082
representing conditional generation-countering network loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA0002766670030000083
representing a reconstruction loss of the visible light image generator from inputting the visible light gray scale image to outputting a visible light color image;
Figure BDA0002766670030000084
representing a texture-preserving loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA00027666700300000814
to representVisible light image generator, z(i)Representing a collection of visible color images.
α1Denotes a first weight parameter, β1Representing a second weight parameter, α2Denotes a third weight parameter, β2Representing a fourth weight parameter.
Stopping training the SAR image noise reduction generator, the first gradient discriminator, the visible light image generator and the second gradient discriminator when the sum of the first training loss function and the second training loss function is less than a set threshold.
The following is a detailed description of a heterogeneous remote sensing image transformation method of the present invention.
As shown in FIG. 2, for ease of illustration, the following convention is made: let X, Y, Z denote the SAR image domain, the visible light grayscale image domain (intermediate visible light image domain) and the visible light color image domain, x, respectively(i)∈X、y(i)∈Y、z(i)E Z are the dataset samples for the corresponding image domain, respectively (i ═ 1,2, …, N denotes the total number of samples of the dataset). Therefore, the transformation of the SAR image into a visible light grayscale image (intermediate-state visible light image) can be defined as a mapping transformation of the source domain X into the target domain Y
Figure BDA0002766670030000085
Let x be(i)Is a random sample taken from the SAR image domain X and has a distribution function of
Figure BDA0002766670030000086
x(i)The random sample mapped to the visible light image domain is y(i). The remote sensing image transformation network provided by the invention is finally realized
Figure BDA0002766670030000087
Final z(i)Is derived from x by a deep learning network(i)Generating and obtaining:
Figure BDA0002766670030000088
remote sensing image changeThe switching network comprises two generators and two discriminators which are respectively used for reducing noise of the SAR image
Figure BDA0002766670030000089
Visible light image generator
Figure BDA00027666700300000815
Gradient discriminator
Figure BDA00027666700300000810
(first gradient discriminator) and gradient discriminator
Figure BDA00027666700300000811
(second gradient discriminator). SAR image noise reduction generator
Figure BDA00027666700300000812
Short for generator
Figure BDA00027666700300000813
Visible light image generator
Figure BDA00027666700300000817
Short for generator
Figure BDA00027666700300000816
Generator
Figure BDA0002766670030000091
The SAR image is converted into a visible light gray level image, so that high-frequency details and low-frequency semantic information of the SAR image are recovered, and a semantic gap from the SAR image to the visible light image is reduced; generator
Figure BDA0002766670030000097
Converting the visible light gray level image into a visible light color image with spectral information, so that the recovered visible light color image has the semantic information of the original SAR image domain and the spectral details of the visible light image; discriminator
Figure BDA0002766670030000092
The gradient details of the visible light gray scale image in the data set Y are used for identification, so that the generator
Figure BDA0002766670030000093
Obtaining stronger noise reduction capability of the SAR image, and simultaneously reserving semantic information in the original SAR image; discriminator
Figure BDA0002766670030000094
And discriminator
Figure BDA0002766670030000095
Similarly, the gradient information of the visible color image is used for discrimination, thus the generator
Figure BDA0002766670030000098
The resulting image allows for better recovery of spectral information while preserving detailed information.
As shown in fig. 3, the generator
Figure BDA0002766670030000096
By adopting an improved U-Net network structure, an input layer and an output layer are single channels, and each adjacent single-layer convolutional layer is connected by a Residual Block and a Skip Connection, so that the learning capability of the generator can be fully enhanced while the network depth is expanded. Meanwhile, instead of converting a multiplicative speckle model into an additive speckle model through input logarithmic conversion, a residual error learning strategy the same as SAR-DRN (SAR-Dual regression networks) is adopted, and a generator
Figure BDA00027666700300000911
The output layer is Gamma noise extracted from the input image, and the clear visible light gray image can be obtained by subtracting the input image from the output image.
As shown in fig. 4, the generator
Figure BDA0002766670030000099
And generator
Figure BDA00027666700300000912
The structure of the network is similar, an improved U-Net network structure is adopted, an input layer is a single Channel, the difference is that the number of output layer channels of the network is 3(Channel is 3), and the input layer is a generator
Figure BDA00027666700300000913
And generating a visible light gray scale image. In principle, the generator
Figure BDA00027666700300000914
The learning is the mapping from the SAR image contaminated by Gamma noise to the visible light gray scale image, its essence is the noise reduction of the image, and the generator
Figure BDA00027666700300000910
Learned is the mapping from the visible light gray-scale image to the visible light color image. Therefore, the progressive cross-mode transformation can obtain a better transformation effect, the expression capability of the model from SAR image reconstruction with less information amount to visible light color image is enhanced by increasing the layer number of the network, and the heterogeneous transformation of the image is realized.
As shown in fig. 5-6, discriminator D1And discriminator D2The Markov judgers are adopted, which are completely composed of convolution layers, and finally, an n-by-n matrix is output. The input of the Markov discriminator is 256 multiplied by 256 to randomly cut a 70 multiplied by 70 image Patch, the output of the Markov discriminator is the mean value of an output matrix (4 multiplied by 4), the final discrimination result is True/False, the True represents that the picture generated by the generator cheats the discriminator, the expected effect is achieved, and the training of the generator is finished; "False" means that the picture generated by the generator has not "spoofed" the discriminator, the generated picture is not as expected, and the generator needs to continue training. Since each output in the output matrix represents a field of view in the original image and corresponds to a slice (patch) of the original image, the GAN constructed in this way may be called patch GAN. In a different aspect, the discriminator
Figure BDA0002766670030000109
Is for discriminating the generator
Figure BDA00027666700300001010
Generated image (True/False), and discriminator
Figure BDA00027666700300001011
Is for discriminating the generator
Figure BDA0002766670030000102
The resulting image (True/False). In the heterogeneous image transformation process, the migration transformation of the content part and the texture part is included. The content portion is the similarity in content (semantic) between the generated image and the original image, and the texture portion is the similarity in texture between the generated image and the target image. Theory shows that with PatchGAN, texture detail can be maintained.
In the remote sensing image transformation network training stage, a generator is trained firstly
Figure BDA00027666700300001012
And a discriminator
Figure BDA00027666700300001013
Generator
Figure BDA00027666700300001014
Intermediate-state visible light image deception discriminator generated by learning speckle noise removal and super-resolution reconstruction (equal-scale) mapping of SAR image from paired SAR and visible light gray level image
Figure BDA00027666700300001015
Making it impossible to distinguish between the visible light gray scale image and the generated intermediate state image; and a discriminator
Figure BDA00027666700300001016
Judging whether the intermediate state image generated by the generator is close to the visible light corresponding to the original input SAR image or notA grayscale image. Then, the generator is fixed
Figure BDA00027666700300001025
And a discriminator
Figure BDA00027666700300001026
Is a training generator
Figure BDA0002766670030000103
And a discriminator
Figure BDA00027666700300001017
Generator
Figure BDA0002766670030000104
And learning the spectral information mapping of the visible light image from the paired visible light gray level image and visible light color image to complete the coloring function of the image. During the training process, the generator
Figure BDA00027666700300001019
And a discriminator
Figure BDA00027666700300001018
After the balance is achieved, the generator is carried out
Figure BDA0002766670030000105
And a discriminator
Figure BDA00027666700300001024
And (4) training.
A testing phase, the input SAR image is processed by a generator
Figure BDA00027666700300001020
Generating a denoised intermediate visible light gray scale image, and inputting the image to a generator
Figure BDA0002766670030000106
Via a generator
Figure BDA0002766670030000107
And generating a visible light color image, and completing the conversion from the SAR image to the visible light image.
Design for loss function
1) cGAN (Conditional generational adaptive networks) loss
Under the premise that paired training data exists, the cGAN is adopted by the method to improve the performance of the generator. By generating a countermeasure network, a generator
Figure BDA00027666700300001021
Will learn the mapping from X to Y and spoof the discriminator
Figure BDA00027666700300001023
Generator
Figure BDA0002766670030000108
Will learn the mapping from Y to Z and spoof the discriminator
Figure BDA00027666700300001022
The network structure of the discriminator has a function of distinguishing a forged image from an actual image. Thus, the cGAN losses from X to Y and Y to Z are:
Figure BDA0002766670030000101
Figure BDA0002766670030000111
in the first half of training, when the discriminator
Figure BDA00027666700300001113
Learn to make
Figure BDA0002766670030000112
When maximizing, the generator
Figure BDA00027666700300001114
Attempt to make
Figure BDA0002766670030000113
Minimized until both reach an optimal state (nash equilibrium); in the second half of training, when the discriminator
Figure BDA00027666700300001115
Learn to make
Figure BDA00027666700300001110
When maximizing, the generator
Figure BDA00027666700300001112
Attempt to make
Figure BDA00027666700300001111
Minimized until both reach an optimum (Nash equilibrium),
Figure BDA00027666700300001118
representing a mathematical expectation, a smaller mathematical expectation is better, and a smaller mathematical expectation is less loss.
2) Loss of reconstruction
Because the invention adopts the improved U-Net structure, the introduced residual block and jump connection increase the network depth, so the L is adopted in the reconstruction loss design1The loss is used for minimizing the distance between the reconstructed image and the original image, and the reconstruction losses of the two training stages are respectively:
Figure BDA0002766670030000114
Figure BDA0002766670030000115
3) texture retention loss
As shown in fig. 2, in the optimum conversion state
Figure BDA00027666700300001116
Output of the network
Figure BDA00027666700300001117
Should approximate the visible light image y(i). The invention provides gradient-guided texture retention loss on the basis of the traditional reconstruction loss. Assuming that M (-) represents an image gradient operation, the texture retention penalty for the two training phases is:
Figure BDA0002766670030000116
Figure BDA0002766670030000117
for the image I, M (-) the operation of extracting the image gradient is calculated as follows:
Ix(x,y)=I(x+1,y)-I(x-1,y),
Iy(x,y)=I(x,y+1)-I(x,y-1),
Figure BDA0002766670030000118
Figure BDA0002766670030000119
in a specific implementation, the M (-) operation is implemented by convolution with a fixed convolution kernel, and (x, y) represents the pixel coordinates.
Thus, the first half of the model training phase, the generator
Figure BDA00027666700300001119
And a discriminator
Figure BDA00027666700300001120
Formation of antagonistic networks GAN1The total training loss is:
Figure BDA0002766670030000121
the second half of the model training phase, the generator
Figure BDA0002766670030000129
And a discriminator
Figure BDA00027666700300001211
Formation of antagonistic networks GAN2The total training loss is:
Figure BDA0002766670030000122
wherein,
Figure BDA0002766670030000123
representing conditional generation-countering network loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000124
representing a reconstruction loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image,
Figure BDA0002766670030000125
representing a texture retention loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA00027666700300001212
representing the SAR image noise reduction generator, x(i)Representing the set of SAR images, y(i)Representing a set of visible light grayscale images.
Figure BDA0002766670030000126
Representing the visible light image generationGenerating-countering network loss from a condition of inputting the visible light gray-scale image to outputting a visible light color image;
Figure BDA0002766670030000127
representing a reconstruction loss of the visible light image generator from inputting the visible light gray scale image to outputting a visible light color image;
Figure BDA0002766670030000128
representing a texture-preserving loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA00027666700300001210
representing a visible light image generator, z(i)Representing a collection of visible color images.
α1Denotes a first weight parameter, β1Representing a second weight parameter, α2Denotes a third weight parameter, β2Representing a fourth weight parameter.
The technical effect of the heterogeneous remote sensing image transformation method of the invention comprises the following points:
1. a heterogeneous remote sensing image transformation method constructs a weighted target function with reconstruction loss, cGAN loss and texture preserving gradient loss, realizes natural transformation from an SAR image to a visible light image by weighting three constraint conditions of comprehensive semantic information, style characteristics and one-dimensional gradient characteristics, has obvious advantages in overcoming noise and distortion, and has strong adaptability, wide application range and strong robustness.
2. A heterogeneous remote sensing image transformation method is characterized in that a residual block (Residualblock) and a Skip Connection (Skip Connection) are used for improving U-Net and a generator for an SAR image to a visible light gray image and a generator for a visible light gray image to a visible light color image for the first time, so that the network depth and the multi-scale feature extraction capability and the mapping capability are increased, the mapping transformation capability of a single-channel gray remote sensing image to an RGB channel remote sensing image is remarkably improved, and the spectrum mapping distortion is greatly reduced.
3. A heterogeneous remote sensing image transformation method optimally designs a total objective function, and based on a traditional method, a result generated by one-dimensional gradient feature constraint retains the detail features of an original SAR image, and improves the quality of heterogeneous transformation.
4. A heterogeneous remote sensing image transformation method overcomes the distortion phenomenon of SAR image to RGB visible light image transformation by adjusting the final transformation effect of network structure constraint aiming at the characteristic of asymmetric information between an SAR image and a visible light color image.
Fig. 7 is a schematic structural diagram of a heterogeneous remote sensing image transformation system according to the present invention, and as shown in fig. 7, the heterogeneous remote sensing image transformation system includes:
a visible light grayscale image obtaining module 201, configured to denoise multiple SAR images to obtain a visible light grayscale image of each SAR image;
a visible light color image obtaining module 202, configured to obtain a visible light color image of each SAR image;
the first training module 203 is configured to use the SAR image as an input of a SAR image noise reduction generator, use the visible light grayscale image as an output of the SAR image noise reduction generator, and use a first gradient discriminator to discriminate the visible light grayscale image output by the SAR image noise reduction generator, so as to train the SAR image noise reduction generator and the first gradient discriminator;
a second training module 204, configured to train the visible light image generator and the second gradient discriminator by using the visible light grayscale image as an input of the visible light image generator, using the visible light color image as an output of the visible light image generator, and discriminating the visible light color image output by the visible light image generator by using a second gradient discriminator;
the SAR image conversion module 205 is configured to input the SAR image to be converted into the trained SAR image noise reduction generator to obtain a visible light grayscale image to be converted, and input the visible light grayscale image to be converted into the trained visible light image generator to obtain a converted visible light color image.
The SAR image noise reduction generator is of an improved U-Net network structure, an input layer and an output layer of the SAR image noise reduction generator are single channels, and each adjacent single-layer convolution layer in the SAR image noise reduction generator is connected through a residual block and a jump connection.
The visible light image generator is of an improved U-Net network structure, and the number of output layer channels of the visible light image generator is 3.
The first gradient discriminator and the second gradient discriminator are both Markov discriminators.
The first training loss function for training the SAR image noise reduction generator and the first gradient discriminator is as follows:
Figure BDA0002766670030000141
wherein,
Figure BDA0002766670030000142
representing conditional generation-countering network loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA0002766670030000148
representing a reconstruction loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image,
Figure BDA0002766670030000149
representing a texture retention loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure BDA00027666700300001410
representing the SAR image noise reduction generator, x(i)Representing the set of SAR images, y(i)A set of visible light grayscale images is represented,
the second training loss function training the visible light image generator and the second gradient discriminator is:
Figure BDA0002766670030000143
wherein,
Figure BDA0002766670030000144
representing conditional generation-countering network loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA0002766670030000145
representing a reconstruction loss of the visible light image generator from inputting the visible light gray scale image to outputting a visible light color image;
Figure BDA0002766670030000146
representing a texture-preserving loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure BDA0002766670030000147
representing a visible light image generator, z(i)Representing a set of visible color images;
α1denotes a first weight parameter, β1Representing a second weight parameter, α2Denotes a third weight parameter, β2Representing a fourth weight parameter;
stopping training the SAR image noise reduction generator, the first gradient discriminator, the visible light image generator and the second gradient discriminator when the sum of the first training loss function and the second training loss function is less than a set threshold.
The embodiments in the present description are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other. For the system disclosed by the embodiment, the description is relatively simple because the system corresponds to the method disclosed by the embodiment, and the relevant points can be referred to the method part for description.
The principles and embodiments of the present invention have been described herein using specific examples, which are provided only to help understand the method and the core concept of the present invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, the specific embodiments and the application range may be changed. In view of the above, the present disclosure should not be construed as limiting the invention.

Claims (10)

1. A heterogeneous remote sensing image transformation method is characterized by comprising the following steps:
denoising the plurality of SAR images to obtain a visible light gray level image of each SAR image;
obtaining a visible light color image of each SAR image;
the SAR image is used as the input of an SAR image noise reduction generator, the visible light gray level image is used as the output of the SAR image noise reduction generator, a first gradient discriminator is used for discriminating the visible light gray level image output by the SAR image noise reduction generator, and the SAR image noise reduction generator and the first gradient discriminator are trained;
training the visible light image generator and a second gradient discriminator by taking the visible light gray scale image as the input of the visible light image generator, taking the visible light color image as the output of the visible light image generator and discriminating the visible light color image output by the visible light image generator by using the second gradient discriminator;
inputting the SAR image to be transformed into the trained SAR image noise reduction generator to obtain a visible light gray image to be transformed, and inputting the visible light gray image to be transformed into the trained visible light image generator to obtain a transformed visible light color image.
2. The heterogeneous remote sensing image transformation method according to claim 1, wherein the SAR image noise reduction generator is of an improved U-Net network structure, an input layer and an output layer of the SAR image noise reduction generator are both single channels, and each adjacent single-layer convolution layer in the SAR image noise reduction generator is connected through a residual block and a jump connection.
3. The heterogeneous remote sensing image transformation method according to claim 1, wherein the visible light image generator is of an improved U-Net network structure, and the number of output layer channels of the visible light image generator is 3.
4. The heterogeneous remote sensing image transformation method of claim 1, wherein the first gradient discriminator and the second gradient discriminator are both Markov discriminators.
5. The heterogeneous remote sensing image transformation method of claim 1, wherein a first training loss function for training the SAR image noise reduction generator and the first gradient discriminator is:
Figure FDA0002766670020000011
wherein,
Figure FDA0002766670020000012
representing conditional generation-countering network loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure FDA0002766670020000021
representing a reconstruction loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image,
Figure FDA0002766670020000022
representing the SAR image noise reduction generator slave inputTexture retention loss from entering the SAR image to outputting a visible light gray level image;
Figure FDA0002766670020000023
representing the SAR image noise reduction generator, x(i)Representing the set of SAR images, y(i)A set of visible light grayscale images is represented,
the second training loss function training the visible light image generator and the second gradient discriminator is:
Figure FDA0002766670020000024
wherein,
Figure FDA0002766670020000025
representing conditional generation-countering network loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure FDA0002766670020000026
representing a reconstruction loss of the visible light image generator from inputting the visible light gray scale image to outputting a visible light color image;
Figure FDA0002766670020000027
representing a texture-preserving loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure FDA0002766670020000028
representing a visible light image generator, z(i)Representing a set of visible color images;
α1denotes a first weight parameter, β1Representing a second weight parameter, α2Denotes a third weight parameter, β2Representing a fourth weight parameter;
stopping training the SAR image noise reduction generator, the first gradient discriminator, the visible light image generator and the second gradient discriminator when the sum of the first training loss function and the second training loss function is less than a set threshold.
6. A heterogeneous remote sensing image transformation system, the system comprising:
the visible light gray level image acquisition module is used for denoising the SAR images to obtain visible light gray level images of the SAR images;
the visible light color image acquisition module is used for acquiring a visible light color image of each SAR image;
the first training module is used for taking the SAR image as the input of the SAR image noise reduction generator, taking the visible light gray image as the output of the SAR image noise reduction generator, identifying the visible light gray image output by the SAR image noise reduction generator by using a first gradient identifier, and training the SAR image noise reduction generator and the first gradient identifier;
the second training module is used for taking the visible light gray-scale image as the input of the visible light image generator, taking the visible light color image as the output of the visible light image generator, and identifying the visible light color image output by the visible light image generator by using a second gradient identifier to train the visible light image generator and the second gradient identifier;
and the SAR image conversion module is used for inputting the SAR image to be converted into the trained SAR image noise reduction generator to obtain the visible light gray image to be converted, and inputting the visible light gray image to be converted into the trained visible light image generator to obtain the converted visible light color image.
7. The heterogeneous remote sensing image transformation system of claim 6, wherein the SAR image noise reduction generator is of an improved U-Net network structure, an input layer and an output layer of the SAR image noise reduction generator are both single channels, and each adjacent single-layer convolution layer in the SAR image noise reduction generator is connected through a residual block and a jump connection.
8. The heterogeneous remote sensing image transformation system of claim 6, wherein the visible light image generator is of a modified U-Net network structure, and the number of output layer channels of the visible light image generator is 3.
9. The heterogeneous remote sensing image transformation system of claim 6, wherein the first gradient discriminator and the second gradient discriminator are both Markov discriminators.
10. The heterogeneous remote sensing image transformation system of claim 6, wherein the first training loss function for training the SAR image noise reduction generator and the first gradient discriminator is:
Figure FDA0002766670020000031
wherein,
Figure FDA0002766670020000032
representing conditional generation-countering network loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image;
Figure FDA0002766670020000033
representing a reconstruction loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image,
Figure FDA0002766670020000034
representing a texture retention loss of the SAR image noise reduction generator from inputting the SAR image to outputting a visible light grayscale image; p denotes the SAR image noise reduction generator, x(i)Representing the set of SAR images, y(i)A set of visible light grayscale images is represented,
the second training loss function training the visible light image generator and the second gradient discriminator is:
Figure FDA0002766670020000035
wherein,
Figure FDA0002766670020000036
representing conditional generation-countering network loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image;
Figure FDA0002766670020000037
representing a reconstruction loss of the visible light image generator from inputting the visible light gray scale image to outputting a visible light color image;
Figure FDA0002766670020000041
representing a texture-preserving loss of the visible light image generator from inputting the visible light grayscale image to outputting a visible light color image; q denotes a visible light image generator, z(i)Representing a set of visible color images;
α1denotes a first weight parameter, β1Representing a second weight parameter, α2Denotes a third weight parameter, β2Representing a fourth weight parameter;
stopping training the SAR image noise reduction generator, the first gradient discriminator, the visible light image generator and the second gradient discriminator when the sum of the first training loss function and the second training loss function is less than a set threshold.
CN202011236018.3A 2020-11-09 2020-11-09 Heterogeneous remote sensing image transformation method and system Active CN112330562B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011236018.3A CN112330562B (en) 2020-11-09 2020-11-09 Heterogeneous remote sensing image transformation method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011236018.3A CN112330562B (en) 2020-11-09 2020-11-09 Heterogeneous remote sensing image transformation method and system

Publications (2)

Publication Number Publication Date
CN112330562A true CN112330562A (en) 2021-02-05
CN112330562B CN112330562B (en) 2022-11-15

Family

ID=74316926

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011236018.3A Active CN112330562B (en) 2020-11-09 2020-11-09 Heterogeneous remote sensing image transformation method and system

Country Status (1)

Country Link
CN (1) CN112330562B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113012251A (en) * 2021-03-17 2021-06-22 厦门大学 SAR image automatic colorization method based on generation countermeasure network
CN113554671A (en) * 2021-06-23 2021-10-26 西安电子科技大学 Method and device for converting SAR image into visible light image based on contour enhancement

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105976345A (en) * 2016-04-27 2016-09-28 北京航空航天大学 Visible light remote sensing image synthesis method
CN109636742A (en) * 2018-11-23 2019-04-16 中国人民解放军空军研究院航空兵研究所 The SAR image of network and the mode conversion method of visible images are generated based on confrontation
CN110473142A (en) * 2019-05-22 2019-11-19 南京理工大学 Single image super resolution ratio reconstruction method based on deep learning
CN110751699A (en) * 2019-10-15 2020-02-04 西安电子科技大学 Color reconstruction method of optical remote sensing image based on convolutional neural network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105976345A (en) * 2016-04-27 2016-09-28 北京航空航天大学 Visible light remote sensing image synthesis method
CN109636742A (en) * 2018-11-23 2019-04-16 中国人民解放军空军研究院航空兵研究所 The SAR image of network and the mode conversion method of visible images are generated based on confrontation
CN110473142A (en) * 2019-05-22 2019-11-19 南京理工大学 Single image super resolution ratio reconstruction method based on deep learning
CN110751699A (en) * 2019-10-15 2020-02-04 西安电子科技大学 Color reconstruction method of optical remote sensing image based on convolutional neural network

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHENG MA ET AL.: ""Structure-Preserving Super Resolution with Gradient Guidance"", 《ARXIV》 *
LONG GUO ET AL.: ""Gray Scale Image Coloring Method Based on GAN"", 《DSIT 2020: PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON DATA SCIENCE AND INFORMATION》 *
PHILLIP ISOLA ET AL.: ""Image-to-Image Translation with Conditional Adversarial Networks"", 《IEEE》 *
PUYANG WANG ET AL.: ""Generating high quality visible images from SAR"", 《IEEE》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113012251A (en) * 2021-03-17 2021-06-22 厦门大学 SAR image automatic colorization method based on generation countermeasure network
CN113012251B (en) * 2021-03-17 2022-05-03 厦门大学 SAR image automatic colorization method based on generation countermeasure network
CN113554671A (en) * 2021-06-23 2021-10-26 西安电子科技大学 Method and device for converting SAR image into visible light image based on contour enhancement

Also Published As

Publication number Publication date
CN112330562B (en) 2022-11-15

Similar Documents

Publication Publication Date Title
CN110956094B (en) RGB-D multi-mode fusion personnel detection method based on asymmetric double-flow network
CN111539316B (en) High-resolution remote sensing image change detection method based on dual-attention twin network
Lei et al. Multiscale superpixel segmentation with deep features for change detection
Liu et al. Remote sensing image change detection based on information transmission and attention mechanism
CN109871902B (en) SAR small sample identification method based on super-resolution countermeasure generation cascade network
CN114841972A (en) Power transmission line defect identification method based on saliency map and semantic embedded feature pyramid
CN109558806A (en) The detection method and system of high score Remote Sensing Imagery Change
CN109741340B (en) Ice cover radar image ice layer refined segmentation method based on FCN-ASPP network
CN113838064B (en) Cloud removal method based on branch GAN using multi-temporal remote sensing data
CN109635726B (en) Landslide identification method based on combination of symmetric deep network and multi-scale pooling
Wang et al. The poor generalization of deep convolutional networks to aerial imagery from new geographic locations: an empirical study with solar array detection
CN114255403A (en) Optical remote sensing image data processing method and system based on deep learning
CN112330562B (en) Heterogeneous remote sensing image transformation method and system
CN113887472B (en) Remote sensing image cloud detection method based on cascade color and texture feature attention
CN117765361B (en) Method for detecting building change area in double-time-phase remote sensing image based on residual neural network
CN115049841A (en) Depth unsupervised multistep anti-domain self-adaptive high-resolution SAR image surface feature extraction method
CN109977968A (en) A kind of SAR change detecting method of deep learning classification and predicting
CN105512622A (en) Visible remote-sensing image sea-land segmentation method based on image segmentation and supervised learning
Ge et al. Coarse-to-fine foraminifera image segmentation through 3D and deep features
CN115861756A (en) Earth background small target identification method based on cascade combination network
CN114842524A (en) Face false distinguishing method based on irregular significant pixel cluster
Zhang et al. Spatial–temporal gray-level co-occurrence aware CNN for SAR image change detection
CN116168240A (en) Arbitrary-direction dense ship target detection method based on attention enhancement
CN112784777B (en) Unsupervised hyperspectral image change detection method based on countermeasure learning
CN117994240A (en) Multi-scale two-level optical remote sensing image stripe noise intelligent detection method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant