CN112163998A - Single-image super-resolution analysis method matched with natural degradation conditions - Google Patents

Single-image super-resolution analysis method matched with natural degradation conditions Download PDF

Info

Publication number
CN112163998A
CN112163998A CN202011015611.5A CN202011015611A CN112163998A CN 112163998 A CN112163998 A CN 112163998A CN 202011015611 A CN202011015611 A CN 202011015611A CN 112163998 A CN112163998 A CN 112163998A
Authority
CN
China
Prior art keywords
image
resolution
network
super
resolution analysis
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011015611.5A
Other languages
Chinese (zh)
Inventor
陈伦强
龙学军
陈东文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhaoqing Boshixin Electronic Technology Co ltd
Original Assignee
Zhaoqing Boshixin Electronic Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhaoqing Boshixin Electronic Technology Co ltd filed Critical Zhaoqing Boshixin Electronic Technology Co ltd
Priority to CN202011015611.5A priority Critical patent/CN112163998A/en
Publication of CN112163998A publication Critical patent/CN112163998A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention discloses a single image super-resolution analysis method matched with natural degradation conditions, which comprises a domain learning network and a super-resolution analysis learning network, wherein the domain learning network comprises a generator network GdAnd discriminator network DxThe super-resolution analysis learning network comprises a super-resolution analysis network GSRAnd discriminator network Dy. Generator network G based on deep learning trainingdTo generate a low resolution image such that it matches the naturally degraded low resolution source domain image x, and through a super resolution analysis network GSRLow resolution through use of generation in generating a countermeasure network framework
Figure DDA0002698956920000011
The image and the corresponding high-resolution image are trained in a clean high-resolution target domain y by pixel-by-pixel supervision, residual error learning is carried out on the image blocks in the high-resolution domain from the generated low resolution by a minimized energy loss function, and countertraining is carried out by pixel-level supervision, so that an image degradation process matched with reality is realized.

Description

Single-image super-resolution analysis method matched with natural degradation conditions
Technical Field
The invention relates to the technical field of image processing, in particular to a single-image super-resolution analysis method matched with natural degradation conditions.
Background
The goal of single image super resolution analysis (SISR) is to recover a High Resolution (HR) image from a Low Resolution (LR) observed image. The single image super resolution problem is a fundamental low-level vision and image processing problem, and has various practical applications in satellite imaging, medical imaging, astronomy, microscopic imaging, seismology, remote sensing, surveillance, biometric identification, image compression, and the like. Recently, many works have solved the task of single-graph super-resolution analysis due to the powerful functional representation capabilities of deep CNNs. Lim et al in 2017 proposed an enhanced deep super resolution analysis network (EDSR) using residual learning. Wang et al in 2018 proposed an enhanced super-resolution analysis-generated countermeasure network model (ESRGAN) to achieve the most advanced perceptual performance at that time, trained on two-three downsampling through HR/LR images, and therefore limited performance in a practical setting. Fritsche et al, 2019, proposed to train the network to learn the degradation mechanism in an unsupervised manner and modify the enhanced super-resolution analysis generation countermeasure network into frequency-separated super-resolution analysis (SR-FS) to further improve its performance in real environments [3 ]. These single image super resolution analysis methods based on deep learning learn a non-linear mapping between low resolution input and high resolution output from a large number of pairs of low resolution/high resolution (LR/HR) training data. The high-resolution images are actually shot high-resolution images, paired low-resolution images are generally obtained by a bicubic down-sampling degradation operation of the high-resolution images, and natural image features cannot be well summarized by using a known bicubic down-sampling degradation model. Therefore, when such degradation process is compared with the actual image damage, such as inherent sensor noise, random noise and compression artifacts, the image degradation process does not match with these natural degradation conditions, resulting in limited performance of super-resolution analysis, and therefore we propose a single-image super-resolution analysis method matching with the natural degradation conditions to solve the above problems.
Disclosure of Invention
The invention aims to provide a single-image super-resolution analysis method matched with natural degradation conditions, so as to solve the problems in the background technology.
In order to achieve the purpose, the invention provides the following technical scheme: a single image super-resolution analysis method matched with natural degradation conditions comprises a domain learning network and a super-resolution analysis learning network, wherein the domain learning network comprises a generator network GdAnd discriminator network DxThe super-resolution analysis learning network comprises a super-resolution analysis network GSRAnd discriminator network DyGenerator network G based on deep learning trainingdAs a sampling network to generate a low resolution image such that it matches, i.e. has the same characteristics, a naturally degraded low resolution source domain image x, and through a super resolution analysis network GSRLow resolution through use of generation in generating a countermeasure network framework
Figure BDA0002698956900000021
The image and the corresponding high-resolution image are trained in a clean high-resolution target domain y by pixel supervision, so that the mapping from a bicubic down-sampled image z of the high-resolution image y to a source domain image x is learned, the image content is kept, residual error learning is carried out on image blocks in the high-resolution domain from the generated low resolution, countertraining is carried out by pixel-level supervision, and an image degradation process matched with reality is realized.
In a preferred embodiment, the model of the degradation process is as follows:
Figure BDA0002698956900000022
wherein the content of the first and second substances,
Figure BDA0002698956900000023
is a low-resolution image obtained by the degradation,
Figure BDA0002698956900000024
is a downsampling operator which scales a high resolution image by a scaling factor s and an additive white gaussian noise η
Figure BDA0002698956900000025
Resizing, where the standard deviation of additive white gaussian noise is σ, N × N is the total number of pixels in the image.
In a preferred embodiment, the model of the degradation process is a training process, whereas in the application process, Y is an observed degraded low-resolution image, and the high-resolution image X is restored from the observed low-resolution image Y by using a variation method combining the observation and the prior information, and the corresponding loss function is as follows:
Figure BDA0002698956900000031
wherein, therein
Figure BDA0002698956900000032
Is a data fidelity term that measures the closeness of a solution to an observation, RW(X) is a regularization term associated with the image prior information and is a trade-off parameter that controls the trade-off between data fidelity and regularization conditions.
In a preferred embodiment, the regularizer used by the regularization term is of the general form:
Figure BDA0002698956900000033
where L corresponds to a first or higher order differential linear operator and ρ (-) represents a function such as L applied to the filtered outputpA potential function of the vector or matrix norm, regularizer RW(X) the parameter is expressed by W in the deep convolutional neural network, so that strong image prior capability can be obtained.
In a preferred embodiment, the energy loss minimization function uses a suitable optimization strategy to find W, and minimizes the energy-based loss function, and the method comprises the following steps:
1) the high resolution image X is restored by the minimization method:
Figure BDA0002698956900000035
by referring to equations (2) and (3), it is written as:
Figure BDA0002698956900000036
2) the image intensity needs to be limited, the values in the natural image need to be non-negative, so (5) it needs to be rewritten in a constraint optimization form:
Figure BDA0002698956900000037
and solving equation (6) by an approximate gradient method to process optimization of the function which cannot be completely differentiated, dividing the function into a smooth part and a non-smooth part, and rewriting equation (6) as follows:
Figure BDA0002698956900000041
wherein, iotacIs a convex set
Figure BDA0002698956900000042
The trainable projection layer calculates a near-end mapping for the index function:
Figure BDA0002698956900000043
wherein
Figure BDA0002698956900000044
Is a parameterized threshold, where α is a trainable parameter, σ is the noise level, C H W is the total number of pixels in the image;
3) performing an iterative solution to the problem in (7) by updating the rules:
Figure BDA0002698956900000045
wherein, γtIs the step size of the frame,
Figure BDA0002698956900000046
is related to the index function iotacA related near-end operator defined as:
Figure BDA0002698956900000047
the gradient of F (X) is calculated as follows:
Figure BDA0002698956900000048
wherein phik(. is a potential function ρ)kGradient of (·);
4) by combining equations (8), (9) and (10), the final form is:
Figure BDA0002698956900000049
formula (11) is at the starting points Y and X(0)A near gradient descent inference step is performed at 0, which has the formula:
Figure BDA00026989569000000410
where α ═ λ γ corresponds to the projection layer trainable parameters, Lk TIs LkIs associated with the filter, and HTIndicating up-scaling operations
In a preferred embodiment, step 4), the formula (12) is also used for designing the generator network GdWherein phik(. to) corresponds to a parameterized rectified linear unit (PReLU) applied to a convolutional signature.
In a preferred embodiment, the generatorNetwork of generators GdTraining by adopting a loss function, so as to learn the image degradation of the domain distribution layer from the source domain x and train the generator network G under the framework of the countermeasure networkdThe loss function of (a) is as follows:
Figure BDA0002698956900000051
wherein L iscolorRepresents the color loss, LtexRepresents the loss of texture, LperIndicating a loss of perception.
In an embodiment, the super-resolution analysis network GSRThe network has the following loss function:
Figure BDA0002698956900000052
wherein the perception loss LperFocusing on the perceptual quality of the output image, it is defined as:
Figure BDA0002698956900000053
where φ is a feature extracted from the pre-trained network;
texture loss LGANFocusing on the high frequency part of the output image, it is defined as:
Figure BDA0002698956900000054
wherein E isyAnd
Figure BDA0002698956900000055
respectively representing all true y sums generated for small image blocks
Figure BDA0002698956900000056
The data averaging operation, then the relative generation of true high-resolution and reconstructed high-resolution image blocks opposes the network score may be defined as:
Figure BDA0002698956900000057
where C is the raw discriminator output and is a sigmoid (sigmoid) function;
content loss L1Is defined as:
Figure BDA0002698956900000058
where N is the size of the small image block;
total variation loss Ltv: focusing on minimizing gradient differences and producing sharpness in the output image, which is defined as;
Figure BDA0002698956900000061
wherein the content of the first and second substances,
Figure BDA0002698956900000062
and
Figure BDA0002698956900000063
representing horizontal and vertical gradations of the image.
In a preferred embodiment, the discriminator network DxThe convolutional layers (Conv) are operated at the image block level, the convolutional layers (Conv) are provided with 64-256 feature mapping kernels supporting 5 x5, the last convolutional layer maps 256 to 1, and a normalization module (BN) and a Leak activation function (LReLU) are applied after each of the rest convolutional layers.
In a preferred embodiment, the discriminator network DyThe true HR image can be distinguished from the generated pseudo-super-resolution analysis image through training, the original discriminator network contains 10 convolutional layers, the number of kernels supporting 3x3 and 4x4 feature mapping is 512, and then a normalization module (BN) and a Leak activation function are connected.
Compared with the prior art, the invention has the beneficial effects that: generating a countermeasure network through deep ultra-high resolution residual convolution, performing residual learning on image blocks in a high resolution domain through a minimized energy loss function from generated low resolution, and performing countermeasure training through pixel level supervision, thereby realizing an image degradation process matched with reality; the method can be used for improving the image quality in the real environment, strictly follows an image observation (physical) model, namely a super-resolution analysis learning method matched with a natural degraded image, so as to overcome the super-resolution challenge in the real world, and uses a large-scale optimization technology of image regularization and inverse problem solving to evaluate the method on a plurality of data sets of synthesized and natural degraded images, and the result shows that the method provided by the inventor is superior to other methods under the natural degraded condition of the real world.
Drawings
FIG. 1 is a schematic structural view of the present invention;
FIG. 2 is a super-resolution analysis network G in the present inventionSRA schematic structural diagram;
FIG. 3 is a diagram of an authenticator network D of the present inventionyThe structure is schematic.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1-3, the present invention provides a technical solution: a single image super-resolution analysis method matched with natural degradation conditions comprises a domain learning network and a super-resolution analysis learning network, wherein the domain learning network comprises a generator network GdAnd discriminator network DxThe super-resolution analysis learning network comprises a super-resolution analysis network GSRAnd discriminator network DySince there is no pair of (low resolution/high resolution) data, the generator network G under deep learning training is first baseddAs a sampling network to generate a low resolution image such that the low resolution image and a naturally degraded low resolution imageThe source domain images x are matched, i.e. have the same characteristics, and are analyzed by the super-resolution analysis network GSRLow resolution through use of generation in generating a countermeasure network framework
Figure BDA0002698956900000071
The image and the corresponding high-resolution image are trained in a clean high-resolution target domain y by pixel supervision, so that the mapping from a bicubic down-sampled image z of the high-resolution image y to a source domain image x is learned, the image content is kept, residual error learning is carried out on image blocks in the high-resolution domain from the generated low resolution, countertraining is carried out by pixel-level supervision, and an image degradation process matched with reality is realized.
Learning a non-linear mapping between a low resolution input and a high resolution output from a large number of pairs of low resolution/high resolution (LR/HR) training data, the high resolution images being high resolution images taken from reality, the pairs of low resolution images being typically obtained from the high resolution images by a degeneration operation, the model of the degeneration process being as follows:
Figure BDA0002698956900000072
wherein the content of the first and second substances,
Figure BDA0002698956900000073
is a low-resolution image obtained by the degradation,
Figure BDA0002698956900000074
is a downsampling operator which scales a high resolution image by a scaling factor s and an additive white gaussian noise η
Figure BDA0002698956900000081
Resizing, where the standard deviation of additive white gaussian noise is σ, N × N is the total number of pixels in the image.
Further, the model of the degradation process is a training process, whereas in the application process, Y is an observed degraded low-resolution image, and a high-resolution image X is restored from the observed low-resolution image Y using a variational method combining observation and prior information, and the corresponding loss function is as follows:
Figure BDA0002698956900000082
wherein, therein
Figure BDA0002698956900000083
Is a data fidelity term that measures the closeness of a solution to an observation, RW(X) is a regularization term associated with the image prior information and is a trade-off parameter that controls the trade-off between data fidelity and regularization conditions.
Further, the regularizer used by the regularization term is of the general form:
Figure BDA0002698956900000084
where L corresponds to a first or higher order differential linear operator and ρ (-) represents a function such as L applied to the filtered outputpA potential function of the vector or matrix norm, regularizer RW(X) the parameter is expressed by W in the deep convolutional neural network, so that strong image prior capability can be obtained.
Further, in addition to properly formulating the loss function and the regularization function, the variational method used in the method also needs to obtain a minimization strategy of the required solution, the derived solution can be described as a penalty maximum likelihood estimation (p) or a Maximum A Posteriori (MAP) estimation, the minimization energy loss function adopts a proper optimization strategy to find W, and the energy-based loss function is minimized, and the method comprises the following steps:
1) the high resolution image X is restored by the minimization method:
Figure BDA0002698956900000086
by referring to equations (2) and (3), it is written as:
Figure BDA0002698956900000091
2) the image intensity needs to be limited, the values in the natural image need to be non-negative, so (5) it needs to be rewritten in a constraint optimization form:
Figure BDA0002698956900000092
and solving equation (6) by an approximate gradient method to process optimization of the function which cannot be completely differentiated, dividing the function into a smooth part and a non-smooth part, and rewriting equation (6) as follows:
Figure BDA0002698956900000093
wherein, iotacIs a convex set
Figure BDA0002698956900000094
The trainable projection layer calculates a near-end mapping for the index function:
Figure BDA0002698956900000095
wherein
Figure BDA0002698956900000096
Is a parameterized threshold, where α is a trainable parameter, σ is the noise level, C H W is the total number of pixels in the image;
3) performing an iterative solution to the problem in (7) by updating the rules:
Figure BDA0002698956900000097
wherein, γtIs the step size of the frame,
Figure BDA0002698956900000098
is related to the index function iotacA related near-end operator defined as:
Figure BDA0002698956900000099
the gradient of F (X) is calculated as follows:
Figure BDA00026989569000000910
where φ k (·) is the potential function ρkGradient of (·);
4) by combining equations (8), (9) and (10), the final form is:
Figure BDA00026989569000000911
formula (11) is at the starting points Y and X(0)A near gradient descent inference step is performed at 0, which has the formula:
Figure BDA0002698956900000101
where α ═ λ γ corresponds to the projection layer trainable parameters, Lk TIs LkIs associated with the filter, and HTIndicating up-scaling operations
In step 4), the formula (12) is also used to design the generator network GdWherein phik(. to) corresponds to a parameterized rectified linear unit (PReLU) applied to a convolutional signature.
Further, most of the parameters in equation (12) come from the antecedents of equation (2), which results in the proposed generator network representing most of its parameters as image prior information, with weights set to zero mean and fixed scale constraints for learning the effective weights of the regularization parameters, and with generator network G for learning the image degradation at the domain distribution level from the source domain xdTraining by adopting a loss function, and training a generator under the condition of generating a countermeasure network frameworkNetwork GdThe loss function of (a) is as follows:
Figure BDA0002698956900000102
wherein L iscolorRepresents the color loss, LtexRepresents the loss of texture, LperIndicating a loss of perception.
Further, a generator network GdConsisting of 8 residual modules (two convolutional layers and a PReLU activation component between them) sandwiched between two convolutional layers (Conv). All convolutional layers (Conv) have 64 feature mapping kernels supporting 3 × 3, while sigmoid (sigmoid) nonlinear functions are applied to GdAnd (4) outputting of the network.
Discriminator network DxThe convolutional network is composed of three layers of convolutional networks, the convolutional layers (Conv) run at the image block level and have 64-256 feature mapping kernels supporting 5 x5, the last convolutional layer maps 256 to 1, and a normalization module (BN) and a Leak activation function (LReLU) are applied after each of the rest convolutional layers.
Training G with 512x512 image blocksdThe network is subjected to secondary down-sampling, and the source domain image (x) is randomly clipped to 128x128 size, specifically, the use parameter is beta1=0.5,β2=0.999,=10-8The Adam optimizer of (1) trains the network.
Super-resolution analysis network GSRThe network has the following loss function:
Figure BDA0002698956900000111
wherein the perception loss LperFocusing on the perceptual quality of the output image, it is defined as:
Figure BDA0002698956900000112
where φ is a feature extracted from the pre-trained network;
textureLoss LGANFocusing on the high frequency part of the output image, it is defined as:
Figure BDA0002698956900000113
wherein E isyAnd
Figure BDA0002698956900000114
respectively representing all true y sums generated for small image blocks
Figure BDA0002698956900000115
The data averaging operation, then the relative generation of true high-resolution and reconstructed high-resolution image blocks opposes the network score may be defined as:
Figure BDA0002698956900000116
where C is the raw discriminator output and is a sigmoid (sigmoid) function;
content loss L1Is defined as:
Figure BDA0002698956900000117
where N is the size of the small image block;
total variation loss Ltv: focusing on minimizing gradient differences and producing sharpness in the output image, which is defined as;
Figure BDA0002698956900000118
wherein the content of the first and second substances,
Figure BDA0002698956900000119
and
Figure BDA00026989569000001110
representing horizontal and vertical gradations of the image.
Further, a super-resolution analysis network GSRArchitecture of (1), encoder layer (L)kFilter) and decoder layers(i.e., L)k TFilters) have 64 eigenmap kernels supporting 5 x5, with a C x H x W tensor, where C is the number of channels of the input image, and inside the encoder the LR image Y is up-sampled by a bilinear kernel with an up-sampling layer (equivalent to H-upsampling layer)TY operation), wherein a random up-sampling kernel is selected, a residual network (Resnet) is composed of 5 residual modules and 2 pre-activation convolutional layers, each of which has 64 feature mapping kernels supporting 3 × 3, and pre-activation (phi)k(. DEG)) refers to a parameterized rectified linear unit (PReLU) supporting 64 functional channels, a trainable projection layer (Proj) inside the decoder, i.e. the near-end operator PC) Computing a near-end mapping using the estimated noise standard deviation sigma and processing the data fidelity and a priori terms, and during training, fine-tuned by the projection layer (Proj) parameters, noise estimation in the residual network (Resnet) sandwiched between encoder and decoder, subtraction of the estimated residual image after decoder from the LR input image, finally, the cropping layer incorporates our a priori knowledge of the effective range of image intensities and forces the pixel values of the reconstructed image to [0, 255%]Interval, reflection filling is used before all convolutional layers to ensure slow variation at the boundaries of the input image.
Discriminator network DyThe true HR image can be distinguished from the generated pseudo-super-resolution analysis image through training, the original discriminator network contains 10 convolutional layers, the number of kernels supporting 3x3 and 4x4 feature mapping is 512, and then a normalization module (BN) and a Leak activation function are connected.
Further, in training, the input LR image block size is set to 32x32 using a parameter β1=0.9,β2=0.999,=10-8The Adam optimizer network of (1) was subjected to 51000 iterative trainings, the number of samples in one training selection was 16, the weights of the generator and discriminator were not reduced to minimize the loss function in equation (14), the learning rate was initially set to 10-4, and then multiplied by 0.5 after 5K, 10K, 20K and 30K iterations, the projection layer parameter σ was estimated from the input LR image, and the projection layer parameter σ was estimated from αmax2 to aminThe projection layer parameters α are initialized on a logarithmic scale value of 1 and then in the training processFurther fine-tuned by back propagation.
Further, the method of the invention is applied to practical test application, which specifically comprises the following steps:
the trained model is evaluated under peak signal-to-noise ratio (PSNR), Structural Similarity (SSIM) and learning-perceptual image patch similarity (LPIPS) indices using source domain data (X: 2650 corrupted HR images) corrupted by unknown degradation, which may be sensor noise, compression artifacts, etc., using target domain data (Y: 800 sharp HR images), using training data augmented by random vertical and horizontal flips and 90 rotations. PSNR and SSIM are distortion-based metrics that have poor correlation to actual perception-based similarity, while LPIPS has better correlation to human perception than distortion-based metrics, comparing the method of the present invention with other most advanced super resolution analysis methods, including enhanced deep super resolution network (EDSR), enhanced super resolution generation countermeasure network (ESRGAN) and frequency separation super resolution analysis (SR-FS), the results of which are shown in the table below,
comparison results
Figure BDA0002698956900000131
The results show a quantitative result comparison of the method of the invention on the selected data sets, tested under synthetic degradation conditions (JPEG compression with 8 sensor noise and quality 30) and natural degradation conditions, respectively. Compared with other methods, the performance of the method is superior to that of other methods in both PSNR and SSIM indexes. On the other hand, the LPIPS index is not as good as SR-FS under the condition of synthetic degradation, but is superior to other methods under the condition of natural degradation in the real world.
In summary, the present invention generates a countermeasure network by a deep ultra-high resolution residual convolution, performs residual learning from the generated low resolution image blocks in a high resolution domain by minimizing an energy loss function, and performs countermeasure training by pixel-level supervision, thereby implementing an image degradation process matching reality, which can be used to improve image quality in such a real environment, strictly follows an image observation (physical) model, i.e., a super-resolution analysis learning method matching a naturally degraded image, to overcome the challenges of super-resolution in the real world, uses a large-scale optimization technique of image regularization and solving an inverse problem, evaluates the method on multiple data sets of synthesized and naturally degraded images, and results show that the method proposed by us is superior to other methods under the naturally degraded conditions in the real world.
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (10)

1. A single image super-resolution analysis method matched with natural degradation conditions comprises a domain learning network and a super-resolution analysis learning network, and is characterized in that: the domain learning network comprises a generator network GdAnd discriminator network DxThe super-resolution analysis learning network comprises a super-resolution analysis network GSRAnd discriminator network DyGenerator network G based on deep learning trainingdAs a sampling network to generate a low resolution image such that it matches, i.e. has the same characteristics, a naturally degraded low resolution source domain image x, and through a super resolution analysis network GSRLow resolution through use of generation in generating a countermeasure network framework
Figure FDA0002698956890000011
The image and the corresponding high-resolution image are trained in a clean high-resolution target domain y by pixel-by-pixel supervision, so that the mapping from a bicubic down-sampled image z of the high-resolution image y to a source domain image x is learned while the image content is retained, and residual error learning is performed on the image blocks in the high-resolution domain by minimizing an energy loss function from the generated low resolutionAnd performing countermeasure training by using pixel-level supervision to realize an image degradation process matched with reality.
2. The single-image super-resolution analysis method matching with natural degradation conditions according to claim 1, wherein: the model of the degradation process is as follows:
Figure FDA0002698956890000012
wherein the content of the first and second substances,
Figure FDA0002698956890000013
is a low-resolution image obtained by the degradation,
Figure FDA0002698956890000014
is a downsampling operator which scales a high resolution image by a scaling factor s and an additive white gaussian noise η
Figure FDA0002698956890000015
Resizing, where the standard deviation of additive white gaussian noise is σ, N × N is the total number of pixels in the image.
3. The single-image super-resolution analysis method matching the natural degradation condition according to claim 2, wherein: the model of the degradation process is a training process, otherwise, in an application process, Y is an observed degraded low-resolution image, a high-resolution image X is recovered by the observed low-resolution image Y, a variation method combining observation and prior information is used, and corresponding loss functions are as follows:
Figure FDA0002698956890000021
wherein, therein
Figure FDA0002698956890000022
Is a data fidelity term that measures the closeness of a solution to an observation, RW(X) is a regularization term associated with the image prior information and is a trade-off parameter that controls the trade-off between data fidelity and regularization conditions.
4. The single-image super-resolution analysis method matching natural degradation conditions according to claim 3, wherein: the regularizer used by the regularization term is generally of the form:
Figure FDA0002698956890000023
where L corresponds to a first or higher order differential linear operator and ρ (-) represents a function such as L applied to the filtered outputpA potential function of the vector or matrix norm, regularizer RW(X) the parameter is expressed by W in the deep convolutional neural network, so that strong image prior capability can be obtained.
5. The single-image super-resolution analysis method matching with natural degradation conditions according to claim 1, wherein: the energy loss minimization function adopts a proper optimization strategy to find W, and minimizes an energy-based loss function, and the method comprises the following steps:
1) the high resolution image X is restored by the minimization method:
Figure FDA0002698956890000024
by referring to equations (2) and (3), it is written as:
Figure FDA0002698956890000025
2) the image intensity needs to be limited, the values in the natural image need to be non-negative, so (5) it needs to be rewritten in a constraint optimization form:
Figure FDA0002698956890000026
and solving equation (6) by an approximate gradient method to process optimization of the function which cannot be completely differentiated, dividing the function into a smooth part and a non-smooth part, and rewriting equation (6) as follows:
Figure FDA0002698956890000031
wherein, iotacIs a convex set
Figure FDA0002698956890000032
The trainable projection layer calculates a near-end mapping for the index function:
Figure FDA0002698956890000033
wherein
Figure FDA0002698956890000034
Is a parameterized threshold, where α is a trainable parameter, σ is the noise level, C H W is the total number of pixels in the image;
3) performing an iterative solution to the problem in (7) by updating the rules:
Figure FDA0002698956890000035
wherein, γtIs the step size of the frame,
Figure FDA0002698956890000036
is related to the index function iotacA related near-end operator defined as:
Figure FDA0002698956890000037
the gradient of F (X) is calculated as follows:
Figure FDA0002698956890000038
wherein phik(. is a potential function ρ)kGradient of (·);
4) by combining equations (8), (9) and (10), the final form is:
Figure FDA0002698956890000039
formula (11) is at the starting points Y and X(0)A near gradient descent inference step is performed at 0, which has the formula:
Figure FDA00026989568900000310
where α ═ λ γ corresponds to the projection layer trainable parameters, Lk TIs LkIs associated with the filter, and HTShowing a scale-up operation.
6. The single-image super-resolution analysis method matching natural degradation conditions according to claim 5, wherein: in step 4), the formula (12) is also used to design the generator network GdWherein phik(. to) corresponds to a parameterized rectified linear unit (PReLU) applied to a convolutional signature.
7. The single-image super-resolution analysis method matching natural degradation conditions according to claim 6, wherein: the generator network GdTraining by adopting a loss function, so as to learn the image degradation of the domain distribution layer from the source domain x and train the generator network G under the framework of the countermeasure networkdThe loss function of (a) is as follows:
Figure FDA0002698956890000048
wherein L iscolorRepresents the color loss, LtexRepresents the loss of texture, LperIndicating a loss of perception.
8. The single-image super-resolution analysis method matching with natural degradation conditions according to claim 1, wherein: the super-resolution analysis network GSRThe network has the following loss function:
Figure FDA0002698956890000041
wherein the perception loss LperFocusing on the perceptual quality of the output image, it is defined as:
Figure FDA0002698956890000042
where φ is a feature extracted from the pre-trained network;
texture loss LGANFocusing on the high frequency part of the output image, it is defined as:
Figure FDA0002698956890000043
wherein E isyAnd
Figure FDA0002698956890000044
respectively representing all true y sums generated for small image blocks
Figure FDA0002698956890000045
The data averaging operation, then the relative generation of true high-resolution and reconstructed high-resolution image blocks opposes the network score may be defined as:
Figure FDA0002698956890000046
where C is the raw discriminator output and is a sigmoid (sigmoid) function;
content loss L1Is defined as:
Figure FDA0002698956890000047
where N is the size of the small image block;
total variation loss Ltv: focusing on minimizing gradient differences and producing sharpness in the output image, which is defined as;
Figure FDA0002698956890000051
wherein the content of the first and second substances,
Figure FDA0002698956890000052
and
Figure FDA0002698956890000053
representing horizontal and vertical gradations of the image.
9. The single-image super-resolution analysis method matching with natural degradation conditions according to claim 1, wherein: said discriminator network DxThe convolutional layers (Conv) are operated at the image block level, the convolutional layers (Conv) are provided with 64-256 feature mapping kernels supporting 5 x5, the last convolutional layer maps 256 to 1, and a normalization module (BN) and a Leak activation function (LReLU) are applied after each of the rest convolutional layers.
10. The single-image super-resolution analysis method matching with natural degradation conditions according to claim 1, wherein: said discriminator network DyThe true HR image can be distinguished from the generated pseudo-super-resolution analysis image through training, the original discriminator network comprises 10 convolutional layers and supports 3x3 and 4x4 feature mappingIs 512, after which a normalization module (BN) and a Leak activation function are connected.
CN202011015611.5A 2020-09-24 2020-09-24 Single-image super-resolution analysis method matched with natural degradation conditions Pending CN112163998A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011015611.5A CN112163998A (en) 2020-09-24 2020-09-24 Single-image super-resolution analysis method matched with natural degradation conditions

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011015611.5A CN112163998A (en) 2020-09-24 2020-09-24 Single-image super-resolution analysis method matched with natural degradation conditions

Publications (1)

Publication Number Publication Date
CN112163998A true CN112163998A (en) 2021-01-01

Family

ID=73863824

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011015611.5A Pending CN112163998A (en) 2020-09-24 2020-09-24 Single-image super-resolution analysis method matched with natural degradation conditions

Country Status (1)

Country Link
CN (1) CN112163998A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113284046A (en) * 2021-05-26 2021-08-20 中国电子科技集团公司第五十四研究所 Remote sensing image enhancement and restoration method and network based on no high-resolution reference image
CN113435334A (en) * 2021-06-28 2021-09-24 中国科学院上海微系统与信息技术研究所 Small target face recognition method based on deep learning
CN113763247A (en) * 2021-09-09 2021-12-07 中国矿业大学 Three-dimensional mine image super-resolution reconstruction method and system based on manifold discrimination

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105528595A (en) * 2016-02-01 2016-04-27 成都通甲优博科技有限责任公司 Method for identifying and positioning power transmission line insulators in unmanned aerial vehicle aerial images
US20180075581A1 (en) * 2016-09-15 2018-03-15 Twitter, Inc. Super resolution using a generative adversarial network
WO2018120329A1 (en) * 2016-12-28 2018-07-05 深圳市华星光电技术有限公司 Single-frame super-resolution reconstruction method and device based on sparse domain reconstruction
CN109978762A (en) * 2019-02-27 2019-07-05 南京信息工程大学 A kind of super resolution ratio reconstruction method generating confrontation network based on condition
CN110111251A (en) * 2019-04-22 2019-08-09 电子科技大学 A kind of combination depth supervision encodes certainly and perceives the image super-resolution rebuilding method of iterative backprojection
CN110136063A (en) * 2019-05-13 2019-08-16 南京信息工程大学 A kind of single image super resolution ratio reconstruction method generating confrontation network based on condition
CN110599401A (en) * 2019-08-19 2019-12-20 中国科学院电子学研究所 Remote sensing image super-resolution reconstruction method, processing device and readable storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105528595A (en) * 2016-02-01 2016-04-27 成都通甲优博科技有限责任公司 Method for identifying and positioning power transmission line insulators in unmanned aerial vehicle aerial images
US20180075581A1 (en) * 2016-09-15 2018-03-15 Twitter, Inc. Super resolution using a generative adversarial network
WO2018120329A1 (en) * 2016-12-28 2018-07-05 深圳市华星光电技术有限公司 Single-frame super-resolution reconstruction method and device based on sparse domain reconstruction
CN109978762A (en) * 2019-02-27 2019-07-05 南京信息工程大学 A kind of super resolution ratio reconstruction method generating confrontation network based on condition
CN110111251A (en) * 2019-04-22 2019-08-09 电子科技大学 A kind of combination depth supervision encodes certainly and perceives the image super-resolution rebuilding method of iterative backprojection
CN110136063A (en) * 2019-05-13 2019-08-16 南京信息工程大学 A kind of single image super resolution ratio reconstruction method generating confrontation network based on condition
CN110599401A (en) * 2019-08-19 2019-12-20 中国科学院电子学研究所 Remote sensing image super-resolution reconstruction method, processing device and readable storage medium

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
UMER, RAO MUHAMMAD;: "Deep Generative Adversarial Residual Convolutional Networks for Real-World Super-Resolution", 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), pages 1769 - 1777 *
丁明航;邓然然;邵恒;: "基于注意力生成对抗网络的图像超分辨率重建方法", 计算机系统应用, no. 02 *
史振威;雷森;: "图像超分辨重建算法综述", 数据采集与处理, no. 01 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113284046A (en) * 2021-05-26 2021-08-20 中国电子科技集团公司第五十四研究所 Remote sensing image enhancement and restoration method and network based on no high-resolution reference image
CN113284046B (en) * 2021-05-26 2023-04-07 中国电子科技集团公司第五十四研究所 Remote sensing image enhancement and restoration method and network based on no high-resolution reference image
CN113435334A (en) * 2021-06-28 2021-09-24 中国科学院上海微系统与信息技术研究所 Small target face recognition method based on deep learning
CN113435334B (en) * 2021-06-28 2024-02-27 中国科学院上海微系统与信息技术研究所 Small target face recognition method based on deep learning
CN113763247A (en) * 2021-09-09 2021-12-07 中国矿业大学 Three-dimensional mine image super-resolution reconstruction method and system based on manifold discrimination
CN113763247B (en) * 2021-09-09 2023-12-05 中国矿业大学 Three-dimensional mine image super-resolution reconstruction method based on manifold discrimination

Similar Documents

Publication Publication Date Title
Molini et al. Deepsum: Deep neural network for super-resolution of unregistered multitemporal images
Zhang et al. Image restoration: From sparse and low-rank priors to deep priors [lecture notes]
CN110119780B (en) Hyper-spectral image super-resolution reconstruction method based on generation countermeasure network
McCann et al. Convolutional neural networks for inverse problems in imaging: A review
Jiang et al. Single image super-resolution via locally regularized anchored neighborhood regression and nonlocal means
CN106952228B (en) Super-resolution reconstruction method of single image based on image non-local self-similarity
CN109741256A (en) Image super-resolution rebuilding method based on rarefaction representation and deep learning
CN111062880A (en) Underwater image real-time enhancement method based on condition generation countermeasure network
CN111080567A (en) Remote sensing image fusion method and system based on multi-scale dynamic convolution neural network
CN112163998A (en) Single-image super-resolution analysis method matched with natural degradation conditions
CN110232653A (en) The quick light-duty intensive residual error network of super-resolution rebuilding
CN113284051B (en) Face super-resolution method based on frequency decomposition multi-attention machine system
CN113222836B (en) Hyperspectral and multispectral remote sensing information fusion method and system
CN112270644A (en) Face super-resolution method based on spatial feature transformation and cross-scale feature integration
Ye et al. CSformer: Bridging convolution and transformer for compressive sensing
CN112837224A (en) Super-resolution image reconstruction method based on convolutional neural network
Guan et al. Srdgan: learning the noise prior for super resolution with dual generative adversarial networks
Rasheed et al. LSR: Lightening super-resolution deep network for low-light image enhancement
CN116029902A (en) Knowledge distillation-based unsupervised real world image super-resolution method
Lin et al. A Deep Neural Network Based on Prior Driven and Structural-Preserving for SAR Image Despeckling
Zhang et al. Enhanced visual perception for underwater images based on multistage generative adversarial network
Zheng et al. Overwater image dehazing via cycle-consistent generative adversarial network
CN113240581A (en) Real world image super-resolution method for unknown fuzzy kernel
CN111899166A (en) Medical hyperspectral microscopic image super-resolution reconstruction method based on deep learning
Wali et al. Recent Progress in Digital Image Restoration Techniques: A Review

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination