WO2021056969A1 - Super-resolution image reconstruction method and device - Google Patents

Super-resolution image reconstruction method and device Download PDF

Info

Publication number
WO2021056969A1
WO2021056969A1 PCT/CN2020/077874 CN2020077874W WO2021056969A1 WO 2021056969 A1 WO2021056969 A1 WO 2021056969A1 CN 2020077874 W CN2020077874 W CN 2020077874W WO 2021056969 A1 WO2021056969 A1 WO 2021056969A1
Authority
WO
WIPO (PCT)
Prior art keywords
generator
image
input
discriminator
output
Prior art date
Application number
PCT/CN2020/077874
Other languages
French (fr)
Chinese (zh)
Inventor
王永成
张欣
张宁
徐东东
王晓东
Original Assignee
中国科学院长春光学精密机械与物理研究所
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国科学院长春光学精密机械与物理研究所 filed Critical 中国科学院长春光学精密机械与物理研究所
Publication of WO2021056969A1 publication Critical patent/WO2021056969A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4046Scaling the whole image or part thereof using neural networks

Definitions

  • the invention belongs to the fields of computer vision, deep learning, and image processing, and particularly relates to a super-resolution image reconstruction method and device.
  • remote sensing imaging equipment usually can only obtain remote sensing images with low spatial resolution.
  • the improvement of the accuracy of imaging equipment, the reduction of image size, and the increase of the number of pixels collected per unit area are all restricted by the current manufacturing level. It is difficult to improve the image resolution from the hardware, and it takes a long time and is extremely costly.
  • the super-resolution reconstruction method based on deep learning can learn the end-to-end mapping relationship between low-resolution images and high-resolution images through a large number of sample training, which can more fully learn the prior information in low-resolution images and reconstruct The speed is fast and the effect is good.
  • the embodiments of the present invention provide a super-resolution image reconstruction method and device, which can effectively solve the problem of difficulty in reconstruction caused by the unsatisfactory resolution of the reference image in the prior art.
  • the first aspect of the embodiments of the present invention provides a super-resolution image reconstruction method, including:
  • Construct a generative confrontation network wherein the generative confrontation network includes at least a generator, a discriminator, and a loss function calculator; construct an original resolution image training set; train the generative confrontation network based on the original resolution image training set, wherein, The original resolution image training set is preprocessed as the first input of the discriminator, and the output of the generator is used as the second input of the discriminator; using the trained generator of the generated confrontation network, Reconstruct the input image.
  • a second aspect of the embodiments of the present invention provides a super-resolution image reconstruction device, which is characterized by comprising: a first building module to construct a generative confrontation network, wherein the generative confrontation network at least includes: a generator and a discriminator And loss function calculator; a second construction module, which constructs an original resolution image training set; a training module, which trains the generative confrontation network based on the original resolution image training set, wherein the original resolution image training set is pre-processed Processing is used as the first input of the discriminator, and the output of the generator is used as the second input of the discriminator; the reconstruction module uses the trained generator of the generated confrontation network to reconstruct the input image .
  • the constructed generation confrontation network includes a generator, a discriminator, and a loss function calculator, and the generation confrontation network is trained using the original resolution image training set, Therefore, it is possible to use the original resolution image as a reference image to reconstruct a super-resolution image, so that even if the original resolution is a lower resolution, a higher resolution image can still be reconstructed.
  • FIG. 1A is a schematic flowchart of a super-resolution image reconstruction method according to an embodiment of the present invention.
  • FIG. 1B is a flowchart of specific training steps for generating a confrontation network model according to an embodiment of the present invention.
  • Fig. 2 is the flow of super-resolution reconstruction performed by unsupervised learning in the present invention.
  • Fig. 3 is a generator structure in a generating confrontation network in an embodiment of the present invention.
  • Fig. 4 is the structure of the discriminator in the generated confrontation network in the embodiment of the present invention.
  • Fig. 5 is a schematic block diagram of a super-resolution image reconstruction device according to an embodiment of the present invention.
  • Figure 6 is an evaluation diagram of the result of ⁇ 2 multiple reconstruction on the test set.
  • Figure 7 is an evaluation diagram of the result of ⁇ 4 multiple reconstruction on the test set.
  • Figure 8 is a visual quality effect diagram of a ⁇ 2 reconstructed picture.
  • Figure 9 is a visual quality effect diagram of a ⁇ 4 reconstructed picture.
  • FIG. 1A is a schematic flowchart of a super-resolution image reconstruction method according to an embodiment of the present invention.
  • the super-resolution image reconstruction method 100 of FIG. 1A includes:
  • the generative confrontation network at least includes a generator, a discriminator, and a loss function calculator.
  • the original resolution image in the embodiment of the present invention may include any resolution image.
  • the technical effect of the embodiment of the present invention can be optimized. That is, a data set that includes a considerable number of high-resolution images is also within the scope of the embodiment of the present invention to achieve the purpose of reconstructing a higher-resolution image, which is not limited in the embodiment of the present invention.
  • the images in the training set according to the embodiment of the present invention may be any images, preferably remote sensing images. As a specific embodiment, in order to achieve the optimal technical effect, for example, as shown in FIG.
  • the training set is a low-resolution remote sensing image data set constructed by down-sampling of UCM, NWPU, and WHU-RS19.
  • the test set is the data set of the super-resolution image evaluation MATLAB toolbox provided by Jaime I University in Spain.
  • the original resolution image training set is input into the generator after preprocessing, and may be input into the generator after interpolation processing, and the embodiment of the present invention does not limit various interpolation methods.
  • bicubic interpolation processing can be performed.
  • the convolutional neural network may include an average pooling layer for down-sampling processing.
  • the image reconstruction device that implements the image reconstruction method can be a server, can be used as a component in a remote sensing device, or can be used as an independent processing device connected to the remote sensing device, which is not limited in the embodiment of the present invention.
  • a PyTorch development environment is built; the hardware platform of the present invention is based on the Intel E5-2690V3 processor, TITAN V GPU, and 64G memory.
  • the experimental software platform is based on the Urbanu16.04 version, using CUDA 9.0, CUDNN 7 and PyTorch 1.0.0 environment.
  • the generator in the embodiment of the present invention may not include a down sampler, that is, during training, the output of the down sampler is input to the discriminator.
  • the generator may include a down-sampler (for example, an average pooling layer), and the output of the generator is input to the discriminator.
  • the training in the embodiment of the present invention adopts the idea of generating a confrontation network, but it is applicable to a variety of architectures of generating a confrontation network. Therefore, any scheme that uses the original resolution image training set as the label of the discriminator for reconstruction is described in the embodiment of the present invention. In the range.
  • the constructed generation confrontation network includes a generator, a discriminator, and a loss function calculator, and the generation confrontation network is trained using the original resolution image training set, Therefore, it is possible to use the original resolution image as a reference image to reconstruct a super-resolution image, so that even if the original resolution is a lower resolution, a higher resolution image can still be reconstructed.
  • the network model trained with natural image is not suitable for direct application to remote sensing image super-resolution reconstruction, and migration learning still requires high-resolution remote sensing image tags; therefore To fundamentally solve the problem of fewer high-resolution image sources for remote sensing images, unsupervised learning methods can be used.
  • the Generative Adversarial Network can conduct adversarial training with the generator by adding a discriminator, so as to achieve the purpose of generating pictures with good quality.
  • the discriminator can discriminate the generated picture and the label picture as much as possible through training, and the generator can generate pictures that the discriminator cannot discriminate as much as possible through the training process.
  • the discriminator can indirectly serve as a loss function to improve the learning ability of the generator.
  • Applying generative confrontation network to image super-resolution reconstruction (SRGAN model) can make the reconstructed image have more texture detail information, and the visual effect is better than other deep learning models.
  • Remote sensing images are usually acquired from a long distance and contain complex scene and feature information. Therefore, in the process of remote sensing image reconstruction, the requirements for detail reconstruction are higher.
  • the introduction of GAN into remote sensing image super-resolution reconstruction can effectively improve the reconstruction. effect.
  • constructing a generative confrontation network includes: constructing a generator using a convolutional neural network.
  • the convolutional neural network includes multiple convolutional layers with successively decreasing convolution kernels, each of which The size of the convolution kernel of each convolutional layer is larger than the step size.
  • the generator structure adopts a codec structure constructed by a convolutional layer and a deconvolutional layer, which can extract and restore features, and introduce skip links to integrate the low-level features into the high-level output, effectively retaining the original picture Low frequency information.
  • the use of a convolutional neural network to construct a generator includes: adding an average pooling layer for downsampling after the output layer of the convolutional neural network, where the average pooling layer
  • the output image of the transformation layer is used as the second input of the discriminator.
  • the input image is reconstructed by the generator of the generated confrontation network after training, including: inputting the input image to the generator, and obtaining the reconstructed image from the output layer .
  • another layer of average pooling is added to the output layer of the network ( ⁇ 4 multiples are two layers of average pooling) to achieve the purpose of down-sampling the reconstructed image.
  • the final output of the generator is the reconstructed picture I SR and the down-sampled reconstructed picture I SR' .
  • the first input of the discriminator corresponds to the first output
  • the second input corresponds to the second output.
  • the construction of a generative confrontation network includes: using a generator loss function calculator , So as to feed back the generator loss function to the generator, where the generator loss function includes at least the cross-entropy loss between the first output and the second output.
  • the generator loss function L G of the embodiment of the present invention may include the following parts: image loss L image , perceptual loss L VGG , confrontation loss L Adv and total variation loss L TV .
  • the calculation formula of the generator loss function is:
  • L G L image +2 ⁇ 10 -3 ⁇ L VGG +L Adv +2 ⁇ 10 -8 ⁇ L TV (1)
  • the image loss is the L1 norm of I SR and I SR':
  • r is the feature extraction ratio
  • W is the pixel-based image width
  • H is the pixel-based image height.
  • VGG VisualGeometryGroup, visual geometry group
  • W' is the width of the image based on pixels
  • H' is the height of the image based on pixels.
  • the confrontation loss is between D(I LR ) and real label 1, D(I SR' ) and false label 0, and D(I LR ) and D(I SR' )
  • the total variation loss is the mean square value of the difference between the horizontal and vertical directions of the image:
  • I x, y represents the xth pixel on the abscissa, and the yth pixel on the ordinate.
  • the discriminator loss in the embodiment of the present invention is:
  • constructing a generative confrontation network includes: constructing a discriminator using a batch normalization layer, and the batch normalization layer is placed between the dense layer of the discriminator and the excitation function layer.
  • the network model can be trained and tested.
  • the Adam optimizer can be used to optimize the loss function.
  • the input data adopts 128*128 randomly cropped image blocks, which can play a certain role in data enhancement.
  • the learning rate is set to 5*10-4, and the Batch_size size is 32.
  • an evaluation index can be used to evaluate the reconstructed image.
  • the embodiment of the present invention uses SAM (spectral angle mapping), RMSE (root mean square error), ERGAS (integrated dimensionless overall relative error), sCC (spatial correlation coefficient), Qindex (quality index), SSIM (structure similarity) ), PSNR (Peak Signal-to-Noise Ratio) is used as an evaluation index to comprehensively evaluate the reconstructed image.
  • SAM spectral angle mapping
  • RMSE root mean square error
  • ERGAS integrated dimensionless overall relative error
  • sCC spatial correlation coefficient
  • Qindex quality index
  • SSIM structure similarity
  • PSNR Peak Signal-to-Noise Ratio
  • Figure 2 is a flow chart of realizing super-resolution reconstruction of remote sensing images through unsupervised learning.
  • the reconstructed image is obtained, and then down-sampled to the same size as the low-resolution image LR, and then input to the discriminator to make it discriminate between LR and SR', and find the loss between the two to update the network parameters.
  • Figure 3 is the generator structure for generating the confrontation network in the present invention.
  • the generator has a codec structure of a convolution layer and a deconvolution layer, and the convolution layer and the deconvolution layer have a skip connection.
  • the generator includes a series of convolutional layers and deconvolutional layers.
  • the three convolution kernels are 7, 5, and 3 convolutional layers (connected to ReLU (linear rectification unit)) for feature extraction.
  • ReLU linear rectification unit
  • 4 consecutive 3 ⁇ 3 convolutional layers for image feature coding then 4 layers of deconvolution are used for feature decoding, and 3, 5, and 7 three-layer deconvolutions are used for feature restoration, and finally a reconstructed image is obtained .
  • a jump connection is added between the input and the output to preserve the low-frequency information.
  • Fig. 4 is the discriminator in the generated confrontation network of the present invention. After inputting the image, the discriminator goes through a series of convolution, Leaky ReLU and BN layers to extract the features of the input image. The final part is densely connected and the sigmoid function is input to obtain the score value of the input image, and the score between the input images The difference.
  • Fig. 5 is a schematic block diagram of an image reconstruction device according to an embodiment of the present invention.
  • the super-resolution image reconstruction device 500 of FIG. 5 includes:
  • the first building module 510 is to build a generative confrontation network, where the generative confrontation network at least includes a generator, a discriminator, and a loss function calculator;
  • the second construction module 520 constructs an original resolution image training set
  • the training module 530 trains and generates a confrontation network based on the original resolution image training set, where the original resolution image training set is preprocessed as the first input of the discriminator, and the output of the generator is used as the second input of the discriminator;
  • the reconstruction module 540 uses the trained generator to generate the confrontation network to reconstruct the input image.
  • the constructed generation confrontation network includes a generator, a discriminator, and a loss function calculator, and the generation confrontation network uses the original resolution image training set for training, so it can
  • the original resolution image is used as the reference image to reconstruct the super-resolution image, so that even if the original resolution is a lower resolution, a higher resolution image can still be reconstructed.
  • the first building module is specifically used to: construct a generator using a convolutional neural network, the convolutional neural network including a plurality of convolutional layers with successively decreasing convolution kernels, each of which The size of the convolution kernel of the convolution layer is larger than the step size.
  • the first building module is specifically used to add an average pooling layer for downsampling after the output layer of the convolutional neural network, wherein the image output from the average pooling layer
  • the reconstruction module is specifically used to input the input image to the generator, and obtain the reconstructed image from the output layer.
  • the first input of the discriminator corresponds to the first output
  • the second input corresponds to the second output.
  • the first building module is specifically used for: using a generator loss function calculator, In order to feed back the generator loss function to the generator, the generator loss function includes at least the cross-entropy loss between the first output and the second output.
  • the first construction module is specifically used to construct a discriminator using a batch normalization layer, and the batch normalization layer is placed between the dense layer of the discriminator and the excitation function layer.
  • Fig. 6 is an average value of various evaluation indexes obtained by performing a ⁇ 2 multiple test on a test set in an embodiment of the present invention.
  • the figure contains 12 unsupervised learning methods for image super-resolution reconstruction. The closer the SAM value is to 0, the better, the smaller the RMSE, the better, the smaller the ERGAS, the better, the closer the sCC is to 1, the larger the Q Well, the closer the SSIM is to 1, the better, and the larger the PSNR, the better. From the figures and the values in the table, it can be seen that, among the above indicators, the reconstructed image of the present invention can obtain the best evaluation result (ERGAS result ranks second), which proves that the network performs unsupervised learning to reconstruct the image. Effectiveness.
  • Fig. 7 is the average value of each evaluation index obtained by the ⁇ 4 multiple test on the test set. It can be seen from the figures and the numerical values in the table that, among the various evaluation indicators, the method achieves the best evaluation results, which proves the effectiveness of the unsupervised learning reconstruction results of the present invention at a larger magnification.
  • Figure 8 is a visual display of the results of the test set ⁇ 2 multiples in the embodiment of the present invention.
  • Figure (a) is the original high-resolution image
  • Figure (b) is the image after bicubic interpolation of the input network
  • Figure (c) The image reconstructed by this method can be seen from the figure that there are many details reconstructed by this method, which is close to a high-resolution image.
  • Figure 9 is a visual display of the results of the test set ⁇ 2 multiples in the embodiment of the present invention.
  • Figure (a) is the original high-resolution image
  • Figure (b) is the image after bicubic interpolation of the input network
  • Figure (c) For the image reconstructed by this method, it can be seen from the figure that the reconstructed image by this method can achieve good visual effects, which is closer to high-resolution images.
  • the disclosed device/terminal device and method may be implemented in other ways.
  • the device/terminal device embodiments described above are merely illustrative.
  • the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units.
  • components can be combined or integrated into another system, or some features can be omitted or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
  • the functional units in the various embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
  • the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
  • the present invention implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program.
  • the computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, it can implement the steps of the foregoing method embodiments.
  • the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
  • the computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signal, telecommunications signal, and software distribution media, etc.
  • ROM Read-Only Memory
  • RAM Random Access Memory
  • electrical carrier signal telecommunications signal
  • software distribution media etc.
  • the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of the legislation and patent practice in the jurisdiction.
  • the computer-readable medium Does not include electrical carrier signals and telecommunication signals.

Abstract

A super-resolution image reconstruction method, comprising: constructing a generative adversarial network; constructing an original resolution image training set; training the generative adversarial network on the basis of the original resolution image training set; and reconstructing an input image by using a generator of the trained generative adversarial network. In the super-resolution image reconstruction method and device in embodiments of the present invention, the constructed generative adversarial network comprises a generator, a discriminator, and a loss function calculator. The generative adversarial network is trained by adopting an original resolution image training set, so that the original resolution image can be used as a reference image to reconstruct a super-resolution image, and even if the original resolution is relatively low, a relatively high-resolution image can still be reconstructed.

Description

超分辨率图像重构方法和装置Super-resolution image reconstruction method and device
本申请要求于2019年9月29日提交中国专利局、申请号为201910935280.8、名称为“超分辨率图像重构方法和装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Chinese Patent Office on September 29, 2019, with application number 201910935280.8 and titled "Super-resolution image reconstruction method and device", the entire content of which is incorporated into this application by reference in.
技术领域Technical field
本发明属于计算机视觉、深度学习、图像处理领域,尤其涉及一种超分辨率图像重构方法和装置。The invention belongs to the fields of computer vision, deep learning, and image processing, and particularly relates to a super-resolution image reconstruction method and device.
背景技术Background technique
高分辨率遥感图像在资源勘探、环境监测、自然灾害预防、军事侦察等方面发挥着重要的作用。然而由于信号传输带宽及成像传感器存储的限制,遥感成像设备通常只能获取低空间分辨率的遥感图像。成像设备精度的提高、图像尺寸的缩小、单位面积内像素采集量的增加,都受到了当前制造水平的制约,从硬件上提高图像分辨率较为困难,且耗时较长,成本极高。High-resolution remote sensing images play an important role in resource exploration, environmental monitoring, natural disaster prevention, and military reconnaissance. However, due to the limitation of signal transmission bandwidth and imaging sensor storage, remote sensing imaging equipment usually can only obtain remote sensing images with low spatial resolution. The improvement of the accuracy of imaging equipment, the reduction of image size, and the increase of the number of pixels collected per unit area are all restricted by the current manufacturing level. It is difficult to improve the image resolution from the hardware, and it takes a long time and is extremely costly.
基于深度学习的超分辨率重构方法,通过大量样本训练学习低分辨率图像和高分辨率图像之间的端到端映射关系,能够更充分地学习低分辨率图像中的先验信息,重建速度快且效果好。然而目前含有大量高分辨率遥感图像的数据集很少,如何利用较少的高分辨率遥感图像训练样本使用深度学习的方法提高遥感图像的空间分辨率仍然是一个亟待解决的问题。The super-resolution reconstruction method based on deep learning can learn the end-to-end mapping relationship between low-resolution images and high-resolution images through a large number of sample training, which can more fully learn the prior information in low-resolution images and reconstruct The speed is fast and the effect is good. However, there are currently few data sets containing a large number of high-resolution remote sensing images. How to use a small number of high-resolution remote sensing image training samples to use deep learning methods to improve the spatial resolution of remote sensing images is still an urgent problem to be solved.
发明内容Summary of the invention
鉴于此,本发明实施例提供了一种超分辨率图像重构方法和装置,可以有效地解决现有技术中因参考图像分辨率不理想而带来的重构困难的问题。In view of this, the embodiments of the present invention provide a super-resolution image reconstruction method and device, which can effectively solve the problem of difficulty in reconstruction caused by the unsatisfactory resolution of the reference image in the prior art.
本发明实施例的第一方面提供了一种超分辨率图像重构方法,包括:The first aspect of the embodiments of the present invention provides a super-resolution image reconstruction method, including:
构建生成对抗网络,其中所述生成对抗网络至少包括生成器、判别器和损失函数计算器;构建原始分辨率图像训练集;基于所述原始分辨率图像训练集训练所述生成对抗网络,其中,所述原始分辨率图像训练集经由预处理作为所述判别器的第一输入,所述生成器的输出作为所述判别器的第二输入;利用训练后的所述生成对抗网络的生成器,对输入图像进行重构。Construct a generative confrontation network, wherein the generative confrontation network includes at least a generator, a discriminator, and a loss function calculator; construct an original resolution image training set; train the generative confrontation network based on the original resolution image training set, wherein, The original resolution image training set is preprocessed as the first input of the discriminator, and the output of the generator is used as the second input of the discriminator; using the trained generator of the generated confrontation network, Reconstruct the input image.
本发明实施例的第二方面提供了一种超分辨率图像重构装置,其特征在于,包括:第一构建模块,构建生成对抗网络,其中所述生成对抗网络至少包括:生成器、判别器和损失函数计算器;第二构建模块,构建原始分辨率图像训练集;训练模块,基于所述原始分辨率图像训练集训练所述生成对抗网络,其中,所述原始分辨率图像训练集经由预处理作为所述判别器的第一输入,所述生成器的输出作为所述判别器的第二输入;重构模块,利用训练后的所述生成对抗网络的生成器,对输入图像进行重构。A second aspect of the embodiments of the present invention provides a super-resolution image reconstruction device, which is characterized by comprising: a first building module to construct a generative confrontation network, wherein the generative confrontation network at least includes: a generator and a discriminator And loss function calculator; a second construction module, which constructs an original resolution image training set; a training module, which trains the generative confrontation network based on the original resolution image training set, wherein the original resolution image training set is pre-processed Processing is used as the first input of the discriminator, and the output of the generator is used as the second input of the discriminator; the reconstruction module uses the trained generator of the generated confrontation network to reconstruct the input image .
在本发明实施例的超分辨率图像重构方法和装置中,所构建的生成对抗网络包括生成器、判别器以及损失函数计算器,并且该生成对抗网络采用原始分辨率图像训练集进行训练,因此能够利用原始分辨率图像作为参考图像来重构超分辨率图像,从而即使原始分辨率为较低分辨率,仍然可以重构出较高分辨率图像。In the super-resolution image reconstruction method and device of the embodiment of the present invention, the constructed generation confrontation network includes a generator, a discriminator, and a loss function calculator, and the generation confrontation network is trained using the original resolution image training set, Therefore, it is possible to use the original resolution image as a reference image to reconstruct a super-resolution image, so that even if the original resolution is a lower resolution, a higher resolution image can still be reconstructed.
附图说明Description of the drawings
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly describe the technical solutions in the embodiments of the present invention, the following will briefly introduce the drawings that need to be used in the description of the embodiments or the prior art. Obviously, the drawings in the following description are merely of the present invention. For some embodiments, those of ordinary skill in the art can obtain other drawings based on these drawings without creative labor.
图1A是本发明实施例的超分辨率图像重构方法的示意性流程图。FIG. 1A is a schematic flowchart of a super-resolution image reconstruction method according to an embodiment of the present invention.
图1B是本发明实施例的生成对抗网络模型的具体训练步骤的流程图。FIG. 1B is a flowchart of specific training steps for generating a confrontation network model according to an embodiment of the present invention.
图2是本发明中无监督学习进行超分辨率重构的流程。Fig. 2 is the flow of super-resolution reconstruction performed by unsupervised learning in the present invention.
图3是本发明实施例中生成对抗网络中的生成器结构。Fig. 3 is a generator structure in a generating confrontation network in an embodiment of the present invention.
图4是本发明实施例中生成对抗网络中的判别器结构。Fig. 4 is the structure of the discriminator in the generated confrontation network in the embodiment of the present invention.
图5是本发明实施例的超分辨率图像重构装置的示意性框图。Fig. 5 is a schematic block diagram of a super-resolution image reconstruction device according to an embodiment of the present invention.
图6是在测试集上×2倍数重构结果评价图。Figure 6 is an evaluation diagram of the result of ×2 multiple reconstruction on the test set.
图7是在测试集上×4倍数重构结果评价图。Figure 7 is an evaluation diagram of the result of ×4 multiple reconstruction on the test set.
图8是×2重构图片的视觉质量效果图。Figure 8 is a visual quality effect diagram of a ×2 reconstructed picture.
图9是×4重构图片的视觉质量效果图。Figure 9 is a visual quality effect diagram of a ×4 reconstructed picture.
具体实施方式detailed description
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本发明实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本发明。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本发明的描述。In the following description, for the purpose of illustration rather than limitation, specific details such as a specific system structure and technology are proposed for a thorough understanding of the embodiments of the present invention. However, it should be clear to those skilled in the art that the present invention can also be implemented in other embodiments without these specific details. In other cases, detailed descriptions of well-known systems, devices, circuits, and methods are omitted to avoid unnecessary details from obstructing the description of the present invention.
为了说明本发明所述的技术方案,下面通过具体实施例来进行说明。In order to illustrate the technical solution of the present invention, specific embodiments are used for description below.
图1A是本发明实施例的超分辨率图像重构方法的示意性流程图。图1A的超分辨率图像重构方法100包括:FIG. 1A is a schematic flowchart of a super-resolution image reconstruction method according to an embodiment of the present invention. The super-resolution image reconstruction method 100 of FIG. 1A includes:
110:构建生成对抗网络,其中生成对抗网络至少包括生成器、判别器和损失函数计算器。110: Build a generative confrontation network, where the generative confrontation network at least includes a generator, a discriminator, and a loss function calculator.
120:构建原始分辨率图像训练集。120: Construct a training set of original resolution images.
应理解,本发明实施例中的原始分辨率图像可以包括任何的分辨率图像,优选地,在存在低分辨率图像时,本发明实施例的技术效果可以达到最优。即,包括相当数量的高分辨率图像的数据集也在本发明实施例的范围内,达到重构更高分辨率图像的目的,本发明实施例对此不作限定。根据本发明实施例的训练集中的图像可以为任何图像,优先地为遥感图像。作为一个具体的实施例,为达到最优技术效果,例如图1B所示,训练集为UCM、NWPU以及WHU-RS19进行下采样构建的低分辨率遥感图像数据集。测试集为西班牙海梅一世大学提供的超分辨率图像评价MATLAB工具箱的数据集。It should be understood that the original resolution image in the embodiment of the present invention may include any resolution image. Preferably, when a low-resolution image exists, the technical effect of the embodiment of the present invention can be optimized. That is, a data set that includes a considerable number of high-resolution images is also within the scope of the embodiment of the present invention to achieve the purpose of reconstructing a higher-resolution image, which is not limited in the embodiment of the present invention. The images in the training set according to the embodiment of the present invention may be any images, preferably remote sensing images. As a specific embodiment, in order to achieve the optimal technical effect, for example, as shown in FIG. 1B, the training set is a low-resolution remote sensing image data set constructed by down-sampling of UCM, NWPU, and WHU-RS19. The test set is the data set of the super-resolution image evaluation MATLAB toolbox provided by Jaime I University in Spain.
130:基于原始分辨率图像训练集训练生成对抗网络,其中,原始分辨率图像训练集经由预处理作为判别器的第一输入,生成器的输出作为判别 器的第二输入;130: Train and generate a confrontation network based on the original resolution image training set, where the original resolution image training set is preprocessed as the first input of the discriminator, and the output of the generator is used as the second input of the discriminator;
140:利用训练后的生成对抗网络的生成器,对输入图像进行重构。140: Use the trained generator of the generated confrontation network to reconstruct the input image.
应理解,原始分辨率图像训练集经过预处理之后输入到生成器中,可以是经过插值处理输入到生成器中,本发明实施例对各种插值的方式不做限定。作为一个优选地的实施例,可以进行双三次插值处理。此外,卷积神经网络可以包括平均池化层,用于进行下采样处理。It should be understood that the original resolution image training set is input into the generator after preprocessing, and may be input into the generator after interpolation processing, and the embodiment of the present invention does not limit various interpolation methods. As a preferred embodiment, bicubic interpolation processing can be performed. In addition, the convolutional neural network may include an average pooling layer for down-sampling processing.
应理解,实施该图像重构方法的图像重构装置可以为服务器,可以作为遥感装置中的部件,也可以作为与遥感装置连接的独立的处理装置,本发明实施例对此不作限定。作为一个具体的实施例,例如图1B所示,搭建PyTorch的开发环境;本发明的硬件平台是基于Intel E5-2690V3处理器,TITAN V GPU,64G内存。实验的软件平台是基于Ubantu16.04版本,采用CUDA 9.0、CUDNN 7以及PyTorch1.0.0环境。It should be understood that the image reconstruction device that implements the image reconstruction method can be a server, can be used as a component in a remote sensing device, or can be used as an independent processing device connected to the remote sensing device, which is not limited in the embodiment of the present invention. As a specific embodiment, for example, as shown in FIG. 1B, a PyTorch development environment is built; the hardware platform of the present invention is based on the Intel E5-2690V3 processor, TITAN V GPU, and 64G memory. The experimental software platform is based on the Urbanu16.04 version, using CUDA 9.0, CUDNN 7 and PyTorch 1.0.0 environment.
应理解,本发明实施例中的生成器可以不包括下采样器,即在训练时,将下采样器的输出输入到判别器。或者,生成器可以包括下采样器(例如,平均池化层),将生成器的输出输入到判别器。本发明实施例的训练采用生成对抗网络的思想,但适用于多种生成对抗网络的架构,因此,利用原始分辨率图像训练集作为判别器的标签进行重构的任何方案都在本发明实施例的范围内。It should be understood that the generator in the embodiment of the present invention may not include a down sampler, that is, during training, the output of the down sampler is input to the discriminator. Alternatively, the generator may include a down-sampler (for example, an average pooling layer), and the output of the generator is input to the discriminator. The training in the embodiment of the present invention adopts the idea of generating a confrontation network, but it is applicable to a variety of architectures of generating a confrontation network. Therefore, any scheme that uses the original resolution image training set as the label of the discriminator for reconstruction is described in the embodiment of the present invention. In the range.
因此,在本发明实施例的超分辨率图像重构方法中,所构建的生成对抗网络包括生成器、判别器以及损失函数计算器,并且该生成对抗网络采用原始分辨率图像训练集进行训练,因此能够利用原始分辨率图像作为参考图像来重构超分辨率图像,从而即使原始分辨率为较低分辨率,仍然可以重构出较高分辨率图像。换句话说,例如,利用本发明实施例构建带有下采样过程的生成对抗网络模型,作为较优选的实施例,通过将较低分辨率图像作为标签,即使用下采样后的重构图片与较低分辨率图片求损失以更新网络参数,达到无监督超分辨率重构的目的。Therefore, in the super-resolution image reconstruction method of the embodiment of the present invention, the constructed generation confrontation network includes a generator, a discriminator, and a loss function calculator, and the generation confrontation network is trained using the original resolution image training set, Therefore, it is possible to use the original resolution image as a reference image to reconstruct a super-resolution image, so that even if the original resolution is a lower resolution, a higher resolution image can still be reconstructed. In other words, for example, using the embodiment of the present invention to construct a generative confrontation network model with a down-sampling process, as a more preferred embodiment, by using a lower resolution image as a label, that is, using the down-sampled reconstructed image and The lower resolution picture seeks the loss to update the network parameters to achieve the purpose of unsupervised super-resolution reconstruction.
目前,遥感图像中通常只存在少量甚至不存在高分辨率遥感图像,针对训练数据源少的问题,可以采取以下几种解决方式:(1)采用数据集扩增方式扩增数据;(2)使用自然图像来训练重构网络模型,再迁移到遥感 图像超分辨率重构中;(3)采用无监督学习的方式重构高分辨率遥感图像。采取数据集扩增的方式只能在一定程度上缓解这一问题带来的影响,与ImageNet大数据集相比,扩增后的数据量并不足以训练出理想的重构模型;而且遥感图像与自然图像成像的降质过程有很大的不同之处,采用自然图像训练的网络模型不适宜直接应用到遥感图像超分辨率重构中,且迁移学习仍然需要高分辨率遥感图像标签;因此,要从根本上解决遥感图像高分辨率图像源少的问题,可以采用无监督学习的方法。At present, there are usually only a few or no high-resolution remote sensing images in remote sensing images. To solve the problem of few training data sources, the following solutions can be adopted: (1) Use data set amplification to amplify data; (2) Use natural images to train the reconstruction network model, and then migrate it to the super-resolution reconstruction of remote sensing images; (3) Use unsupervised learning to reconstruct high-resolution remote sensing images. The use of data set augmentation can only alleviate the impact of this problem to a certain extent. Compared with the ImageNet large data set, the amount of augmented data is not enough to train an ideal reconstruction model; and remote sensing images It is very different from the degradation process of natural image imaging. The network model trained with natural image is not suitable for direct application to remote sensing image super-resolution reconstruction, and migration learning still requires high-resolution remote sensing image tags; therefore To fundamentally solve the problem of fewer high-resolution image sources for remote sensing images, unsupervised learning methods can be used.
生成对抗网络(GAN)通过增加判别器,可以与生成器进行对抗训练,从而达到质量好的生成图片的目的。判别器通过训练能够尽量判别出生成图片与标签图片,而生成器通过训练过程能生成尽可能使判别器判别不出来的图片。在二者对抗训练过程中,判别器能够间接作为损失函数提高生成器的学习能力。将生成对抗网络应用于图像超分辨率重构(SRGAN模型)能够使得重构图像具有更多的纹理细节信息,视觉效果优于其他深度学习模型。遥感图像通常由远距离获取,包含复杂的场景和地物信息,因此在遥感图像重构过程中,对于细节的重构要求更高,将GAN引入遥感图像超分辨率重构能够有效提高重构效果。The Generative Adversarial Network (GAN) can conduct adversarial training with the generator by adding a discriminator, so as to achieve the purpose of generating pictures with good quality. The discriminator can discriminate the generated picture and the label picture as much as possible through training, and the generator can generate pictures that the discriminator cannot discriminate as much as possible through the training process. In the training process of the two confrontation, the discriminator can indirectly serve as a loss function to improve the learning ability of the generator. Applying generative confrontation network to image super-resolution reconstruction (SRGAN model) can make the reconstructed image have more texture detail information, and the visual effect is better than other deep learning models. Remote sensing images are usually acquired from a long distance and contain complex scene and feature information. Therefore, in the process of remote sensing image reconstruction, the requirements for detail reconstruction are higher. The introduction of GAN into remote sensing image super-resolution reconstruction can effectively improve the reconstruction. effect.
根据图1所述的超分辨率图像重构方法,构建生成对抗网络,包括:利用卷积神经网络构建生成器,卷积神经网络包括卷积核大小依次递减的多个卷积层,其中每个卷积层的卷积核大小都大于步长。具体地,生成器结构采用了卷积层解卷积层构建的编解码器结构,能够进行特征的提取与恢复,并且引入跳跃链接,将底层特征融入高层输出中,有效的保留了原始图片的低频信息。According to the super-resolution image reconstruction method described in Figure 1, constructing a generative confrontation network includes: constructing a generator using a convolutional neural network. The convolutional neural network includes multiple convolutional layers with successively decreasing convolution kernels, each of which The size of the convolution kernel of each convolutional layer is larger than the step size. Specifically, the generator structure adopts a codec structure constructed by a convolutional layer and a deconvolutional layer, which can extract and restore features, and introduce skip links to integrate the low-level features into the high-level output, effectively retaining the original picture Low frequency information.
根据图1所述的超分辨率图像重构方法,利用卷积神经网络构建生成器,包括:将卷积神经网络的输出层之后增加用于下采样的平均池化层,其中将从平均池化层输出的图像作为判别器的第二输入,其中,利用训练后的生成对抗网络的生成器,对输入图像进行重构,包括:将输入图像输入到生成器,从输出层得到重构图像。具体地,在网络中输出层再加入一层平均池化层(×4倍数为两层平均池化),达到将重构图像进行下采样的目的。平均池化与最大池化相比能够在下采样过程尽可能多的保留邻域像 素信息,而且能够增加模型对于不同过程降质图片的泛化能力。生成器的最终输出为重构图片I SR与下采样的重构图片I SR'According to the super-resolution image reconstruction method described in Figure 1, the use of a convolutional neural network to construct a generator includes: adding an average pooling layer for downsampling after the output layer of the convolutional neural network, where the average pooling layer The output image of the transformation layer is used as the second input of the discriminator. Among them, the input image is reconstructed by the generator of the generated confrontation network after training, including: inputting the input image to the generator, and obtaining the reconstructed image from the output layer . Specifically, another layer of average pooling is added to the output layer of the network (×4 multiples are two layers of average pooling) to achieve the purpose of down-sampling the reconstructed image. Compared with maximum pooling, average pooling can retain as much neighborhood pixel information as possible in the down-sampling process, and it can increase the generalization ability of the model for degraded images in different processes. The final output of the generator is the reconstructed picture I SR and the down-sampled reconstructed picture I SR' .
根据图1所述的超分辨率图像重构方法,判别器的第一输入对应于第一输出,第二输入对应于第二输出,其中构建生成对抗网络,包括:利用生成器损失函数计算器,以便向生成器反馈生成器损失函数,其中生成器损失函数至少包括第一输出与第二输出之间的交叉熵损失。首先,本发明实施例的生成器损失函数L G可以包含以下几个部分,图像损失L image,感知损失L VGG,对抗损失L Adv以及全变分损失L TV。生成器损失函数计算公式为: According to the super-resolution image reconstruction method described in Fig. 1, the first input of the discriminator corresponds to the first output, and the second input corresponds to the second output. The construction of a generative confrontation network includes: using a generator loss function calculator , So as to feed back the generator loss function to the generator, where the generator loss function includes at least the cross-entropy loss between the first output and the second output. First, the generator loss function L G of the embodiment of the present invention may include the following parts: image loss L image , perceptual loss L VGG , confrontation loss L Adv and total variation loss L TV . The calculation formula of the generator loss function is:
L G=L image+2×10 -3×L VGG+L Adv+2×10 -8×L TV       (1) L G =L image +2×10 -3 ×L VGG +L Adv +2×10 -8 ×L TV (1)
其中,图像损失为I SR与I SR'的L1范数: Among them, the image loss is the L1 norm of I SR and I SR':
Figure PCTCN2020077874-appb-000001
Figure PCTCN2020077874-appb-000001
其中,r为特征提取比,W为基于像素的图像宽度,而H为基于像素的图像高度。Among them, r is the feature extraction ratio, W is the pixel-based image width, and H is the pixel-based image height.
VGG(VisualGeometryGroup,视觉几何组)模型损失为I LR与I SR’经过VGG16特征提取网络的最后一层卷积层(包括激活函数)的输出特征之间的L 1范数: The VGG (VisualGeometryGroup, visual geometry group) model loss is the L 1 norm between the output features of I LR and I SR' through the last layer of the convolutional layer (including the activation function) of the VGG16 feature extraction network:
Figure PCTCN2020077874-appb-000002
Figure PCTCN2020077874-appb-000002
其中,W’为基于像素的图像宽度,而H’为基于像素的图像高度。Among them, W'is the width of the image based on pixels, and H'is the height of the image based on pixels.
基于D(.)的判别器输出函数,对抗损失为D(I LR)与真实标签1、D(I SR')与虚假标签0以及D(I LR)与D(I SR')之间的交叉熵损失的加权和,其中w 1和w 2为可调权重: Based on the output function of the discriminator of D(.), the confrontation loss is between D(I LR ) and real label 1, D(I SR' ) and false label 0, and D(I LR ) and D(I SR' ) The weighted sum of cross entropy loss, where w 1 and w 2 are adjustable weights:
Figure PCTCN2020077874-appb-000003
Figure PCTCN2020077874-appb-000003
全变分损失为图像水平方向与垂直方向的差分均方值:The total variation loss is the mean square value of the difference between the horizontal and vertical directions of the image:
Figure PCTCN2020077874-appb-000004
Figure PCTCN2020077874-appb-000004
其中,I x,y表示横坐标第x个像素,纵坐标为第y个像素。 Among them, I x, y represents the xth pixel on the abscissa, and the yth pixel on the ordinate.
除了生成器损失函数之外,本发明实施例的判别器损失为:In addition to the generator loss function, the discriminator loss in the embodiment of the present invention is:
Figure PCTCN2020077874-appb-000005
Figure PCTCN2020077874-appb-000005
根据图1所述的超分辨率图像重构方法,构建生成对抗网络,包括:利用批归一化层构建判别器,批归一化层置于判别器的密集层与激励函数层之间。具体而言,判别器由一系列不同步长的卷积层构成,每个卷积层后接Leaky ReLU(α=0.2)激活函数以及BN(归一化)层,最后在两个密集连结层后加入一层BN再输入给Sigmoid函数(一种激励函数),得到判别器对一张图片的评分。According to the super-resolution image reconstruction method described in Fig. 1, constructing a generative confrontation network includes: constructing a discriminator using a batch normalization layer, and the batch normalization layer is placed between the dense layer of the discriminator and the excitation function layer. Specifically, the discriminator is composed of a series of non-synchronized convolutional layers. Each convolutional layer is followed by a Leaky ReLU (α=0.2) activation function and a BN (normalization) layer, and finally in two densely connected layers Then add a layer of BN and then input it to the Sigmoid function (an excitation function) to get the discriminator's score for a picture.
此外,训练结束之后,并且在利用训练后的生成对抗网络的生成器对输入图像进行重构之前,可以对网络模型进行训练与测试,例如,本发明实施例可以采用Adam优化器优化损失函数,输入数据采用128*128随机裁剪的图像块,能够起到一定的数据增强作用。学习率设置为5*10-4,Batch_size大小为32。利用训练集对所构建的生成对抗网络进行训练,并使用测试集对训练好的重构网络进行测试,完成遥感图像的超分辨率重构。In addition, after the training is completed, and before the input image is reconstructed using the trained generator that generates the adversarial network, the network model can be trained and tested. For example, in the embodiment of the present invention, the Adam optimizer can be used to optimize the loss function. The input data adopts 128*128 randomly cropped image blocks, which can play a certain role in data enhancement. The learning rate is set to 5*10-4, and the Batch_size size is 32. Use the training set to train the constructed generative confrontation network, and use the test set to test the trained reconstruction network to complete the super-resolution reconstruction of remote sensing images.
此外,可以采用评价指标对重构图像进行评价。例如,本发明实施例采用SAM(光谱角制图),RMSE(均方根误差),ERGAS(综合无量纲整体相对误差),sCC(空间相关系数),Qindex(质量指标),SSIM(结构相似度),PSNR(峰值信噪比)作为评价指标对重构图像进行全面评价。例如,表1给出了这些评价指标的计算公式。In addition, an evaluation index can be used to evaluate the reconstructed image. For example, the embodiment of the present invention uses SAM (spectral angle mapping), RMSE (root mean square error), ERGAS (integrated dimensionless overall relative error), sCC (spatial correlation coefficient), Qindex (quality index), SSIM (structure similarity) ), PSNR (Peak Signal-to-Noise Ratio) is used as an evaluation index to comprehensively evaluate the reconstructed image. For example, Table 1 shows the calculation formulas for these evaluation indicators.
表1 评价指标计算公式Table 1 Evaluation index calculation formula
Figure PCTCN2020077874-appb-000006
Figure PCTCN2020077874-appb-000006
Figure PCTCN2020077874-appb-000007
Figure PCTCN2020077874-appb-000007
图2是通过无监督学习实现遥感图像超分辨率重构的流程图。作为一个具体的实施例,通过输入插值后的低分辨率图像从生成器输出后,得到重构图像,然后将下采样至与低分辨率图像LR同尺寸,然后输入判别器使其判别LR和SR’,并求二者之间的损失以更新网络参数。Figure 2 is a flow chart of realizing super-resolution reconstruction of remote sensing images through unsupervised learning. As a specific embodiment, after inputting the interpolated low-resolution image and outputting it from the generator, the reconstructed image is obtained, and then down-sampled to the same size as the low-resolution image LR, and then input to the discriminator to make it discriminate between LR and SR', and find the loss between the two to update the network parameters.
图3是本发明中生成对抗网络的生成器结构。生成器具有卷积层和解卷积层的编解码器结构,并且所述卷积层和所述解卷积层具有跳跃连接。具体而言,该生成器包含一系列卷积层和解卷积层,首先三个卷积核分别为7,5和3的卷积层(接ReLU(线性整流单元))用于特征的提取,其后接连续4层3×3的卷积层用于图像特征编码;然后4层解卷积用于特征解码,3,5,7三层解卷积用于特征恢复,最终得到重构图像。为解决图像连续卷积低频信息丢失的问题,在输入与输出之间加入跳跃连接,保留低频信息。Figure 3 is the generator structure for generating the confrontation network in the present invention. The generator has a codec structure of a convolution layer and a deconvolution layer, and the convolution layer and the deconvolution layer have a skip connection. Specifically, the generator includes a series of convolutional layers and deconvolutional layers. First, the three convolution kernels are 7, 5, and 3 convolutional layers (connected to ReLU (linear rectification unit)) for feature extraction. This is followed by 4 consecutive 3×3 convolutional layers for image feature coding; then 4 layers of deconvolution are used for feature decoding, and 3, 5, and 7 three-layer deconvolutions are used for feature restoration, and finally a reconstructed image is obtained . In order to solve the problem of the loss of low-frequency information in continuous convolution of the image, a jump connection is added between the input and the output to preserve the low-frequency information.
图4是本发明生成对抗网络中的判别器。输入图像后,判别器经过一 系列卷积、Leaky ReLU以及BN层,对输入图片进行特征提取,最后部分进行密集连接并输入sigmoid函数得到该输入图片的得分值,根据得分判断输入图片之间的差异。Fig. 4 is the discriminator in the generated confrontation network of the present invention. After inputting the image, the discriminator goes through a series of convolution, Leaky ReLU and BN layers to extract the features of the input image. The final part is densely connected and the sigmoid function is input to obtain the score value of the input image, and the score between the input images The difference.
图5是本发明实施例的图像重构装置的示意性框图。图5的超分辨率图像重构装置500包括:Fig. 5 is a schematic block diagram of an image reconstruction device according to an embodiment of the present invention. The super-resolution image reconstruction device 500 of FIG. 5 includes:
第一构建模块510,构建生成对抗网络,其中生成对抗网络至少包括生成器、判别器和损失函数计算器;The first building module 510 is to build a generative confrontation network, where the generative confrontation network at least includes a generator, a discriminator, and a loss function calculator;
第二构建模块520,构建原始分辨率图像训练集;The second construction module 520 constructs an original resolution image training set;
训练模块530,基于原始分辨率图像训练集训练生成对抗网络,其中,原始分辨率图像训练集经由预处理作为判别器的第一输入,生成器的输出作为判别器的第二输入;The training module 530 trains and generates a confrontation network based on the original resolution image training set, where the original resolution image training set is preprocessed as the first input of the discriminator, and the output of the generator is used as the second input of the discriminator;
重构模块540,利用训练后的生成对抗网络的生成器,对输入图像进行重构。The reconstruction module 540 uses the trained generator to generate the confrontation network to reconstruct the input image.
在本发明实施例的超分辨率图像重构装置中,所构建的生成对抗网络包括生成器、判别器以及损失函数计算器,并且该生成对抗网络采用原始分辨率图像训练集进行训练,因此能够利用原始分辨率图像作为参考图像来重构超分辨率图像,从而即使原始分辨率为较低分辨率,仍然可以重构出较高分辨率图像。In the super-resolution image reconstruction device of the embodiment of the present invention, the constructed generation confrontation network includes a generator, a discriminator, and a loss function calculator, and the generation confrontation network uses the original resolution image training set for training, so it can The original resolution image is used as the reference image to reconstruct the super-resolution image, so that even if the original resolution is a lower resolution, a higher resolution image can still be reconstructed.
根据图5的超分辨率图像重构装置,第一构建模块具体用于:利用卷积神经网络构建生成器,卷积神经网络包括卷积核大小依次递减的多个卷积层,其中每个卷积层的卷积核大小都大于步长。According to the super-resolution image reconstruction device of Fig. 5, the first building module is specifically used to: construct a generator using a convolutional neural network, the convolutional neural network including a plurality of convolutional layers with successively decreasing convolution kernels, each of which The size of the convolution kernel of the convolution layer is larger than the step size.
根据图5的超分辨率图像重构装置,第一构建模块具体用于:将卷积神经网络的输出层之后增加用于下采样的平均池化层,其中将从平均池化层输出的图像作为判别器的第二输入,其中重构模块具体用于:将输入图像输入到生成器,从输出层得到重构图像。According to the super-resolution image reconstruction device of FIG. 5, the first building module is specifically used to add an average pooling layer for downsampling after the output layer of the convolutional neural network, wherein the image output from the average pooling layer As the second input of the discriminator, the reconstruction module is specifically used to input the input image to the generator, and obtain the reconstructed image from the output layer.
根据图5的超分辨率图像重构装置,判别器的第一输入对应于第一输出,第二输入对应于第二输出,其中第一构建模块具体用于:利用生成器损失函数计算器,以便向生成器反馈生成器损失函数,其中生成器损失函数至少包括第一输出与第二输出之间的交叉熵损失。According to the super-resolution image reconstruction device of FIG. 5, the first input of the discriminator corresponds to the first output, and the second input corresponds to the second output. The first building module is specifically used for: using a generator loss function calculator, In order to feed back the generator loss function to the generator, the generator loss function includes at least the cross-entropy loss between the first output and the second output.
根据图5的超分辨率图像重构装置,第一构建模块具体用于:利用批归一化层构建判别器,批归一化层置于判别器的密集层与激励函数层之间。According to the super-resolution image reconstruction device of FIG. 5, the first construction module is specifically used to construct a discriminator using a batch normalization layer, and the batch normalization layer is placed between the dense layer of the discriminator and the excitation function layer.
图6是本发明实施例中对测试集进行×2倍数测试得到的各项评价指标平均值。图中包含了12种无监督学习进行图像超分辨率重构的方法,其中SAM值越接近0越好,RMSE越小越好,ERGAS越小越好,sCC越接近1越好,Q越大越好,SSIM越接近1越好,PSNR越大越好。由图和表中数值可知,在以上各项指标中,本发明重构的图像均能获取最好的评价结果(ERGAS结果排第二),从而证明了该网络进行无监督学习重构图像的有效性。Fig. 6 is an average value of various evaluation indexes obtained by performing a ×2 multiple test on a test set in an embodiment of the present invention. The figure contains 12 unsupervised learning methods for image super-resolution reconstruction. The closer the SAM value is to 0, the better, the smaller the RMSE, the better, the smaller the ERGAS, the better, the closer the sCC is to 1, the larger the Q Well, the closer the SSIM is to 1, the better, and the larger the PSNR, the better. From the figures and the values in the table, it can be seen that, among the above indicators, the reconstructed image of the present invention can obtain the best evaluation result (ERGAS result ranks second), which proves that the network performs unsupervised learning to reconstruct the image. Effectiveness.
图7对测试集进行×4倍数测试得到的各项评价指标平均值。由图和表中数值可知,在各评价指标中,本方法均取得了最好的评价结果,证明了本发明在较大放大倍数下无监督学习重构结果的有效性。Fig. 7 is the average value of each evaluation index obtained by the ×4 multiple test on the test set. It can be seen from the figures and the numerical values in the table that, among the various evaluation indicators, the method achieves the best evaluation results, which proves the effectiveness of the unsupervised learning reconstruction results of the present invention at a larger magnification.
图8是本发明实施例中对于测试集×2倍数部分结果的视觉显示,图(a)为原始高分辨率图像,图(b)为输入网络的双三次插值后的图像,图(c)为本方法重构的图像,由图可以看出本方法重构的细节多,接近高分辨率图像。Figure 8 is a visual display of the results of the test set × 2 multiples in the embodiment of the present invention. Figure (a) is the original high-resolution image, Figure (b) is the image after bicubic interpolation of the input network, and Figure (c) The image reconstructed by this method can be seen from the figure that there are many details reconstructed by this method, which is close to a high-resolution image.
图9是本发明实施例中对于测试集×2倍数部分结果的视觉显示,图(a)为原始高分辨率图像,图(b)为输入网络的双三次插值后的图像,图(c)为本方法重构的图像,由图可以看出本方法的重构的图片能取得良好的视觉效果,更加接近高分辨率图像。Figure 9 is a visual display of the results of the test set × 2 multiples in the embodiment of the present invention. Figure (a) is the original high-resolution image, Figure (b) is the image after bicubic interpolation of the input network, and Figure (c) For the image reconstructed by this method, it can be seen from the figure that the reconstructed image by this method can achieve good visual effects, which is closer to high-resolution images.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系统中单元、模块的具体工作过程, 可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and conciseness of description, only the division of the above functional units and modules is used as an example. In practical applications, the above functions can be allocated to different functional units and modules as needed. Module completion, that is, the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above. The functional units and modules in the embodiments can be integrated into one processing unit, or each unit can exist alone physically, or two or more units can be integrated into one unit. The above-mentioned integrated units can be hardware-based Formal realization can also be realized in the form of software functional units. In addition, the specific names of the functional units and modules are only used to facilitate distinguishing from each other, and are not used to limit the protection scope of the present application. For the specific working process of the units and modules in the foregoing system, reference may be made to the corresponding process in the foregoing method embodiment, and details are not described herein again.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above-mentioned embodiments, the description of each embodiment has its own emphasis. For parts that are not described in detail or recorded in an embodiment, reference may be made to related descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。A person of ordinary skill in the art may realize that the units and algorithm steps of the examples described in combination with the embodiments disclosed herein can be implemented by electronic hardware or a combination of computer software and electronic hardware. Whether these functions are performed by hardware or software depends on the specific application and design constraint conditions of the technical solution. Professionals and technicians can use different methods for each specific application to implement the described functions, but such implementation should not be considered as going beyond the scope of the present invention.
在本发明所提供的实施例中,应该理解到,所揭露的装置/终端设备和方法,可以通过其它的方式实现。例如,以上所描述的装置/终端设备实施例仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided by the present invention, it should be understood that the disclosed device/terminal device and method may be implemented in other ways. For example, the device/terminal device embodiments described above are merely illustrative. For example, the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units. Or components can be combined or integrated into another system, or some features can be omitted or not implemented. In addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, the functional units in the various embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
所述集成的模块/单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明实现上述实施例方法中的全部或部分流程,也可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法 实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、电载波信号、电信信号以及软件分发介质等。需要说明的是,所述计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括电载波信号和电信信号。If the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium. Based on this understanding, the present invention implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program. The computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, it can implement the steps of the foregoing method embodiments. Wherein, the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signal, telecommunications signal, and software distribution media, etc. It should be noted that the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of the legislation and patent practice in the jurisdiction. For example, in some jurisdictions, according to the legislation and patent practice, the computer-readable medium Does not include electrical carrier signals and telecommunication signals.
以上所述实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围,均应包含在本发明的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present invention, but not to limit them; although the present invention has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that it can still implement the foregoing various embodiments. The technical solutions recorded in the examples are modified, or some of the technical features are equivalently replaced; these modifications or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the present invention, and should be included in Within the protection scope of the present invention.

Claims (10)

  1. 一种超分辨率图像重构方法,其特征在于,包括:A super-resolution image reconstruction method, characterized in that it comprises:
    构建生成对抗网络,其中所述生成对抗网络至少包括生成器、判别器和损失函数计算器;Constructing a generative confrontation network, wherein the generative confrontation network at least includes a generator, a discriminator, and a loss function calculator;
    构建原始分辨率图像训练集;Construct the original resolution image training set;
    基于所述原始分辨率图像训练集训练所述生成对抗网络,其中,所述原始分辨率图像训练集经由预处理作为所述判别器的第一输入,所述生成器的输出作为所述判别器的第二输入;The generative confrontation network is trained based on the original resolution image training set, wherein the original resolution image training set is preprocessed as the first input of the discriminator, and the output of the generator is used as the discriminator The second input;
    利用训练后的所述生成对抗网络的生成器,对输入图像进行重构。The input image is reconstructed using the trained generator of the generated confrontation network.
  2. 根据权利要求1所述的图像重构方法,其特征在于,所述构建生成对抗网络,包括:The image reconstruction method according to claim 1, wherein said constructing a generative confrontation network comprises:
    利用卷积神经网络构建所述生成器,所述卷积神经网络包括卷积核大小依次递减的多个卷积层,其中每个卷积层的卷积核大小都大于步长。The generator is constructed using a convolutional neural network. The convolutional neural network includes a plurality of convolutional layers with successively decreasing convolution kernel sizes, wherein the convolution kernel size of each convolutional layer is greater than the step size.
  3. 根据权利要求2所述的图像重构方法,其特征在于,利用卷积神经网络构建所述生成器,包括:The image reconstruction method according to claim 2, wherein the construction of the generator using a convolutional neural network comprises:
    将卷积神经网络的输出层之后增加用于下采样的平均池化层,其中将从所述平均池化层输出的图像作为所述判别器的所述第二输入,Add an average pooling layer for downsampling after the output layer of the convolutional neural network, where the image output from the average pooling layer is used as the second input of the discriminator,
    其中,所述利用训练后的所述生成对抗网络的生成器,对输入图像进行重构,包括:Wherein, the reconstructing the input image by using the trained generator for generating the confrontation network includes:
    将所述输入图像输入到所述生成器,从输出层得到重构图像。The input image is input to the generator, and a reconstructed image is obtained from the output layer.
  4. 根据权利要求1所述的图像重构方法,其特征在于,所述判别器的所述第一输入对应于第一输出,所述第二输入对应于第二输出,其中所述构建生成对抗网络,包括:The image reconstruction method according to claim 1, wherein the first input of the discriminator corresponds to a first output, and the second input corresponds to a second output, wherein the constructing a confrontation network ,include:
    利用生成器损失函数计算器,以便向所述生成器反馈所述生成器损失函数,其中所述生成器损失函数至少包括所述第一输出与所述第二输出之间的交叉熵损失。A generator loss function calculator is used to feed back the generator loss function to the generator, wherein the generator loss function includes at least a cross-entropy loss between the first output and the second output.
  5. 根据权利要求1所述的图像重构方法,其特征在于,所述构建生成对抗网络,包括:The image reconstruction method according to claim 1, wherein said constructing a generative confrontation network comprises:
    利用批归一化层构建所述判别器,所述批归一化层置于所述判别器的 密集层与激励函数层之间。The discriminator is constructed using a batch normalization layer, and the batch normalization layer is placed between the dense layer and the excitation function layer of the discriminator.
  6. 一种超分辨率图像重构装置,其特征在于,包括:A super-resolution image reconstruction device, characterized in that it comprises:
    第一构建模块,构建生成对抗网络,其中所述生成对抗网络至少包括生成器、判别器和损失函数计算器;The first building module is to build a generative confrontation network, wherein the generative confrontation network at least includes a generator, a discriminator, and a loss function calculator;
    第二构建模块,构建原始分辨率图像训练集;The second building module is to build the original resolution image training set;
    训练模块,基于所述原始分辨率图像训练集训练所述生成对抗网络,其中,所述原始分辨率图像训练集经由预处理作为所述判别器的第一输入,所述生成器的输出作为所述判别器的第二输入;The training module trains the generative confrontation network based on the original resolution image training set, wherein the original resolution image training set is preprocessed as the first input of the discriminator, and the generator output is used as the first input of the discriminator. The second input of the discriminator;
    重构模块,利用训练后的所述生成对抗网络的生成器,对输入图像进行重构。The reconstruction module uses the trained generator for generating the confrontation network to reconstruct the input image.
  7. 根据权利要求6所述的图像重构装置,其特征在于,所述第一构建模块具体用于:The image reconstruction device according to claim 6, wherein the first building module is specifically configured to:
    利用卷积神经网络构建所述生成器,所述卷积神经网络包括卷积核大小依次递减的多个卷积层,其中每个卷积层的卷积核大小都大于步长。The generator is constructed using a convolutional neural network. The convolutional neural network includes a plurality of convolutional layers with successively decreasing convolution kernel sizes, wherein the convolution kernel size of each convolutional layer is greater than the step size.
  8. 根据权利要求7所述的图像重构装置,其特征在于,所述第一构建模块具体用于:The image reconstruction device according to claim 7, wherein the first construction module is specifically configured to:
    将卷积神经网络的输出层之后增加用于下采样的平均池化层,其中将从所述平均池化层输出的图像作为所述判别器的所述第二输入,Add an average pooling layer for downsampling after the output layer of the convolutional neural network, where the image output from the average pooling layer is used as the second input of the discriminator,
    其中,所述重构模块具体用于:Wherein, the reconstruction module is specifically used for:
    将所述输入图像输入到所述生成器,从输出层得到重构图像。The input image is input to the generator, and a reconstructed image is obtained from the output layer.
  9. 根据权利要求6所述的图像重构装置,其特征在于,所述判别器的所述第一输入对应于第一输出,所述第二输入对应于第二输出,其中所述第一构建模块具体用于:The image reconstruction device according to claim 6, wherein the first input of the discriminator corresponds to a first output, and the second input corresponds to a second output, wherein the first building module Specifically used for:
    利用生成器损失函数计算器,以便向所述生成器反馈所述生成器损失函数,其中所述生成器损失函数至少包括所述第一输出与所述第二输出之间的交叉熵损失。A generator loss function calculator is used to feed back the generator loss function to the generator, wherein the generator loss function includes at least a cross-entropy loss between the first output and the second output.
  10. 根据权利要求6所述的图像重构装置,其特征在于,所述第一构建模块具体用于:The image reconstruction device according to claim 6, wherein the first building module is specifically configured to:
    利用批归一化层构建所述判别器,所述批归一化层置于所述判别器的 密集层与激励函数层之间。The discriminator is constructed using a batch normalization layer, and the batch normalization layer is placed between the dense layer and the excitation function layer of the discriminator.
PCT/CN2020/077874 2019-09-29 2020-03-05 Super-resolution image reconstruction method and device WO2021056969A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910935280.8A CN110717857A (en) 2019-09-29 2019-09-29 Super-resolution image reconstruction method and device
CN201910935280.8 2019-09-29

Publications (1)

Publication Number Publication Date
WO2021056969A1 true WO2021056969A1 (en) 2021-04-01

Family

ID=69211158

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/077874 WO2021056969A1 (en) 2019-09-29 2020-03-05 Super-resolution image reconstruction method and device

Country Status (2)

Country Link
CN (1) CN110717857A (en)
WO (1) WO2021056969A1 (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113298719A (en) * 2021-07-02 2021-08-24 西安电子科技大学 Super-resolution reconstruction method of low-resolution fuzzy face image based on feature separation learning
CN113538238A (en) * 2021-07-09 2021-10-22 深圳市深光粟科技有限公司 High-resolution photoacoustic image imaging method and device and electronic equipment
CN113592715A (en) * 2021-08-05 2021-11-02 昆明理工大学 Super-resolution image reconstruction method for small sample image set
CN113643182A (en) * 2021-08-20 2021-11-12 中国地质大学(武汉) Remote sensing image super-resolution reconstruction method based on dual learning graph network
CN113689360A (en) * 2021-09-30 2021-11-23 合肥工业大学 Image restoration method based on generation countermeasure network
CN113888406A (en) * 2021-08-24 2022-01-04 厦门仟易网络科技有限公司 Camera super-resolution method through deep learning
CN114596285A (en) * 2022-03-09 2022-06-07 南京邮电大学 Multitask medical image enhancement method based on generation countermeasure network
CN114693547A (en) * 2022-03-03 2022-07-01 大连海事大学 Radio frequency image enhancement method and radio frequency image identification method based on image super-resolution
CN114913086A (en) * 2022-05-05 2022-08-16 上海云思智慧信息技术有限公司 Face image quality enhancement method based on generation countermeasure network
CN115545110A (en) * 2022-10-12 2022-12-30 中国电信股份有限公司 High resolution data reconstruction method for generating countermeasure network and related method and device
CN115841423A (en) * 2022-12-12 2023-03-24 之江实验室 Wide-field illumination fluorescence super-resolution microscopic imaging method based on deep learning
CN115936983A (en) * 2022-11-01 2023-04-07 青岛哈尔滨工程大学创新发展中心 Method and device for super-resolution of nuclear magnetic image based on style migration and computer storage medium
CN116681604A (en) * 2023-04-24 2023-09-01 吉首大学 Qin simple text restoration method based on condition generation countermeasure network
CN116863016A (en) * 2023-05-31 2023-10-10 北京长木谷医疗科技股份有限公司 Medical image reconstruction method and device for generating countermeasure network based on deep learning
CN117131348A (en) * 2023-10-27 2023-11-28 深圳中科保泰科技有限公司 Data quality analysis method and system based on differential convolution characteristics
CN113888406B (en) * 2021-08-24 2024-04-23 厦门仟易网络科技有限公司 Camera super-resolution method through deep learning

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110717857A (en) * 2019-09-29 2020-01-21 中国科学院长春光学精密机械与物理研究所 Super-resolution image reconstruction method and device
CN112508929A (en) * 2020-12-16 2021-03-16 奥比中光科技集团股份有限公司 Method and device for training generation of confrontation network
CN115760563A (en) * 2021-09-02 2023-03-07 深圳市中兴微电子技术有限公司 Image super-resolution model training method and device and computer-readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170365038A1 (en) * 2016-06-16 2017-12-21 Facebook, Inc. Producing Higher-Quality Samples Of Natural Images
CN108171762A (en) * 2017-12-27 2018-06-15 河海大学常州校区 System and method for is reconfigured quickly in a kind of similar image of the compressed sensing of deep learning
CN109785258A (en) * 2019-01-10 2019-05-21 华南理工大学 A kind of facial image restorative procedure generating confrontation network based on more arbiters
CN110189253A (en) * 2019-04-16 2019-08-30 浙江工业大学 A kind of image super-resolution rebuilding method generating confrontation network based on improvement
CN110717857A (en) * 2019-09-29 2020-01-21 中国科学院长春光学精密机械与物理研究所 Super-resolution image reconstruction method and device

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10621695B2 (en) * 2017-10-31 2020-04-14 Disney Enterprises, Inc. Video super-resolution using an artificial neural network
CN109903223B (en) * 2019-01-14 2023-08-25 北京工商大学 Image super-resolution method based on dense connection network and generation type countermeasure network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170365038A1 (en) * 2016-06-16 2017-12-21 Facebook, Inc. Producing Higher-Quality Samples Of Natural Images
CN108171762A (en) * 2017-12-27 2018-06-15 河海大学常州校区 System and method for is reconfigured quickly in a kind of similar image of the compressed sensing of deep learning
CN109785258A (en) * 2019-01-10 2019-05-21 华南理工大学 A kind of facial image restorative procedure generating confrontation network based on more arbiters
CN110189253A (en) * 2019-04-16 2019-08-30 浙江工业大学 A kind of image super-resolution rebuilding method generating confrontation network based on improvement
CN110717857A (en) * 2019-09-29 2020-01-21 中国科学院长春光学精密机械与物理研究所 Super-resolution image reconstruction method and device

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113298719A (en) * 2021-07-02 2021-08-24 西安电子科技大学 Super-resolution reconstruction method of low-resolution fuzzy face image based on feature separation learning
CN113298719B (en) * 2021-07-02 2024-04-05 西安电子科技大学 Feature separation learning-based super-resolution reconstruction method for low-resolution fuzzy face image
CN113538238A (en) * 2021-07-09 2021-10-22 深圳市深光粟科技有限公司 High-resolution photoacoustic image imaging method and device and electronic equipment
CN113592715A (en) * 2021-08-05 2021-11-02 昆明理工大学 Super-resolution image reconstruction method for small sample image set
CN113643182A (en) * 2021-08-20 2021-11-12 中国地质大学(武汉) Remote sensing image super-resolution reconstruction method based on dual learning graph network
CN113643182B (en) * 2021-08-20 2024-03-19 中国地质大学(武汉) Remote sensing image super-resolution reconstruction method based on dual learning graph network
CN113888406B (en) * 2021-08-24 2024-04-23 厦门仟易网络科技有限公司 Camera super-resolution method through deep learning
CN113888406A (en) * 2021-08-24 2022-01-04 厦门仟易网络科技有限公司 Camera super-resolution method through deep learning
CN113689360A (en) * 2021-09-30 2021-11-23 合肥工业大学 Image restoration method based on generation countermeasure network
CN113689360B (en) * 2021-09-30 2024-02-20 合肥工业大学 Image restoration method based on generation countermeasure network
CN114693547A (en) * 2022-03-03 2022-07-01 大连海事大学 Radio frequency image enhancement method and radio frequency image identification method based on image super-resolution
CN114596285A (en) * 2022-03-09 2022-06-07 南京邮电大学 Multitask medical image enhancement method based on generation countermeasure network
CN114913086A (en) * 2022-05-05 2022-08-16 上海云思智慧信息技术有限公司 Face image quality enhancement method based on generation countermeasure network
CN114913086B (en) * 2022-05-05 2023-05-02 上海云思智慧信息技术有限公司 Face image quality enhancement method based on generation countermeasure network
CN115545110A (en) * 2022-10-12 2022-12-30 中国电信股份有限公司 High resolution data reconstruction method for generating countermeasure network and related method and device
CN115545110B (en) * 2022-10-12 2024-02-02 中国电信股份有限公司 High resolution data reconstruction method for generating an antagonism network and related method and apparatus
CN115936983A (en) * 2022-11-01 2023-04-07 青岛哈尔滨工程大学创新发展中心 Method and device for super-resolution of nuclear magnetic image based on style migration and computer storage medium
CN115841423A (en) * 2022-12-12 2023-03-24 之江实验室 Wide-field illumination fluorescence super-resolution microscopic imaging method based on deep learning
CN116681604B (en) * 2023-04-24 2024-01-02 吉首大学 Qin simple text restoration method based on condition generation countermeasure network
CN116681604A (en) * 2023-04-24 2023-09-01 吉首大学 Qin simple text restoration method based on condition generation countermeasure network
CN116863016A (en) * 2023-05-31 2023-10-10 北京长木谷医疗科技股份有限公司 Medical image reconstruction method and device for generating countermeasure network based on deep learning
CN117131348B (en) * 2023-10-27 2024-02-09 深圳中科保泰科技有限公司 Data quality analysis method and system based on differential convolution characteristics
CN117131348A (en) * 2023-10-27 2023-11-28 深圳中科保泰科技有限公司 Data quality analysis method and system based on differential convolution characteristics

Also Published As

Publication number Publication date
CN110717857A (en) 2020-01-21

Similar Documents

Publication Publication Date Title
WO2021056969A1 (en) Super-resolution image reconstruction method and device
CN108830796B (en) Hyperspectral image super-resolution reconstruction method based on spectral-spatial combination and gradient domain loss
CN109886871B (en) Image super-resolution method based on channel attention mechanism and multi-layer feature fusion
CN112001847A (en) Method for generating high-quality image by relatively generating antagonistic super-resolution reconstruction model
Du et al. Fully convolutional measurement network for compressive sensing image reconstruction
CN111311490A (en) Video super-resolution reconstruction method based on multi-frame fusion optical flow
CN108734660A (en) A kind of image super-resolution rebuilding method and device based on deep learning
CN111598778B (en) Super-resolution reconstruction method for insulator image
CN110675336A (en) Low-illumination image enhancement method and device
CN109146788A (en) Super-resolution image reconstruction method and device based on deep learning
CN111242846B (en) Fine-grained scale image super-resolution method based on non-local enhancement network
CN110533607B (en) Image processing method and device based on deep learning and electronic equipment
CN112801904B (en) Hybrid degraded image enhancement method based on convolutional neural network
CN111784570A (en) Video image super-resolution reconstruction method and device
CN110288524B (en) Deep learning super-resolution method based on enhanced upsampling and discrimination fusion mechanism
CN111738954B (en) Single-frame turbulence degradation image distortion removal method based on double-layer cavity U-Net model
CN114677304A (en) Image deblurring algorithm based on knowledge distillation and deep neural network
Guan et al. Srdgan: learning the noise prior for super resolution with dual generative adversarial networks
CN108288256A (en) A kind of multispectral mosaic image restored method
CN112949636A (en) License plate super-resolution identification method and system and computer readable medium
Li et al. Hdrnet: Single-image-based hdr reconstruction using channel attention cnn
CN115393191A (en) Method, device and equipment for reconstructing super-resolution of lightweight remote sensing image
CN111640067B (en) Single image super-resolution reconstruction method based on three-channel convolutional neural network
CN115115514A (en) Image super-resolution reconstruction method based on high-frequency information feature fusion
CN110047038B (en) Single-image super-resolution reconstruction method based on hierarchical progressive network

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20867060

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20867060

Country of ref document: EP

Kind code of ref document: A1