US20210133925A1 - Device and Method of Handling Image Super-Resolution - Google Patents

Device and Method of Handling Image Super-Resolution Download PDF

Info

Publication number
US20210133925A1
US20210133925A1 US16/796,872 US202016796872A US2021133925A1 US 20210133925 A1 US20210133925 A1 US 20210133925A1 US 202016796872 A US202016796872 A US 202016796872A US 2021133925 A1 US2021133925 A1 US 2021133925A1
Authority
US
United States
Prior art keywords
feature maps
image
module
parameter
computing device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/796,872
Other versions
US11010871B1 (en
Inventor
Wei-Yu Lee
Po-Yu CHUANG
Yu-Chiang Wang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Moxa Inc
Original Assignee
Moxa Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Moxa Inc filed Critical Moxa Inc
Priority to US16/796,872 priority Critical patent/US11010871B1/en
Assigned to MOXA INC. reassignment MOXA INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHUANG, PO-YU, LEE, WEI-YU, WANG, YU-CHIANG
Priority to CN202010226039.0A priority patent/CN112785496A/en
Priority to TW109115806A priority patent/TWI765264B/en
Publication of US20210133925A1 publication Critical patent/US20210133925A1/en
Application granted granted Critical
Publication of US11010871B1 publication Critical patent/US11010871B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • G06T3/4076Super resolution, i.e. output image resolution higher than sensor resolution by iteratively correcting the provisional high resolution image using the original low-resolution image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4046Scaling the whole image or part thereof using neural networks

Definitions

  • the present invention relates to a device and a method used in a multimedia system, and more particularly, to a device and a method of handling image super-resolution.
  • Image super-resolution aims to reconstruct a high-resolution (HR) image from a low resolution (LR) image.
  • HR high-resolution
  • LR low resolution
  • Various methods have been proposed for minimizing differences between the HR image and an output image, when handling the ISR.
  • some of the methods have a high peak signal-to-noise ratio (PSNR) performance while having a low perceptual quality (e.g., the output image is overly smooth).
  • PSNR peak signal-to-noise ratio
  • the others of the methods have a high perceptual quality while having a low PSNR performance (e.g., the output image has structure distortion).
  • PSNR peak signal-to-noise ratio
  • the output image has structure distortion
  • the present invention therefore provides a device and a method for handling image super-resolution to solve the abovementioned problem.
  • a computing device for handling image super-resolution comprises a generator module, for receiving at least one input image, to perform the following operations: generating a first plurality of feature maps of the at least one input image by at least one first channel attention (CA); and generating an output image according to the first plurality of feature maps and at least one first parameter of the generator module; a discriminator module, coupled to the generator module, for receiving the output image and a high resolution (HR) image, to perform the following operations: generating a second plurality of feature maps of the output image and a third plurality of feature maps of the HR image by at least one second CA; and generating at least one score according to the second plurality of feature maps, the third plurality of feature maps and at least one second parameter of the discriminator module; and a feedback module, couple to the generator module and the discriminator module, for receiving the at least one score, to update the at least one first parameter and the at least one second parameter according to the at least one score and an objective function.
  • CA channel attention
  • HR high resolution
  • FIG. 1 is a schematic diagram of a computing device according to an example of the present invention.
  • FIG. 2 is a schematic diagram of a generator according to an example of the present invention.
  • FIG. 3 is a schematic diagram of a discriminator according to an example of the present invention.
  • FIG. 4 is a schematic diagram of a CA according to an example of the present invention.
  • FIG. 5 is a comparison table of experiment results according to an example of the present invention.
  • FIG. 6 is a flowchart of a process according to an example of the present invention.
  • FIG. 1 is a schematic diagram of a computing device 10 according to an example of the present invention.
  • the computing device 10 includes a generator module 100 , a discriminator module 110 and a feedback module 120 .
  • the generator module 100 receives at least one input image, and generates a first plurality of feature maps of the at least one input image by at least one first channel attention (CA).
  • CA channel attention
  • the generator module 100 generates an output image according to at least one first parameter of the generator module 100 and the first plurality of feature maps.
  • the discriminator module 110 is coupled to the generator module 100 , and receives the output image and a high resolution (HR) image.
  • HR high resolution
  • the discriminator module 110 generates a second plurality of feature maps of the output image and a third plurality of feature maps of the HR image by at least one second CA, and generates at least one score (e.g., at least one probability or at least one distance) according to at least one second parameter of the discriminator module 110 , the second plurality of feature maps and the third plurality of feature maps.
  • the feedback module 120 is coupled to the generator module 100 and the discriminator module 110 , and receives the at least one score.
  • the feedback module 120 updates the at least one first parameter and the at least one second parameter according to the at least one score and an objective function.
  • the feedback module 120 generates a first value to update the at least one first parameter of the generator module 100 according to the first value.
  • the feedback module 120 generates a second value to update the at least one second parameter of the discriminator module 110 according to the second value.
  • the feedback module 120 may generate the first value when updating the at least one first parameter of the generator module 100 is requested (e.g., configurated), and the at least one second parameter of the discriminator module 110 may be fixed. In one example, the feedback module 120 may generate the second value when updating the at least one second parameter of the discriminator module 110 is requested (e.g., configurated), and the at least one first parameter of the generator module 100 may be fixed. That is, the feedback module 120 may only update parameter (s) of a module (i.e., the generator module 100 or the discriminator module 110 ) in an iteration, and parameter(s) of the other module (i.e., the discriminator module 110 or the generator module 100 ) is fixed.
  • the feedback module 120 uses an Adaptive Moment Estimation (ADAM) optimizer to update the at least one first parameter according to the first value in an iteration.
  • ADAM Adaptive Moment Estimation
  • the generator module 100 may generate a new output image according to at least one new first parameter of the generator module 100 that is updated, wherein the new output image has a higher quality than that of a present output image (i.e., the new output image with higher PSNR performance).
  • the discriminator module 110 generates at least one new score after receiving the new output image and the HR image, wherein new difference between the at least one new score may be smaller than present difference between at least one present score.
  • the feedback module 120 uses an ADAM optimizer to update the at least one second parameter according to the second value in an iteration. After the iteration for updating the at least one second parameter, the discriminator module 110 generates at least one new score according to at least one new second parameter of the discriminator module 110 that is updated when receiving the same output image and HR image. New difference between the at least one new score may be larger than present difference between at least one present score.
  • the at least one first parameter of the generator module 100 is updated after the at least one second parameter of the discriminator module 110 is updated N times, wherein the N is a positive integer.
  • the at least one second parameter of the discriminator module 110 is updated after the at least one first parameter of the generator module 100 is updated M times, wherein the M is a positive integer.
  • the discriminator module 110 generates a first score according to the second plurality of feature maps of the output image and the at least one second parameter, after receiving the output image.
  • the discriminator module 110 generates a second score according to the third plurality of feature maps of the HR image and the at least one second parameter, after receiving the HR image.
  • the feedback module 120 receives the first score and the second score, to update the at least one first parameter and the at least one second parameter according to an objective function and these scores.
  • the discriminator module 110 generates a score for distinguishing (e.g., discriminating) how real (or fake) an image is (i.e., for distinguishing whether the image is more likely to be a HR image or an output image generated by the generator module 100 ), after receiving the image. Larger difference between at least one score (e.g., the first score and the second score) represents that the discriminator module 110 distinguishes a HR image and an output image more successfully.
  • a target of the discriminator module 110 is to increase difference between at least one score when receiving the same output image and HR image.
  • a target of the generator module 100 is to generate an output image with a quality the same as that of the HR image (i.e., to reduce difference between at least one score after generating a new output image with higher PSNR performance than before).
  • the generator module 100 independently operates without the discriminator module 110 and the feedback module 120 , e.g., when the generator module 100 is not able to generate an output image with a higher PSNR performance than before, and/or when an output image and the HR image cannot be distinguished according to at least one score generated by the discriminator module 110 (e.g., difference of the at least one score is close to zero).
  • the generator module 100 independently operates without the discriminator module 110 and the feedback module 120 , e.g., when a termination condition (e.g., a number of iterations is equal to a default value) is satisfied. In this situation, the generator module 100 can start to generate an output image with a high quality close to that of a HR image.
  • the at least one first CA is the same or different.
  • the at least one second CA is the same or different.
  • the at least one first CA is the same as or different from the at least one second CA.
  • the at least one first CA comprises at least one of a global pooling layer, a convolution layer and an activation function layer.
  • the at least one second CA comprises at least one of a global pooling layer, a convolution layer and an activation function layer.
  • the activation function layer for the both examples may be a leaky Rectified Linear Unit (ReLU) function layer or a sigmoid function layer.
  • the generator module 100 generates the first plurality of feature maps by the at least one first CA to focus on (e.g., extract) at least one first feature of the at least one input image in a single scale.
  • the discriminator module 110 generates the second plurality of feature maps and the third plurality of feature maps by the at least one second CA to preserve at least one second feature of the output image and at least one third feature of the HR image in at least one scale (e.g., different scale).
  • ISR image super-resolution
  • the two ways may be pixel-wise differences minimization and perceptual quality optimization.
  • Peak signal-to-noise ratio (PSNR)-based methods based on the pixel-wise differences minimization for optimizing the output images are designed to minimize pixel-wise differences between HR images and the output images.
  • PSNR-based methods achieve outstanding performance on PSNR and Structure Similarity (SSIM)
  • SSIM Structure Similarity
  • the output images generated by the PSNR-based methods may not satisfy evaluation of Human Visual System (HVS).
  • HVS Human Visual System
  • GAN generative adversarial networks
  • GAN generative adversarial networks
  • the output images generated by the GAN-based methods may have structure distortion and artifacts.
  • the present invention provides a refiner for the ISR methods.
  • the refiner aims to overcome an overly smoothed problem of the PSNR-based methods and a structure distortion problem of the GAN-based methods.
  • the problems of the PSNR-based methods and the GAN-based methods are solved.
  • the at least one input image is generated according to a processing method for the ISR.
  • the processing method may be a residual channel attention networks (RCAN), a deep back-projection networks (DBPN) or an enhanced deep super-resolution network (EDSR).
  • the output image is generated according to a structure of the at least one input image via an identity mapping.
  • the at least one input image is generated according to a PSNR-based method.
  • the generator module 100 may only need to focus on missing details (e.g., features) of the at least one input image to generate an output image.
  • FIG. 2 is a schematic diagram of a generator 20 according to an example of the present invention.
  • the generator 20 may be utilized for realizing the generator module 100 in FIG. 1 .
  • the generator 20 includes a partial generator network 200 , an identity mapping 210 (e.g., an identity mapping shortcut) and a last convolution layer 220 .
  • the identity mapping 210 is established between at least one input image and the last convolution layer 220 .
  • the partial network 200 receives the at least one input image, and generates a last plurality of feature maps according to at least one third parameter of the generator 20 and at least one first CA.
  • the generator 20 uses the identity mapping 210 to preserve a structure of the at least one input image.
  • the last convolution layer 220 receives the last plurality of feature maps and the at least one input image, and generates an output image according the at least one third parameter, the last plurality of feature maps and the structure of the at least one input image.
  • the generator 20 uses the identity mapping 210 to maintain a structure of the output image according to the structure of the at least one input image.
  • FIG. 3 is a schematic diagram of a discriminator 30 according to an example of the present invention.
  • the discriminator 30 may be utilized for realizing the discriminator module 110 in FIG. 1 .
  • the discriminator 30 includes networks 300 , 310 , 320 and at least one convolution layer 330 .
  • the networks 300 , 310 and 320 include a first CA, a second CA and a third CA, respectively, and generate different feature maps of an output image and a HR image with different scales according to the CAs.
  • the at least one convolution layer 330 receives feature maps generated by the network 320 , and generates at least one score. That is, the discriminator 30 uses the CAs in three scales, to generate at least one score according to the different feature maps with the different scales.
  • the discriminator 30 extracts at least one second feature of the output image and at least one third feature of the HR image by the CAs in at least one scale (e.g., different scales).
  • the network 300 may extract lowest-level features of the output image and the HR image by the first CA in a finest scale.
  • the network 310 may extract median-level features of the output image and the HR image by the second CA in a median scale.
  • the network 320 may extract highest-level features of the output image and the HR image by the third CA in a coarsest scale.
  • the first CA, the second CA and the third CA may be the same or different.
  • the discriminator 30 can downscale the feature maps to reduce computational complexity.
  • weightings e.g., priorities
  • the CA is used in the generator module 100 to weight the plurality of feature maps without the weightings, to compensate details of the at least one input image for generating the output image according to the plurality of weighted feature maps (i.e., the first plurality of feature maps).
  • the CA is used in the discriminator module 110 to weight the plurality of feature maps without the weightings, to extract (e.g., focus on) at least one second feature of the output image and at least one third feature of the HR image for generating the at least one score according to the plurality of weighted feature maps (i.e., the second plurality of feature maps).
  • FIG. 4 is a schematic diagram of a CA 40 according to an example of the present invention.
  • the CA 40 may be utilized in the at least one first CA and the at least one second CA of FIG. 1 .
  • the CA 40 includes a global pooling layer 400 , a first convolution layer 410 , a leaky ReLU function layer 420 , a second convolution layer 430 and a sigmoid function layer 440 and a multiplier 450 .
  • the global pooling layer 400 processes a plurality of feature maps without weightings F: ⁇ f 1 , f 2 , . . .
  • the objective function comprises at least one regularizer (e.g., at least one regularization function) and a Wasserstein Generative Adversarial Network (WGAN) loss function.
  • the at least one regularizer may be a mean square error (MSE) function L 2 or a perceptual loss function L p .
  • the perceptual loss function L p may be another MSE function of a certain layer output of the discriminator module 110 (after activation function layer), which not only allows the discriminator module 110 to focus on the at least one second feature of the output image and the at least one third feature of the HR image, but also helps the generator module 100 to generate the output image that satisfy a perceptual quality.
  • the MSE function L 2 and the perceptual loss function L p may be performed according to the following equations:
  • E I HR , ⁇ SR ( ) is an expected value formula
  • I HR is the HR image
  • G( ⁇ SR ) is the output image generated by the generator module 100
  • ⁇ i indicates the ith layer of feature maps of the discriminator module 110 after the activation function layer.
  • the objective function L total is optimized for solving an adversarial min-max problem, to update the at least one first parameter of the generator module 100 or the at least one second parameter of the discriminator module 110 according to the objective function.
  • the objective function may be performed according to the following equation:
  • FIG. 5 is a comparison table 50 of experiment results according to an example of the present invention.
  • three methods generate the experiment results with different datasets: EnhanceNet, ESRGAN and an example of the present invention, and the experiment results of the methods are then compared.
  • Benchmarks of the experiment results are Perceptual Index (PI) and Root Mean Square Error (RMSE), and the qualitative and quantitative experiment results are shown.
  • PI Perceptual Index
  • RMSE Root Mean Square Error
  • the qualitative and quantitative experiment results are shown.
  • at least one input image is generated according to RCAN.
  • the comparison table 50 the experiment results show that the example of the present invention has better performance than the other methods.
  • the process 60 is utilized in the computing device 10 , and includes the following steps:
  • Step 600 Start.
  • Step 602 Receive at least one input image.
  • Step 604 Generate a first plurality of feature maps of the at least one input image by at least one first CA.
  • Step 606 Generate an output image according to the first plurality of feature maps and at least one first parameter of the generator module.
  • Step 608 Receive a HR image.
  • Step 610 Generate a second plurality of feature maps of the output image and a third plurality of feature maps of the HR image by at least one second CA.
  • Step 612 Generate at least one score according to the second plurality of feature maps, the third plurality of feature maps and at least one second parameter of the discriminator module.
  • Step 614 Update the at least one first parameter and the at least one second parameter according to the at least one score and an objective function
  • Step 616 End.
  • the process 60 is used for illustrating the operations of the computing device 10 . Detailed description and variations of the process 60 can be referred to the previous description, and are not narrated herein.
  • Examples of the hardware may include analog circuit(s), digital circuit (s) and/or mixed circuit (s).
  • the hardware may include application-specific integrated circuit(s) (ASIC(s)), field programmable gate array(s) (FPGA(s)), programmable logic device(s), coupled hardware components or combination thereof.
  • ASIC application-specific integrated circuit
  • FPGA field programmable gate array
  • programmable logic device(s) programmable logic device(s)
  • coupled hardware components or combination thereof the hardware includes general-purpose processor(s), microprocessor(s), controller(s), digital signal processor(s) (DSP(s)) or combination thereof.
  • DSP(s) digital signal processor
  • Examples of the software may include set(s) of codes, set(s) of instructions and/or set(s) of functions retained (e.g., stored) in a storage unit, e.g., a computer-readable medium.
  • the computer-readable medium may include Subscriber Identity Module (SIM), Read-Only Memory (ROM), flash memory, Random Access Memory (RAM), CD-ROM/DVD-ROM/BD-ROM, magnetic tape, hard disk, optical data storage device, non-volatile storage unit, or combination thereof.
  • SIM Subscriber Identity Module
  • ROM Read-Only Memory
  • RAM Random Access Memory
  • CD-ROM/DVD-ROM/BD-ROM Compact Disc-Read Only Memory
  • magnetic tape e.g., hard disk
  • optical data storage device e.g., optical data storage unit, or combination thereof.
  • the computer-readable medium e.g., storage unit
  • the at least one processor which may include one or more modules may (e.g., be configured to) execute the software in the computer-readable medium.
  • the set(s) of codes, the set(s) of instructions and/or the set(s) of functions may cause the at least one processor, the module(s), the hardware and/or the electronic system to perform the related steps.
  • the present invention provides a computing device for handling ISR. Comparing with the prior art, the present invention not only improves perceptual quality of output image but also overcomes structure distortion problem.

Abstract

A computing device for handling image super-resolution (ISR), comprises a generator module, for receiving at least one input image, to generate an output image according to at least one first parameter and a first plurality of feature maps generated by at least one first channel attention (CA); a discriminator module, for receiving the output image and a high resolution (HR) image, to generate a second plurality of feature maps and a third plurality of feature maps by at least one second CA, and to generate at least one score according to the second plurality of feature maps, the third plurality of feature maps and at least one second parameter; and a feedback module, for receiving the at least one score, to update the at least one first parameter and the at least one second parameter according to the at least one score and an objective function.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application No. 62/930,581 filed on Nov. 5, 2019, which is incorporated herein by reference.
  • BACKGROUND OF THE INVENTION 1. Field of the Invention
  • The present invention relates to a device and a method used in a multimedia system, and more particularly, to a device and a method of handling image super-resolution.
  • 2. Description of the Prior Art
  • Image super-resolution (ISR) aims to reconstruct a high-resolution (HR) image from a low resolution (LR) image. Various methods have been proposed for minimizing differences between the HR image and an output image, when handling the ISR. However, some of the methods have a high peak signal-to-noise ratio (PSNR) performance while having a low perceptual quality (e.g., the output image is overly smooth). The others of the methods have a high perceptual quality while having a low PSNR performance (e.g., the output image has structure distortion). Thus, a method for handing the ISR with a better tradeoff between the PSNR performance and the perceptual quality is needed.
  • SUMMARY OF THE INVENTION
  • The present invention therefore provides a device and a method for handling image super-resolution to solve the abovementioned problem.
  • A computing device for handling image super-resolution (ISR), comprises a generator module, for receiving at least one input image, to perform the following operations: generating a first plurality of feature maps of the at least one input image by at least one first channel attention (CA); and generating an output image according to the first plurality of feature maps and at least one first parameter of the generator module; a discriminator module, coupled to the generator module, for receiving the output image and a high resolution (HR) image, to perform the following operations: generating a second plurality of feature maps of the output image and a third plurality of feature maps of the HR image by at least one second CA; and generating at least one score according to the second plurality of feature maps, the third plurality of feature maps and at least one second parameter of the discriminator module; and a feedback module, couple to the generator module and the discriminator module, for receiving the at least one score, to update the at least one first parameter and the at least one second parameter according to the at least one score and an objective function.
  • These and other objectives of the present invention will no doubt become obvious to those of ordinary skill in the art after reading the following detailed description of the preferred embodiment that is illustrated in the various figures and drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a schematic diagram of a computing device according to an example of the present invention.
  • FIG. 2 is a schematic diagram of a generator according to an example of the present invention.
  • FIG. 3 is a schematic diagram of a discriminator according to an example of the present invention.
  • FIG. 4 is a schematic diagram of a CA according to an example of the present invention.
  • FIG. 5 is a comparison table of experiment results according to an example of the present invention.
  • FIG. 6 is a flowchart of a process according to an example of the present invention.
  • DETAILED DESCRIPTION
  • FIG. 1 is a schematic diagram of a computing device 10 according to an example of the present invention. The computing device 10 includes a generator module 100, a discriminator module 110 and a feedback module 120. In detail, the generator module 100 receives at least one input image, and generates a first plurality of feature maps of the at least one input image by at least one first channel attention (CA). The generator module 100 generates an output image according to at least one first parameter of the generator module 100 and the first plurality of feature maps. The discriminator module 110 is coupled to the generator module 100, and receives the output image and a high resolution (HR) image. The discriminator module 110 generates a second plurality of feature maps of the output image and a third plurality of feature maps of the HR image by at least one second CA, and generates at least one score (e.g., at least one probability or at least one distance) according to at least one second parameter of the discriminator module 110, the second plurality of feature maps and the third plurality of feature maps. The feedback module 120 is coupled to the generator module 100 and the discriminator module 110, and receives the at least one score. The feedback module 120 updates the at least one first parameter and the at least one second parameter according to the at least one score and an objective function.
  • In one example, the feedback module 120 generates a first value to update the at least one first parameter of the generator module 100 according to the first value. The feedback module 120 generates a second value to update the at least one second parameter of the discriminator module 110 according to the second value.
  • In one example, the feedback module 120 may generate the first value when updating the at least one first parameter of the generator module 100 is requested (e.g., configurated), and the at least one second parameter of the discriminator module 110 may be fixed. In one example, the feedback module 120 may generate the second value when updating the at least one second parameter of the discriminator module 110 is requested (e.g., configurated), and the at least one first parameter of the generator module 100 may be fixed. That is, the feedback module 120 may only update parameter (s) of a module (i.e., the generator module 100 or the discriminator module 110) in an iteration, and parameter(s) of the other module (i.e., the discriminator module 110 or the generator module 100) is fixed.
  • In one example, the feedback module 120 uses an Adaptive Moment Estimation (ADAM) optimizer to update the at least one first parameter according to the first value in an iteration. After the iteration for updating the at least one first parameter, the generator module 100 may generate a new output image according to at least one new first parameter of the generator module 100 that is updated, wherein the new output image has a higher quality than that of a present output image (i.e., the new output image with higher PSNR performance). Thus, the discriminator module 110 generates at least one new score after receiving the new output image and the HR image, wherein new difference between the at least one new score may be smaller than present difference between at least one present score.
  • In one example, the feedback module 120 uses an ADAM optimizer to update the at least one second parameter according to the second value in an iteration. After the iteration for updating the at least one second parameter, the discriminator module 110 generates at least one new score according to at least one new second parameter of the discriminator module 110 that is updated when receiving the same output image and HR image. New difference between the at least one new score may be larger than present difference between at least one present score.
  • In one example, the at least one first parameter of the generator module 100 is updated after the at least one second parameter of the discriminator module 110 is updated N times, wherein the N is a positive integer. In one example, the at least one second parameter of the discriminator module 110 is updated after the at least one first parameter of the generator module 100 is updated M times, wherein the M is a positive integer.
  • In one example, the discriminator module 110 generates a first score according to the second plurality of feature maps of the output image and the at least one second parameter, after receiving the output image. The discriminator module 110 generates a second score according to the third plurality of feature maps of the HR image and the at least one second parameter, after receiving the HR image. Then, the feedback module 120 receives the first score and the second score, to update the at least one first parameter and the at least one second parameter according to an objective function and these scores.
  • In one example, the discriminator module 110 generates a score for distinguishing (e.g., discriminating) how real (or fake) an image is (i.e., for distinguishing whether the image is more likely to be a HR image or an output image generated by the generator module 100), after receiving the image. Larger difference between at least one score (e.g., the first score and the second score) represents that the discriminator module 110 distinguishes a HR image and an output image more successfully.
  • In one example, a target of the discriminator module 110 is to increase difference between at least one score when receiving the same output image and HR image.
  • In one example, a target of the generator module 100 is to generate an output image with a quality the same as that of the HR image (i.e., to reduce difference between at least one score after generating a new output image with higher PSNR performance than before).
  • In one example, the generator module 100 independently operates without the discriminator module 110 and the feedback module 120, e.g., when the generator module 100 is not able to generate an output image with a higher PSNR performance than before, and/or when an output image and the HR image cannot be distinguished according to at least one score generated by the discriminator module 110 (e.g., difference of the at least one score is close to zero). In one example, the generator module 100 independently operates without the discriminator module 110 and the feedback module 120, e.g., when a termination condition (e.g., a number of iterations is equal to a default value) is satisfied. In this situation, the generator module 100 can start to generate an output image with a high quality close to that of a HR image.
  • In one example, the at least one first CA is the same or different. In one example, the at least one second CA is the same or different. In one example, the at least one first CA is the same as or different from the at least one second CA.
  • In one example, the at least one first CA comprises at least one of a global pooling layer, a convolution layer and an activation function layer. In one example, the at least one second CA comprises at least one of a global pooling layer, a convolution layer and an activation function layer. The activation function layer for the both examples may be a leaky Rectified Linear Unit (ReLU) function layer or a sigmoid function layer.
  • In one example, the generator module 100 generates the first plurality of feature maps by the at least one first CA to focus on (e.g., extract) at least one first feature of the at least one input image in a single scale. In one example, the discriminator module 110 generates the second plurality of feature maps and the third plurality of feature maps by the at least one second CA to preserve at least one second feature of the output image and at least one third feature of the HR image in at least one scale (e.g., different scale).
  • In the prior art, methods for image super-resolution (ISR) may be divided to two categories according to ways for optimizing output images. The two ways may be pixel-wise differences minimization and perceptual quality optimization. Peak signal-to-noise ratio (PSNR)-based methods based on the pixel-wise differences minimization for optimizing the output images are designed to minimize pixel-wise differences between HR images and the output images. Although the PSNR-based methods achieve outstanding performance on PSNR and Structure Similarity (SSIM), the output images generated by the PSNR-based methods may not satisfy evaluation of Human Visual System (HVS). On the other hand, generative adversarial networks (GAN)-based methods based on the perceptual quality optimization for optimizing the output images are designed to reconstruct texture details of input images. Although the GAN-based methods produce a better perceptual quality, the output images generated by the GAN-based methods may have structure distortion and artifacts.
  • The present invention provides a refiner for the ISR methods. The refiner aims to overcome an overly smoothed problem of the PSNR-based methods and a structure distortion problem of the GAN-based methods. Thus, the problems of the PSNR-based methods and the GAN-based methods are solved.
  • In one example, the at least one input image is generated according to a processing method for the ISR. Further, the processing method may be a residual channel attention networks (RCAN), a deep back-projection networks (DBPN) or an enhanced deep super-resolution network (EDSR). In one example, the output image is generated according to a structure of the at least one input image via an identity mapping. In one example, the at least one input image is generated according to a PSNR-based method. Thus, the generator module 100 may only need to focus on missing details (e.g., features) of the at least one input image to generate an output image.
  • FIG. 2 is a schematic diagram of a generator 20 according to an example of the present invention. The generator 20 may be utilized for realizing the generator module 100 in FIG. 1. The generator 20 includes a partial generator network 200, an identity mapping 210 (e.g., an identity mapping shortcut) and a last convolution layer 220. The identity mapping 210 is established between at least one input image and the last convolution layer 220. The partial network 200 receives the at least one input image, and generates a last plurality of feature maps according to at least one third parameter of the generator 20 and at least one first CA. The generator 20 uses the identity mapping 210 to preserve a structure of the at least one input image. The last convolution layer 220 receives the last plurality of feature maps and the at least one input image, and generates an output image according the at least one third parameter, the last plurality of feature maps and the structure of the at least one input image. In other words, the generator 20 uses the identity mapping 210 to maintain a structure of the output image according to the structure of the at least one input image.
  • In one example, the partial network 200 includes others convolution layers, wherein the convolution layers may have stride=1. In one example, the last convolution layer 220 may have stride=1.
  • FIG. 3 is a schematic diagram of a discriminator 30 according to an example of the present invention. The discriminator 30 may be utilized for realizing the discriminator module 110 in FIG. 1. The discriminator 30 includes networks 300, 310, 320 and at least one convolution layer 330. The networks 300, 310 and 320 include a first CA, a second CA and a third CA, respectively, and generate different feature maps of an output image and a HR image with different scales according to the CAs. The at least one convolution layer 330 receives feature maps generated by the network 320, and generates at least one score. That is, the discriminator 30 uses the CAs in three scales, to generate at least one score according to the different feature maps with the different scales.
  • In one example, according to FIG. 3, the discriminator 30 extracts at least one second feature of the output image and at least one third feature of the HR image by the CAs in at least one scale (e.g., different scales). For example, the network 300 may extract lowest-level features of the output image and the HR image by the first CA in a finest scale. The network 310 may extract median-level features of the output image and the HR image by the second CA in a median scale. The network 320 may extract highest-level features of the output image and the HR image by the third CA in a coarsest scale. In one example, the first CA, the second CA and the third CA may be the same or different.
  • In one example, the network 300, 310 and 320 include convolution layers, wherein the convolution layers may have stride=2. The at least one convolution layer 330 may have stride=2. Thus, the discriminator 30 can downscale the feature maps to reduce computational complexity.
  • In one example, operations of the CA are described as follows. The CA receives a plurality of feature maps without weightings (e.g., priorities) F: {f1, f2, . . . , fn}, and generates a plurality of weighted feature maps F:{f′1, f′2, . . . , f′n}=W·F, wherein W:{ω1, ω2, . . . , ωn} are the weightings, and n is a number of the plurality of feature maps. That is, the CA is used in the generator module 100 to weight the plurality of feature maps without the weightings, to compensate details of the at least one input image for generating the output image according to the plurality of weighted feature maps (i.e., the first plurality of feature maps). The CA is used in the discriminator module 110 to weight the plurality of feature maps without the weightings, to extract (e.g., focus on) at least one second feature of the output image and at least one third feature of the HR image for generating the at least one score according to the plurality of weighted feature maps (i.e., the second plurality of feature maps).
  • FIG. 4 is a schematic diagram of a CA 40 according to an example of the present invention. The CA 40 may be utilized in the at least one first CA and the at least one second CA of FIG. 1. The CA 40 includes a global pooling layer 400, a first convolution layer 410, a leaky ReLU function layer 420, a second convolution layer 430 and a sigmoid function layer 440 and a multiplier 450. The global pooling layer 400 processes a plurality of feature maps without weightings F:{f1, f2, . . . , fn} and compresses the plurality of feature maps to generate initial weightings Z=HGP(F), wherein HGP( ) is a global pooling function and a number of the initial weightings is n (i.e., Z:{z1, z2, . . . , zn}). The first convolution layer 410 processes the initial weightings Z=HGP(F), and generates channel-downscaled weightings WD*Z, wherein a number of the channel-downscaled weightings is smaller than the n. The leaky ReLU function layer 420 processes the channel-downscaled weightings WD*Z, and generates leaky ReLU weightings Y=δ(WD*Z), wherein δ( ) is a leaky ReLU function. The second convolution layer 430 processes the leaky ReLU weightings Y=δ(WD*Z), and generates channel-upscaled weightings WU*Y, wherein a number of the channel-upscaled weightings is equal to the n. The sigmoid function layer 440 processes the channel-upscaled weightings WU*Y, and generates final weightings W=S(WU*Y), wherein S( ) is a sigmoid function. The multiplier 450 multiplies the plurality of feature maps without the weightings F:{f1, f2, . . . , fn} and the final weightings W=S(WU*Y), and generates a plurality of weighted features maps F′=W·F.
  • In one example, the objective function comprises at least one regularizer (e.g., at least one regularization function) and a Wasserstein Generative Adversarial Network (WGAN) loss function. The at least one regularizer may be a mean square error (MSE) function L2 or a perceptual loss function Lp. The perceptual loss function Lp may be another MSE function of a certain layer output of the discriminator module 110 (after activation function layer), which not only allows the discriminator module 110 to focus on the at least one second feature of the output image and the at least one third feature of the HR image, but also helps the generator module 100 to generate the output image that satisfy a perceptual quality. The MSE function L2 and the perceptual loss function Lp may be performed according to the following equations:

  • L 2 =E I HR SR [(I HR −G(Ĩ SR))2],  (Eq. 1)

  • L p =E I HR SR [(θi(I HR)−θi(G(Ĩ SR)))2]  (Eq. 2)
  • wherein EI HR SR ( ) is an expected value formula, and IHR is the HR image, and G(ĨSR) is the output image generated by the generator module 100, and θi indicates the ith layer of feature maps of the discriminator module 110 after the activation function layer.
  • The objective function Ltotal is optimized for solving an adversarial min-max problem, to update the at least one first parameter of the generator module 100 or the at least one second parameter of the discriminator module 110 according to the objective function. The objective function may be performed according to the following equation:

  • L total =L+λ 1 L 22 L p,  (Eq. 3)
  • wherein L is the WGAN loss function, and the coefficients λ1 and λ2 in the equation (Eq. 3) are two fixed values.
  • FIG. 5 is a comparison table 50 of experiment results according to an example of the present invention. In FIG. 5, three methods generate the experiment results with different datasets: EnhanceNet, ESRGAN and an example of the present invention, and the experiment results of the methods are then compared. Benchmarks of the experiment results are Perceptual Index (PI) and Root Mean Square Error (RMSE), and the qualitative and quantitative experiment results are shown. In the example of the present invention, at least one input image is generated according to RCAN. According to the comparison table 50, the experiment results show that the example of the present invention has better performance than the other methods.
  • Operations of the computing device 10 in the above examples can be summarized into a process 60 shown in FIG. 6. The process 60 is utilized in the computing device 10, and includes the following steps:
  • Step 600: Start.
  • Step 602: Receive at least one input image.
  • Step 604: Generate a first plurality of feature maps of the at least one input image by at least one first CA.
  • Step 606: Generate an output image according to the first plurality of feature maps and at least one first parameter of the generator module.
  • Step 608: Receive a HR image.
  • Step 610: Generate a second plurality of feature maps of the output image and a third plurality of feature maps of the HR image by at least one second CA.
  • Step 612: Generate at least one score according to the second plurality of feature maps, the third plurality of feature maps and at least one second parameter of the discriminator module.
  • Step 614: Update the at least one first parameter and the at least one second parameter according to the at least one score and an objective function
  • Step 616: End.
  • The process 60 is used for illustrating the operations of the computing device 10. Detailed description and variations of the process 60 can be referred to the previous description, and are not narrated herein.
  • Those skilled in the art should readily make combinations, modifications and/or alterations on the abovementioned description and examples. The abovementioned generator, discriminator, description, functions, modules and/or processes including suggested steps can be realized by means that could be hardware, software, firmware (known as a combination of a hardware device and computer instructions and data that reside as read-only software on the hardware device), an electronic system, or combination thereof.
  • Examples of the hardware may include analog circuit(s), digital circuit (s) and/or mixed circuit (s). For example, the hardware may include application-specific integrated circuit(s) (ASIC(s)), field programmable gate array(s) (FPGA(s)), programmable logic device(s), coupled hardware components or combination thereof. In one example, the hardware includes general-purpose processor(s), microprocessor(s), controller(s), digital signal processor(s) (DSP(s)) or combination thereof.
  • Examples of the software may include set(s) of codes, set(s) of instructions and/or set(s) of functions retained (e.g., stored) in a storage unit, e.g., a computer-readable medium. The computer-readable medium may include Subscriber Identity Module (SIM), Read-Only Memory (ROM), flash memory, Random Access Memory (RAM), CD-ROM/DVD-ROM/BD-ROM, magnetic tape, hard disk, optical data storage device, non-volatile storage unit, or combination thereof. The computer-readable medium (e.g., storage unit) may be coupled to at least one processor internally (e.g., integrated) or externally (e.g., separated). The at least one processor which may include one or more modules may (e.g., be configured to) execute the software in the computer-readable medium. The set(s) of codes, the set(s) of instructions and/or the set(s) of functions may cause the at least one processor, the module(s), the hardware and/or the electronic system to perform the related steps.
  • To sum up, the present invention provides a computing device for handling ISR. Comparing with the prior art, the present invention not only improves perceptual quality of output image but also overcomes structure distortion problem.
  • Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.

Claims (9)

What is claimed is:
1. A computing device for handling image super-resolution (ISR), comprising:
a generator module, for receiving at least one input image, to perform the following operations:
generating a first plurality of feature maps of the at least one input image by at least one first channel attention (CA); and
generating an output image according to the first plurality of feature maps and at least one first parameter of the generator module;
a discriminator module, coupled to the generator module, for receiving the output image and a high resolution (HR) image, to perform the following operations:
generating a second plurality of feature maps of the output image and a third plurality of feature maps of the HR image by at least one second CA; and
generating at least one score according to the second plurality of feature maps, the third plurality of feature maps and at least one second parameter of the discriminator module; and
a feedback module, couple to the generator module and the discriminator module, for receiving the at least one score, to update the at least one first parameter and the at least one second parameter according to the at least one score and an objective function.
2. The computing device of claim 1, wherein the at least one input image is generated according to a processing method for the ISR.
3. The computing device of claim 2, wherein the processing method is a residual channel attention network (RCAN), a deep back-projection network (DBPN) or an enhanced deep super-resolution network (EDSR).
4. The computing device of claim 1, wherein the generator module generates the output image according to a structure of the at least one input image via an identity mapping.
5. The computing device of claim 1, wherein the discriminator module generates the second plurality of feature maps and the third plurality of feature maps by the at least one second CA in at least one scale.
6. The computing device of claim 1, wherein the at least one first CA comprises at least one of a global pooling layer, a convolution layer and an activation function layer.
7. The computing device of claim 6, wherein the activation function layer is a leaky Rectified Linear Unit (ReLU) function layer or a sigmoid function layer.
8. The computing device of claim 1, wherein the objective function comprises at least one regularizer and a Wasserstein Generative Adversarial Network (WGAN) loss function.
9. The computing device of claim 8, wherein the at least one regularizer is a mean square error (MSE) function or a perceptual loss function.
US16/796,872 2019-11-05 2020-02-20 Device and method of handling image super-resolution Active US11010871B1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US16/796,872 US11010871B1 (en) 2019-11-05 2020-02-20 Device and method of handling image super-resolution
CN202010226039.0A CN112785496A (en) 2019-11-05 2020-03-26 Device and method for processing image super-resolution
TW109115806A TWI765264B (en) 2019-11-05 2020-05-13 Device and method of handling image super-resolution

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201962930581P 2019-11-05 2019-11-05
US16/796,872 US11010871B1 (en) 2019-11-05 2020-02-20 Device and method of handling image super-resolution

Publications (2)

Publication Number Publication Date
US20210133925A1 true US20210133925A1 (en) 2021-05-06
US11010871B1 US11010871B1 (en) 2021-05-18

Family

ID=75687561

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/796,872 Active US11010871B1 (en) 2019-11-05 2020-02-20 Device and method of handling image super-resolution

Country Status (3)

Country Link
US (1) US11010871B1 (en)
CN (1) CN112785496A (en)
TW (1) TWI765264B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113783643A (en) * 2021-09-13 2021-12-10 中国电信股份有限公司 Channel parameter generation method and related equipment
CN114170089A (en) * 2021-09-30 2022-03-11 成都大学附属医院 Method and electronic device for diabetic retinopathy classification
CN114547017A (en) * 2022-04-27 2022-05-27 南京信息工程大学 Meteorological big data fusion method based on deep learning
CN115131242A (en) * 2022-06-28 2022-09-30 闽江学院 Lightweight super-resolution reconstruction method based on attention and distillation mechanism

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170256033A1 (en) * 2016-03-03 2017-09-07 Mitsubishi Electric Research Laboratories, Inc. Image Upsampling using Global and Local Constraints

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103514580B (en) * 2013-09-26 2016-06-08 香港应用科技研究院有限公司 For obtaining the method and system of the super-resolution image that visual experience optimizes
CN105069825B (en) * 2015-08-14 2018-06-12 厦门大学 Image super-resolution rebuilding method based on depth confidence network
CN105678698A (en) * 2015-12-30 2016-06-15 南京信息工程大学 Non-local regularization remote sensing image super-resolution reconstruction method
CN106204449B (en) * 2016-07-06 2019-09-10 安徽工业大学 A kind of single image super resolution ratio reconstruction method based on symmetrical depth network
WO2018212599A1 (en) * 2017-05-17 2018-11-22 Samsung Electronics Co., Ltd. Super-resolution processing method for moving image and image processing apparatus therefor
CN107481188A (en) * 2017-06-23 2017-12-15 珠海经济特区远宏科技有限公司 A kind of image super-resolution reconstructing method
CN107492070B (en) * 2017-07-10 2019-12-03 华北电力大学 A kind of single image super-resolution calculation method of binary channels convolutional neural networks
CN109146784B (en) * 2018-07-27 2020-11-20 徐州工程学院 Image super-resolution reconstruction method based on multi-scale generation countermeasure network
CN109886871B (en) * 2019-01-07 2023-04-07 国家新闻出版广电总局广播科学研究院 Image super-resolution method based on channel attention mechanism and multi-layer feature fusion
CN110276721A (en) * 2019-04-28 2019-09-24 天津大学 Image super-resolution rebuilding method based on cascade residual error convolutional neural networks
CN110363716B (en) * 2019-06-25 2021-11-19 北京工业大学 High-quality reconstruction method for generating confrontation network composite degraded image based on conditions

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170256033A1 (en) * 2016-03-03 2017-09-07 Mitsubishi Electric Research Laboratories, Inc. Image Upsampling using Global and Local Constraints

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
"Advances in Cryptology - CRYPTO 2018, Part III", vol. 11211, 1 January 2018, SPRINGER, Cham, ISBN: 978-3-030-71592-2, ISSN: 0302-9743, article ZHANG YULUN, LI KUNPENG, LI KAI, WANG LICHEN, ZHONG BINENG, FU YUN: "Image Super-Resolution Using Very Deep Residual Channel Attention Networks : 15th European Conference, Munich, Germany, September 8–14, 2018, Proceedings, Part VII", pages: 294 - 310, XP055792295, DOI: 10.1007/978-3-030-01234-2_18 *
"Advances in Cryptology - CRYPTO 2018, Part III", vol. 27, 26 September 2018, SPRINGER, Cham, ISBN: 978-3-030-71592-2, ISSN: 0302-9743, article SEITZER MAXIMILIAN; YANG GUANG; SCHLEMPER JO; OKTAY OZAN; WÜRFL TOBIAS; CHRISTLEIN VINCENT; WONG TOM; MOHIADDIN RAAD; FIRMIN DAVID: "Adversarial and Perceptual Refinement for Compressed Sensing MRI Reconstruction", pages: 232 - 240, XP047557613, DOI: 10.1007/978-3-030-00928-1_27 *
"Advances in Cryptology - CRYPTO 2018, Part III", vol. 38, 17 September 2016, SPRINGER, Cham, ISBN: 978-3-030-71592-2, ISSN: 0302-9743, article HE KAIMING; ZHANG XIANGYU; REN SHAOQING; SUN JIAN: "Identity Mappings in Deep Residual Networks", pages: 630 - 645, XP047565992, DOI: 10.1007/978-3-319-46493-0_38 *
BROWNLEE JASON .: "How to Code the GAN Training Algorithm and Loss Functions", MACHINE LEARNING MASTERY, 12 July 2019 (2019-07-12), XP055792303, Retrieved from the Internet <URL:https://machinelearningmastery.com/how-to-code-the-generative-adversarial-network-training-algorithm-and-loss-functions/> *
CAO FEILONG, HUAN LIU: "Single image super-resolution via multi-scale residual channel attention network", NEUROCOMPUTING, vol. 358, 25 May 2019 (2019-05-25), pages 424 - 436, XP055792305, DOI: 10.1016/j.neucom.2019.05.066 *
JIE HU, LI SHEN, GANG SUN: "Squeeze-and-Excitation Networks", EYE IN-PAINTING WITH EXEMPLAR GENERATIVE ADVERSARIAL NETWORKS, 1 June 2018 (2018-06-01) - 23 June 2018 (2018-06-23), pages 7132 - 7141, XP055617919, ISBN: 978-1-5386-6420-9, DOI: 10.1109/CVPR.2018.00745 *
MARTIN ARJOVSKY, CHINTALA SOUMITH, BOTTOU LÉON: "Wasserstein Generative Adversarial Networks", 11 August 2017 (2017-08-11), XP055524181, Retrieved from the Internet <URL:https://leon.bottou.org/publications/pdf/icml-2017.pdf> [retrieved on 20181115] *
RONG CHEN ; YUAN XIE ; XIAOTONG LUO ; YANYUN QU ; CUIHUA LI: "Joint-attention Discriminator for Accurate Super-resolution via Adversarial Training", MULTIMEDIA, ACM, 2 PENN PLAZA, SUITE 701NEW YORKNY10121-0701USA, 15 October 2019 (2019-10-15) - 25 October 2019 (2019-10-25), 2 Penn Plaza, Suite 701New YorkNY10121-0701USA, pages 711 - 719, XP058442577, ISBN: 978-1-4503-6889-6, DOI: 10.1145/3343031.3351008 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113783643A (en) * 2021-09-13 2021-12-10 中国电信股份有限公司 Channel parameter generation method and related equipment
CN114170089A (en) * 2021-09-30 2022-03-11 成都大学附属医院 Method and electronic device for diabetic retinopathy classification
CN114170089B (en) * 2021-09-30 2023-07-07 成都市第二人民医院 Method for classifying diabetic retinopathy and electronic equipment
CN114547017A (en) * 2022-04-27 2022-05-27 南京信息工程大学 Meteorological big data fusion method based on deep learning
CN115131242A (en) * 2022-06-28 2022-09-30 闽江学院 Lightweight super-resolution reconstruction method based on attention and distillation mechanism

Also Published As

Publication number Publication date
US11010871B1 (en) 2021-05-18
TW202119294A (en) 2021-05-16
TWI765264B (en) 2022-05-21
CN112785496A (en) 2021-05-11

Similar Documents

Publication Publication Date Title
US11010871B1 (en) Device and method of handling image super-resolution
CN109087273B (en) Image restoration method, storage medium and system based on enhanced neural network
US20190294931A1 (en) Systems and Methods for Generative Ensemble Networks
CN111899163A (en) Efficient structure preservation for generating single image super-resolution in countermeasure networks
WO2020098422A1 (en) Encoded pattern processing method and device , storage medium and electronic device
JP5159844B2 (en) Image processing device
CN111340820B (en) Image segmentation method and device, electronic equipment and storage medium
CN114037888B (en) Target detection method and system based on joint attention and adaptive NMS
CN114612289A (en) Stylized image generation method and device and image processing equipment
CN108921801B (en) Method and apparatus for generating image
CN113221925A (en) Target detection method and device based on multi-scale image
JP2019164618A (en) Signal processing apparatus, signal processing method and program
CN115984570A (en) Video denoising method and device, storage medium and electronic device
CN112801890B (en) Video processing method, device and equipment
JP2020177582A (en) Leaning device, learning method, program, and recognition device
WO2020248706A1 (en) Image processing method, device, computer storage medium, and terminal
US20120038785A1 (en) Method for producing high resolution image
Xiao et al. Video denoising algorithm based on improved dual‐domain filtering and 3D block matching
Kokkinos et al. Pixel adaptive filtering units
Chang et al. R2p: Recomposition and retargeting of photographic images
Brateanu et al. LYT-Net: Lightweight YUV Transformer-based Network for Low-Light Image Enhancement
CN115760658A (en) Image processing method, image processing device, storage medium and electronic equipment
CN113888419A (en) Method for removing dark corners of image
CN113284059A (en) Model training method, image enhancement method, device, electronic device and medium
CN114363532A (en) Focusing method and related device

Legal Events

Date Code Title Description
AS Assignment

Owner name: MOXA INC., TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEE, WEI-YU;CHUANG, PO-YU;WANG, YU-CHIANG;REEL/FRAME:051882/0823

Effective date: 20200219

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE