US11449751B2 - Training method for generative adversarial network, image processing method, device and storage medium - Google Patents

Training method for generative adversarial network, image processing method, device and storage medium Download PDF

Info

Publication number
US11449751B2
US11449751B2 US16/759,669 US201916759669A US11449751B2 US 11449751 B2 US11449751 B2 US 11449751B2 US 201916759669 A US201916759669 A US 201916759669A US 11449751 B2 US11449751 B2 US 11449751B2
Authority
US
United States
Prior art keywords
image
resolution
network
generative
output
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/759,669
Other versions
US20200285959A1 (en
Inventor
Hanwen Liu
Dan Zhu
Pablo Navarrete Michelini
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BOE Technology Group Co Ltd
Original Assignee
BOE Technology Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN201811155326.6A external-priority patent/CN109345455B/en
Priority claimed from CN201811155930.9A external-priority patent/CN109345456B/en
Priority claimed from CN201811155147.2A external-priority patent/CN109360151B/en
Priority claimed from CN201811155252.6A external-priority patent/CN109255390B/en
Application filed by BOE Technology Group Co Ltd filed Critical BOE Technology Group Co Ltd
Publication of US20200285959A1 publication Critical patent/US20200285959A1/en
Application granted granted Critical
Publication of US11449751B2 publication Critical patent/US11449751B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2148Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the process organisation or structure, e.g. boosting cascade
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06K9/6232
    • G06K9/6257
    • G06K9/6268
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • G06N3/0454
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0475Generative networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • G06N3/0481
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/094Adversarial learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/20Linear translation of whole images or parts thereof, e.g. panning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4046Scaling of whole images or parts thereof, e.g. expanding or contracting using neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • G06T3/4076Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution using the original low-resolution images to iteratively correct the high-resolution images
    • G06T5/002
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • G06V10/449Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
    • G06V10/451Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
    • G06V10/454Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/751Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination

Definitions

  • the present disclosure relates to, but is not limited to, the field of image processing, and in particular, to a training method for generative adversarial network, an image processing method using a generative adversarial network obtained by the training method, a computer device, and a computer-readable storage medium.
  • a convolutional neural network is a common deep learning network, and has been widely applied to the field of image processing nowadays to achieve image identification, image classification, super-resolution image reconstruction, and so on.
  • a second-resolution image reconstructed based on a first-resolution image usually lacks detail information, which makes the second-resolution image look unreal.
  • the present disclosure provides a training method for generative adversarial network
  • the generative adversarial network includes a generative network and a discriminative network
  • the generative network is configured to convert a first-resolution image into a second-resolution image, with a resolution of the second-resolution image higher than that of the first-resolution image
  • the training method includes a generative network training procedure, which includes:
  • the first input image includes the first-resolution sample image and a first noise image corresponding to a noise sample with a first amplitude
  • the second input image includes the first-resolution sample image and a second noise image corresponding to a noise sample with a second amplitude
  • the first amplitude is greater than 0, and the second amplitude is equal to 0;
  • the loss function of the generative network includes a first loss, a second loss and a third loss, the first loss of the loss function is based on a reconstruction error between the second output image and the second-resolution sample image; the second loss of the loss function is based on a perceptual error between the first output image and the second-resolution sample image; and the third loss of the loss function is based on the first discrimination result and the second discrimination result.
  • the reconstruction error between the second output image and the second-resolution sample image is determined according to any one of L1 norm of a difference image matrix between the second output image and the second-resolution sample image, a mean square error between the second output image and the second-resolution sample image, and a structural similarity index between the second output image and the second-resolution sample image.
  • X denotes the second-resolution sample image
  • L denotes a total number of times of the resolution enhancement procedure in the iteration process
  • LR denotes the first-resolution sample image
  • E[ ] denotes calculation of matrix energy
  • ⁇ 1 is a preset weight.
  • L CX ( ) is a contextual loss calculation function
  • ⁇ 2 is a preset weight
  • ⁇ 3 is a preset weight.
  • ⁇ 1 : ⁇ 2 : ⁇ 3 10:0.1:0.001
  • the noise sample is random noise.
  • the training method further includes a discriminative network training procedure, which includes: separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a discrimination result based on the first output image and a discrimination result based on the second-resolution sample image, respectively; and adjusting parameters of the discriminative network to reduce a loss function of the discriminative network;
  • the discriminative network training procedure and the generative network training procedure are alternately performed until a preset training condition is met.
  • both the first output image and the second output image are generated by the generative network through an iteration process of the resolution enhancement procedure, and a total number of times of the resolution enhancement procedure in the iteration process is L; and when L is greater than 1, in the previous L ⁇ 1 times of the resolution enhancement procedure in the iteration process performed by the generative network based on the first input image, the generative network generates an intermediate image each time the resolution enhancement procedure is performed;
  • each intermediate image generated by the generative network based on the first input image is provided for the discriminative network, while the first output image is being provided for the discriminative network; and while the second-resolution sample image is being provided for the discriminative network, and third-resolution sample images obtained by downsampling the second-resolution sample image are provided for the discriminative network, the third-resolution sample images being in one-to-one correspondence with intermediate images, and each having a resolution the same as that of the corresponding intermediate image
  • the present disclosure further provides an image processing method using the generative network of the generative adversarial network obtained by the training method, and the image processing method is used for increasing a resolution of an image, and includes:
  • an amplitude of the reference noise ranges from 0 to the first amplitude.
  • the reference noise is random noise.
  • the present disclosure further provides a computer device including a memory having computer programs stored thereon, and a processor, and the above training method is performed when the computer programs are executed by the processor.
  • the present disclosure further provides a computer-readable storage medium having computer programs stored thereon, and the above training method is performed when the computer programs are executed by a processor.
  • FIG. 1 is a schematic diagram illustrating a relationship between reconstruction distortion and perceptual distortion
  • FIG. 2 is a flowchart illustrating a generative network training procedure according to the embodiments of the present disclosure.
  • FIG. 3 is a schematic structural diagram of a generative network according to the embodiments of the present disclosure.
  • Super-resolution image reconstruction is a technology for increasing a resolution of an initial image to obtain an image with a higher resolution.
  • reconstruction distortion and perceptual distortion are used for evaluating a super-resolution reconstruction effect.
  • the reconstruction distortion is used for measuring a difference between a reconstructed image and a reference image, and specific evaluation criteria include mean square error (MSE), structural similarity index (SSIM), and peak signal-to-noise ratio (PSNR); and the perceptual distortion mainly focuses on making the image more look like a natural image.
  • FIG. 1 is a schematic diagram illustrating a relationship between the reconstruction distortion and the perceptual distortion. As shown in FIG.
  • the perceptual distortion when the reconstruction distortion is relatively small, the perceptual distortion is relatively large, in which case the reconstructed image looks smoother but lacks details.
  • the reconstruction distortion is relatively large, in which case the reconstructed image has more details.
  • the current super-resolution image reconstruction methods usually aim at relatively small reconstruction distortion, but people prefer to obtaining reconstructed images with rich details in some application scenarios.
  • the present disclosure provides a training method for generative adversarial network
  • the generative adversarial network includes a generative network and a discriminative network
  • the generative network is configured to convert a first-resolution image into a second-resolution image to obtain the second-resolution image having a target resolution, and the resolution of the second-resolution image is higher than that of the first-resolution image.
  • the generative network can obtain the second-resolution image by performing a resolution enhancement procedure once or iterating a resolution enhancement procedure for a plurality of times.
  • an image to be processed (i.e., a first-resolution image) has a resolution of 128 ⁇ 128 and the target resolution is 1024 ⁇ 1024, the generative network may obtain the second-resolution image having a resolution of 1024 ⁇ 1024 by performing once the resolution enhancement procedure which increases a resolution by 8 times; or the generative network may obtain a 256 ⁇ 256 image, a 512 ⁇ 512 image and a 1024 ⁇ 1024 image in sequence by iterating the resolution enhancement procedure, which increases a resolution by 2 times, three times.
  • the training method for generative adversarial network includes a generative network training procedure.
  • FIG. 2 is a flowchart illustrating the generative network training procedure according to the embodiments of the present disclosure. As shown in FIG. 2 , the generative network training procedure includes following S 1 through S 4 .
  • the first-resolution sample image may be obtained by downsampling the second-resolution sample image.
  • the amplitude of the noise sample is an average fluctuation amplitude of the noise sample.
  • the noise sample is random noise
  • a mean of an image corresponding to the noise sample is ⁇
  • a variance of the image corresponding to the noise sample is ⁇ , that is, most pixel values of the image corresponding to the noise sample fluctuate from ⁇ - ⁇ to ⁇ + ⁇ , in which case a noise amplitude is ⁇ .
  • any image is shown in the form of matrix in an image processing process, and the pixel values represent element values of an image matrix.
  • each element value of the image matrix may be considered to be 0.
  • the training method for generative adversarial network includes a plurality of generative network training procedures; and in a single generative network training procedure, the first-resolution sample image is the single one, and model parameters of the generative network when receiving the first input image and the second input image are the same.
  • the first discrimination result is used for representing a matching degree between the first output image and the second-resolution sample image, for example, the first discrimination result is used for representing a probability determined by the discriminative network that the first output image is identical to the second-resolution sample image; and the second discrimination result is used for representing a probability determined by the discriminative network that the second-resolution sample image is indeed the second-resolution sample image.
  • the discriminative network may be regarded as a classifier having a scoring function.
  • the discriminative network can score a received to-be-discriminated image, and output a score which indicates a probability that the to-be-discriminated image (the first output image) is identical to the second-resolution sample image, that is, indicating the matching degree mentioned above, which may range from 0 to 1.
  • the output score of the discriminative network is 0 or close to 0, it is indicated that the discriminative network classifies the received to-be-discriminated image as a non-high-resolution sample image; and when the output score of the discriminative network is 1 or close to 1, it is indicated that the received to-be-discriminated image is identical to the second-resolution sample image.
  • the scoring function of the discriminative network may be trained by use of a “true” sample and a “false” sample with predetermined scores.
  • the “false” sample is an image generated by the generative network and the “true” sample is the second-resolution sample image.
  • a training process of the discriminative network is a process of adjusting parameters of the discriminative network to enable the discriminative network to output a score close to 1 when receiving the “true” sample, and output a score close to 0 when receiving the “false” sample.
  • the loss function of the generative network includes a first loss, a second loss, and a third loss; specifically, the loss function is superposition of the first loss, the second loss and the third loss, and the first loss is based on a reconstruction error between the second output image and the second-resolution sample image; the second loss is based on a perceptual error between the first output image and the second-resolution sample image; and the third loss is based on the first discrimination result and the second discrimination result.
  • the second input image including a noise image with an amplitude of 0 and the first input image including a noise image with an amplitude of 1 are separately provided for the generative network for training, and the first loss of the loss function reflects the reconstruction distortion of a result generated by the generative network, and the second loss reflects the perceptual distortion of the result generated by the generative network, that is, the loss function combines two distortion evaluation criteria.
  • an amplitude of input noise can be adjusted according to actual needs (i.e., whether details of the image need to be emphasized and to what extent the details are emphasized), so that a reconstructed image can meet the actual needs. For example, within a given range of reconstruction distortion, minimum perceptual distortion is achieved by adjusting the amplitude of the input noise; or within a given range of perceptual distortion, minimum reconstruction distortion is achieved by adjusting the amplitude of the input noise.
  • the amplitude of the noise image of the first input image which is 1 in the embodiment, is an amplitude value obtained by normalizing the amplitude of the noise image. In other embodiments of the present disclosure, it is possible not to normalize the amplitude of the noise image, so that the amplitude of the noise image of the first input image may be not equal to 1.
  • the noise sample is random noise; and a mean of the first noise image is 1.
  • a mean of the first noise image is a mean of a normalized image of the first noise image. For example, if the first noise image is a grayscale image, an average of all pixel values in an image obtained by normalizing the first noise image is the mean of the first noise image; as another example, if the first noise image is a color image, an average of all pixel values in an image obtained by normalizing every channel of the first noise image is the mean of the first noise image.
  • the channel of the image in the embodiment of the present disclosure indicates one or more channels obtained by dividing an image for processing, for example, an RGB-mode color image may be divided into three channels, i.e., a red channel, a green channel, and a blue channel; if the image is a grayscale image, it is a one-channel image; and if the color image is divided according to an HSV color system, the image may be divided into three channels, i.e., a hue (H) channel, a saturation (S) channel, and a value (V) channel.
  • H hue
  • S saturation
  • V value
  • ⁇ 1 : ⁇ 2 : ⁇ 3 may be set according to continuity of local images. While in some other embodiments, ⁇ 1 : ⁇ 2 : ⁇ 3 may be set according to target pixels of an image.
  • both the first output image and the second output image are generated by the generative network through an iteration process of a resolution enhancement procedure; and a total number of times of the resolution enhancement procedure in the iteration process is L, and L ⁇ 1.
  • LR denotes the first-resolution sample image
  • the downsampling may be performed in a way the same as that for extracting the first-resolution sample image from the second-resolution sample image in the step S 1 .
  • E[ ] denotes calculation of matrix energy.
  • E[ ] can calculate a maximum or average of the elements in a matrix in “[ ]”.
  • a third-resolution image i.e., HR 1 , HR 2 , . . . , HR L ⁇ 1
  • L1 norm of a difference image between the third-resolution image, the image generated by downsampling the second output image, and the first-resolution sample image is also calculated.
  • the resolution of the third-resolution image is higher than that of the first-resolution sample image, and is the same as that of the third-resolution sample image.
  • MSE mean square error
  • SSIM structural similarity index
  • the downsampling may be performed in a way the same as that for extracting the first-resolution sample image from the second-resolution sample image in the step S 1 .
  • L CX ( ) denotes a contextual loss calculation function
  • the training method further includes a discriminative network training procedure, which includes: separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a discrimination result based on the first output image and a discrimination result based on the second-resolution sample image, respectively; and adjusting the parameters of the discriminative network to reduce a loss function of the discriminative network.
  • a discriminative network training procedure which includes: separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a discrimination result based on the first output image and a discrimination result based on the second-resolution sample image, respectively; and adjusting the parameters of the discriminative network to reduce a loss function of the discriminative network.
  • the discriminative network training procedure and the generative network training procedure are alternately performed until a preset training condition is met.
  • the preset training condition may be that the number of alternation times reaches a predetermined value.
  • the parameters of the generative network and the discriminative network are preset or random.
  • both the first output image and the second output image are generated by the generative network through an iteration process of the resolution enhancement procedure, and a total number of times of the resolution enhancement procedure in the iteration process is L.
  • L a total number of times of the resolution enhancement procedure in the iteration process
  • L>1 it is possible to supply the first output image or the second-resolution sample image alone to the discriminative network each time an image is supplied to the discriminative network.
  • L>1 in the previous L ⁇ 1 times of the resolution enhancement procedure performed by the generative network based on the first input image, the generative network generates an intermediate image each time the resolution enhancement procedure is performed; and when the resolution enhancement procedure is iterated for the L th time, the image generated by the generative network is the first output image.
  • the discriminative network is provided with a plurality of input terminals to receive a plurality of images simultaneously, and determines a matching degree between one of the received plurality of images, which has a highest resolution, and the second-resolution sample image.
  • each intermediate image generated by the generative network based on the first input image is provided for the discriminative network, while the first output image is being provided for the discriminative network; and while the second-resolution sample image is being provided for the discriminative network, third-resolution sample images obtained by downsampling the second-resolution sample image are provided for the discriminative network, the third-resolution sample images being in one-to-one correspondence with intermediate images, and each having a resolution the same as that of the corresponding intermediate image.
  • the parameters of the generative network are adjusted to enable the discriminative network to output a matching degree as close to 1 as possible as a discrimination result after an output result of the generative network is input into the discriminative network, that is, to enable the discriminative network to regard the output result of the generative network as the second-resolution sample image.
  • the parameters of the discriminative network are adjusted to enable the discriminative network to output a matching degree as close to 1 as possible after the second-resolution sample image is input into the discriminative network, and also enable the discriminative network to output a matching degree as close to 0 as possible after an output result of the generative network is input into the discriminative network; that is, the discriminative network can be trained to be capable of determining whether a received image is the second-resolution sample image.
  • the discriminative network is continuously optimized to improve discrimination capability, and the generative network is continuously optimized to output a result as close to the second-resolution sample image as possible.
  • the two “opposing” models compete with each other and each is improved based on an increasingly better result from the other one in each training process, so that the generative adversarial network model obtained are getting better and better.
  • the present disclosure further provides an image processing method using a generative adversarial network obtained by the above training method, and the image processing method is used for increasing a resolution of an image by using a generative network of the generative adversarial network, and includes providing an input image and a noise image corresponding to reference noise for the generative network to allow the generative network to generate an image having a higher resolution than the input image.
  • An amplitude of the reference noise ranges from 0 to a first amplitude.
  • the reference noise is random noise.
  • the noise sample with an amplitude of 0 and the noise sample with a first amplitude are separately provided for the generative network, and the loss function of the generative network combines two distortion evaluation criteria for evaluating the reconstruction distortion and the perceptual distortion, so that the amplitude of reference noise can be adjusted according to the actual needs when the generative network is used to increase a resolution of an image, so as to meet the actual needs. For example, within a given range of reconstruction distortion, minimum perceptual distortion is achieved by adjusting the amplitude of the reference noise; or within a given range of perceptual distortion, minimum reconstruction distortion is achieved by adjusting the amplitude of the reference noise.
  • FIG. 3 is a schematic structural diagram of a generative network according to the embodiments of the present disclosure.
  • the generative network is described below in conjunction with FIG. 3 .
  • the generative network is used for iterating a resolution enhancement procedure, and a resolution of a to-be-processed image I l ⁇ 1 is increased each time the resolution enhancement procedure is performed, so as to obtain an image I l with an increased resolution.
  • the to-be-processed image I l ⁇ 1 is an initial input image; when the total number of times of the iteration of the resolution enhancement procedure is L and L>1, the to-be-processed image I l ⁇ 1 is an image output after iterating the resolution enhancement procedure for the (l ⁇ 1) th time.
  • the to-be-processed image I l ⁇ 1 in FIG. 3 is a 256 ⁇ 256 image obtained after performing the resolution enhancement procedure once.
  • the generative network includes a first analysis module 11 , a second analysis module 12 , a first concatenating module 21 , a second concatenating module 22 , an interpolation module 31 , a first upsampling module 41 , a first downsampling module 51 , a superposition module 70 , and a residual correction system for iteration.
  • the first analysis module 11 is configured to generate a feature image R l ⁇ 1 ⁇ of the to-be-processed image I l ⁇ 1 , and a number of channels of the feature image R l ⁇ 1 ⁇ is greater than that of the to-be-processed image I l ⁇ 1 .
  • the first concatenating module 21 is configured to concatenate the feature image R l ⁇ 1 ⁇ of the to-be-processed image and a noise image to obtain a first merged image RC l ⁇ 1 ⁇ ; and a number of channels of the first merged image RC l ⁇ 1 ⁇ is a sum of the number of the channels of the feature image R l ⁇ 1 ⁇ and a number of channels of the noise image.
  • each of the first input image and the second input image provided for the generative network may include the first-resolution sample image and a plurality of noise sample images having different resolutions; or each of the first input image and the second input image may include the first-resolution sample image and one noise sample image, and when the resolution enhancement procedure is iterated for the l th time, the generative network generates the noise sample image at a required magnification according to an amplitude of a noise sample.
  • the interpolation module 31 is configured to perform interpolation on the to-be-processed image I l ⁇ 1 to obtain a fourth-resolution image based thereon, the fourth-resolution image having a resolution of 512 ⁇ 512.
  • the interpolation module may perform the interpolation by using traditional interpolation methods, such as bicubic interpolation.
  • the resolution of the fourth-resolution image is higher than that of the to-be-processed image I l ⁇ 1 .
  • the second analysis module 12 is configured to generate a feature image of the fourth-resolution image, a number of channels of the feature image being greater than that of the fourth-resolution image.
  • the first downsampling module 51 is configured to downsample the feature image of the fourth-resolution image to obtain a first downsampled feature image having a resolution of 256 ⁇ 256.
  • the second concatenating module 22 is configured to concatenate the first merged image RC l ⁇ 1 ⁇ and the first downsampled feature image to obtain a second merged image.
  • the first upsampling module 41 is configured to upsample the second merged image to obtain a first upsampled feature image R l 0 .
  • the residual correction system for iteration is configured to perform residual correction on the first upsampled feature image through back-projection for at least one time, so as to obtain a residual-corrected feature image.
  • the residual correction system for iteration includes a second downsampling module 52 , a second upsampling module 42 , and a residual determination module 60 .
  • the second downsampling module 52 is configured to downsample a received image by 2 times
  • the second upsampling module 42 is configured to upsample a received image by 2 times
  • the residual determination module 60 is configured to determine a difference image between two received images.
  • the first upsampled feature image R l 0 is downsampled by 2 times by the first one second downsampling module 52 to obtain a feature image R l 01 ;
  • the feature image R l 01 is downsampled by 2 times by the second one second downsampling module 52 to obtain a feature image R l 02 having a same resolution as the initial input image;
  • one residual determination module is used to obtain a difference image between the feature image R l 02 and the first merged image RC ⁇ 0 obtained in the first time of the resolution enhancement procedure (i.e., the first merged image RC ⁇ 0 obtained by merging the feature image of the initial input image and a noise image); then, the difference image is upsampled by the second upsampling module to obtain a feature image, and the obtained upsampled feature image is superposed on the feature image R l 01 by the superposition module 70 , so as to obtain a feature image R 03 l having a same resolution as a first merged image R l
  • the feature image R l 1 may be subjected to the second residual correction in the same way to obtain a feature image R l 2 subjected to the second residual correction; and the feature image R l 2 may be subjected to the third residual correction in the same way, and so on.
  • represents a number of times of the residual correction.
  • the generative network further includes a synthesis module 80 configured to synthesize a feature image R l ⁇ obtained after a plurality of times of residual correction, so as to obtain a fifth-resolution image, a number of channels of the fifth-resolution image being the same as that of the fourth-resolution image; and the fifth-resolution image and the fourth-resolution image are superposed to obtain an output image I l after the resolution enhancement procedure is performed for the l th time.
  • a resolution of the fifth-resolution image is the same as that of the fourth-resolution image.
  • the first analysis module 11 , the second analysis module 12 , the first upsampling module 41 , the second upsampling module 42 , the first downsampling module 51 , the second downsampling module 52 , and the synthesis module 80 can perform corresponding functions through a convolutional layer.
  • the present disclosure further provides a computer device including a memory having computer programs stored thereon, and a processor, and the above training method for generative adversarial network is performed when the computer programs are executed by the processor.
  • the present disclosure further provides a computer-readable storage medium having computer programs stored thereon, and the above training method for generative adversarial network is performed when the computer programs are executed by a processor.
  • the above memory and computer-readable storage medium include, but are not limited to, the following readable media: random access memories (RAMs), read-only memories (ROMs), non-volatile random access memories (NVRAMs), programmable read-only memories (PROMs), erasable programmable read-only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), flash memories, magnetic or optical data memories, registers, magnetic disks or tapes, optical storage media such as compact discs (CDs) or digital versatile discs (DVDs), and other non-transitory media.
  • the processor include, but are not limited to, a general-purpose processor, a central processing unit (CPU), a microprocessor, a digital signal processor (DSP), a controller, a microcontroller, a state machine, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Computational Mathematics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Algebra (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The present disclosure provides a training method for generative adversarial network, which includes: extracting a first-resolution sample image from a second-resolution sample image; separately providing a first input image and a second input image for a generative network to generate a first output image and a second output image respectively, the first input image including a first-resolution sample image and a first noise image, the second input image including the first-resolution sample image and a second noise image; separately providing the first output image and a second-resolution sample image for a discriminative network to output a first discrimination result and a second discrimination result; and adjusting parameters of the generative network to reduce a loss function. The present disclosure further provides an image processing method using the generative adversarial network, a computer device, and a computer-readable storage medium.

Description

CROSS REFERENCE TO RELATED APPLICATIONS
This is a National Phase Application filed under 35 U.S.C. 371 as a national stage of PCT/CN2019/107761, filed Sep. 25, 2019, an application claiming the benefit of Chinese Application No. 201811155930.9, filed Sep. 30, 2018, Chinese Application No. 201811155326.6, filed Sep. 30, 2018, Chinese Application No. 201811155147.2, filed Sep. 30, 2018 and Chinese Application No. 201811155252.6, filed Sep. 30, 2018 the content of each of which is hereby incorporated by reference in its entirety.
TECHNICAL FIELD
The present disclosure relates to, but is not limited to, the field of image processing, and in particular, to a training method for generative adversarial network, an image processing method using a generative adversarial network obtained by the training method, a computer device, and a computer-readable storage medium.
BACKGROUND
A convolutional neural network is a common deep learning network, and has been widely applied to the field of image processing nowadays to achieve image identification, image classification, super-resolution image reconstruction, and so on.
With current methods of super-resolution reconstruction, a second-resolution image reconstructed based on a first-resolution image (a resolution of the second-resolution image is higher than that of the first-resolution image) usually lacks detail information, which makes the second-resolution image look unreal.
SUMMARY
The present disclosure provides a training method for generative adversarial network, the generative adversarial network includes a generative network and a discriminative network, the generative network is configured to convert a first-resolution image into a second-resolution image, with a resolution of the second-resolution image higher than that of the first-resolution image, and the training method includes a generative network training procedure, which includes:
extracting a first-resolution sample image from a second-resolution sample image, with a resolution of the second-resolution sample image higher than that of the first-resolution sample image;
separately providing a first input image and a second input image for the generative network to generate a first output image based on the first input image, and a second output image based on the second input image respectively; the first input image includes the first-resolution sample image and a first noise image corresponding to a noise sample with a first amplitude, and the second input image includes the first-resolution sample image and a second noise image corresponding to a noise sample with a second amplitude; and the first amplitude is greater than 0, and the second amplitude is equal to 0;
separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a first discrimination result based on the first output image, and a second discrimination result based on the second-resolution sample image; and
adjusting parameters of the generative network to reduce a loss function of the generative network; the loss function of the generative network includes a first loss, a second loss and a third loss, the first loss of the loss function is based on a reconstruction error between the second output image and the second-resolution sample image; the second loss of the loss function is based on a perceptual error between the first output image and the second-resolution sample image; and the third loss of the loss function is based on the first discrimination result and the second discrimination result.
In some implementations, the reconstruction error between the second output image and the second-resolution sample image is determined according to any one of L1 norm of a difference image matrix between the second output image and the second-resolution sample image, a mean square error between the second output image and the second-resolution sample image, and a structural similarity index between the second output image and the second-resolution sample image.
In some implementations, both the first output image and the second output image are generated by the generative network through an iteration process of a resolution enhancement procedure; and the first loss of the loss function of the generative network is λ1Lrec(X, Yn=0), and
L rec ( X , Y n = 0 ) = E [ l = 1 L Y n = 0 l - HR l 1 ] + E [ l = 1 L D bic l ( Y n = 0 l ) - LR 1 ]
where, X denotes the second-resolution sample image;
Yn=0 denotes the second output image;
Lrec(X, Yn=0) denotes the reconstruction error between the second output image and the second-resolution sample image;
L denotes a total number of times of the resolution enhancement procedure in the iteration process; and L≥1
Yn=0 l denotes an image generated at the end of the lth time of the resolution enhancement procedure in an iteration process performed by the generative network based on the second input image; and l≤L;
LR denotes the first-resolution sample image;
Dbic l (Yn=0 l) denotes an image obtained by downsampling Yn=0 l, a resolution of the image being the same as that of the first-resolution sample image;
HRl denotes an image obtained by downsampling the second-resolution sample image, a resolution of the image being the same as that of Yn=0 l;
E[ ] denotes calculation of matrix energy; and
λ1 is a preset weight.
In some implementations, the second loss of the loss function of the generative network is λ2Lper(X, Yn=1), and
L per ( X , Y n = 1 ) = E [ l = 1 L L CX ( Y n = 1 l , HR l ) ] + E [ l = 1 L L CX ( D bic l ( Y n = 1 l ) , LR ) ]
where, Yn=1 denotes the first output image;
Lper (X, Yn=1) denotes the perceptual error between the first output image and the second-resolution sample image;
Yn=1 l denotes an image generated at the end of the lth time of the resolution enhancement procedure in an iteration process performed by the generative network based on the first input image;
Dbic l(Yn=1 l) denotes an image obtained by downsampling Yn=1 l, a resolution of the image being the same as that of the first-resolution sample image;
LCX( ) is a contextual loss calculation function; and
λ2 is a preset weight.
In some implementations, the third loss of the loss function of the generative network is λ3LGAN(Yn=1), and
L GAN(Y n=1)=E[log(1−D(Y n=1 1, 2, . . . L))]+E[log(D(HR 1, 2, . . . L))]
where, Yn=1 1, 2 . . . L denotes an image group generated in the iteration process performed by the generative network based on the first input image, the image group including images each generated at the end of each time of the resolution enhancement procedure;
HR1, 2, . . . L denotes images obtained by downsampling the second-resolution sample image, the images being in one-to-one correspondence with the images in Yn=1 1, 2, . . . L, and each having a resolution the same as that of the corresponding image;
D(Yn=1 1, 2, . . . L) denotes the first discrimination result;
D(HR1, 2, . . . L denotes the second discrimination result; and
λ3 is a preset weight.
In some implementations, λ1: λ2: λ3=10:0.1:0.001
In some implementations, the noise sample is random noise. In some implementations, the training method further includes a discriminative network training procedure, which includes: separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a discrimination result based on the first output image and a discrimination result based on the second-resolution sample image, respectively; and adjusting parameters of the discriminative network to reduce a loss function of the discriminative network;
the discriminative network training procedure and the generative network training procedure are alternately performed until a preset training condition is met.
In some implementations, both the first output image and the second output image are generated by the generative network through an iteration process of the resolution enhancement procedure, and a total number of times of the resolution enhancement procedure in the iteration process is L; and when L is greater than 1, in the previous L−1 times of the resolution enhancement procedure in the iteration process performed by the generative network based on the first input image, the generative network generates an intermediate image each time the resolution enhancement procedure is performed;
in the discriminative network training procedure, each intermediate image generated by the generative network based on the first input image is provided for the discriminative network, while the first output image is being provided for the discriminative network; and while the second-resolution sample image is being provided for the discriminative network, and third-resolution sample images obtained by downsampling the second-resolution sample image are provided for the discriminative network, the third-resolution sample images being in one-to-one correspondence with intermediate images, and each having a resolution the same as that of the corresponding intermediate image
Accordingly, the present disclosure further provides an image processing method using the generative network of the generative adversarial network obtained by the training method, and the image processing method is used for increasing a resolution of an image, and includes:
providing an input image and a noise image corresponding to reference noise for the generative network to allow the generative network to generate a second-resolution image based on the input image.
In some implementations, an amplitude of the reference noise ranges from 0 to the first amplitude.
In some implementations, the reference noise is random noise.
Accordingly, the present disclosure further provides a computer device including a memory having computer programs stored thereon, and a processor, and the above training method is performed when the computer programs are executed by the processor.
Accordingly, the present disclosure further provides a computer-readable storage medium having computer programs stored thereon, and the above training method is performed when the computer programs are executed by a processor.
BRIEF DESCRIPTION OF THE DRAWINGS
The accompanying drawings are intended to provide a further understanding of the present disclosure, and are incorporated in and constitute a part of the specification. The drawings, together with the specific embodiments below, are intended to explain the present disclosure, but do not make any limitation to the present disclosure. In the drawings:
FIG. 1 is a schematic diagram illustrating a relationship between reconstruction distortion and perceptual distortion;
FIG. 2 is a flowchart illustrating a generative network training procedure according to the embodiments of the present disclosure; and
FIG. 3 is a schematic structural diagram of a generative network according to the embodiments of the present disclosure.
DETAILED DESCRIPTION
The specific embodiments of the present disclosure are described in detail below with reference to the accompanying drawings. It should be understood that the specific embodiments described herein are merely for illustrating and explaining the present disclosure, but do not make any limitation to the present disclosure.
Super-resolution image reconstruction is a technology for increasing a resolution of an initial image to obtain an image with a higher resolution. In the super-resolution image reconstruction, reconstruction distortion and perceptual distortion are used for evaluating a super-resolution reconstruction effect. The reconstruction distortion is used for measuring a difference between a reconstructed image and a reference image, and specific evaluation criteria include mean square error (MSE), structural similarity index (SSIM), and peak signal-to-noise ratio (PSNR); and the perceptual distortion mainly focuses on making the image more look like a natural image. FIG. 1 is a schematic diagram illustrating a relationship between the reconstruction distortion and the perceptual distortion. As shown in FIG. 1, when the reconstruction distortion is relatively small, the perceptual distortion is relatively large, in which case the reconstructed image looks smoother but lacks details. When the perceptual distortion is relatively small, the reconstruction distortion is relatively large, in which case the reconstructed image has more details. The current super-resolution image reconstruction methods usually aim at relatively small reconstruction distortion, but people prefer to obtaining reconstructed images with rich details in some application scenarios.
The present disclosure provides a training method for generative adversarial network, the generative adversarial network includes a generative network and a discriminative network, the generative network is configured to convert a first-resolution image into a second-resolution image to obtain the second-resolution image having a target resolution, and the resolution of the second-resolution image is higher than that of the first-resolution image. The generative network can obtain the second-resolution image by performing a resolution enhancement procedure once or iterating a resolution enhancement procedure for a plurality of times. For example, an image to be processed (i.e., a first-resolution image) has a resolution of 128×128 and the target resolution is 1024×1024, the generative network may obtain the second-resolution image having a resolution of 1024×1024 by performing once the resolution enhancement procedure which increases a resolution by 8 times; or the generative network may obtain a 256×256 image, a 512×512 image and a 1024×1024 image in sequence by iterating the resolution enhancement procedure, which increases a resolution by 2 times, three times.
The training method for generative adversarial network includes a generative network training procedure. FIG. 2 is a flowchart illustrating the generative network training procedure according to the embodiments of the present disclosure. As shown in FIG. 2, the generative network training procedure includes following S1 through S4.
S1, extracting a first-resolution sample image from a second-resolution sample image, with a resolution of the second-resolution sample image higher than that of the first-resolution sample image. Specifically, the first-resolution sample image may be obtained by downsampling the second-resolution sample image.
S2, separately providing a first input image and a second input image for the generative network to generate a first output image based on the first input image, and a second output image based on the second input image respectively, with the first input image including the first-resolution sample image and a first noise image corresponding to a noise sample with a first amplitude, and the second input image including the first-resolution sample image and a second noise image corresponding to a noise sample with a second amplitude. The first amplitude is greater than 0, and the second amplitude is equal to 0.
The amplitude of the noise sample is an average fluctuation amplitude of the noise sample. For example, the noise sample is random noise, a mean of an image corresponding to the noise sample is μ, and a variance of the image corresponding to the noise sample is σ, that is, most pixel values of the image corresponding to the noise sample fluctuate from μ-σ to μ+σ, in which case a noise amplitude is μ. It should be understood that any image is shown in the form of matrix in an image processing process, and the pixel values represent element values of an image matrix. When the amplitude of the noise sample is 0, since no element value of the image matrix is less than 0, each element value of the image matrix may be considered to be 0.
In addition, it should be noted that the training method for generative adversarial network includes a plurality of generative network training procedures; and in a single generative network training procedure, the first-resolution sample image is the single one, and model parameters of the generative network when receiving the first input image and the second input image are the same.
S3, separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a first discrimination result based on the first output image, and a second discrimination result based on the second-resolution sample image. The first discrimination result is used for representing a matching degree between the first output image and the second-resolution sample image, for example, the first discrimination result is used for representing a probability determined by the discriminative network that the first output image is identical to the second-resolution sample image; and the second discrimination result is used for representing a probability determined by the discriminative network that the second-resolution sample image is indeed the second-resolution sample image.
The discriminative network may be regarded as a classifier having a scoring function. The discriminative network can score a received to-be-discriminated image, and output a score which indicates a probability that the to-be-discriminated image (the first output image) is identical to the second-resolution sample image, that is, indicating the matching degree mentioned above, which may range from 0 to 1. When the output score of the discriminative network is 0 or close to 0, it is indicated that the discriminative network classifies the received to-be-discriminated image as a non-high-resolution sample image; and when the output score of the discriminative network is 1 or close to 1, it is indicated that the received to-be-discriminated image is identical to the second-resolution sample image.
The scoring function of the discriminative network may be trained by use of a “true” sample and a “false” sample with predetermined scores. For example, the “false” sample is an image generated by the generative network and the “true” sample is the second-resolution sample image. A training process of the discriminative network is a process of adjusting parameters of the discriminative network to enable the discriminative network to output a score close to 1 when receiving the “true” sample, and output a score close to 0 when receiving the “false” sample.
S4, adjusting parameters of the generative network to reduce a loss function of the generative network. “To reduce a loss function of the generative network” means that a value of the loss function is reduced as compared to that in previous generative network training procedure, or that values of the loss function in a plurality of generative network training procedures show decreasing trend. The loss function of the generative network includes a first loss, a second loss, and a third loss; specifically, the loss function is superposition of the first loss, the second loss and the third loss, and the first loss is based on a reconstruction error between the second output image and the second-resolution sample image; the second loss is based on a perceptual error between the first output image and the second-resolution sample image; and the third loss is based on the first discrimination result and the second discrimination result.
In super-resolution reconstruction, detail features (e.g. hairs, lines, etc.) in a reconstructed second-resolution image are usually related to noise. When no noise is added in the training of the generative network, the second-resolution image generated by the generative network has small reconstruction distortion and large perceptual distortion, so that the second-resolution image looks unreal by the naked eye; when noise is added in the training of the generative network, the reconstructed second-resolution image has obvious detail features, but the reconstruction distortion thereof is relatively large. In the generative network training procedure of the present disclosure, the second input image including a noise image with an amplitude of 0 and the first input image including a noise image with an amplitude of 1 are separately provided for the generative network for training, and the first loss of the loss function reflects the reconstruction distortion of a result generated by the generative network, and the second loss reflects the perceptual distortion of the result generated by the generative network, that is, the loss function combines two distortion evaluation criteria. When a trained generative network is used to perform resolution enhancement on an image, an amplitude of input noise can be adjusted according to actual needs (i.e., whether details of the image need to be emphasized and to what extent the details are emphasized), so that a reconstructed image can meet the actual needs. For example, within a given range of reconstruction distortion, minimum perceptual distortion is achieved by adjusting the amplitude of the input noise; or within a given range of perceptual distortion, minimum reconstruction distortion is achieved by adjusting the amplitude of the input noise.
It should be noted that the amplitude of the noise image of the first input image, which is 1 in the embodiment, is an amplitude value obtained by normalizing the amplitude of the noise image. In other embodiments of the present disclosure, it is possible not to normalize the amplitude of the noise image, so that the amplitude of the noise image of the first input image may be not equal to 1.
In some implementations, the noise sample is random noise; and a mean of the first noise image is 1. In some implementations, a mean of the first noise image is a mean of a normalized image of the first noise image. For example, if the first noise image is a grayscale image, an average of all pixel values in an image obtained by normalizing the first noise image is the mean of the first noise image; as another example, if the first noise image is a color image, an average of all pixel values in an image obtained by normalizing every channel of the first noise image is the mean of the first noise image. It should be noted that the channel of the image in the embodiment of the present disclosure indicates one or more channels obtained by dividing an image for processing, for example, an RGB-mode color image may be divided into three channels, i.e., a red channel, a green channel, and a blue channel; if the image is a grayscale image, it is a one-channel image; and if the color image is divided according to an HSV color system, the image may be divided into three channels, i.e., a hue (H) channel, a saturation (S) channel, and a value (V) channel.
In some implementations, the loss function of the generative network is represented in a formula as follows:
Loss=λL rec(X, Y n=0)+λ2 L per(X, Y n=1)+λ3 L GAN(Y n=1)
where, in the first loss λ1Lrec(X, Yn=0) of the loss function Loss, Lrec(X, Yn=0) denotes the reconstruction error between the second output image and the second-resolution sample image; in the second loss λ2Lper (X, Yn=1) of the loss function Loss, Lper(X, Yn=1) denotes the perceptual error between the second output image and the second-resolution sample image; in the third loss λ3LGAN (Yn=1) of the loss function Loss, LGAN(Yn=1) denotes a sum of the first discrimination result and the second discrimination result; and λ1, λ2, λ3 are all preset weights. λ1: λ2: λ3 may be adjusted according to actual needs, for example, λ1: λ2: λ3=10:0.1:0.001, or λ1: λ2: λ3=1:1:0.5. In some embodiments, λ1: λ2: λ3 may be set according to continuity of local images. While in some other embodiments, λ1: λ2: λ3 may be set according to target pixels of an image.
Specifically, the reconstruction error Lrec(X, Yn=0) between the second output image Yn=0 and the second-resolution sample image X is calculated according to the following formula:
L rec ( X , Y n = 0 ) = E [ l = 1 L Y n = 0 l - HR l 1 ] + E [ l = 1 L D bic l ( Y n = 0 l ) - LR 1 ]
where both the first output image and the second output image are generated by the generative network through an iteration process of a resolution enhancement procedure; and a total number of times of the resolution enhancement procedure in the iteration process is L, and L≥1.
Yn=0 l denotes an image generated at the end of the lth time of the resolution enhancement procedure in an iteration process performed by the generative network based on the second input image; and l≤L. It should be understood that the generative network generates the second output image Yn=0 when l=L.
LR denotes the first-resolution sample image; and Dbic l(Yn=0 l) denotes an image obtained by downsampling Yn=0 l, a resolution of the image being the same as that of the first-resolution sample image. The downsampling may be performed in a way the same as that for extracting the first-resolution sample image from the second-resolution sample image in the step S1.
HRl denotes an image obtained by downsampling the second-resolution sample image, a resolution of the image being the same as that of Yn=0 l. It should be noted that, when l=L, Yn=0 l is the second output image Yn=0, and HRl is the second-resolution sample image itself, or may be considered as an image obtained by downsampling the second-resolution sample image by 1 time.
E[ ] denotes calculation of matrix energy. For example, E[ ] can calculate a maximum or average of the elements in a matrix in “[ ]”.
In the case of iterating a resolution enhancement procedure by the generative network for a plurality of times, when calculating the reconstruction error, L1 norm of a difference image matrix between the second output image itself and the second-resolution sample image is calculated, L1 norm of a difference image matrix between a third-resolution image (i.e., Yn=0 1, Yn=0 2, . . . Yn=0 L−l) generated by the generative network and a third-resolution sample image (i.e., HR1, HR2, . . . , HRL−1) is also calculated, a resolution of the third-resolution sample image being the same as that of the third-resolution image. At the same time, L1 norm of a difference image between the third-resolution image, the image generated by downsampling the second output image, and the first-resolution sample image is also calculated. In this way, when the generative network is used for resolution enhancement and noise with an amplitude of 0 is input, an image finally output by the generative network can achieve minimum reconstruction distortion. It should be noted that the resolution of the third-resolution image is higher than that of the first-resolution sample image, and is the same as that of the third-resolution sample image.
In the above embodiment, the reconstruction error Lrec(X, Yn=0) between the second output image and the second-resolution sample image is obtained based on the L1 norm of the difference image matrix between the second output image and the second-resolution sample image, or the reconstruction error may be obtained based on a mean square error (MSE) between the second output image and the second-resolution sample image, or may be obtained based on a structural similarity index (SSIM) between the second output image and the second-resolution sample image.
In some implementations, the perceptual error Lper(X, Yn=1) between the first output image Yn=1 and the second-resolution sample image X is calculated according to the following formula:
L per ( X , Y n = 1 ) = E [ l = 1 L L CX ( Y n = 1 l , HR l ) ] + E [ l = 1 L L CX ( D bic l ( Y n = 1 l ) , LR ) ]
Yn=1 l denotes an image generated at the end of the lth time of the resolution enhancement procedure in an iteration process performed by the generative network based on the first input image; and l≤L. It should be understood that the generative network generates the first output image Yn=1 when l=L.
Dbic l(Yn=1 l) denotes an image obtained by downsampling Yn=1 l, a resolution of the image being the same as that of the first-resolution sample image LR. The downsampling may be performed in a way the same as that for extracting the first-resolution sample image from the second-resolution sample image in the step S1. Reference may be made to the above description for the meanings of HRl and E[ ], which will not be repeated herein.
LCX( ) denotes a contextual loss calculation function.
Similar to the calculation of the reconstruction error, calculation of the perceptual error not only involves calculation of a different between the first output image and the second-resolution sample image by use of the contextual loss calculation function, but also involves calculation of a difference between a third-resolution image (i.e., Yn= 1
Figure US11449751-20220920-P00001
Yn=1 2
Figure US11449751-20220920-P00001
. . . Yn=1 L−1 generated by the generative network based on the first input image and a third-resolution sample image (i.e., HR1
Figure US11449751-20220920-P00001
HR2
Figure US11449751-20220920-P00001
. . . HRL−1), a resolution of the third-resolution sample image being the same as that of the third-resolution image, and further involves calculation of a difference between the third-resolution image, the image generated by downsampling the second output image, and the first-resolution sample image. In this way, when the generative network is used for resolution enhancement and noise with the first amplitude is input, an image finally output by the generative network can achieve minimum perceptual distortion.
In some implementations, LGAN(Yn=1) in the third loss of the loss function of the generative network is calculated according to the following formula:
L GAN(Y n=1)=E[log(1−D(Y n=1 1, 2, . . . L))]E[log(D(HR 1, 2, . . . L))]
where, Yn=1 1, 2, . . . L denotes an image group generated in an iteration process performed by the generative network based on the first input image, the image group including images each generated at the end of each time of the resolution enhancement procedure. When L=1, the image group includes the first output image alone; and when L>1, the image group includes Yn=1 1 to Yn=1 L−1, and the first output image Yn=1.
HR1, 2, . . . L denotes images obtained by downsampling the second-resolution sample image, the images being in one-to-one correspondence with the images in Yn=1 1, 2, . . . L and each having a resolution the same as that of the corresponding image. HRL is the second-resolution sample image itself.
D(Yn=1 1, 2, . . . L) denotes a discrimination result generated by the discriminative network based on Yn=1 1, 2, . . . L, that is, the first discrimination result; and D(HR1, 2, . . . L) denotes a discrimination result generated by the discriminative network based on HR1, 2, . . . L that is, the second discrimination result.
In the training method of the present disclosure, in addition to the generative network training procedure, the training method further includes a discriminative network training procedure, which includes: separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a discrimination result based on the first output image and a discrimination result based on the second-resolution sample image, respectively; and adjusting the parameters of the discriminative network to reduce a loss function of the discriminative network.
The discriminative network training procedure and the generative network training procedure are alternately performed until a preset training condition is met. For example, the preset training condition may be that the number of alternation times reaches a predetermined value.
In an initialization process, the parameters of the generative network and the discriminative network are preset or random.
As described above, both the first output image and the second output image are generated by the generative network through an iteration process of the resolution enhancement procedure, and a total number of times of the resolution enhancement procedure in the iteration process is L. When L=1, it is possible to supply the first output image or the second-resolution sample image alone to the discriminative network each time an image is supplied to the discriminative network. When L>1, in the previous L−1 times of the resolution enhancement procedure performed by the generative network based on the first input image, the generative network generates an intermediate image each time the resolution enhancement procedure is performed; and when the resolution enhancement procedure is iterated for the Lth time, the image generated by the generative network is the first output image. In such case, the discriminative network is provided with a plurality of input terminals to receive a plurality of images simultaneously, and determines a matching degree between one of the received plurality of images, which has a highest resolution, and the second-resolution sample image. In the discriminative network training procedure, each intermediate image generated by the generative network based on the first input image is provided for the discriminative network, while the first output image is being provided for the discriminative network; and while the second-resolution sample image is being provided for the discriminative network, third-resolution sample images obtained by downsampling the second-resolution sample image are provided for the discriminative network, the third-resolution sample images being in one-to-one correspondence with intermediate images, and each having a resolution the same as that of the corresponding intermediate image.
In the training process of the generative network, the parameters of the generative network are adjusted to enable the discriminative network to output a matching degree as close to 1 as possible as a discrimination result after an output result of the generative network is input into the discriminative network, that is, to enable the discriminative network to regard the output result of the generative network as the second-resolution sample image. In the training process of the discriminative network, the parameters of the discriminative network are adjusted to enable the discriminative network to output a matching degree as close to 1 as possible after the second-resolution sample image is input into the discriminative network, and also enable the discriminative network to output a matching degree as close to 0 as possible after an output result of the generative network is input into the discriminative network; that is, the discriminative network can be trained to be capable of determining whether a received image is the second-resolution sample image. By alternately training the generative network and the discriminative network, the discriminative network is continuously optimized to improve discrimination capability, and the generative network is continuously optimized to output a result as close to the second-resolution sample image as possible. With this method, the two “opposing” models compete with each other and each is improved based on an increasingly better result from the other one in each training process, so that the generative adversarial network model obtained are getting better and better.
The present disclosure further provides an image processing method using a generative adversarial network obtained by the above training method, and the image processing method is used for increasing a resolution of an image by using a generative network of the generative adversarial network, and includes providing an input image and a noise image corresponding to reference noise for the generative network to allow the generative network to generate an image having a higher resolution than the input image. An amplitude of the reference noise ranges from 0 to a first amplitude. Specifically, the reference noise is random noise.
In the training process of the generative network of the generative adversarial network according to the present disclosure, the noise sample with an amplitude of 0 and the noise sample with a first amplitude are separately provided for the generative network, and the loss function of the generative network combines two distortion evaluation criteria for evaluating the reconstruction distortion and the perceptual distortion, so that the amplitude of reference noise can be adjusted according to the actual needs when the generative network is used to increase a resolution of an image, so as to meet the actual needs. For example, within a given range of reconstruction distortion, minimum perceptual distortion is achieved by adjusting the amplitude of the reference noise; or within a given range of perceptual distortion, minimum reconstruction distortion is achieved by adjusting the amplitude of the reference noise.
FIG. 3 is a schematic structural diagram of a generative network according to the embodiments of the present disclosure. The generative network is described below in conjunction with FIG. 3. The generative network is used for iterating a resolution enhancement procedure, and a resolution of a to-be-processed image Il−1 is increased each time the resolution enhancement procedure is performed, so as to obtain an image Il with an increased resolution. When a total number of times of the iteration of the resolution enhancement procedure is 1, the to-be-processed image Il−1 is an initial input image; when the total number of times of the iteration of the resolution enhancement procedure is L and L>1, the to-be-processed image Il−1 is an image output after iterating the resolution enhancement procedure for the (l−1)th time. The generative network is illustrated below by an example as follows: the initial input image has a resolution of 128×128, the resolution is increased by 2 times in each resolution enhancement procedure, and l=2. In this example, the to-be-processed image Il−1 in FIG. 3 is a 256×256 image obtained after performing the resolution enhancement procedure once.
As shown in FIG. 3, the generative network includes a first analysis module 11, a second analysis module 12, a first concatenating module 21, a second concatenating module 22, an interpolation module 31, a first upsampling module 41, a first downsampling module 51, a superposition module 70, and a residual correction system for iteration.
The first analysis module 11 is configured to generate a feature image Rl−1 μ of the to-be-processed image Il−1, and a number of channels of the feature image Rl−1 μ is greater than that of the to-be-processed image Il−1.
The first concatenating module 21 is configured to concatenate the feature image Rl−1 μ of the to-be-processed image and a noise image to obtain a first merged image RCl−1 μ; and a number of channels of the first merged image RCl−1 μ is a sum of the number of the channels of the feature image Rl−1 μ and a number of channels of the noise image.
It should be noted that a resolution of the noise image is the same as that of the to-be-processed image Il−1. Therefore, when the total number of times of the iteration of the resolution enhancement procedure performed by the generative network is greater than 1, in the generative network training procedure, each of the first input image and the second input image provided for the generative network may include the first-resolution sample image and a plurality of noise sample images having different resolutions; or each of the first input image and the second input image may include the first-resolution sample image and one noise sample image, and when the resolution enhancement procedure is iterated for the lth time, the generative network generates the noise sample image at a required magnification according to an amplitude of a noise sample.
The interpolation module 31 is configured to perform interpolation on the to-be-processed image Il−1 to obtain a fourth-resolution image based thereon, the fourth-resolution image having a resolution of 512×512. The interpolation module may perform the interpolation by using traditional interpolation methods, such as bicubic interpolation. The resolution of the fourth-resolution image is higher than that of the to-be-processed image Il−1.
The second analysis module 12 is configured to generate a feature image of the fourth-resolution image, a number of channels of the feature image being greater than that of the fourth-resolution image.
The first downsampling module 51 is configured to downsample the feature image of the fourth-resolution image to obtain a first downsampled feature image having a resolution of 256×256.
The second concatenating module 22 is configured to concatenate the first merged image RCl−1 μ and the first downsampled feature image to obtain a second merged image.
The first upsampling module 41 is configured to upsample the second merged image to obtain a first upsampled feature image Rl 0.
The residual correction system for iteration is configured to perform residual correction on the first upsampled feature image through back-projection for at least one time, so as to obtain a residual-corrected feature image.
The residual correction system for iteration includes a second downsampling module 52, a second upsampling module 42, and a residual determination module 60. The second downsampling module 52 is configured to downsample a received image by 2 times, and the second upsampling module 42 is configured to upsample a received image by 2 times; and the residual determination module 60 is configured to determine a difference image between two received images.
In first residual correction, the first upsampled feature image Rl 0 is downsampled by 2 times by the first one second downsampling module 52 to obtain a feature image Rl 01; the feature image Rl 01 is downsampled by 2 times by the second one second downsampling module 52 to obtain a feature image Rl 02 having a same resolution as the initial input image; then, one residual determination module is used to obtain a difference image between the feature image Rl 02 and the first merged image RCμ 0 obtained in the first time of the resolution enhancement procedure (i.e., the first merged image RCμ 0 obtained by merging the feature image of the initial input image and a noise image); then, the difference image is upsampled by the second upsampling module to obtain a feature image, and the obtained upsampled feature image is superposed on the feature image Rl 01 by the superposition module 70, so as to obtain a feature image R03 l having a same resolution as a first merged image Rl l−1; then, another residual determination module is used to obtain a difference image between the feature image R03 l and the first merged image RCμ l−1; then, the difference image is upsampled by 2 times by the second upsampling module 42 to obtain an upsampled image, and the upsampled image is superposed on the first upsampled feature image Rl 0, so as to obtain a feature image Rl 1 subjected to the first residual correction.
Then, the feature image Rl 1 may be subjected to the second residual correction in the same way to obtain a feature image Rl 2 subjected to the second residual correction; and the feature image Rl 2 may be subjected to the third residual correction in the same way, and so on. In FIG. 3, μ represents a number of times of the residual correction.
The generative network further includes a synthesis module 80 configured to synthesize a feature image Rl μ obtained after a plurality of times of residual correction, so as to obtain a fifth-resolution image, a number of channels of the fifth-resolution image being the same as that of the fourth-resolution image; and the fifth-resolution image and the fourth-resolution image are superposed to obtain an output image Il after the resolution enhancement procedure is performed for the lth time. A resolution of the fifth-resolution image is the same as that of the fourth-resolution image.
In the generative network, the first analysis module 11, the second analysis module 12, the first upsampling module 41, the second upsampling module 42, the first downsampling module 51, the second downsampling module 52, and the synthesis module 80 can perform corresponding functions through a convolutional layer.
The second time of the resolution enhancement procedure in an iteration process is illustrated above by an example of l=2; and other times of the resolution enhancement procedure are similar to the second time, and thus will not be described in detail herein.
The present disclosure further provides a computer device including a memory having computer programs stored thereon, and a processor, and the above training method for generative adversarial network is performed when the computer programs are executed by the processor.
The present disclosure further provides a computer-readable storage medium having computer programs stored thereon, and the above training method for generative adversarial network is performed when the computer programs are executed by a processor.
The above memory and computer-readable storage medium include, but are not limited to, the following readable media: random access memories (RAMs), read-only memories (ROMs), non-volatile random access memories (NVRAMs), programmable read-only memories (PROMs), erasable programmable read-only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), flash memories, magnetic or optical data memories, registers, magnetic disks or tapes, optical storage media such as compact discs (CDs) or digital versatile discs (DVDs), and other non-transitory media. Examples of the processor include, but are not limited to, a general-purpose processor, a central processing unit (CPU), a microprocessor, a digital signal processor (DSP), a controller, a microcontroller, a state machine, etc.
It should be understood that the above embodiments are merely exemplary embodiments for illustrating the principles of the present disclosure, and the present disclosure is not limited thereto. Without departing from the spirit and essence of the present disclosure, various changes and modifications can be made by those skilled in the art, and should be considered to fall within the scope of the present disclosure.

Claims (14)

What is claimed is:
1. A training method for a generative adversarial network, the generative adversarial network comprising a generative network and a discriminative network, the generative network being configured to convert a first-resolution image into a second-resolution image, a resolution of the second-resolution image being higher than that of the first-resolution image, and the training method comprising a generative network training procedure, which comprises:
extracting a first-resolution sample image from a second-resolution sample image, with a resolution of the second-resolution sample image higher than that of the first-resolution sample image;
separately providing a first input image and a second input image for the generative network to generate a first output image based on the first input image, and a second output image based on the second input image respectively; wherein, the first input image comprises the first-resolution sample image and a first noise image corresponding to a noise sample with a first amplitude; the second input image comprises the first-resolution sample image and a second noise image corresponding to a noise sample with a second amplitude; and the first amplitude is greater than 0, and the second amplitude is equal to 0;
separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a first discrimination result based on the first output image, and a second discrimination result based on the second-resolution sample image; and
adjusting parameters of the generative network to reduce a loss function of the generative network, wherein the loss function of the generative network comprises a first loss, a second loss and a third loss, the first loss is based on a reconstruction error between the second output image and the second-resolution sample image; the second loss is based on a perceptual error between the first output image and the second-resolution sample image; and the third loss is based on the first discrimination result and the second discrimination result.
2. The training method of claim 1, wherein the reconstruction error between the second output image and the second-resolution sample image is determined according to any one of L1 norm of a difference image matrix between the second output image and the second-resolution sample image, a mean square error between the second output image and the second-resolution sample image, and a structural similarity index between the second output image and the second-resolution sample image.
3. The training method of claim 1, wherein both the first output image and the second output image are generated by the generative network through an iteration process of a resolution enhancement procedure, and the first loss of the loss function of the generative network is λ1Lrec(X, Yn=0), and
L rec ( X , Y n = 0 ) = E [ l = 1 L Y n = 0 l - HR l 1 ] + E [ l = 1 L D bio l ( Y n = 0 l ) - LR 1 ]
where X denotes the second-resolution sample image;
Yn=0 denotes the second output image;
Lrec(X, Yn=0) denotes the reconstruction error between the second output image and the second-resolution sample image;
L denotes a total number of times of the resolution enhancement procedure in the iteration process, and L≥1;
Yn=0 l denotes an image generated at the end of the lth time of the resolution enhancement procedure in the iteration process performed by the generative network based on the second input image, and l≤L;
LR denotes the first-resolution sample image;
Dbic l(Yn=0 l) denotes an image obtained by downsampling Yn=0 l, a resolution of the image being the same as that of the first-resolution sample image;
HRl denotes an image obtained by downsampling the second-resolution sample image, a resolution of the image being the same as that of Yn=0 l;
E[ ] denotes calculation of matrix energy; and
λ1 is a preset weight.
4. The training method of claim 3, wherein the second loss of the loss function of the generative network is λ2Lper(X, Yn=1), and
L per ( X , Y n = 1 ) = E [ l = 1 L L CX ( Y n = 1 l , HR l ) ] + E [ l = 1 L L CX ( D bic l ( Y n = 1 l ) , LR ) ]
where, Yn=1 denotes the first output image;
Lper(X, Yn=1) denotes the perceptual error between the first output image and the second-resolution sample image;
Yn=1 l denotes an image generated at the end of the lth time of the resolution enhancement procedure in the iteration process performed by the generative network based on the first input image;
Dbic l(Yn=1 l) denotes an image obtained by downsampling Yn=1 l, a resolution of the image being the same as that of the first-resolution sample image;
LCX( ) is a contextual loss calculation function; and
λ2 is a preset weight.
5. The training method of claim 4, wherein the third loss of the loss function of the generative network is λ3LGAN(Yn=1), and

L GAN(Y n=1)=E[log(1−D(Y n=1 1, 2, . . . L))]+E[log(D(HR 1, 2, . . . L))]
where, Yn=1 1, 2, . . . L denotes an image group generated in the iteration process performed by the generative network based on the first input image, the image group comprising images each generated at the end of each time of the resolution enhancement procedure;
HR1, 2, . . . L denotes images obtained by downsampling the second-resolution sample image, the images being in one-to-one correspondence with the images in Yn=1 1, 2, . . . L, and each having a resolution the same as that of the corresponding image;
D(Yn=1 1, 2, . . . L) denotes the first discrimination result; D(HR1, 2, . . . L) denotes the second discrimination result; and
λ3 is a preset weight.
6. The training method of claim 5, wherein λ123=10:0.1:0.001.
7. The training method of claim 1, wherein the noise sample is random noise.
8. The training method of claim 1, further comprising a discriminative network training procedure, which comprises: separately providing the first output image and the second-resolution sample image for the discriminative network to allow the discriminative network to output a discrimination result based on the first output image and a discrimination result based on the second-resolution sample image, respectively; and adjusting parameters of the discriminative network to reduce a loss function of the discriminative network; and
the discriminative network training procedure and the generative network training procedure are alternately performed until a preset training condition is met.
9. The training method of claim 8, wherein,
both the first output image and the second output image are generated by the generative network through an iteration process of a resolution enhancement procedure, and a total number of times of the resolution enhancement procedure in the iteration process is L; and when L is greater than 1, in previous L−1 times of the resolution enhancement procedure in the iteration process performed by the generative network based on the first input image, the generative network generates an intermediate image each time the resolution enhancement procedure is performed; and
in the discriminative network training procedure, each intermediate image generated by the generative network based on the first input image is provided for the discriminative network, while the first output image is being provided for the discriminative network; and while the second-resolution sample image is being provided for the discriminative network, third-resolution sample images obtained by downsampling the second-resolution sample image are provided for the discriminative network, the third-resolution sample images being in one-to-one correspondence with intermediate images, and each having a resolution the same as that of the corresponding intermediate image.
10. An image processing method using the generative network of the generative adversarial network obtained by the training method of claim 1, wherein the image processing method is used for increasing a resolution of an image, and comprises:
providing an input image and a noise image corresponding to reference noise for the generative network to allow the generative network to generate a second-resolution image based on the input image.
11. The image processing method of claim 10, wherein an amplitude of the reference noise ranges from 0 to the first amplitude.
12. The image processing method of claim 10, wherein the reference noise is random noise.
13. A computer device, comprising a memory having computer programs stored thereon, and a processor, wherein the training method of claim 1 is performed when the computer programs are executed by the processor.
14. A non-transitory computer-readable storage medium having computer programs stored thereon, wherein the training method of claim 1 is performed when the computer programs are executed by a processor.
US16/759,669 2018-09-30 2019-09-25 Training method for generative adversarial network, image processing method, device and storage medium Active 2040-02-08 US11449751B2 (en)

Applications Claiming Priority (9)

Application Number Priority Date Filing Date Title
CN201811155252.6 2018-09-30
CN201811155930.9 2018-09-30
CN201811155326.6A CN109345455B (en) 2018-09-30 2018-09-30 Image authentication method, authenticator and computer-readable storage medium
CN201811155147.2 2018-09-30
CN201811155930.9A CN109345456B (en) 2018-09-30 2018-09-30 Generative confrontation network training method, image processing method, device and storage medium
CN201811155147.2A CN109360151B (en) 2018-09-30 2018-09-30 Image processing method and system, resolution improving method and readable storage medium
CN201811155252.6A CN109255390B (en) 2018-09-30 2018-09-30 Preprocessing method and module for training image, discriminator, and readable storage medium
CN201811155326.6 2018-09-30
PCT/CN2019/107761 WO2020063648A1 (en) 2018-09-30 2019-09-25 Training method, image processing method, device and storage medium for generative adversarial network

Publications (2)

Publication Number Publication Date
US20200285959A1 US20200285959A1 (en) 2020-09-10
US11449751B2 true US11449751B2 (en) 2022-09-20

Family

ID=69950197

Family Applications (4)

Application Number Title Priority Date Filing Date
US16/604,410 Active 2041-05-11 US11615505B2 (en) 2018-09-30 2019-04-23 Apparatus and method for image processing, and system for training neural network
US16/614,547 Active 2040-04-01 US11361222B2 (en) 2018-09-30 2019-06-20 System, method, and computer-readable medium for image classification
US16/614,558 Active 2039-12-30 US11348005B2 (en) 2018-09-30 2019-06-20 Apparatus, method, and computer-readable medium for image processing, and system for training a neural network
US16/759,669 Active 2040-02-08 US11449751B2 (en) 2018-09-30 2019-09-25 Training method for generative adversarial network, image processing method, device and storage medium

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US16/604,410 Active 2041-05-11 US11615505B2 (en) 2018-09-30 2019-04-23 Apparatus and method for image processing, and system for training neural network
US16/614,547 Active 2040-04-01 US11361222B2 (en) 2018-09-30 2019-06-20 System, method, and computer-readable medium for image classification
US16/614,558 Active 2039-12-30 US11348005B2 (en) 2018-09-30 2019-06-20 Apparatus, method, and computer-readable medium for image processing, and system for training a neural network

Country Status (9)

Country Link
US (4) US11615505B2 (en)
EP (4) EP3857447A4 (en)
JP (3) JP7415251B2 (en)
KR (2) KR102661434B1 (en)
AU (1) AU2019350918B2 (en)
BR (1) BR112020022560A2 (en)
MX (1) MX2020013580A (en)
RU (1) RU2762144C1 (en)
WO (4) WO2020062846A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220230276A1 (en) * 2019-05-23 2022-07-21 Deepmind Technologies Limited Generative Adversarial Networks with Temporal and Spatial Discriminators for Efficient Video Generation

Families Citing this family (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111986127B (en) * 2019-05-22 2022-03-08 腾讯科技(深圳)有限公司 Image processing method and device, computer equipment and storage medium
JP7312026B2 (en) * 2019-06-12 2023-07-20 キヤノン株式会社 Image processing device, image processing method and program
EP3788933B1 (en) * 2019-09-05 2025-12-24 BSH Hausgeräte GmbH Method for controlling a home appliance
CN120563345A (en) * 2020-01-23 2025-08-29 三星电子株式会社 Electronic device, method for controlling electronic device, and medium
US11507831B2 (en) * 2020-02-24 2022-11-22 Stmicroelectronics International N.V. Pooling unit for deep learning acceleration
WO2021220343A1 (en) * 2020-04-27 2021-11-04 日本電気株式会社 Data generation device, data generation method, learning device, and recording medium
CN111695605B (en) * 2020-05-20 2024-05-10 平安科技(深圳)有限公司 OCT image-based image recognition method, server and storage medium
EP4383184A3 (en) * 2020-07-08 2024-06-26 Sartorius Stedim Data Analytics AB Computer-implemented method, computer program product and system for processing images
JP7332810B2 (en) * 2020-07-09 2023-08-23 株式会社日立ハイテク pattern matching device, pattern measurement system, pattern matching program
US11887279B2 (en) * 2020-08-25 2024-01-30 Sharif University Of Technology Machine learning-based denoising of an image
US20220067519A1 (en) * 2020-08-28 2022-03-03 Affectiva, Inc. Neural network synthesis architecture using encoder-decoder models
US11455811B2 (en) * 2020-08-28 2022-09-27 Check it out Co., Ltd. System and method for verifying authenticity of an anti-counterfeiting element, and method for building a machine learning model used to verify authenticity of an anti-counterfeiting element
CN112132012B (en) * 2020-09-22 2022-04-26 中国科学院空天信息创新研究院 High-resolution SAR ship image generation method based on generation countermeasure network
JP7304484B2 (en) * 2020-11-09 2023-07-06 グーグル エルエルシー Portrait Relighting Using Infrared Light
US11893710B2 (en) * 2020-11-16 2024-02-06 Boe Technology Group Co., Ltd. Image reconstruction method, electronic device and computer-readable storage medium
CN112419200B (en) * 2020-12-04 2024-01-19 宁波舜宇仪器有限公司 Image quality optimization method and display method
US11895330B2 (en) * 2021-01-25 2024-02-06 Lemon Inc. Neural network-based video compression with bit allocation
CN113012064B (en) * 2021-03-10 2023-12-12 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN112884673B (en) * 2021-03-11 2025-02-11 西安建筑科技大学 Improved loss function SinGAN method for reconstructing missing information between tomb mural blocks
US12142016B2 (en) * 2021-06-17 2024-11-12 Nvidia Corporation Fused processing of a continuous mathematical operator
US12141941B2 (en) 2021-06-17 2024-11-12 Nvidia Corporation Generative neural networks with reduced aliasing
US12111919B2 (en) * 2021-08-23 2024-10-08 Fortinet, Inc. Systems and methods for quantifying file access risk exposure by an endpoint in a network environment
JP2023040928A (en) * 2021-09-10 2023-03-23 富士フイルム株式会社 Learning device, method of operating the same, and medical image processing terminal
JP2023041375A (en) * 2021-09-13 2023-03-24 キヤノン株式会社 Information processing device, information processing method and program
CN113962360B (en) * 2021-10-09 2024-04-05 西安交通大学 Sample data enhancement method and system based on GAN network
US12394024B2 (en) 2021-11-15 2025-08-19 Samsung Electronics Co., Ltd. System and method for training of noise model using noisy signal pairs
CN114169002B (en) * 2021-12-07 2025-04-04 杭州电子科技大学 A privacy protection method for face images driven by key point differential privacy
WO2023106723A1 (en) * 2021-12-08 2023-06-15 주식회사 딥엑스 Neural processing unit for image fusion, and artificial neural network system
KR102548283B1 (en) * 2021-12-22 2023-06-27 (주)뉴로컴즈 Convolutional neural network computing device
CN114331903B (en) * 2021-12-31 2023-05-12 电子科技大学 Image restoration method and storage medium
CN115063492B (en) * 2022-04-28 2023-08-08 宁波大学 Method for generating countermeasure sample for resisting JPEG compression
US12450495B2 (en) 2022-06-13 2025-10-21 International Business Machines Corporation Neural capacitance: neural network selection via edge dynamics
KR20240033619A (en) 2022-09-05 2024-03-12 삼성에스디에스 주식회사 Method and apparatus for extracting area of interest in documents
CN115393242A (en) * 2022-09-30 2022-11-25 国网电力空间技术有限公司 Method and device for enhancing foreign matter image data of power grid based on GAN
CN115631178B (en) * 2022-11-03 2023-11-10 昆山润石智能科技有限公司 Automatic wafer defect detection method, system, equipment and storage medium
CN115995021A (en) * 2022-12-31 2023-04-21 深圳云天励飞技术股份有限公司 Road disease identification method, device, electronic equipment and storage medium
CN116051382B (en) * 2023-03-02 2025-08-08 浙江工业大学 A data enhancement method based on deep reinforcement learning generative adversarial neural network and super-resolution reconstruction
US20250061697A1 (en) * 2023-08-18 2025-02-20 Mohamed bin Zayed University of Artificial Intelligence A train-time loss in a system and method for calibrating object detection
CN117196985A (en) * 2023-09-12 2023-12-08 军事科学院军事医学研究院军事兽医研究所 Visual rain and fog removing method based on deep reinforcement learning
US20250225400A1 (en) * 2024-01-05 2025-07-10 ProrataAI, Inc. Systems and methods for improving performance of a large language model by controlling training content

Citations (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5754697A (en) 1994-12-02 1998-05-19 Fu; Chi-Yung Selective document image data compression technique
US5781196A (en) 1990-10-19 1998-07-14 Eidos Plc Of The Boat House Video compression by extracting pixel changes exceeding thresholds
WO2002089046A1 (en) 2001-04-26 2002-11-07 Georgia Tech Research Corporation Video enhancement using multiple frame techniques
WO2003060823A2 (en) 2001-12-26 2003-07-24 Yeda Research And Development Co.Ltd. A system and method for increasing space or time resolution in video
US6766067B2 (en) 2001-04-20 2004-07-20 Mitsubishi Electric Research Laboratories, Inc. One-pass super-resolution images
CN101593269A (en) 2008-05-29 2009-12-02 汉王科技股份有限公司 Face identification device and method
CN101872472A (en) 2010-06-02 2010-10-27 中国科学院自动化研究所 A face image super-resolution reconstruction method based on sample learning
US20120328210A1 (en) 2010-01-28 2012-12-27 Yissum Research Development Company Of The Hebrew University Of Jerusalem Method and system for generating an output image of increased pixel resolution from an input image
CN102915527A (en) 2012-10-15 2013-02-06 中山大学 Face image super-resolution reconstruction method based on morphological component analysis
US20130301933A1 (en) 2012-05-10 2013-11-14 Thomson Licensing Method and device for generating a super-resolution version of a low resolution input data structure
CN103514580A (en) 2013-09-26 2014-01-15 香港应用科技研究院有限公司 Method and system for obtaining super-resolution images optimized for viewing experience
US8675999B1 (en) 2012-09-28 2014-03-18 Hong Kong Applied Science And Technology Research Institute Co., Ltd. Apparatus, system, and method for multi-patch based super-resolution from an image
CN103903236A (en) 2014-03-10 2014-07-02 北京信息科技大学 Method and device for reconstructing super-resolution facial image
CN104853059A (en) 2014-02-17 2015-08-19 台达电子工业股份有限公司 Super-resolution image processing method and device
US20150235345A1 (en) 2014-02-17 2015-08-20 Delta Electronics, Inc. Method and device for processing a super-resolution image
US20150296232A1 (en) 2012-11-27 2015-10-15 Lg Electronics Inc. Signal transceiving apparatus and signal transceiving method
CN105144232A (en) 2014-03-25 2015-12-09 展讯通信(上海)有限公司 Methods and systems for denoising images
CN105975931A (en) 2016-05-04 2016-09-28 浙江大学 Convolutional neural network face recognition method based on multi-scale pooling
CN105976318A (en) 2016-04-28 2016-09-28 北京工业大学 Image super-resolution reconstruction method
CN105975968A (en) 2016-05-06 2016-09-28 西安理工大学 Caffe architecture based deep learning license plate character recognition method
US20170178293A1 (en) 2014-02-13 2017-06-22 Thomson Licensing Method for performing super-resolution on single images and apparatus for performing super-resolution on single images
WO2017100903A1 (en) 2015-12-14 2017-06-22 Motion Metrics International Corp. Method and apparatus for identifying fragmented material portions within an image
US9727959B2 (en) 2011-09-28 2017-08-08 The United States Of America As Represented By The Secretary Of The Army System and processor implemented method for improved image quality and generating an image of a target illuminated by quantum particles
CN107133601A (en) 2017-05-13 2017-09-05 五邑大学 A kind of pedestrian's recognition methods again that network image super-resolution technique is resisted based on production
CN107154023A (en) 2017-05-17 2017-09-12 电子科技大学 Face super-resolution reconstruction method based on generation confrontation network and sub-pix convolution
RU2635883C1 (en) 2016-06-02 2017-11-16 Самсунг Электроникс Ко., Лтд. Image processing method and system for forming superhigh-resolution images
CN107369189A (en) 2017-07-21 2017-11-21 成都信息工程大学 The medical image super resolution ratio reconstruction method of feature based loss
US20170365038A1 (en) 2016-06-16 2017-12-21 Facebook, Inc. Producing Higher-Quality Samples Of Natural Images
CN107527044A (en) 2017-09-18 2017-12-29 北京邮电大学 A kind of multiple car plate clarification methods and device based on search
US9865036B1 (en) 2015-02-05 2018-01-09 Pixelworks, Inc. Image super resolution via spare representation of multi-class sequential and joint dictionaries
CN107767343A (en) 2017-11-09 2018-03-06 京东方科技集团股份有限公司 Image processing method, processing unit and processing equipment
CN107766860A (en) 2017-10-31 2018-03-06 武汉大学 Natural scene image Method for text detection based on concatenated convolutional neutral net
US20180075581A1 (en) * 2016-09-15 2018-03-15 Twitter, Inc. Super resolution using a generative adversarial network
CN107977932A (en) 2017-12-28 2018-05-01 北京工业大学 It is a kind of based on can differentiate attribute constraint generation confrontation network face image super-resolution reconstruction method
WO2018086354A1 (en) 2016-11-09 2018-05-17 京东方科技集团股份有限公司 Image upscaling system, training method therefor, and image upscaling method
CN108052940A (en) 2017-12-17 2018-05-18 南京理工大学 SAR remote sensing images waterborne target detection methods based on deep learning
CN108122197A (en) 2017-10-27 2018-06-05 江西高创保安服务技术有限公司 A kind of image super-resolution rebuilding method based on deep learning
CN108154499A (en) 2017-12-08 2018-06-12 东华大学 A kind of woven fabric texture flaw detection method based on K-SVD study dictionaries
CN108268870A (en) 2018-01-29 2018-07-10 重庆理工大学 Multi-scale feature fusion ultrasonoscopy semantic segmentation method based on confrontation study
CN108334848A (en) 2018-02-06 2018-07-27 哈尔滨工业大学 A kind of small face identification method based on generation confrontation network
CN108416428A (en) 2018-02-28 2018-08-17 中国计量大学 A kind of robot visual orientation method based on convolutional neural networks
US20180240257A1 (en) 2017-02-21 2018-08-23 Adobe Systems Incorporated Deep high-resolution style synthesis
CN108476291A (en) 2017-09-26 2018-08-31 深圳市大疆创新科技有限公司 Image generating method, video generation device and machine readable storage medium
CN108596830A (en) 2018-04-28 2018-09-28 国信优易数据有限公司 A kind of image Style Transfer model training method and image Style Transfer method
CN109255390A (en) 2018-09-30 2019-01-22 京东方科技集团股份有限公司 Preprocess method and module, discriminator, the readable storage medium storing program for executing of training image
CN109345455A (en) 2018-09-30 2019-02-15 京东方科技集团股份有限公司 Image identification method, discriminator, and computer-readable storage medium
CN109345456A (en) 2018-09-30 2019-02-15 京东方科技集团股份有限公司 Generative confrontation network training method, image processing method, device and storage medium
CN109360151A (en) 2018-09-30 2019-02-19 京东方科技集团股份有限公司 Image processing method and system, resolution enhancement method, and readable storage medium
US20190114742A1 (en) * 2017-10-13 2019-04-18 Adobe Inc. Image upscaling with controllable noise reduction using a neural network
US20190129858A1 (en) 2016-04-26 2019-05-02 Cambricon Technologies Corporation Limited Apparatus and methods for circular shift operations
US20190156201A1 (en) 2016-04-27 2019-05-23 Commissariat A L'energie Atomique Et Aux Energies Alternatives Device and method for distributing convolutional data of a convolutional neural network
US20190302290A1 (en) 2018-03-27 2019-10-03 Westerngeco Llc Generative adversarial network seismic data processor
US20190333198A1 (en) 2018-04-25 2019-10-31 Adobe Inc. Training and utilizing an image exposure transformation neural network to generate a long-exposure image from a single short-exposure image
US20190333199A1 (en) 2018-04-26 2019-10-31 The Regents Of The University Of California Systems and methods for deep learning microscopy
US20190370608A1 (en) * 2018-05-31 2019-12-05 Seoul National University R&Db Foundation Apparatus and method for training facial locality super resolution deep neural network
US20200034948A1 (en) * 2018-07-27 2020-01-30 Washington University Ml-based methods for pseudo-ct and hr mr image estimation

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130085893A1 (en) * 2011-09-30 2013-04-04 Ebay Inc. Acquisition and use of query images with image feature data
KR102338372B1 (en) * 2015-09-30 2021-12-13 삼성전자주식회사 Device and method to segment object from image
US10360477B2 (en) * 2016-01-11 2019-07-23 Kla-Tencor Corp. Accelerating semiconductor-related computations using learning based models
JP2018063504A (en) 2016-10-12 2018-04-19 株式会社リコー Generation model learning method, device and program
KR20180057096A (en) * 2016-11-21 2018-05-30 삼성전자주식회사 Device and method to perform recognizing and training face expression
CN108229508B (en) 2016-12-15 2022-01-04 富士通株式会社 Training apparatus and training method for training image processing apparatus
KR101854071B1 (en) * 2017-01-13 2018-05-03 고려대학교 산학협력단 Method of generating image of interest region using deep learning and apparatus thereof
KR101947782B1 (en) * 2017-02-22 2019-02-13 한국과학기술원 Apparatus and method for depth estimation based on thermal image, and neural network learning method
JP2018139071A (en) 2017-02-24 2018-09-06 株式会社リコー Generation model learning method, generation model learning apparatus and program
KR102499396B1 (en) * 2017-03-03 2023-02-13 삼성전자 주식회사 Neural network device and operating method of neural network device
RU2652722C1 (en) * 2017-05-03 2018-04-28 Самсунг Электроникс Ко., Лтд. Data processing for super-resolution

Patent Citations (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5781196A (en) 1990-10-19 1998-07-14 Eidos Plc Of The Boat House Video compression by extracting pixel changes exceeding thresholds
US5754697A (en) 1994-12-02 1998-05-19 Fu; Chi-Yung Selective document image data compression technique
US6766067B2 (en) 2001-04-20 2004-07-20 Mitsubishi Electric Research Laboratories, Inc. One-pass super-resolution images
WO2002089046A1 (en) 2001-04-26 2002-11-07 Georgia Tech Research Corporation Video enhancement using multiple frame techniques
WO2003060823A2 (en) 2001-12-26 2003-07-24 Yeda Research And Development Co.Ltd. A system and method for increasing space or time resolution in video
CN101593269A (en) 2008-05-29 2009-12-02 汉王科技股份有限公司 Face identification device and method
US20120328210A1 (en) 2010-01-28 2012-12-27 Yissum Research Development Company Of The Hebrew University Of Jerusalem Method and system for generating an output image of increased pixel resolution from an input image
CN101872472A (en) 2010-06-02 2010-10-27 中国科学院自动化研究所 A face image super-resolution reconstruction method based on sample learning
US9727959B2 (en) 2011-09-28 2017-08-08 The United States Of America As Represented By The Secretary Of The Army System and processor implemented method for improved image quality and generating an image of a target illuminated by quantum particles
US20130301933A1 (en) 2012-05-10 2013-11-14 Thomson Licensing Method and device for generating a super-resolution version of a low resolution input data structure
CN103426148A (en) 2012-05-10 2013-12-04 汤姆逊许可公司 Method and device for generating a super-resolution version of a low resolution input data structure
US8675999B1 (en) 2012-09-28 2014-03-18 Hong Kong Applied Science And Technology Research Institute Co., Ltd. Apparatus, system, and method for multi-patch based super-resolution from an image
CN102915527A (en) 2012-10-15 2013-02-06 中山大学 Face image super-resolution reconstruction method based on morphological component analysis
US20150296232A1 (en) 2012-11-27 2015-10-15 Lg Electronics Inc. Signal transceiving apparatus and signal transceiving method
CN103514580A (en) 2013-09-26 2014-01-15 香港应用科技研究院有限公司 Method and system for obtaining super-resolution images optimized for viewing experience
US20170178293A1 (en) 2014-02-13 2017-06-22 Thomson Licensing Method for performing super-resolution on single images and apparatus for performing super-resolution on single images
CN104853059A (en) 2014-02-17 2015-08-19 台达电子工业股份有限公司 Super-resolution image processing method and device
US20150235345A1 (en) 2014-02-17 2015-08-20 Delta Electronics, Inc. Method and device for processing a super-resolution image
CN103903236B (en) 2014-03-10 2016-08-31 北京信息科技大学 The method and apparatus of face image super-resolution rebuilding
CN103903236A (en) 2014-03-10 2014-07-02 北京信息科技大学 Method and device for reconstructing super-resolution facial image
CN105144232A (en) 2014-03-25 2015-12-09 展讯通信(上海)有限公司 Methods and systems for denoising images
US9865036B1 (en) 2015-02-05 2018-01-09 Pixelworks, Inc. Image super resolution via spare representation of multi-class sequential and joint dictionaries
WO2017100903A1 (en) 2015-12-14 2017-06-22 Motion Metrics International Corp. Method and apparatus for identifying fragmented material portions within an image
US20190129858A1 (en) 2016-04-26 2019-05-02 Cambricon Technologies Corporation Limited Apparatus and methods for circular shift operations
US20190156201A1 (en) 2016-04-27 2019-05-23 Commissariat A L'energie Atomique Et Aux Energies Alternatives Device and method for distributing convolutional data of a convolutional neural network
CN105976318A (en) 2016-04-28 2016-09-28 北京工业大学 Image super-resolution reconstruction method
CN105975931A (en) 2016-05-04 2016-09-28 浙江大学 Convolutional neural network face recognition method based on multi-scale pooling
CN105975968A (en) 2016-05-06 2016-09-28 西安理工大学 Caffe architecture based deep learning license plate character recognition method
RU2635883C1 (en) 2016-06-02 2017-11-16 Самсунг Электроникс Ко., Лтд. Image processing method and system for forming superhigh-resolution images
US20170365038A1 (en) 2016-06-16 2017-12-21 Facebook, Inc. Producing Higher-Quality Samples Of Natural Images
US20180075581A1 (en) * 2016-09-15 2018-03-15 Twitter, Inc. Super resolution using a generative adversarial network
WO2018086354A1 (en) 2016-11-09 2018-05-17 京东方科技集团股份有限公司 Image upscaling system, training method therefor, and image upscaling method
US20190005619A1 (en) 2016-11-09 2019-01-03 Boe Technology Group Co., Ltd. Image upscaling system, training method thereof, and image upscaling method
US20180240257A1 (en) 2017-02-21 2018-08-23 Adobe Systems Incorporated Deep high-resolution style synthesis
CN107133601A (en) 2017-05-13 2017-09-05 五邑大学 A kind of pedestrian's recognition methods again that network image super-resolution technique is resisted based on production
CN107154023A (en) 2017-05-17 2017-09-12 电子科技大学 Face super-resolution reconstruction method based on generation confrontation network and sub-pix convolution
CN107369189A (en) 2017-07-21 2017-11-21 成都信息工程大学 The medical image super resolution ratio reconstruction method of feature based loss
CN107527044A (en) 2017-09-18 2017-12-29 北京邮电大学 A kind of multiple car plate clarification methods and device based on search
CN108476291A (en) 2017-09-26 2018-08-31 深圳市大疆创新科技有限公司 Image generating method, video generation device and machine readable storage medium
US20190114742A1 (en) * 2017-10-13 2019-04-18 Adobe Inc. Image upscaling with controllable noise reduction using a neural network
CN108122197A (en) 2017-10-27 2018-06-05 江西高创保安服务技术有限公司 A kind of image super-resolution rebuilding method based on deep learning
CN107766860A (en) 2017-10-31 2018-03-06 武汉大学 Natural scene image Method for text detection based on concatenated convolutional neutral net
CN107767343A (en) 2017-11-09 2018-03-06 京东方科技集团股份有限公司 Image processing method, processing unit and processing equipment
US10430683B2 (en) 2017-11-09 2019-10-01 Boe Technology Group Co., Ltd. Image processing method and processing device
CN108154499A (en) 2017-12-08 2018-06-12 东华大学 A kind of woven fabric texture flaw detection method based on K-SVD study dictionaries
CN108052940A (en) 2017-12-17 2018-05-18 南京理工大学 SAR remote sensing images waterborne target detection methods based on deep learning
CN107977932A (en) 2017-12-28 2018-05-01 北京工业大学 It is a kind of based on can differentiate attribute constraint generation confrontation network face image super-resolution reconstruction method
CN108268870A (en) 2018-01-29 2018-07-10 重庆理工大学 Multi-scale feature fusion ultrasonoscopy semantic segmentation method based on confrontation study
CN108334848A (en) 2018-02-06 2018-07-27 哈尔滨工业大学 A kind of small face identification method based on generation confrontation network
CN108416428A (en) 2018-02-28 2018-08-17 中国计量大学 A kind of robot visual orientation method based on convolutional neural networks
US20190302290A1 (en) 2018-03-27 2019-10-03 Westerngeco Llc Generative adversarial network seismic data processor
US20190333198A1 (en) 2018-04-25 2019-10-31 Adobe Inc. Training and utilizing an image exposure transformation neural network to generate a long-exposure image from a single short-exposure image
US20190333199A1 (en) 2018-04-26 2019-10-31 The Regents Of The University Of California Systems and methods for deep learning microscopy
CN108596830A (en) 2018-04-28 2018-09-28 国信优易数据有限公司 A kind of image Style Transfer model training method and image Style Transfer method
US20190370608A1 (en) * 2018-05-31 2019-12-05 Seoul National University R&Db Foundation Apparatus and method for training facial locality super resolution deep neural network
US20200034948A1 (en) * 2018-07-27 2020-01-30 Washington University Ml-based methods for pseudo-ct and hr mr image estimation
CN109360151A (en) 2018-09-30 2019-02-19 京东方科技集团股份有限公司 Image processing method and system, resolution enhancement method, and readable storage medium
CN109345456A (en) 2018-09-30 2019-02-15 京东方科技集团股份有限公司 Generative confrontation network training method, image processing method, device and storage medium
CN109345455A (en) 2018-09-30 2019-02-15 京东方科技集团股份有限公司 Image identification method, discriminator, and computer-readable storage medium
CN109255390A (en) 2018-09-30 2019-01-22 京东方科技集团股份有限公司 Preprocess method and module, discriminator, the readable storage medium storing program for executing of training image

Non-Patent Citations (65)

* Cited by examiner, † Cited by third party
Title
"Circshift"; The Wayback Machine; MATLAB; Jul. 7, 2022.
"ImTranslate"; The Wayback Machine; MATLAB; Jul. 7, 2022
"Rgb2gray"; MATLAB; The Wayback Machine; Jul. 7, 2022.
"Translate an Image Using Imtranslate Function"; MATLAB & Simulink; The Wayback Machine; Jul. 6, 2022.
Blau, Yochai, and Tomer Michaeli. "The Perception-Distortion Tradeoff." arXiv preprint arXiv:1711.06077v2 (2018). (Year: 2018). *
Blau, Yochai, et al. "The Perception-DistortionTradeoff"; IEEE; Oct. 25, 2020.
Bulat, et al. "To Learn Image Super-Resolution, Use A GAN to Learn How to do Image Degradation First"; Jul. 30, 2018.
Chen, Li, et al. "Joint denoising and super-resolution via generative adversarial training." 2018 24th International Conference on Pattern Recognition (ICPR). IEEE, 2018. (Year: 2018). *
Cheon, Manri, et al. "Generative adversarial network-based image super-resolution using perceptual content losses." arXiv preprint arXiv:1809.04783v2 (2018). (Year: 2018). *
Extended European Search Report dated Dec. 17, 2021 for application No. EP 18889945.4.
First Office Action dated Nov. 8, 2021 for application No. IN 202047021736.
First Office Action dated Oct. 29, 2021 for application No. KR 10-2020-7014462 with English translation attached.
Goodfellow, et al. "Deep learning. MIT Press—Chapter 9 Convolutional Networks"; 2016.
He, et al. "Deep Residual Learning for Image Recognition"; 2016.
He, et al. "Deep Residual Learning for Image Recognition"; Dec. 10, 2015.
He, Kaiming; Zhang, Xiangyu; Ren, Shaoqing; Sun, Jian: Deep Residual Learning for Image Recognition: Microsoft Research: Dec. 10, 2015.
Huang, et al. "Densely Connected Convolutional Networks"; Jan. 28, 2018.
Huang, Gao; Chen, Danlu; Li, Tianhong; Wu, Felix; Van Der Maaten, Laurens; Weinberger, Kilian: Multi-Scale Dense Networks for Resource Efficient Image Classification: Published as a conference paper at ICLR 2018: Jun. 7, 2018.
Indian First Office Action dated Dec. 27, 2021 corresponding to application No. 202027055323.
International Search Report dated Mar. 15, 2019 corresponding to application No. PCT/CN2018/121466.
Khan, Amir, et al. "Implementation and Experiments on Face Detection System (FDS) Using Perceptual Quality Aware Features"; Eastern Mediterranean University; Feb. 2017.
Lai, et al. "Fast and Accurate Image Super-Resolution with Deep Laplacian Pyramid Networks"; Aug. 9, 2018.
Lai, Wei-Sheng; Huang, Jia-Bin; Ahuja, Narendra; Yang, Ming-Hsuan: Fast and Accurate Image Super-Resolution with Deep Laplacian Pyramid Networks: Aug. 9, 2018.
Ledig, Christian, et al. "Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network"; Twitter, Inc.; Sep. 15, 2016.
Lim, Bee; Son, Sanghyun; Kim, Heewon; Nah, Seungjun; Lee, Kyoung Mu: Enhanced Deep Residual Networks for Single Image Super-Resolution: Jul. 10, 2017: Seoul, Korea.
Liu, et al. "Deep Networks for Image-to-Image Translation with Mux and Demux Layers"; 2018.
Liu, Hanwen, et al. "Deep Networks for Image-to-Image Translation with Mux and Demux Layers"; BOE Technology Group Co., Ltd.; Springer Nature Switzerland; 2019.
Liu, Hanwen; Michelini, Pablo Navarrete; Zhu, Dan: Deep Networks for Image-to-Image Translation with Mux and Demux Layers: ECCV 2018 workshop paper: Beijing, China.
Mechrez, Roey, Itamar Talmi, and Lihi Zelnik-Manor. "The Contextual Loss for Image Transformation with Non-Aligned Data." arXiv preprint arXiv:1803.02077v4 (2018). (Year: 2018). *
Michelini, et al. "Multi-Scale Recursive and Perception-Distortion Controllable Image Super-Resolution"; 2018.
Michelini, Pablo Navarrete; Zhu, Dan; Liu, Hanwen: Multi-Scale Recursive and Perception-Distortion Controllable Image Super-Resolution: ECCV 2018 workshop paper: Jan. 29, 2019.
Mittal, Anish, et al. "No-Reference Image Quality Assessment in the Spatial Domain"; IEEE Transactions on Image Processing ; vol. 21, No. 12; Dec. 2012.
Navarrete, Pablo, et al. "Multi-Scale Recursive and Perception-Distortion Image Super Resolution"; BOE Technology Group Co., Ltd.; Sep. 27, 2018.
Notice of Acceptance dated Sep. 28, 2021 issued in corresponding Australian Application No. 2019350918.
Notice of Allowance dated Jul. 27, 2021 issued in corresponding U.S. Appl. No. 16/465,294.
Notification of a Grant dated Jan. 18, 2022 corresponding to Korean application No. 9-5-2022-005086425.
Office Action dated Apr. 23, 2021 issued in corresponding U.S. Appl. No. 16/465,294.
Office Action dated May 13, 2021 issued in corresponding Russian Application No. 2020136214.
Office Action dated May 21, 2021 issued in corresponding Australian Application No. 2019350918.
Office Action dated Sep. 9, 2021 issued in corresponding U.S. Appl. No. 16/614,558.
Rasti, Pejman; Demirel, Hasan; Anbarjafari, Gholamreza: Iterative Back Projection based Image Resolution Enhancement: 2013 8th Iranian Conference of Machine Vision and Image Processing: Zanjan, Iran.
Ruderman, Daniel L. "The Statistics of Natural Images"; Computation in Neural Systems 5 (1994), 517-548.
Salimans, et al. "Improved Techniques for Training GANs"; Jun. 10, 2016.
Samangouei, et al., "Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models", Published as a conference paper at ICLR 2018, pp. 1-17.
Search Report dated Dec. 26, 2019 issued in corresponding International Application No. PCT/CN2019/107761.
Search Report dated Jul. 24, 2019 issued in corresponding International Application No. PCT/CN2019/083872.
Search Report dated Sep. 26, 2019 issued in corresponding International Application No. PCT/CN2019/092113.
Search Report dated Sep. 27, 2019 issued in corresponding International Application No. PCT/CN2019/092042.
Takeki, et al. "Parallel Grid Pooling for Data Augmentation"; Mar. 30, 2018.
The Extedned European Search Report dated May 31, 2022 corresponding to application No. 19850782.4-1210.
The Extended European Search Report dated Jul. 13, 2022 corresponding to application No. 19864756.2-1210.
The Extended European Search Report dated Jul. 8, 2022 corresponding to application No. 19850805.3-1210.
The Extended European Search Report dated Jun. 17, 2022 corresponding to application No. 19850757.6-1207.
The First Office Action dated Apr. 26, 2020 corresponding to Chinese application No. 201811155326.6.
The First Office Action dated Apr. 27, 2020 corresponding to Chinese application No. 201811155147.2.
The First Office Action dated Feb. 11, 2020 corresponding to Chinese application No. 201811155930.9.
The First Office Action dated Jul. 2, 2020 corresponding to Chinese application No. 201810280478.2.
The First Office Action dated Jun. 30, 2020 corresponding to Chinese application No. 201811155252.6.
United States of America Non-Final Office Action dated Jul. 13, 2022 corresponding to U.S. Appl. No. 16/604,410.
United States of America Non-Final Office Action dated Nov. 15, 2021 corresponding to U.S. Appl. No. 16/614,547.
Wang, et al. "High-Resolution Image Synthesis and Semantic Manipulation with Conditional GAMS"; Aug. 20, 2018.
Wang, Xintao, et al. "ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks." arXiv preprint arXiv:1809.00219v2 (2018). (Year: 2018). *
Yang, et al. "Convolutional Neural Networks with Alternately Updated Clique"; Apr. 3, 2018.
Yu, et al., "Ultra-Resolving Face Images by Discriminative Generative Networks", ECCV-16 submission ID 1002, pp. 1-16.
Zhang, Dongyang, et al. "Sharp and real image super-resolution using generative adversarial network." International Conference on Neural Information Processing. Springer, Cham, 2017. (Year: 2017). *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220230276A1 (en) * 2019-05-23 2022-07-21 Deepmind Technologies Limited Generative Adversarial Networks with Temporal and Spatial Discriminators for Efficient Video Generation
US12277672B2 (en) * 2019-05-23 2025-04-15 Deepmind Technologies Limited Generative adversarial networks with temporal and spatial discriminators for efficient video generation

Also Published As

Publication number Publication date
EP3859655A4 (en) 2022-08-10
EP3857447A4 (en) 2022-06-29
RU2762144C1 (en) 2021-12-16
US11615505B2 (en) 2023-03-28
AU2019350918A1 (en) 2020-11-19
EP3857504A4 (en) 2022-08-10
KR20200073267A (en) 2020-06-23
US20210365744A1 (en) 2021-11-25
JP2022501663A (en) 2022-01-06
JP7415251B2 (en) 2024-01-17
MX2020013580A (en) 2021-02-26
EP3857503A1 (en) 2021-08-04
WO2020062846A1 (en) 2020-04-02
JP2022501662A (en) 2022-01-06
US11348005B2 (en) 2022-05-31
KR102661434B1 (en) 2024-04-29
EP3857503A4 (en) 2022-07-20
EP3859655A1 (en) 2021-08-04
JP2022501661A (en) 2022-01-06
AU2019350918B2 (en) 2021-10-07
EP3857504A1 (en) 2021-08-04
KR102389173B1 (en) 2022-04-21
EP3857447A1 (en) 2021-08-04
US20210334642A1 (en) 2021-10-28
WO2020062958A1 (en) 2020-04-02
US20200285959A1 (en) 2020-09-10
EP3859655B1 (en) 2025-04-23
JP7446997B2 (en) 2024-03-11
US20210342976A1 (en) 2021-11-04
WO2020063648A1 (en) 2020-04-02
KR20210012009A (en) 2021-02-02
BR112020022560A2 (en) 2021-06-01
US11361222B2 (en) 2022-06-14
JP7463643B2 (en) 2024-04-09
WO2020062957A1 (en) 2020-04-02

Similar Documents

Publication Publication Date Title
US11449751B2 (en) Training method for generative adversarial network, image processing method, device and storage medium
CN109345456B (en) Generative confrontation network training method, image processing method, device and storage medium
US20220164601A1 (en) Methods and Apparatuses of Contrastive Learning for Color Constancy
CN110838119B (en) Human face image quality evaluation method, computer device and computer readable storage medium
US20240420502A1 (en) Facial image processing method and related device
US8103058B2 (en) Detecting and tracking objects in digital images
CN109978882A (en) A kind of medical imaging object detection method based on multi-modal fusion
CN113591831B (en) Font identification method, system and storage medium based on deep learning
CN117527983A (en) Image information hiding method based on Transformer
JP2008033424A (en) Image processing apparatus, image processing method, program, and storage medium
US8873839B2 (en) Apparatus of learning recognition dictionary, and method of learning recognition dictionary
CN111259792A (en) Face liveness detection method based on DWT-LBP-DCT feature
US20220414827A1 (en) Training apparatus, training method, and medium
CN111046893A (en) Image similarity determination method and device, image processing method and device
CN111753714A (en) A multi-directional natural scene text detection method based on character segmentation
Grijalva et al. Smartphone recognition of the US banknotes' denomination, for visually impaired people
JP4588575B2 (en) Method, apparatus and program for detecting multiple objects in digital image
JP2011170890A (en) Face detecting method, face detection device, and program
CN116612521A (en) A face recognition method, device, chip and terminal
CN116246330A (en) A Fine-Grained Face Age Estimation Method Based on Horizontal Pyramid Matching
Pal et al. Super-resolution of textual images using autoencoders for text identification
JP4795737B2 (en) Face detection method, apparatus, and program
US20250005720A1 (en) Resizing for enhanced inference
LU505861B1 (en) Method and system for image to image translation
JP2006285959A (en) Learning method of face discriminating apparatus, face discriminating method and apparatus, and program

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: BOE TECHNOLOGY GROUP CO., LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIU, HANWEN;ZHU, DAN;NAVARRETE MICHELINI, PABLO;SIGNING DATES FROM 20200313 TO 20200319;REEL/FRAME:052529/0534

STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED

STCF Information on status: patent grant

Free format text: PATENTED CASE

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE

STCF Information on status: patent grant

Free format text: PATENTED CASE