CN111696036A - Residual error neural network based on cavity convolution and two-stage image demosaicing method - Google Patents

Residual error neural network based on cavity convolution and two-stage image demosaicing method Download PDF

Info

Publication number
CN111696036A
CN111696036A CN202010447460.4A CN202010447460A CN111696036A CN 111696036 A CN111696036 A CN 111696036A CN 202010447460 A CN202010447460 A CN 202010447460A CN 111696036 A CN111696036 A CN 111696036A
Authority
CN
China
Prior art keywords
image
neural network
channel
residual
mosaic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010447460.4A
Other languages
Chinese (zh)
Other versions
CN111696036B (en
Inventor
朱树元
王岩
王忠荣
刘光辉
曾辽原
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN202010447460.4A priority Critical patent/CN111696036B/en
Publication of CN111696036A publication Critical patent/CN111696036A/en
Application granted granted Critical
Publication of CN111696036B publication Critical patent/CN111696036B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4015Image demosaicing, e.g. colour filter arrays [CFA] or Bayer patterns
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4046Scaling of whole images or parts thereof, e.g. expanding or contracting using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • G06V10/449Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
    • G06V10/451Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
    • G06V10/454Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the field of digital image processing, and particularly relates to a residual error neural network based on cavity convolution and a two-stage image demosaicing method; the method introduces a shallow feature extraction unit, a local residual error unit and a deep feature extraction unit based on the residual neural network, three basic units interact with each other to greatly enhance the learning capability and the modeling capability of the target neural network, and can establish accurate mapping from a mosaic image to an RGB color image aiming at the problem of image demosaicing, and finally, the mosaic image in a Bayer CFA mode can be processed through the established effective mapping to obtain the RGB color image; meanwhile, a two-stage image demosaicing model is introduced, prior information is fully utilized, the modeling capability of the network is improved, and the knowledge space is optimized; the image demosaicing method can obviously improve the peak signal-to-noise ratio of the image, greatly improve the efficiency, quality and robustness of demosaicing the image, and has profound significance in the field of image processing.

Description

Residual error neural network based on cavity convolution and two-stage image demosaicing method
Technical Field
The invention belongs to the field of digital image processing, and particularly relates to a residual error neural network based on cavity convolution and a two-stage image demosaicing method.
Background
Single sensor color imaging technology using CFA is widely used in the current digital camera industry. In a single sensor camera with CFA, each pixel point only records one pixel point in an R channel image, a G channel image and a B channel image, and when an RGB color image is recovered, the other two lost pixel values need to be estimated; this process, commonly referred to as image demosaicing, plays a crucial role in the field of obtaining high quality RGB color images. The most popular and widely used CFA is the bayer CFA, in which green pixel points are sampled on a quincunx grid and red and blue pixel points are sampled on a rectangular grid, where the number of green sample points is twice the number of red or blue sample points.
The bayer cfa-based image demosaicing method has been widely studied, and most bayer cfa-based image demosaicing methods first interpolate G-channel image values; for the R channel image and the B channel image, they are usually first converted into color ratios or color differences, and then interpolated in the transform domain. Another popular method is a frequency domain algorithm, which first converts the mosaic image to the frequency domain and then separates the luminance and chrominance components by frequency filtering. In addition, an image demosaicing method based on a compressed sensing theory is also provided.
In recent years, deep learning has been excellent in solving many problems such as image classification, object detection, and natural language processing; among the different types of neural networks, Convolutional Neural Networks (CNNs) are the most extensively studied. The CNN can automatically extract effective representations in the target image, namely the CNN can directly carry out effective modeling on data from the original pixels through few preprocessing. The CNN is adopted in the literature 'Color Image demosaicing via Deep Residual Learning', so that the Image demosaicing is realized, and compared with the traditional method, the method has good performance gain, but due to the introduction of Batch Normalization (BN), the depth of the network is reduced on the premise of the same operation cost, and the perception domain of the network is reduced. A three-Stage Image demosaicing Using a 3-Stage conditional Neural network structure is constructed in a Color Image demosaicing Using document, which fully embodies the excellent modeling capability and learning capability of a Neural network, but the restored G-channel Image is used as prior information to guide the restoration of an R-channel Image and a B-channel Image to be suboptimal.
Disclosure of Invention
The invention aims to provide a residual error neural network based on cavity convolution and a two-stage image demosaicing method aiming at the technical problems, the relation between a sensing domain and the network operation cost is fully considered, a residual error block is constructed by adopting the cavity convolution, and the sensing domain of the network is increased on the premise of not increasing the operation cost; in the network training stage, the original G channel image is used as prior information to guide the recovery of the R channel image and the B channel image, so that the modeling capability of the network is better, and the learning space is optimized; the residual error neural network based on the cavity convolution and the two-stage image demosaicing method can obviously improve the peak signal-to-noise ratio (PSNR) of the image and have the advantages of good demosaicing effect, high speed, strong robustness and the like.
In order to achieve the purpose, the invention adopts the following technical scheme:
a residual error neural network based on cavity convolution and a two-stage image demosaicing method comprise the following steps:
step 1: building a residual error neural network model based on cavity convolution;
step 2: converting the RGB color image into a mosaic image through Bayer CFA, carrying out data preprocessing to form a training set, and setting parameters of a training target neural network;
and step 3: according to the residual error neural network model based on the cavity convolution, training a corresponding neural network model by taking a minimized loss function as a target in two stages;
and 4, step 4: and (3) according to the trained target neural network model, performing the same data preprocessing as that in the step (2) on the mosaic image of the Bayer CFA mode to be processed, inputting the mosaic image into the target neural network model, and outputting the RGB color image without mosaic.
Further, in step 1, the residual neural network model based on the hole convolution includes: the device comprises a neural network G, a neural network R and a neural network B, wherein the neural network G, the neural network R and the neural network B all adopt a residual error neural network based on cavity convolution; the method comprises the following steps that a G channel image is restored by inputting a mosaic image into a neural network G, an R channel image is restored by inputting the mosaic image into the neural network R, the G channel image restored by the neural network G and an image only containing R channel sampling pixels, a B channel image is restored by inputting the mosaic image into a neural network B, the G channel image restored by the neural network G and the image only containing B channel sampling pixels, and the restored R channel image, the restored G channel image and the restored B channel image are synthesized into a restored RGB color image.
Still further, in step 1, the residual neural network based on the hole convolution includes: 1 shallow layer feature extraction unit, N local residual error units and 1 deep layer feature extraction unit, wherein N is more than or equal to 1; the method is characterized in that an input image is converted into shallow features through a shallow feature extraction unit, the shallow features sequentially pass through N local residual error units to form main features, and the main features pass through a deep feature extraction unit to output residual error images.
Still further, the shallow feature extraction unit includes: a3 x 3 convolutional layers with ReLU activation function and 1 3 x 3 convolutional layers without activation function; the input image is converted into shallow layer characteristics through A3 x 3 convolutional layers with ReLU activation functions and 1 3 x 3 convolutional layer without the activation functions in sequence;
the local residual unit includes: b residual blocks and 1 dimension reduction block; wherein, the residual block contains 1 3 × 3 convolutional layer with ReLU activation function, C3 × 3 void convolutional layer with ReLU activation function and 1 3 × 3 convolutional layer without activation function, and the layers are connected end to end in sequence; the residual block forms a local residual using a jump join; the dimension reduction block comprises: 1 cascade layer and 1X 1 convolution layer without activation function, and the layers are connected end to end in sequence; inputting shallow layer characteristics and sequentially passing through B residual blocks, and outputting the output of each residual block through dimension reduction of a dimension reduction block;
the deep feature extraction unit includes: d3 x 3 convolutional layers with ReLU activation functions and 1 3 x 3 convolutional layer without activation functions, which are connected end to end in sequence; the input main features sequentially pass through D3 x 3 convolutional layers with ReLU activation functions and 1 3 x 3 convolutional layer without activation functions to output residual images.
In all the local residual units, the local residual in the residual block is subjected to a feature scaling technique and then the input and the output are connected in an identical manner.
Further, in step 2, the data preprocessing process includes:
firstly, setting a G channel sampling pixel and a B channel sampling pixel in a mosaic image to be 0 to obtain an image only containing the R channel sampling pixel, and setting the G channel sampling pixel and the R channel sampling pixel in the mosaic image to be 0 to obtain an image only containing the B channel sampling pixel;
then, the mosaic image, the image only containing R-channel sampling pixels, the image only containing B-channel sampling pixels, the original R-channel image, the original G-channel image and the original B-channel image are respectively divided into a plurality of mosaic image blocks, image blocks only containing R-channel sampling pixels, image blocks only containing B-channel sampling pixels, original R-channel image blocks, original G-channel image blocks and original B-channel image blocks, wherein the number and the size of the mosaic image blocks, the image blocks only containing R-channel sampling pixels, the image blocks only containing B-channel sampling pixels, the original R-channel image blocks, the original G-channel image blocks and the original B-channel image blocks are the same.
Further, in step 3, the first stage of the two-stage image demosaicing method is: training a neural network R by using a loss function R, wherein the neural network R constructs a mapping from a mosaic image, an image only containing R channel sampling pixels and an original G channel image to an R channel image; training a neural network G by using a loss function G, wherein the neural network G constructs a mapping from a mosaic image to a G channel image; training a neural network B by using a loss function B, wherein the neural network B constructs a mapping from a mosaic image, an image only containing B-channel sampling pixels and an original G-channel image to a B-channel image; the second stage of the two-stage image demosaicing method is as follows: and (3) jointly training a neural network R, a neural network G and a neural network B by using a loss function RGB, wherein the neural network R, the neural network G and the neural network B together construct a mapping from the mosaic image to the RGB color image.
Further, in step 3, the loss function of the first stage includes: a loss function R, a loss function G and a loss function B; the loss function of the second stage includes: the loss function RGB, and all of the loss functions take the form of a Mean Absolute Error (MAE) function:
Figure BDA0002506333530000041
Figure BDA0002506333530000042
Figure BDA0002506333530000043
Figure BDA0002506333530000044
wherein L isR(·)、LG(·)、LB(. and L)RGB(. The) represents a loss function R, a loss function G, a loss function B, and a loss function RGB, θ, respectivelyR、θGAnd thetaBRespectively representing parameters of a neural network R, parameters of a neural network G and parameters of a neural network B, wherein M represents the total number of image blocks; f. ofR(. -) represents a trained mapping from a mosaic image, an image containing only R-channel sample pixels, and an original G-channel image to an R-channel image, fG(. represents a trained mapping from mosaic images to G-channel images, fB(. to) shows a trained slave mosaic image, containing only B channelsMapping of the image of the sampled pixels and the original G-channel image to a B-channel image, fR,G,B(. represents f)R(·),fG(. and f)B(ii) a combined mapping of; x is the number ofmos、xmos_rAnd xmos_bRespectively representing a mosaic image, an image containing only R-channel sample pixels and an image containing only B-channel sample pixels, xrgbRepresenting an RGB color image recovered by synthesizing an R channel image recovered by a neural network R, a G channel image recovered by a neural network G and a B channel image recovered by a neural network B; y isR、yG、yBAnd yRGBRepresenting an original R-channel image, an original G-channel image, an original B-channel image, and an original RGB color image, respectively.
Still further, in the step 3, in the process of training the residual error neural network model based on cavity convolution, a parameter θ in the residual error neural network based on cavity convolutionR、θGAnd thetaBAdopts an Xavier method, namely that the mean value is 0 and the variance is Var (theta)R)、Var(θG) And Var (θ)B) Distribution of (a):
Figure BDA0002506333530000045
Figure BDA0002506333530000051
Figure BDA0002506333530000052
wherein the content of the first and second substances,
Figure BDA0002506333530000053
and
Figure BDA0002506333530000054
respectively representing the number of the input neurons of the layer in the neural network R, the number of the input neurons of the layer in the neural network G and the number of the input neurons of the layer in the neural network B,
Figure BDA0002506333530000055
and
Figure BDA0002506333530000056
respectively representing the number of the output neurons of the layer in the neural network R, the number of the output neurons of the layer in the neural network G and the number of the output neurons of the layer in the neural network B; the minimization loss function employs an Adam optimization method.
Compared with the prior art, the invention has the beneficial effects that:
according to the residual error neural network based on cavity convolution and the two-stage image demosaicing method, the shallow feature extraction unit, the local residual error unit and the deep feature extraction unit are introduced, three basic units interact with each other to greatly enhance the learning capability and the modeling capability of a target neural network, accurate mapping from a mosaic image to an RGB color image can be established aiming at the image demosaicing problem, and finally the mosaic image of a Bayer CFA mode can be processed through the established effective mapping to obtain the RGB color image; meanwhile, a two-stage image demosaicing model is introduced, prior information is fully utilized, the modeling capability of the network is improved, and the knowledge space is optimized; the image demosaicing method can obviously improve the peak signal-to-noise ratio (PSNR) of the image, greatly improve the efficiency, quality and robustness of image demosaicing, and has profound significance in the field of image processing.
Further, the invention also has the following beneficial effects:
the built residual error neural network model based on the cavity convolution has the modularized property, different units have different functions, a shallow feature extraction unit aims at converting an image into a shallow feature, a local residual error unit is a network main body and aims at converting the shallow feature into a main feature, a deep feature extraction unit aims at converting the main feature into a residual error image, and the shallow feature extraction unit, a plurality of local residual error units and the deep feature extraction unit are sequentially connected end to end; according to the connection mode, the network can stack a plurality of local residual error units, and the network depth is increased on the premise of ensuring efficient feature extraction, so that the nonlinear modeling capability and learning capability of the network are improved.
Each local residual error unit in the residual error neural network model based on the cavity convolution built by the invention contains a plurality of residual error blocks, each residual error block is provided with a jump connection, and each jump connection is subjected to residual error addition once; local residual learning in the local residual unit and global residual learning of the whole network interact, so that the performance and the convergence speed of the network are further improved; meanwhile, the residual block introduces a hole convolution, and the sensing domain of the network is expanded on the premise of ensuring the operation cost; the similar structure of the hole convolution and the Bayer CFA further promotes the extraction of the characteristics by the residual block, thereby improving the network performance.
Drawings
Fig. 1 is a schematic diagram of an internal structure of a residual neural network G based on hole convolution according to an embodiment of the present invention.
FIG. 2 is a schematic diagram of the internal structure of the residual neural network R/B based on the hole convolution according to the embodiment of the invention.
Fig. 3 is a schematic structural diagram of the shallow feature extraction unit in fig. 1.
Fig. 4 is a schematic structural diagram of the local residual unit in fig. 1.
Fig. 5 is a schematic structural diagram of the deep layer feature extraction unit in fig. 1.
FIG. 6 is a flow chart of the residual neural network based on hole convolution and the two-stage image demosaicing method according to the preferred embodiment of the invention.
Fig. 7 is a schematic diagram of the ReLU function in fig. 3.
Detailed Description
The invention will be further described with reference to the accompanying drawings and preferred embodiments.
The present embodiment provides a residual neural network based on void convolution and a two-stage image demosaicing method, and a flow diagram thereof is shown in fig. 6; the method comprises the following steps:
step 1: building a residual error neural network model based on cavity convolution;
the residual error neural network G based on the hole convolution is shown in fig. 1, and includes: 1 shallow layer feature extraction unit, 3 local residual error units and 1 deep layer feature extraction unit; the 3 local residual error units are sequentially connected end to end; the input image is converted into shallow layer characteristics through a shallow layer characteristic extraction unit, the shallow layer characteristics sequentially pass through 3 local residual error units to form main characteristics, and the main characteristics pass through a deep layer characteristic extraction unit to output residual error images;
the residual error neural network R/B based on the hole convolution is shown in FIG. 2 and comprises: 1 shallow layer feature extraction unit, 1 local residual error unit and 1 deep layer feature extraction unit; the input image is converted into a shallow feature through a shallow feature extraction unit, the shallow feature forms a main feature through a local residual error unit, and the main feature outputs a residual error image through a deep feature extraction unit;
in this embodiment, the reason that the neural network R and the neural network B only use 1 local residual unit is that, as the network depth increases, the restored R channel image and the restored B channel image generate a frame effect, and 1 local residual block can avoid the frame effect and obtain good subjective/objective performance;
as shown in fig. 3, the shallow feature extraction unit includes: 1 3 × 3 convolutional layer with ReLU activation function and 1 3 × 3 convolutional layer without activation function; the input image is converted into shallow layer characteristics through 1 3 x 3 convolutional layer with a ReLU activation function and 1 3 x 3 convolutional layer without the activation function in sequence; in this embodiment, the number of convolution kernels of the 3 × 3 convolution layers with the ReLU activation function (32 in this embodiment) is half of the number of convolution kernels of the 3 × 3 convolution layers without the activation function (64 in this embodiment), so as to achieve a gradual structure characteristic, which is advantageous for achieving the transformation from the image domain to the feature domain.
As shown in fig. 4, the local residual unit includes: a first residual block, a second residual block, a third residual block and a first dimension reduction block; wherein, the residual block contains 1 3 × 3 convolutional layer with ReLU activation function, 2 3 × 3 void convolutional layers with ReLU activation function and 1 3 × 3 convolutional layer without activation function, and the layers are connected end to end in sequence; the dimension reduction block comprises: 1 cascade layer and 1X 1 convolution layer without activation function, and the layers are connected end to end in sequence; the output characteristic of each residual block is input into a cascade layer, and the cascade layer is followed by 1 multiplied by 1 convolutional layer dimensionality reduction output without an activation function; in this embodiment, the local residual units have 3 skip connections, the skip connection unit in each residual block does not directly adopt identity mapping to connect input and output, but introduces a feature scaling technique, before identity mapping, a coefficient α smaller than 1 is used to scale features, and the feature scaling makes the training process more stable when training a deep network (α is 0.1 in this embodiment); in addition, the 3 × 3 void convolution layer (with an expansion coefficient of 2) is selected because of the small number of parameters and the low calculation cost, and because of the similarity to the bayer cfa pattern to a certain extent, the method is favorable for feature extraction.
As shown in fig. 5, the deep feature extraction unit includes: 3 x 3 convolutional layers with ReLU activation function and 1 3 x 3 convolutional layer without activation function; the input image sequentially passes through 3 × 3 convolutional layers with a ReLU activation function and 1 3 × 3 convolutional layer without an activation function to output a residual image.
Generally, a sensing domain has a direct relation to the improvement of network performance, and the expansion of the sensing domain usually uses a larger convolution kernel size or a deeper network structure, but the excessively large convolution kernel size or the excessively deeper network structure increases the network operation cost, and even causes the disappearance of a gradient and the generation of gradient explosion, which affects the network performance. The local residual error unit introduced by the invention uses the cavity convolution, and expands the perception domain of the network on the premise of not increasing the operation cost; in addition, the cavity convolution with the expansion coefficient of 2 and the Bayer CFA have similar physical properties to a certain extent, so that more effective characteristics can be extracted from the mosaic image, and the performance of the network is improved.
Step 2: converting the RGB color image into a mosaic image through Bayer CFA to form a training set, and setting parameters of a training target neural network;
the training set selects 800 images in DIV2K as a training set, RGB color images are converted into mosaic images through Bayer CFA, G channel sampling pixels and B channel sampling pixels in the mosaic images are set to be 0, images only containing R channel sampling pixels are obtained, and G channel sampling pixels and R channel sampling pixels in the mosaic images are set to be 0, and images only containing B channel sampling pixels are obtained; then setting training parameters of a residual error neural network model based on the cavity convolution, wherein the training parameters comprise the number of image blocks input into the model training each time, the sizes of input image blocks and output image blocks, learning rate and the like; respectively dividing a mosaic image, an image only containing R-channel sampling pixels, an image only containing B-channel sampling pixels, an original R-channel image, an original G-channel image and an original B-channel image in a training set into image blocks with the same resolution, wherein the number and the size of the mosaic image block, the image block only containing R-channel sampling pixels, the image block only containing B-channel sampling pixels, the original R-channel image block, the original G-channel image block and the original B-channel image block are the same; zero-padding operations are performed for each convolution (i.e., the image size is not reduced according to the size of the convolution kernel, i.e., the input and output sizes are consistent).
In this embodiment, a mosaic image in a training set, an image containing only R-channel sampling pixels, an image containing only B-channel sampling pixels, an original R-channel image, an original G-channel image, and an original B-channel image are respectively divided into a mosaic image block of 64 × 64, an image block containing only R-channel sampling pixels, an image block containing only B-channel sampling pixels, an original R-channel image block, an original G-channel image block, and an original B-channel image block, so that structural information and detail information of the images can be better captured when a model is trained; the number of image blocks trained by the model each time is 32 (in other embodiments, any value of 16, 32 and 128 can be taken); the learning rate is set to 0.0001 (in other embodiments, any value from 0.01 to 0.00001 may be taken), and the decay rate per training is set to 0.9 (in other embodiments, any value from 0.1 to 0.9 may be taken); the test is performed every 1000 times (in other embodiments, any value in 500-. Wherein, the training set can be selected and simultaneously the test set can be selected, and the Kodak or McMaster image set is selected in the test set.
And step 3: according to the residual error neural network model based on the cavity convolution, training a corresponding neural network model by taking a minimized loss function as a target in two stages;
specifically, the first stage: the minimum loss function R obtains the weight and the deviation of the neural network R, the minimum loss function G obtains the weight and the deviation of the neural network G, and the minimum loss function B obtains the weight and the deviation of the neural network B; and a second stage: the minimization of the loss function RGB yields the weights and biases of the neural network R, the neural network G, and the neural network B, so that a target neural network model of the image demosaicing problem can be established.
Wherein the loss function R, the loss function G, the loss function B and the loss function RGB all take the form of Mean Absolute Error (MAE) functions:
Figure BDA0002506333530000081
Figure BDA0002506333530000082
Figure BDA0002506333530000091
Figure BDA0002506333530000092
wherein L isR(·)、LG(·)、LB(. and L)RGB(. The) represents a loss function R, a loss function G, a loss function B, and a loss function RGB, θ, respectivelyR、θGAnd thetaBRespectively representing parameters of a neural network R, G and B, M representing an imageTotal number of blocks, fR(. -) represents a trained mapping from a mosaic image, an image containing only R-channel sample pixels, and an original G-channel image to an R-channel image, fG(. represents a trained mapping from mosaic images to G-channel images, fB(. -) represents a trained mapping from a mosaic image, an image containing only B-channel sample pixels, and an original G-channel image to a B-channel image, fR,G,B(. represents f)R(·),fG(. and f)B(ii) combined mapping of (x)mos、xmos_rAnd xmos_bRespectively representing a mosaic image, an image containing only R-channel sample pixels and an image containing only B-channel sample pixels, xrgbA recovered RGB color image, y, representing a synthesis of an R-channel image recovered by a neural network R, a G-channel image recovered by a neural network G, and a B-channel image recovered by a neural network BR、yG、yBAnd yRGBRepresenting an original R-channel image, an original G-channel image, an original B-channel image, and an original RGB color image, respectively.
Since the peak signal-to-noise ratio (PSNR) is formulated as:
Figure BDA0002506333530000093
Figure BDA0002506333530000094
wherein x isiRepresenting the recovered image, yiRepresenting the corresponding original image, and theta represents the parameter of the neural network; from the above equation, it can be seen that a high peak signal-to-noise ratio (PSNR) can be obtained by continuously minimizing the loss function, i.e. the higher the objective quality of the image.
In this embodiment, the minimization loss function adopts an Adam optimization method, wherein the calculation method of the Adam optimization method is as follows:
s←ρ1s+(1-ρ1)g
γ←ρ2γ+(1-ρ2)g⊙g
Figure BDA0002506333530000101
Figure BDA0002506333530000102
Figure BDA0002506333530000103
where ρ is1、ρ2And all are constants (the default values are respectively rho)1=0.9、ρ2=0.999、=0.001、=10-8) G represents the gradient of the loss function with respect to the parameter theta, s represents the biased first order moment estimate, gamma represents the biased second order moment estimate,
Figure BDA0002506333530000104
represents the deviation of the corrected first moment,
Figure BDA0002506333530000105
represents the deviation of the second moment of correction, and delta theta represents the change of the parameter theta; the updating of parameters of the Adam optimization method is not influenced by the expansion and contraction transformation of the gradient, and the learning rate can be adjusted in a self-adaptive manner; in addition, the algorithm is simple to implement, high in calculation efficiency and low in memory requirement.
In the embodiment, the parameter theta of the residual neural network model based on the cavity convolutionR、θGAnd thetaBThe method comprises the following steps:
Figure BDA0002506333530000106
Figure BDA0002506333530000107
Figure BDA0002506333530000108
wherein the content of the first and second substances,
Figure BDA0002506333530000109
and
Figure BDA00025063335300001010
respectively representing the number of the input neurons of the layer in the neural network R, the number of the input neurons of the layer in the neural network G and the number of the input neurons of the layer in the neural network B,
Figure BDA00025063335300001011
and
Figure BDA00025063335300001012
respectively representing the number of the output neurons of the layer in the neural network R, the number of the output neurons of the layer in the neural network G and the number of the output neurons of the layer in the neural network B; parameter thetaR、θGAnd thetaBInitialization obeys a mean of 0 and a variance of Var (theta)R)、Var(θG) And Var (θ)B) The distribution of (a); the Xavier parameter initialization method can improve the training efficiency of the network and can improve the performance of the network.
And 4, step 4: and inputting the mosaic image of the Bayer CFA mode into the target neural network model according to the trained target neural network model, and outputting the RGB color image after mosaic removal.
Converting the RGB color images into mosaic images through Bayer CFA in the training set in the step 2 to form a training set, training the neural network R, the neural network G and the neural network B in two stages in the step 3 to obtain an image demosaicing target neural network model aiming at a Bayer CFA mode, and inputting the mosaic images of the Bayer CFA mode into the target neural network model in the step 4 to obtain corresponding RGB color images.
In this embodiment, RGB color images in a Kodak test set (including 24 RGB color images with an image size of 768 × 512) are converted into a mosaic image by Bayer CFA, and after model mapping, the average PSNR of the obtained RGB color images is 42.94 dB; converting RGB color images in an McMaster test set (comprising 18 RGB color images with the image size of 500 multiplied by 500) into mosaic images through Bayer CFA, and obtaining the average PSNR of the color RGB images to be 39.76dB after model mapping; by the residual error neural network based on the cavity convolution and the two-stage image demosaicing method trained by the invention, the objective quality of the image is greatly improved, and the visual effect is satisfactory; the results are shown in the following table:
Figure BDA0002506333530000111
according to the image demosaicing method, a target neural network model can be trained in advance, the target neural network model is end-to-end mapping from an input Bayer CFA mode mosaic image to an output RGB color image, demosaicing speed of the mosaic image through the target neural network model is extremely high, the practical value is very high, and the method can be applied to occasions needing real-time deblocking effect; besides the advantages of high speed, good mosaic removing effect and the like, the mosaic removing method has strong robustness, and objective gain and subjective gain of mosaic removing do not fluctuate greatly aiming at mosaic images of different types and different scenes. Therefore, the residual error neural network based on the cavity convolution and the two-stage image demosaicing method have the advantages of good demosaicing effect, high speed, strong robustness, strong practicability and real-time performance, wide market prospect and particularly high requirement on the real-time performance.
The invention is based on the residual error neural network of the void convolution and the two-stage image demosaicing method, and can accurately learn the mapping from the mosaic image to the RGB color image; the hole convolution can enlarge the perception domain of the network on the premise of not increasing the operation cost; the convergence speed of the network can be accelerated by combining the local residual error and the global residual error; the neural network of the invention has interpretability by adopting a modular network structure; the two-stage training mode fully utilizes prior information and optimizes the solution space of the network model.
While the invention has been described with reference to specific embodiments, any feature disclosed in this specification may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise; all of the disclosed features, or all of the method or process steps, may be combined in any combination, except mutually exclusive features and/or steps.

Claims (9)

1. A residual error neural network based on cavity convolution and a two-stage image demosaicing method comprise the following steps:
step 1: building a residual error neural network model based on cavity convolution;
step 2: converting the RGB color image into a mosaic image through Bayer CFA, carrying out data preprocessing to form a training set, and setting parameters of a training target neural network;
and step 3: according to the residual error neural network model based on the cavity convolution, training a corresponding neural network model by taking a minimized loss function as a target in two stages;
and 4, step 4: and (3) according to the trained target neural network model, performing the same data preprocessing as that in the step (2) on the mosaic image of the Bayer CFA mode to be processed, inputting the mosaic image into the target neural network model, and outputting the RGB color image without mosaic.
2. The method for demosaicing a hole convolution-based residual neural network and a two-stage image as claimed in claim 1, wherein the step 1, the residual neural network model based on hole convolution comprises: the device comprises a neural network G, a neural network R and a neural network B, wherein the neural network G, the neural network R and the neural network B all adopt a residual error neural network based on cavity convolution; the method comprises the following steps that a G channel image is restored by inputting a mosaic image into a neural network G, an R channel image is restored by inputting the mosaic image into the neural network R, the G channel image restored by the neural network G and an image only containing R channel sampling pixels, a B channel image is restored by inputting the mosaic image into a neural network B, the G channel image restored by the neural network G and the image only containing B channel sampling pixels, and the restored R channel image, the restored G channel image and the restored B channel image are synthesized into a restored RGB color image.
3. The residual neural network and two-stage image demosaicing method based on hole convolution of claim 2, wherein the residual neural network based on hole convolution comprises: 1 shallow layer feature extraction unit, N local residual error units and 1 deep layer feature extraction unit, wherein N is more than or equal to 1; the method is characterized in that an input image is converted into shallow features through a shallow feature extraction unit, the shallow features sequentially pass through N local residual error units to form main features, and the main features pass through a deep feature extraction unit to output residual error images.
4. The residual neural network and two-stage image demosaicing method based on hole convolution as claimed in claim 3, wherein said shallow feature extraction unit comprises: a3 x 3 convolutional layers with ReLU activation function and 1 3 x 3 convolutional layers without activation function; the input image is converted into shallow layer characteristics through A3 x 3 convolutional layers with ReLU activation functions and 1 3 x 3 convolutional layer without the activation functions in sequence;
the local residual unit includes: b residual blocks and 1 dimension reduction block; wherein, the residual block contains 1 3 × 3 convolutional layer with ReLU activation function, C3 × 3 void convolutional layer with ReLU activation function and 1 3 × 3 convolutional layer without activation function, and the layers are connected end to end in sequence; the residual block forms a local residual using a jump join; the dimension reduction block comprises: 1 cascade layer and 1X 1 convolution layer without activation function, and the layers are connected end to end in sequence; inputting shallow layer characteristics and sequentially passing through B residual blocks, and outputting the output of each residual block through dimension reduction of a dimension reduction block;
the deep feature extraction unit includes: d3 x 3 convolutional layers with ReLU activation functions and 1 3 x 3 convolutional layer without activation functions, which are connected end to end in sequence; the input main features sequentially pass through D3 x 3 convolutional layers with ReLU activation functions and 1 3 x 3 convolutional layer without activation functions to output residual images.
5. The residual neural network and two-stage image demosaicing method based on hole convolution of claim 4, wherein the local residuals in the residual block are connected with input and output identities after a feature scaling technique is applied to all the local residual units.
6. The residual neural network and two-stage image demosaicing method based on hole convolution as claimed in claim 1, wherein in step 2, the data preprocessing process is as follows:
firstly, setting a G channel sampling pixel and a B channel sampling pixel in a mosaic image to be 0 to obtain an image only containing the R channel sampling pixel, and setting the G channel sampling pixel and the R channel sampling pixel in the mosaic image to be 0 to obtain an image only containing the B channel sampling pixel;
then, the mosaic image, the image only containing R-channel sampling pixels, the image only containing B-channel sampling pixels, the original R-channel image, the original G-channel image and the original B-channel image are respectively divided into a plurality of mosaic image blocks, image blocks only containing R-channel sampling pixels, image blocks only containing B-channel sampling pixels, original R-channel image blocks, original G-channel image blocks and original B-channel image blocks, wherein the number and the size of the mosaic image blocks, the image blocks only containing R-channel sampling pixels, the image blocks only containing B-channel sampling pixels, the original R-channel image blocks, the original G-channel image blocks and the original B-channel image blocks are the same.
7. The residual neural network and two-stage image demosaicing method based on hole convolution as claimed in claim 1, wherein in step 3, the first stage of the two-stage image demosaicing method is: training a neural network R by using a loss function R, wherein the neural network R constructs a mapping from a mosaic image, an image only containing R channel sampling pixels and an original G channel image to an R channel image; training a neural network G by using a loss function G, wherein the neural network G constructs a mapping from a mosaic image to a G channel image; training a neural network B by using a loss function B, wherein the neural network B constructs a mapping from a mosaic image, an image only containing B-channel sampling pixels and an original G-channel image to a B-channel image; the second stage of the two-stage image demosaicing method is as follows: and (3) jointly training a neural network R, a neural network G and a neural network B by using a loss function RGB, wherein the neural network R, the neural network G and the neural network B together construct a mapping from the mosaic image to the RGB color image.
8. The method for residual neural network and two-stage image demosaicing based on hole convolution of claim 1, wherein in the step 3, the loss function of the first stage comprises: a loss function R, a loss function G and a loss function B; the loss function of the second stage includes: the loss function RGB, and all of the loss functions take the form of a Mean Absolute Error (MAE) function:
Figure FDA0002506333520000031
Figure FDA0002506333520000032
Figure FDA0002506333520000033
Figure FDA0002506333520000034
wherein L isR(·)、LG(·)、LB(. and L)RGB(. The) represents a loss function R, a loss function G, a loss function B, and a loss function RGB, θ, respectivelyR、θGAnd thetaBRespectively representing parameters of a neural network R, parameters of a neural network G and parameters of a neural network B, wherein M represents the total number of image blocks; f. ofR(. -) represents a trained mapping from a mosaic image, an image containing only R-channel sample pixels, and an original G-channel image to an R-channel image, fG(. represents a trained mapping from mosaic images to G-channel images, fB(. table)Showing the trained mapping from mosaic image, image containing only B-channel sample pixels and original G-channel image to B-channel image, fR,G,B(. represents f)R(·),fG(. and f)B(ii) a combined mapping of; x is the number ofmos、xmos_rAnd xmos_bRespectively representing a mosaic image, an image containing only R-channel sample pixels and an image containing only B-channel sample pixels, xrgbRepresenting an RGB color image recovered by synthesizing an R channel image recovered by a neural network R, a G channel image recovered by a neural network G and a B channel image recovered by a neural network B; y isR、yG、yBAnd yRGBRepresenting an original R-channel image, an original G-channel image, an original B-channel image, and an original RGB color image, respectively.
9. The residual error neural network and two-stage image demosaicing method based on hole convolution of claim 1, wherein in step 3, during the training of the residual error neural network model based on hole convolution, the parameter θ in the residual error neural network based on hole convolution isR、θGAnd thetaBAdopts an Xavier method, namely that the mean value is 0 and the variance is Var (theta)R)、Var(θG) And Var (θ)B) Distribution of (a):
Figure FDA0002506333520000035
Figure FDA0002506333520000041
Figure FDA0002506333520000042
wherein the content of the first and second substances,
Figure FDA0002506333520000043
and
Figure FDA0002506333520000044
respectively representing the number of the input neurons of the layer in the neural network R, the number of the input neurons of the layer in the neural network G and the number of the input neurons of the layer in the neural network B,
Figure FDA0002506333520000045
and
Figure FDA0002506333520000046
respectively representing the number of the output neurons of the layer in the neural network R, the number of the output neurons of the layer in the neural network G and the number of the output neurons of the layer in the neural network B; the minimization loss function employs an Adam optimization method.
CN202010447460.4A 2020-05-25 2020-05-25 Residual error neural network based on cavity convolution and two-stage image demosaicing method Active CN111696036B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010447460.4A CN111696036B (en) 2020-05-25 2020-05-25 Residual error neural network based on cavity convolution and two-stage image demosaicing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010447460.4A CN111696036B (en) 2020-05-25 2020-05-25 Residual error neural network based on cavity convolution and two-stage image demosaicing method

Publications (2)

Publication Number Publication Date
CN111696036A true CN111696036A (en) 2020-09-22
CN111696036B CN111696036B (en) 2023-03-28

Family

ID=72478175

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010447460.4A Active CN111696036B (en) 2020-05-25 2020-05-25 Residual error neural network based on cavity convolution and two-stage image demosaicing method

Country Status (1)

Country Link
CN (1) CN111696036B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112488956A (en) * 2020-12-14 2021-03-12 南京信息工程大学 Method for image restoration based on WGAN network
CN112926692A (en) * 2021-04-09 2021-06-08 四川翼飞视科技有限公司 Target detection device and method based on non-uniform mixed convolution and storage medium
CN113076804A (en) * 2021-03-09 2021-07-06 武汉理工大学 Target detection method, device and system based on YOLOv4 improved algorithm
CN113850269A (en) * 2021-12-01 2021-12-28 西南石油大学 Method for nested connection of residual error network based on multi-branch selective kernel
CN114240776A (en) * 2021-12-12 2022-03-25 西北工业大学 Demosaic and compression fusion framework for MSFA hyperspectral image
CN114612299A (en) * 2022-02-17 2022-06-10 北京理工大学 Space self-adaptive real image demosaicing method and system
CN116128735A (en) * 2023-04-17 2023-05-16 中国工程物理研究院电子工程研究所 Multispectral image demosaicing structure and method based on densely connected residual error network
CN116503671A (en) * 2023-06-25 2023-07-28 电子科技大学 Image classification method based on residual network compression of effective rank tensor approximation
CN117939309A (en) * 2024-03-25 2024-04-26 荣耀终端有限公司 Image demosaicing method, electronic device and storage medium
CN114612299B (en) * 2022-02-17 2024-06-04 北京理工大学 Space self-adaptive real image demosaicing method and system

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107578392A (en) * 2017-09-25 2018-01-12 华北电力大学 A kind of convolutional neural networks demosaicing algorithms based on remaining interpolation
CN108492265A (en) * 2018-03-16 2018-09-04 西安电子科技大学 CFA image demosaicing based on GAN combines denoising method
CN108765295A (en) * 2018-06-12 2018-11-06 腾讯科技(深圳)有限公司 Image processing method, image processing apparatus and storage medium
US10235601B1 (en) * 2017-09-07 2019-03-19 7D Labs, Inc. Method for image analysis
CN109886875A (en) * 2019-01-31 2019-06-14 深圳市商汤科技有限公司 Image super-resolution rebuilding method and device, storage medium
CN110009590A (en) * 2019-04-12 2019-07-12 北京理工大学 A kind of high-quality colour image demosaicing methods based on convolutional neural networks
CN110120019A (en) * 2019-04-26 2019-08-13 电子科技大学 A kind of residual error neural network and image deblocking effect method based on feature enhancing
WO2019222951A1 (en) * 2018-05-24 2019-11-28 Nokia Technologies Oy Method and apparatus for computer vision
CN110706181A (en) * 2019-10-09 2020-01-17 中国科学技术大学 Image denoising method and system based on multi-scale expansion convolution residual error network
CN111047515A (en) * 2019-12-29 2020-04-21 兰州理工大学 Cavity convolution neural network image super-resolution reconstruction method based on attention mechanism

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10235601B1 (en) * 2017-09-07 2019-03-19 7D Labs, Inc. Method for image analysis
CN107578392A (en) * 2017-09-25 2018-01-12 华北电力大学 A kind of convolutional neural networks demosaicing algorithms based on remaining interpolation
CN108492265A (en) * 2018-03-16 2018-09-04 西安电子科技大学 CFA image demosaicing based on GAN combines denoising method
WO2019222951A1 (en) * 2018-05-24 2019-11-28 Nokia Technologies Oy Method and apparatus for computer vision
CN108765295A (en) * 2018-06-12 2018-11-06 腾讯科技(深圳)有限公司 Image processing method, image processing apparatus and storage medium
CN109886875A (en) * 2019-01-31 2019-06-14 深圳市商汤科技有限公司 Image super-resolution rebuilding method and device, storage medium
CN110009590A (en) * 2019-04-12 2019-07-12 北京理工大学 A kind of high-quality colour image demosaicing methods based on convolutional neural networks
CN110120019A (en) * 2019-04-26 2019-08-13 电子科技大学 A kind of residual error neural network and image deblocking effect method based on feature enhancing
CN110706181A (en) * 2019-10-09 2020-01-17 中国科学技术大学 Image denoising method and system based on multi-scale expansion convolution residual error network
CN111047515A (en) * 2019-12-29 2020-04-21 兰州理工大学 Cavity convolution neural network image super-resolution reconstruction method based on attention mechanism

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KAI CUI等: "COLOR IMAGE DEMOSAICKING USING A 3-STAGE CONVOLUTIONAL NEURAL NETWORK STRUCTURE", 《2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP)》 *
YAN WANG等: "Color Image Compression with Transform Domain Down-Sampling and Deep Convolutional Reconstruction", 《2019 IEEE VISUAL COMMUNICATIONS AND IMAGE PROCESSING(VCIP)》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112488956A (en) * 2020-12-14 2021-03-12 南京信息工程大学 Method for image restoration based on WGAN network
CN113076804A (en) * 2021-03-09 2021-07-06 武汉理工大学 Target detection method, device and system based on YOLOv4 improved algorithm
CN113076804B (en) * 2021-03-09 2022-06-17 武汉理工大学 Target detection method, device and system based on YOLOv4 improved algorithm
CN112926692A (en) * 2021-04-09 2021-06-08 四川翼飞视科技有限公司 Target detection device and method based on non-uniform mixed convolution and storage medium
CN112926692B (en) * 2021-04-09 2023-05-09 四川翼飞视科技有限公司 Target detection device, method and storage medium based on non-uniform mixed convolution
CN113850269A (en) * 2021-12-01 2021-12-28 西南石油大学 Method for nested connection of residual error network based on multi-branch selective kernel
CN113850269B (en) * 2021-12-01 2022-03-15 西南石油大学 Denoising method based on multi-branch selective kernel nested connection residual error network
CN114240776B (en) * 2021-12-12 2024-03-12 西北工业大学 Demosaicing and compression fusion framework for MSFA hyperspectral image
CN114240776A (en) * 2021-12-12 2022-03-25 西北工业大学 Demosaic and compression fusion framework for MSFA hyperspectral image
CN114612299A (en) * 2022-02-17 2022-06-10 北京理工大学 Space self-adaptive real image demosaicing method and system
CN114612299B (en) * 2022-02-17 2024-06-04 北京理工大学 Space self-adaptive real image demosaicing method and system
CN116128735A (en) * 2023-04-17 2023-05-16 中国工程物理研究院电子工程研究所 Multispectral image demosaicing structure and method based on densely connected residual error network
CN116503671A (en) * 2023-06-25 2023-07-28 电子科技大学 Image classification method based on residual network compression of effective rank tensor approximation
CN116503671B (en) * 2023-06-25 2023-08-29 电子科技大学 Image classification method based on residual network compression of effective rank tensor approximation
CN117939309A (en) * 2024-03-25 2024-04-26 荣耀终端有限公司 Image demosaicing method, electronic device and storage medium

Also Published As

Publication number Publication date
CN111696036B (en) 2023-03-28

Similar Documents

Publication Publication Date Title
CN111696036B (en) Residual error neural network based on cavity convolution and two-stage image demosaicing method
US8229212B2 (en) Interpolation system and method
CN110120019B (en) Residual error neural network based on feature enhancement and image deblocking method
US8717460B2 (en) Methods and systems for automatic white balance
CN102254301B (en) Demosaicing method for CFA (color filter array) images based on edge-direction interpolation
CN112804561A (en) Video frame insertion method and device, computer equipment and storage medium
CN110136060B (en) Image super-resolution reconstruction method based on shallow dense connection network
CN112308803B (en) Self-supervision low-illumination image enhancement and denoising method based on deep learning
CN101287130A (en) Apparatus and method for generating wide colour gamut signal in image capturing device
CN112465727A (en) Low-illumination image enhancement method without normal illumination reference based on HSV color space and Retinex theory
CN107169946B (en) Image fusion method based on nonnegative sparse matrix and hypersphere color transformation
CN111383200A (en) CFA image demosaicing method based on generative antagonistic neural network
CN112215767B (en) Anti-blocking effect image video enhancement method
CN106447632A (en) RAW image denoising method based on sparse representation
CN111405206B (en) Poisson-Gaussian combined noise image sequence separation noise reduction method
CN111583129A (en) Screen shot image moire removing method based on convolutional neural network AMNet
CN106709874B (en) Compressed low-resolution face image restoration method based on face structure correlation
CN113824945B (en) Rapid automatic white balance and color correction method based on deep learning
CN110728643A (en) Low-illumination band noise image optimization method based on convolutional neural network
CN112019704B (en) Video denoising method based on prior information and convolutional neural network
Guo et al. Joint demosaicking and denoising benefits from a two-stage training strategy
CN113129300A (en) Drainage pipeline defect detection method, device, equipment and medium for reducing false detection rate
CN100420271C (en) Device for eliminating flicker noise between frames in image sequence
CN111654705A (en) Mosaic image compression method based on novel color space conversion
Chen et al. A method to improve robustness of the gray world algorithm

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant