CN113129236A - Single low-light image enhancement method and system based on Retinex and convolutional neural network - Google Patents

Single low-light image enhancement method and system based on Retinex and convolutional neural network Download PDF

Info

Publication number
CN113129236A
CN113129236A CN202110449727.8A CN202110449727A CN113129236A CN 113129236 A CN113129236 A CN 113129236A CN 202110449727 A CN202110449727 A CN 202110449727A CN 113129236 A CN113129236 A CN 113129236A
Authority
CN
China
Prior art keywords
component
image
neural network
illumination
retinex
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110449727.8A
Other languages
Chinese (zh)
Other versions
CN113129236B (en
Inventor
张强
王健
付飞琪
张凯
蒲亦非
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Petroleum East China
Original Assignee
China University of Petroleum East China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Petroleum East China filed Critical China University of Petroleum East China
Priority to CN202110449727.8A priority Critical patent/CN113129236B/en
Publication of CN113129236A publication Critical patent/CN113129236A/en
Application granted granted Critical
Publication of CN113129236B publication Critical patent/CN113129236B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the technical field of computer vision images, and provides a single low-illumination image enhancement method and system based on Retinex and a convolutional neural network. The method comprises the steps of obtaining an image, preprocessing the image, separating three channels, and obtaining a hue component, a saturation component and a brightness component; obtaining an illumination component according to the lightness component by adopting the trained deep convolution neural network model; calculating a reflection component of the illumination component by utilizing a Retinex theory; and recombining the reflection component with the hue component and the saturation component to obtain a three-channel image in the HSV color space.

Description

Single low-light image enhancement method and system based on Retinex and convolutional neural network
Technical Field
The invention belongs to the technical field of computer vision images, and particularly relates to a single low-illumination image enhancement method and system based on Retinex and a convolutional neural network.
Background
The statements in this section merely provide background information related to the present disclosure and may not necessarily constitute prior art.
With the development of science and technology, it is a common means to acquire information from images. Many computer vision algorithms such as object detection, recognition and tracking are becoming more common. However, not all images acquired by the sensors can be used directly for these algorithms. For example, the quality of the image acquired in a low-light environment is degraded, such as low visibility, low contrast, color distortion, high noise, and the like. The information that humans can get through these photographs is very little. The direct application of these photographs to computer vision algorithms can affect the performance of the algorithms. Therefore, how to improve the quality of images acquired under complex environments is a research focus in recent years in the field of computer vision.
A low-light image generally refers to an image of low quality that is acquired in a low-light environment. The main purpose of low-light image enhancement is to increase the brightness and contrast of a low-light image (or an underexposed image), highlight the main information of the image, and mainly realize the technology from the perspective of software. The enhanced image can be better used for computer vision tasks and can provide valuable information. The low-illumination image enhancement method has wide application prospects in the fields of monitoring, automatic driving and the like.
Traditional low-illumination image enhancement methods are mainly classified into four categories: (1) a method for enhancing images based on spatial domain. Such methods mainly change the distribution range of image pixel values to achieve image enhancement. Such methods are mostly based on histogram equalization, gamma correction and fuzzy logic transformations, etc., such as histogram equalization and limited contrast adaptive histogram equalization. (2) Transform domain based image enhancement methods, such methods mainly convert the image into the frequency domain where it is enhanced with a suitable filter function. Most of the methods are based on frequency domain and wavelet transform domain, and the used filter functions mainly include low-pass filter, band-pass filter, high-pass filter and the like. A representative method is a wavelet transform method. (3) An image enhancement method based on image fusion. The algorithm mainly realizes image enhancement in a mode of fusing a plurality of images. The method is based on a plurality of different images of the same scene, and combines parts with good visual effect in the different images to form a high-brightness image. Representative methods such as high dynamic lighting rendering. (4) An image enhancement method based on Retinex theory. The Retinex theory refers to that an image is formed by the joint action of illumination and an object, and the image can be expressed by the product of illumination components and reflection components, wherein the reflection components are inherent properties of the object and have consistent invariance under different illumination conditions. Most of these methods use the original image to obtain the reflection component (also the illumination component and the reflection component). Representative methods are single-scale Retinex using gaussian filtering of the illumination component of the image; the method comprises the following steps of performing Gaussian filtering on an image in multiple scales and increasing a color influence factor to obtain a multi-scale Retinex with color recovery; the illumination component is estimated using a structural prior, the reflection component is solved using Retinex theory and used as LIME of the final result, and so on.
The methods can achieve better effect on partial images, but are limited by models, lack of generalization capability and difficult to be applied to wider scenes.
Convolutional neural networks have great advantages in processing image tasks, have been applied to various types of computer vision tasks, and have been with great success. In the field of low-light image processing, there are also many methods based on convolutional neural networks. LLNet establishes an autoencoder to enhance the image; msr-net realizes image enhancement by learning the mapping relation between light and dark images; Retinex-Net establishes a decomposition network and an enhancement network, trains the network using a paired image dataset, and the like. The method has strong generalization capability and can adapt to a plurality of scenes, but the effect of the method has a great relationship with a paired data set used for training a network model, the establishment of the bright and dark image to the data set is a difficult work, and the brightness of a real image (namely a normal image) does not have a specific standard.
Disclosure of Invention
In order to solve the technical problems in the background art, the invention provides a method and a system for enhancing a single low-light image based on Retinex and a convolutional neural network, which can effectively enhance the single low-light image, do not generate color distortion, can keep the texture details of an original image, and have good generalization capability on different data sets.
In order to achieve the purpose, the invention adopts the following technical scheme:
the invention provides a single low-light image enhancement method based on Retinex and a convolutional neural network.
A single low-light image enhancement method based on Retinex and a convolutional neural network comprises the following steps:
acquiring an image, preprocessing the image, separating three channels and acquiring a hue component, a saturation component and a brightness component;
obtaining an illumination component according to the lightness component by adopting the trained deep convolution neural network model;
calculating a reflection component of the illumination component by utilizing a Retinex theory;
and recombining the reflection component with the hue component and the saturation component to obtain a three-channel image in the HSV color space.
And converting the three-channel image in the HSV color space into an RGB color space to obtain an enhanced low-illumination image.
Further, the process of obtaining the enhanced low-light image comprises: and converting the three-channel image in the HSV color space into an RGB color space, and adjusting the pixel range of the three-channel image from [0,1] to [0,255] to obtain the enhanced low-illumination image.
Further, the process of obtaining the hue component, the saturation component and the brightness component includes: normalizing the pixels of the image to [0,1], converting the normalized image from a color space RGB to an HSV space, and separating three channels to obtain hue components, saturation and brightness components.
Further, the process of training the deep convolutional neural network model includes:
constructing a deep convolutional neural network model, and establishing a target loss function of the neural network based on Retinex theory and prior hypothesis;
obtaining a logarithm brightness component after the brightness component is subjected to logarithm and normalization, and obtaining a bright channel prior component through the brightness component;
and (3) bringing the illumination component, the lightness component and the bright channel prior component into a target loss function, calculating an error, and realizing gradient updating on the weight and the parameters of the neural network by using an Adam optimization algorithm through the target loss function until the error is smaller than a set threshold or iteration reaches a preset number of times, and ending the model training.
The target loss function is:
E=Lossis1Lossr2Lossrs3Losslc
therein, LossisRepresenting the Loss function of illumination smoothness, LossrRepresenting the reflection Loss function, LossrsRepresenting Loss of reflection smoothness function, LosslcRepresenting the prior loss function of the bright channel, λ123Respectively, reflection loss, reflection component smoothness loss, and bright channel prior loss.
As an embodiment, three weight values may be taken: lambda [ alpha ]1=0.05,λ2=0.1,λ30.5. It should be noted that the numerical values of the three weight values are only one embodiment of the present invention, and should not be construed as limiting the present invention.
The second aspect of the invention provides a single low-light image enhancement system based on Retinex and a convolutional neural network.
A single low-light image enhancement system based on Retinex and a convolutional neural network comprises:
an acquisition and pre-processing module configured to: acquiring an image, preprocessing the image, separating three channels and acquiring a hue component, a saturation component and a brightness component;
an illumination component obtaining module configured to: obtaining an illumination component according to the lightness component by adopting the trained deep convolution neural network model;
a reflected component obtaining module configured to: calculating a reflection component of the illumination component by utilizing a Retinex theory;
a reassembly module configured to: and recombining the reflection component with the hue component and the saturation component to obtain a three-channel image in the HSV color space.
An output module configured to: and converting the three-channel image in the HSV color space into an RGB color space to obtain an enhanced low-illumination image.
A third aspect of the invention provides a computer-readable storage medium.
A computer-readable storage medium, on which a computer program is stored, which when executed by a processor implements the steps in the Retinex and convolutional neural network-based single-low-illumination image enhancement method as described in the first aspect above.
A fourth aspect of the invention provides a computer apparatus.
A computer device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, the processor implementing the steps of the Retinex and convolutional neural network-based single low-light image enhancement method according to the first aspect when executing the program.
Compared with the prior art, the invention has the beneficial effects that:
1. according to the invention, the low-illumination image is converted from the color space RGB to the color space HSV, the channels are separated, only the brightness component is enhanced, the color information is kept in the hue (H) and the saturation (S), and the color distortion in the image enhancing process is avoided.
2. The invention establishes the objective function of the deep learning neural network through Retinex theory and a series of prior, realizes the estimation of illumination component from the brightness component of a single image, further estimates the reflection component, thereby obtaining the enhanced image, completes the unsupervised image enhancement process, and has strong generalization capability.
3. The invention uses fractional order differentiation in the objective function of the neural network, and improves the retention capacity of the enhanced image details.
Advantages of additional aspects of the invention will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, are included to provide a further understanding of the invention, and are incorporated in and constitute a part of this specification, illustrate exemplary embodiments of the invention and together with the description serve to explain the invention and not to limit the invention.
FIG. 1 is a schematic flow chart of a single low-light image enhancement method based on Retinex and a convolutional neural network;
FIG. 2 is a diagram of a deep convolutional network architecture in an embodiment of the present invention;
FIG. 3 is a schematic illustration of an object imaging process of the present invention;
FIG. 4(a) is an input low-light image;
FIG. 4(b) is a visualization effect diagram after the image enhancement processing of the DICM data set of the BIMEF algorithm;
FIG. 4(c) is a visualization effect diagram after the image enhancement processing of the DICM data set of the DONG algorithm;
FIG. 4(d) is a visualization effect diagram after the image enhancement processing of the DICM data set of the LIME algorithm;
FIG. 4(e) is a visualization effect diagram after the image enhancement processing of the DICM data set by the MF algorithm;
FIG. 4(f) is a visualization effect diagram after the image enhancement processing of the DICM data set of the NPE algorithm;
FIG. 4(g) is a visualization effect diagram after SIRE algorithm DICM data set image enhancement processing;
FIG. 4(h) is a visualization effect diagram after ULE algorithm DICM data set image enhancement processing;
FIG. 4(i) is a visualization effect diagram after the image enhancement processing of the Retinex-Net algorithm DICM data set;
FIG. 4(j) is a visualization effect diagram after image enhancement processing of the KinD algorithm DICM data set;
fig. 4(k) is a visualization effect diagram after the DICM dataset image enhancement processing according to the method of the present invention.
Detailed Description
The invention is further described with reference to the following figures and examples.
It is to be understood that the following detailed description is exemplary and is intended to provide further explanation of the invention as claimed. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of exemplary embodiments according to the invention. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
Example one
As shown in fig. 1, the embodiment provides a single low-light image enhancement method based on Retinex and convolutional neural network, and the embodiment is illustrated by applying the method to a server, it is understood that the method may also be applied to a terminal, and may also be applied to a system including a terminal and a server, and is implemented by interaction between the terminal and the server. The server may be an independent physical server, a server cluster or a distributed system formed by a plurality of physical servers, or a cloud server providing basic cloud computing services such as a cloud service, a cloud database, cloud computing, a cloud function, cloud storage, a network server, cloud communication, middleware service, a domain name service, a security service CDN, a big data and artificial intelligence platform, and the like. The terminal may be, but is not limited to, a smart phone, a tablet computer, a laptop computer, a desktop computer, a smart speaker, a smart watch, and the like. The terminal and the server may be directly or indirectly connected through wired or wireless communication, and the application is not limited herein. In this embodiment, the method includes the steps of:
step 1: acquiring an image, preprocessing the image, separating three channels and acquiring a hue component, a saturation component and a brightness component;
the method comprises the following steps of preprocessing the acquired low-illumination image:
step 1.1, in the stage of reading in the low-illumination image, the pixel value range of the read image is changed into [0,1] by using a special read-in function; or create a normalization function to normalize pixels of the low-light image from [0,255] to [0,1 ].
Step 1.2: converting the image obtained in the step 1.1 from a color space RGB to a color space HSV, and extracting three channel components: hue (H), saturation (S), lightness (V). The specific implementation method comprises the following steps:
step 1.2.1: calculating the maximum channel (c) of the imagemax) Minimum channel (c)min) And a contrast (Δ), the calculation formula being: c. Cmax=max(R,G,B),cmin=min(R,G,B),Δ=cmax-cmin(wherein R, G, B are the three channels of the color space RGB, with values at [0,1]]In (d) of (a);
step 1.2.2: the hue (H), saturation (S), lightness (V) are calculated according to the following formula:
Figure BDA0003038197660000081
Figure BDA0003038197660000082
V=cmax
step 1.3: the luminance component is logarithmized using a function y ═ log (x +1), and the result is normalized to obtain a logarithmized luminance component (v).
Step 1.4: calculating to obtain a prior component (v) of a bright channel by a formulalight) The formula is as follows:
Figure BDA0003038197660000091
where Ω is a k × k region centered at (i, j),
Figure BDA0003038197660000092
and vp,qRepresenting the pixel at the specified location.
Step 2: obtaining an illumination component according to the lightness component by adopting the trained deep convolution neural network model;
the method comprises the following specific steps of constructing a deep convolutional neural network model:
step 2.1: and constructing a deep convolutional network. The network structure of the invention is shown in fig. 2, the input is the lightness component (v) obtained in step 1.3, the output is the illumination component (l), the first 4 layers of the network model are composed of a convolution operation and a ReLU function, the latter layer only has 1 convolution operation, and finally, a sigmoid layer is connected. The specific information of each layer of the network model is as follows:
TABLE 1 deep convolutional network
Figure BDA0003038197660000093
Step 2.2: and establishing an objective function of the neural network based on Retinex theory and prior hypothesis. Fig. 3 shows the imaging process of an object, which the Retinex theory shows as:
Figure BDA0003038197660000101
where S represents the acquired (or observed) image, R is the reflected component of the object, L is the illumination component of the environment,
Figure BDA0003038197660000102
representing multiplication by element. Under the assumption that the spatial variation of the illumination is smooth, a loss of illumination smoothness is proposed:
Figure BDA0003038197660000103
wherein l is the illumination component output by the network,
Figure BDA0003038197660000104
is v is1(being positive) derivative of order, v2Is a positive number and N is the total number of pixels of the image. By applying constraint to the fractional order gradient of the illumination component, the illumination component obtained by the neural network is smooth when the space changes.
In order to avoid too large difference between the obtained illumination image and the original image, the reflection loss is designed:
Figure BDA0003038197660000105
where l is the illumination component of the network output, v is the lightness component of the input, v is the luminance component of the network output2Is a positive number.
In order to make the spatial variation of the reflection component as smooth as possible, to guarantee its clarity and visual effect, a loss of reflection smoothness is proposed:
Figure BDA0003038197660000106
in order to have the value of the reflection component (R) between [0,1], it is necessary to guarantee that the illumination component (L) is greater than (V), thus proposing a bright channel prior penalty:
Figure BDA0003038197660000107
wherein v islightIs the bright channel prior component calculated from the luma component v,
in the above loss, the present invention uses fractional order gradient and fractional order differential to improve the retention of image texture details by the model.
And synthesizing the loss functions of all parts, wherein the target loss function of the model is as follows:
E=Lossis1Lossr2Lossrs3Losslc
wherein λ is123The weights are respectively reflection loss, reflection component smoothness loss and bright channel prior loss, and the values of the three weights are respectively: lambda [ alpha ]1=0.05,λ2=0.1,λ30.5. The resulting illumination components can be spatially smoothed with 4 lossy constraints, preserving texture details.
Step 2.3: randomly initializing a network weight, and determining a weight optimization algorithm as an Adam optimization algorithm.
Wherein a neural network is trained. Inputting the lightness component (v) obtained in the step 1.3 into a convolution neural network, and obtaining an illumination component (l), a lightness component (v) and a bright channel prior component (v) by the neural networklight) And (3) carrying the error into the objective function in the step 2.2, calculating the error, and realizing gradient updating on the weight and the parameters of the neural network by using an Adam optimization algorithm through the objective function, wherein the iteration is stopped when the error meets the requirement or the iteration reaches a preset number.
And step 3: calculating a reflection component of the illumination component by utilizing a Retinex theory;
step 3.1: and (3) inputting the brightness component (v) obtained in the step (3.1) into the deep convolutional neural network trained in the step (3) to obtain an output illumination component (l).
Step 3.2: and (3) gamma correction is carried out on the illumination component obtained in the step (3.1), and then the lightness component (V) and the illumination component (L) obtained in the step (1.3) are subjected to indexing processing to obtain a lightness component (V) and an illumination component (L) after indexing. Obtaining a calculation formula of the reflection component according to Retinex theory: and R is V/L. The reflection component (R) is calculated by a calculation formula.
And 4, step 4: recombining the reflection component with the hue component and the saturation component to obtain a three-channel image in an HSV color space; and converting the three-channel image in the HSV color space into an RGB color space to obtain an enhanced low-illumination image.
Specifically, the reflection component (R) is used as lightness component (V), and combined with hue (H) and saturation (S) obtained in step 1.2 to restore an image enhanced in color space HSV, and then the image is converted from color space HSV to color space RGB, and then the image value range is changed from [0,1] to [0,255] by a special read-out function or a created mapping function.
Through the steps, the enhanced image can be obtained.
We performed tests on DICM low-light image datasets and evaluated network performance using peak Signal-to-Noise ratio psnr (peak Signal to Noise ratio) and structural similarity ssim (structural similarity index). Meanwhile, visualization and quantitative comparison are carried out with the current advanced algorithm, including a natural color retention enhancement algorithm NPE, a DONG based on defogging, an MF based on fusion, a LIME based on illumination component estimation, a BIMEF based on illumination component estimation and multi-exposure fusion, an SRIE based on reflection component and illumination component estimation, and an ULE, Retinex-Net and KinD based on deep learning.
The environment of the experiment is CPU processor Intel (R) Xeon (R) CPU E5-2620 v3@2.40GHz, memory 128G, video card Nvidia GeForce GTX TITAN X, and video memory 12G. The software system is Windows10, the deep learning framework of the pytorech, python3.8, CUDA version 10.0, cuDNN version 7.4. The development software used was Pycharm2020 and Matlab2020 a.
The learning rate of the parameter setting of this experiment is 0.001, and the total iteration number is 1000.
Fig. 4(a) - (k) illustrate the visual comparison results of different algorithms on a DICM dataset. Fig. 4(a) is an input low-light image, fig. 4(b) -fig. 4(j) are results of other methods, and fig. 4(k) is a method proposed by the present invention. Fig. 4 (b): BIMEF, fig. 4 (d): LIME, fig. 4 (e): MF, fig. 4 (f): NPE, fig. 4 (h): ULE and fig. 4 (i): the Retinex-Net method over-enhances the image and the enhanced image suffers from color distortion. Fig. 4 (c): DONG and fig. 4 (g): the consequences of SIRE are unnatural and in some places, excessive enhancement occurs. As can be seen from the image fig. 4(k), the method of the present embodiment retains the image details, makes the image more natural, and does not generate color distortion and excessive enhancement. By contrast, the method of the invention has better overall effect representation, and fine color and structure can be better recovered.
The quantitative test results on the DICM dataset are as follows:
TABLE 2 quantitative comparison on DICM datasets
Figure BDA0003038197660000131
Example two
The embodiment provides a single low-light image enhancement system based on Retinex and a convolutional neural network.
A single low-light image enhancement system based on Retinex and a convolutional neural network comprises:
an acquisition and pre-processing module configured to: acquiring an image, preprocessing the image, separating three channels and acquiring a hue component, a saturation component and a brightness component;
an illumination component obtaining module configured to: obtaining an illumination component according to the lightness component by adopting the trained deep convolution neural network model;
a reflected component obtaining module configured to: calculating a reflection component of the illumination component by utilizing a Retinex theory;
a reassembly module configured to: and recombining the reflection component with the hue component and the saturation component to obtain a three-channel image in the HSV color space.
An output module configured to: and converting the three-channel image in the HSV color space into an RGB color space to obtain an enhanced low-illumination image.
EXAMPLE III
The present embodiment provides a computer-readable storage medium, on which a computer program is stored, which when executed by a processor implements the steps in the single-low-illumination image enhancement method based on Retinex and convolutional neural network as described in the first embodiment above.
Example four
The embodiment provides a computer device, which includes a memory, a processor, and a computer program stored in the memory and executable on the processor, and the processor executes the program to implement the steps in the single-low-light image enhancement method based on Retinex and convolutional neural network as described in the first embodiment.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of a hardware embodiment, a software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, optical storage, and the like) having computer-usable program code embodied therein.
The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), or the like.
The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (10)

1. A single low-light image enhancement method based on Retinex and a convolutional neural network is characterized by comprising the following steps:
acquiring an image, preprocessing the image, separating three channels and acquiring a hue component, a saturation component and a brightness component;
obtaining an illumination component according to the lightness component by adopting the trained deep convolution neural network model;
calculating a reflection component of the illumination component by utilizing a Retinex theory;
and recombining the reflection component with the hue component and the saturation component to obtain a three-channel image in the HSV color space.
2. The single-sheet low-illumination image enhancement method based on Retinex and the convolutional neural network as claimed in claim 1, wherein the three-channel image in HSV color space is converted into RGB color space, resulting in an enhanced low-illumination image.
3. The single low-light image enhancement method based on Retinex and convolutional neural network of claim 2, wherein the process of obtaining the enhanced low-light image comprises: and converting the three-channel image in the HSV color space into an RGB color space, and adjusting the pixel range of the three-channel image from [0,1] to [0,255] to obtain the enhanced low-illumination image.
4. The single low-light image enhancement method based on Retinex and convolutional neural network of claim 1, wherein the process of obtaining hue component, saturation component and brightness component comprises: normalizing the pixels of the image to [0,1], converting the normalized image from a color space RGB to an HSV space, and separating three channels to obtain hue components, saturation and brightness components.
5. The single-low-light image enhancement method based on Retinex and the convolutional neural network as claimed in claim 1, wherein the process of deep convolutional neural network model training comprises:
constructing a deep convolutional neural network model, and establishing a target loss function of the neural network based on Retinex theory and prior hypothesis;
obtaining a logarithm brightness component after the brightness component is subjected to logarithm and normalization, and obtaining a bright channel prior component through the brightness component;
and (3) bringing the illumination component, the lightness component and the bright channel prior component into a target loss function, calculating an error, and realizing gradient updating on the weight and the parameters of the neural network by using an Adam optimization algorithm through the target loss function until the error is smaller than a set threshold or iteration reaches a preset number of times, and ending the model training.
6. The single-low-illumination image enhancement method based on Retinex and convolutional neural network of claim 5, wherein the target loss function is:
E=Lossis1Lossr2Lossrs3Losslc
therein, LossisRepresenting the Loss function of illumination smoothness, LossrRepresenting the reflection Loss function, LossrsRepresenting Loss of reflection smoothness function, LosslcRepresenting the prior loss function of the bright channel, λ123The weights are respectively reflection loss, reflection component smoothness loss and bright channel prior loss, and the values of the three weights are respectively: lambda [ alpha ]1=0.05,λ2=0.1,λ3=0.5。
7. The single-sheet low-illumination image enhancement method based on Retinex and the convolutional neural network as claimed in claim 1, wherein the deep convolutional neural network model comprises five convolutional layers and one sigmoid activation layer.
8. A single low-light image enhancement system based on Retinex and a convolutional neural network is characterized by comprising:
an acquisition and pre-processing module configured to: acquiring an image, preprocessing the image, separating three channels and acquiring a hue component, a saturation component and a brightness component;
an illumination component obtaining module configured to: obtaining an illumination component according to the lightness component by adopting the trained deep convolution neural network model;
a reflected component obtaining module configured to: calculating a reflection component of the illumination component by utilizing a Retinex theory;
a reassembly module configured to: and recombining the reflection component with the hue component and the saturation component to obtain a three-channel image in the HSV color space.
An output module configured to: and converting the three-channel image in the HSV color space into an RGB color space to obtain an enhanced low-illumination image.
9. A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, carries out the steps of the method for single-low-illumination image enhancement based on Retinex and convolutional neural network according to any of claims 1 to 7.
10. A computer device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, wherein the processor implements the steps in the Retinex and convolutional neural network based single low-light image enhancement method of any of claims 1-7 when executing the program.
CN202110449727.8A 2021-04-25 2021-04-25 Single low-light image enhancement method and system based on Retinex and convolutional neural network Active CN113129236B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110449727.8A CN113129236B (en) 2021-04-25 2021-04-25 Single low-light image enhancement method and system based on Retinex and convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110449727.8A CN113129236B (en) 2021-04-25 2021-04-25 Single low-light image enhancement method and system based on Retinex and convolutional neural network

Publications (2)

Publication Number Publication Date
CN113129236A true CN113129236A (en) 2021-07-16
CN113129236B CN113129236B (en) 2022-07-12

Family

ID=76779830

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110449727.8A Active CN113129236B (en) 2021-04-25 2021-04-25 Single low-light image enhancement method and system based on Retinex and convolutional neural network

Country Status (1)

Country Link
CN (1) CN113129236B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114202475A (en) * 2021-11-24 2022-03-18 北京理工大学 Adaptive image enhancement method and system
CN114266705A (en) * 2021-11-04 2022-04-01 山东师范大学 Low-light image enhancement method and system
CN114283288A (en) * 2021-12-24 2022-04-05 合肥工业大学智能制造技术研究院 Method, system, equipment and storage medium for enhancing night vehicle image
CN114862737A (en) * 2022-06-02 2022-08-05 大连海事大学 Underwater image enhancement method based on image reflection component and deep learning model
CN116824511A (en) * 2023-08-03 2023-09-29 行为科技(北京)有限公司 Tool identification method and device based on deep learning and color space
CN116993636A (en) * 2023-07-10 2023-11-03 中国地质大学(武汉) Image enhancement method and device for underground low-illumination deep stratum empty area
CN117853783A (en) * 2023-12-12 2024-04-09 济南大学 Single board defect identification method and system based on deep learning

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080101719A1 (en) * 2006-10-30 2008-05-01 Samsung Electronics Co., Ltd. Image enhancement method and system
CN106600564A (en) * 2016-12-23 2017-04-26 潘敏 Novel image enhancement method
CN110298796A (en) * 2019-05-22 2019-10-01 中山大学 Based on the enhancement method of low-illumination image for improving Retinex and Logarithmic image processing
WO2020001164A1 (en) * 2018-06-28 2020-01-02 Boe Technology Group Co., Ltd. Image enhancement method and apparatus
CN110930341A (en) * 2019-10-17 2020-03-27 杭州电子科技大学 Low-illumination image enhancement method based on image fusion
CN112465727A (en) * 2020-12-07 2021-03-09 北京邮电大学 Low-illumination image enhancement method without normal illumination reference based on HSV color space and Retinex theory

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080101719A1 (en) * 2006-10-30 2008-05-01 Samsung Electronics Co., Ltd. Image enhancement method and system
CN106600564A (en) * 2016-12-23 2017-04-26 潘敏 Novel image enhancement method
WO2020001164A1 (en) * 2018-06-28 2020-01-02 Boe Technology Group Co., Ltd. Image enhancement method and apparatus
CN110298796A (en) * 2019-05-22 2019-10-01 中山大学 Based on the enhancement method of low-illumination image for improving Retinex and Logarithmic image processing
CN110930341A (en) * 2019-10-17 2020-03-27 杭州电子科技大学 Low-illumination image enhancement method based on image fusion
CN112465727A (en) * 2020-12-07 2021-03-09 北京邮电大学 Low-illumination image enhancement method without normal illumination reference based on HSV color space and Retinex theory

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LIYUN ZHUANG ET.AL: "Image Enhancement by Deep Learning Network Based on derived image and Retinex", 《2019 IEEE 3RD ADVANCED INFORMATION MANAGEMENT,COMMUNICATES,ELECTRONIC AND AUTOMATION CONTROL CONFERENCE》 *
张红颖 等: "空间的RetinexNet低照度图像增强算法", 《激光与光电子学进展》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114266705A (en) * 2021-11-04 2022-04-01 山东师范大学 Low-light image enhancement method and system
CN114202475A (en) * 2021-11-24 2022-03-18 北京理工大学 Adaptive image enhancement method and system
CN114283288A (en) * 2021-12-24 2022-04-05 合肥工业大学智能制造技术研究院 Method, system, equipment and storage medium for enhancing night vehicle image
CN114862737A (en) * 2022-06-02 2022-08-05 大连海事大学 Underwater image enhancement method based on image reflection component and deep learning model
CN116993636A (en) * 2023-07-10 2023-11-03 中国地质大学(武汉) Image enhancement method and device for underground low-illumination deep stratum empty area
CN116993636B (en) * 2023-07-10 2024-02-13 中国地质大学(武汉) Image enhancement method and device for underground low-illumination deep stratum empty area
CN116824511A (en) * 2023-08-03 2023-09-29 行为科技(北京)有限公司 Tool identification method and device based on deep learning and color space
CN117853783A (en) * 2023-12-12 2024-04-09 济南大学 Single board defect identification method and system based on deep learning

Also Published As

Publication number Publication date
CN113129236B (en) 2022-07-12

Similar Documents

Publication Publication Date Title
CN113129236B (en) Single low-light image enhancement method and system based on Retinex and convolutional neural network
Lv et al. Attention guided low-light image enhancement with a large scale low-light simulation dataset
US20240062530A1 (en) Deep perceptual image enhancement
Liu et al. Survey of natural image enhancement techniques: Classification, evaluation, challenges, and perspectives
CN111079764B (en) Low-illumination license plate image recognition method and device based on deep learning
Zhou et al. Multi-scale retinex-based adaptive gray-scale transformation method for underwater image enhancement
Guo et al. Image dehazing via enhancement, restoration, and fusion: A survey
Wang et al. Joint iterative color correction and dehazing for underwater image enhancement
Steffens et al. Cnn based image restoration: Adjusting ill-exposed srgb images in post-processing
Fan et al. Multi-scale depth information fusion network for image dehazing
CN114581318B (en) Low-illumination image enhancement method and system
CN114529709A (en) Method and system for training convolutional neural network
CN116157805A (en) Camera image or video processing pipeline using neural embedding
Wen et al. Autonomous robot navigation using Retinex algorithm for multiscale image adaptability in low-light environment
Lei et al. Low-light image enhancement using the cell vibration model
Wang et al. Single Underwater Image Enhancement Based on $ L_ {P} $-Norm Decomposition
Saleem et al. A non-reference evaluation of underwater image enhancement methods using a new underwater image dataset
CN115797205A (en) Unsupervised single image enhancement method and system based on Retinex fractional order variation network
Zhou et al. Sparse representation with enhanced nonlocal self-similarity for image denoising
Chen et al. High-dynamic range, night vision, image-fusion algorithm based on a decomposition convolution neural network
CN117974459A (en) Low-illumination image enhancement method integrating physical model and priori
CN113628143A (en) Weighted fusion image defogging method and device based on multi-scale convolution
CN117391987A (en) Dim light image processing method based on multi-stage joint enhancement mechanism
Singh et al. Multiscale reflection component based weakly illuminated nighttime image enhancement
Soumya et al. Self-organized night video enhancement for surveillance systems

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant