CN112085681A - Image enhancement method, system, device and storage medium based on deep learning - Google Patents

Image enhancement method, system, device and storage medium based on deep learning Download PDF

Info

Publication number
CN112085681A
CN112085681A CN202010941865.3A CN202010941865A CN112085681A CN 112085681 A CN112085681 A CN 112085681A CN 202010941865 A CN202010941865 A CN 202010941865A CN 112085681 A CN112085681 A CN 112085681A
Authority
CN
China
Prior art keywords
convolutional neural
image
neural network
training
image enhancement
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010941865.3A
Other languages
Chinese (zh)
Other versions
CN112085681B (en
Inventor
肖潇
晋兆龙
邹文艺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou Keda Technology Co Ltd
Original Assignee
Suzhou Keda Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Suzhou Keda Technology Co Ltd filed Critical Suzhou Keda Technology Co Ltd
Priority to CN202010941865.3A priority Critical patent/CN112085681B/en
Publication of CN112085681A publication Critical patent/CN112085681A/en
Priority to PCT/CN2021/082761 priority patent/WO2022052445A1/en
Application granted granted Critical
Publication of CN112085681B publication Critical patent/CN112085681B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/80Geometric correction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention provides an image enhancement method, a system, equipment and a storage medium based on deep learning, wherein the method comprises the following steps: collecting sample images for model training; training a plurality of first convolutional neural networks connected in parallel based on the sample image, wherein each first convolutional neural network corresponds to at least one enhancement item; training a second convolutional neural network, which is connected in series after the plurality of first convolutional neural networks connected in parallel; combining the trained first convolutional neural network and the trained second convolutional neural network to obtain an image enhancement model; and inputting the image to be enhanced into the image enhancement model to obtain an enhanced image output by the image enhancement model. By adopting the invention, the image enhancement of a plurality of enhancement items can be realized simultaneously by adopting the image enhancement model based on deep learning, and the image enhancement effect is improved.

Description

Image enhancement method, system, device and storage medium based on deep learning
Technical Field
The present invention relates to the field of data processing technologies, and in particular, to an image enhancement method, system, device, and storage medium based on deep learning.
Background
With the increasing complexity of camera imaging environments, based on the rapid development of deep learning and Artificial Intelligence (AI), the method can intelligently enhance the challenging imaging environments, and becomes an extremely important key part especially in important application scenes with high requirements on vehicles, human images, human faces and the like, thereby not only improving the solid foundation for subsequent analysis, but also enabling the captured images to better accord with the perception of human beings and improving the experience of customers.
The image enhancement method based on the traditional algorithm needs to design more independent modules, such as a demosaicing module, a denoising module, a white balance module and the like, and needs to carefully adjust each parameter based on scenes of various environments, so that a certain effect can be ensured, and the obtained enhanced image cannot necessarily reach a perfect state, and the subsequent analysis and the experience of customers are damaged.
Disclosure of Invention
In view of the problems in the prior art, an object of the present invention is to provide an image enhancement method, system, device and storage medium based on deep learning, which can simultaneously enhance images of multiple items and improve image optimization effect.
The embodiment of the invention provides an image enhancement method based on deep learning, which comprises the following steps:
collecting sample images for model training;
training a plurality of first convolutional neural networks connected in parallel based on the sample image, wherein each first convolutional neural network corresponds to at least one enhancement item;
training a second convolutional neural network, which is connected in series after the plurality of first convolutional neural networks connected in parallel;
combining the trained first convolutional neural network and the trained second convolutional neural network to obtain an image enhancement model;
and inputting the image to be enhanced into the image enhancement model to obtain an enhanced image output by the image enhancement model.
Optionally, training a plurality of first convolutional neural networks connected in parallel based on the sample image comprises the following steps:
training each of the first convolutional neural networks individually using the sample image as an input to the first convolutional neural network;
jointly training the plurality of first convolutional neural networks based on the sample images.
Optionally, jointly training the plurality of first convolutional neural networks comprises the following steps:
dividing the first convolutional neural network into a plurality of network groups;
performing joint training on a plurality of first convolutional neural networks in the same network group;
multiple network groups are jointly trained.
Optionally, the jointly training the plurality of first convolutional neural networks in the same network group includes the following steps:
inputting the sample images into each first convolutional neural network, and constructing a loss function of each first convolutional neural network based on an output image and a label image of each first convolutional neural network;
combining the loss functions of the first convolutional neural networks in the same network group to obtain the loss function of the network group;
iteratively training each first convolutional neural network in the network group based on the loss function of the network group;
the joint training of the plurality of network groups comprises the following steps:
inputting the sample images into each first convolution neural network, and constructing a loss function of each convolution neural network based on an output image and a label image of each first convolution neural network;
combining the loss functions of the first convolutional neural networks to obtain an overall loss function;
each convolutional neural network in each network group is iteratively trained based on the overall loss function.
Optionally, dividing the first convolutional neural network into a plurality of network groups, including the following steps: acquiring a label image of each first convolutional neural network;
comparing the color histogram of the label image of the first convolutional neural network with the color histogram of the corresponding sample image;
and grouping the first convolutional neural network according to the comparison result of the color histogram.
Optionally, training the second convolutional neural network comprises the steps of:
combining label images based on each first convolution neural network to obtain a first multichannel image;
and taking the first multichannel image as an input image of a second convolutional neural network, and iteratively training the second convolutional neural network.
Optionally, after iteratively training the second convolutional neural network by using the first multichannel image as an input image of the second convolutional neural network, the method further includes the following steps:
combining the output images of the first convolution neural networks to obtain a second multi-channel image;
and taking the second multichannel image as an input image of a second convolutional neural network, and iteratively training the second convolutional neural network.
By adopting the image enhancement method based on deep learning, the convolutional neural network based on deep learning replaces a manual design module in an image signal processing flow, a plurality of parallel first convolutional neural networks and a series second convolutional neural network are adopted to enhance the image, and image enhancement processing in multiple aspects is realized at the same time, so that the respective image enhancement effect of each convolutional neural network can be ensured, and the effect enhancement can be realized mutually, and the overall image enhancement effect is improved.
The embodiment of the invention also provides an image enhancement system based on deep learning, which is applied to the image enhancement method based on deep learning, and the system comprises:
the sample acquisition module is used for acquiring sample images for model training;
the first training module is used for training a plurality of first convolutional neural networks which are connected in parallel based on the sample image, and each first convolutional neural network corresponds to at least one enhancement item;
a second training module for training a second convolutional neural network, the second convolutional neural network being connected in series after the plurality of first convolutional neural networks connected in parallel;
the model combination module is used for combining the trained first convolutional neural network and the trained second convolutional neural network to obtain an image enhancement model;
and the image enhancement model is used for inputting the image to be enhanced into the image enhancement model to obtain an enhanced image output by the image enhancement model.
By adopting the image enhancement system based on deep learning, the convolutional neural network based on deep learning replaces a manual design module in an image signal processing flow, a plurality of parallel first convolutional neural networks and a series second convolutional neural network are respectively trained by the first training module and the second training module, the obtained image enhancement model can simultaneously realize image enhancement processing in multiple aspects, the respective image enhancement effect of each convolutional neural network can be ensured, the effect improvement can be realized mutually, and the overall image enhancement effect is improved.
The embodiment of the present invention further provides an image enhancement device based on deep learning, including: a processor;
a memory having stored therein executable instructions of the processor;
wherein the processor is configured to perform the steps of the deep learning based image enhancement method via execution of the executable instructions.
By adopting the image enhancement device based on deep learning provided by the invention, the processor executes the image enhancement method based on deep learning when executing the executable instruction, thereby obtaining the beneficial effects of the image enhancement method based on deep learning.
An embodiment of the present invention further provides a computer-readable storage medium for storing a program, where the program implements the steps of the image enhancement method based on deep learning when executed.
By adopting the computer-readable storage medium provided by the invention, the stored program realizes the steps of the image enhancement method based on the deep learning when being executed, thereby the beneficial effects of the image enhancement method based on the deep learning can be obtained.
Drawings
Other features, objects and advantages of the present invention will become more apparent upon reading of the following detailed description of non-limiting embodiments thereof, with reference to the accompanying drawings.
FIG. 1 is a flowchart of an image enhancement method based on deep learning according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a process after an image is input into an image enhancement model according to an embodiment of the present invention;
FIG. 3 is a flow chart of training of an image enhancement model according to an embodiment of the invention;
FIG. 4 is a schematic diagram of an image enhancement system based on deep learning according to an embodiment of the present invention;
FIG. 5 is a schematic structural diagram of an image enhancement apparatus based on deep learning according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of a computer storage medium according to an embodiment of the present invention.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art. The same reference numerals in the drawings denote the same or similar structures, and thus their repetitive description will be omitted.
As shown in fig. 1, in this embodiment, the image enhancement method based on deep learning includes the following steps:
s100: collecting sample images for model training;
the sample image can be a sample image for face recognition, a sample image for vehicle recognition of a user, and the like, a large number of original images (raw format images) in various different scenes can be collected as the sample image, for example, images in various challenging scenes are collected, the sample image is composed of original images of landscapes, traffic vehicles, figures, faces, and the like in various time periods, and the robustness of a network can be improved by using a large number of original image data;
s200: training a plurality of first convolutional neural networks connected in parallel based on the sample image, wherein each first convolutional neural network corresponds to at least one enhancement item;
s300: training a second convolutional neural network, which is connected in series after the plurality of first convolutional neural networks connected in parallel;
s400: combining the trained first convolutional neural network and the trained second convolutional neural network to obtain an image enhancement model;
s500: and inputting the image to be enhanced into the image enhancement model to obtain an enhanced image output by the image enhancement model.
By adopting the image enhancement method based on deep learning, a plurality of first convolutional neural networks connected in parallel are trained through the steps S100 and S200, a second convolutional neural network connected in series is trained through the step S300, the image enhancement model is obtained through the combination of the step S400, and the image to be enhanced is input into the image enhancement model through the step S500, so that the image enhancement in multiple aspects can be realized simultaneously, the respective image enhancement effect of each convolutional neural network can be ensured, the effect improvement can be realized mutually, and the overall image enhancement effect is improved.
As shown in fig. 2, a schematic diagram of the processing procedure after the image is input into the image enhancement model is shown, wherein the relationship of each convolutional neural network in the image enhancement model is shown. The first convolutional neural network and the enhancement item may have a one-to-one correspondence, for example, a first convolutional neural network is used for denoising, a first convolutional neural network is used for contrast enhancement, a first convolutional neural network is used for color enhancement, and the like. Therefore, the method can perform intelligent image enhancement processing by utilizing the deep convolution neural network based on the original image obtained by the camera photosensitive element under the condition of the challenging conditions of insufficient illumination, excessive noise, uneven exposure and the like, so that the processed image is objective and undistorted and can conform to the human perception system. Because each convolution neural network in the image enhancement model forms a complete neural network, when the image enhancement model is applied to image enhancement, the inference is carried out once without intermediate operation.
The step S100: after the sample images for model training are acquired, the number of the directly acquired sample images and related scenes are limited, so that the trained sample images can be appropriately augmented to increase the richness of data. The method comprises the steps of cutting the image at random positions, randomly turning the image horizontally and randomly turning the image vertically. The augmentations occur with independent probability and are cascaded, so that the diversity of training data can be greatly enriched, the robustness of the model is improved, and the phenomenon of overfitting is prevented.
After the sample image is obtained, label images of the sample image corresponding to the convolutional neural networks are further obtained. Specifically, the label image corresponding to the sample image can be obtained by labeling with tools such as Adobe and Photoshop, and the label image corresponding to the sample image is obtained for each of the first convolutional neural network and the second convolutional neural network. The label images of each first convolutional neural network are different from each other. For example, the label image of the first convolutional neural network with adjusted contrast is the label image obtained by adjusting the contrast of the sample image, the label image of the first convolutional neural network with enhanced color is the label image obtained by color enhancing the sample image, and the label image of the second convolutional neural network is the label image obtained by performing targeted adjustment on each enhanced item in the sample image.
As shown in fig. 3, in this embodiment, step S200: training a plurality of first convolutional neural networks connected in parallel based on the sample image comprises the following steps:
s210: training each of the first convolutional neural networks individually using the sample image as an input to the first convolutional neural network;
specifically, the sample image is input into each first convolution neural network to obtain an output image of each first convolution neural network, a loss function is constructed according to a label image and the output image of the first convolution neural network, and the first convolution neural network is iteratively trained according to the loss function;
s220: jointly training the plurality of first convolutional neural networks based on the sample images.
According to the invention, the parallel first convolution neural networks are distributed and jointly trained through the steps S210 and S220 respectively, so that the combined first convolution neural network can achieve a good image enhancement effect.
Specifically, in constructing each first convolutional neural network, the convolutional layer, the pooling layer, the activation function layer, and the batch normalization layer may adopt a hierarchical structure in the existing convolutional neural network. All data of the data layer may be further normalized to a data distribution of zero mean unit variance, which is normalized by the following formula:
Figure BDA0002673918850000071
wherein,
Figure BDA0002673918850000072
is the mean value, v is the variance, and the calculation formula is as follows:
Figure BDA0002673918850000073
wherein x isiIs the data to be normalized, x'iFor normalized data, n is the number of data.
In the invention, when each convolutional neural network is trained, the size of the predicted output image is consistent with that of the label image, the predicted output image and the label image are sent to a loss function layer to calculate loss, and the loss is iteratively minimized, so that the training purpose is achieved. The loss function used may be any of various existing loss functions. For example, in this embodiment, the loss function may be a smooth L1 loss function, which is calculated as follows:
Figure BDA0002673918850000074
the smooth L1 loss function is adopted, so that the simplicity, the efficiency and the practicability are high. The invention is not limited thereto and other types of loss functions are possible.
As shown in fig. 3, in this embodiment, the step S220: jointly training the plurality of first convolutional neural networks, comprising the steps of:
s221: the first convolutional neural network is divided into a plurality of network groups, for example, the first convolutional neural network 1, the first convolutional neural network 2, and the first convolutional neural network 3 shown in fig. 2 may be divided into one group, and the first convolutional neural network 4 to the first convolutional neural network n shown in fig. 2 may be divided into one group;
s222: the method comprises the following steps of performing combined training on a plurality of first convolutional neural networks in the same network group to realize that each first convolutional neural network combination in the same network group can well enhance an image;
s223: and performing joint training on a plurality of network groups to realize that all the first convolution neural network combinations can well enhance the image.
In this embodiment, the step S221: dividing the first convolutional neural network into a plurality of network groups, comprising the steps of:
acquiring a label image of each first convolutional neural network;
comparing the color histogram of the label image of the first convolutional neural network with the color histogram of the corresponding sample image;
and grouping the first convolution neural network according to the comparison result of the color histogram, namely grouping according to the change condition of the first convolution neural network to the color histogram.
Specifically, the first convolutional neural network may be divided into two groups: group1 and Group 2. The first convolutional neural network in Group1 is a network that does not change the color histogram of the image, e.g., the first convolutional neural network corresponding to denoising, the first convolutional neural network corresponding to super-resolution reconstruction, etc. Thus, the network in Group1 may focus primarily on the learning of image content, without concern for other aspects. The first convolutional neural network in Group2 is a network that can change the color histogram of the image, such as the first convolutional neural network corresponding to color enhancement, the first convolutional neural network corresponding to contrast enhancement, etc., and the label image of the network in Group2 mainly uses human subjective factors as criteria and better conforms to human visual perception. Thus, the network within Group2 may focus on fitting the perception of the human eye, enabling learning information that more closely conforms to the subjective perception of the human, such as hue, exposure, contrast, etc.
The method is only an example of the first convolution neural network grouping, and the network training purposes in the same network group have basic uniformity through the division of the color histogram, so that the difficulty of training each network in each group is naturally reduced, and the networks in the group have the effect of mutual promotion on the basis of the uniform purposes. In other alternative embodiments, the first convolutional neural network may also be grouped based on other factors, and the number of groups is not limited to two groups.
In this embodiment, the step S222: the joint training of a plurality of first convolutional neural networks in the same network group comprises the following steps:
inputting the sample images into each first convolutional neural network, and constructing a loss function of each first convolutional neural network based on an output image and a label image of each first convolutional neural network;
combining the loss functions of the first convolutional neural networks in the same network group to obtain the loss function of the network group, wherein the loss function can adopt smooth L1 loss function or other types of loss functions, and when in combination, the weight of the loss function of each first convolutional neural network in the loss function of the network group can be equal, is equal to the reciprocal of the number of the first convolutional neural networks in the network group, and can also be set as different weights, which all belong to the protection scope of the invention;
each first convolutional neural network in the network set is iteratively trained based on the loss functions of the network set.
The step S223: performing joint training on a plurality of network groups, comprising the following steps:
inputting the sample images into each first convolution neural network, and constructing a loss function of each convolution neural network based on an output image and a label image of each first convolution neural network;
combining the loss functions of the first convolutional neural networks to obtain an overall loss function, wherein the weight of the loss function of each first convolutional neural network in the overall loss function can be equal to the reciprocal of the number of all the first convolutional neural networks during combination, and can also be set as different weights, and the weights belong to the protection scope of the invention;
each convolutional neural network in each network group is iteratively trained based on the overall loss function.
Through the iterative training of step S222, the networks in each network group have substantially reached the purpose of designing the respective modules. In the iterative training in step S223, the iteration may use a small learning rate, which is only a process of adapting each network group, but cannot change the model parameters obtained by the training in step S222 too much.
As shown in fig. 3, in this embodiment, the step S300: training the second convolutional neural network comprises the following steps:
s310: combining label images based on each first convolution neural network to obtain a first multichannel image;
s320: and taking the first multichannel image as an input image of a second convolutional neural network, and iteratively training the second convolutional neural network, specifically, constructing a loss function according to an output image of the second convolutional neural network and a label image in iterative training, wherein the loss function may adopt a smooth L1 loss function or other types of loss functions.
The steps S310 and S320 first depend on the label image of the first convolutional neural network connected in parallel before, but not the predicted output image, to perform iterative training on the second convolutional neural network, so that the second convolutional neural network does not depend on the training result of the previous part and can not be interfered by the first convolutional neural network connected in parallel before.
In this embodiment, the step S320: after the first multichannel image is used as an input image of a second convolutional neural network and the second convolutional neural network is iteratively trained, the method further comprises the following steps:
s330: combining the output images of the first convolution neural networks to obtain a second multi-channel image;
s340: and taking the second multichannel image as an input image of a second convolutional neural network, and iteratively training the second convolutional neural network, specifically, constructing a loss function according to an output image of the second convolutional neural network and a label image in iterative training, where the loss function may be a smooth L1 loss function or other types of loss functions.
According to the invention, the previous part of label images are used as input for training through the steps S310 and S320, so that the training of the second convolutional neural network can be ensured in the early stage, the interference of the previous part of network is avoided, and the convergence of the current second convolutional neural network can be ensured and accelerated. And through step S330 and step S340, after the network of this part converges, the network parameters of the first convolutional neural network of the previous part of the robust are still needed in the later training period. Therefore, in the later training stage, the training of the second convolutional neural network is performed based on the output image predicted by the first convolutional neural network, and the iterative training can be performed at a small learning rate, so that the whole image enhancement model has good compatibility before and after.
In the invention, the iterative training of each convolutional neural network can be performed by uniformly adopting SGD (Stochastic Gradient Descent) and using a learning rate change mode with linear Descent. The training platform of the convolutional neural network can adopt caffe or tensorflow, etc. The sample data augmentation and special loss functions after image acquisition in the image enhancement method are added into a training platform, and the method can be used for training each convolutional neural network after compiling and debugging.
By adopting the image enhancement method based on deep learning, the image enhancement effect can be greatly improved, the method can be applied to intelligent image enhancement in different scenes and different environments, a powerful basis is provided for subsequent analysis, and the perception degree of human eyes is met.
As shown in fig. 4, an embodiment of the present invention further provides a deep learning based image enhancement system, which is applied to the deep learning based image enhancement method, and the system includes:
the sample acquisition module M100 is used for acquiring sample images for model training;
a first training module M200, configured to train, based on the sample image, a plurality of first convolutional neural networks connected in parallel, where each first convolutional neural network corresponds to at least one enhancement item;
a second training module M300, configured to train a second convolutional neural network, where the second convolutional neural network is connected in series after the plurality of parallel first convolutional neural networks;
the model combination module M400 is used for combining the trained first convolutional neural network and the trained second convolutional neural network to obtain an image enhancement model;
and the image enhancement model M500 is used for inputting the image to be enhanced into the image enhancement model to obtain an enhanced image output by the image enhancement model.
By adopting the image enhancement system based on deep learning, the sample acquisition module M100 and the first training module M200 train a plurality of first convolutional neural networks which are connected in parallel, the second training module M300 trains a second convolutional neural network which is connected in series, the image enhancement model obtained by combining the model combining module M400 is input into the image enhancement model through the image enhancement model M500, so that the image enhancement in multiple aspects can be realized simultaneously, the respective image enhancement effect of each convolutional neural network can be ensured, the effect enhancement can be realized mutually, and the overall image enhancement effect is improved.
In the image enhancement system based on deep learning, the functions of the modules can be realized by adopting the specific implementation of the steps in the image enhancement method based on deep learning. For example, the sample acquisition module M100 may acquire a sample image by using the specific implementation manner of step S100, and may further perform an augmentation process on the sample image, the first training module M200 may train each first convolutional neural network by using the specific implementation manner of step S200, the second training module M300 may train each second convolutional neural network by using the specific implementation manner of step S300, and the image enhancement model M500 may implement the application of image enhancement by using the specific implementation manner of step S500.
The embodiment of the invention also provides image enhancement equipment based on deep learning, which comprises a processor; a memory having stored therein executable instructions of the processor; wherein the processor is configured to perform the steps of the deep learning based image enhancement method via execution of the executable instructions.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or program product. Thus, various aspects of the invention may be embodied in the form of: an entirely hardware embodiment, an entirely software embodiment (including firmware, microcode, etc.) or an embodiment combining hardware and software aspects that may all generally be referred to herein as a "circuit," module "or" system.
An electronic device 600 according to this embodiment of the invention is described below with reference to fig. 5. The electronic device 600 shown in fig. 5 is only an example and should not bring any limitation to the functions and the scope of use of the embodiments of the present invention.
As shown in fig. 5, the electronic device 600 is embodied in the form of a general purpose computing device. The components of the electronic device 600 may include, but are not limited to: at least one processing unit 610, at least one storage unit 620, a bus 630 that connects the various system components (including the storage unit 620 and the processing unit 610), a display unit 640, and the like.
Wherein the storage unit stores program code executable by the processing unit 610 to cause the processing unit 610 to perform steps according to various exemplary embodiments of the present invention described in the above-mentioned electronic prescription flow processing method section of the present specification. For example, the processing unit 610 may perform the steps as shown in fig. 1.
The storage unit 620 may include readable media in the form of volatile memory units, such as a random access memory unit (RAM)6201 and/or a cache memory unit 6202, and may further include a read-only memory unit (ROM) 6203.
The memory unit 620 may also include a program/utility 6204 having a set (at least one) of program modules 6205, such program modules 6205 including, but not limited to: an operating system, one or more application programs, other program modules, and program data, each of which, or some combination thereof, may comprise an implementation of a network environment.
Bus 630 may be one or more of several types of bus structures, including a memory unit bus or memory unit controller, a peripheral bus, an accelerated graphics port, a processing unit, or a local bus using any of a variety of bus architectures.
The electronic device 600 may also communicate with one or more external devices 700 (e.g., keyboard, pointing device, bluetooth device, etc.), with one or more devices that enable a user to interact with the electronic device 600, and/or with any devices (e.g., router, modem, etc.) that enable the electronic device 600 to communicate with one or more other computing devices. Such communication may occur via an input/output (I/O) interface 650. Also, the electronic device 600 may communicate with one or more networks (e.g., a Local Area Network (LAN), a Wide Area Network (WAN), and/or a public network such as the Internet) via the network adapter 660. The network adapter 660 may communicate with other modules of the electronic device 600 via the bus 630. It should be appreciated that although not shown in the figures, other hardware and/or software modules may be used in conjunction with the electronic device 600, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems, among others.
By adopting the image enhancement device based on deep learning provided by the invention, the processor executes the image enhancement method based on deep learning when executing the executable instruction, thereby obtaining the beneficial effects of the image enhancement method based on deep learning.
An embodiment of the present invention further provides a computer-readable storage medium for storing a program, where the program implements the steps of the image enhancement method based on deep learning when executed. In some possible embodiments, aspects of the present invention may also be implemented in the form of a program product comprising program code for causing a terminal device to perform the steps according to various exemplary embodiments of the present invention described in the above-mentioned electronic prescription flow processing method section of this specification, when the program product is run on the terminal device.
Referring to fig. 6, a program product 800 for implementing the above method according to an embodiment of the present invention is described, which may employ a portable compact disc read only memory (CD-ROM) and include program code, and may be run on a terminal device, such as a personal computer. However, the program product of the present invention is not limited in this regard and, in the present document, a readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
The program product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. A readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium include: an electrical connection having one or more wires, a portable disk, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The computer readable storage medium may include a propagated data signal with readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A readable storage medium may also be any readable medium that is not a readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a readable storage medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computing device, partly on the user's device, as a stand-alone software package, partly on the user's computing device and partly on a remote computing device, or entirely on the remote computing device or cluster. In the case of a remote computing device, the remote computing device may be connected to the user computing device through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computing device (e.g., through the internet using an internet service provider).
By adopting the computer-readable storage medium provided by the invention, the stored program realizes the steps of the image enhancement method based on the deep learning when being executed, thereby the beneficial effects of the image enhancement method based on the deep learning can be obtained.
The foregoing is a more detailed description of the invention in connection with specific preferred embodiments and it is not intended that the invention be limited to these specific details. For those skilled in the art to which the invention pertains, several simple deductions or substitutions can be made without departing from the spirit of the invention, and all shall be considered as belonging to the protection scope of the invention.

Claims (10)

1. An image enhancement method based on deep learning is characterized by comprising the following steps:
collecting sample images for model training;
training a plurality of first convolutional neural networks connected in parallel based on the sample image, wherein each first convolutional neural network corresponds to at least one enhancement item;
training a second convolutional neural network, which is connected in series after the plurality of first convolutional neural networks connected in parallel;
combining the trained first convolutional neural network and the trained second convolutional neural network to obtain an image enhancement model;
and inputting the image to be enhanced into the image enhancement model to obtain an enhanced image output by the image enhancement model.
2. The deep learning-based image enhancement method according to claim 1, wherein training a plurality of first convolutional neural networks connected in parallel based on the sample image comprises the following steps:
training each of the first convolutional neural networks individually using the sample image as an input to the first convolutional neural network;
jointly training the plurality of first convolutional neural networks based on the sample images.
3. The deep learning based image enhancement method according to claim 2, wherein jointly training the plurality of first convolutional neural networks comprises the following steps:
dividing the first convolutional neural network into a plurality of network groups;
performing joint training on a plurality of first convolutional neural networks in the same network group;
multiple network groups are jointly trained.
4. The deep learning-based image enhancement method according to claim 3, wherein the joint training of the plurality of first convolutional neural networks in the same network group comprises the following steps:
inputting the sample images into each first convolutional neural network, and constructing a loss function of each first convolutional neural network based on an output image and a label image of each first convolutional neural network;
combining the loss functions of the first convolutional neural networks in the same network group to obtain the loss function of the network group;
iteratively training each first convolutional neural network in the network group based on the loss function of the network group;
the joint training of the plurality of network groups comprises the following steps:
inputting the sample images into each first convolution neural network, and constructing a loss function of each convolution neural network based on an output image and a label image of each first convolution neural network;
combining the loss functions of the first convolutional neural networks to obtain an overall loss function;
each convolutional neural network in each network group is iteratively trained based on the overall loss function.
5. The deep learning based image enhancement method according to claim 3, wherein the dividing the first convolutional neural network into a plurality of network groups comprises the following steps:
acquiring a label image of each first convolutional neural network;
comparing the color histogram of the label image of the first convolutional neural network with the color histogram of the corresponding sample image;
and grouping the first convolutional neural network according to the comparison result of the color histogram.
6. The deep learning based image enhancement method of claim 1, wherein training the second convolutional neural network comprises the steps of:
combining label images based on each first convolution neural network to obtain a first multichannel image;
and taking the first multichannel image as an input image of a second convolutional neural network, and iteratively training the second convolutional neural network.
7. The deep learning-based image enhancement method according to claim 6, wherein after iteratively training a second convolutional neural network with the first multi-channel image as an input image of the second convolutional neural network, the method further comprises the following steps:
combining the output images of the first convolution neural networks to obtain a second multi-channel image;
and taking the second multichannel image as an input image of a second convolutional neural network, and iteratively training the second convolutional neural network.
8. An image enhancement system based on deep learning, which is applied to the image enhancement method based on deep learning of any one of claims 1 to 7, the system comprising:
the sample acquisition module is used for acquiring sample images for model training;
the first training module is used for training a plurality of first convolutional neural networks which are connected in parallel based on the sample image, and each first convolutional neural network corresponds to at least one enhancement item;
a second training module for training a second convolutional neural network, the second convolutional neural network being connected in series after the plurality of first convolutional neural networks connected in parallel;
the model combination module is used for combining the trained first convolutional neural network and the trained second convolutional neural network to obtain an image enhancement model;
and the image enhancement model is used for inputting the image to be enhanced into the image enhancement model to obtain an enhanced image output by the image enhancement model.
9. An image enhancement device based on deep learning, characterized by comprising:
a processor;
a memory having stored therein executable instructions of the processor;
wherein the processor is configured to perform the steps of the deep learning based image enhancement method of any one of claims 1 to 7 via execution of the executable instructions.
10. A computer-readable storage medium storing a program, wherein the program is configured to implement the steps of the deep learning based image enhancement method according to any one of claims 1 to 7 when executed.
CN202010941865.3A 2020-09-09 2020-09-09 Image enhancement method, system, device and storage medium based on deep learning Active CN112085681B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202010941865.3A CN112085681B (en) 2020-09-09 2020-09-09 Image enhancement method, system, device and storage medium based on deep learning
PCT/CN2021/082761 WO2022052445A1 (en) 2020-09-09 2021-03-24 Deep-learning-based image enhancement method, system and device, and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010941865.3A CN112085681B (en) 2020-09-09 2020-09-09 Image enhancement method, system, device and storage medium based on deep learning

Publications (2)

Publication Number Publication Date
CN112085681A true CN112085681A (en) 2020-12-15
CN112085681B CN112085681B (en) 2023-04-07

Family

ID=73732207

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010941865.3A Active CN112085681B (en) 2020-09-09 2020-09-09 Image enhancement method, system, device and storage medium based on deep learning

Country Status (2)

Country Link
CN (1) CN112085681B (en)
WO (1) WO2022052445A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022052445A1 (en) * 2020-09-09 2022-03-17 苏州科达科技股份有限公司 Deep-learning-based image enhancement method, system and device, and storage medium
CN114782830A (en) * 2022-06-22 2022-07-22 天津大学 Image deep learning feature enhancement method, system and device
CN115496131A (en) * 2022-08-30 2022-12-20 北京华控智加科技有限公司 Equipment health state classification method based on multiple pre-training neural networks
WO2023061116A1 (en) * 2021-10-12 2023-04-20 腾讯科技(深圳)有限公司 Training method and apparatus for image processing network, computer device, and storage medium

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114677566B (en) * 2022-04-08 2023-10-17 北京百度网讯科技有限公司 Training method of deep learning model, object recognition method and device
CN116363009B (en) * 2023-03-31 2024-03-12 哈尔滨工业大学 Method and system for enhancing rapid light-weight low-illumination image based on supervised learning
CN117173503B (en) * 2023-08-16 2024-04-19 安徽派睿太赫兹医疗器械技术开发有限公司 Fuzzy terahertz image recognition method and device based on deep learning and electronic equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109583569A (en) * 2018-11-30 2019-04-05 中控智慧科技股份有限公司 A kind of multi-modal Feature fusion and device based on convolutional neural networks
CN111047532A (en) * 2019-12-06 2020-04-21 广东启迪图卫科技股份有限公司 Low-illumination video enhancement method based on 3D convolutional neural network
CN111415304A (en) * 2020-02-26 2020-07-14 中国农业大学 Underwater vision enhancement method and device based on cascade deep network

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10140544B1 (en) * 2018-04-02 2018-11-27 12 Sigma Technologies Enhanced convolutional neural network for image segmentation
CN110909578A (en) * 2018-09-18 2020-03-24 中国移动通信有限公司研究院 Low-resolution image recognition method and device and storage medium
CN109753938B (en) * 2019-01-10 2021-11-05 京东方科技集团股份有限公司 Image recognition method and device, application of image recognition method and device, and training method of neural network
CN110675336A (en) * 2019-08-29 2020-01-10 苏州千视通视觉科技股份有限公司 Low-illumination image enhancement method and device
CN111079761B (en) * 2019-11-05 2023-07-18 北京航空航天大学青岛研究院 Image processing method, device and computer storage medium
CN112085681B (en) * 2020-09-09 2023-04-07 苏州科达科技股份有限公司 Image enhancement method, system, device and storage medium based on deep learning

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109583569A (en) * 2018-11-30 2019-04-05 中控智慧科技股份有限公司 A kind of multi-modal Feature fusion and device based on convolutional neural networks
CN111047532A (en) * 2019-12-06 2020-04-21 广东启迪图卫科技股份有限公司 Low-illumination video enhancement method based on 3D convolutional neural network
CN111415304A (en) * 2020-02-26 2020-07-14 中国农业大学 Underwater vision enhancement method and device based on cascade deep network

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022052445A1 (en) * 2020-09-09 2022-03-17 苏州科达科技股份有限公司 Deep-learning-based image enhancement method, system and device, and storage medium
WO2023061116A1 (en) * 2021-10-12 2023-04-20 腾讯科技(深圳)有限公司 Training method and apparatus for image processing network, computer device, and storage medium
CN114782830A (en) * 2022-06-22 2022-07-22 天津大学 Image deep learning feature enhancement method, system and device
CN114782830B (en) * 2022-06-22 2022-09-06 天津大学 Image deep learning feature enhancement method, system and device
CN115496131A (en) * 2022-08-30 2022-12-20 北京华控智加科技有限公司 Equipment health state classification method based on multiple pre-training neural networks
CN115496131B (en) * 2022-08-30 2023-06-13 北京华控智加科技有限公司 Equipment health state classification method based on multiple pre-training neural networks

Also Published As

Publication number Publication date
CN112085681B (en) 2023-04-07
WO2022052445A1 (en) 2022-03-17

Similar Documents

Publication Publication Date Title
CN112085681B (en) Image enhancement method, system, device and storage medium based on deep learning
CN110163080B (en) Face key point detection method and device, storage medium and electronic equipment
WO2022247005A1 (en) Method and apparatus for identifying target object in image, electronic device and storage medium
US11069033B2 (en) Neural network evolution using expedited genetic algorithm for medical image denoising
CN109416758A (en) The method of neural network and neural metwork training
CN110738102A (en) face recognition method and system
CN111292262B (en) Image processing method, device, electronic equipment and storage medium
JP2008542911A (en) Image comparison by metric embedding
CN111292334B (en) Panoramic image segmentation method and device and electronic equipment
CN113012068B (en) Image denoising method, image denoising device, electronic equipment and computer-readable storage medium
CN116363261B (en) Training method of image editing model, image editing method and device
US20220156987A1 (en) Adaptive convolutions in neural networks
CN112419169B (en) CNN medical CT image denoising method based on noise priori
CN111160096A (en) Method, device and system for identifying poultry egg abnormality, storage medium and electronic device
WO2021139351A1 (en) Image segmentation method, apparatus, medium, and electronic device
CN115526803A (en) Non-uniform illumination image enhancement method, system, storage medium and device
CN110580726A (en) Dynamic convolution network-based face sketch generation model and method in natural scene
WO2024060839A9 (en) Object operation method and apparatus, computer device, and computer storage medium
CN117315369A (en) Fundus disease classification method and device based on neural network
CN116992941A (en) Convolutional neural network pruning method and device based on feature similarity and feature compensation
CN111724306A (en) Image reduction method and system based on convolutional neural network
KR20210038027A (en) Method for Training to Compress Neural Network and Method for Using Compressed Neural Network
CN114418064B (en) Target detection method, terminal equipment and storage medium
CN114240794A (en) Image processing method, system, device and storage medium
CN113688945A (en) Image processing hyper-parameter optimization method, system, device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant