CN112819787A - Multi-light source prediction method - Google Patents

Multi-light source prediction method Download PDF

Info

Publication number
CN112819787A
CN112819787A CN202110136230.0A CN202110136230A CN112819787A CN 112819787 A CN112819787 A CN 112819787A CN 202110136230 A CN202110136230 A CN 202110136230A CN 112819787 A CN112819787 A CN 112819787A
Authority
CN
China
Prior art keywords
light source
network
light
prediction
value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110136230.0A
Other languages
Chinese (zh)
Other versions
CN112819787B (en
Inventor
董宇涵
邢晓岩
李志德
余澄
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen International Graduate School of Tsinghua University
Original Assignee
Shenzhen International Graduate School of Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen International Graduate School of Tsinghua University filed Critical Shenzhen International Graduate School of Tsinghua University
Priority to CN202110136230.0A priority Critical patent/CN112819787B/en
Publication of CN112819787A publication Critical patent/CN112819787A/en
Application granted granted Critical
Publication of CN112819787B publication Critical patent/CN112819787B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

A multi-light-source prediction method, comprising the steps of: decoupling the color and semantic features of the image by a feature extraction network to obtain a high-dimensional feature matrix; predicting the light source value of the principal light source by the principal light source prediction network through a high-dimensional feature matrix obtained by the feature extraction network; and the light source distribution weight map prediction network realizes the pixel-level light source distribution weight map prediction through a high-dimensional characteristic matrix obtained in the characteristic extraction network. The method overcomes the problem of constant color of multiple light sources when a digital camera images, can effectively complete the multiple light source prediction based on scene semantic information, has better adaptability to different illumination modes, and can realize accurate light source distribution and light source color information obtained from multiple light source pictures under various real conditions.

Description

Multi-light source prediction method
Technical Field
The invention relates to computational photography, in particular to a multi-light-source prediction method.
Background
Color Constancy (CC) is a classic problem faced by cameras during imaging, and human eyes can accurately restore the Color of an object when affected by light sources of different colors, which benefits from the prior knowledge of human brain on a specific environment, but for cameras, the Color of the object cannot be accurately restored when facing the interference of an ambient light source.
Common color constancy approaches can be divided into two categories, one is traditional statistical-based algorithms and the other is data-driven-based methods. The method has the advantages of high speed and small calculation amount, and has the defect that the method is extremely susceptible to the influence of a large-area pure color area in the face of the condition of lacking white points (such as a large-area pure color scene and the like), thereby causing poor adaptability. With the rise of deep learning in the field of computer vision, data-driven algorithms are also applied to Color Constancy, and mainly include Fast Fourier Color Constancy (FFCC) for learning frequency domain spectrum as the main and full-connected convolution network Color Constancy (FC) based on picture semantic information4) Compared with the traditional algorithm, the data-driven method has better adaptability in the face of complex scenes, for example, when the method faces large-area pure-color scenes, the data-driven method can often realize better light source prediction.
However, most of the current color constancy methods focus on the recovery of a single light source, while light sources which ignore common scenes are often multi-light sources, and although the conventional methods and the data-driven methods can realize the prediction of the multi-light sources by a clustering or partitioning method, the prediction means often ignore semantic information of pictures and cannot truly restore information of the multi-light source scenes.
Some researchers propose to realize multi-light source restoration from a picture to a picture by using a generated countermeasure network, but the method may affect the structural information of the picture, and a scene which does not belong to a real picture may be introduced in the countermeasure generation, so that the restoration result cannot be really applied to a mobile photographing terminal.
It is to be noted that the information disclosed in the above background section is only for understanding the background of the present application and thus may include information that does not constitute prior art known to a person of ordinary skill in the art.
Disclosure of Invention
The main objective of the present invention is to overcome the above-mentioned drawbacks of the background art, and to provide a multi-light-source prediction method.
In order to achieve the purpose, the invention adopts the following technical scheme:
a multi-light-source prediction method, comprising the steps of:
decoupling the color and semantic features of the image by a feature extraction network to obtain a high-dimensional feature matrix;
predicting the light source value of the principal light source by the principal light source prediction network through a high-dimensional feature matrix obtained by the feature extraction network;
and the light source distribution weight map prediction network realizes the pixel-level light source distribution weight map prediction through a high-dimensional characteristic matrix obtained in the characteristic extraction network.
Further:
the feature extraction network comprises a shallow semantic extraction branch, a deep semantic extraction branch and a color preference extraction branch, the shallow semantic extraction branch extracts shallow semantic information in an image through a small receptive field, the deep semantic extraction branch extracts deep semantic information in the image through a large receptive field so as to extract the incidence relation of different composition structures in the image, and the color preference extraction branch extracts color preference in the image so as to decouple color and semantic features.
One or more of the following settings are used:
the shallow semantic extraction branch comprises 5 convolutional layers and 4 pooling layers, all convolutional layers of the branch are convolution kernels of 3 x 3, and the step length is 2;
the deep semantic extraction branch adopts a front 5-layer network of AlexNet, the network comprises 5 convolutional layers and 3 pooling layers, convolution kernels of the front two convolutional layers are respectively 11 multiplied by 11 and 5 multiplied by 5, and convolution kernels of the rear three convolutional layers are respectively 3 multiplied by 3;
the color preference extraction branch adopts 5 convolutional layers and 4 pooling layers, and all convolutional layer convolutional kernels are 1 × 1.
The main light source prediction network comprises a light source position selection module and a light source regression module, wherein the light source position selection module is composed of convolution networks with the same number as that of light sources to be predicted, and the determination of the light source position is realized by adopting convolution layers with convolution kernel size of 1 multiplied by 1 and combining 2 times of pooling downsampling; the light source regression module adopts a full convolution network, and realizes the regression of the light source by extracting the characteristics of each channel after the light source position selection module determines the position of the light source.
The principal illuminant prediction network further comprises a channel attention module, and the channel attention module reweighs channels in the original high-dimensional feature matrix F to obtain a feature matrix G with different channel weight values.
The feature matrix G is calculated by equation (1):
Figure BDA0002926769840000031
where ω is the weight of the channel weighting, k is the dimension of the corresponding matrix channel, i is the pixel point of the corresponding characteristic map,
Figure BDA0002926769840000032
representing the multiplication of the channels.
The light source distribution weight graph prediction network comprises four upper sampling layers, and the four upper sampling layers are in short connection with the corresponding feature graph size and position in the feature extraction network.
Using predicted illuminant values
Figure BDA0002926769840000033
Angular error L from true nominal light source value EangularThe evaluation indexes were as follows:
Figure BDA0002926769840000034
wherein the predicted value
Figure BDA0002926769840000035
The light source value is actually a 3-dimensional matrix with the size of (1,1,3), which represents the size of R, G, B three-channel values in RGB space, and is calculated by using an inner product (·) with the real calibration value E;
to minimize the angular error L between the predicted and true valuesangularAs an optimization target and the angle error LangularThe loss function of the predicted network as a principal light source is used to optimize the iteration of the network.
Further, the mean square error MSE is also used as a loss function of the light source distribution weight map network as follows:
Figure BDA0002926769840000036
where N is the number of samples, i is the current sample number, xiFor the purpose of the current sample,
Figure BDA0002926769840000037
is the mean of all samples.
And multiplying the predicted light source value by a light source weight map to obtain light source distribution of a pixel level, and supervising the light source distribution by using the angle errors similarly, wherein the average value of the angle errors of all pixel points of the whole map is supervised:
Figure BDA0002926769840000038
wherein the content of the first and second substances,
Figure BDA0002926769840000039
for the predicted illuminant value of each pixel point i in the illuminant distribution map, N is the total number of full image pixels, EiThe real light source value of the point i; the final loss function L of the network is:
Figure BDA0002926769840000041
where k represents k candidate principal illuminant k e { 2.
A computer-readable storage medium, in which a computer program is stored which, when executed by a processor, carries out the method.
Compared with the prior art, the invention has the following beneficial effects:
the invention provides a multi-light source prediction method based on deep learning, aiming at the multi-light source estimation problem required by the multi-light source local white balance problem of a camera, the multi-light source prediction method can realize effective separation of main component light sources and distribution thereof from a multi-light source scene picture, and is a multi-light source estimation scheme with good application prospect.
The multi-light source prediction method based on the light source distribution weight graph solves the problem of constant multi-light source color during imaging of a digital camera, can effectively complete multi-light source prediction based on scene semantic information, has good adaptability to different illumination modes, and can realize accurate light source distribution and light source color information obtained from multi-light source pictures in various real conditions.
Drawings
Fig. 1 is a diagram of a deep neural network structure of a multi-light-source prediction method according to an embodiment of the present invention.
Fig. 2 is a light source distribution weight diagram and a different light source recovery diagram according to an embodiment of the invention.
Detailed Description
The embodiments of the present invention will be described in detail below. It should be emphasized that the following description is merely exemplary in nature and is not intended to limit the scope of the invention or its application.
The invention provides a multi-light-source prediction method based on a light source distribution probability map, which is used for realizing multi-light-source color constancy in the field of camera imaging. The core idea of the invention is to establish an end-to-end neural network, realize the prediction of the probability distribution diagram of the light source and the prediction of the corresponding light source, and realize the color constancy of the pixel level multi-light source by combining the two.
Referring to fig. 1 and fig. 2, the multi-light-source prediction method based on light source probability distribution according to the embodiment of the present invention mainly includes:
decoupling the color and semantic features of the image by a feature extraction network to obtain a high-dimensional feature matrix;
predicting the light source value of the principal light source by the principal light source prediction network through a high-dimensional feature matrix obtained by the feature extraction network;
and the light source distribution weight map prediction network realizes the pixel-level light source distribution weight map prediction through a high-dimensional characteristic matrix obtained in the characteristic extraction network.
The structure of the deep neural network used in the embodiment of the present invention is shown in fig. 1.
Feature extraction network
The feature extraction is important for the prediction of multiple light sources, so a brand-new feature extraction network is designed in the embodiment of the invention. The network comprises three branches; a shallow semantic extraction branch, a deep semantic extraction branch, and a color preference extraction branch.
The shallow semantic extraction branch comprises 5 convolutional layers and 4 pooling layers, all convolutional layers of the branch are convolution kernels of 3 x 3, and the step length is 2. The purpose of the layer is to extract shallow semantic information such as the outline of the picture content through a small receptive field.
The deep semantic extraction branch selects a first 5-layer network (Krizhevsky A, Sutskeeper I, Hinton G E. imaging classification with deep semantic network [ J ]. Communications of the ACM,2017,60(6):84-90) of AlexNet, the network comprises 5 convolutional layers and 3 pooling layers, convolution kernels of the first two layers are 11 × 11 and 5 × 5 respectively, convolution kernels of the last three layers are 3 × 3 respectively, and the first five layers of network introduced into AlexNet as the deep semantic extraction branch depends on a larger-size receptive field which can be provided by the deep semantic extraction branch, so that the neural network can better extract the incidence relation of different composition structures in the picture.
The color preference extraction branch also adopts 5 convolutional layers and 4 pooling layers, but all convolutional layer convolutional kernels of the branch are 1 × 1, so that the decoupling of color and semantic features is implicitly realized in order to avoid the interference of image content on feature extraction.
Principal light source prediction network
The purpose of the principal light source prediction network is to predict the value of the principal light source in the picture, and the network performs the prediction of the light source value through a high-dimensional feature matrix obtained by a feature extraction network. The prediction network consists of two parts: the device comprises a light source position selection module and a light source regression module. The light source position selection module is composed of convolution networks with the same number as the light sources to be predicted, the convolution layer with the convolution kernel size of 1 multiplied by 1 is adopted in the part, and the determination of the light source position is realized by combining 2 times of pooling downsampling; the light source regression module uses a full convolution network, and realizes the regression of the light source by extracting the characteristics of each channel after the selection module determines the position of the light source. In particular, in order to make the network prediction effect more accurate, in the preferred embodiment of the present invention, a channel attention module is added in the prediction network, and the channel attention module can re-weight the channels in the original feature matrix F to obtain the feature matrix G with different channel weight values, and the module can be calculated by formula (1).
Figure BDA0002926769840000051
Where ω is the weight of the channel weighting, k is the dimension of the corresponding matrix channel, i is the pixel point of the corresponding characteristic map,
Figure BDA0002926769840000061
representing the multiplication of the channels.
Pixel-level light source distribution weight map prediction network
In order to obtain more accurate light source distribution situation, the embodiment of the invention designs a light source weight prediction network at pixel level, which can realize the prediction of a weight map at pixel level through a high-dimensional feature matrix obtained in a feature extraction network, and such a prediction process can also be considered as a map-to-map change. The network mainly comprises four upper sampling layers and a feature extraction network part, wherein in order to keep the structure information of pictures as much as possible and reduce the loss caused by the increase of the number of network layers, short connection is carried out on the four upper sampling layers and the feature extraction network corresponding to the size and the position of a feature map.
Optimization and loss function
The preferred embodiment of the present invention uses predicted light source values
Figure BDA0002926769840000064
Angular error L from true nominal light source value EangularAs evaluation indexes, the following were used
Figure BDA0002926769840000065
Wherein the predicted value
Figure BDA0002926769840000066
The illuminant value is actually a 3-dimensional matrix of size (1,1,3) representing the size of the R, G, B three-channel value in RGB space, calculated using the inner product (-) with the true calibration value E。
In order to obtain the optimal prediction effect, the optimization goal of the multi-light-source prediction scheme of the preferred embodiment is to minimize the angle error L between the predicted value and the true valueangularAnd used as a loss function of the principal light source prediction network to optimize iteration of the network. Meanwhile, the preferred embodiment uses Mean Square Error (MSE) as a loss function for the light source distribution weight map network:
Figure BDA0002926769840000062
where N is the number of samples, i is the current sample number, xiFor the purpose of the current sample,
Figure BDA0002926769840000063
is the mean of all samples.
Further, multiplying the predicted light source value by the light source weight map can obtain the light source distribution at the pixel level, and the light source distribution is also supervised by using the angle error, and the actual supervision should be that the average value of the angle errors of all pixel points in the whole map is:
Figure BDA0002926769840000071
wherein the content of the first and second substances,
Figure BDA0002926769840000072
for the predicted illuminant value of each pixel point i in the illuminant distribution map, N is the total number of full image pixels, EiThe real light source value of the point i; the final loss function L of the network is:
Figure BDA0002926769840000073
where k represents k candidate principal illuminant k e { 2.
Performance analysis
The angular errors were used for analysis and the Mean (Mean), Median (Median) and tertile (Trimean) of the angular errors of all test data were used for calculation and comparison in the actual evaluation, as shown in table 1. Compared with the traditional algorithm Gray-World, the method provided by the embodiment of the invention has the advantages that the three indexes are all improved, and compared with 3du-awb and side-awb which are deep learning methods, the method provided by the invention has larger lead on each index. Meanwhile, compared with the existing color constancy algorithm, the method of the embodiment of the invention realizes the prediction of a plurality of principal light sources and the prediction of a pixel-level light source distribution diagram, and is a multi-light-source estimation scheme with good application prospect.
TABLE 1
Figure BDA0002926769840000074
The background of the present invention may contain background information related to the problem or environment of the present invention and does not necessarily describe the prior art. Accordingly, the inclusion in the background section is not an admission of prior art by the applicant.
The foregoing is a more detailed description of the invention in connection with specific/preferred embodiments and is not intended to limit the practice of the invention to those descriptions. It will be apparent to those skilled in the art that various substitutions and modifications can be made to the described embodiments without departing from the spirit of the invention, and these substitutions and modifications should be considered to fall within the scope of the invention. In the description herein, references to the description of the term "one embodiment," "some embodiments," "preferred embodiments," "an example," "a specific example," or "some examples" or the like are intended to mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above are not necessarily intended to refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Various embodiments or examples and features of various embodiments or examples described in this specification can be combined and combined by one skilled in the art without contradiction. Although embodiments of the present invention and their advantages have been described in detail, it should be understood that various changes, substitutions and alterations can be made herein without departing from the scope of the claims.

Claims (10)

1. A multi-light source prediction method is characterized by comprising the following steps:
decoupling the color and semantic features of the image by a feature extraction network to obtain a high-dimensional feature matrix;
predicting the light source value of the principal light source by the principal light source prediction network through a high-dimensional feature matrix obtained by the feature extraction network;
and the light source distribution weight map prediction network realizes the pixel-level light source distribution weight map prediction through a high-dimensional characteristic matrix obtained in the characteristic extraction network.
2. The multi-light-source prediction method of claim 1, wherein the feature extraction network comprises a shallow semantic extraction branch, a deep semantic extraction branch and a color preference extraction branch, the shallow semantic extraction branch extracts shallow semantic information in the image through a smaller receptive field, the deep semantic extraction branch extracts deep semantic information in the image through a larger receptive field so as to extract correlations of different constituent structures in the image, and the color preference extraction branch extracts color preferences in the image so as to achieve color and semantic feature decoupling.
3. The multi-light-source prediction method of claim 2, wherein one or more of the following settings are used:
the shallow semantic extraction branch comprises 5 convolutional layers and 4 pooling layers, all convolutional layers of the branch are convolution kernels of 3 x 3, and the step length is 2;
the deep semantic extraction branch adopts a front 5-layer network of AlexNet, the network comprises 5 convolutional layers and 3 pooling layers, convolution kernels of the front two convolutional layers are respectively 11 multiplied by 11 and 5 multiplied by 5, and convolution kernels of the rear three convolutional layers are respectively 3 multiplied by 3;
the color preference extraction branch adopts 5 convolutional layers and 4 pooling layers, and all convolutional layer convolutional kernels are 1 × 1.
4. The multi-light-source prediction method according to any one of claims 1 to 3, wherein the principal component light source prediction network comprises a light source position selection module and a light source regression module, the light source position selection module is composed of convolution networks with the same number as the light sources to be predicted, the determination of the light source positions is realized by adopting convolution layers with convolution kernel size of 1 x 1 and combining 2 times of pooling down-sampling; the light source regression module adopts a full convolution network, and realizes the regression of the light source by extracting the characteristics of each channel after the light source position selection module determines the position of the light source.
5. The multi-light-source prediction method of claim 4, wherein the principal light-source prediction network further comprises a channel attention module, and the channel attention module reweighs channels in the original high-dimensional feature matrix F to obtain a feature matrix G with different channel weight values.
6. The multi-light-source prediction method according to claim 5, wherein the feature matrix G is calculated by formula (1):
Figure FDA0002926769830000021
where ω is the weight of the channel weighting, k is the dimension of the corresponding matrix channel, i is the pixel point of the corresponding characteristic map,
Figure FDA0002926769830000022
representing the multiplication of the channels.
7. The multi-light-source prediction method of any one of claims 1 to 6, wherein the light source distribution weight map prediction network includes four upsampling layers, and the four upsampling layers are short-connected with corresponding feature map size positions in the feature extraction network.
8. The multi-light-source prediction method according to any one of claims 1 to 7, wherein a predicted light source value is used
Figure FDA0002926769830000023
Angular error L from true nominal light source value EangularThe evaluation indexes were as follows:
Figure FDA0002926769830000024
wherein the predicted value
Figure FDA0002926769830000025
The light source value is actually a 3-dimensional matrix with the size of (1,1,3), which represents the size of R, G, B three-channel values in RGB space, and is calculated by using an inner product (·) with the real calibration value E;
to minimize the angular error L between the predicted and true valuesangularAs an optimization target and the angle error LangularThe loss function of the predicted network as a principal light source is used to optimize the iteration of the network.
9. The multi-light-source prediction method of claim 8 further using mean square error, MSE, as a loss function of the light source distribution weight map network as follows:
Figure FDA0002926769830000026
where N is the number of samples, i is the current sample number, xiFor the purpose of the current sample,
Figure FDA0002926769830000027
is the mean of all samples;
preferably, the predetermined light source value is multiplied by a light source weight map to obtain a light source distribution at pixel level, which is also supervised using angular errors, and the average value of angular errors of all pixel points of the whole map is supervised:
Figure FDA0002926769830000031
wherein the content of the first and second substances,
Figure FDA0002926769830000032
for the predicted illuminant value of each pixel point i in the illuminant distribution map, N is the total number of full image pixels, EiThe real light source value of the point i; the final loss function L of the network is:
Figure FDA0002926769830000033
where k represents k candidate principal illuminant k e { 2.
10. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the method according to any one of claims 1 to 9.
CN202110136230.0A 2021-02-01 2021-02-01 Multi-light source prediction method Active CN112819787B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110136230.0A CN112819787B (en) 2021-02-01 2021-02-01 Multi-light source prediction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110136230.0A CN112819787B (en) 2021-02-01 2021-02-01 Multi-light source prediction method

Publications (2)

Publication Number Publication Date
CN112819787A true CN112819787A (en) 2021-05-18
CN112819787B CN112819787B (en) 2023-12-26

Family

ID=75860905

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110136230.0A Active CN112819787B (en) 2021-02-01 2021-02-01 Multi-light source prediction method

Country Status (1)

Country Link
CN (1) CN112819787B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6480299B1 (en) * 1997-11-25 2002-11-12 University Technology Corporation Color printer characterization using optimization theory and neural networks
US20060276966A1 (en) * 2002-07-30 2006-12-07 Cotton Symon D Mehtod and apparatus for quantifying tissue histology
US20170169276A1 (en) * 2011-09-27 2017-06-15 The Board Of Regents Of The University Of Texas System Systems and methods for automated screening and prognosis of cancer from whole-slide biopsy images
CN108388905A (en) * 2018-03-21 2018-08-10 合肥工业大学 A kind of Illuminant estimation method based on convolutional neural networks and neighbourhood context
CN111710049A (en) * 2020-06-18 2020-09-25 三星电子(中国)研发中心 Method and device for determining ambient illumination in AR scene
CN111813913A (en) * 2019-11-27 2020-10-23 上海交通大学 Two-stage problem generation system with problem as guide

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6480299B1 (en) * 1997-11-25 2002-11-12 University Technology Corporation Color printer characterization using optimization theory and neural networks
US20060276966A1 (en) * 2002-07-30 2006-12-07 Cotton Symon D Mehtod and apparatus for quantifying tissue histology
US20170169276A1 (en) * 2011-09-27 2017-06-15 The Board Of Regents Of The University Of Texas System Systems and methods for automated screening and prognosis of cancer from whole-slide biopsy images
CN108388905A (en) * 2018-03-21 2018-08-10 合肥工业大学 A kind of Illuminant estimation method based on convolutional neural networks and neighbourhood context
CN111813913A (en) * 2019-11-27 2020-10-23 上海交通大学 Two-stage problem generation system with problem as guide
CN111710049A (en) * 2020-06-18 2020-09-25 三星电子(中国)研发中心 Method and device for determining ambient illumination in AR scene

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
HUI YANG ET AL.: "Color Constancy of Rare-Earth Solution Image with Adaptive Parameter", 《COLOR CONSTANCY OF RARE-EARTH SOLUTION IMAGE WITH ADAPTIVE PARAMETER》 *
王飞: "图像颜色恒常性计算研究", 《中国博士学位论文全文数据库信息科技辑》, pages 33 - 84 *

Also Published As

Publication number Publication date
CN112819787B (en) 2023-12-26

Similar Documents

Publication Publication Date Title
WO2021164234A1 (en) Image processing method and image processing device
Cao et al. Underwater image restoration using deep networks to estimate background light and scene depth
CN110210608B (en) Low-illumination image enhancement method based on attention mechanism and multi-level feature fusion
CN111523521B (en) Remote sensing image classification method for double-branch fusion multi-scale attention neural network
Zhang et al. A new haze removal approach for sky/river alike scenes based on external and internal clues
US20190244362A1 (en) Differentiable Jaccard Loss Approximation for Training an Artificial Neural Network
Yang et al. Single image haze removal via region detection network
WO2022021999A1 (en) Image processing method and image processing apparatus
CN111160407B (en) Deep learning target detection method and system
CN110796009A (en) Method and system for detecting marine vessel based on multi-scale convolution neural network model
CN109034184B (en) Grading ring detection and identification method based on deep learning
Liu et al. Image de-hazing from the perspective of noise filtering
CN111429433A (en) Multi-exposure image fusion method based on attention generation countermeasure network
CN113518210A (en) Method and device for automatic white balance of image
CN112581379A (en) Image enhancement method and device
CN115223004A (en) Method for generating confrontation network image enhancement based on improved multi-scale fusion
CN113239825B (en) High-precision tobacco beetle detection method in complex scene
CN113554032A (en) Remote sensing image segmentation method based on multi-path parallel network of high perception
CN111914938A (en) Image attribute classification and identification method based on full convolution two-branch network
CN116757988A (en) Infrared and visible light image fusion method based on semantic enrichment and segmentation tasks
Cui et al. A novel underwater image restoration method based on decomposition network and physical imaging model
CN112861987A (en) Target detection method under dark light environment
CN116664446A (en) Lightweight dim light image enhancement method based on residual error dense block
CN112819787A (en) Multi-light source prediction method
Huang et al. An end-to-end dehazing network with transitional convolution layer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant