WO2021082088A1 - 一种色调映射方法、装置及电子设备 - Google Patents
一种色调映射方法、装置及电子设备 Download PDFInfo
- Publication number
- WO2021082088A1 WO2021082088A1 PCT/CN2019/118585 CN2019118585W WO2021082088A1 WO 2021082088 A1 WO2021082088 A1 WO 2021082088A1 CN 2019118585 W CN2019118585 W CN 2019118585W WO 2021082088 A1 WO2021082088 A1 WO 2021082088A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- component
- dynamic range
- range image
- high dynamic
- network
- Prior art date
Links
- 238000013507 mapping Methods 0.000 title claims abstract description 69
- 238000000034 method Methods 0.000 title claims abstract description 52
- 238000013528 artificial neural network Methods 0.000 claims abstract description 20
- 238000000354 decomposition reaction Methods 0.000 claims description 29
- 238000006243 chemical reaction Methods 0.000 claims description 23
- 230000006870 function Effects 0.000 claims description 20
- 230000004913 activation Effects 0.000 claims description 14
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 10
- 230000004927 fusion Effects 0.000 claims description 7
- 238000004590 computer program Methods 0.000 claims description 6
- 238000012549 training Methods 0.000 claims description 6
- 238000010606 normalization Methods 0.000 claims description 5
- 230000004075 alteration Effects 0.000 abstract description 6
- 238000012545 processing Methods 0.000 description 21
- 238000005516 engineering process Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 11
- 238000013135 deep learning Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 5
- 230000009286 beneficial effect Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 4
- 238000001914 filtration Methods 0.000 description 4
- 239000011159 matrix material Substances 0.000 description 3
- 230000000717 retained effect Effects 0.000 description 3
- 238000011161 development Methods 0.000 description 2
- 125000001475 halogen functional group Chemical group 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 230000000593 degrading effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000001575 pathological effect Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/92—Dynamic range modification of images or parts thereof based on global image properties
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/04—Context-preserving transformations, e.g. by using an importance map
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/60—Image enhancement or restoration using machine learning, e.g. neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/088—Non-supervised learning, e.g. competitive learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20208—High dynamic range [HDR] image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Definitions
- This specification relates to the field of digital image processing technology, and in particular to a tone mapping method, device and electronic equipment.
- High Dynamic Range With the rapid development of High Dynamic Range (HDR) technology, various high dynamic range videos, images and other content are increasing. Compared with ordinary dynamic range images, high dynamic range images can provide more dynamic range and Image details, so high dynamic range images can better restore the visual effects in the real environment. However, since most multimedia devices still display images with limited dynamic range (ie low dynamic range), high dynamic range images cannot be displayed normally on such multimedia devices, so how to display high dynamic range images on such devices Normal display, that is, tone mapping technology has become a more important technology in the field of digital image processing.
- tone mapping is limited by the bit depth of multimedia devices and other conditions, it is impossible to reproduce high dynamic range images on multimedia devices completely, so how to compress the dynamic range while retaining as many local details as possible, that is, restore as much as possible High dynamic range images have become the focus of research.
- a high dynamic range image is divided into a basic layer and a detail layer through a filter.
- the basic layer contains low-frequency information such as brightness of the image, and the detail layer contains high-frequency information such as image edges.
- the basic layer is compressed, and the detail layer is compressed. It is enhanced and finally merged into a low dynamic range image.
- the filtering process will introduce noises such as halos and artifacts, and these noises will have a serious impact on the result of tone mapping, easily cause color difference and reduce the naturalness of the image.
- the existing tone mapping method cannot robustly complete the high Conversion of dynamic range image to low dynamic range image.
- the purpose of the present invention is to provide a tone mapping method, device and electronic device to solve the problems of the prior art that the tone mapping will produce chromatic aberration and the conversion is not robust enough.
- An embodiment of the present specification provides a tone mapping method, the method includes:
- the storage form of the high dynamic range image is a predetermined storage form, performing a decomposition operation on the high dynamic range image to obtain the first component, the second component, and the third component of the high dynamic range image;
- the first component and the second component after the mapping are merged with the third component to obtain a merged low dynamic range image corresponding to the high dynamic range image, so as to complete tone mapping.
- the method before performing the decomposition operation on the high dynamic range image, the method further includes:
- the storage format of the high dynamic range image is a non-predetermined storage format
- a conversion operation is performed on the high dynamic range image to convert it into a high dynamic range image in a predetermined storage format, and the converted Perform a decomposition operation on the high dynamic range image.
- the predetermined storage form includes an HSV color space
- the performing a decomposition operation on the high dynamic range image to obtain the first component, the second component, and the third component of the high dynamic range image includes:
- the components in the HSV color space corresponding to the high dynamic range image are extracted to obtain the first component, the second component, and the third component; wherein the first component includes saturation information, and the first component includes saturation information.
- the second component includes brightness information, and the third component includes hue information.
- the predetermined deep neural network is a generative adversarial network
- the generative adversarial network includes a generative network and a discriminant network, wherein:
- the generation network is established based on the U-Net network, the generation network includes an encoder and a decoder, the encoder includes at least one convolution block and a plurality of residual blocks, and the decoder includes a plurality of deconvolutions.
- the discriminant network includes a plurality of convolutional blocks, and each convolutional block includes a convolutional layer, a normalization layer, and an activation layer arranged in sequence.
- the generative confrontation network is obtained by training a predetermined loss function, and the loss function includes one or more of the generative confrontation loss function, the mean square error function, and the multi-scale structural similarity loss function.
- the fusing the mapped first component and the second component with the third component to obtain the fused low dynamic range image corresponding to the high dynamic range image includes:
- the first component and the second component after the mapping are superimposed with the third component to obtain a low dynamic range image conforming to a predetermined storage format.
- the method further includes:
- a conversion operation is performed on the low dynamic range image, so as to convert it into a low dynamic range image corresponding to the RGB color space.
- An embodiment of the present specification provides a tone mapping device, the device includes:
- the acquisition module is used to acquire one or more high dynamic range images and judge the storage form of the high dynamic range images
- the decomposition module is used to perform a decomposition operation on the high dynamic range image when it is determined that the storage form of the high dynamic range image is a predetermined storage form to obtain the first component, the second component, and the high dynamic range image.
- Third component
- the mapping module is used to input the first component and the second component into a predetermined deep neural network, and use the deep neural network to map the first component and the second component respectively to obtain the mapped first component The first component and the second component;
- a fusion module for fusing the mapped first component and second component with the third component to obtain a fused low dynamic range image corresponding to the high dynamic range image, so as to complete tone mapping .
- the device further includes:
- the first conversion module is configured to perform a conversion operation on the high dynamic range image when it is determined that the storage form of the high dynamic range image is a non-predetermined storage form before performing the decomposition operation on the high dynamic range image, In order to convert it into a high dynamic range image in a predetermined storage form, and perform a decomposition operation on the converted high dynamic range image.
- the predetermined storage form includes an HSV color space
- the decomposition module is specifically configured to:
- the components in the HSV color space corresponding to the high dynamic range image are extracted to obtain the first component, the second component, and the third component; wherein the first component includes saturation information, and the first component includes saturation information.
- the second component includes brightness information, and the third component includes hue information.
- the fusion module is specifically used for:
- the first component and the second component after the mapping are superimposed with the third component to obtain a low dynamic range image conforming to a predetermined storage format.
- the device further includes:
- the second conversion module is configured to perform a conversion operation on the low dynamic range image after the low dynamic range image conforming to the predetermined storage format is obtained, so as to convert it into a low dynamic range image corresponding to the RGB color space.
- An electronic device provided by an embodiment of this specification includes a memory, a processor, and a computer program stored in the memory and capable of running on the processor, and the processor implements the above-mentioned tone mapping method when the program is executed.
- the present invention obtains one or more high dynamic range images and determines the storage form of the high dynamic range image.
- the storage form of the high dynamic range image is a predetermined storage form
- the high dynamic range image is decomposed into the first component and the second component.
- Two components and a third component input the first component and the second component into a predetermined deep neural network, and use the deep neural network to map the first component and the second component respectively to obtain the mapped first component and the second component
- Two components; the mapped first component and the second component and the third component are fused to obtain a low dynamic range image corresponding to the high dynamic range image after fusion, so as to complete tone mapping.
- FIG. 1 is a schematic flowchart of a tone mapping method provided by an embodiment of this specification
- FIG. 2 is a schematic flowchart of tone mapping using a generative adversarial network in a specific application scenario provided by an embodiment of this specification;
- Fig. 3 is a schematic structural diagram of a tone mapping device provided by an embodiment of this specification.
- High dynamic range (HDR) technology As one of the important branches in the field of image processing technology, high dynamic range (HDR) technology has also risen, and various high dynamic range videos and images are increasing. . High dynamic range images can be considered to provide more dynamic range and detailed images than ordinary dynamic range images. Therefore, high dynamic range images can better restore the visual effects in the real environment. Dynamic range is the ratio of the highest luminance in the scene and the lowest luminance, in practical applications, the dynamic range of the image may be more than 105 is considered a high dynamic range image.
- Tone mapping refers to a computer graphics technology that approximately displays high dynamic range images on a limited dynamic range medium.
- the limited dynamic range medium includes LCD display devices, projection devices, and so on. Because tone mapping is a pathological problem, limited by the bit depth of multimedia devices and other conditions, it is impossible to reproduce high dynamic range images on multimedia devices completely, so how to compress the dynamic range while retaining as much local details as possible Therefore, it has become the focus of research to restore high dynamic range images as much as possible.
- a high dynamic range image is divided into a basic layer and a detail layer through a filter.
- the basic layer contains low-frequency information such as brightness of the image
- the detail layer contains high-frequency information such as image edges. Layers are enhanced, and finally merged into a low dynamic range image.
- this existing processing method has many drawbacks. For example, the filtering process will introduce noises such as halos and artifacts. These noises are difficult to eliminate, and the noise will have a serious impact on the result of tone mapping, easily causing chromatic aberration, and degrading the image. Naturalness.
- the existing deep learning method is based on direct tone mapping in the RGB color space, so the color difference problem is still unavoidable; in addition, the existing deep learning method
- the tone-mapped image obtained by the traditional filtering method is still used as the label for deep learning training, but the low dynamic range image obtained by the traditional filtering method has a relatively large color difference, resulting in the quality of the image label used for deep learning training The overall situation is poor, so it is difficult to learn a high-quality tone-mapped image.
- the following embodiments of this specification are performed on high dynamic range images as the processing object.
- the embodiments of this specification do not limit the storage format of high dynamic range images.
- the storage format can be the high dynamic range of the RGB color space.
- the image is the processing object, and the high dynamic range image in the RGB color space is only an embodiment in the actual application scenario of this specification, and does not constitute a limitation on the application scope of the embodiment of this specification.
- FIG. 1 is a schematic flowchart of a tone mapping method provided by an embodiment of this specification. The method may specifically include the following steps:
- step S110 one or more high dynamic range images are acquired, and the storage form of the high dynamic range images is judged.
- the high dynamic range image can be regarded as the object of tone mapping processing. Therefore, acquiring one or more high dynamic range images can be regarded as acquiring one or more original processing objects or Target image.
- the original processing object in the embodiment of this specification can be a high dynamic range image stored in any storage form.
- the storage form of the high dynamic range image includes but is not limited to: RGB, HSV, CMY, CMYK , YIQ, Lab and other color spaces (or called color spaces).
- the storage form of different color spaces can be considered as using different matrices and color variables, so the high dynamic range image can be analyzed by The matrix structure or color judges the storage form of the high dynamic range image.
- the HSV color space its spatial matrix structure is a hexagonal pyramid model, and the color of an image is described by hue, saturation, and brightness.
- step S120 when it is determined that the storage format of the high dynamic range image is a predetermined storage format, a decomposition operation is performed on the high dynamic range image to obtain the first component, the second component, and the high dynamic range image.
- the third component is a decomposition operation performed on the high dynamic range image to obtain the first component, the second component, and the high dynamic range image.
- the next step is determined according to the judgment result, which may specifically include the following situations:
- Case 1 When it is determined that the storage format of the high dynamic range image is a predetermined storage format, the decomposition operation is performed on the high dynamic range image to obtain the first component, the second component, and the third component of the high dynamic range image.
- the predetermined storage format may be HSV color space.
- the target image that is, the high dynamic range image
- the target image can be directly executed. Decompose the operation to obtain the first component, the second component, and the third component of the target image.
- Case 2 When it is judged that the storage format of the high dynamic range image is not a predetermined storage format, that is, when the storage format of the target image does not use the HSV color space, for example, it is determined that the storage format of the target image is the RGB color space; Before performing the decomposition operation on the high dynamic range image, it is also necessary to perform a conversion operation on the high dynamic range image in order to convert it into a high dynamic range image in a predetermined storage format (ie HSV color space), so that the converted high dynamic range image Perform the decomposition operation.
- a predetermined storage format ie HSV color space
- the high dynamic range image can be converted from the RGB color space based on the computer vision processing technology under Opencv To the HSV color space. Therefore, by converting the storage form of the high dynamic range image, a high dynamic range image conforming to the predetermined storage form is obtained, so as to convert the original processing object into a to-be-processed image that can be directly used for decomposition.
- the following method can be used to perform the decomposition operation on the high dynamic range image, so as to obtain the first component, the second component, and the high dynamic range image.
- the third component can specifically include the following:
- the HSV color space uses Hue, Saturation, and Value to describe the color of an image
- the HSV color space contains the hue component (H channel), saturation component (S channel) and Luminance component (V channel), so the above three components can be extracted directly from the HSV color space and denoted as the first component, the second component and the third component.
- the first component can be used to represent the saturation information
- the second The component represents brightness information
- the third component represents hue information
- the “first”, “second”, and “third” in the above-mentioned first, second, and third components are only for distinguishing different components, not as a pair The specific component name and content limitation.
- the reason why the embodiment of this specification converts the original processing object into the HSV color space and decomposes the components of the high dynamic range image in the HSV color space is that it is considered that the tone mapping is mainly for the dynamic range. Compression, the hue problem is generally solved by color gamut mapping. Therefore, the high dynamic range image is converted from RGB color space to HSV color space and decomposed into H channel, S channel and V channel. Among them, H channel contains hue information, S The channel contains the saturation information, the V channel contains the brightness information. The saturation component and the brightness component are learned and mapped, and the hue component is not processed temporarily. The hue component is retained, and then merged to form a low dynamic range image. Because the hue component is retained, It reduces the impact on colors and reduces the color difference of the image after tone mapping.
- step S130 the first component and the second component are input into a predetermined deep neural network, and the deep neural network is used to map the first component and the second component respectively to obtain the mapped first component One component and second component.
- the predetermined deep neural network is a generative adversarial network.
- the generative adversarial network can include a generative network and a discriminant network.
- the structure of the generative network and the discriminant network will be further described below. Include the following:
- the generation network is established based on the U-Net network.
- the generation network includes an encoder and a decoder.
- the encoder contains at least one convolution block and multiple residual blocks, and the decoder contains multiple deconvolution blocks;
- the generating network can also be called a generator, and the generating network is established based on the U-Net network structure; the encoder contains one convolution block and four residual blocks arranged in sequence, where the The convolution block contains a convolution layer and an activation layer.
- the size of the convolution kernel of the convolution layer is 3*3, the step size is 2, the padding is 1, and the number of channels is 64; each residual block contains sequentially arranged Convolutional layer, activation layer, convolutional layer and activation layer, and before the second activation layer, it also includes the addition of the input information of the current residual block and the output information of the second convolutional layer, where,
- the convolution kernel size of the convolution layer in the residual block is 3*3, the step size is 2, and the number of channels of each residual block is doubled from 64.
- the activation layer in the encoder uses the RELU activation function, in order to Keep the size of the feature map unchanged and use mirror symmetry to do edge filling; after the last residual block of the encoder, there is also a convolutional layer with a channel of 512 and a convolution kernel of 1*1 for feature transformation;
- the decoder contains five deconvolution blocks arranged in sequence for upsampling.
- the convolution kernel of the deconvolution layer (transposed convolution layer) in the deconvolution block is 3*3, the step size is 2, and the number of channels is as follows Decrease by one-half.
- a skip connection is added between the convolutional blocks of the same resolution of the encoder and the decoder to recover the loss of spatial structure information due to the halving of the resolution.
- After the decoder connect two convolution blocks for fine adjustment.
- the convolution kernel of the convolution layer in the two convolution blocks is 3*3, the step size is 1, and the channels are 64 and 2, respectively.
- the RELU activation function is used for the rest except that the activation layer of the last layer uses the Sigmoid activation function.
- the discriminant network includes multiple convolutional blocks, and each convolutional block contains a convolutional layer, a normalization layer, and an activation layer arranged in sequence. Further, in the embodiment of this specification, the discriminant network can also be called a discriminator, the discriminant network is composed of four convolution blocks, the size of the convolution kernel of the convolution layer in the convolution block is 3*3, and the step size is 2. , The normalization layer in the discrimination network adopts layer normalization, and the activation layer adopts the RELU activation function.
- the generative adversarial network can be trained by a predetermined loss function, and the loss function includes one or more of the generative adversarial loss function, the mean square error function, and the multi-scale structural similarity loss function.
- step S140 the mapped first component and the second component are merged with the third component to obtain a merged low dynamic range image corresponding to the high dynamic range image, so as to complete tone mapping .
- the brightness component and saturation component are input to generate the anti-network learning map, the mapped brightness component and saturation component are output, and the mapped brightness component is output. And the saturation component and the hue component are fused, and then the fused low dynamic range image corresponding to the original processing object (high dynamic range image) can be obtained, that is, the hue mapping is completed.
- the above-mentioned components may be merged in the following manner to obtain a low dynamic range image, specifically:
- the mapped first component, the second component and the third component are superimposed to obtain a low dynamic range image conforming to a predetermined storage format.
- the S channel and the V channel obtained after learning and mapping will be learned.
- the low dynamic range image corresponding to the HSV color space is still obtained.
- the low dynamic range image may also include: performing a conversion operation on the low dynamic range image , In order to convert it into a low dynamic range image corresponding to the RGB color space; of course, it is easy to understand that there are no specific restrictions on the color space corresponding to the original processing object (high dynamic range image) in the embodiments of this specification, so the low The color space to which the dynamic range image is converted can be determined according to actual needs.
- FIG. 2 shows a schematic flowchart of tone mapping using a generative adversarial network in a specific application scenario provided by an embodiment of this specification.
- this figure shows a schematic flowchart of tone mapping using a generative adversarial network in a specific application scenario provided by an embodiment of this specification.
- sufficient multi-scale information is learned by using the U-Net network structure in the generator part; since the tone mapping is mainly correct
- the brightness mapping the structure of the object and other information will not change, so the residual block is introduced into the encoder to reduce the difficulty of network learning while maintaining structural integrity and avoiding information loss.
- tone mapping often results in unrealistic mapping results, the use of generative adversarial networks and the introduction of adversarial loss to learn at the perceptual level can improve the naturalness of the mapped image.
- the saturation component and the brightness component of the high dynamic range image are simultaneously input into the generation confrontation network to learn the mapping, and the original hue components are retained, and finally they are merged to form a low dynamic range image.
- the brightness component and saturation component obtained after the learning and mapping of the confrontation network are generated by the present invention are fused with the original hue component.
- the structure is highly consistent with the original high dynamic range image, but also has a high degree of naturalness. It can avoid problems such as color difference while learning brightness and saturation mapping.
- Using the image obtained by tone mapping in the embodiment of this specification as the data set for training and generating the confrontation network can improve the learning effect of the neural network, and it is also possible to obtain a high-quality tone mapping label data set by adjusting the parameters.
- Fig. 3 is a tone mapping device provided by the embodiment of this specification.
- the device 300 mainly includes:
- the acquiring module 301 is configured to acquire one or more high dynamic range images, and judge the storage form of the high dynamic range images;
- the decomposition module 302 is configured to perform a decomposition operation on the high dynamic range image when it is determined that the storage form of the high dynamic range image is a predetermined storage form to obtain the first component and the second component of the high dynamic range image And the third component;
- the mapping module 303 is configured to input the first component and the second component into a predetermined deep neural network, and use the deep neural network to respectively map the first component and the second component to obtain the mapped The first component and the second component;
- the fusion module 304 is configured to merge the mapped first component and the second component with the third component to obtain a fused low dynamic range image corresponding to the high dynamic range image, so as to complete the tone Mapping.
- the device further includes:
- the first conversion module 305 is configured to perform a conversion operation on the high dynamic range image when it is determined that the storage form of the high dynamic range image is a non-predetermined storage form before performing the decomposition operation on the high dynamic range image , So as to convert it into a high dynamic range image in a predetermined storage form, and perform a decomposition operation on the converted high dynamic range image.
- the predetermined storage form includes an HSV color space
- the decomposition module 302 is specifically configured to:
- the components in the HSV color space corresponding to the high dynamic range image are extracted to obtain the first component, the second component, and the third component; wherein the first component includes saturation information, and the first component includes saturation information.
- the second component includes brightness information, and the third component includes hue information.
- the fusion module 304 is specifically configured to:
- the first component and the second component after the mapping are superimposed with the third component to obtain a low dynamic range image conforming to a predetermined storage format.
- the device further includes:
- the second conversion module 306 is configured to perform a conversion operation on the low dynamic range image after the low dynamic range image conforming to the predetermined storage format is obtained, so as to convert it into a low dynamic range image corresponding to the RGB color space.
- the embodiments of the present specification also provide an electronic device including a memory, a processor, and a computer program stored in the memory and capable of running on the processor, and the processor implements the above-mentioned tone mapping method when the program is executed.
- the device, electronic device, and method provided in the embodiments of this specification are corresponding. Therefore, the device and electronic device also have beneficial technical effects similar to the corresponding method. Since the beneficial technical effects of the method have been described in detail above, therefore, here The beneficial technical effects of the corresponding devices and electronic equipment will not be repeated.
- program modules include routines, programs, objects, components, data structures, etc. that perform specific tasks or implement specific abstract data types.
- the instructions can also be practiced in distributed computing environments where tasks are performed by remote processing devices connected through a communication network.
- program modules can be located in local and remote computer storage media including storage devices.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (13)
- 一种色调映射方法,所述方法包括:获取一个或多个高动态范围图像,并对所述高动态范围图像的存储形式进行判断;当判断所述高动态范围图像的存储形式为预定的存储形式时,对所述高动态范围图像执行分解操作,得到所述高动态范围图像的第一分量、第二分量以及第三分量;将所述第一分量和第二分量输入到预定的深度神经网络中,并利用所述深度神经网络分别对所述第一分量和第二分量进行映射,得到映射后的第一分量和第二分量;将所述映射后的第一分量以及第二分量与所述第三分量进行融合,得到融合后的与所述高动态范围图像相对应的低动态范围图像,以便完成色调映射。
- 如权利要求1所述的方法,所述对所述高动态范围图像执行分解操作之前,还包括:当判断所述高动态范围图像的存储形式为非预定的存储形式时,对所述高动态范围图像执行转换操作,以便将其转换为预定存储形式的高动态范围图像,并对所述转换后的高动态范围图像执行分解操作。
- 如权利要求1所述的方法,所述预定的存储形式包括HSV颜色空间,所述对所述高动态范围图像执行分解操作,得到所述高动态范围图像的第一分量、第二分量以及第三分量,包括:对所述高动态范围图像所对应的HSV颜色空间中的分量进行提取,以便获取所述第一分量、第二分量以及第三分量;其中,所述第一分量包括饱和度信息,所述第二分量包括亮度信息,所述第三分量包括色相信息。
- 如权利要求1所述的方法,所述预定的深度神经网络为生成对抗网络,所述生成对抗网络包括生成网络和判别网络,其中:所述生成网络基于U-Net网络建立,所述生成网络包括编码器和解码器,所述编码器内包含至少一个卷积块以及多个残差块,所述解码器内包含多个反卷积块;所述判别网络包括多个卷积块,每个所述卷积块内包含依次排列的卷积层、归一化层和激活层。
- 如权利要求4所述的方法,所述生成对抗网络由预定的损失函数训练得到,所述损失函数包括生成对抗损失函数、均方误差函数和多尺度结构相似性损失函数中的一种或几种。
- 如权利要求1所述的方法,所述将所述映射后的第一分量以及第二分量与所述第三分量进行融合,得到融合后的与所述高动态范围图像相对应的低动态范围图像,包括:将所述映射后的第一分量以及第二分量与所述第三分量进行叠加,得到符合预定存储形式的低动态范围图像。
- 如权利要求6所述的方法,所述得到符合预定存储形式的低动态范围图像之后,还包括:对所述低动态范围图像执行转换操作,以便将其转换为RGB颜色空间所对应的低动态范围图像。
- 一种色调映射装置,所述装置包括:获取模块,用于获取一个或多个高动态范围图像,并对所述高动态范围图像的存储形式进行判断;分解模块,用于当判断所述高动态范围图像的存储形式为预定的存储形式时,对所述高动态范围图像执行分解操作,得到所述高动态范围图像的第一分量、第二分量以及第三分量;映射模块,用于将所述第一分量和第二分量输入到预定的深度神经网络中,并利用所述深度神经网络分别对所述第一分量和第二分量进行映射,得到映射 后的第一分量和第二分量;融合模块,用于将所述映射后的第一分量以及第二分量与所述第三分量进行融合,得到融合后的与所述高动态范围图像相对应的低动态范围图像,以便完成色调映射。
- 如权利要求8所述的装置,所述装置还包括:第一转换模块,用于在对所述高动态范围图像执行分解操作之前,当判断所述高动态范围图像的存储形式为非预定的存储形式时,对所述高动态范围图像执行转换操作,以便将其转换为预定存储形式的高动态范围图像,并对所述转换后的高动态范围图像执行分解操作。
- 如权利要求8所述的装置,所述预定的存储形式包括HSV颜色空间,所述分解模块具体用于:对所述高动态范围图像所对应的HSV颜色空间中的分量进行提取,以便获取所述第一分量、第二分量以及第三分量;其中,所述第一分量包括饱和度信息,所述第二分量包括亮度信息,所述第三分量包括色相信息。
- 如权利要求8所述的装置,所述融合模块具体用于:将所述映射后的第一分量以及第二分量与所述第三分量进行叠加,得到符合预定存储形式的低动态范围图像。
- 如权利要求11所述的装置,所述装置还包括:第二转换模块,用于在所述得到符合预定存储形式的低动态范围图像之后,对所述低动态范围图像执行转换操作,以便将其转换为RGB颜色空间所对应的低动态范围图像。
- 一种电子设备,包括存储器,处理器及存储在存储器上并可在处理器上运行的计算机程序,其特征在于,所述处理器执行所述程序时实现权利要求1 至7中任一项所述的方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/725,334 US20220245775A1 (en) | 2019-10-31 | 2022-04-20 | Tone mapping method and electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911057461.1A CN110796595B (zh) | 2019-10-31 | 2019-10-31 | 一种色调映射方法、装置及电子设备 |
CN201911057461.1 | 2019-10-31 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/725,334 Continuation-In-Part US20220245775A1 (en) | 2019-10-31 | 2022-04-20 | Tone mapping method and electronic device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021082088A1 true WO2021082088A1 (zh) | 2021-05-06 |
Family
ID=69440621
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/118585 WO2021082088A1 (zh) | 2019-10-31 | 2019-11-14 | 一种色调映射方法、装置及电子设备 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20220245775A1 (zh) |
CN (1) | CN110796595B (zh) |
WO (1) | WO2021082088A1 (zh) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11908109B1 (en) * | 2019-11-04 | 2024-02-20 | madVR Holdings LLC | Enhanced video processor |
US11893482B2 (en) * | 2019-11-14 | 2024-02-06 | Microsoft Technology Licensing, Llc | Image restoration for through-display imaging |
CN111667430B (zh) * | 2020-06-09 | 2022-11-22 | 展讯通信(上海)有限公司 | 图像的处理方法、装置、设备以及存储介质 |
CN111784598B (zh) * | 2020-06-18 | 2023-06-02 | Oppo(重庆)智能科技有限公司 | 色调映射模型的训练方法、色调映射方法及电子设备 |
CN113066019A (zh) * | 2021-02-27 | 2021-07-02 | 华为技术有限公司 | 一种图像增强方法及相关装置 |
CN115529834A (zh) * | 2021-04-27 | 2022-12-27 | 京东方科技集团股份有限公司 | 图像处理方法和图像处理装置 |
CN116993997A (zh) * | 2022-04-22 | 2023-11-03 | 戴尔产品有限公司 | 用于训练编码器和处理数据的方法、电子设备和程序产品 |
TWI844030B (zh) * | 2022-06-07 | 2024-06-01 | 鴻海精密工業股份有限公司 | 瑕疵檢測裝置及方法 |
CN116029914B (zh) * | 2022-07-27 | 2023-10-20 | 荣耀终端有限公司 | 图像处理方法与电子设备 |
CN115205157B (zh) * | 2022-07-29 | 2024-04-26 | 如你所视(北京)科技有限公司 | 图像处理方法和系统、电子设备和存储介质 |
CN115631428B (zh) * | 2022-11-01 | 2023-08-11 | 西南交通大学 | 一种基于结构纹理分解的无监督图像融合方法和系统 |
CN117474816B (zh) * | 2023-12-26 | 2024-03-12 | 中国科学院宁波材料技术与工程研究所 | 高动态范围图像色调映射方法、系统及可读存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8766999B2 (en) * | 2010-05-20 | 2014-07-01 | Aptina Imaging Corporation | Systems and methods for local tone mapping of high dynamic range images |
CN108885782A (zh) * | 2017-08-09 | 2018-11-23 | 深圳市大疆创新科技有限公司 | 图像处理方法、设备及计算机可读存储介质 |
CN110197463A (zh) * | 2019-04-25 | 2019-09-03 | 深圳大学 | 基于深度学习的高动态范围图像色调映射方法及其系统 |
CN110232669A (zh) * | 2019-06-19 | 2019-09-13 | 湖北工业大学 | 一种高动态范围图像的色调映射方法及系统 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8237813B2 (en) * | 2009-04-23 | 2012-08-07 | Csr Technology Inc. | Multiple exposure high dynamic range image capture |
US20110028850A1 (en) * | 2009-07-28 | 2011-02-03 | Thomas Schuhrke | Process for quantitative display of blood flow |
CN110383802B (zh) * | 2017-03-03 | 2021-05-25 | 杜比实验室特许公司 | 利用逼近函数的彩色图像修改方法 |
CN107657594A (zh) * | 2017-09-22 | 2018-02-02 | 武汉大学 | 一种高质量的快速色调映射方法和系统 |
CN108010024B (zh) * | 2017-12-11 | 2021-12-07 | 宁波大学 | 一种盲参考色调映射图像质量评价方法 |
CN108024104B (zh) * | 2017-12-12 | 2020-02-28 | 上海顺久电子科技有限公司 | 一种对输入的高动态范围图像进行处理的方法和显示设备 |
CN108805836A (zh) * | 2018-05-31 | 2018-11-13 | 大连理工大学 | 基于深度往复式hdr变换的图像校正方法 |
WO2019241155A1 (en) * | 2018-06-11 | 2019-12-19 | Arterys Inc. | Simulating abnormalities in medical images with generative adversarial networks |
CN110223256A (zh) * | 2019-06-10 | 2019-09-10 | 北京大学深圳研究生院 | 一种逆色调映射方法、装置及电子设备 |
US11150605B1 (en) * | 2019-07-22 | 2021-10-19 | Facebook Technologies, Llc | Systems and methods for generating holograms using deep learning |
-
2019
- 2019-10-31 CN CN201911057461.1A patent/CN110796595B/zh active Active
- 2019-11-14 WO PCT/CN2019/118585 patent/WO2021082088A1/zh active Application Filing
-
2022
- 2022-04-20 US US17/725,334 patent/US20220245775A1/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8766999B2 (en) * | 2010-05-20 | 2014-07-01 | Aptina Imaging Corporation | Systems and methods for local tone mapping of high dynamic range images |
CN108885782A (zh) * | 2017-08-09 | 2018-11-23 | 深圳市大疆创新科技有限公司 | 图像处理方法、设备及计算机可读存储介质 |
CN110197463A (zh) * | 2019-04-25 | 2019-09-03 | 深圳大学 | 基于深度学习的高动态范围图像色调映射方法及其系统 |
CN110232669A (zh) * | 2019-06-19 | 2019-09-13 | 湖北工业大学 | 一种高动态范围图像的色调映射方法及系统 |
Also Published As
Publication number | Publication date |
---|---|
CN110796595A (zh) | 2020-02-14 |
US20220245775A1 (en) | 2022-08-04 |
CN110796595B (zh) | 2022-03-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021082088A1 (zh) | 一种色调映射方法、装置及电子设备 | |
US10148871B2 (en) | Advanced raw conversion to produce high dynamic range, wide color gamut output | |
CN108933933B (zh) | 一种视频信号处理方法及装置 | |
CN110717868B (zh) | 视频高动态范围反色调映射模型构建、映射方法及装置 | |
CN105850114A (zh) | 用于图像的逆色调映射的方法 | |
CN106780417A (zh) | 一种光照不均图像的增强方法及系统 | |
CN111145290B (zh) | 一种图像彩色化方法、系统和计算机可读存储介质 | |
CN114998145B (zh) | 一种基于多尺度和上下文学习网络的低照度图像增强方法 | |
CN112508812B (zh) | 图像色偏校正方法、模型训练方法、装置及设备 | |
JP2016046706A (ja) | 画像処理装置、画像処理方法、記憶媒体及びプログラム | |
CN113284064A (zh) | 一种基于注意力机制跨尺度上下文低照度图像增强方法 | |
CN110232669A (zh) | 一种高动态范围图像的色调映射方法及系统 | |
CN115393227A (zh) | 基于深度学习的微光全彩视频图像自适应增强方法及系统 | |
CN114202460B (zh) | 面对不同损伤图像的超分辨率高清重建方法和系统及设备 | |
CN106709888B (zh) | 一种基于人眼视觉模型的高动态范围图像产生方法 | |
CN114862707A (zh) | 一种多尺度特征恢复图像增强方法、装置及存储介质 | |
CN111161189A (zh) | 一种基于细节弥补网络的单幅图像再增强方法 | |
CN114638764B (zh) | 基于人工智能的多曝光图像融合方法及系统 | |
US11647298B2 (en) | Image processing apparatus, image capturing apparatus, image processing method, and storage medium | |
CN114549386A (zh) | 一种基于自适应光照一致性的多曝光图像融合方法 | |
CN109598686B (zh) | 一种基于bayer模板的卫星影像重建方法 | |
Kim et al. | Efficient-HDRTV: Efficient SDR to HDR Conversion for HDR TV | |
CN103077396B (zh) | 一种彩色图像的向量空间特征点提取方法及装置 | |
CN113077385A (zh) | 基于对抗生成网络和边缘增强的视频超分辨率方法及系统 | |
Kar et al. | Statistical approach for color image detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19950262 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19950262 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19950262 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 14.02.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19950262 Country of ref document: EP Kind code of ref document: A1 |