US20190304072A1 - Method and apparatus for converting low dynamic range video to high dynamic range video - Google Patents

Method and apparatus for converting low dynamic range video to high dynamic range video Download PDF

Info

Publication number
US20190304072A1
US20190304072A1 US15/936,459 US201815936459A US2019304072A1 US 20190304072 A1 US20190304072 A1 US 20190304072A1 US 201815936459 A US201815936459 A US 201815936459A US 2019304072 A1 US2019304072 A1 US 2019304072A1
Authority
US
United States
Prior art keywords
current frame
pixel
dynamic range
mapping functions
breathing parameters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/936,459
Other versions
US10445865B1 (en
Inventor
Hon Wah Wong
Hon Tung Luk
Chi Keung Fong
Yin SZE
Hok Kwan CHEUNG
Yiu Fai YUEN
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tfi Digital Media Ltd
Original Assignee
Tfi Digital Media Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tfi Digital Media Ltd filed Critical Tfi Digital Media Ltd
Priority to US15/936,459 priority Critical patent/US10445865B1/en
Assigned to TFI DIGITAL MEDIA LIMITED reassignment TFI DIGITAL MEDIA LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHEUNG, Hok Kwan, FONG, CHI KEUNG, LUK, HON TUNG, SZE, Yin, WONG, HON WAH, YUEN, Yiu Fai
Priority to CN201810410603.7A priority patent/CN110310231B/en
Publication of US20190304072A1 publication Critical patent/US20190304072A1/en
Application granted granted Critical
Publication of US10445865B1 publication Critical patent/US10445865B1/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06T5/92
    • G06T5/90
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/007Dynamic range modification
    • G06T5/009Global, i.e. based on properties of the image as a whole
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration by the use of histogram techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20208High dynamic range [HDR] image processing

Definitions

  • the present invention relates to digital image and video processing and conversion of dynamic ranges of images and videos, and more particularly methods for enhancing standard dynamic range images and videos to high dynamic range images and videos.
  • No. 8,265,378 discloses how to convert and represent image data from lower bit depth to higher bit depth for rendering HDR image data that are typically coded in 10 bits to 12 bits, instead of 8 bits for SDR image data.
  • U.S. Pat. No. 8,050,512 discloses a conversion performed during the displaying process, where the conversion does not depend on other images.
  • U.S. Pat. No. 7,573,533 an adaptive contrast enhancement method by generating the transfer curves is proposed.
  • Human eyes are highly adaptive to a wide range of luminance levels. Human visual perception adjusts automatically according to the target display for comfortable viewing experience. It is essential to utilize the maximum dynamic range of the target display without losing details, and at the same time present a majority of the content at a luminance level that is most sensitive to human eyes. Thus, the conversion from a SDR video to a HDR video is actually an enhancement of the dynamic range from SDR to HDR. The perceptual responses of human eyes to different dynamic ranges and colors are different. It would be difficult to find a universal mapping for all pixels in a video to facilitate a pleasant perceptual viewing experience after the video is converted to HDR.
  • a static conversion in which a universal mapping function is used for the whole video without taking the spatial and temporal characteristics of the video into account, is not the optimal way in most cases.
  • a static conversion may result in overly bright HDR images from some bright SDR images or overly dark HDR images from some dark SDR images.
  • An adaptive conversion based on the spatial statistics of the video may do a better job.
  • the adaptive conversion with only spatial information may result in loss of continuity of luminance changes from frame to frame due to the different spatial statistics of individual frames. Except for screen change, such adaptive conversion may introduce flickering effect.
  • the conversion is adaptive and takes both spatial and temporal information of a current frame and previous frames into consideration such that the majority of pixels falls into the most sensitive regions of human eyes in the target dynamic range, while at the same time, the continuity of luminance is maintained in temporal domain to prevent flickering.
  • pleasant visual perceptual experience can then be facilitated.
  • the method for converting a SDR video to a HDR video comprises: generating breathing parameters of a current frame from temporal information and spatial information of the current frame and one or more previous frames of a first dynamic range video; determining one or more mapping functions according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one or more weights for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixels of the current frame to form a converted frame in the second dynamic range video.
  • FIG. 1 illustrates a block diagram of the conversion method according to one embodiment of the invention
  • FIG. 2 illustrates a block diagram of the step of generating the breathing parameters according to one embodiment of the invention
  • FIG. 3 illustrates a block diagram of the step of determining the mapping functions according to one embodiment of the invention
  • FIG. 4 illustrates a block diagram of the step of determining weights according to one embodiment of the invention.
  • FIG. 5 illustrates a block diagram of the step of converting the pixel values with the determined mapping functions and weights according to one embodiment of the invention.
  • FIG. 1 shows a flowchart of a method for converting a SDR video to a HDR video according to one embodiment of the present invention.
  • the method comprises the: generating breathing parameters 103 of a current frame from temporal information 101 and spatial information 102 of the current frame and one or more previous frames of the SDR video; determining one or more mapping functions 104 according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one or more weights 105 for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixel of the current frame to form a converted frame 106 in the HDR video.
  • FIG. 2 shows how the breathing parameters are generated according to one embodiment of the present invention.
  • a histogram 203 or a cumulative frequency curve of pixel values, such as luminance, chrominance, any one or combination of red, green, and Blue (RGB) channels, or any post-processing results based on any one or combination of the RGB channels, of the current frame is analyzed and transformed to temporary breathing parameters 204 of the current frame.
  • Previous breathing parameters 202 which were extracted from histograms or cumulative frequency curves of previous frames 201 , are retrieved.
  • the temporary breathing parameters 204 and previous breathing parameters 202 are then combined to form current breathing parameters 205 for the current frame.
  • the current breathing parameters of the current frame may be obtained by using a temporal function to calculate a weighted sum of the temporary breathing parameters of the current frame and the breathing parameters of previous frames.
  • the breathing parameters derived solely from the current frame are only temporary breathing parameters.
  • the difference between breathing parameters of the current frame and previous frames is controlled by the temporal function and restricted to below a threshold to maintain continuity.
  • the temporal function may be a linear function along a temporal domain.
  • the algorithm in the process steps of determining the mapping functions and determining the weights for each pixel, may use the breathing parameters as a source of temporal information, the continuity of the breathing parameters can be propagated to the whole conversion process.
  • FIG. 3 shows how the mapping functions are determined according to one embodiment of the invention.
  • a set of spatial criteria 303 . . . 306 are applied on the current breathing parameters 301 of the current frame to estimate the percentages of which the spatial criteria are satisfied by the current frame and cluster the pixels of the current frame into N pixel-groups depending on whether any one of the spatial criteria and which spatial criterion is satisfied by each pixel.
  • the mapping functions 304 . . . 307 are then determined for the pixel-groups 302 . . . 305 according to the satisfied spatial criteria.
  • Various approaches can be used to define the spatial criteria, such as skin tone detection, high saturation color detection or any other machine learning methods.
  • mapping functions may be used to construct an N-dimensional model as shown below:
  • M conversion is the N-dimensional conversion model
  • M 1 . . . M N are the independent mapping functions
  • w 1 . . . w N are the weights corresponding to the mapping functions and also representing the suitability of a pixel to the corresponding mapping functions
  • N representing the number of spatial criteria taken into consideration.
  • the N-dimensional model may be a linear model to simulate a perfect conversion, with the mapping functions being the vectors.
  • mapping functions are designed and generated so as to match the characteristics of different sets of pixels, depending on the luminance and color statistics, as well as the dynamic ranges of the SDR video and HDR video. These mapping functions can also be dependent on the breathing parameters, which contain the temporal and spatial information of the frames. These mapping functions can be independent from each other, such that the conversion model may cover as much scenarios as possible. Typically, the mapping functions may be non-deceasing convex functions.
  • various spatial criteria are used in generating the mapping functions which can model different scenarios. Different spatial criteria may be applied to derive different sets of mapping functions for different groups of pixels with different pixel characteristics. For example, the brightness of bright pixels can be suppressed by applying a less steep mapping function, whereas the brightness of dark pixels can be increased by applying a steep mapping function.
  • FIG. 4 illustrates how the weights are determined.
  • a same set of spatial criteria 403 . . . 407 as those used in determining the mapping functions are applied to estimate the suitability of each of pixels to each of the mapping functions 402 . . . 406 .
  • the respective weights 404 . . . 408 for each pixel to each of the mapping functions are then determined according to the respectively estimated suitability of the pixel to each of the mapping functions.
  • the mapping functions are frame dependent, which means that there are N mapping functions for each frame in total, while the weights are pixel dependent, which means that for each pixel, there are N weights corresponding to N mapping functions, and totally there are N*(number of pixels in the frame) weights for each frame. This is equivalent to clustering the pixels into different groups and applying different mapping functions to different pixel-groups.
  • the value indicating whether a pixel falls into a pixel-group can be a soft decision ratio to ensure smooth transition between adjacent pixels. They can be continuous ratios ranged between 0 and 1 after normalization.
  • this soft decision ratio namely the weight for each mapping function, is determined by the pixel characteristics.
  • Each pixel is covered by at least one pixel-group. In other words, all pixels are covered by the union of the different groups of pixels corresponding to the different mapping functions.
  • the conversion process not only uses the spatial and temporal information from the breathing parameters to generate the mapping functions, but also uses the spatial information of the pixels to determine the weights. Therefore, the conversion and combination are in a more smooth and continuous way in spatial domain.
  • FIG. 5 illustrates how the mapping functions are applied on the current frame to obtain a converted frame in the HDR video.
  • a pixel 501 of the current frame are converted to one or more temporary converted pixel values 506 . . . 511 with the mapping functions 505 . . . 510 corresponding to pixel-groups 503 . . . 508 respectively.
  • weighted sums of the temporary converted pixel values are calculated with the corresponding weights 504 . . . 509 to obtain a converted pixel 512 which will be used to construct the HDR video frame.
  • the converted frame is a weighted sum of converted pixel-values passing through different mapping functions, where the weights are the soft decision ratios determined in previous steps.
  • the pixel-groups may also be converted through multiplying a constant, wherein the constant is an effective gain of applying the mapping functions on the maximum value of the pixels.
  • the method and system for enhancing the dynamic range of a video may be implemented in high definition televisions, mobile or personal computing devices (e.g. “tablet” computer, laptop computer, and personal computer) or user terminals having built-in or peripheral electronic displays, or computer processors having data connectivity to any device having built-in or peripheral electronic displays, and having image processing electronics specifically configured to execute machine instructions; wherein the specifically configured image processing electronics may comprise one or more of general purpose or specialized computing devices, computer processors, or electronic circuitries including, but not limited to, digital signal processors (DSP), application specific integrated circuits (ASIC), field programmable gate arrays (FPGA), and other programmable logic devices.
  • DSP digital signal processors
  • ASIC application specific integrated circuits
  • FPGA field programmable gate arrays
  • the method and system may also comprise the use of and various forms of computer storage media having computer instructions or software codes stored therein which can be used to program computers or microprocessors to perform any of the processes of the present invention.
  • the storage media can include, but are not limited to, floppy disks, optical discs, Blu-ray Disc, DVD, CD-ROMs, and magneto-optical disks, ROMs, RAMs, flash memory devices, or any type of media or devices suitable for storing instructions, codes, and/or data.

Abstract

A method and apparatus for converting a standard dynamic range (SDR) video to a high dynamic range (HDR) video. The conversion is adaptive and takes both spatial and temporal information of a current frame and previous frames into consideration such that the majority of pixels falls into the most sensitive regions of human eyes in the target dynamic range, while at the same time, the continuity of luminance is maintained in temporal domain to prevent flickering.

Description

    FIELD OF THE INVENTION
  • The present invention relates to digital image and video processing and conversion of dynamic ranges of images and videos, and more particularly methods for enhancing standard dynamic range images and videos to high dynamic range images and videos.
  • BACKGROUND OF THE INVENTION
  • In the past, due to the limitations of the camera and monitor systems, most videos were captured in low dynamic ranges as relative to the human perceptual sensitivity and encoded as standard dynamic range (SDR) videos. However, the rapid development of the high dynamic range (HDR) display technology has driven the needs of HDR content. Typical HDR video formats include SMPTE-2084, Dolby Vision, HLG, and HDR10+. Various image and video conversion methods have been developed to convert the mass existing SDR videos to HDR videos. For example, U.S. Pat. Nos. 8,948,537, 8,824,829, 8,582,913, and 8,233,738 disclose various methods for enhancing the input low dynamic range image to produce image data that have higher dynamic range in a real-time implementation. U.S. Pat. No. 8,265,378 discloses how to convert and represent image data from lower bit depth to higher bit depth for rendering HDR image data that are typically coded in 10 bits to 12 bits, instead of 8 bits for SDR image data. U.S. Pat. No. 8,050,512 discloses a conversion performed during the displaying process, where the conversion does not depend on other images. In U.S. Pat. No. 7,573,533, an adaptive contrast enhancement method by generating the transfer curves is proposed.
  • Human eyes are highly adaptive to a wide range of luminance levels. Human visual perception adjusts automatically according to the target display for comfortable viewing experience. It is essential to utilize the maximum dynamic range of the target display without losing details, and at the same time present a majority of the content at a luminance level that is most sensitive to human eyes. Thus, the conversion from a SDR video to a HDR video is actually an enhancement of the dynamic range from SDR to HDR. The perceptual responses of human eyes to different dynamic ranges and colors are different. It would be difficult to find a universal mapping for all pixels in a video to facilitate a pleasant perceptual viewing experience after the video is converted to HDR. Obviously, using a static conversion, in which a universal mapping function is used for the whole video without taking the spatial and temporal characteristics of the video into account, is not the optimal way in most cases. For instance, a static conversion may result in overly bright HDR images from some bright SDR images or overly dark HDR images from some dark SDR images. An adaptive conversion based on the spatial statistics of the video may do a better job. However, the adaptive conversion with only spatial information may result in loss of continuity of luminance changes from frame to frame due to the different spatial statistics of individual frames. Except for screen change, such adaptive conversion may introduce flickering effect.
  • SUMMARY OF THE INVENTION:
  • It is an objective of the present invention to provide a method for converting a standard dynamic range (SDR) video to a high dynamic range (HDR) video. The conversion is adaptive and takes both spatial and temporal information of a current frame and previous frames into consideration such that the majority of pixels falls into the most sensitive regions of human eyes in the target dynamic range, while at the same time, the continuity of luminance is maintained in temporal domain to prevent flickering. Pleasant visual perceptual experience can then be facilitated. It is a further objective of the present invention to provide more freedom in adjusting the conversion pixel by pixel using the spatial information, as well as adjusting the conversion smoothly in time domain using the temporal information.
  • In accordance to one embodiment of the present invention, the method for converting a SDR video to a HDR video comprises: generating breathing parameters of a current frame from temporal information and spatial information of the current frame and one or more previous frames of a first dynamic range video; determining one or more mapping functions according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one or more weights for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixels of the current frame to form a converted frame in the second dynamic range video.
  • BRIEF DESCRIPTION OF THE DRAWINGS:
  • Embodiments of the invention are described in more detail hereinafter with reference to the drawings, in which:
  • FIG. 1 illustrates a block diagram of the conversion method according to one embodiment of the invention;
  • FIG. 2 illustrates a block diagram of the step of generating the breathing parameters according to one embodiment of the invention;
  • FIG. 3 illustrates a block diagram of the step of determining the mapping functions according to one embodiment of the invention;
  • FIG. 4 illustrates a block diagram of the step of determining weights according to one embodiment of the invention; and
  • FIG. 5 illustrates a block diagram of the step of converting the pixel values with the determined mapping functions and weights according to one embodiment of the invention.
  • DETAILED DESCRIPTION:
  • In the following description, conversion methods for enhancing the dynamic range of a video are set forth as preferred examples. It will be apparent to those skilled in the art that modifications, including additions and/or substitutions may be made without departing from the scope and spirit of the invention. Specific details may be omitted so as not to obscure the invention; however, the disclosure is written to enable one skilled in the art to practice the teachings herein without undue experimentation.
  • FIG. 1 shows a flowchart of a method for converting a SDR video to a HDR video according to one embodiment of the present invention. The method comprises the: generating breathing parameters 103 of a current frame from temporal information 101 and spatial information 102 of the current frame and one or more previous frames of the SDR video; determining one or more mapping functions 104 according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one or more weights 105 for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixel of the current frame to form a converted frame 106 in the HDR video.
  • FIG. 2 shows how the breathing parameters are generated according to one embodiment of the present invention. A histogram 203 or a cumulative frequency curve of pixel values, such as luminance, chrominance, any one or combination of red, green, and Blue (RGB) channels, or any post-processing results based on any one or combination of the RGB channels, of the current frame is analyzed and transformed to temporary breathing parameters 204 of the current frame. Previous breathing parameters 202, which were extracted from histograms or cumulative frequency curves of previous frames 201, are retrieved. The temporary breathing parameters 204 and previous breathing parameters 202 are then combined to form current breathing parameters 205 for the current frame. Preferably, the current breathing parameters of the current frame may be obtained by using a temporal function to calculate a weighted sum of the temporary breathing parameters of the current frame and the breathing parameters of previous frames.
  • It should be noted that the breathing parameters derived solely from the current frame are only temporary breathing parameters. The difference between breathing parameters of the current frame and previous frames is controlled by the temporal function and restricted to below a threshold to maintain continuity. The temporal function may be a linear function along a temporal domain. As the algorithm, in the process steps of determining the mapping functions and determining the weights for each pixel, may use the breathing parameters as a source of temporal information, the continuity of the breathing parameters can be propagated to the whole conversion process.
  • FIG. 3 shows how the mapping functions are determined according to one embodiment of the invention. A set of spatial criteria 303 . . . 306 are applied on the current breathing parameters 301 of the current frame to estimate the percentages of which the spatial criteria are satisfied by the current frame and cluster the pixels of the current frame into N pixel-groups depending on whether any one of the spatial criteria and which spatial criterion is satisfied by each pixel. The mapping functions 304 . . . 307 are then determined for the pixel-groups 302 . . . 305 according to the satisfied spatial criteria. Various approaches can be used to define the spatial criteria, such as skin tone detection, high saturation color detection or any other machine learning methods.
  • A multiple number of mapping functions may be used to construct an N-dimensional model as shown below:

  • Mconversion =w 1 M 1 +w 2 M 2 + . . . +w N M N,
  • where Mconversion is the N-dimensional conversion model, M1 . . . MN are the independent mapping functions, and w1 . . . wN are the weights corresponding to the mapping functions and also representing the suitability of a pixel to the corresponding mapping functions, and N representing the number of spatial criteria taken into consideration. The N-dimensional model may be a linear model to simulate a perfect conversion, with the mapping functions being the vectors.
  • The mapping functions are designed and generated so as to match the characteristics of different sets of pixels, depending on the luminance and color statistics, as well as the dynamic ranges of the SDR video and HDR video. These mapping functions can also be dependent on the breathing parameters, which contain the temporal and spatial information of the frames. These mapping functions can be independent from each other, such that the conversion model may cover as much scenarios as possible. Typically, the mapping functions may be non-deceasing convex functions. In addition to the statistical information, various spatial criteria are used in generating the mapping functions which can model different scenarios. Different spatial criteria may be applied to derive different sets of mapping functions for different groups of pixels with different pixel characteristics. For example, the brightness of bright pixels can be suppressed by applying a less steep mapping function, whereas the brightness of dark pixels can be increased by applying a steep mapping function.
  • FIG. 4 illustrates how the weights are determined. A same set of spatial criteria 403 . . . 407 as those used in determining the mapping functions are applied to estimate the suitability of each of pixels to each of the mapping functions 402 . . . 406. The respective weights 404 . . . 408 for each pixel to each of the mapping functions are then determined according to the respectively estimated suitability of the pixel to each of the mapping functions. It should be noted that the mapping functions are frame dependent, which means that there are N mapping functions for each frame in total, while the weights are pixel dependent, which means that for each pixel, there are N weights corresponding to N mapping functions, and totally there are N*(number of pixels in the frame) weights for each frame. This is equivalent to clustering the pixels into different groups and applying different mapping functions to different pixel-groups.
  • In general, the value indicating whether a pixel falls into a pixel-group can be a soft decision ratio to ensure smooth transition between adjacent pixels. They can be continuous ratios ranged between 0 and 1 after normalization. In this embodiment, when the mapping functions are generated and applied, this soft decision ratio, namely the weight for each mapping function, is determined by the pixel characteristics. Each pixel is covered by at least one pixel-group. In other words, all pixels are covered by the union of the different groups of pixels corresponding to the different mapping functions. The conversion process not only uses the spatial and temporal information from the breathing parameters to generate the mapping functions, but also uses the spatial information of the pixels to determine the weights. Therefore, the conversion and combination are in a more smooth and continuous way in spatial domain.
  • FIG. 5 illustrates how the mapping functions are applied on the current frame to obtain a converted frame in the HDR video. A pixel 501 of the current frame are converted to one or more temporary converted pixel values 506 . . . 511 with the mapping functions 505 . . . 510 corresponding to pixel-groups 503 . . . 508 respectively. Then, weighted sums of the temporary converted pixel values are calculated with the corresponding weights 504 . . . 509 to obtain a converted pixel 512 which will be used to construct the HDR video frame. In other words, the converted frame is a weighted sum of converted pixel-values passing through different mapping functions, where the weights are the soft decision ratios determined in previous steps. Instead of using the mapping functions, the pixel-groups may also be converted through multiplying a constant, wherein the constant is an effective gain of applying the mapping functions on the maximum value of the pixels.
  • The method and system for enhancing the dynamic range of a video may be implemented in high definition televisions, mobile or personal computing devices (e.g. “tablet” computer, laptop computer, and personal computer) or user terminals having built-in or peripheral electronic displays, or computer processors having data connectivity to any device having built-in or peripheral electronic displays, and having image processing electronics specifically configured to execute machine instructions; wherein the specifically configured image processing electronics may comprise one or more of general purpose or specialized computing devices, computer processors, or electronic circuitries including, but not limited to, digital signal processors (DSP), application specific integrated circuits (ASIC), field programmable gate arrays (FPGA), and other programmable logic devices. The method and system may also comprise the use of and various forms of computer storage media having computer instructions or software codes stored therein which can be used to program computers or microprocessors to perform any of the processes of the present invention. The storage media can include, but are not limited to, floppy disks, optical discs, Blu-ray Disc, DVD, CD-ROMs, and magneto-optical disks, ROMs, RAMs, flash memory devices, or any type of media or devices suitable for storing instructions, codes, and/or data.
  • The foregoing description of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations will be apparent to the practitioner skilled in the art.
  • The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalence.

Claims (12)

What is claimed is:
1. A method for converting a first dynamic range video to a second dynamic range video, comprising providing image processing electronics configured to:
receive the first dynamic range video;
generate one or more breathing parameters of a current frame from temporal information and spatial information of the current frame and one or more previous frames of the first dynamic range video;
determining one or more mapping functions according to the generated breathing parameters of the current frame and one or more spatial criteria;
determining one or more weights for each pixel of the current frame corresponding to the determined mapping functions;
converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and
repeating the conversion for all pixels of the current frame to form a converted frame in the second dynamic range video.
2. The method of claim 1, wherein the spatial information of the current frame includes one or more of distributions of luminance, chrominance, red, green, and blue (RGB) channels, and combinations thereof, post-processing results based on one or more of the RGB channels and combinations thereof of the current frame.
3. The method of claim 1, wherein the mapping functions are non-deceasing convex functions.
4. The method of claim 1, wherein the generation of the breathing parameters of the current frame comprises:
retrieving one or more previous breathing parameters of the previous frames;
creating a histogram or cumulative frequency curve for the spatial information of the current frame;
transforming the histogram or cumulative frequency curve into temporary breathing parameters of the current frame; and
generating the current breathing parameters of the current frame by combining the previous breathing parameters of the previous frames and the temporary breathing parameters of the current frame with a temporal function so as to control the difference between the breathing parameters of the current frame and the previous frames to below a threshold.
5. The method of claim 4, wherein the temporal function is a linear function along a temporal domain.
6. The method of claim 1, wherein the determination of the mapping functions comprises:
applying one or more spatial criteria on the current breathing parameters of the current frame to estimate one or more percentages of which the spatial criteria are satisfied by the current frame;
clustering the pixels of the current frame into one or more pixel-groups depending on whether any one of the spatial criteria and which spatial criterion is satisfied by each pixel; and
determining the mapping functions for each of the pixel-groups according to the satisfied spatial criteria.
7. The method of claim 6, wherein the pixels satisfying one of the spatial criteria are pixels of a skin tone color.
8. The method of claim 6, wherein the pixels satisfying one of the spatial criteria are pixels of a high saturation color.
9. The method of claim 1, wherein the determination of the weights comprises:
applying the spatial criteria to estimate the suitability of each of the pixels to each of the mapping functions; and
determining the respective weights for each pixel corresponding to each of the mapping functions, according to the estimated suitability.
12. The method of claim 1, wherein the conversion of a pixel of the current frame to a converted pixel comprises:
mapping a pixel of the current frame to one or more temporary converted pixel values with the mapping functions; and
calculating a weighted sum of the temporary converted pixel values with the weights corresponding to the mapping functions to obtain a converted pixel in the converted frame in the second dynamic range video.
13. The method of claim 12, wherein the conversion of the pixel-groups comprises multiplying a constant instead of using the mapping functions, wherein the constant is an effective gain of applying the mapping functions on the maximum value of the pixels.
14. An apparatus for converting a first dynamic range video to a second dynamic range video, configured to:
receive the first dynamic range video;
generate one or more breathing parameters of a current frame from temporal information and spatial information of the current frame and one or more previous frames of the first dynamic range video;
determining one or more mapping functions according to the generated breathing parameters of the current frame and one or more spatial criteria;
determining one or more weights for each pixel of the current frame corresponding to the determined mapping functions;
converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and
repeating the conversion for all pixels of the current frame to form a converted frame in the second dynamic range video.
US15/936,459 2018-03-27 2018-03-27 Method and apparatus for converting low dynamic range video to high dynamic range video Active 2038-06-14 US10445865B1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US15/936,459 US10445865B1 (en) 2018-03-27 2018-03-27 Method and apparatus for converting low dynamic range video to high dynamic range video
CN201810410603.7A CN110310231B (en) 2018-03-27 2018-05-02 Device and method for converting first dynamic range video into second dynamic range video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/936,459 US10445865B1 (en) 2018-03-27 2018-03-27 Method and apparatus for converting low dynamic range video to high dynamic range video

Publications (2)

Publication Number Publication Date
US20190304072A1 true US20190304072A1 (en) 2019-10-03
US10445865B1 US10445865B1 (en) 2019-10-15

Family

ID=68056426

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/936,459 Active 2038-06-14 US10445865B1 (en) 2018-03-27 2018-03-27 Method and apparatus for converting low dynamic range video to high dynamic range video

Country Status (2)

Country Link
US (1) US10445865B1 (en)
CN (1) CN110310231B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210125317A1 (en) * 2018-06-11 2021-04-29 Sony Interactive Entertainment Inc. Learning device, image generation device, learning method, image generation method, and program
WO2024067461A1 (en) * 2022-09-30 2024-04-04 腾讯科技(深圳)有限公司 Image processing method and apparatus, and computer device and storage medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112218156B (en) * 2020-10-09 2022-07-29 海信视像科技股份有限公司 Method for adjusting video dynamic contrast and display equipment
CN113781319A (en) * 2021-08-02 2021-12-10 中国科学院深圳先进技术研究院 HDR video conversion method, device, equipment and computer storage medium

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6879731B2 (en) * 2003-04-29 2005-04-12 Microsoft Corporation System and process for generating high dynamic range video
CA2563520C (en) 2004-04-15 2014-12-16 The University Of British Columbia Methods and systems for converting images from low dynamic range to high dynamic range
US7573533B2 (en) 2004-10-15 2009-08-11 Genesis Microchip Inc. Method of generating transfer curves for adaptive contrast enhancement
US8050512B2 (en) 2004-11-16 2011-11-01 Sharp Laboratories Of America, Inc. High dynamic range images from low dynamic range images
US8784318B1 (en) * 2005-07-22 2014-07-22 Zonare Medical Systems, Inc. Aberration correction using channel data in ultrasound imaging system
US7558436B2 (en) * 2006-07-20 2009-07-07 Max-Viz, Inc. Image dynamic range control for visual display
US8233738B2 (en) 2007-07-30 2012-07-31 Dolby Laboratories Licensing Corporation Enhancing dynamic ranges of images
EP2144444B1 (en) * 2008-07-10 2012-06-27 The University Of Warwick HDR video data compression devices and methods
JP5589006B2 (en) * 2009-03-13 2014-09-10 ドルビー ラボラトリーズ ライセンシング コーポレイション Hierarchical compression of high dynamic range, visual dynamic range and wide color gamut videos
TWI559779B (en) * 2010-08-25 2016-11-21 杜比實驗室特許公司 Extending image dynamic range
EP2681710B1 (en) * 2011-03-02 2018-09-12 Dolby Laboratories Licensing Corporation Local multiscale tone-mapping operator
US9591347B2 (en) * 2012-10-31 2017-03-07 Google Inc. Displaying simulated media content item enhancements on mobile devices
US8958649B2 (en) * 2013-03-13 2015-02-17 Wisconsin Alumni Research Foundation Video generation with temporally-offset sampling
JP2016529747A (en) * 2013-05-23 2016-09-23 トムソン ライセンシングThomson Licensing How to tone map a video sequence
US9818422B2 (en) * 2013-08-30 2017-11-14 The Regents Of The University Of California Method and apparatus for layered compression of multimedia signals for storage and transmission over heterogeneous networks
US9558538B2 (en) * 2015-01-09 2017-01-31 Vixs Systems, Inc. Dynamic range converter with frame by frame adaptation and methods for use therewith
CN107211076B (en) * 2015-01-19 2018-10-30 杜比实验室特许公司 The method, apparatus and storage medium of display management for high dynamic range video
KR20180128888A (en) * 2015-07-24 2018-12-04 에테하 취리히 Image processing system for downscaling an image using a perceptual downscaling method
US9824442B2 (en) * 2015-08-20 2017-11-21 Siemens Medical Solutions Usa, Inc. View direction adaptive volume ultrasound imaging
US10043251B2 (en) * 2015-10-09 2018-08-07 Stmicroelectronics Asia Pacific Pte Ltd Enhanced tone mapper for high dynamic range images and video
BR112017016037A2 (en) * 2015-12-17 2018-03-20 Koninklijke Philips N.V. HDR video decoder and encoder; video decoding method; method for hdr video encoding; and computer readable memory
GB201611253D0 (en) * 2016-06-29 2016-08-10 Dolby Laboratories Licensing Corp Efficient Histogram-based luma look matching
CN107154059B (en) * 2017-06-26 2020-08-04 杭州当虹科技股份有限公司 High dynamic range video processing method

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210125317A1 (en) * 2018-06-11 2021-04-29 Sony Interactive Entertainment Inc. Learning device, image generation device, learning method, image generation method, and program
US11625816B2 (en) * 2018-06-11 2023-04-11 Sony Interactive Entertainment Inc. Learning device, image generation device, learning method, image generation method, and program
WO2024067461A1 (en) * 2022-09-30 2024-04-04 腾讯科技(深圳)有限公司 Image processing method and apparatus, and computer device and storage medium

Also Published As

Publication number Publication date
US10445865B1 (en) 2019-10-15
CN110310231A (en) 2019-10-08
CN110310231B (en) 2023-08-29

Similar Documents

Publication Publication Date Title
US9826149B2 (en) Machine learning of real-time image capture parameters
US10445865B1 (en) Method and apparatus for converting low dynamic range video to high dynamic range video
Gu et al. The analysis of image contrast: From quality assessment to automatic enhancement
KR102049481B1 (en) Electronic device for performing image conversion and method thereof
Rivera et al. Content-aware dark image enhancement through channel division
US8666156B2 (en) Image-based backgrounds for images
CN113518185B (en) Video conversion processing method and device, computer readable medium and electronic equipment
CN103403759A (en) Local multiscale tone-mapping operator
Jung et al. Optimized perceptual tone mapping for contrast enhancement of images
CN109686342B (en) Image processing method and device
JPWO2017203942A1 (en) IMAGE PROCESSING APPARATUS, IMAGE PROCESSING METHOD, AND PROGRAM
US20110058736A1 (en) Image processing apparatus, method, and program
CN114511479A (en) Image enhancement method and device
US20180084180A1 (en) Display apparatus and method of processing image thereof
US7885458B1 (en) Illuminant estimation using gamut mapping and scene classification
Trongtirakul et al. Single backlit image enhancement
Abebe et al. Evaluating the color fidelity of itmos and hdr color appearance models
Song et al. Luminance enhancement and detail preservation of images and videos adapted to ambient illumination
Cyriac et al. Optimized tone curve for in-camera image processing
Mohammadi et al. An entropy-based inverse tone mapping operator for high dynamic range applications
Luzardo et al. Fully-automatic inverse tone mapping algorithm based on dynamic mid-level tone mapping
KR20190058753A (en) Image processing method and image processor performing the same
CN105631812B (en) Control method and control device for color enhancement of display image
CN112243118B (en) White balance correction method, device, equipment and storage medium
Mohammadi et al. A high contrast video inverse tone mapping operator for high dynamic range applications

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

AS Assignment

Owner name: TFI DIGITAL MEDIA LIMITED, HONG KONG

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WONG, HON WAH;LUK, HON TUNG;FONG, CHI KEUNG;AND OTHERS;REEL/FRAME:045420/0161

Effective date: 20180327

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment: 4