US20190304072A1 - Method and apparatus for converting low dynamic range video to high dynamic range video - Google Patents
Method and apparatus for converting low dynamic range video to high dynamic range video Download PDFInfo
- Publication number
- US20190304072A1 US20190304072A1 US15/936,459 US201815936459A US2019304072A1 US 20190304072 A1 US20190304072 A1 US 20190304072A1 US 201815936459 A US201815936459 A US 201815936459A US 2019304072 A1 US2019304072 A1 US 2019304072A1
- Authority
- US
- United States
- Prior art keywords
- current frame
- pixel
- dynamic range
- mapping functions
- breathing parameters
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 30
- 238000006243 chemical reaction Methods 0.000 claims abstract description 30
- 230000002123 temporal effect Effects 0.000 claims abstract description 20
- 230000006870 function Effects 0.000 claims description 64
- 238000013507 mapping Methods 0.000 claims description 59
- 230000029058 respiratory gaseous exchange Effects 0.000 claims description 34
- 230000001186 cumulative effect Effects 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 4
- 238000012886 linear function Methods 0.000 claims description 2
- 238000012805 post-processing Methods 0.000 claims description 2
- 238000009826 distribution Methods 0.000 claims 1
- 230000001131 transforming effect Effects 0.000 claims 1
- 230000003044 adaptive effect Effects 0.000 abstract description 7
- 238000010586 diagram Methods 0.000 description 5
- 230000002708 enhancing effect Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000001419 dependent effect Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 238000007792 addition Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000008825 perceptual sensitivity Effects 0.000 description 1
- 230000000644 propagated effect Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 239000013598 vector Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Images
Classifications
-
- G06T5/92—
-
- G06T5/90—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/007—Dynamic range modification
- G06T5/009—Global, i.e. based on properties of the image as a whole
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/40—Image enhancement or restoration by the use of histogram techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20208—High dynamic range [HDR] image processing
Definitions
- the present invention relates to digital image and video processing and conversion of dynamic ranges of images and videos, and more particularly methods for enhancing standard dynamic range images and videos to high dynamic range images and videos.
- No. 8,265,378 discloses how to convert and represent image data from lower bit depth to higher bit depth for rendering HDR image data that are typically coded in 10 bits to 12 bits, instead of 8 bits for SDR image data.
- U.S. Pat. No. 8,050,512 discloses a conversion performed during the displaying process, where the conversion does not depend on other images.
- U.S. Pat. No. 7,573,533 an adaptive contrast enhancement method by generating the transfer curves is proposed.
- Human eyes are highly adaptive to a wide range of luminance levels. Human visual perception adjusts automatically according to the target display for comfortable viewing experience. It is essential to utilize the maximum dynamic range of the target display without losing details, and at the same time present a majority of the content at a luminance level that is most sensitive to human eyes. Thus, the conversion from a SDR video to a HDR video is actually an enhancement of the dynamic range from SDR to HDR. The perceptual responses of human eyes to different dynamic ranges and colors are different. It would be difficult to find a universal mapping for all pixels in a video to facilitate a pleasant perceptual viewing experience after the video is converted to HDR.
- a static conversion in which a universal mapping function is used for the whole video without taking the spatial and temporal characteristics of the video into account, is not the optimal way in most cases.
- a static conversion may result in overly bright HDR images from some bright SDR images or overly dark HDR images from some dark SDR images.
- An adaptive conversion based on the spatial statistics of the video may do a better job.
- the adaptive conversion with only spatial information may result in loss of continuity of luminance changes from frame to frame due to the different spatial statistics of individual frames. Except for screen change, such adaptive conversion may introduce flickering effect.
- the conversion is adaptive and takes both spatial and temporal information of a current frame and previous frames into consideration such that the majority of pixels falls into the most sensitive regions of human eyes in the target dynamic range, while at the same time, the continuity of luminance is maintained in temporal domain to prevent flickering.
- pleasant visual perceptual experience can then be facilitated.
- the method for converting a SDR video to a HDR video comprises: generating breathing parameters of a current frame from temporal information and spatial information of the current frame and one or more previous frames of a first dynamic range video; determining one or more mapping functions according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one or more weights for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixels of the current frame to form a converted frame in the second dynamic range video.
- FIG. 1 illustrates a block diagram of the conversion method according to one embodiment of the invention
- FIG. 2 illustrates a block diagram of the step of generating the breathing parameters according to one embodiment of the invention
- FIG. 3 illustrates a block diagram of the step of determining the mapping functions according to one embodiment of the invention
- FIG. 4 illustrates a block diagram of the step of determining weights according to one embodiment of the invention.
- FIG. 5 illustrates a block diagram of the step of converting the pixel values with the determined mapping functions and weights according to one embodiment of the invention.
- FIG. 1 shows a flowchart of a method for converting a SDR video to a HDR video according to one embodiment of the present invention.
- the method comprises the: generating breathing parameters 103 of a current frame from temporal information 101 and spatial information 102 of the current frame and one or more previous frames of the SDR video; determining one or more mapping functions 104 according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one or more weights 105 for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixel of the current frame to form a converted frame 106 in the HDR video.
- FIG. 2 shows how the breathing parameters are generated according to one embodiment of the present invention.
- a histogram 203 or a cumulative frequency curve of pixel values, such as luminance, chrominance, any one or combination of red, green, and Blue (RGB) channels, or any post-processing results based on any one or combination of the RGB channels, of the current frame is analyzed and transformed to temporary breathing parameters 204 of the current frame.
- Previous breathing parameters 202 which were extracted from histograms or cumulative frequency curves of previous frames 201 , are retrieved.
- the temporary breathing parameters 204 and previous breathing parameters 202 are then combined to form current breathing parameters 205 for the current frame.
- the current breathing parameters of the current frame may be obtained by using a temporal function to calculate a weighted sum of the temporary breathing parameters of the current frame and the breathing parameters of previous frames.
- the breathing parameters derived solely from the current frame are only temporary breathing parameters.
- the difference between breathing parameters of the current frame and previous frames is controlled by the temporal function and restricted to below a threshold to maintain continuity.
- the temporal function may be a linear function along a temporal domain.
- the algorithm in the process steps of determining the mapping functions and determining the weights for each pixel, may use the breathing parameters as a source of temporal information, the continuity of the breathing parameters can be propagated to the whole conversion process.
- FIG. 3 shows how the mapping functions are determined according to one embodiment of the invention.
- a set of spatial criteria 303 . . . 306 are applied on the current breathing parameters 301 of the current frame to estimate the percentages of which the spatial criteria are satisfied by the current frame and cluster the pixels of the current frame into N pixel-groups depending on whether any one of the spatial criteria and which spatial criterion is satisfied by each pixel.
- the mapping functions 304 . . . 307 are then determined for the pixel-groups 302 . . . 305 according to the satisfied spatial criteria.
- Various approaches can be used to define the spatial criteria, such as skin tone detection, high saturation color detection or any other machine learning methods.
- mapping functions may be used to construct an N-dimensional model as shown below:
- M conversion is the N-dimensional conversion model
- M 1 . . . M N are the independent mapping functions
- w 1 . . . w N are the weights corresponding to the mapping functions and also representing the suitability of a pixel to the corresponding mapping functions
- N representing the number of spatial criteria taken into consideration.
- the N-dimensional model may be a linear model to simulate a perfect conversion, with the mapping functions being the vectors.
- mapping functions are designed and generated so as to match the characteristics of different sets of pixels, depending on the luminance and color statistics, as well as the dynamic ranges of the SDR video and HDR video. These mapping functions can also be dependent on the breathing parameters, which contain the temporal and spatial information of the frames. These mapping functions can be independent from each other, such that the conversion model may cover as much scenarios as possible. Typically, the mapping functions may be non-deceasing convex functions.
- various spatial criteria are used in generating the mapping functions which can model different scenarios. Different spatial criteria may be applied to derive different sets of mapping functions for different groups of pixels with different pixel characteristics. For example, the brightness of bright pixels can be suppressed by applying a less steep mapping function, whereas the brightness of dark pixels can be increased by applying a steep mapping function.
- FIG. 4 illustrates how the weights are determined.
- a same set of spatial criteria 403 . . . 407 as those used in determining the mapping functions are applied to estimate the suitability of each of pixels to each of the mapping functions 402 . . . 406 .
- the respective weights 404 . . . 408 for each pixel to each of the mapping functions are then determined according to the respectively estimated suitability of the pixel to each of the mapping functions.
- the mapping functions are frame dependent, which means that there are N mapping functions for each frame in total, while the weights are pixel dependent, which means that for each pixel, there are N weights corresponding to N mapping functions, and totally there are N*(number of pixels in the frame) weights for each frame. This is equivalent to clustering the pixels into different groups and applying different mapping functions to different pixel-groups.
- the value indicating whether a pixel falls into a pixel-group can be a soft decision ratio to ensure smooth transition between adjacent pixels. They can be continuous ratios ranged between 0 and 1 after normalization.
- this soft decision ratio namely the weight for each mapping function, is determined by the pixel characteristics.
- Each pixel is covered by at least one pixel-group. In other words, all pixels are covered by the union of the different groups of pixels corresponding to the different mapping functions.
- the conversion process not only uses the spatial and temporal information from the breathing parameters to generate the mapping functions, but also uses the spatial information of the pixels to determine the weights. Therefore, the conversion and combination are in a more smooth and continuous way in spatial domain.
- FIG. 5 illustrates how the mapping functions are applied on the current frame to obtain a converted frame in the HDR video.
- a pixel 501 of the current frame are converted to one or more temporary converted pixel values 506 . . . 511 with the mapping functions 505 . . . 510 corresponding to pixel-groups 503 . . . 508 respectively.
- weighted sums of the temporary converted pixel values are calculated with the corresponding weights 504 . . . 509 to obtain a converted pixel 512 which will be used to construct the HDR video frame.
- the converted frame is a weighted sum of converted pixel-values passing through different mapping functions, where the weights are the soft decision ratios determined in previous steps.
- the pixel-groups may also be converted through multiplying a constant, wherein the constant is an effective gain of applying the mapping functions on the maximum value of the pixels.
- the method and system for enhancing the dynamic range of a video may be implemented in high definition televisions, mobile or personal computing devices (e.g. “tablet” computer, laptop computer, and personal computer) or user terminals having built-in or peripheral electronic displays, or computer processors having data connectivity to any device having built-in or peripheral electronic displays, and having image processing electronics specifically configured to execute machine instructions; wherein the specifically configured image processing electronics may comprise one or more of general purpose or specialized computing devices, computer processors, or electronic circuitries including, but not limited to, digital signal processors (DSP), application specific integrated circuits (ASIC), field programmable gate arrays (FPGA), and other programmable logic devices.
- DSP digital signal processors
- ASIC application specific integrated circuits
- FPGA field programmable gate arrays
- the method and system may also comprise the use of and various forms of computer storage media having computer instructions or software codes stored therein which can be used to program computers or microprocessors to perform any of the processes of the present invention.
- the storage media can include, but are not limited to, floppy disks, optical discs, Blu-ray Disc, DVD, CD-ROMs, and magneto-optical disks, ROMs, RAMs, flash memory devices, or any type of media or devices suitable for storing instructions, codes, and/or data.
Abstract
Description
- The present invention relates to digital image and video processing and conversion of dynamic ranges of images and videos, and more particularly methods for enhancing standard dynamic range images and videos to high dynamic range images and videos.
- In the past, due to the limitations of the camera and monitor systems, most videos were captured in low dynamic ranges as relative to the human perceptual sensitivity and encoded as standard dynamic range (SDR) videos. However, the rapid development of the high dynamic range (HDR) display technology has driven the needs of HDR content. Typical HDR video formats include SMPTE-2084, Dolby Vision, HLG, and HDR10+. Various image and video conversion methods have been developed to convert the mass existing SDR videos to HDR videos. For example, U.S. Pat. Nos. 8,948,537, 8,824,829, 8,582,913, and 8,233,738 disclose various methods for enhancing the input low dynamic range image to produce image data that have higher dynamic range in a real-time implementation. U.S. Pat. No. 8,265,378 discloses how to convert and represent image data from lower bit depth to higher bit depth for rendering HDR image data that are typically coded in 10 bits to 12 bits, instead of 8 bits for SDR image data. U.S. Pat. No. 8,050,512 discloses a conversion performed during the displaying process, where the conversion does not depend on other images. In U.S. Pat. No. 7,573,533, an adaptive contrast enhancement method by generating the transfer curves is proposed.
- Human eyes are highly adaptive to a wide range of luminance levels. Human visual perception adjusts automatically according to the target display for comfortable viewing experience. It is essential to utilize the maximum dynamic range of the target display without losing details, and at the same time present a majority of the content at a luminance level that is most sensitive to human eyes. Thus, the conversion from a SDR video to a HDR video is actually an enhancement of the dynamic range from SDR to HDR. The perceptual responses of human eyes to different dynamic ranges and colors are different. It would be difficult to find a universal mapping for all pixels in a video to facilitate a pleasant perceptual viewing experience after the video is converted to HDR. Obviously, using a static conversion, in which a universal mapping function is used for the whole video without taking the spatial and temporal characteristics of the video into account, is not the optimal way in most cases. For instance, a static conversion may result in overly bright HDR images from some bright SDR images or overly dark HDR images from some dark SDR images. An adaptive conversion based on the spatial statistics of the video may do a better job. However, the adaptive conversion with only spatial information may result in loss of continuity of luminance changes from frame to frame due to the different spatial statistics of individual frames. Except for screen change, such adaptive conversion may introduce flickering effect.
- It is an objective of the present invention to provide a method for converting a standard dynamic range (SDR) video to a high dynamic range (HDR) video. The conversion is adaptive and takes both spatial and temporal information of a current frame and previous frames into consideration such that the majority of pixels falls into the most sensitive regions of human eyes in the target dynamic range, while at the same time, the continuity of luminance is maintained in temporal domain to prevent flickering. Pleasant visual perceptual experience can then be facilitated. It is a further objective of the present invention to provide more freedom in adjusting the conversion pixel by pixel using the spatial information, as well as adjusting the conversion smoothly in time domain using the temporal information.
- In accordance to one embodiment of the present invention, the method for converting a SDR video to a HDR video comprises: generating breathing parameters of a current frame from temporal information and spatial information of the current frame and one or more previous frames of a first dynamic range video; determining one or more mapping functions according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one or more weights for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixels of the current frame to form a converted frame in the second dynamic range video.
- Embodiments of the invention are described in more detail hereinafter with reference to the drawings, in which:
-
FIG. 1 illustrates a block diagram of the conversion method according to one embodiment of the invention; -
FIG. 2 illustrates a block diagram of the step of generating the breathing parameters according to one embodiment of the invention; -
FIG. 3 illustrates a block diagram of the step of determining the mapping functions according to one embodiment of the invention; -
FIG. 4 illustrates a block diagram of the step of determining weights according to one embodiment of the invention; and -
FIG. 5 illustrates a block diagram of the step of converting the pixel values with the determined mapping functions and weights according to one embodiment of the invention. - In the following description, conversion methods for enhancing the dynamic range of a video are set forth as preferred examples. It will be apparent to those skilled in the art that modifications, including additions and/or substitutions may be made without departing from the scope and spirit of the invention. Specific details may be omitted so as not to obscure the invention; however, the disclosure is written to enable one skilled in the art to practice the teachings herein without undue experimentation.
-
FIG. 1 shows a flowchart of a method for converting a SDR video to a HDR video according to one embodiment of the present invention. The method comprises the: generatingbreathing parameters 103 of a current frame fromtemporal information 101 andspatial information 102 of the current frame and one or more previous frames of the SDR video; determining one or more mapping functions 104 according to the generated breathing parameters of the current frame and one or more spatial criteria; determining one ormore weights 105 for each pixel of the current frame that correspond to the determined mapping functions; and converting a pixel of the current frame to a converted pixel according to the determined mapping functions and corresponding weights; and repeating the conversion for all pixel of the current frame to form aconverted frame 106 in the HDR video. -
FIG. 2 shows how the breathing parameters are generated according to one embodiment of the present invention. Ahistogram 203 or a cumulative frequency curve of pixel values, such as luminance, chrominance, any one or combination of red, green, and Blue (RGB) channels, or any post-processing results based on any one or combination of the RGB channels, of the current frame is analyzed and transformed to temporary breathing parameters 204 of the current frame. Previous breathing parameters 202, which were extracted from histograms or cumulative frequency curves ofprevious frames 201, are retrieved. The temporary breathing parameters 204 and previous breathing parameters 202 are then combined to formcurrent breathing parameters 205 for the current frame. Preferably, the current breathing parameters of the current frame may be obtained by using a temporal function to calculate a weighted sum of the temporary breathing parameters of the current frame and the breathing parameters of previous frames. - It should be noted that the breathing parameters derived solely from the current frame are only temporary breathing parameters. The difference between breathing parameters of the current frame and previous frames is controlled by the temporal function and restricted to below a threshold to maintain continuity. The temporal function may be a linear function along a temporal domain. As the algorithm, in the process steps of determining the mapping functions and determining the weights for each pixel, may use the breathing parameters as a source of temporal information, the continuity of the breathing parameters can be propagated to the whole conversion process.
-
FIG. 3 shows how the mapping functions are determined according to one embodiment of the invention. A set of spatial criteria 303 . . . 306 are applied on thecurrent breathing parameters 301 of the current frame to estimate the percentages of which the spatial criteria are satisfied by the current frame and cluster the pixels of the current frame into N pixel-groups depending on whether any one of the spatial criteria and which spatial criterion is satisfied by each pixel. Themapping functions 304 . . . 307 are then determined for the pixel-groups 302 . . . 305 according to the satisfied spatial criteria. Various approaches can be used to define the spatial criteria, such as skin tone detection, high saturation color detection or any other machine learning methods. - A multiple number of mapping functions may be used to construct an N-dimensional model as shown below:
-
Mconversion =w 1 M 1 +w 2 M 2 + . . . +w N M N, - where Mconversion is the N-dimensional conversion model, M1 . . . MN are the independent mapping functions, and w1 . . . wN are the weights corresponding to the mapping functions and also representing the suitability of a pixel to the corresponding mapping functions, and N representing the number of spatial criteria taken into consideration. The N-dimensional model may be a linear model to simulate a perfect conversion, with the mapping functions being the vectors.
- The mapping functions are designed and generated so as to match the characteristics of different sets of pixels, depending on the luminance and color statistics, as well as the dynamic ranges of the SDR video and HDR video. These mapping functions can also be dependent on the breathing parameters, which contain the temporal and spatial information of the frames. These mapping functions can be independent from each other, such that the conversion model may cover as much scenarios as possible. Typically, the mapping functions may be non-deceasing convex functions. In addition to the statistical information, various spatial criteria are used in generating the mapping functions which can model different scenarios. Different spatial criteria may be applied to derive different sets of mapping functions for different groups of pixels with different pixel characteristics. For example, the brightness of bright pixels can be suppressed by applying a less steep mapping function, whereas the brightness of dark pixels can be increased by applying a steep mapping function.
-
FIG. 4 illustrates how the weights are determined. A same set of spatial criteria 403 . . . 407 as those used in determining the mapping functions are applied to estimate the suitability of each of pixels to each of the mapping functions 402 . . . 406. Therespective weights 404 . . . 408 for each pixel to each of the mapping functions are then determined according to the respectively estimated suitability of the pixel to each of the mapping functions. It should be noted that the mapping functions are frame dependent, which means that there are N mapping functions for each frame in total, while the weights are pixel dependent, which means that for each pixel, there are N weights corresponding to N mapping functions, and totally there are N*(number of pixels in the frame) weights for each frame. This is equivalent to clustering the pixels into different groups and applying different mapping functions to different pixel-groups. - In general, the value indicating whether a pixel falls into a pixel-group can be a soft decision ratio to ensure smooth transition between adjacent pixels. They can be continuous ratios ranged between 0 and 1 after normalization. In this embodiment, when the mapping functions are generated and applied, this soft decision ratio, namely the weight for each mapping function, is determined by the pixel characteristics. Each pixel is covered by at least one pixel-group. In other words, all pixels are covered by the union of the different groups of pixels corresponding to the different mapping functions. The conversion process not only uses the spatial and temporal information from the breathing parameters to generate the mapping functions, but also uses the spatial information of the pixels to determine the weights. Therefore, the conversion and combination are in a more smooth and continuous way in spatial domain.
-
FIG. 5 illustrates how the mapping functions are applied on the current frame to obtain a converted frame in the HDR video. A pixel 501 of the current frame are converted to one or more temporary converted pixel values 506 . . . 511 with the mapping functions 505 . . . 510 corresponding to pixel-groups 503 . . . 508 respectively. Then, weighted sums of the temporary converted pixel values are calculated with the correspondingweights 504 . . . 509 to obtain a convertedpixel 512 which will be used to construct the HDR video frame. In other words, the converted frame is a weighted sum of converted pixel-values passing through different mapping functions, where the weights are the soft decision ratios determined in previous steps. Instead of using the mapping functions, the pixel-groups may also be converted through multiplying a constant, wherein the constant is an effective gain of applying the mapping functions on the maximum value of the pixels. - The method and system for enhancing the dynamic range of a video may be implemented in high definition televisions, mobile or personal computing devices (e.g. “tablet” computer, laptop computer, and personal computer) or user terminals having built-in or peripheral electronic displays, or computer processors having data connectivity to any device having built-in or peripheral electronic displays, and having image processing electronics specifically configured to execute machine instructions; wherein the specifically configured image processing electronics may comprise one or more of general purpose or specialized computing devices, computer processors, or electronic circuitries including, but not limited to, digital signal processors (DSP), application specific integrated circuits (ASIC), field programmable gate arrays (FPGA), and other programmable logic devices. The method and system may also comprise the use of and various forms of computer storage media having computer instructions or software codes stored therein which can be used to program computers or microprocessors to perform any of the processes of the present invention. The storage media can include, but are not limited to, floppy disks, optical discs, Blu-ray Disc, DVD, CD-ROMs, and magneto-optical disks, ROMs, RAMs, flash memory devices, or any type of media or devices suitable for storing instructions, codes, and/or data.
- The foregoing description of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations will be apparent to the practitioner skilled in the art.
- The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalence.
Claims (12)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/936,459 US10445865B1 (en) | 2018-03-27 | 2018-03-27 | Method and apparatus for converting low dynamic range video to high dynamic range video |
CN201810410603.7A CN110310231B (en) | 2018-03-27 | 2018-05-02 | Device and method for converting first dynamic range video into second dynamic range video |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/936,459 US10445865B1 (en) | 2018-03-27 | 2018-03-27 | Method and apparatus for converting low dynamic range video to high dynamic range video |
Publications (2)
Publication Number | Publication Date |
---|---|
US20190304072A1 true US20190304072A1 (en) | 2019-10-03 |
US10445865B1 US10445865B1 (en) | 2019-10-15 |
Family
ID=68056426
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/936,459 Active 2038-06-14 US10445865B1 (en) | 2018-03-27 | 2018-03-27 | Method and apparatus for converting low dynamic range video to high dynamic range video |
Country Status (2)
Country | Link |
---|---|
US (1) | US10445865B1 (en) |
CN (1) | CN110310231B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210125317A1 (en) * | 2018-06-11 | 2021-04-29 | Sony Interactive Entertainment Inc. | Learning device, image generation device, learning method, image generation method, and program |
WO2024067461A1 (en) * | 2022-09-30 | 2024-04-04 | 腾讯科技(深圳)有限公司 | Image processing method and apparatus, and computer device and storage medium |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112218156B (en) * | 2020-10-09 | 2022-07-29 | 海信视像科技股份有限公司 | Method for adjusting video dynamic contrast and display equipment |
CN113781319A (en) * | 2021-08-02 | 2021-12-10 | 中国科学院深圳先进技术研究院 | HDR video conversion method, device, equipment and computer storage medium |
Family Cites Families (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6879731B2 (en) * | 2003-04-29 | 2005-04-12 | Microsoft Corporation | System and process for generating high dynamic range video |
CA2563520C (en) | 2004-04-15 | 2014-12-16 | The University Of British Columbia | Methods and systems for converting images from low dynamic range to high dynamic range |
US7573533B2 (en) | 2004-10-15 | 2009-08-11 | Genesis Microchip Inc. | Method of generating transfer curves for adaptive contrast enhancement |
US8050512B2 (en) | 2004-11-16 | 2011-11-01 | Sharp Laboratories Of America, Inc. | High dynamic range images from low dynamic range images |
US8784318B1 (en) * | 2005-07-22 | 2014-07-22 | Zonare Medical Systems, Inc. | Aberration correction using channel data in ultrasound imaging system |
US7558436B2 (en) * | 2006-07-20 | 2009-07-07 | Max-Viz, Inc. | Image dynamic range control for visual display |
US8233738B2 (en) | 2007-07-30 | 2012-07-31 | Dolby Laboratories Licensing Corporation | Enhancing dynamic ranges of images |
EP2144444B1 (en) * | 2008-07-10 | 2012-06-27 | The University Of Warwick | HDR video data compression devices and methods |
JP5589006B2 (en) * | 2009-03-13 | 2014-09-10 | ドルビー ラボラトリーズ ライセンシング コーポレイション | Hierarchical compression of high dynamic range, visual dynamic range and wide color gamut videos |
TWI559779B (en) * | 2010-08-25 | 2016-11-21 | 杜比實驗室特許公司 | Extending image dynamic range |
EP2681710B1 (en) * | 2011-03-02 | 2018-09-12 | Dolby Laboratories Licensing Corporation | Local multiscale tone-mapping operator |
US9591347B2 (en) * | 2012-10-31 | 2017-03-07 | Google Inc. | Displaying simulated media content item enhancements on mobile devices |
US8958649B2 (en) * | 2013-03-13 | 2015-02-17 | Wisconsin Alumni Research Foundation | Video generation with temporally-offset sampling |
JP2016529747A (en) * | 2013-05-23 | 2016-09-23 | トムソン ライセンシングThomson Licensing | How to tone map a video sequence |
US9818422B2 (en) * | 2013-08-30 | 2017-11-14 | The Regents Of The University Of California | Method and apparatus for layered compression of multimedia signals for storage and transmission over heterogeneous networks |
US9558538B2 (en) * | 2015-01-09 | 2017-01-31 | Vixs Systems, Inc. | Dynamic range converter with frame by frame adaptation and methods for use therewith |
CN107211076B (en) * | 2015-01-19 | 2018-10-30 | 杜比实验室特许公司 | The method, apparatus and storage medium of display management for high dynamic range video |
KR20180128888A (en) * | 2015-07-24 | 2018-12-04 | 에테하 취리히 | Image processing system for downscaling an image using a perceptual downscaling method |
US9824442B2 (en) * | 2015-08-20 | 2017-11-21 | Siemens Medical Solutions Usa, Inc. | View direction adaptive volume ultrasound imaging |
US10043251B2 (en) * | 2015-10-09 | 2018-08-07 | Stmicroelectronics Asia Pacific Pte Ltd | Enhanced tone mapper for high dynamic range images and video |
BR112017016037A2 (en) * | 2015-12-17 | 2018-03-20 | Koninklijke Philips N.V. | HDR video decoder and encoder; video decoding method; method for hdr video encoding; and computer readable memory |
GB201611253D0 (en) * | 2016-06-29 | 2016-08-10 | Dolby Laboratories Licensing Corp | Efficient Histogram-based luma look matching |
CN107154059B (en) * | 2017-06-26 | 2020-08-04 | 杭州当虹科技股份有限公司 | High dynamic range video processing method |
-
2018
- 2018-03-27 US US15/936,459 patent/US10445865B1/en active Active
- 2018-05-02 CN CN201810410603.7A patent/CN110310231B/en active Active
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210125317A1 (en) * | 2018-06-11 | 2021-04-29 | Sony Interactive Entertainment Inc. | Learning device, image generation device, learning method, image generation method, and program |
US11625816B2 (en) * | 2018-06-11 | 2023-04-11 | Sony Interactive Entertainment Inc. | Learning device, image generation device, learning method, image generation method, and program |
WO2024067461A1 (en) * | 2022-09-30 | 2024-04-04 | 腾讯科技(深圳)有限公司 | Image processing method and apparatus, and computer device and storage medium |
Also Published As
Publication number | Publication date |
---|---|
US10445865B1 (en) | 2019-10-15 |
CN110310231A (en) | 2019-10-08 |
CN110310231B (en) | 2023-08-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9826149B2 (en) | Machine learning of real-time image capture parameters | |
US10445865B1 (en) | Method and apparatus for converting low dynamic range video to high dynamic range video | |
Gu et al. | The analysis of image contrast: From quality assessment to automatic enhancement | |
KR102049481B1 (en) | Electronic device for performing image conversion and method thereof | |
Rivera et al. | Content-aware dark image enhancement through channel division | |
US8666156B2 (en) | Image-based backgrounds for images | |
CN113518185B (en) | Video conversion processing method and device, computer readable medium and electronic equipment | |
CN103403759A (en) | Local multiscale tone-mapping operator | |
Jung et al. | Optimized perceptual tone mapping for contrast enhancement of images | |
CN109686342B (en) | Image processing method and device | |
JPWO2017203942A1 (en) | IMAGE PROCESSING APPARATUS, IMAGE PROCESSING METHOD, AND PROGRAM | |
US20110058736A1 (en) | Image processing apparatus, method, and program | |
CN114511479A (en) | Image enhancement method and device | |
US20180084180A1 (en) | Display apparatus and method of processing image thereof | |
US7885458B1 (en) | Illuminant estimation using gamut mapping and scene classification | |
Trongtirakul et al. | Single backlit image enhancement | |
Abebe et al. | Evaluating the color fidelity of itmos and hdr color appearance models | |
Song et al. | Luminance enhancement and detail preservation of images and videos adapted to ambient illumination | |
Cyriac et al. | Optimized tone curve for in-camera image processing | |
Mohammadi et al. | An entropy-based inverse tone mapping operator for high dynamic range applications | |
Luzardo et al. | Fully-automatic inverse tone mapping algorithm based on dynamic mid-level tone mapping | |
KR20190058753A (en) | Image processing method and image processor performing the same | |
CN105631812B (en) | Control method and control device for color enhancement of display image | |
CN112243118B (en) | White balance correction method, device, equipment and storage medium | |
Mohammadi et al. | A high contrast video inverse tone mapping operator for high dynamic range applications |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
AS | Assignment |
Owner name: TFI DIGITAL MEDIA LIMITED, HONG KONG Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WONG, HON WAH;LUK, HON TUNG;FONG, CHI KEUNG;AND OTHERS;REEL/FRAME:045420/0161 Effective date: 20180327 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |