US10650501B2 - Method and apparatus for encoding and decoding HDR images - Google Patents

Method and apparatus for encoding and decoding HDR images Download PDF

Info

Publication number
US10650501B2
US10650501B2 US15/122,149 US201515122149A US10650501B2 US 10650501 B2 US10650501 B2 US 10650501B2 US 201515122149 A US201515122149 A US 201515122149A US 10650501 B2 US10650501 B2 US 10650501B2
Authority
US
United States
Prior art keywords
dynamic range
range images
images
high dynamic
tone mapping
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US15/122,149
Other versions
US20160371822A1 (en
Inventor
Mikael LE PENDU
Ronan Boitard
Dominique Thoreau
Christine Guillemot
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
InterDigital VC Holdings Inc
Original Assignee
InterDigital VC Holdings Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from EP14305480.7A external-priority patent/EP2927865A1/en
Application filed by InterDigital VC Holdings Inc filed Critical InterDigital VC Holdings Inc
Publication of US20160371822A1 publication Critical patent/US20160371822A1/en
Assigned to THOMSON LICENSING reassignment THOMSON LICENSING ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GUILLEMOT, CHRISTINE, BOITARD, Ronan, LE PENDU, Mikael, THOREAU, DOMINIQUE
Assigned to INTERDIGITAL VC HOLDINGS, INC. reassignment INTERDIGITAL VC HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: THOMSON LICENSING
Priority to US16/784,124 priority Critical patent/US11727548B2/en
Application granted granted Critical
Publication of US10650501B2 publication Critical patent/US10650501B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/98Adaptive-dynamic-range coding [ADRC]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • G06T5/92Dynamic range modification of images or parts thereof based on global image properties
    • G06T5/009
    • CCHEMISTRY; METALLURGY
    • C10PETROLEUM, GAS OR COKE INDUSTRIES; TECHNICAL GASES CONTAINING CARBON MONOXIDE; FUELS; LUBRICANTS; PEAT
    • C10GCRACKING HYDROCARBON OILS; PRODUCTION OF LIQUID HYDROCARBON MIXTURES, e.g. BY DESTRUCTIVE HYDROGENATION, OLIGOMERISATION, POLYMERISATION; RECOVERY OF HYDROCARBON OILS FROM OIL-SHALE, OIL-SAND, OR GASES; REFINING MIXTURES MAINLY CONSISTING OF HYDROCARBONS; REFORMING OF NAPHTHA; MINERAL WAXES
    • C10G2/00Production of liquid hydrocarbon mixtures of undefined composition from oxides of carbon
    • C10G2/30Production of liquid hydrocarbon mixtures of undefined composition from oxides of carbon from carbon monoxide with hydrogen
    • C10G2/31Production of liquid hydrocarbon mixtures of undefined composition from oxides of carbon from carbon monoxide with hydrogen thermal, non catalytic conversion
    • CCHEMISTRY; METALLURGY
    • C10PETROLEUM, GAS OR COKE INDUSTRIES; TECHNICAL GASES CONTAINING CARBON MONOXIDE; FUELS; LUBRICANTS; PEAT
    • C10GCRACKING HYDROCARBON OILS; PRODUCTION OF LIQUID HYDROCARBON MIXTURES, e.g. BY DESTRUCTIVE HYDROGENATION, OLIGOMERISATION, POLYMERISATION; RECOVERY OF HYDROCARBON OILS FROM OIL-SHALE, OIL-SAND, OR GASES; REFINING MIXTURES MAINLY CONSISTING OF HYDROCARBONS; REFORMING OF NAPHTHA; MINERAL WAXES
    • C10G2/00Production of liquid hydrocarbon mixtures of undefined composition from oxides of carbon
    • C10G2/30Production of liquid hydrocarbon mixtures of undefined composition from oxides of carbon from carbon monoxide with hydrogen
    • C10G2/32Production of liquid hydrocarbon mixtures of undefined composition from oxides of carbon from carbon monoxide with hydrogen with the use of catalysts
    • CCHEMISTRY; METALLURGY
    • C10PETROLEUM, GAS OR COKE INDUSTRIES; TECHNICAL GASES CONTAINING CARBON MONOXIDE; FUELS; LUBRICANTS; PEAT
    • C10GCRACKING HYDROCARBON OILS; PRODUCTION OF LIQUID HYDROCARBON MIXTURES, e.g. BY DESTRUCTIVE HYDROGENATION, OLIGOMERISATION, POLYMERISATION; RECOVERY OF HYDROCARBON OILS FROM OIL-SHALE, OIL-SAND, OR GASES; REFINING MIXTURES MAINLY CONSISTING OF HYDROCARBONS; REFORMING OF NAPHTHA; MINERAL WAXES
    • C10G45/00Refining of hydrocarbon oils using hydrogen or hydrogen-generating compounds
    • C10G45/58Refining of hydrocarbon oils using hydrogen or hydrogen-generating compounds to change the structural skeleton of some of the hydrocarbon content without cracking the other hydrocarbons present, e.g. lowering pour point; Selective hydrocracking of normal paraffins
    • CCHEMISTRY; METALLURGY
    • C10PETROLEUM, GAS OR COKE INDUSTRIES; TECHNICAL GASES CONTAINING CARBON MONOXIDE; FUELS; LUBRICANTS; PEAT
    • C10GCRACKING HYDROCARBON OILS; PRODUCTION OF LIQUID HYDROCARBON MIXTURES, e.g. BY DESTRUCTIVE HYDROGENATION, OLIGOMERISATION, POLYMERISATION; RECOVERY OF HYDROCARBON OILS FROM OIL-SHALE, OIL-SAND, OR GASES; REFINING MIXTURES MAINLY CONSISTING OF HYDROCARBONS; REFORMING OF NAPHTHA; MINERAL WAXES
    • C10G47/00Cracking of hydrocarbon oils, in the presence of hydrogen or hydrogen- generating compounds, to obtain lower boiling fractions
    • CCHEMISTRY; METALLURGY
    • C10PETROLEUM, GAS OR COKE INDUSTRIES; TECHNICAL GASES CONTAINING CARBON MONOXIDE; FUELS; LUBRICANTS; PEAT
    • C10GCRACKING HYDROCARBON OILS; PRODUCTION OF LIQUID HYDROCARBON MIXTURES, e.g. BY DESTRUCTIVE HYDROGENATION, OLIGOMERISATION, POLYMERISATION; RECOVERY OF HYDROCARBON OILS FROM OIL-SHALE, OIL-SAND, OR GASES; REFINING MIXTURES MAINLY CONSISTING OF HYDROCARBONS; REFORMING OF NAPHTHA; MINERAL WAXES
    • C10G49/00Treatment of hydrocarbon oils, in the presence of hydrogen or hydrogen-generating compounds, not provided for in a single one of groups C10G45/02, C10G45/32, C10G45/44, C10G45/58 or C10G47/00
    • CCHEMISTRY; METALLURGY
    • C10PETROLEUM, GAS OR COKE INDUSTRIES; TECHNICAL GASES CONTAINING CARBON MONOXIDE; FUELS; LUBRICANTS; PEAT
    • C10GCRACKING HYDROCARBON OILS; PRODUCTION OF LIQUID HYDROCARBON MIXTURES, e.g. BY DESTRUCTIVE HYDROGENATION, OLIGOMERISATION, POLYMERISATION; RECOVERY OF HYDROCARBON OILS FROM OIL-SHALE, OIL-SAND, OR GASES; REFINING MIXTURES MAINLY CONSISTING OF HYDROCARBONS; REFORMING OF NAPHTHA; MINERAL WAXES
    • C10G5/00Recovery of liquid hydrocarbon mixtures from gases, e.g. natural gas
    • CCHEMISTRY; METALLURGY
    • C10PETROLEUM, GAS OR COKE INDUSTRIES; TECHNICAL GASES CONTAINING CARBON MONOXIDE; FUELS; LUBRICANTS; PEAT
    • C10GCRACKING HYDROCARBON OILS; PRODUCTION OF LIQUID HYDROCARBON MIXTURES, e.g. BY DESTRUCTIVE HYDROGENATION, OLIGOMERISATION, POLYMERISATION; RECOVERY OF HYDROCARBON OILS FROM OIL-SHALE, OIL-SAND, OR GASES; REFINING MIXTURES MAINLY CONSISTING OF HYDROCARBONS; REFORMING OF NAPHTHA; MINERAL WAXES
    • C10G67/00Treatment of hydrocarbon oils by at least one hydrotreatment process and at least one process for refining in the absence of hydrogen only
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • H04N19/126Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/184Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/19Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding using optimisation based on Lagrange multipliers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/88Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving rearrangement of data among different coding units, e.g. shuffling, interleaving, scrambling or permutation of pixel data or permutation of transform coefficient data among different blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20208High dynamic range [HDR] image processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/177Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a group of pictures [GOP]

Definitions

  • This invention relates to a method and an apparatus for encoding and decoding of High Dynamic Range (HDR) images and videos.
  • HDR High Dynamic Range
  • Tools for compressing image sequences are often designed to encode integer data with a limited dynamic range.
  • compression standards such as MPEG-4 AVC/H.264 and HEVC (High Efficiency Video Coding)
  • MPEG-4 AVC/H.264 and HEVC High Efficiency Video Coding
  • HEVC High Efficiency Video Coding
  • data with higher bit depths can be taken as input.
  • a bit depth up to 16 bits is supported.
  • the high bit-depth extensions may require increased implementation and computational costs.
  • image types whose bit depth is higher than 16 bits.
  • An alternative solution for encoding high bit-depth images is to apply a tone mapping operator (TMO) on the HDR images to reduce the bit depth and to generate a Low Dynamic Range (LDR) version of the HDR images.
  • TMO tone mapping operator
  • LDR Low Dynamic Range
  • a low bit-depth encoder can then be used to compress the LDR version.
  • the TMO is usually invertible and the inverse tone mapping should be known to the decoder.
  • An example of this approach is described in an article by Z. Mai et al., titled “On-the-Fly Tone Mapping for Backward-Compatible High Dynamic Range Image/Video Compression,” ISCAS 2010, which defines a tone mapping curve that minimizes the data loss caused by both the tone mapping and the encoder error.
  • the present principles provide a method for decoding HDR images, comprising: accessing a bitstream having the HDR images included therein; decoding Low Dynamic Range (LDR) images from the bitstream; accessing information from the bitstream, the accessed information including a parameter used to encode the LDR images; and generating the HDR images from the LDR images responsive to the accessed parameter as described below.
  • LDR Low Dynamic Range
  • the present principles also provide an apparatus for performing these steps.
  • the present principles also provide a method for encoding HDR images, comprising: determining a tone mapping function responsive to the HDR images and at least one encoding parameter; determining LDR images from the HDR images responsive to the tone mapping function; and encoding the determined LDR images and information indicative of the determined tone mapping function, wherein the at least one encoding parameter is used to encode the LDR images as described below.
  • the present principles also provide an apparatus for performing these steps.
  • the present principles also provide a computer readable storage medium having stored thereon instructions for encoding or decoding HDR images according to the methods described above.
  • the present principles also provide a computer readable storage medium having stored thereon a bitstream generated according to the methods described above.
  • FIG. 1 is a diagram depicting an exemplary mathematical model of a compression scheme, in accordance with an embodiment of the present principles.
  • FIG. 2 is a diagram depicting an exemplary scheme of encoding HDR images using tone mapping, in accordance with an embodiment of the present principles.
  • FIG. 3A is a pictorial example depicting a histogram
  • FIG. 3B is a pictorial example depicting the fitted Gaussian Mixture Model (GMM) for the histogram in FIG. 3A with Expectation Maximization
  • FIG. 3C is a pictorial example depicting the probability distribution function computed with 100 intervals, in accordance with an embodiment of the present principles.
  • GMM Gaussian Mixture Model
  • FIG. 4 is a pictorial example depicting rate distortion results for the encoding of an HDR image over varying lambda values and QP parameters.
  • FIG. 5 is a diagram depicting an exemplary scheme of decoding HDR images using inverse tone mapping, in accordance with an embodiment of the present principles.
  • FIG. 6A shows image “Peniches”
  • FIG. 6B shows the estimated probability distribution function for image “Peniches”
  • FIG. 7A shows image “Mongolfière”
  • FIG. 7B shows the estimated probability distribution function for image “Mongolfière”
  • FIG. 8 is a flow diagram depicting an exemplary method for encoding HDR images, in accordance with an embodiment of the present principles.
  • FIG. 9 is a flow diagram depicting an exemplary method for decoding HDR images, in accordance with an embodiment of the present principles.
  • FIG. 10 is a block diagram depicting an exemplary system for encoding HDR images, in accordance with an embodiment of the present principles.
  • FIG. 11 is a block diagram depicting an exemplary system for decoding HDR images, in accordance with an embodiment of the present principles.
  • FIG. 12 is a block diagram depicting an example of an image processing system that may be used with one or more implementations.
  • FIG. 13 is a block diagram depicting another example of an image processing system that may be used with one or more implementations.
  • a tone mapping function takes a high bit-depth integer as input and returns a low bit-depth integer
  • an inverse tone mapping function takes a low bit-depth integer as input and returns a high bit-depth integer.
  • tone mapping function takes a high bit-depth integer as input and returns a low bit-depth integer
  • inverse tone mapping function takes a low bit-depth integer as input and returns a high bit-depth integer.
  • FIG. 1 we define a mathematical model of a compression scheme, as illustrated in FIG. 1 .
  • the input image I has a probability distribution function (pdf) p and its minimum and maximum pixel values are respectively x min and x max .
  • PDF probability distribution function
  • the compressor function denotes a continuous and strictly monotonous function that has real valued input and output. These properties ensure that F has an inverse function F ⁇ 1 (denoted as an expander function).
  • the compressor function corresponds to a tone mapping function whose output is integer.
  • the tone mapping function can be considered as a concatenation of a compressor function and a uniform quantization (rounding).
  • a random variable ( ⁇ ) is added to each pixel value to model the encoder error introduced when encoding I F and rounding error from format conversion if there is any.
  • a random variable
  • the distributions of the random variables do not depend on the positions or the values of the pixels.
  • random variable ⁇ has a zero mean, and a variance ⁇ 2 .
  • F * ⁇ ( x ) ⁇ x min x ⁇ - 2 ⁇ ⁇ ⁇ 2 . p ⁇ ( t ) c + ⁇ . p ⁇ ( t ) . log 2 ⁇ ( p ⁇ ( t ) ) 3 ⁇ ⁇ d ⁇ ⁇ t ( 2 )
  • FIG. 2 illustrates an exemplary embodiment of encoding HDR images using tone mapping according to the present principles.
  • the pixel values of the input HDR image may be originally represented by high bit-depth integers or floating point values. If the HDR images are represented by floating point values, a conversion to integer can be used. For example, if the input image is defined in the half float format, taking the bit pattern (i.e., 1 sign bit, 5 exponent bits and 10 mantissa bits) interpreted as an integer provides a lossless conversion from floating point to 16-bit integer. If the image contains only positive values, this conversion approximates a log encoding of the original floating point values.
  • n HDR the bit depth of the integer version of the HDR image.
  • the corresponding inverse conversion function should be applied to the decoded integer HDR image in order to convert it back to floating point data.
  • Expectation Maximization is performed to obtain a Gaussian mixture model of the probability distribution function p of the pixel values.
  • the GMM model is described by parameter set ⁇ (mean, variance, weight of the Gaussians).
  • ⁇ , x min and x max (minimum and maximum pixel values of the image are used to compute the tone mapping LUT.
  • a Lagrangian multiplier ⁇ i computed from encoding parameter QP is also taken into account for the LUT computation.
  • the LUT is applied to the image to obtain an LDR version which is sent to an LDR encoder using encoding parameter QP.
  • the parameters ⁇ , x min and x max are indicated in the bitstream representative of the LDR image.
  • the probability distribution function (p(x)) of the pixel values needs to be determined.
  • p(x) can be determined by computing the histogram of the image.
  • the parameters defining the probability distribution function need to be known at the decoder. To reduce the number of parameters that need to be encoded to represent p(x), we may parameterize the histogram.
  • a Gaussian mixture model to fit the histogram, where a GMM is a weighted sum of several Gaussians.
  • the model parameters are variance v j , mean value ⁇ j and weight ⁇ j of each Gaussian j in the mixture model.
  • the Expectation Maximization (EM) algorithm can be used for the fitting operation.
  • This algorithm is defined in an article by Dempster et al., titled “Maximum Likelihood from Incomplete Data via the EM Algorithm,” Journal of the Royal Statistical Society, 1977.
  • Each iteration k of the algorithm is composed of the following steps:
  • FIG. 3A shows an exemplary histogram
  • FIG. 3B shows the fitted GMM with EM.
  • the probability of the pixels beyond the range [x min , x max ] is null.
  • the minimum and maximum values x min and x max are also to be indicated in the bitstream for the decoder to generate the probability distribution function.
  • an image contains a small number of pixels with very low values. As those pixels may disturb the computation of the algorithm, we may clip the values below a given percentile (for example, at 0.1%) before performing the EM algorithm. In that case, the value of x low is used as x min . Similarly, if the image contains a small number of pixels with very high values, we may clip the high values below a given percentile before performing the EM algorithm. In that case, the value of x high is used as x max .
  • ⁇ i may depend on the encoder in use (e.g., HEVC, MPEG-2, MPEG-4 AVC/H.264, or JPEG), the bit depth of the input data format accepted by the LDR encoder, and encoding parameters (e.g., quantization parameter QP in HEVC).
  • the encoder in use e.g., HEVC, MPEG-2, MPEG-4 AVC/H.264, or JPEG
  • encoding parameters e.g., quantization parameter QP in HEVC.
  • the function F*′ can be computed from Eq. (9) given a value ⁇ i and the probability distribution p:
  • F * ′ ⁇ ( x ) - 2 ⁇ . p ⁇ ( x ) - 1 + ⁇ i . p ⁇ ( x ) . log 2 ⁇ ( p ⁇ ( x ) ) 3 ( 9 )
  • F*′ the value of F*′ is computed for each integer value x ⁇ [x min , x max ]
  • F num a cumulative sum can be performed as:
  • LUT lookup table
  • the tone mapping operation can apply the LUT to every pixel of the original HDR image.
  • the image obtained is then compressed with an LDR encoder.
  • the parameters used for the construction of the tone mapping curve (for example, probability distribution function parameters, x min , x max ) need to be communicated to the decoder, for example, using lossless coding.
  • FIG. 5 illustrates an exemplary embodiment of decoding HDR images using inverse tone mapping according to the present principles.
  • the input bitstream of FIG. 5 may be generated according to FIG. 2 .
  • the LDR images and the model parameters e.g., ⁇ , x min , x max and the QP
  • the operations described in Eqs. (5), (9), (10) and (11) can be performed to generate a tone mapping lookup table.
  • the LUT is inverted to generate the inverse tone mapping lookup table.
  • the inverse LUT is applied to the decoded LDR image to reconstruct the HDR image. If the original image is defined in a floating point format, the reconstructed HDR image can be converted back from integer to floating point.
  • FIGS. 6A-6H and 7A-7H illustrate examples of input 16-bit HDR images, probability distribution functions, and tone mapping curves and inversion tone mapping curves for conversion between 16-bit images and 8-bit images, when an HEVC encoder is used as an LDR encoder.
  • FIG. 6B shows the estimated probability distribution function for image “Pouches.”
  • FIG. 7B shows the estimated probability distribution function for image “Mongolfière.”
  • tone mapping curves are smoother at higher QP values (and higher ⁇ i ).
  • the tone mapping curve is more dependent on the probability distribution function, and the slope of the curve is steeper at high probable values. As a consequence, those values will be tone mapped with more precision.
  • FIG. 3A shows the probability distribution function computed with 100 intervals.
  • LDR encoder may also be a base layer encoder of a bit depth scalable video encoder.
  • bitrate used for encoding parameters indicative of the tone mapping function is negligible and is not included in the Rate Distortion minimization problem.
  • the present principles have the advantage of taking both the distortion of the reconstructed HDR image and the rate for encoding the HDR image into account when designing a tone mapping curve with respect to the rate distortion performance.
  • the tone mapping curve and inverse tone mapping curve according to our principles not only depend on the input HDR image characteristics, but also take into consideration of the impact of the encoding settings.
  • we indicate in the bitstream a set of parameters (i.e., ⁇ , x min , x max ) that are inherent to the input images.
  • an encoding parameter for example, a quantization parameter
  • the inverse tone mapping curve necessary for the HDR reconstruction does not need to be transmitted explicitly. Rather, it can be computed by the decoder from the encoding parameter and the set of parameters that are inherent to the input images. Consequently, the methods according to the present principles are adaptive to the encoding settings (for example, the bitrate) and may improve the compression efficiency when encoding HDR images.
  • FIG. 8 illustrates an exemplary method 800 for encoding HDR images according to the present principles.
  • Method 800 starts at initialization step 810 , which may determine the format of the input HDR images and the accepted format of the LDR images.
  • step 820 it models the probability distribution function of HDR images, for example, using Expectation Maximization with GMM or using a light version of the histogram.
  • the Lagrangian multiplier is estimated, for example, as a function of a quantization parameter. Since the Lagrangian multiplier may depend on encoder settings, for example, the video compression standard in use and the bit depth of the format of the LDR encoder, we may communicate the function to compute the Lagrangian multiplier to the decoder, or both the encoder and decoder may store the Lagrangian multiplier as a pre-determined number.
  • a tone mapping function can be generated, for example, according to Eq. (4).
  • a lookup table may be generated.
  • the HDR images are converted to LDR images using the tone mapping function.
  • the LDR images and parameters indicative of the tone mapping function, such as the parameters indicating the probability distribution function, are encoded in the bitstream at step 860 .
  • Method 800 may proceed in a different order from what is shown in FIG. 8 , for example, step 830 may be performed before step 820 . Since the Lagrangian multiplier may not depend on the input images, it may be derived before encoding the input HDR images. Thus, step 830 can be optional.
  • FIG. 9 illustrates an exemplary method 900 for decoding HDR images according to the present principles.
  • the input of method 900 may be a bitstream generated according to method 800 .
  • Method 900 starts at initialization step 910 , which may determine the format of the reconstructed HDR images.
  • the LDR images and parameters indicative of the tone mapping function are decoded. Based on the parameters indicative of the tone mapping function, a tone mapping function can be generated, for example, according to Eq. (4).
  • an inverse tone mapping function can be generated at step 930 . Similar to what is performed at method 800 , a lookup table and an inverse lookup table can be generated to perform numerical operations.
  • the LDR images are converted to HDR images using the inverse tone mapping function.
  • FIG. 10 depicts a block diagram of an exemplary encoder 1000 for encoding HDR images.
  • the input of encoder 1000 includes the HDR images to be encoded, encoding parameters, and/or other images that may be used to estimate the Lagrangian multiplier.
  • Probability estimator 1010 estimates the probability distribution function of the input images.
  • Tone curve estimator 1020 estimates the Lagrangian multiplier, for example, based on a quantization parameter, and generates the tone mapping curve based on probability distribution function parameters and the Lagrangian multiplier, for example, using Eq. (4).
  • LDR image generator 1030 converts the input HDR images into LDR images.
  • LDR encoder 1040 encodes the LDR images and tone curve parameter encoder 1050 encodes the parameters indicative of the tone mapping function, such as probability distribution function parameters and the minimum and maximum pixel values of the input images. If parameters for describing Lagrangian multiplier are to be indicated in the bitstream, these parameters will also be encoded by tone curve parameter encoder 1050 .
  • Tone curve parameter encoder 1050 can be a standalone module, or it can be included within LDR encoder 1040 .
  • FIG. 11 depicts a block diagram of an exemplary decoder 1100 for decoding HDR images.
  • the input of decoder 1100 includes a bitstream to be decoded.
  • the input bitstream may be generated by encoder 1000 .
  • LDR decoder 1100 decodes the LDR images and tone curve parameter decoder 1120 decodes the parameters indicative of the tone mapping function, such as probability distribution function parameters and the minimum and maximum pixel values of the input images.
  • Tone curve parameter decoder 1120 can be a standalone module, or it can be included within LDR decoder 1110 .
  • Inverse tone curve estimator 1130 estimates parameters, such as the Lagrangian multiplier based on a quantization parameter decoded from the bitstream, and generates the tone mapping curve, for example, using Eq. (4). Based on the inverse tone mapping curve, HDR image generator 1140 converts the decoded LDR images to HDR images.
  • the data transmission system or apparatus 1200 may be, for example, a head-end or transmission system for transmitting a signal using any of a variety of media, such as, for example, satellite, cable, telephone-line, or terrestrial broadcast.
  • the data transmission system or apparatus 1200 also, or alternatively, may be used, for example, to provide a signal for storage.
  • the transmission may be provided over the Internet or some other network.
  • the data transmission system or apparatus 1200 is capable of generating and delivering, for example, video content and other contents.
  • the data transmission system or apparatus 1200 receives processed data and other information from a processor 1201 .
  • the processor 1201 converts HDR images to LDR images.
  • the data transmission system or apparatus 1200 includes an encoder 1202 and a transmitter 1204 capable of transmitting the encoded signal.
  • the encoder 1202 receives data information from the processor 1201 .
  • the encoder 1202 generates an encoded signal(s).
  • the encoder 1202 includes the processor 1201 and therefore performs the operations of the processor 1201 .
  • the transmitter 1204 receives the encoded signal(s) from the encoder 1202 and transmits the encoded signal(s) in one or more output signals.
  • the transmitter 1204 may be, for example, adapted to transmit a program signal having one or more bitstreams representing encoded pictures and/or information related thereto.
  • Typical transmitters perform functions such as, for example, one or more of providing error-correction coding, interleaving the data in the signal, randomizing the energy in the signal, and modulating the signal onto one or more carriers using a modulator 12012 .
  • the transmitter 1204 may include, or interface with, an antenna (not shown). Further, implementations of the transmitter 1204 may be limited to the modulator 12012 .
  • the data transmission system or apparatus 1200 is also communicatively coupled to a storage unit 1208 .
  • the storage unit 1208 is coupled to the encoder 1202 , and stores an encoded bitstream from the encoder 1202 .
  • the storage unit 1208 is coupled to the transmitter 1204 , and stores a bitstream from the transmitter 1204 .
  • the bitstream from the transmitter 1204 may include, for example, one or more encoded bitstream that has been further processed by the transmitter 1204 .
  • the storage unit 1208 is, in different implementations, one or more of a standard DVD, a Blu-Ray disc, a hard drive, or some other storage device.
  • the data receiving system or apparatus 1300 may be configured to receive signals over a variety of media, such as, for example, storage device, satellite, cable, telephone-line, or terrestrial broadcast.
  • the signals may be received over the Internet or some other network.
  • the data receiving system or apparatus 1300 may be, for example, a cell-phone, a computer, a set-top box, a television, or other device that receives encoded video and provides, for example, decoded video signal for display (display to a user, for example), for processing, or for storage.
  • the data receiving system or apparatus 1300 may provide its output to, for example, a screen of a television, a computer monitor, a computer (for storage, processing, or display), or some other storage, processing, or display device.
  • the data receiving system or apparatus 1300 includes a receiver 1302 for receiving an encoded signal, such as, for example, the signals described in the implementations of this application.
  • the receiver 1302 may receive, for example, a signal output from the data transmission system 1200 of FIG. 12 .
  • the receiver 1302 may be, for example, adapted to receive a program signal having a plurality of bitstreams representing encoded pictures. Typical receivers perform functions such as, for example, one or more of receiving a modulated and encoded data signal, demodulating the data signal from one or more carriers using a demodulator 1304 , de-randomizing the energy in the signal, de-interleaving the data in the signal, and error-correction decoding the signal.
  • the receiver 1302 may include, or interface with, an antenna (not shown). Implementations of the receiver 1302 may be limited to the demodulator 1304 .
  • the data receiving system or apparatus 1300 includes a decoder 1306 .
  • the receiver 1302 provides a received signal to the decoder 1306 .
  • the signal provided to the decoder 1306 by the receiver 1302 may include one or more encoded bitstreams.
  • the decoder 1306 outputs a decoded signal, such as, for example, decoded video signals including video information.
  • the decoder 1306 may be, for example, decoder 1100 described in FIG. 11 .
  • the data receiving system or apparatus 1300 is also communicatively coupled to a storage unit 1307 .
  • the storage unit 1307 is coupled to the receiver 1302 , and the receiver 1302 accesses a bitstream from the storage unit 1307 .
  • the storage unit 1307 is coupled to the decoder 1306 , and the decoder 1306 accesses a bitstream from the storage unit 1307 .
  • the bitstream accessed from the storage unit 1307 includes, in different implementations, one or more encoded bitstreams.
  • the storage unit 1307 is, in different implementations, one or more of a standard DVD, a Blu-Ray disc, a hard drive, or some other storage device.
  • the output data from the decoder 1306 is provided, in one implementation, to a processor 1308 .
  • the decoder 1306 includes the processor 1308 and therefore performs the operations of the processor 1308 .
  • the processor 1308 is part of a downstream device such as, for example, a set-top box or a television.
  • the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed may also be implemented in other forms (for example, an apparatus or program).
  • An apparatus and constituents included therein, for example, a processor, an encoder and a decoder may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants (“PDAs”), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • the appearances of the phrase “in one embodiment” or “in an embodiment” or “in one implementation” or “in an implementation”, as well any other variations, appearing in various places throughout the specification are not necessarily all referring to the same embodiment.
  • Determining the information may include one or more of, for example, estimating the information, calculating the information, predicting the information, or retrieving the information from memory.
  • Accessing the information may include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
  • Receiving is, as with “accessing”, intended to be a broad term.
  • Receiving the information may include one or more of, for example, accessing the information, or retrieving the information (for example, from memory).
  • “receiving” is typically involved, in one way or another, during operations such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
  • implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted.
  • the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
  • a signal may be formatted to carry the bitstream of a described embodiment.
  • Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
  • the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
  • the information that the signal carries may be, for example, analog or digital information.
  • the signal may be transmitted over a variety of different wired or wireless links, as is known.
  • the signal may be stored on a processor-readable medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Chemical & Material Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Oil, Petroleum & Natural Gas (AREA)
  • Organic Chemistry (AREA)
  • Chemical Kinetics & Catalysis (AREA)
  • General Chemical & Material Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Crystallography & Structural Chemistry (AREA)
  • Thermal Sciences (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Processing (AREA)
  • Picture Signal Circuits (AREA)
  • Facsimile Image Signal Circuits (AREA)
  • Algebra (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)

Abstract

To encode High Dynamic Range (HDR) images, the HDR images can be converted to Low Dynamic Range (LDR) images through tone mapping operation, and the LDR images can be encoded with an LDR encoder. The present principles formulates a rate distortion minimization problem when designing the tone mapping curve. In particular, the tone mapping curve is formulated as a function of the probability distribution function of the HDR images to be encoded and a Lagrangian multiplier that depends on encoding parameters. At the decoder, based on the parameters indicative of the tone mapping function, an inverse tone mapping function can be derived to reconstruct HDR images from decoded LDR images.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application claims the benefit, under 35 U.S.C. § 365 of International Application of PCT/EP2015/053747, filed Feb. 23, 2015, which was published in accordance with PCT Article 21(2) on Sep. 3, 2015, in English, and which claims the benefit of European patent application No. 14305266.0, filed Feb. 26, 2014, and European patent application No. 14305480.7, filed Apr. 1, 2014.
TECHNICAL FIELD
This invention relates to a method and an apparatus for encoding and decoding of High Dynamic Range (HDR) images and videos.
BACKGROUND
Tools for compressing image sequences are often designed to encode integer data with a limited dynamic range. For example, compression standards, such as MPEG-4 AVC/H.264 and HEVC (High Efficiency Video Coding), can be used to compress images and videos whose color components are represented by 8 bit integers. In extended versions of these standards, data with higher bit depths can be taken as input. For example, in the Range Extension version of HEVC, a bit depth up to 16 bits is supported. The high bit-depth extensions may require increased implementation and computational costs. Moreover, for some applications such as 3D computer graphics, there also exist image types whose bit depth is higher than 16 bits.
An alternative solution for encoding high bit-depth images is to apply a tone mapping operator (TMO) on the HDR images to reduce the bit depth and to generate a Low Dynamic Range (LDR) version of the HDR images. A low bit-depth encoder can then be used to compress the LDR version. In this approach, the TMO is usually invertible and the inverse tone mapping should be known to the decoder. An example of this approach is described in an article by Z. Mai et al., titled “On-the-Fly Tone Mapping for Backward-Compatible High Dynamic Range Image/Video Compression,” ISCAS 2010, which defines a tone mapping curve that minimizes the data loss caused by both the tone mapping and the encoder error.
SUMMARY
The present principles provide a method for decoding HDR images, comprising: accessing a bitstream having the HDR images included therein; decoding Low Dynamic Range (LDR) images from the bitstream; accessing information from the bitstream, the accessed information including a parameter used to encode the LDR images; and generating the HDR images from the LDR images responsive to the accessed parameter as described below. The present principles also provide an apparatus for performing these steps.
The present principles also provide a method for encoding HDR images, comprising: determining a tone mapping function responsive to the HDR images and at least one encoding parameter; determining LDR images from the HDR images responsive to the tone mapping function; and encoding the determined LDR images and information indicative of the determined tone mapping function, wherein the at least one encoding parameter is used to encode the LDR images as described below. The present principles also provide an apparatus for performing these steps.
The present principles also provide a computer readable storage medium having stored thereon instructions for encoding or decoding HDR images according to the methods described above.
The present principles also provide a computer readable storage medium having stored thereon a bitstream generated according to the methods described above.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a diagram depicting an exemplary mathematical model of a compression scheme, in accordance with an embodiment of the present principles.
FIG. 2 is a diagram depicting an exemplary scheme of encoding HDR images using tone mapping, in accordance with an embodiment of the present principles.
FIG. 3A is a pictorial example depicting a histogram, FIG. 3B is a pictorial example depicting the fitted Gaussian Mixture Model (GMM) for the histogram in FIG. 3A with Expectation Maximization, and FIG. 3C is a pictorial example depicting the probability distribution function computed with 100 intervals, in accordance with an embodiment of the present principles.
FIG. 4 is a pictorial example depicting rate distortion results for the encoding of an HDR image over varying lambda values and QP parameters.
FIG. 5 is a diagram depicting an exemplary scheme of decoding HDR images using inverse tone mapping, in accordance with an embodiment of the present principles.
FIG. 6A shows image “Peniches,” FIG. 6B shows the estimated probability distribution function for image “Peniches,” FIGS. 6C and 6D show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=0, FIGS. 6E and 6F show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=16, and FIGS. 6G and 6H show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=32.
FIG. 7A shows image “Mongolfière,” FIG. 7B shows the estimated probability distribution function for image “Mongolfière,” FIGS. 7C and 7D show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=0, FIGS. 7E and 7F show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=16, and FIGS. 7G and 7H show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=32.
FIG. 8 is a flow diagram depicting an exemplary method for encoding HDR images, in accordance with an embodiment of the present principles.
FIG. 9 is a flow diagram depicting an exemplary method for decoding HDR images, in accordance with an embodiment of the present principles.
FIG. 10 is a block diagram depicting an exemplary system for encoding HDR images, in accordance with an embodiment of the present principles.
FIG. 11 is a block diagram depicting an exemplary system for decoding HDR images, in accordance with an embodiment of the present principles.
FIG. 12 is a block diagram depicting an example of an image processing system that may be used with one or more implementations.
FIG. 13 is a block diagram depicting another example of an image processing system that may be used with one or more implementations.
DETAILED DESCRIPTION
The present principles are directed to encoding and decoding of HDR images and videos using tone mapping techniques. A tone mapping function takes a high bit-depth integer as input and returns a low bit-depth integer, and an inverse tone mapping function takes a low bit-depth integer as input and returns a high bit-depth integer. In the present application, we use the terms “tone mapping function,” “tone mapping curve,” and “tone curve” interchangeably, and use the terms “inverse tone mapping function” and “inverse tone mapping curve” interchangeably.
In one embodiment, we define a mathematical model of a compression scheme, as illustrated in FIG. 1. In this model, we consider an image is compressed and that pixels have real values (not necessarily integers). The input image I has a probability distribution function (pdf) p and its minimum and maximum pixel values are respectively xmin and xmax.
First, a compressor function F is applied to the pixel values, where F is defined on the interval [xmin, xmax] and F(xmin)=0 and F(xmax)=2n−1, where n is the bit depth of the LDR image to be encoded by an LDR encoder. In this mathematical model, the compressor function denotes a continuous and strictly monotonous function that has real valued input and output. These properties ensure that F has an inverse function F−1 (denoted as an expander function). When the mathematical model is applied to encoding HDR images, the compressor function corresponds to a tone mapping function whose output is integer. The tone mapping function can be considered as a concatenation of a compressor function and a uniform quantization (rounding). After applying function F, theoretically no data is lost since F is mathematically invertible. We denote the output image from the compressor function as IF, and the probability distribution function of IF as pF.
Then a random variable (ε) is added to each pixel value to model the encoder error introduced when encoding IF and rounding error from format conversion if there is any. Here we suppose that the distributions of the random variables do not depend on the positions or the values of the pixels. We also assume that random variable ε has a zero mean, and a variance σ2. After IF is encoded, the expander F1 is applied to reconstruct HDR image Irec.
In the mathematical model, we assume real values are used as input and output. When the mathematical model is actually applied, the input is usually integer, and the output is also integer because we need to generate low bit-depth (for example, 8 bits) integer values to be encoded by an LDR encoder.
To design a tone mapping curve for the model shown in FIG. 1, we formulate a rate distortion function J=Dtot0·R, where Dtot is the total distortion between the original HDR image I and reconstructed HDR image Irec, R is the bit rate of encoding the LDR image, and λ0 is a Lagrangian multiplier that is adjusted to improve rate distortion performance. Given an image I and an encoder, we assume that bit rate R of the encoded image is proportional to the entropy of IF. Thus, minimizing Dtot0·R is equivalent to minimizing Dtot+λ·entropy(IF), where λ is another Lagrangian multiplier.
In one embodiment, we assume the distortion is measured by mean square error (MSE), that is, Dtot=E{(Irec−I)2}, where E{·} calculates the expectation of a random variable. Our goal is to find function F* that minimizes the rate distortion function Dtot+λ·entropy(IF). It can be shown analytically that the derivative of function F* can be formulated as:
F * ( x ) = - 2 σ 2 . p ( x ) c + λ . p ( x ) . log 2 ( p ( x ) ) 3 ( 1 )
Thus,
F * ( x ) = x min x - 2 σ 2 . p ( t ) c + λ . p ( t ) . log 2 ( p ( t ) ) 3 d t ( 2 )
where c is a constant that can be adjusted so that F*(xmax)=2n−1.
However, we don't have an analytical solution to determine the value of c given λ and σ. Moreover, a model of the actual encoder used and the encoding parameters (for example, but not limited to, quantization parameter QP in HEVC and bit depth of the LDR image) would be needed to find the value of σ. To eliminate the dependence of function F*(x) on variables c and σ, we define a function
S ( x , λ i ) = x min x - 2 . p ( t ) - 1 + λ i . p ( t ) . log 2 ( p ( t ) ) 3 d t ( 3 )
It can be shown that for any positive value λ and σ, there exist a value λi
Figure US10650501-20200512-P00001
such that:
x [ x min , x max ] , F * ( x ) = ( 2 n - 1 ) S ( x , λ i ) S ( x max , λ i ) ( 4 )
Consequently, only parameter λi is needed to derive the tone mapping function. In one embodiment, we can compute S(x, λi) by numerical integration and divide the result by S(xmax, λi).
FIG. 2 illustrates an exemplary embodiment of encoding HDR images using tone mapping according to the present principles. The pixel values of the input HDR image may be originally represented by high bit-depth integers or floating point values. If the HDR images are represented by floating point values, a conversion to integer can be used. For example, if the input image is defined in the half float format, taking the bit pattern (i.e., 1 sign bit, 5 exponent bits and 10 mantissa bits) interpreted as an integer provides a lossless conversion from floating point to 16-bit integer. If the image contains only positive values, this conversion approximates a log encoding of the original floating point values.
In another example, we may apply a logarithm function to the floating point values and round the result to integers. In that case, the function used should be adjusted so that the zero and the maximum floating point values are mapped respectively to 0 and 2n HDR −1, where nHDR is the bit depth of the integer version of the HDR image. A reasonable choice for the value of nHDR is 16 if the original image is in the half float format (i.e., 16 bits floating point values), or nHDR=32 if the original image is in the single precision float format (i.e., 32 bits floating point values). The corresponding inverse conversion function should be applied to the decoded integer HDR image in order to convert it back to floating point data.
Then, Expectation Maximization is performed to obtain a Gaussian mixture model of the probability distribution function p of the pixel values. The GMM model is described by parameter set Θ (mean, variance, weight of the Gaussians). Θ, xmin and xmax (minimum and maximum pixel values of the image are used to compute the tone mapping LUT. A Lagrangian multiplier λi computed from encoding parameter QP is also taken into account for the LUT computation. Then, the LUT is applied to the image to obtain an LDR version which is sent to an LDR encoder using encoding parameter QP. Finally, the parameters Θ, xmin and xmax are indicated in the bitstream representative of the LDR image. In the following, we describe the determination of the probability distribution function and the Lagrangian multiplier in further detail.
Probability Distribution Function Determination
To derive the tone mapping function as described in Eq. (4), the probability distribution function (p(x)) of the pixel values needs to be determined. In one example, p(x) can be determined by computing the histogram of the image. In order for the decoder to compute the inverse tone mapping curve, the parameters defining the probability distribution function need to be known at the decoder. To reduce the number of parameters that need to be encoded to represent p(x), we may parameterize the histogram.
In one embodiment, we can use a Gaussian mixture model to fit the histogram, where a GMM is a weighted sum of several Gaussians. The model parameters are variance vj, mean value μj and weight αj of each Gaussian j in the mixture model. The model of the probability distribution function can be formulated as:
p(x)=Σj=1 mαj g(x,μ jj)  (5)
where m is the number of Gaussians used in the model, and g is a Gaussian function:
g ( x , μ j , v j ) = 1 2 π v j · e - ( x - μ j ) 2 2 v j ( 6 )
The Expectation Maximization (EM) algorithm can be used for the fitting operation. This algorithm is defined in an article by Dempster et al., titled “Maximum Likelihood from Incomplete Data via the EM Algorithm,” Journal of the Royal Statistical Society, 1977. Given a number m of Gaussian models and an initial parameter set θ0=(αj 0, μj 0, νj 0) (j∈[1, m]), the objective is to find θ that maximizes the log-likelihood function:
J(θ)=ln Πi=1 N p(x i)  (7)
where N is the number of pixels and xi is the value of the pixel i.
Each iteration k of the algorithm is composed of the following steps:
1. For each j in [ 1 , m ] , and i in [ 1 , N ] , compute : γ i , j = α j k - 1 g ( x i , μ j k - 1 , v j k - 1 ) l = 1 m α l k - 1 g ( x i , μ l k - 1 , v l k - 1 ) and N j = i = 1 N γ i , j
2. Compute the new means : μ j k = 1 N j i = 1 N γ i , j x i
3. Compute the new variance : v j k = 1 N j i = 1 N γ i , j ( x i - μ j k ) 2
4. Compute the new weights : α j k = N j N
5. Stop if convergence is reached (i.e. |J(θk−1) − J(θk)| < Threshold),
(e.g., Threshold = 10−9). Otherwise, do next iteration.
The GMM parameters obtained by EM can then be used to compute the probability distribution function at any integer value x from xmin to xmax with Eq. (5). FIG. 3A shows an exemplary histogram, and FIG. 3B shows the fitted GMM with EM.
We consider that the probability of the pixels beyond the range [xmin, xmax] is null. Thus, in addition to the GMM parameters, the minimum and maximum values xmin and xmax are also to be indicated in the bitstream for the decoder to generate the probability distribution function.
In HDR images, it is possible that an image contains a small number of pixels with very low values. As those pixels may disturb the computation of the algorithm, we may clip the values below a given percentile (for example, at 0.1%) before performing the EM algorithm. In that case, the value of xlow is used as xmin. Similarly, if the image contains a small number of pixels with very high values, we may clip the high values below a given percentile before performing the EM algorithm. In that case, the value of xhigh is used as xmax.
Lagrangian Multiplier Determination
To derive the tone mapping function as described in Eq. (4), we also need to determine the Lagrangian multiplier λi. The value of λi may depend on the encoder in use (e.g., HEVC, MPEG-2, MPEG-4 AVC/H.264, or JPEG), the bit depth of the input data format accepted by the LDR encoder, and encoding parameters (e.g., quantization parameter QP in HEVC). In the following, we discuss how to determine Lagrangian multiplier λi in further detail.
In one embodiment, we may encode several images with several quality parameters (e.g., QP value for HEVC) and λi values. For a given image at each QP, the encoding is performed several times by varying the value of λi over a large range. Given a QP value, the Rate Distortion (RD) point obtained with the optimal λi is on the convex hull of the set of all the RD points, as illustrated in FIG. 4. Thus, we can derive the Lagrangian multiplier λi* as a function of QP. For example, when 16-bit images are tone mapped to 8-bit and encoded with HEVC, the following exponential function was found to represent the Lagrangian multiplier:
λi*=100*20.37QP  (8)
Given the function described in Eq. (8), the same value of λi can be derived at the encoder and the decoder, since the QP value is known by the decoder.
Lookup Table Generation
In one embodiment, we may need to perform numerical integration to implement the present principles. Since computing the sum for every pixel would be very time consuming, we may compute the sum, as shown below in Eq. (10) for a value x by cumulating the result at x−1. Other methods may be used to perform numerical operations.
As explained before, the function F*′ can be computed from Eq. (9) given a value λi and the probability distribution p:
F * ( x ) = - 2 . p ( x ) - 1 + λ i . p ( x ) . log 2 ( p ( x ) ) 3 ( 9 )
When the value of F*′ is computed for each integer value x∈[xmin, xmax], it can be integrated numerically to compute Fnum, which approximates the function S defined in Eq. (3). For example, a cumulative sum can be performed as:
F num(x)=Σi=x min x-1 F*′  (10)
We may generate a lookup table (LUT) to represent the tone mapping curve. In particular, function F may be scaled and rounded to obtain integers with the required bit depth n for the LDR image:
LUT ( x ) [ ( 2 n - 1 ) F num ( x ) F num ( x max ) ] ( 11 )
Note that
F num ( x ) F num ( x max )
is the numerical approximation of
S ( x , λ i ) S ( x max , λ i )
from Eq. (4). After the LUT is generated, the tone mapping operation can apply the LUT to every pixel of the original HDR image. The image obtained is then compressed with an LDR encoder. The parameters used for the construction of the tone mapping curve (for example, probability distribution function parameters, xmin, xmax) need to be communicated to the decoder, for example, using lossless coding.
FIG. 5 illustrates an exemplary embodiment of decoding HDR images using inverse tone mapping according to the present principles. The input bitstream of FIG. 5 may be generated according to FIG. 2. The LDR images and the model parameters (e.g., Θ, xmin, xmax and the QP) can be decoded from the bitstream. Knowing the parameters, the operations described in Eqs. (5), (9), (10) and (11) can be performed to generate a tone mapping lookup table. Then the LUT is inverted to generate the inverse tone mapping lookup table. The inverse LUT is applied to the decoded LDR image to reconstruct the HDR image. If the original image is defined in a floating point format, the reconstructed HDR image can be converted back from integer to floating point.
FIGS. 6A-6H and 7A-7H illustrate examples of input 16-bit HDR images, probability distribution functions, and tone mapping curves and inversion tone mapping curves for conversion between 16-bit images and 8-bit images, when an HEVC encoder is used as an LDR encoder. FIG. 6A shows image “Péniches,” with xmin=0 and xmax=27024. FIG. 6B shows the estimated probability distribution function for image “Péniches.” FIGS. 6C and 6D show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=0 (λi=100). FIGS. 6E and 6F show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=16 (λi=6055). FIGS. 6G and 6H show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=32 (λi=3.67e+05).
FIG. 7A shows image “Mongolfière,” with xmin=1834 and xmax=23634. FIG. 7B shows the estimated probability distribution function for image “Mongolfière.” FIGS. 7C and 7D show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=0 (λi=100). FIGS. 7E and 7F show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=16 (λi=6055). FIGS. 7G and 7H show tone mapping LUT and inverse tone mapping LUT, respectively, when QP=32 (λi=3.67e+05).
We observe from FIGS. 6A-6H and 7A-7H that the tone mapping curves are smoother at higher QP values (and higher λi). At lower QPs, the tone mapping curve is more dependent on the probability distribution function, and the slope of the curve is steeper at high probable values. As a consequence, those values will be tone mapped with more precision.
When determining the tone mapping curve at the encoder, we discussed above that MSE can be used to measure the distortion and Expectation Maximization with GMM can be used to estimate the probability distribution. In a variation, instead of the mean square error, we can choose the sum of absolute difference (SAD) as a distortion metric. In this case, the cube root in Eqs. (1)-(3) will be replaced by a square root.
In another variation, we may compute a light version of the histogram and model the probability distribution function with a few parameters. Specifically, instead of computing the full histogram containing the number of pixels of the image at each integer value, we can divide the range [xmin, xmax] into a small number of intervals of the same length. For each interval K, the number NK of pixels of the image falling into this interval can be computed. Then for each value x in the interval K the probability is estimated by
p ( x ) = N K total number of pixels .
For example, we can take 100 intervals, and 100 values NK (K∈[1,100]) and the values xmin and xmax need to be communicated to the decoder. For the exemplary histogram shown in FIG. 3A, FIG. 3C shows the probability distribution function computed with 100 intervals.
Various encoders, for example, encoders conforming to MPEG-2, MPEG-4 AVC/H.264, and HEVC can be used as an LDR encoder. The LDR encoder may also be a base layer encoder of a bit depth scalable video encoder. In the above examples, we assume that the bitrate used for encoding parameters indicative of the tone mapping function is negligible and is not included in the Rate Distortion minimization problem. We may also include the bitrate for these parameters in our model if it becomes more significant. In the above, we have used one image as an input. The present principles can also be applied when a video sequence is used as input and we may vary the parameters from image to image.
Several numerical values have been used above, for example, to provide examples for λi computation, histogram intervals, and the very low and high pixel values with a small percentage. These numerical values can be changed as the input images and/or encoding settings change.
The present principles have the advantage of taking both the distortion of the reconstructed HDR image and the rate for encoding the HDR image into account when designing a tone mapping curve with respect to the rate distortion performance. The tone mapping curve and inverse tone mapping curve according to our principles not only depend on the input HDR image characteristics, but also take into consideration of the impact of the encoding settings. In one embodiment, we indicate in the bitstream a set of parameters (i.e., Θ, xmin, xmax) that are inherent to the input images. Based on this set of inherent parameters and an encoding parameter (for example, a quantization parameter), we can generate different sets of tone mapping curves and inverse tone mapping curves adaptive to the quantization parameter, and thus the bitrate. The inverse tone mapping curve necessary for the HDR reconstruction does not need to be transmitted explicitly. Rather, it can be computed by the decoder from the encoding parameter and the set of parameters that are inherent to the input images. Consequently, the methods according to the present principles are adaptive to the encoding settings (for example, the bitrate) and may improve the compression efficiency when encoding HDR images.
FIG. 8 illustrates an exemplary method 800 for encoding HDR images according to the present principles. Method 800 starts at initialization step 810, which may determine the format of the input HDR images and the accepted format of the LDR images. At step 820, it models the probability distribution function of HDR images, for example, using Expectation Maximization with GMM or using a light version of the histogram.
At step 830, the Lagrangian multiplier is estimated, for example, as a function of a quantization parameter. Since the Lagrangian multiplier may depend on encoder settings, for example, the video compression standard in use and the bit depth of the format of the LDR encoder, we may communicate the function to compute the Lagrangian multiplier to the decoder, or both the encoder and decoder may store the Lagrangian multiplier as a pre-determined number.
Based on the probability distribution function modelled at step 820 and the Lagrangian multiplier estimated at step 830, a tone mapping function can be generated, for example, according to Eq. (4). To perform numerical operation, a lookup table may be generated. At step 850, the HDR images are converted to LDR images using the tone mapping function. The LDR images and parameters indicative of the tone mapping function, such as the parameters indicating the probability distribution function, are encoded in the bitstream at step 860.
Method 800 may proceed in a different order from what is shown in FIG. 8, for example, step 830 may be performed before step 820. Since the Lagrangian multiplier may not depend on the input images, it may be derived before encoding the input HDR images. Thus, step 830 can be optional.
FIG. 9 illustrates an exemplary method 900 for decoding HDR images according to the present principles. The input of method 900 may be a bitstream generated according to method 800. Method 900 starts at initialization step 910, which may determine the format of the reconstructed HDR images. At step 920, the LDR images and parameters indicative of the tone mapping function are decoded. Based on the parameters indicative of the tone mapping function, a tone mapping function can be generated, for example, according to Eq. (4). Subsequently, an inverse tone mapping function can be generated at step 930. Similar to what is performed at method 800, a lookup table and an inverse lookup table can be generated to perform numerical operations. At step 940, the LDR images are converted to HDR images using the inverse tone mapping function.
FIG. 10 depicts a block diagram of an exemplary encoder 1000 for encoding HDR images. The input of encoder 1000 includes the HDR images to be encoded, encoding parameters, and/or other images that may be used to estimate the Lagrangian multiplier. Probability estimator 1010 estimates the probability distribution function of the input images. Tone curve estimator 1020 estimates the Lagrangian multiplier, for example, based on a quantization parameter, and generates the tone mapping curve based on probability distribution function parameters and the Lagrangian multiplier, for example, using Eq. (4). Based on the tone mapping curve, LDR image generator 1030 converts the input HDR images into LDR images. LDR encoder 1040 encodes the LDR images and tone curve parameter encoder 1050 encodes the parameters indicative of the tone mapping function, such as probability distribution function parameters and the minimum and maximum pixel values of the input images. If parameters for describing Lagrangian multiplier are to be indicated in the bitstream, these parameters will also be encoded by tone curve parameter encoder 1050. Tone curve parameter encoder 1050 can be a standalone module, or it can be included within LDR encoder 1040.
FIG. 11 depicts a block diagram of an exemplary decoder 1100 for decoding HDR images. The input of decoder 1100 includes a bitstream to be decoded. The input bitstream may be generated by encoder 1000. LDR decoder 1100 decodes the LDR images and tone curve parameter decoder 1120 decodes the parameters indicative of the tone mapping function, such as probability distribution function parameters and the minimum and maximum pixel values of the input images. Tone curve parameter decoder 1120 can be a standalone module, or it can be included within LDR decoder 1110.
Inverse tone curve estimator 1130 estimates parameters, such as the Lagrangian multiplier based on a quantization parameter decoded from the bitstream, and generates the tone mapping curve, for example, using Eq. (4). Based on the inverse tone mapping curve, HDR image generator 1140 converts the decoded LDR images to HDR images.
Referring now to FIG. 12, a data transmission system or apparatus 1200 is shown, to which the features and principles described above may be applied. The data transmission system or apparatus 1200 may be, for example, a head-end or transmission system for transmitting a signal using any of a variety of media, such as, for example, satellite, cable, telephone-line, or terrestrial broadcast. The data transmission system or apparatus 1200 also, or alternatively, may be used, for example, to provide a signal for storage. The transmission may be provided over the Internet or some other network. The data transmission system or apparatus 1200 is capable of generating and delivering, for example, video content and other contents.
The data transmission system or apparatus 1200 receives processed data and other information from a processor 1201. In one implementation, the processor 1201 converts HDR images to LDR images. The data transmission system or apparatus 1200 includes an encoder 1202 and a transmitter 1204 capable of transmitting the encoded signal. The encoder 1202 receives data information from the processor 1201. The encoder 1202 generates an encoded signal(s). In some implementations, the encoder 1202 includes the processor 1201 and therefore performs the operations of the processor 1201.
The transmitter 1204 receives the encoded signal(s) from the encoder 1202 and transmits the encoded signal(s) in one or more output signals. The transmitter 1204 may be, for example, adapted to transmit a program signal having one or more bitstreams representing encoded pictures and/or information related thereto. Typical transmitters perform functions such as, for example, one or more of providing error-correction coding, interleaving the data in the signal, randomizing the energy in the signal, and modulating the signal onto one or more carriers using a modulator 12012. The transmitter 1204 may include, or interface with, an antenna (not shown). Further, implementations of the transmitter 1204 may be limited to the modulator 12012.
The data transmission system or apparatus 1200 is also communicatively coupled to a storage unit 1208. In one implementation, the storage unit 1208 is coupled to the encoder 1202, and stores an encoded bitstream from the encoder 1202. In another implementation, the storage unit 1208 is coupled to the transmitter 1204, and stores a bitstream from the transmitter 1204. The bitstream from the transmitter 1204 may include, for example, one or more encoded bitstream that has been further processed by the transmitter 1204. The storage unit 1208 is, in different implementations, one or more of a standard DVD, a Blu-Ray disc, a hard drive, or some other storage device.
Referring now to FIG. 13, a data receiving system or apparatus 1300 is shown to which the features and principles described above may be applied. The data receiving system or apparatus 1300 may be configured to receive signals over a variety of media, such as, for example, storage device, satellite, cable, telephone-line, or terrestrial broadcast. The signals may be received over the Internet or some other network.
The data receiving system or apparatus 1300 may be, for example, a cell-phone, a computer, a set-top box, a television, or other device that receives encoded video and provides, for example, decoded video signal for display (display to a user, for example), for processing, or for storage. Thus, the data receiving system or apparatus 1300 may provide its output to, for example, a screen of a television, a computer monitor, a computer (for storage, processing, or display), or some other storage, processing, or display device.
The data receiving system or apparatus 1300 includes a receiver 1302 for receiving an encoded signal, such as, for example, the signals described in the implementations of this application. The receiver 1302 may receive, for example, a signal output from the data transmission system 1200 of FIG. 12.
The receiver 1302 may be, for example, adapted to receive a program signal having a plurality of bitstreams representing encoded pictures. Typical receivers perform functions such as, for example, one or more of receiving a modulated and encoded data signal, demodulating the data signal from one or more carriers using a demodulator 1304, de-randomizing the energy in the signal, de-interleaving the data in the signal, and error-correction decoding the signal. The receiver 1302 may include, or interface with, an antenna (not shown). Implementations of the receiver 1302 may be limited to the demodulator 1304.
The data receiving system or apparatus 1300 includes a decoder 1306. The receiver 1302 provides a received signal to the decoder 1306. The signal provided to the decoder 1306 by the receiver 1302 may include one or more encoded bitstreams. The decoder 1306 outputs a decoded signal, such as, for example, decoded video signals including video information. The decoder 1306 may be, for example, decoder 1100 described in FIG. 11.
The data receiving system or apparatus 1300 is also communicatively coupled to a storage unit 1307. In one implementation, the storage unit 1307 is coupled to the receiver 1302, and the receiver 1302 accesses a bitstream from the storage unit 1307. In another implementation, the storage unit 1307 is coupled to the decoder 1306, and the decoder 1306 accesses a bitstream from the storage unit 1307. The bitstream accessed from the storage unit 1307 includes, in different implementations, one or more encoded bitstreams. The storage unit 1307 is, in different implementations, one or more of a standard DVD, a Blu-Ray disc, a hard drive, or some other storage device.
The output data from the decoder 1306 is provided, in one implementation, to a processor 1308. In some implementations, the decoder 1306 includes the processor 1308 and therefore performs the operations of the processor 1308. In other implementations, the processor 1308 is part of a downstream device such as, for example, a set-top box or a television.
The implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed may also be implemented in other forms (for example, an apparatus or program). An apparatus and constituents included therein, for example, a processor, an encoder and a decoder, may be implemented in, for example, appropriate hardware, software, and firmware. The methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants (“PDAs”), and other devices that facilitate communication of information between end-users.
Reference to “one embodiment” or “an embodiment” or “one implementation” or “an implementation” of the present principles, as well as other variations thereof, mean that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment of the present principles. Thus, the appearances of the phrase “in one embodiment” or “in an embodiment” or “in one implementation” or “in an implementation”, as well any other variations, appearing in various places throughout the specification are not necessarily all referring to the same embodiment.
Additionally, this application or its claims may refer to “determining” various pieces of information. Determining the information may include one or more of, for example, estimating the information, calculating the information, predicting the information, or retrieving the information from memory.
Further, this application or its claims may refer to “accessing” various pieces of information. Accessing the information may include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
Additionally, this application or its claims may refer to “receiving” various pieces of information. Receiving is, as with “accessing”, intended to be a broad term. Receiving the information may include one or more of, for example, accessing the information, or retrieving the information (for example, from memory). Further, “receiving” is typically involved, in one way or another, during operations such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
As will be evident to one of skill in the art, implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted. The information may include, for example, instructions for performing a method, or data produced by one of the described implementations. For example, a signal may be formatted to carry the bitstream of a described embodiment. Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal. The formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream. The information that the signal carries may be, for example, analog or digital information. The signal may be transmitted over a variety of different wired or wireless links, as is known. The signal may be stored on a processor-readable medium.

Claims (21)

The invention claimed is:
1. A method for decoding High Dynamic Range images, comprising:
accessing a bitstream having the High Dynamic Range images included therein;
decoding Low Dynamic Range images from the bitstream;
accessing information from the bitstream, the accessed information including a decoding parameter used to decode the Low Dynamic Range images;
generating an inverse tone-mapping function responsive to said decoding parameter used to decode the Low Dynamic Range images; and
generating the High Dynamic Range images by applying said inverse tone mapping curve.
2. The method of claim 1, wherein the decoding parameter used to decode the Low Dynamic Range image is a quantization parameter.
3. The method of claim 1, wherein the accessed information includes parameters indicative of the dynamic range of original High Dynamic Range images corresponding to the High Dynamic Range images to be decoded.
4. The method of claim 1, wherein the accessed information further includes parameters indicative of a probability distribution function of original High Dynamic Range images corresponding to the High Dynamic Range images to be decoded.
5. A method for encoding High Dynamic Range images, comprising:
determining a tone mapping function responsive to the High Dynamic Range images and an encoding parameter;
determining Low Dynamic Range images from the High Dynamic Range images responsive to the tone mapping function; and
encoding the determined Low Dynamic Range images by using said encoding parameter and information including said encoding parameter, said encoding parameter being intended to be used for generating an inverse tone-mapping function.
6. The method of claim 5, wherein the encoding parameter used to encode the Low Dynamic Range images is a quantization parameter.
7. The method of claim 6, wherein the determining a tone mapping function comprises determining a Lagrangian multiplier responsive to the quantization parameter.
8. The method of claim 5, wherein the determining a tone mapping function comprises determining at least one of the dynamic range of the High Dynamic Range images and a probability distribution function of the High Dynamic Range images.
9. The method of claim 5, wherein the determining a tone mapping function is responsive to a rate distortion function.
10. An apparatus for decoding High Dynamic Range images, comprising a processor configured to:
decode Low Dynamic Range images from a bitstream having the High Dynamic Range images included therein;
access information from the bitstream, the accessed information including a decoding parameter used to decode the Low Dynamic Range images; and
generate an inverse tone-mapping function responsive to said decoding parameter used to decode the Low Dynamic Range images, and to generate the High Dynamic Range images by applying said inverse tone mapping curve.
11. The apparatus of claim 10, wherein the decoding parameter used to decode the Low Dynamic Range image is a quantization parameter.
12. The apparatus of claim 10, wherein the accessed information includes parameters indicative of the dynamic range of original High Dynamic Range images corresponding to the High Dynamic Range images to be decoded.
13. The apparatus of claim 10, wherein the accessed information further includes parameters indicative of a probability distribution function of original High Dynamic Range images corresponding to the HDR images to be decoded.
14. An apparatus for encoding High Dynamic Range images, comprising a processor configured to:
determine a tone mapping function responsive to the High Dynamic Range images and an encoding parameter;
determine Low Dynamic Range images from the High Dynamic Range images responsive to the tone mapping function; and
encode the determined Low Dynamic Range images by using said encoding parameter and information including said encoding parameter, said encoding parameter being intended to be used for generating an inverse tone-mapping function.
15. The apparatus of claim 14, wherein the encoding parameter used to encode the Low Dynamic Range images is a quantization parameter.
16. The apparatus of claim 15, wherein the tone curve estimator determines a Lagrangian multiplier responsive to the quantization parameter.
17. The apparatus of claim 14, wherein the tone curve estimator determines at least one of the dynamic range of the High Dynamic Range images and a probability distribution function of the High Dynamic Range images.
18. The apparatus of claim 14, wherein the tone curve estimator determines the tone mapping function responsive to a rate distortion function.
19. A non-transitory computer readable storage medium having stored thereon instructions for encoding High Dynamic Range images, according to claim 5.
20. A non-transitory computer readable storage medium having stored thereon a bitstream generated according to claim 5.
21. A non-transitory computer readable storage medium having stored thereon instructions for decoding High Dynamic Range images, according to claim 1.
US15/122,149 2014-02-26 2015-02-23 Method and apparatus for encoding and decoding HDR images Active US10650501B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/784,124 US11727548B2 (en) 2014-02-26 2020-02-06 Method and apparatus for encoding and decoding HDR images

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
EP14305266 2014-02-26
EP14305266.0 2014-02-26
EP14305266 2014-02-26
EP14305480.7 2014-04-01
EP14305480.7A EP2927865A1 (en) 2014-04-01 2014-04-01 Method and apparatus for encoding and decoding HDR images
EP14305480 2014-04-01
PCT/EP2015/053747 WO2015128295A1 (en) 2014-02-26 2015-02-23 Method and apparatus for encoding and decoding hdr images

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2015/053747 A-371-Of-International WO2015128295A1 (en) 2014-02-26 2015-02-23 Method and apparatus for encoding and decoding hdr images

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/784,124 Continuation US11727548B2 (en) 2014-02-26 2020-02-06 Method and apparatus for encoding and decoding HDR images

Publications (2)

Publication Number Publication Date
US20160371822A1 US20160371822A1 (en) 2016-12-22
US10650501B2 true US10650501B2 (en) 2020-05-12

Family

ID=52595323

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/122,149 Active US10650501B2 (en) 2014-02-26 2015-02-23 Method and apparatus for encoding and decoding HDR images
US16/784,124 Active 2036-11-03 US11727548B2 (en) 2014-02-26 2020-02-06 Method and apparatus for encoding and decoding HDR images

Family Applications After (1)

Application Number Title Priority Date Filing Date
US16/784,124 Active 2036-11-03 US11727548B2 (en) 2014-02-26 2020-02-06 Method and apparatus for encoding and decoding HDR images

Country Status (7)

Country Link
US (2) US10650501B2 (en)
EP (1) EP3111416A1 (en)
JP (1) JP6546599B2 (en)
KR (3) KR20220025153A (en)
CN (2) CN110708545B (en)
TW (1) TWI677235B (en)
WO (1) WO2015128295A1 (en)

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3111416A1 (en) * 2014-02-26 2017-01-04 Thomson Licensing Method and apparatus for encoding and decoding hdr images
JP6358718B2 (en) 2014-05-28 2018-07-18 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Method and apparatus for encoding HDR images, and method and apparatus for use of such encoded images
JP6635312B2 (en) * 2014-07-03 2020-01-22 ソニー株式会社 Image processing apparatus and method
EP3026917A1 (en) * 2014-11-27 2016-06-01 Thomson Licensing Methods and apparatus for model-based visual descriptors compression
JP2018525905A (en) * 2015-07-22 2018-09-06 アリス エンタープライジズ エルエルシーArris Enterprises Llc System for encoding high dynamic range and wide color gamut sequences
KR101909867B1 (en) 2015-12-09 2018-10-18 돌비 레버러토리즈 라이쎈싱 코오포레이션 Low Complexity Lookup Table Configuration with Reduced Interpolation Error
CN108370442B (en) * 2015-12-15 2020-02-14 华为技术有限公司 Method and device for processing high dynamic range image and computer readable storage medium
JP6237797B2 (en) * 2016-01-05 2017-11-29 ソニー株式会社 Video system, video processing method, program, and video converter
GB2549696A (en) * 2016-04-13 2017-11-01 Sony Corp Image processing method and apparatus, integrated circuitry and recording medium
EP3507981B1 (en) * 2016-08-30 2023-11-29 Dolby Laboratories Licensing Corporation Real-time reshaping of single-layer backwards-compatible codec
KR102594201B1 (en) * 2016-09-22 2023-10-27 삼성디스플레이 주식회사 Method of processing image and display apparatus performing the same
JP6822121B2 (en) * 2016-12-19 2021-01-27 ソニー株式会社 Image processing equipment, image processing methods and programs
KR102416867B1 (en) * 2017-03-07 2022-07-05 중앙대학교 산학협력단 Method and its Apparatus for offering solution of vision problem
EP3639238B1 (en) * 2017-06-16 2022-06-15 Dolby Laboratories Licensing Corporation Efficient end-to-end single layer inverse display management coding
KR102412007B1 (en) * 2017-07-20 2022-06-22 삼성전자주식회사 Electronic device and method for compressing high dynamic range image data in the electronic device
US10972767B2 (en) * 2017-11-01 2021-04-06 Realtek Semiconductor Corp. Device and method of handling multiple formats of a video sequence
CN108171762B (en) * 2017-12-27 2021-10-12 河海大学常州校区 Deep learning compressed sensing same-class image rapid reconstruction system and method
WO2019192484A1 (en) 2018-04-02 2019-10-10 Huawei Technologies Co., Ltd. Video coding with successive codecs
JP7189230B2 (en) 2018-04-09 2022-12-13 ドルビー ラボラトリーズ ライセンシング コーポレイション HDR image representation using neural network mapping
CN108900838B (en) * 2018-06-08 2021-10-15 宁波大学 Rate distortion optimization method based on HDR-VDP-2 distortion criterion
TWI843747B (en) * 2018-09-05 2024-06-01 荷蘭商皇家飛利浦有限公司 Multi-range hdr video coding
US10841356B2 (en) * 2018-11-28 2020-11-17 Netflix, Inc. Techniques for encoding a media title while constraining bitrate variations
US10880354B2 (en) 2018-11-28 2020-12-29 Netflix, Inc. Techniques for encoding a media title while constraining quality variations
JP7344654B2 (en) * 2019-03-07 2023-09-14 キヤノン株式会社 Imaging device and playback device and their control method and program
CN110769210B (en) * 2019-11-07 2022-06-03 湖南国科微电子股份有限公司 Method and device for generating HDR image
US11398017B2 (en) * 2020-10-09 2022-07-26 Samsung Electronics Co., Ltd. HDR tone mapping based on creative intent metadata and ambient light
US20230114798A1 (en) * 2021-09-28 2023-04-13 Samsung Electronics Co., Ltd. Joint auto-exposure tone-mapping system
CN117440154B (en) * 2023-12-21 2024-04-19 之江实验室 Depth map sequence compression method considering floating point digital splitting

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030103459A1 (en) 2001-11-16 2003-06-05 Connors Dennis P. Method and implementation for a flow specific modified selective-repeat ARQ communication system
US7020822B2 (en) 2001-08-02 2006-03-28 Texas Instruments Incorporated Automatic repeat request for centralized channel access
US20070183451A1 (en) 2004-01-22 2007-08-09 Joachim Lohr Method of harq retransmission timing control
US20080144550A1 (en) 2006-12-15 2008-06-19 Motorola, Inc. Retransmission scheme for maintaining performance for wireless communications in the presence of periodic intermittent interference
US20080209297A1 (en) 2005-10-21 2008-08-28 Interdigital Technology Corporation Method and apparatus for retransmission management for reliable hybrid arq process
US7436854B2 (en) 2004-01-23 2008-10-14 Ntt Docomo, Inc. Transmitter device for controlling data transmission
US7715314B2 (en) 2006-03-30 2010-05-11 Ntt Docomo, Inc. Communication terminal and retransmission control method
WO2010105036A1 (en) 2009-03-13 2010-09-16 Dolby Laboratories Licensing Corporation Layered compression of high dynamic range, visual dynamic range, and wide color gamut video
WO2011002505A1 (en) 2009-06-29 2011-01-06 Thomson Licensing Zone-based tone mapping
US7983496B2 (en) * 2007-06-26 2011-07-19 Mitsubishi Electric Research Laboratories, Inc. Inverse tone mapping for bit-depth scalable image coding adapted to variable block sizes
CN102341825A (en) 2009-03-03 2012-02-01 微软公司 Multi-modal tone-mapping of images
US20120201456A1 (en) * 2009-10-08 2012-08-09 International Business Machines Corporation Transforming a digital image from a low dynamic range (ldr) image to a high dynamic range (hdr) image
CN102970536A (en) 2012-11-15 2013-03-13 上海交通大学 Improved video coding method with predicted residual regulation
CN103891294A (en) 2011-04-28 2014-06-25 皇家飞利浦有限公司 Apparatuses and methods for hdr image encoding and decodng

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100220789A1 (en) * 2007-10-19 2010-09-02 Wu Yuwen Combined spatial and bit-depth scalability
JP2011010108A (en) 2009-06-26 2011-01-13 Seiko Epson Corp Imaging control apparatus, imaging apparatus, and imaging control method
TWI586150B (en) 2012-06-29 2017-06-01 新力股份有限公司 Image processing device and non-transitory computer readable storage medium
EP3111416A1 (en) * 2014-02-26 2017-01-04 Thomson Licensing Method and apparatus for encoding and decoding hdr images

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7020822B2 (en) 2001-08-02 2006-03-28 Texas Instruments Incorporated Automatic repeat request for centralized channel access
US20030103459A1 (en) 2001-11-16 2003-06-05 Connors Dennis P. Method and implementation for a flow specific modified selective-repeat ARQ communication system
US20070183451A1 (en) 2004-01-22 2007-08-09 Joachim Lohr Method of harq retransmission timing control
US7436854B2 (en) 2004-01-23 2008-10-14 Ntt Docomo, Inc. Transmitter device for controlling data transmission
US20080209297A1 (en) 2005-10-21 2008-08-28 Interdigital Technology Corporation Method and apparatus for retransmission management for reliable hybrid arq process
US7715314B2 (en) 2006-03-30 2010-05-11 Ntt Docomo, Inc. Communication terminal and retransmission control method
US20080144550A1 (en) 2006-12-15 2008-06-19 Motorola, Inc. Retransmission scheme for maintaining performance for wireless communications in the presence of periodic intermittent interference
US7983496B2 (en) * 2007-06-26 2011-07-19 Mitsubishi Electric Research Laboratories, Inc. Inverse tone mapping for bit-depth scalable image coding adapted to variable block sizes
CN102341825A (en) 2009-03-03 2012-02-01 微软公司 Multi-modal tone-mapping of images
CN102388612A (en) 2009-03-13 2012-03-21 杜比实验室特许公司 Layered compression of high dynamic range, visual dynamic range, and wide color gamut video
WO2010105036A1 (en) 2009-03-13 2010-09-16 Dolby Laboratories Licensing Corporation Layered compression of high dynamic range, visual dynamic range, and wide color gamut video
WO2011002505A1 (en) 2009-06-29 2011-01-06 Thomson Licensing Zone-based tone mapping
CN102473295A (en) 2009-06-29 2012-05-23 汤姆森特许公司 Zone-based tone mapping
US20120201456A1 (en) * 2009-10-08 2012-08-09 International Business Machines Corporation Transforming a digital image from a low dynamic range (ldr) image to a high dynamic range (hdr) image
CN102696220A (en) 2009-10-08 2012-09-26 国际商业机器公司 Method and system for transforming a digital image from a low dynamic range (LDR) image to a high dynamic range (HDR) image
CN103891294A (en) 2011-04-28 2014-06-25 皇家飞利浦有限公司 Apparatuses and methods for hdr image encoding and decodng
US20140247870A1 (en) 2011-04-28 2014-09-04 Koninklijke Philips N.V. Apparatuses and methods for hdr image encoding and decodng
CN102970536A (en) 2012-11-15 2013-03-13 上海交通大学 Improved video coding method with predicted residual regulation

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Bross et al., "High Efficiency Video Coding (HEVC) text specification draft 9", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document JCTCV-K1003, Version 13, 11th Meeting: Shanghai, China, Oct. 10, 2012, pp. 1-317.
Ferradans et al., "A Multi-Modal Approach to Perceptual Tone Mapping", 2009 6th European Conference for Visual Media Production, London, United Kingdom, Nov. 12, 2009, pp. 1-10.
Mai et al., "On-the-Fly Tone Mapping for Backward-Compatible High Dynamic Range Image/Video Compression", 2010 IEEE International Symposium on Circuits and Systems, Paris, France, May 30, 2010, pp. 1831-1834.
Mai et al., "Optimizing a Tone Curve for Backward-Compatible High Dynamic Range Image and Video Compression", IEEE Transactions on Image Processing, vol. 20, No. 6, Jun. 2011, pp. 1558-1571. *
Takeuchi et al., "A Coding Method for High Dynamic Range Video Sequence Using Multiple Look-Up-Tables", Workshop on Picture Coding and Image Processing, Nagoya, Japan, Dec. 7, 2010, pp. 33-34.
Wang et al., "Memory Efficient Hierarchical Lookup Tables for Mass Arbitrary-Side Growing Huffman Trees Decoding", IEEE Transactions on Circuits and Systems for Video Technology, vol. 18, No. 10, Oct. 2008, pp. 1335-1346.

Also Published As

Publication number Publication date
US11727548B2 (en) 2023-08-15
CN106062816A (en) 2016-10-26
CN110708545A (en) 2020-01-17
US20160371822A1 (en) 2016-12-22
JP6546599B2 (en) 2019-07-17
KR20160125399A (en) 2016-10-31
JP2017509057A (en) 2017-03-30
KR102361927B1 (en) 2022-02-11
CN106062816B (en) 2019-11-22
WO2015128295A1 (en) 2015-09-03
US20210062095A9 (en) 2021-03-04
EP3111416A1 (en) 2017-01-04
KR20230062888A (en) 2023-05-09
TWI677235B (en) 2019-11-11
KR20220025153A (en) 2022-03-03
US20200172813A1 (en) 2020-06-04
TW201543867A (en) 2015-11-16
CN110708545B (en) 2022-04-01

Similar Documents

Publication Publication Date Title
US11727548B2 (en) Method and apparatus for encoding and decoding HDR images
US9781449B2 (en) Rate distortion optimization in image and video encoding
US9930329B2 (en) Video encoding and decoding based on image refinement
US8531321B1 (en) Signal processing and inheritance in a tiered signal quality hierarchy
US10009611B2 (en) Visual quality measure for real-time video processing
US10063874B2 (en) Picture coding device, picture coding method, and picture coding program, and picture decoding device, picture decoding method, and picture decoding program
CN101606320A (en) The distortion of quantized data is estimated
EP2927865A1 (en) Method and apparatus for encoding and decoding HDR images
US20160360237A1 (en) Method and apparatus for encoding, decoding a video signal using additional control of quantizaton error
CN117480778A (en) Residual coding and video coding methods, devices, equipment and systems
US8428380B2 (en) Apparatus and method of embedded quantization for the improved SNR scalbility
CN107666472B (en) Method and apparatus for hybrid digital-analog coding
US10771820B2 (en) Image encoding method and apparatus using artifact reduction filter, and image decoding method and apparatus using artifact reduction filter
US11303916B2 (en) Motion compensation techniques for video
US10536703B2 (en) Method and apparatus for video quality assessment based on content complexity
CN117751575A (en) Method or apparatus for estimating film grain parameters
JP6019215B2 (en) Vertex correction method and apparatus for rotated three-dimensional (3D) components
Le Pendu et al. Template based inter-layer prediction for high dynamic range scalable compression
Andre et al. Entropy-based distortion measure for image coding
Kaur et al. An Analysis of Relationship Between Image Characteristics and Compression Quality for High-Resolution Satellite
KR20110041962A (en) Method and apparatus for video quality assessment

Legal Events

Date Code Title Description
AS Assignment

Owner name: THOMSON LICENSING, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LE PENDU, MIKAEL;BOITARD, RONAN;GUILLEMOT, CHRISTINE;AND OTHERS;SIGNING DATES FROM 20180118 TO 20180124;REEL/FRAME:045082/0972

AS Assignment

Owner name: INTERDIGITAL VC HOLDINGS, INC., DELAWARE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:THOMSON LICENSING;REEL/FRAME:047289/0698

Effective date: 20180730

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP, ISSUE FEE PAYMENT VERIFIED

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4