WO2008019524A1 - Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data - Google Patents

Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data Download PDF

Info

Publication number
WO2008019524A1
WO2008019524A1 PCT/CN2006/001699 CN2006001699W WO2008019524A1 WO 2008019524 A1 WO2008019524 A1 WO 2008019524A1 CN 2006001699 W CN2006001699 W CN 2006001699W WO 2008019524 A1 WO2008019524 A1 WO 2008019524A1
Authority
WO
WIPO (PCT)
Prior art keywords
video image
transfer function
color
values
histogram
Prior art date
Application number
PCT/CN2006/001699
Other languages
French (fr)
Inventor
Yongying Gao
Yuwen Wu
Ying Chen
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing filed Critical Thomson Licensing
Priority to DE200660019006 priority Critical patent/DE602006019006D1/en
Priority to US12/309,335 priority patent/US8761249B2/en
Priority to EP20060761443 priority patent/EP2041983B1/en
Priority to PCT/CN2006/001699 priority patent/WO2008019524A1/en
Priority to JP2009519775A priority patent/JP4991851B2/en
Priority to CN2006800553696A priority patent/CN101507284B/en
Priority to KR20097001015A priority patent/KR101196975B1/en
Publication of WO2008019524A1 publication Critical patent/WO2008019524A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N11/00Colour television systems
    • H04N11/04Colour television systems using pulse code modulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • G06T5/92Dynamic range modification of images or parts thereof based on global image properties
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/46Colour picture communication systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/14Coding unit complexity, e.g. amount of activity or edge presence estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/36Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image

Definitions

  • This invention relates to a method and an apparatus for encoding video color enhancement data. Further, the invention relates to a method and an apparatus for decoding video color enhancement data.
  • Color bit-depth scalability is potentially useful considering the fact that in a considerably long period in the future standard 8-bit and higher-bit digital imaging systems will simultaneously exist in consumer marketplaces . Different color bit-depths are of particular importance for example for terminal display devices during multimedia content deliveries .
  • the present invention provides methods and device for enabling video color space scalability.
  • a method and a device for encoding a color enhancement layer is provided, which is encoded differentially.
  • Another aspect of the invention is a method and a device for decoding a video signal to obtain either a conventional color bit-depth image or an enhanced color bit-depth image.
  • the encoding aspect of the invention comprises the following steps: generating a transfer function, for example in the form of a look-up table (LUT) , which is suitable for mapping input color values to output color values, both consisting of 2 M different colors, applying the transfer function to a first video picture with low or conventional color bit-depth, generating a difference picture or residual between the transferred video picture and a second video picture with higher color bit-depth (N bit, with N>M; but same spatial resolution as the first video picture) and encoding the residual. Then, the encoded first video picture, parameters of the transfer function (e.g. the LUT itself) and the encoded residual are transmitted to a receiver. The parameters of the transfer function may also be encoded. Further, the parameters of the transfer function are indicated as such.
  • the first and second images can be regarded as being a color base layer and a color enhancement layer, respectively.
  • the transfer function may be obtained by comparing color histograms of the first and the second video pictures, for which purpose the color histogram of the first picture, which has 2 M bins, is transformed into a "smoothed" color histogram with 2 N bins (N>M) , and determining a transfer function from the smoothed histogram and the color enhancement layer histogram which defines a transfer between the values of the smoothed color histogram and the values of the color enhancement layer histogram.
  • the described procedure is done separately for the basic display colors red, green and blue.
  • a method for decoding comprises extracting from a bit stream video data for a first and a second video image, and extracting color enhancement control data, furthermore decoding and reconstructing the first video image, wherein a reconstructed first video image is obtained having color pixel values with M bit each, and constructing from the color enhancement control data a mapping table that implements a transfer function. Then the mapping table is applied to each of the pixels of the reconstructed first video image, and the resulting transferred video image serves as prediction image which is then updated with the decoded second video image.
  • the decoded second video image is a> residual image
  • the updating results in an enhanced video image which has pixel values with N bit each (N>M) , and therefore a higher color space than the reconstructed first video image.
  • the above steps are performed separately for each of the basic video colors red, green and blue.
  • a complete video signal may comprise for each picture an encoded low color-resolution image, and for each of these colors an encoded residual image and parameters of a transfer function, both for generating a higher color-resolution image.
  • generating the transfer function and the residual image is performed on the R-G-B values of the raw video image, and is therefore independent from the further video encoding.
  • the low color-resolution image can then be encoded using any conventional encoding, for example according to an MPEG or JVT standard (AVC, SVC etc.) .
  • AVC MPEG or JVT standard
  • the color enhancement is performed on top of the conventional decoding, and therefore independent from its encoding format.
  • devices with lower color bit-depth display capability need only decode the color base layer having lower color bit-depth
  • advanced devices with enhanced color bit-depth display capability e.g. 12-bit displays
  • the invention is related to a new type of video scalability, namely color bit-depth scalability, and provides a color bit-depth prediction solution that can be applied in the inter-layer prediction of a color bit-depth scalable CODEC to improve the coding efficiency.
  • the advantage of using a transfer function to generate a prediction image before generating a residual image is that the encoding is more efficient, because the prediction image matches the respective color enhancement layer image better.
  • a value for a particular green tone which is described by an 8-bit value of 9Ahe ⁇ i- n the color base layer, may map to sixteen different 12-bit color values in the color enhancement layer, from 9A0 he ⁇ to 9AF hex - While in one picture one of these sixteen values may dominate in the color enhancement layer, it may be another value in another picture.
  • the invention enables optimized encoding of the color enhancement layer.
  • Fig.l the structure of color bit-depth scalable video coding and decoding
  • Fig.2 a flow diagram of the encoding process
  • Fig.3 a flow diagram of smoothed histogram specification
  • Fig.4 an example of histogram smoothing
  • Fig.5 a flow diagram of the decoding process
  • Fig.6 an overview of a complete enhancement layer encoder. Detailed description of the invention
  • Fig.l shows the general structure of color bit-depth scalable video coding and decoding.
  • the encoder ENC receives each video image twice, in conventional color resolution of M bit and in enhanced color resolution of N bit (N>M) . Instead of generating two independent separate video streams, a color base layer and a color enhancement layer are generated, where the color enhancement layer is dependent on the color base layer. As shown in Fig.l, an inter-layer prediction is performed in both the encoder and the decoder for predicting colors of higher bit-depth.
  • the base layer encoder block in Fig.l contains all the conventional encoding steps, for example any spatial and/or temporal redundancy reduction such as generating separate luminance and chrominance values, DCT transformations etc.
  • the base layer decoder performs the respective decoding, "base layer” and “enhancement layer” refer only to color bit-depth.
  • any video encoding and decoding method can be used for the base layer, independent from the color enhancement process enabled by the invention.
  • the output of a base layer decoder is suitable for a conventional display device, for example with M bit color resolution, while the output of an enhancement layer decoder is suitable for a high-quality display that is capable of displaying 2 N (N>M) different color values for each of the red, green and blue partial image.
  • N are 10, 12 or 16 bits if M is e.g. 8, but in principle all combinations of M and N (N > M) can be implemented.
  • the inter-layer prediction impacts significantly the coding efficiency.
  • the present invention proposes a spatially uniform approach for color bit depth prediction based on smoothed histogram specification.
  • the corresponding pixels here the "corresponding pixels” mean two pixels that belong to the two images respectively but have the same coordinates in the image coordinate system
  • the only difference between the corresponding pixels is the color bit depth.
  • each color of one image is encoded with code words of M bit length while the other image it is encoded with code words of N bit length, with M ⁇ N.
  • the task of inverse tone mapping is to generate a predicted version of the N-bit image from the M-bit image, following the criterion that the difference between the predicted N-bit image and the original N-bit image is minimized.
  • the difference between the predicted N-bit image and the original N-bit image may be measured by any method, for example PSNR, which is widely accepted and used in the field of video compression. Further, in this case the most important aspect is how effective the predicted image works for the following residual data compression, rather than how the predicted image looks.
  • PSNR which is widely accepted and used in the field of video compression.
  • Fig.2 shows the general encoding procedure of one embodiment of the invention.
  • a base layer image IM BL is transformed TR using a look-up table LUT fina i that implements a mapping function.
  • the result of the transformation TR is a transformed base layer image IM TR , BL -
  • the look-up table LUT f i na i is generated from another, intermediate look-up table LUTinter, which is generated from color histograms of the base layer picture, preferably the reconstructed base layer picture, and of the enhancement layer picture.
  • each bin of the base layer histogram corresponds to 2 N ⁇ M bins of the enhancement layer histogram, i.e. each base layer pixel value would conventionally serve to predict one out of 2 N ⁇ M pixel values. This prediction is optimized by considering the color histograms of the two pictures.
  • the intermediate look-up table LUTi nte r implements a histogram transformation function describing a transformation between the color histogram Hist EL of the enhancement layer picture and a color histogram SmHist BL generated from the base layer picture IM B L r as described below.
  • the transformed base layer picture IM TR , BL is used to predict the enhancement layer picture IM EL , and the difference or residual ⁇ is calculated, encoded IM res and transmitted. All these processes are separate for R, G and B.
  • Fig.3 shows details of the proposed spatially uniform inverse tone mapping.
  • the base layer (M-bit) image is named "input image” NI
  • the enhancement layer (N-bit) image is renamed "desired image” DI.
  • One basic idea is to apply the algorithm of histogram specification 1 to NI and DI.
  • a lookup table is generated to convert NI such that it has a particular histogram, as specified by the histogram of DI.
  • Fig.3 for the low-bit (M-bit) image NI, the histogram of one color channel is p x (x) .
  • the histogram is smoothed by using eq. (5), and a smoothed histogram p x ' (x) results, which has the same bin size as the histogram of the high- bit (N-bit) image DI does.
  • the high-bit (N-bit) image DI has the histogram p z (z) of the same color channel, which is accumulated to result in v k (same k as above) . It can be expressed as a function G(z k ) . Then the distribution v-, is determined, which gives for each value of s k the best-matching value vi (from the v k values of the high-bit image DI) . This distribution v-, sets up a transformation from the values s k (therefore x k ' ) to the values v k , and it is the inverse of the transform G.
  • histogram smoothing is described.
  • the goal of histogram smoothing is to "stretch" the input histogram so that it will have the same bin size as the desired histogram. This is a prerequisite for the histogram specification process.
  • the value of p x (x k ) is uniformly distributed into a corresponding range of [x k ,x k ' +1, ⁇ ••,x k + 2 ( - N"M) -1] , as described in eq. (5) :
  • the number of values in a single bin of the input picture is the number of pixels within the input picture that have this particular color value.
  • a value of 40 in the left-most bin corresponding to the value 0 means that 40 pixels of the input picture have a "red" value of 0 (in this histogram refers to the color "red") .
  • the smoothing process distributes these values to 2 N ⁇ M corresponding bins of the smoothed histogram.
  • the first four bins of the smoothed histogram correspond to the first bin of the original histogram.
  • the 40 values of the original histogram are evenly distributed to the four corresponding bins of the smoothed histogram, as shown in Fig.4 b) .
  • the input histogram can be specified as exactly the same as the desired histogram.
  • the smoothness procedure is equivalent to increasing the sampling rate during the discretization of the input histogram, and it outperforms the direct histogram specification between two histograms that have different bin size.
  • the smoothed histogram is just an intermediate step in this algorithm. There is not an image corresponding to it.
  • Eq. (6) says that among the 2 N ⁇ M values that x k corresponds to, we select the one that has the highest value in the desired histogram p z (y_/) •
  • Eq. (7) says that among the 2 N ⁇ M values that x k corresponds to, we use the nearest integer less than or equal to their mean as the finally mapped value .
  • the LUT m ter is a "one-to-one" mapping, because it maps the smoothed histogram of the input image to the histogram of the desired image.
  • each x k corresponds to 2 N ⁇ M values.
  • the task of "Probability Mapping" is to choose only one value from the 2 N ⁇ M values as the finally mapped value of x k .
  • the LUT fina i is still a "one-to-one" mapping: it maps each value x k to a value y k .
  • the mapping of the LUT is invertible because only 2 M values of the total values (2 N ) of the desired image have the corresponding x k .
  • the histograms of both images are calculated. Then the input histogram is smoothed, resulting in the "smoothed histogram". After the remaining steps (classical histogram specification and probability mapping) are finished, a final LUT is generated to map the levels of the input histogram to the levels of that of the desired histogram. Then the predicted image is generated by applying the LUT to each pixel of the input image .
  • Fig.6 shows an overview of a complete enhancement layer encoder, with separate branches for R, G and B images.
  • the base layer is conventionally encoded.
  • the invention is usable for video encoders and video decoders, and particularly for encoding / decoding a color base layer and a color enhancement layer of the same spatial resolution with optimized efficiency.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Processing (AREA)

Abstract

A highly accurate reproduction of visual intensity and contrast rather than the conventional 8-bit color depth is more and more used, motivating the development of an enhanced dynamic range called high bit-depth. A method for encoding a first, low bit-depth image (IMBL) of M bit RGB pixels and a second, high bit-depth video image (IMEL) of N bit RGB pixels, M<N, both having same spatial resolution, comprises generating a first transfer function (LUTinter) based on color histograms (HistBL, HistEL) of the first and the second video image, generating a second transfer function (LUTfinal) based on the first transfer function (LUTinter), applying (TR) the second transfer function (LUTfinal) to the first video image, wherein a transformed first video image (IMTR, BL) is generated, calculating and encoding the residual (IMres) f encoding the first video image (IMBL) and transmitting the encoded first video image (IMBL), the second transfer function (LUTfinal) and the encoded difference.

Description

Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data
Field of the invention
This invention relates to a method and an apparatus for encoding video color enhancement data. Further, the invention relates to a method and an apparatus for decoding video color enhancement data.
Background
In recent years, highly accurate reproduction of visual intensity and contrast rather than the conventional 8-bit color depth is more and more used in many fields, such as medical imaging, high-quality video-enabled computer games and professional studio and home theatre related applications. This process motivates the development of an enhanced dynamic range, which is called high bit-depth herein, for the convenience of comparison with the conventional 8-bit color depth. On contrast to the fact that advances in electronic sensors, processors and storage devices have resulted in very high pixel resolutions for both capturing and display devices, the color capacities of digital imaging systems have evolved in a very slow pace. 8-bit color depth has been dominant for capturing and display devices since the first generation of digitalized visual contents emerged.
Color bit-depth scalability is potentially useful considering the fact that in a considerably long period in the future standard 8-bit and higher-bit digital imaging systems will simultaneously exist in consumer marketplaces . Different color bit-depths are of particular importance for example for terminal display devices during multimedia content deliveries .
Summary of the Invention
The present invention provides methods and device for enabling video color space scalability. According to one aspect of the invention, a method and a device for encoding a color enhancement layer is provided, which is encoded differentially. Another aspect of the invention is a method and a device for decoding a video signal to obtain either a conventional color bit-depth image or an enhanced color bit-depth image.
In principle, the encoding aspect of the invention comprises the following steps: generating a transfer function, for example in the form of a look-up table (LUT) , which is suitable for mapping input color values to output color values, both consisting of 2M different colors, applying the transfer function to a first video picture with low or conventional color bit-depth, generating a difference picture or residual between the transferred video picture and a second video picture with higher color bit-depth (N bit, with N>M; but same spatial resolution as the first video picture) and encoding the residual. Then, the encoded first video picture, parameters of the transfer function (e.g. the LUT itself) and the encoded residual are transmitted to a receiver. The parameters of the transfer function may also be encoded. Further, the parameters of the transfer function are indicated as such. The first and second images can be regarded as being a color base layer and a color enhancement layer, respectively.
In particular, the transfer function may be obtained by comparing color histograms of the first and the second video pictures, for which purpose the color histogram of the first picture, which has 2M bins, is transformed into a "smoothed" color histogram with 2N bins (N>M) , and determining a transfer function from the smoothed histogram and the color enhancement layer histogram which defines a transfer between the values of the smoothed color histogram and the values of the color enhancement layer histogram. The described procedure is done separately for the basic display colors red, green and blue.
According to the decoding aspect of the invention, a method for decoding comprises extracting from a bit stream video data for a first and a second video image, and extracting color enhancement control data, furthermore decoding and reconstructing the first video image, wherein a reconstructed first video image is obtained having color pixel values with M bit each, and constructing from the color enhancement control data a mapping table that implements a transfer function. Then the mapping table is applied to each of the pixels of the reconstructed first video image, and the resulting transferred video image serves as prediction image which is then updated with the decoded second video image. The decoded second video image is a> residual image, and the updating results in an enhanced video image which has pixel values with N bit each (N>M) , and therefore a higher color space than the reconstructed first video image. The above steps are performed separately for each of the basic video colors red, green and blue. Thus, a complete video signal may comprise for each picture an encoded low color-resolution image, and for each of these colors an encoded residual image and parameters of a transfer function, both for generating a higher color-resolution image. Advantageously, generating the transfer function and the residual image is performed on the R-G-B values of the raw video image, and is therefore independent from the further video encoding. Thus, the low color-resolution image can then be encoded using any conventional encoding, for example according to an MPEG or JVT standard (AVC, SVC etc.) . Also on the decoding side the color enhancement is performed on top of the conventional decoding, and therefore independent from its encoding format.
Thus, devices with lower color bit-depth display capability (e.g. 8-bit displays) need only decode the color base layer having lower color bit-depth, while advanced devices with enhanced color bit-depth display capability (e.g. 12-bit displays) may decode also the color enhancement layer and the transfer tables for red, green and blue, and generate pictures with full color space utilization.
The invention is related to a new type of video scalability, namely color bit-depth scalability, and provides a color bit-depth prediction solution that can be applied in the inter-layer prediction of a color bit-depth scalable CODEC to improve the coding efficiency. The advantage of using a transfer function to generate a prediction image before generating a residual image is that the encoding is more efficient, because the prediction image matches the respective color enhancement layer image better. For example, a value for a particular green tone, which is described by an 8-bit value of 9Aheχ i-n the color base layer, may map to sixteen different 12-bit color values in the color enhancement layer, from 9A0heχ to 9AFhex- While in one picture one of these sixteen values may dominate in the color enhancement layer, it may be another value in another picture. Thus, the invention enables optimized encoding of the color enhancement layer.
Advantageous embodiments of the invention are disclosed in the dependent claims, the following description and the figures .
Brief description of the drawings
Exemplary embodiments of the invention are described with reference to the accompanying drawings, which show in
Fig.l the structure of color bit-depth scalable video coding and decoding;
Fig.2 a flow diagram of the encoding process;
Fig.3 a flow diagram of smoothed histogram specification;
Fig.4 an example of histogram smoothing;
Fig.5 a flow diagram of the decoding process; and
Fig.6 an overview of a complete enhancement layer encoder. Detailed description of the invention
Fig.l shows the general structure of color bit-depth scalable video coding and decoding. The encoder ENC receives each video image twice, in conventional color resolution of M bit and in enhanced color resolution of N bit (N>M) . Instead of generating two independent separate video streams, a color base layer and a color enhancement layer are generated, where the color enhancement layer is dependent on the color base layer. As shown in Fig.l, an inter-layer prediction is performed in both the encoder and the decoder for predicting colors of higher bit-depth.
The base layer encoder block in Fig.l contains all the conventional encoding steps, for example any spatial and/or temporal redundancy reduction such as generating separate luminance and chrominance values, DCT transformations etc. On the decoder side, the base layer decoder performs the respective decoding, "base layer" and "enhancement layer" refer only to color bit-depth. Advantageously, any video encoding and decoding method can be used for the base layer, independent from the color enhancement process enabled by the invention. The output of a base layer decoder is suitable for a conventional display device, for example with M bit color resolution, while the output of an enhancement layer decoder is suitable for a high-quality display that is capable of displaying 2N (N>M) different color values for each of the red, green and blue partial image. Examples for N are 10, 12 or 16 bits if M is e.g. 8, but in principle all combinations of M and N (N > M) can be implemented. Like other type of scalability, such as quality scalability, spatial scalability and temporal scalability, the inter-layer prediction impacts significantly the coding efficiency.
The present invention proposes a spatially uniform approach for color bit depth prediction based on smoothed histogram specification. Consider two images that describe the same scene. For the two images, the corresponding pixels (here the "corresponding pixels" mean two pixels that belong to the two images respectively but have the same coordinates in the image coordinate system) refer to the same scene location. The only difference between the corresponding pixels is the color bit depth. Assume each color of one image is encoded with code words of M bit length while the other image it is encoded with code words of N bit length, with M < N. The task of inverse tone mapping is to generate a predicted version of the N-bit image from the M-bit image, following the criterion that the difference between the predicted N-bit image and the original N-bit image is minimized. The difference between the predicted N-bit image and the original N-bit image may be measured by any method, for example PSNR, which is widely accepted and used in the field of video compression. Further, in this case the most important aspect is how effective the predicted image works for the following residual data compression, rather than how the predicted image looks.
Fig.2 shows the general encoding procedure of one embodiment of the invention. A base layer image IMBL is transformed TR using a look-up table LUTfinai that implements a mapping function. The result of the transformation TR is a transformed base layer image IMTR,BL- The look-up table LUTfinai is generated from another, intermediate look-up table LUTinter, which is generated from color histograms of the base layer picture, preferably the reconstructed base layer picture, and of the enhancement layer picture. Due to the different color resolution of these pictures, their color histograms have not only a different amount of bins (one for each possible color) , but the distribution of the values among the bins of the enhancement layer picture may be different for each base layer color, and for each picture of a sequence. If for example the base layer has M bit and the enhancement layer has N bit, each bin of the base layer histogram corresponds to 2N~M bins of the enhancement layer histogram, i.e. each base layer pixel value would conventionally serve to predict one out of 2N~M pixel values. This prediction is optimized by considering the color histograms of the two pictures. The intermediate look-up table LUTinter implements a histogram transformation function describing a transformation between the color histogram HistEL of the enhancement layer picture and a color histogram SmHistBL generated from the base layer picture IMBL r as described below.
The transformed base layer picture IMTR,BL is used to predict the enhancement layer picture IMEL , and the difference or residual Δ is calculated, encoded IMres and transmitted. All these processes are separate for R, G and B.
Fig.3 shows details of the proposed spatially uniform inverse tone mapping. For better understanding, the base layer (M-bit) image is named "input image" NI while the enhancement layer (N-bit) image is renamed "desired image" DI. One basic idea is to apply the algorithm of histogram specification1 to NI and DI. A lookup table is generated to convert NI such that it has a particular histogram, as specified by the histogram of DI.
However, there are two major drawbacks when employing the classical histogram specification for color bit depth prediction. First, because the histograms have discrete values, the converted NI does not have exactly the same histogram as DI. Instead, the histogram of the converted NI is an approximation to that of DI. Second, the different bin size of the two histograms (caused by the different bit depth) deteriorates the matching between them (bin size means the number of the levels of each color component e.g. the bin size of 8-bit images is 256) . This is a particular drawback in the considered color bit depth prediction. For instance, in the case that NI is simply bit-shift from DI, the PSNR of the converted NI which was obtained by histogram specification can often be lower than that obtained by simple inverse bit-shift.
To overcome these drawbacks, we propose to "smooth" the histogram of NI before it is specified. The smoothed histogram SmHistBL is of the same bin size as the histogram HistEL of DI, which serves as the desired histogram for the process of histogram specification. The classical histogram specification algorithm is applied on the smoothed histogram and the desired histogram. Finally, a post processing called "Probability mapping" is carried out to obtain the LUT. The flowchart of the smoothed histogram specification is shown in Fig.3. px(xk),k = 0,1,---,2^ -1
1 R. C. Gonzalez and R. E. Woods, "Digital image processing", 2nd edition, pp. 94-102, Prentice Hall, 2002. represents the histogram of NI HistBL, and the histogram of DI HistEL is described by p(zk),k = 0,l,---,2M -1 .
In Fig.3, for the low-bit (M-bit) image NI, the histogram of one color channel is px(x) . The histogram is smoothed by using eq. (5), and a smoothed histogram px ' (x) results, which has the same bin size as the histogram of the high- bit (N-bit) image DI does. The values are accumulated, resulting in sk (k=l,...,16 for the example of Fig.3) . It can be expressed as a function F(xk') .
The high-bit (N-bit) image DI has the histogram pz(z) of the same color channel, which is accumulated to result in vk (same k as above) . It can be expressed as a function G(zk) . Then the distribution v-, is determined, which gives for each value of sk the best-matching value vi (from the vk values of the high-bit image DI) . This distribution v-, sets up a transformation from the values sk (therefore x k' ) to the values vk, and it is the inverse of the transform G.
In the following, the process of histogram smoothing is described. The goal of histogram smoothing is to "stretch" the input histogram so that it will have the same bin size as the desired histogram. This is a prerequisite for the histogram specification process. Following the denotations in Fig.3, the value of px(xk) is uniformly distributed into a corresponding range of [xk,xk ' +1,■ ••,xk + 2(-N"M) -1] , as described in eq. (5) :
Figure imgf000011_0001
0, otherwise To ensure "uniform distribution", un-normalized histograms are used. A simple example of a smoothed histogram is shown in Fig.4. Fig.4 a) shows a color histogram of a base layer input picture with a color bit depth of 2 bit, i.e. 22=4 possible color values of a particular color, e.g. red. The histogram smoothing process "stretches" the histogram to a color bit depth of 4 bit, i.e. 24=16 possible color values. The number of values in a single bin of the input picture is the number of pixels within the input picture that have this particular color value. E.g. a value of 40 in the left-most bin corresponding to the value 0 means that 40 pixels of the input picture have a "red" value of 0 (in this histogram refers to the color "red") . The smoothing process distributes these values to 2N~M corresponding bins of the smoothed histogram. In this example with 2N~M=4 the first four bins of the smoothed histogram correspond to the first bin of the original histogram. According to eq.(5), the 40 values of the original histogram are evenly distributed to the four corresponding bins of the smoothed histogram, as shown in Fig.4 b) .
The idea behind the smoothing is that it improves the continuity of the input histogram, so that the histogram specification will be more efficient. In detail, we write down the alternatives of eq. (1) , (2) and (4) shown in Fig.3 for continuous gray level images:
s = F(x)= [px(μ)du (1')
Figure imgf000012_0001
y = G~l(s) = G~l[F(x)] (4') In the case of continuous gray levels, the input histogram can be specified as exactly the same as the desired histogram. However, as aforementioned, for discrete gray levels only an approximation to the desired histogram can be achieved. Furthermore, as the number of the gray levels is approaching infinity, the approximation is approaching exact match. Therefore, in theory the smoothness procedure is equivalent to increasing the sampling rate during the discretization of the input histogram, and it outperforms the direct histogram specification between two histograms that have different bin size.
The smoothed histogram is just an intermediate step in this algorithm. There is not an image corresponding to it.
The following describes the process of Probability Mapping. Once the classical histogram specification algorithm has been applied to the smoothed histogram px(x) and the desired histogram p~(z) , an intermediate LUT yk - LUTinter(xk) is generated. The next problem is to choose the unique mapped value of xk from the multiple mapped values of its corresponding range [xk,xk +!,■■■,xk +2('N~M^> -I] . Exemplarily, two different criteria are proposed as criterion for probability mapping, as described in eq.(6) and (7) below:
yk e[xk,xk +l,- - -,xk +2{N~M) -1]}, ( 6)
Figure imgf000013_0001
yk
Figure imgf000013_0002
-1]}J , ( 7 )
where yk is the mapped value of χk . A final LUT
yk = LUTβm!(xk), k = 0,l,-,2M~l, yk ≡{zhl = 0X-,2N -1} is generated to map the original histogram to the desired histogram. Eq. (6) says that among the 2N~M values that xk corresponds to, we select the one that has the highest value in the desired histogram pz(y_/) •
Eq. (7) says that among the 2N~M values that xk corresponds to, we use the nearest integer less than or equal to their mean as the finally mapped value .
The LUTmter is a "one-to-one" mapping, because it maps the smoothed histogram of the input image to the histogram of the desired image. However, if we consider the straight histogram of the input image, we can see that each xk corresponds to 2N~M values. The task of "Probability Mapping" is to choose only one value from the 2N~M values as the finally mapped value of xk. Hence, the LUTfinai is still a "one-to-one" mapping: it maps each value xk to a value yk. The mapping of the LUT is invertible because only 2M values of the total values (2N) of the desired image have the corresponding xk.
When the above-described algorithm is applied on an input image and a desired image, the histograms of both images are calculated. Then the input histogram is smoothed, resulting in the "smoothed histogram". After the remaining steps (classical histogram specification and probability mapping) are finished, a final LUT is generated to map the levels of the input histogram to the levels of that of the desired histogram. Then the predicted image is generated by applying the LUT to each pixel of the input image .
Fig.6 shows an overview of a complete enhancement layer encoder, with separate branches for R, G and B images. The base layer is conventionally encoded. The invention is usable for video encoders and video decoders, and particularly for encoding / decoding a color base layer and a color enhancement layer of the same spatial resolution with optimized efficiency.

Claims

Claims
1. A method for encoding a first video image (IMBL) consisting of RGB pixel values having M bit each and a second video image (IMEL) consisting of RGB pixel values having N bit each, with M<N, wherein the first and second video image have the same spatial resolution, comprising the steps of generating a first transfer function (LUTlnter) having 2N possible input values and being based on color histograms (HistBL, HistEL) of the first and the second video image; generating a second transfer function (LUTfinai) based on the first transfer function (LUTlnter) , wherein the second transfer function has 2M possible input values and comprises a function for mapping each of its input values to one out of
2 (N-M) correSpOnding values of the first transfer function; - applying (TR) the second transfer function
(LUTfmai) to the first video image, wherein the second transfer function is used for mapping the pixels, and wherein a transformed first video image (IMTRjBL) is generated; - calculating the difference (IMres) between the transformed first video image (IMTRfBL) and the second video image (IMEL) ; encoding the difference (IMres) ; encoding the first video image (IMBL) ; and - transmitting the encoded first video image (IMBL) , the second transfer function (LUTflnal) and the encoded difference (IMres) •
2. Method according to claim 1, wherein the step of generating a first transfer function (LUTinter) comprises the steps of generating a first color histogram (HistBL) of the reconstructed first image, the first color histogram having 2M bins; generating a second color histogram (HistEL) of the second image, the second color histogram having 2N bins; and - generating (sm) from the first color histogram (HistBL) a smoothed color histogram (SmHistBL) having 2N bins, wherein the values of each bin of the first histogram are distributed to 2 (N~M> bins, and wherein the first transfer function (LUTinter) defines a transfer between the smoothed color histogram (SmHistBL) and the second color histogram (HistEL) .
3. Method according to claim 1 or 2, further comprising the initial steps of encoding the first video image (IMBL); and decoding the encoded first video image to obtain a reconstructed first video image (IMBL,rec) , wherein the step of applying (TR) the second transfer function refers to the reconstructed first video image.
4. Method according to any of claims 1-3, wherein the function for mapping each of the input values of the second transfer function (LUTfα.nai) to one out of 2 (N-M) corresponding values of the first transfer function is a probability mapping function.
5. Method according to claim 4, wherein the probability mapping function selects for each of the 2M values of the second transfer function one value from the 2N~M corresponding values of the first transfer function, being the one that has the highest value in the second histogram (HistEL) ■
6. Method according to claim 4, wherein the probability mapping function selects for each of the 2M values of the second transfer function one value from the 2N~M corresponding values of the first transfer function, being the one that is the nearest integer less than or equal to the mean value of said 2N~M corresponding values .
7. Method for decoding a first and a second video image received in a bitstream (bs), wherein the first and the second video image have the same spatial resolution, comprising the steps of - extracting from the bitstream (bs) video data for the first and the second video image, and color enhancement control data; decoding and reconstructing (decBi) the first video image, wherein a reconstructed first video image (IMBL) is obtained having color pixel values with M bit each; constructing (decLϋT) a mapping table (LUTfinai) from the color enhancement control data; applying (TRdec) the mapping table (LUTfinai) to the pixels of the reconstructed first video image
(IMBL) r wherein a transformed first video image (IMTR,BL) is generated having a different color histogram than the reconstructed first video image; decoding the second video image (IMres) i wherein the second video image comprises pixel difference values; and adding the decoded second video image (IMres) to the transformed first video image (IMTR,BL) r wherein a reconstructed second video image (IMEL) is obtained having color pixel values with N bit each, with M<N .
8. Method according to claim 7, wherein the transformed first video image (IMTR/BL) has color histograms for red, green and/or blue that differ from the respective color histograms of the reconstructed first video image (IMBL) •
9. Method according to claim 7 or 8, wherein three separate mapping tables (LUTfinai) or mapping table areas are generated from the color enhancement control data, wherein each of the separate mapping tables or mapping table areas relates to one of the colors red, green and blue.
10. Apparatus for encoding a first video image (IMBL) consisting of RGB pixel values having M bit each and a second video image (IMEL) consisting of RGB pixel values having N bit each, with M<N, wherein the first and second video image have the same spatial resolution, comprising means for generating a first transfer function (LUTmter) having 2N possible input values and being based on color histograms (HistBL, HistE]j) of the first and the second video image; means for generating a second transfer function (LUTflnai) based on the first transfer function (LUTmter) r wherein the second transfer function has 2M possible input values and comprises a function for mapping each of its input values to one out of 2 <N~M) corresponding values of the first transfer function; - means for applying (TR) the second transfer function (LUTflnai) to the first video image, wherein the second transfer function is used for mapping the pixels, and wherein a transformed first video image (IMTR,BL) is generated; - means for calculating the difference (IMres) between the transformed first video image (IMTR,BL) and the second video image (IMEL); means for encoding the difference (IMres); means for encoding the first video image (IMBL) ; and means for transmitting the encoded first video image (IMBL) , the second transfer function (LUTfmai) and the encoded difference (IMres) •
11. Apparatus according to claim 10, wherein the means for generating a first transfer function (LUTmter) comprises means for generating a first color histogram (HistBL) of the reconstructed first image, the first color histogram having 2M bins; means for generating a second color histogram (HistEL) of the second image, the second color histogram having 2N bins; and means for generating (sm) from the first color histogram (HistBL) a smoothed color histogram (SmHistBL) having 2N bins, wherein the values of each bin of the first histogram are distributed to 2(N~M) bins, and wherein the first transfer function (LUTinter) defines a transfer between the smoothed color histogram (SmHistBL) and the second color histogram (HistEL) .
12. Apparatus for decoding a first and a second video image received in a bitstream (bs) , wherein the first and the second video image have the same spatial resolution, comprising means for extracting from the bitstream (bs) video data for the first and the second video image, and color enhancement control data; means for decoding and reconstructing (decBL) the first video image, wherein a reconstructed first video image (IMBL) is obtained having color pixel values with M bit each; means for constructing (decLUτ) a mapping table (LUTfmai) from the color enhancement control data; means for applying (TRdec) the mapping table (LUTfλnai) to the pixels of the reconstructed first video image (IMBL) , wherein a transformed first video image (IMTR,BL) is generated having a different color histogram than the reconstructed first video image; means for decoding the second video image (IMres) , wherein the second video image comprises pixel difference values; and means for adding the decoded second video image (IMres) to the transformed first video image (IMTR,BL), wherein a reconstructed second video image (IMEL) is obtained having color pixel values with N bit each, with M<N .
PCT/CN2006/001699 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data WO2008019524A1 (en)

Priority Applications (7)

Application Number Priority Date Filing Date Title
DE200660019006 DE602006019006D1 (en) 2006-07-17 2006-07-17 METHOD AND DEVICE FOR CODING VIDEO COLOR EXPANSION DATA AND METHOD AND DEVICE FOR DECODING VIDEO COLOR EXPANSION DATA
US12/309,335 US8761249B2 (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data
EP20060761443 EP2041983B1 (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data
PCT/CN2006/001699 WO2008019524A1 (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data
JP2009519775A JP4991851B2 (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data and method and apparatus for decoding video color enhancement data
CN2006800553696A CN101507284B (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data
KR20097001015A KR101196975B1 (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2006/001699 WO2008019524A1 (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data

Publications (1)

Publication Number Publication Date
WO2008019524A1 true WO2008019524A1 (en) 2008-02-21

Family

ID=39081897

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2006/001699 WO2008019524A1 (en) 2006-07-17 2006-07-17 Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data

Country Status (7)

Country Link
US (1) US8761249B2 (en)
EP (1) EP2041983B1 (en)
JP (1) JP4991851B2 (en)
KR (1) KR101196975B1 (en)
CN (1) CN101507284B (en)
DE (1) DE602006019006D1 (en)
WO (1) WO2008019524A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008083521A1 (en) 2007-01-10 2008-07-17 Thomson Licensing Video encoding method and video decoding method for enabling bit depth scalability
WO2010021705A1 (en) * 2008-08-22 2010-02-25 Thomson Licensing Method and system for content delivery
JP2010199925A (en) * 2009-02-25 2010-09-09 Nippon Telegr & Teleph Corp <Ntt> Method, apparatus and program for encoding image
CN101577828B (en) * 2008-04-16 2013-06-26 英特尔公司 Tone mapping for bit-depth scalable video codec
US8867616B2 (en) 2009-02-11 2014-10-21 Thomson Licensing Methods and apparatus for bit depth scalable video encoding and decoding utilizing tone mapping and inverse tone mapping
US8982963B2 (en) 2009-03-13 2015-03-17 Dolby Laboratories Licensing Corporation Compatible compression of high dynamic range, visual dynamic range, and wide color gamut video
US9549194B2 (en) 2012-01-09 2017-01-17 Dolby Laboratories Licensing Corporation Context based inverse mapping method for layered codec
EP3110154A4 (en) * 2014-02-21 2017-10-11 Sony Corporation Transmission device, transmission method, reception device, and reception method
US20210019868A1 (en) * 2016-03-18 2021-01-21 Koninklijke Philips N.V. Encoding and decoding hdr videos

Families Citing this family (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101328547B1 (en) 2004-11-01 2013-11-13 테크니컬러, 인크. Method and system for mastering and distributing enhanced color space content
US8773494B2 (en) 2006-08-29 2014-07-08 Microsoft Corporation Techniques for managing visual compositions for a multimedia conference call
US8085852B2 (en) * 2007-06-26 2011-12-27 Mitsubishi Electric Research Laboratories, Inc. Inverse tone mapping for bit-depth scalable image coding
BRPI0817767A2 (en) * 2007-10-16 2020-01-28 Thomson Licensing artifact removal methods and apparatus for bit depth scaling capability
US9538176B2 (en) 2008-08-08 2017-01-03 Dolby Laboratories Licensing Corporation Pre-processing for bitdepth and color format scalable video coding
US8432413B2 (en) * 2008-11-17 2013-04-30 Xrfiles, Inc. System and method for the display of extended bit depth high resolution images
JP2010278724A (en) * 2009-05-28 2010-12-09 Olympus Corp Image processing apparatus, image processing method, and image processing program
TWI407426B (en) * 2010-04-15 2013-09-01 Nuvoton Technology Corp Display device, control circuit thereof, and method of displaying image data
TWI431512B (en) * 2010-06-23 2014-03-21 Ind Tech Res Inst Method for recognizing three-dimensional control point and computer readable medium using the same thereof
CN102331883B (en) * 2010-07-14 2013-11-06 财团法人工业技术研究院 Identification method for three-dimensional control end point and computer readable medium adopting same
CN101923708B (en) * 2010-08-09 2012-07-25 福建华映显示科技有限公司 Depth map reinforcement method
JP2012060261A (en) 2010-09-06 2012-03-22 Sony Corp Image processing apparatus and method
KR20120087233A (en) 2010-12-03 2012-08-07 한국전자통신연구원 Method for tile-map service using image fusion, and system for the same
CN104054338B (en) * 2011-03-10 2019-04-05 杜比实验室特许公司 Locating depth and color scalable video
WO2012125892A1 (en) 2011-03-17 2012-09-20 Dolby Laboratories Licensing Corporation Generating alternative versions of image content using histograms
US9036042B2 (en) 2011-04-15 2015-05-19 Dolby Laboratories Licensing Corporation Encoding, decoding, and representing high dynamic range images
TWI690211B (en) 2011-04-15 2020-04-01 美商杜比實驗室特許公司 Decoding method for high dynamic range images, processor non-transistory readable medium and computer program product thereof
US8334911B2 (en) 2011-04-15 2012-12-18 Dolby Laboratories Licensing Corporation Encoding, decoding, and representing high dynamic range images
CN103106668B (en) * 2011-11-09 2016-08-03 佳能株式会社 The method and system of image-region is described based on color histogram
US8934726B2 (en) 2011-12-19 2015-01-13 Dolby Laboratories Licensing Corporation Video codecs with integrated gamut management
CN102542528B (en) * 2011-12-26 2013-10-09 Tcl集团股份有限公司 Image conversion processing method and system
TWI556629B (en) 2012-01-03 2016-11-01 杜比實驗室特許公司 Specifying visual dynamic range coding operations and parameters
BR112015003707A2 (en) * 2012-08-27 2017-07-04 Sony Corp device and method of transmission, and device and method of reception.
TWI545964B (en) * 2012-09-14 2016-08-11 鴻海精密工業股份有限公司 Image white balancing method and image capturing device using same
KR101901184B1 (en) 2012-09-20 2018-09-21 삼성전자주식회사 Apparatus and method for processing color image using depth image
EP2898694B1 (en) * 2012-09-24 2019-06-19 Huawei Technologies Co., Ltd. Video compression with color space scalability
US9532057B2 (en) * 2012-12-28 2016-12-27 Qualcomm Incorporated Inter-layer prediction using sample-adaptive adjustments for bit depth scalable video coding
US10469866B2 (en) * 2013-04-05 2019-11-05 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding video with respect to position of integer pixel
US9445713B2 (en) 2013-09-05 2016-09-20 Cellscope, Inc. Apparatuses and methods for mobile imaging and analysis
US9948916B2 (en) 2013-10-14 2018-04-17 Qualcomm Incorporated Three-dimensional lookup table based color gamut scalability in multi-layer video coding
US10531105B2 (en) 2013-12-17 2020-01-07 Qualcomm Incorporated Signaling partition information for 3D lookup table for color gamut scalability in multi-layer video coding
US9756337B2 (en) 2013-12-17 2017-09-05 Qualcomm Incorporated Signaling color values for 3D lookup table for color gamut scalability in multi-layer video coding
US9445111B2 (en) * 2014-01-08 2016-09-13 Sony Corporation Image processing apparatus and image processing method
US10531109B2 (en) 2015-01-27 2020-01-07 Dolby Laboratories Licensing Corporation Predictive image encoding and decoding with pixel group based quantization
US11178204B1 (en) * 2017-02-23 2021-11-16 Cox Communications, Inc. Video processor to enhance color space and/or bit-depth
CN111145100B (en) * 2018-11-02 2023-01-20 深圳富泰宏精密工业有限公司 Dynamic image generation method and system, computer device and readable storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1447285A (en) * 2002-03-27 2003-10-08 微软公司 System and method for progressively changing and coding digital data
CN1764277A (en) * 2001-09-18 2006-04-26 微软公司 Improved block transform and quantization for image and video coding

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4857993A (en) * 1988-04-27 1989-08-15 Universal Video Communications Corp. Method and system for decompressing digital color video statistically encoded data
US6185327B1 (en) * 1998-05-04 2001-02-06 Eastman Kodak Company Method and system for the reduction of memory capacity required for a digital representation of an image
JP2005027168A (en) 2003-07-04 2005-01-27 Canon Inc Image processor and image processing method
EP1591963B1 (en) * 2004-04-29 2008-07-09 Mitsubishi Electric Information Technology Centre Europe B.V. Adaptive quantisation of a depth map
US7991238B2 (en) * 2004-04-30 2011-08-02 Neiversan Networks Co. Llc Adaptive compression of multi-level images
CN100544398C (en) * 2004-04-30 2009-09-23 微软公司 The method and system of the adaptability compression of multi-level images
US20050259729A1 (en) * 2004-05-21 2005-11-24 Shijun Sun Video coding with quality scalability
US8537893B2 (en) 2006-01-23 2013-09-17 Max-Planck-Gesellschaft Zur Forderung Der Wissenschaften E.V. High dynamic range codecs

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1764277A (en) * 2001-09-18 2006-04-26 微软公司 Improved block transform and quantization for image and video coding
CN1447285A (en) * 2002-03-27 2003-10-08 微软公司 System and method for progressively changing and coding digital data

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2041983A4 *

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2127395A4 (en) * 2007-01-10 2013-01-16 Thomson Licensing Video encoding method and video decoding method for enabling bit depth scalability
EP2127395A1 (en) * 2007-01-10 2009-12-02 Thomson Licensing Video encoding method and video decoding method for enabling bit depth scalability
WO2008083521A1 (en) 2007-01-10 2008-07-17 Thomson Licensing Video encoding method and video decoding method for enabling bit depth scalability
US9338475B2 (en) 2008-04-16 2016-05-10 Intel Corporation Tone mapping for bit-depth scalable video codec
CN101577828B (en) * 2008-04-16 2013-06-26 英特尔公司 Tone mapping for bit-depth scalable video codec
WO2010021705A1 (en) * 2008-08-22 2010-02-25 Thomson Licensing Method and system for content delivery
US8867616B2 (en) 2009-02-11 2014-10-21 Thomson Licensing Methods and apparatus for bit depth scalable video encoding and decoding utilizing tone mapping and inverse tone mapping
JP2010199925A (en) * 2009-02-25 2010-09-09 Nippon Telegr & Teleph Corp <Ntt> Method, apparatus and program for encoding image
US8982963B2 (en) 2009-03-13 2015-03-17 Dolby Laboratories Licensing Corporation Compatible compression of high dynamic range, visual dynamic range, and wide color gamut video
US9549194B2 (en) 2012-01-09 2017-01-17 Dolby Laboratories Licensing Corporation Context based inverse mapping method for layered codec
EP3110154A4 (en) * 2014-02-21 2017-10-11 Sony Corporation Transmission device, transmission method, reception device, and reception method
RU2683628C1 (en) * 2014-02-21 2019-03-29 Сони Корпорейшн Transmission device, transmission method, reception device and reception method
US10674181B2 (en) 2014-02-21 2020-06-02 Sony Corporation Transmission device, transmission method, reception device, and reception method
US10735771B2 (en) 2014-02-21 2020-08-04 Sony Corporation Transmission device, transmission method, reception device, and reception method
US11330303B2 (en) 2014-02-21 2022-05-10 Sony Corporation Transmission device, transmission method, reception device, and reception method
US20210019868A1 (en) * 2016-03-18 2021-01-21 Koninklijke Philips N.V. Encoding and decoding hdr videos
US11593922B2 (en) * 2016-03-18 2023-02-28 Koninklijke Philips N.V. Encoding and decoding HDR videos
US20230177659A1 (en) * 2016-03-18 2023-06-08 Koninklijke Philips N.V. Encoding and decoding hdr videos
US11887285B2 (en) 2016-03-18 2024-01-30 Koninklijke Philips N.V. Encoding and decoding HDR videos

Also Published As

Publication number Publication date
US8761249B2 (en) 2014-06-24
JP4991851B2 (en) 2012-08-01
US20090285283A1 (en) 2009-11-19
DE602006019006D1 (en) 2011-01-27
CN101507284B (en) 2012-06-27
EP2041983A1 (en) 2009-04-01
JP2009544198A (en) 2009-12-10
CN101507284A (en) 2009-08-12
KR20090052849A (en) 2009-05-26
EP2041983B1 (en) 2010-12-15
EP2041983A4 (en) 2010-04-07
KR101196975B1 (en) 2012-11-02

Similar Documents

Publication Publication Date Title
EP2041983B1 (en) Method and apparatus for encoding video color enhancement data, and method and apparatus for decoding video color enhancement data
KR102129541B1 (en) Color volume transformations in coding of high dynamic range and wide color gamut sequences
JP6609056B2 (en) System for reconstruction and encoding of high dynamic range and wide color gamut sequences
US20230276061A1 (en) Scalable video coding system with parameter signaling
US10158835B2 (en) Extending image dynamic range
US10575005B2 (en) Video coding and delivery with both spatial and dynamic range scalability
CN106488246B (en) Method for enhancing the adaptive shaping of the hierarchical coding of dynamic range signal
CN101622878B (en) Video encoding method and video decoding method for enabling bit depth scalability
WO2020074897A1 (en) Dynamic range support within a multi-layer hierarchical coding scheme
WO2008071542A1 (en) Method and apparatus for encoding and/or decoding video data using adaptive prediction order for spatial and bit depth prediction
CN106412595B (en) Method and apparatus for encoding high dynamic range frames and applied low dynamic range frames
EP3251351B1 (en) Predictive image encoding and decoding with pixel group based quantization
Nascimento et al. H. 264/AVC backward compatible bit-depth scalable video coding
Hsiao et al. Dual-Layer Lossless Coding for Infrared Video

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200680055369.6

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 06761443

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2006761443

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 12309335

Country of ref document: US

Ref document number: 2009519775

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 1020097001015

Country of ref document: KR

NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: RU