WO2015193117A1 - Procédé et dispositif de décodage d'une image hdr à partir d'un flux binaire représentant une image ldr et une image d'éclairage - Google Patents

Procédé et dispositif de décodage d'une image hdr à partir d'un flux binaire représentant une image ldr et une image d'éclairage Download PDF

Info

Publication number
WO2015193117A1
WO2015193117A1 PCT/EP2015/062477 EP2015062477W WO2015193117A1 WO 2015193117 A1 WO2015193117 A1 WO 2015193117A1 EP 2015062477 W EP2015062477 W EP 2015062477W WO 2015193117 A1 WO2015193117 A1 WO 2015193117A1
Authority
WO
WIPO (PCT)
Prior art keywords
picture
decoded version
color
ldr
sample
Prior art date
Application number
PCT/EP2015/062477
Other languages
English (en)
Inventor
Pierre Andrivon
Sebastien Lasserre
Edouard Francois
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing filed Critical Thomson Licensing
Publication of WO2015193117A1 publication Critical patent/WO2015193117A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/127Prioritisation of hardware or computational resources
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/98Adaptive-dynamic-range coding [ADRC]

Definitions

  • the present disclosure generally relates to picture/video decoding. Particularly, but not exclusively, the technical field of the present disclosure is related to decoding of a picture whose pixels values belong to a high- dynamic range.
  • a picture (sometimes called an image or frame in prior art) contains one or several arrays of samples (pixel values) in a specific picture/video format which specifies all information relative to the pixel values of a picture (or a video) and all information which may be used by a display and/or a decoding device to visualize and/or decode a picture (or video).
  • a picture comprises at least one component, in the shape of a first array of samples, usually a luma (or luminance) component, and, possibly, at least one other component, in the shape of at least one other array of samples, usually a color component.
  • LDR pictures Low-Dynamic-Range pictures
  • HDR pictures high-dynamic range pictures
  • luma samples are usually represented in floating-point format (either 32-bit or 16-bit for each component, namely float or half-float), the most popular format being openEXR half-float format (16-bit per RGB component, i.e. 48 bits per sample) or in integers with a long representation, typically at least 1 6 bits.
  • a dual modulation scheme is a typical approach for encoding an input
  • HDR picture in a bitstream and for obtaining a decoded version of the input HDR picture by decoding the bitstream at least partially.
  • an illumination picture IF (also called illumination map or backlight picture) is obtained from the input HDR picture.
  • a residual picture Res is then obtained by dividing the input HDR picture by the illumination picture IF and both the illumination picture IF (or illumination data representing the illumination picture) and the residual picture Res are then encoded in a bitstream F.
  • Encoding an input HDR picture using this approach leads to encode two components in a bitstream F: a residual picture Res (called the LDR picture in the following), which may be a viewable picture, i.e. a picture with reduced visual artifacts and adapted to be viewed on a display, and illumination picture IF (or illumination data representing the illumination picture).
  • Fig. 1 shows a diagram of the steps of a method for decoding an HDR picture from the bitstream F according to the prior art (for example, "High Dynamic Range Video Coding", Lasserre et al., JCTVC-P0159/m32076,16 th MPEG meeting, San Jose (CA), 9 -17 Jan. 2014).
  • a decoder DEC is configured for obtaining the sample values of a decoded version Res of the LDR picture Res and the sample values of a decoded version TP of the illumination picture IF by directly decoding the bitstream F at least partially.
  • a decoded version / of the HDR picture to be decoded is then obtained by multiplying the decoded version of the LDR picture Res by the decoded version of the illumination picture IF.
  • some specific post-processing are applied to the decoded version Res of the LDR picture Res and to the decoded version TP of the illumination picture IF.
  • a module PIF is configured for applying some post- processing to the decoded version TP of the illumination picture IF which may be, for a non limitative example, an upsampling and/or an inverse- gamma correction.
  • step 12 the chroma components of the decoded version Res of the LDR picture Res are upsampled in order to convert the usual 4:2:0 format of the decoded version Res of the LDR picture Res to a 4:4:4 format.
  • This step 1 2 is of course optional.
  • sample values of the decoded version Res of the LDR picture Res are usually expressed in the well-known YCbCr color space.
  • a module CSC1 is configured for obtaining a RGB color value for each YCbCr sample value of the decoded version Res of the LDR picture Res.
  • step 14 some other post-processing are applied to the decoded version Res of the LDR picture.
  • a module SCL scales the decoded version Res of the LDR picture Res by dividing each sample value of the decoded version Res of the LDR picture Res by a scaling factor cst sca iing- The resulting LDR picture Res s is then given by
  • the scaling factor cst sca iing is defined to map the sample values of the decoded version Res of the LDR picture Res from 0 to the maximum value 2 N -1 , where N is the number of bits allowed as input for further post-processing. This is naturally obtained by mapping the value 1 (which is roughly the mean value of the decoded version Res of the LDR picture Res) to the mid- gray value 2 N 1 .
  • N the number of bits allowed as input for further post-processing.
  • a module ITMO applied an inverse-tone- mapping to the decoded version Res of the LDR picture.
  • Tone mapping a LDR picture at the encoding side, comprises either a gamma correction or a SLog correction according to the sample values of the LDR picture Res.
  • a tone-mapped version Res v of the LDR picture Res is given, for example, by:
  • being a coefficient of a gamma curve equal, for example, to 1 /2.4.
  • the tone-mapped version Res v of the LDR picture Res is given, for example, by:
  • a,b and c are coefficients of a SLog curve determined such that 0 and 1 are invariant, and the derivative of the SLog curve is continuous in 1 when prolonged by a gamma curve below 1 .
  • a,b and c are functions of the parameter y.
  • tone-mapping a LDR picture Res comprises either the gamma correction or the SLog correction according to the sample values of the LDR picture Res. For example, when the pixel value of the LDR picture Res is below a threshold (for example equal to 1 ), then the gamma correction is applied and otherwise the SLog correction is applied.
  • Inverse-tone-mapping the decoded version Res of the LDR picture Res comprises applying either an inverse Slog-correction or inverse gamma- correction. For example, the inverse-tone-mapping is just to find, from the gamma curve, the values which correspond to the sample values of the decoded version Res of the LDR picture Res using the gamma curve.
  • a module RGBF is configured for applying a RGB factor to the decoded version Res of the LDR picture Res, i.e. multiplying the sample values of the decoded version Res of the LDR picture by a coefficient, adding an offset to get resulting values and finally right- shifting the resulting values.
  • a module CSC2 is configured for obtaining a color value expressed in a specific output color space for each sample value of the decoded version / of the HDR picture.
  • module CSC2 may be configured for applying other processing such a specific gamma correction and/or OETF (Opto-Electrical Transfer Function).
  • OETF Opto-Electrical Transfer Function
  • the specific output space is linear. Any RGB or XYZ (also called CIEXYZ) color spaces may be used as output color space.
  • a XYZ color space may be combined with a RGB color space with primaries compliant with Rec. 2020.
  • the module CSC2 is then configured for obtaining a XYZ color value for each RGB sample value of the decoded version / of the HDR picture.
  • the dynamic range of the intermediate values reaches up to 18 bits which cannot be implemented on a typical 1 6-bits integer color converter.
  • Other examples show that 24-bits input requires 34-bits intermediate values.
  • aspects of the present disclosure are directed to creating and maintaining semantic relationships between data objects on a computer system.
  • the following presents a simplified summary of the disclosure in order to provide a basic understanding of some aspects of the disclosure. This summary is not an extensive overview of the disclosure. It is not intended to identify key or critical elements of the disclosure. The following summary merely presents some aspects of the disclosure in a simplified form as a prelude to the more detailed description provided below.
  • the disclosure sets out to remedy some of the drawbacks of the prior art with a method for decoding a HDR picture from a bitstream representing a LDR picture and an illumination picture.
  • the method comprises:
  • Color-converting the sample values of the decoded version of the LDR picture just before multiplying rather than color-converting the decoded version of the HDR picture reduces significantly the dynamic range of the sample values to be color-converted (and of the intermediate values requires during the color conversion processing) and allows a 16 (or 32) bit integer implementation of the complete decoding method taking into account the constraints of the typical 16 (or 32) bit color converters.
  • the color conversion (module CSC2) which usually occurs after the multiplication of the decoded version of the LDR picture by the decoded version of the illumination picture has been replaced by a color conversion of the decoded version of the LDR picture, in accordance with the disclosure.
  • Note such a modification of the decoding method swaps linear operations (color conversion, multiplication).
  • such a modification of the decoding method of a usual dual modulation scheme allows an integer implementation of the modified decoding method of a usual dual modulation that preserves the performance of a whole floating point processing workflow.
  • the method further comprising scaling and color-converting the sample values of the decoded version of the LDR, the sample values of the decoded version of the LDR are color-converted before scaled in accordance with this variant.
  • Fig. 1 shows a diagram of the steps of a method for decoding an HDR picture from the bitstream F according to the prior art
  • Fig. 2 shows a diagram of the steps of a method for decoding an
  • FIG. 3 shows a diagram of the steps of a method for decoding an HDR picture from the bitstream F in accordance with a variant of the embodiment of the disclosure described in relation with Fig. 1 ;
  • - Fig. 4 shows an illustration of the dynamic range of both a YCbCr and RGB color space
  • - Fig. 5 shows an example of an architecture of a device in accordance with an embodiment of the disclosure.
  • each block represents a circuit element, module, or portion of code which comprises one or more executable instructions for implementing the specified logical function(s).
  • the function(s) noted in the blocks may occur out of the order noted. For example, two blocks shown in succession may, in fact, be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending on the functionality involved.
  • a picture (sometimes called an image or frame in prior art) contains an array of samples (pixel values) in a specific picture/video format which specified all information relative to the pixel values of the picture (or a video) and all information which may be used by a display and/or a decoding device.
  • a picture comprises at least one component, usually a luma (or luminance) component, and, possibly, at least one other component, usually a color component.
  • the disclosure is described for decoding a picture but extends to the decoding of a sequence of pictures (video) because each picture of the sequence is sequentially decoded as described below.
  • Fig. 2 shows a diagram of the steps of a method for decoding an HDR picture from the bitstream F in accordance with an embodiment of the disclosure.
  • the module CSC2 now applied to the decoded version Res of the LDR picture Res (possibly post-processed) and not to the decoded version / of the HDR picture as described in prior art (Fig. 1 )-
  • a color value expressed in the output color space e.g. CIEXYZ
  • a color space e.g. RGB
  • Fig. 3 shows a diagram of the steps of a method for decoding an HDR picture from the bitstream F in accordance with a variant of the embodiment of the disclosure described in relation with Fig. 1.
  • the decoding method further comprises scaling (step 141 ) and color-converting (step 13) the decoded version Res of the LDR picture Res .
  • the sample values of the decoded version Res of the LDR picture Res are scaled (step 141 ) before being color- converted (step 13). More precisely, a RGB color value is obtained for each scaled YCbCr sample value following the above example.
  • Scaling YCbCr sample values rather than scaling RGB sample values reduced the dynamic range of the resulting scaled and color-converted sample values because the dynamic range of the YCbCr sample values is smaller than the dynamic range of the RGB sample values as explained in Fig. 4.
  • Fig. 4 shows an illustration of the volume of both YCbCr and RGB color spaces.
  • the color values are expressed in a wider volume when these color values are expressed in the YCbCr color space rather than in the RGB color space.
  • 3 ⁇ 4 or more of the YCbCr code combinations do not represent colors. For example, when the 8-bit Rec.
  • the YCbCr color space represents fewer colors (or equivalently more quantization noise), than the RGB color space (SIGGRAPH2004: Color science and color appearance models for CG, HDTV, D-Cinema, Charles
  • the dynamic range of the sample values expressed as YCbCr colors is smaller than the dynamic range of these sample values when they are expressed as RGB colors.
  • the modules are functional units, which may or not be in relation with distinguishable physical units. For example, these modules or some of them may be brought together in a unique component or circuit, or contribute to functionalities of a software. A contrario, some modules may potentially be composed of separate physical entities.
  • the apparatus which are compatible with the invention are implemented using either pure hardware, for example using dedicated hardware such ASIC or FPGA or VLSI, respectively « Application Specific Integrated Circuit " , " Field- Programmable Gate Array » « Very Large Scale Integration » or from several integrated electronic components embedded in a device or from a blend of hardware and software components.
  • Fig. 5 represents an exemplary architecture of a device 50 which may be configured to implement a method described in relation with Fig. 1-3.
  • Device 50 comprises following elements that are linked together by a data and address bus 51 :
  • microprocessor 52 which is, for example, a DSP (or Digital Signal Processor);
  • RAM or Random Access Memory
  • the battery 56 is external to the device.
  • the word « register » used in the specification can correspond to area of small capacity (some bits) or to very large area (e.g. a whole program or large amount of received or decoded data).
  • ROM 53 comprises at least a program and parameters. Algorithm of the methods according to the invention is stored in the ROM 53. When switched on, the CPU 52 uploads the program in the RAM and executes the corresponding instructions.
  • RAM 54 comprises, in a register, the program executed by the CPU 52 and uploaded after switch on of the device 50, input data in a register, intermediate data in different states of the method in a register, and other variables used for the execution of the method in a register.
  • the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method or a device), the implementation of features discussed may also be implemented in other forms (for example a program).
  • An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • the decoded picture ⁇ is sent to a destination; specifically, the destination belongs to a set comprising:
  • a local memory e.g. a video memory or a RAM, a flash memory, a hard disk ;
  • a storage interface e.g. an interface with a mass storage, a RAM, a flash memory, a ROM, an optical disc or a magnetic support;
  • a communication interface e.g. a wireline interface (for example a bus interface (e.g. USB (or Universal Serial Bus)), a wide area network interface, a local area network interface, a HDMI (High Definition Multimedia Interface) interface) or a wireless interface (such as a IEEE 802.1 1 interface, WiFi ® or a Bluetooth ® interface); and
  • a wireline interface for example a bus interface (e.g. USB (or Universal Serial Bus)
  • a wide area network interface e.g. USB (or Universal Serial Bus)
  • a wide area network interface e.g. USB (or Universal Serial Bus)
  • a local area network interface e.g. USB (or Universal Serial Bus)
  • HDMI High Definition Multimedia Interface
  • a wireless interface such as a IEEE 802.1 1 interface, WiFi ® or a Bluetooth ® interface
  • the bitstream F is obtained from a source.
  • the bitstream is read from a local memory, e.g. a video memory (54), a RAM (54), a ROM (53), a flash memory (53) or a hard disk (53).
  • the bitstream is received from a storage interface (55), e.g. an interface with a mass storage, a RAM, a ROM, a flash memory, an optical disc or a magnetic support and/or received from a communication interface (55), e.g. an interface to a point to point link, a bus, a point to multipoint link or a broadcast network.
  • device 50 being configured to implement a decoding method described in relation with Fig. 1-3, belongs to a set comprising:
  • Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications, particularly, for example, equipment or applications.
  • equipment or applications include a decoder, a post-processor processing output from a decoder, a video decoder, a web server, a set-top box, a laptop, a personal computer, a cell phone, a PDA, and other communication devices.
  • the equipment may be mobile and even installed in a mobile vehicle.
  • a computer readable storage medium can take the form of a computer readable program product embodied in one or more computer readable medium(s) and having computer readable program code embodied thereon that is executable by a computer.
  • a computer readable storage medium as used herein is considered a non-transitory storage medium given the inherent capability to store the information therein as well as the inherent capability to provide retrieval of the information therefrom.
  • a computer readable storage medium can be, for example, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. It is to be appreciated that the following, while providing more specific examples of computer readable storage mediums to which the present principles can be applied, is merely an illustrative and not exhaustive listing as is readily appreciated by one of ordinary skill in the art: a portable computer diskette; a hard disk; a read-only memory (ROM); an erasable programmable read-only memory (EPROM or Flash memory); a portable compact disc read-only memory (CD-ROM); an optical storage device; a magnetic storage device; or any suitable combination of the foregoing.
  • the instructions may form an application program tangibly embodied on a processor-readable medium.
  • Instructions may be, for example, in hardware, firmware, software, or a combination. Instructions may be found in, for example, an operating system, a separate application, or a combination of the two.
  • a processor may be characterized, therefore, as, for example, both a device configured to carry out a process and a device that includes a processor-readable medium (such as a storage device) having instructions for carrying out a process. Further, a processor-readable medium may store, in addition to or in lieu of instructions, data values produced by an implementation.
  • implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted.
  • the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
  • a signal may be formatted to carry as data the rules for writing or reading the syntax of a described embodiment, or to carry as data the actual syntax-values written by a described embodiment.
  • Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
  • the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
  • the information that the signal carries may be, for example, analog or digital information.
  • the signal may be transmitted over a variety of different wired or wireless links, as is known.
  • the signal may be stored on a processor-readable medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computing Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

La présente invention concerne de manière générale un procédé et un dispositif de décodage d'une image HDR à partir d'un flux binaire représentant une image LDR et une image d'éclairage. Le procédé comprend les étapes consistant à : - obtenir une version décodée de l'image HDR en multipliant les valeurs d'échantillons d'une version décodée de l'image LDR par les valeurs d'échantillons de la version décodée de l'image d'éclairage; et - obtenir une valeur chromatique exprimée dans un espace chromatique de sortie pour chaque valeur d'échantillon de la version décodée de l'image HDR. Le procédé est caractérisé en ce qu'il comprend en outre l'étape consistant à : - obtenir une valeur chromatique exprimée dans l'espace chromatique de sortie pour chaque valeur d'échantillon d'une version décodée de l'image LDR juste avant de multiplier lesdites valeurs chromatiques par les valeurs d'échantillons de la version décodée de l'image d'éclairage.
PCT/EP2015/062477 2014-06-20 2015-06-04 Procédé et dispositif de décodage d'une image hdr à partir d'un flux binaire représentant une image ldr et une image d'éclairage WO2015193117A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP14305951.7 2014-06-20
EP14305951 2014-06-20

Publications (1)

Publication Number Publication Date
WO2015193117A1 true WO2015193117A1 (fr) 2015-12-23

Family

ID=51162633

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2015/062477 WO2015193117A1 (fr) 2014-06-20 2015-06-04 Procédé et dispositif de décodage d'une image hdr à partir d'un flux binaire représentant une image ldr et une image d'éclairage

Country Status (2)

Country Link
TW (1) TW201601521A (fr)
WO (1) WO2015193117A1 (fr)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8248486B1 (en) * 2011-04-15 2012-08-21 Dolby Laboratories Licensing Corporation Encoding, decoding, and representing high dynamic range images

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8248486B1 (en) * 2011-04-15 2012-08-21 Dolby Laboratories Licensing Corporation Encoding, decoding, and representing high dynamic range images

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
GREG WARD ET AL: "JPEG-HDR", ACM SIGGRAPH 2006 COURSES ON , SIGGRAPH '06, 1 January 2006 (2006-01-01), New York, New York, USA, pages 3, XP055117692, ISBN: 978-1-59-593364-5, DOI: 10.1145/1185657.1185685 *
LASSERRE S ET AL: "High Dynamic Range video coding", 16. JCT-VC MEETING; 9-1-2014 - 17-1-2014; SAN JOSE; (JOINT COLLABORATIVE TEAM ON VIDEO CODING OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ); URL: HTTP://WFTP3.ITU.INT/AV-ARCH/JCTVC-SITE/,, no. JCTVC-P0159, 5 January 2014 (2014-01-05), XP030115677 *
SEWOONG OH: "High Dynamic Range Image Encoding for BrightSide Display", 1 January 2006 (2006-01-01), XP055130767, Retrieved from the Internet <URL:http://scien.stanford.edu/pages/labsite/2007/psych221/projects/07/HDR_encoding/SewoongOh_report.pdf> [retrieved on 20140722] *

Also Published As

Publication number Publication date
TW201601521A (zh) 2016-01-01

Similar Documents

Publication Publication Date Title
US10873764B2 (en) Method for encoding an input video comprising a luma component and two chroma components, the method comprising reshaping of said input video based on reshaping functions
US20180005358A1 (en) A method and apparatus for inverse-tone mapping a picture
US11062432B2 (en) Method and device for reconstructing an HDR image
US20220318965A1 (en) Method and device for obtaining a second image from a first image when the dynamic range of the luminance of the first image is greater than the dynamic range of the luminance of the second image
US9699426B2 (en) Method and device for estimating a color mapping between two different color-graded versions of a picture
US20180220190A1 (en) Methods and devices for encoding/decoding videos
US10972698B2 (en) Method and device for selecting a process to be applied on video data from a set of candidate processes driven by a common set of information data
WO2015193116A1 (fr) Procédé et dispositif pour décoder une image hdr à partir d&#39;un flux binaire représentant une image ldr et une image d&#39;éclairage
US10148958B2 (en) Method and device for encoding and decoding a HDR picture and a LDR picture
US20170118482A1 (en) Method and device for decoding a hdr picture from a bitstream representing a ldr picture and an illumination picture
WO2015193117A1 (fr) Procédé et dispositif de décodage d&#39;une image hdr à partir d&#39;un flux binaire représentant une image ldr et une image d&#39;éclairage
WO2015177123A1 (fr) Méthode et dispositif pour coder une trame et/ou décoder un train de bits représentant une trame
US20170188049A1 (en) Method and device for encoding a frame and/or decoding a bitstream representing a frame
EP2938083A1 (fr) Procédé et dispositif de traitement de données d&#39;informations indiquant qu&#39;une trame contient des échantillons d&#39;au moins deux trames constitutives conditionnées distinctes selon un schéma d&#39;arrangement de conditionnement de trame spécifique
EP3104610A1 (fr) Procede et dispositif de codage de differences de couleurs pour des donnees d&#39;image en couleur
WO2015177119A1 (fr) Procédé et dispositif pour coder une trame et/ou de décoder un train de bits représentant une trame
WO2015177126A1 (fr) Procédé et dispositif de codage d&#39;une trame et/ou de décodage d&#39;un flux binaire représentant une trame
WO2015177125A1 (fr) Procédé et dispositif pour coder une trame et/ou de décoder un train de bits représentant une trame

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15725081

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15725081

Country of ref document: EP

Kind code of ref document: A1