WO2015177139A1 - Method and device for encoding a high-dynamic range frame and/or decoding a bitstream - Google Patents
Method and device for encoding a high-dynamic range frame and/or decoding a bitstream Download PDFInfo
- Publication number
- WO2015177139A1 WO2015177139A1 PCT/EP2015/060983 EP2015060983W WO2015177139A1 WO 2015177139 A1 WO2015177139 A1 WO 2015177139A1 EP 2015060983 W EP2015060983 W EP 2015060983W WO 2015177139 A1 WO2015177139 A1 WO 2015177139A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- frame
- backlight
- decoded
- shape functions
- residual
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/12—Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/154—Measured or subjectively estimated visual quality after decoding, e.g. measurement of distortion
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/186—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/187—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/36—Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability
Definitions
- the present disclosure generally relates to frame/video encoding and decoding.
- the technical field of the present disclosure is related to encoding/decoding of a frame whose pixels values belong to a high- dynamic range.
- LDR frames Low-Dynamic-Range frames
- HDR frames high- dynamic range frames
- pixel values are usually represented in floating-point format (either 32-bit or 1 6-bit for each component, namely float or half-float), the most popular format being openEXR half-float format (16-bit per RGB component, i.e. 48 bits per pixel) or in integers with a long representation, typically at least 1 6 bits.
- a typical approach for encoding an HDR frame is to reduce the dynamic range of the frame in order to encode the frame by means of a legacy encoding scheme (initially configured to encode LDR frames).
- a backlight frame is determined from the luminance component of the input HDR frame.
- a residual frame is then obtained by dividing the input HDR frame by the backlight frame and both the backlight frame and the residual frame are encoded by a legacy encoder such as H.264/AVC (("Advanced video coding for generic audiovisual Services", SERIES H: AUDIOVISUAL AND MULTIMEDIA SYSTEMS, Recommendation ITU-T H.264, Telecommunication Standardization Sector of ITU, January 2012).) or HEVC ("High Efficiency Video Coding", SERIES H: AUDIOVISUAL AND MULTIMEDIA SYSTEMS, Recommendation ITU-T H.265, Telecommunication Standardization Sector of ITU, April 2013).
- the capability of providing automatically a viewable LDR frame from the encoded residual frame is a key advantage as it allows distributing the HDR frame (video) to customers equipped with standard LDR TV sets and reception device not dedicated to post-process both the backlight frame and the residual frame to decode an HDR frame (video).
- the method is characterized in that the processor is further configured in order that the backlight frame is represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
- the 2D shape functions are defined according to constraints of smoothness on the boundaries of their supports.
- the 2D shape functions of a set have the shape of a model symmetric function.
- the 2D shape functions are defined over a regular grid.
- the model symmetric function is the product of two separable 1 D polynomials.
- the backlight frame is represented by a weighted linear combination of several sets of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
- coefficients of the weighted linear combination and/or the 2D shape functions are encoded in a bitstream.
- the disclosure further relates to a method for decoding a frame from at least one bitstream representing a backlight frame obtained from the frame and a residual frame calculated by dividing the frame by the backlight frame.
- the decoding method comprising a processor configured for:
- the method is characterized in that the processor is further configured in order that the decoded backlight frame is represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
- the backlight frame is represented by a weighted linear combination of several sets of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
- the processor is further configured for obtaining coefficients of the weighted linear combination of 2D shape functions by a at least partially decoding of a bitstream and/or the 2D shape functions of a set of 2D shape functions by a at least partially decoding of a bitstream.
- the disclosure relates to an encoding and decoding device, a computer program product, a processor readable medium and non-transitory storage medium.
- FIG. 1 shows a block diagram of the steps of a method for encoding a frame I in accordance with an embodiment of the disclosure
- FIG. 2 shows an example of support for the 2D shape functions
- FIG. 3 show an example of the model symmetric function ⁇ over a triangular meshing
- - Fig. 4 shows an example of a representation using four sets of 2D shape functions defined over a rectangular grid
- - Fig. 5 shows a block diagram of a step of the method in accordance with an embodiment of the disclosure
- FIG. 6 shows a block diagram of a step of the method in accordance with an embodiment of the disclosure
- Fig. 7 shows a block diagram of a step of the method in accordance with an embodiment of the disclosure
- FIG. 8 shows a block diagram of the steps of a method, in accordance with an embodiment of the disclosure, for decoding a bitstream representing a residual frame calculated by dividing an frame by a backlight frame;
- FIG. 9 shows an example of an architecture of a device in accordance with an embodiment of the disclosure.
- FIG. 10 shows two remote devices communicating over a communication network in accordance with an embodiment of the disclosure.
- each block represents a circuit element, module, or portion of code which comprises one or more executable instructions for implementing the specified logical function(s).
- the function(s) noted in the blocks may occur out of the order noted. For example, two blocks shown in succession may, in fact, be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending on the functionality involved.
- the disclosure is described for encoding/decoding a frame but extends to the encoding/decoding of a sequence of frames (video) because each frame of the sequence is sequentially encoded/decoded as described below.
- Fig. 1 shows a block diagram of the steps of a method for encoding a frame I in accordance with an embodiment of the disclosure.
- a module IC obtains the luminance component L and potentially at least one color component C(i) of the frame I to be encoded.
- the luminance component L is obtained, for instance in the 709 gamut, by a linear combination which is given by:
- a module BAM determines a backlight frame Bal from the luminance component L of the frame I.
- a module Bl determines a backlight frame Ba as being represented by a weighted linear combination of at least one set of 2D shape functions.
- the backlight frame Ba is given by
- each set j form a partition of unity of a domain Do defined by the union of their supports, i.e, by definition, the 2D shape functions ⁇ have compact supports and their sum over the domain Do equals 1 :
- the 2D shape functions are defined over a regular grid.
- a rectangular grid is considered by using multiplicative stretching in the two directions of the square grid.
- the 2D shape functions ⁇ ( are obtained as follows.
- each of these four 2D shape functions has its support contained in the four grid elements around its centers.
- the domain Do is thus defined by the union of the four grid elements around the center of each 2D shape functions.
- the partition of unity condition on the domain Do is then given by:
- the 2D shape functions of a set j are the translated functions of a model symmetric function ⁇ , i.e.
- the 2D shape functions xfjj are defined according to constraints of smoothness on the boundaries of their supports.
- the constraints on the boundaries of the 2D shape functions are:
- Fig. 3 show an example of the model symmetric function ⁇ when the regular grid is a triangular meshing.
- each of the three 2D shape functions has its support contained in the six triangles around its center (Fig. 3).
- the domain DO is thus defined here by the union of the six triangles around the center of each 2D shape function.
- the 2D shape functions have the shape of a model symmetric function ⁇ whose the support is a single triangle.
- the model symmetric function ⁇ whose the support is a triangle is defined to constraints on its boundaries.
- the constraints on the boundaries of the model symmetric function ⁇ are:
- the backlight frame is represented by a weighted linear combination of several sets of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
- Fig. 4 shows an example of a representation using four sets of 2D shape functions defined over a rectangular grid.
- the black circles represent the centers of the 2D shape functions of a first set, the triangles the centers of a second one, the ovals the centers of a third one and the crosses the centers of a forth one.
- the sum of all the 2D shape functions of all the sets is constant everywhere on the grid, with value 4 here. So one gets a partition of unity, up to a multiplicative factor.
- the key point here is that the 2D shape functions overlap much more than on the example using only a single set of 2D shape functions as described in relation with Fig. 2 and 3.
- By using several set of 2D shape functions it is possible to get even more overlapping and to use 2D shape functions with closer centers, i.e. a higher density of shape functions to follow the HDR signal smoothly while remaining precise.
- the disclosure is not limited to a specific number of sets of 2D shape functions nor to the way to combine them together on a mesh nor the kind of mesh used to locate the centers of the 2D shape functions.
- the backlight frame Bal output from step 1 1 , is the backlight frame Ba given by equation (1 ).
- a module BM modulates the backlight frame Ba (given by equation (1 )) with a mean luminance value L mean of the frame I obtained by the means of a module HL.
- the backlight frame Bal output from step 1 1 , is the modulated backlight frame.
- the module HL is configured to calculate the mean luminance value L mean over the whole luminance component L.
- the module HL is configured to calculate the mean luminance value L mean by
- This last embodiment is advantageous because it avoids that the mean luminance value L mean be influenced by a few pixels with extreme high values which usually leads to very annoying temporal mean brightness instability when the frame I belongs to a sequence of frames.
- the disclosure is not limited to a specific embodiment for calculating the mean luminance value L mean .
- a module N normalizes the backlight frame Ba (given by equation (1 )) by its mean value E(Ba) such that one gets a mid-gray-at-one backlight frame Ba gray for the frame (or for all frames if the frame I belongs to a sequence of frames):
- the module BM is configured to modulate the mid-gray-at-one backlight frame Ba gray with the mean luminance value L mean of the frame L, by using the following relation
- cst mod being a modulation coefficient and a being another modulation coefficient less than 1 , typically 1 /3.
- the backlight frame Bal output from step 1 1 , is the modulated backlight frame Ba mod given by equation (4).
- the modulation coefficient cst mod is tuned to get a good looking brightness for the residual frame and highly depends on the process to obtain the backlight frame. For example, cst m0 d s 1 .7 for a backlight frame obtained by least means squares.
- step 12 the data needed to determine the backlight frame Bal, output from step 1 1 , are encoded by means of an encoder ENC1 and added in a bitstream BF which may be stored on a local or remote memory and/or transmitted through a communication interface (e.g. to a bus or over a communication network or a broadcast network).
- a communication interface e.g. to a bus or over a communication network or a broadcast network.
- the data to be encoded are limited to the weighting coefficients a ⁇ or a ⁇ when known non-adaptive shape functions are used, but the 2D shape functions ⁇ ( may also be a priori unknown and then encoded in the bitstream BF, for instance in a case of a somewhat optimal mathematical construction for better fitting. So, all the weighting coefficients a[ or a ⁇ (and potentially shape functions ⁇ () are encoded in the bitstream BF.
- the weighting coefficients a[ or a ⁇ are quantized before encoded in order to reduce the size of the bitstream BF.
- step 1 3 a residual frame Res is calculated by dividing the frame by a decoded version Ba of the backlight frame.
- the luminance component L and potentially each colour component C(i) of the frame I, obtained from the module IC, is divided by the decoded version Ba of the backlight frame. This division is done pixel per pixel.
- the decoded version Ba of the backlight frame is obtained by decoding at least partially the bitstream BF by means of a decoder DEC1 .
- a decoder DEC1 some data needed to obtain the backlight frame, output of step 1 1 , have been encoded (step 12) and then obtained by at least partially decoding the bitstream BF.
- weighting coefficients a ⁇ (and potentially shape functions ⁇ () are then obtained as output of step 14.
- a module BAG generates a decoded version Ba of the backlight frame from the weighting coefficients a ⁇ and either some known non-adaptive shape functions or the shape functions ⁇ ( by:
- a module TMO tone-maps the residual frame Res in order to get a viewable residual frame Res v .
- the residual frame Res may not be viewable because its dynamic range is too high and because a decoded version of this residual frame Res shows too visible artifacts. Tone-mapping the residual frame remedies to at least one of these drawbacks.
- the disclosure is not limited to any specific tone-mapping operator.
- tone-mapping operator shall be reversible.
- the tone-mapping operator defined by Reinhard may be used ⁇ Reinhard, E., Stark, M., Shirley, P., and Ferwerda, J., ⁇ Photographic tone reproduction for digital frames, " ACM Transactions on Graphics 21 (July 2002)), or Boitard, R., Bouatouch, K., Cozot, R., Thoreau, D., & Gruson, A. (2012). Temporal coherency for video tone mapping. In A. M. J. van Eijk, C. C. Davis, S. M. Hammel, & A. K. Majumdar (Eds.), Proc. SPIE 8499, Applications of Digital Frame Processing (p. 84990 D-84990D-10)).
- step 19 the viewable residual frame Res v is encoded by means of an encoder ENC2 in a bitstream F which may be stored on a local or remote memory and/or transmitted through a communication interface (e.g. on a bus or over a communication network or a broadcast network).
- tone mapping the residual frame comprises either a gamma correction or a SLog correction according to the pixel values of the residual frame.
- the viewable residual frame Res v is then given, for example, by:
- ⁇ being a coefficient of a gamma curve equal, for example, to 1 /2.4.
- the viewable residual frame Res v is given, for example, by:
- a,b and c are coefficients of a SLog curve determined such that 0 and 1 are invariant, and the derivative of the SLog curve is continuous in 1 when prolonged by a gamma curve below 1 .
- a,b and c are functions of the parameter y.
- the parameter y of the gamma-Slog curve is encoded in the bitstream BF.
- TMO applies either the gamma correction or the SLog correction according to the pixel values of the residual frame Res.
- the gamma correction is applied and otherwise the SLog correction is applied.
- the viewable residual frame Res v usually has a mean value more or less close to 1 depending on the brightness of the frame I, making the use of the above gamma-Slog combination particularly efficient.
- a module SCA scales the viewable residual frame Res v before encoding (step 19) by multiplying each component of the viewable residual frame Res v by a scaling factor cstscaiing- The resulting residual frame Res s is then given by
- the scaling factor cst sca iing is defined to map the values of the viewable residual frame Res v between from 0 to the maximum value 2 N -1 , where N is the number of bits allowed as input for the coding by the encoder ENC2.
- step 19 the residual frame Res s is encoded by means of an encoder ENC2.
- a module CLI clips the viewable residual frame Res v before encoding to limit its dynamic range to a targeted dynamic range TDR which is defined, for example, according to the capabilities of the encoder ENC2.
- the resulting residual frame Res c is given, for example, by:
- the disclosure is not limited to such clipping (max(.)) but extends to any kind of clipping.
- step 19 the residual frame Res c is encoded by means of an encoder ENC2.
- Ressc max(2 N , cst S caiing * Res v )
- Ressc max(2 N , cst SC aiing * Res s ) according to the embodiments of the method.
- the residual frame Res sc is encoded by means of an encoder ENC2.
- the tone-mapping and scaling of the viewable residual frame Res v is a parametric process.
- the parameters may be fixed or not and in the latter case they may be encoded in the bitstream BF by means of the encoder ENC1 .
- the constant value ⁇ of the gamma correction, the scaling factor cst sca iing may be parameters which are encoded in the bitstream BF.
- Fig. 8 shows a block diagram of the steps of a method, in accordance with an embodiment of the disclosure, for decoding a bitstream representing a residual frame calculated by dividing an frame by a backlight frame.
- a backlight frame B a ⁇ is obtained for example by at least partially decoding a bitstream BF by means of the decoder DEC1 .
- the bitstream BF may have been stored locally or received from a communication network.
- a decoded residual frame Res ' is obtained by a at least partial decoding of a bitstream F by means of a decoder DEC2.
- the bitstream F may have been stored locally or received from a communication network.
- the decoded residual frame Res ' is viewable by a traditional apparatus.
- a decoded frame / is obtained by multiplying the decoded residual frame Res ⁇ by the backlight frame B ⁇ .
- the parameters ⁇ and/or are a'so obtained either from a local memory or by a at least partial decoding of the bitstream BF by means of the decoder DEC1 .
- a module ISCA applied an inverse scaling to the decoded residual frame Res ⁇ by dividing the decoded residual frame Res ⁇ by the parameter cst ⁇ ing .
- step 83 a module ITMO applied an inverse-tone-mapping to the decoded residual frame Res ' , by means of the parameters ⁇ .
- the parameter ⁇ defines a gamma curve and the inverse-tone-mapping is just to find, from the gamma curve, the values which correspond to the pixel values of the decoded residual frame Res ⁇ .
- the decoders DEC1 is configured to decode data which have been encoded by the encoder ENC1 , respectively DEC2.
- the encoders ENC1 and ENC2 are not limited to a specific encoder (decoder) but when an entropy encoder (decoder) is required, an entropy encoder such as a Huffmann coder, an arithmetic coder or a context adaptive coder like Cabac used in h264/AVC or HEVC is advantageous.
- the encoders ENC1 and ENC2 are not limited to a specific encoder which may be, for example, an frame/video coder with loss like JPEG, JPEG2000, MPEG2, h264/AVC or HEVC.
- the modules are functional units, which may or not be in relation with distinguishable physical units. For example, these modules or some of them may be brought together in a unique component or circuit, or contribute to functionalities of a software. A contrario, some modules may potentially be composed of separate physical entities.
- the apparatus which are compatible with the disclosure are implemented using either pure hardware, for example using dedicated hardware such ASIC or FPGA or VLSI, respectively « Application Specific Integrated Circuit » « Field- Programmable Gate Array » « Very Large Scale Integration » or from several integrated electronic components embedded in a device or from a blend of hardware and software components.
- Fig. 9 represents an exemplary architecture of a device 90 which may be configured to implement a method described in relation with Fig. 1-8.
- Device 90 comprises following elements that are linked together by a data and address bus 91 :
- microprocessor 92 which is, for example, a DSP (or Digital Signal Processor);
- RAM or Random Access Memory
- the battery 96 is external to the device.
- the word « register » used in the specification can correspond to area of small capacity (some bits) or to very large area (e.g. a whole program or large amount of received or decoded data).
- ROM 93 comprises at least a program and parameters. Algorithm of the methods according to the disclosure is stored in the ROM 93. When switched on, the CPU 92 uploads the program in the RAM and executes the corresponding instructions.
- RAM 94 comprises, in a register, the program executed by the CPU 92 and uploaded after switch on of the device 90, input data in a register, intermediate data in different states of the method in a register, and other variables used for the execution of the method in a register.
- the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method or a device), the implementation of features discussed may also be implemented in other forms (for example a program).
- An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
- the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
- PDAs portable/personal digital assistants
- the frame I is obtained from a source.
- the source belongs to a set comprising:
- a local memory e.g. a video memory or a RAM (or
- Random Access Memory Random Access Memory
- flash memory a ROM (or Read Only Memory), a hard disk ;
- a storage interface (95), e.g. an interface with a mass storage, a RAM, a flash memory, a ROM, an optical disc or a magnetic support;
- a communication interface e.g. a wireline interface (for example a bus interface, a wide area network interface, a local area network interface) or a wireless interface (such as a IEEE 802.1 1 interface or a Bluetooth® interface); and
- a wireline interface for example a bus interface, a wide area network interface, a local area network interface
- a wireless interface such as a IEEE 802.1 1 interface or a Bluetooth® interface
- an frame capturing circuit e.g. a sensor such as, for example, a
- CCD Charge-Coupled Device
- CMOS Complementary Metal-Oxide-Semiconductor
- the decoded frame ⁇ is sent to a destination; specifically, the destination belongs to a set comprising:
- a local memory e.g. a video memory or a RAM, a flash memory, a hard disk ;
- a storage interface (95), e.g. an interface with a mass storage, a RAM, a flash memory, a ROM, an optical disc or a magnetic support;
- a communication interface (95), e.g. a wireline interface (for example a bus interface (e.g. USB (or Universal Serial Bus)), a wide area network interface, a local area network interface, a HDMI (High Definition Multimedia Interface) interface) or a wireless interface (such as a IEEE 802.1 1 interface, WiFi ® or a Bluetooth ® interface); and
- a wireline interface for example a bus interface (e.g. USB (or Universal Serial Bus)
- a wide area network interface e.g. USB (or Universal Serial Bus)
- a wide area network interface e.g. USB (or Universal Serial Bus)
- a local area network interface e.g. HDMI (High Definition Multimedia Interface) interface
- a wireless interface such as a IEEE 802.1 1 interface, WiFi ® or a Bluetooth ® interface
- bitstream BF and/or F are sent to a destination.
- bitstream F and BF or both bitstreams F and BF are stored in a local or remote memory, e.g. a video memory (94) or a RAM (94), a hard disk (93).
- a storage interface e.g. an interface with a mass storage, a flash memory, ROM, an optical disc or a magnetic support and/or transmitted over a communication interface (95), e.g. an interface to a point to point link, a communication bus, a point to multipoint link or a broadcast network.
- the bitstream BF and/or F is obtained from a source.
- the bitstream is read from a local memory, e.g. a video memory (94), a RAM (94), a ROM (93), a flash memory (93) or a hard disk (93).
- the bitstream is received from a storage interface (95), e.g. an interface with a mass storage, a RAM, a ROM, a flash memory, an optical disc or a magnetic support and/or received from a communication interface (95), e.g. an interface to a point to point link, a bus, a point to multipoint link or a broadcast network.
- device 90 being configured to implement an encoding method described in relation with Fig. 1-7, belongs to a set comprising:
- a video server e.g. a broadcast server, a video-on-demand server or a web server.
- device 90 being configured to implement a decoding method described in relation with Fig. 8, belongs to a set comprising:
- the device A comprises means which are configured to implement a method for encoding an frame as described in relation with the Fig. 1 and the device B comprises means which are configured to implement a method for decoding as described in relation with Fig. 8.
- the network is a broadcast network, adapted to broadcast still frames or video frames from device A to decoding devices including the device B.
- Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications, particularly, for example, equipment or applications.
- equipment examples include an encoder, a decoder, a post-processor processing output from a decoder, a pre-processor providing input to an encoder, a video coder, a video decoder, a video codec, a web server, a set-top box, a laptop, a personal computer, a cell phone, a PDA, and other communication devices.
- the equipment may be mobile and even installed in a mobile vehicle.
- the methods may be implemented by instructions being performed by a processor, and such instructions (and/or data values produced by an implementation) may be stored on a processor-readable medium such as, for example, an integrated circuit, a software carrier or other storage device such as, for example, a hard disk, a compact diskette (“CD"), an optical disc (such as, for example, a DVD, often referred to as a digital versatile disc or a digital video disc), a random access memory (“RAM”), or a read-only memory (“ROM”).
- the instructions may form an application program tangibly embodied on a processor-readable medium. Instructions may be, for example, in hardware, firmware, software, or a combination.
- a processor may be characterized, therefore, as, for example, both a device configured to carry out a process and a device that includes a processor-readable medium (such as a storage device) having instructions for carrying out a process. Further, a processor-readable medium may store, in addition to or in lieu of instructions, data values produced by an implementation.
- implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted.
- the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
- a signal may be formatted to carry as data the rules for writing or reading the syntax of a described embodiment, or to carry as data the actual syntax-values written by a described embodiment.
- Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
- the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
- the information that the signal carries may be, for example, analog or digital information.
- the signal may be transmitted over a variety of different wired or wireless links, as is known.
- the signal may be stored on a processor-readable medium.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Discrete Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
The present disclosure generally relates to a method and device for encoding a frame, comprising a processor configured for: encoding (12) a backlight frame determined (11) from the frame to be encoded; obtaining (13) at least one component of a residual frame by dividing each component of the frame by a decoded version of the backlight frame; encoding (19) the residual frame; wherein the processor is further configured in order that the backlight frame is represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports, and wherein coefficients of the weighted linear combination and/or the 2D shape functionsare encoded in a bitstream. The disclosure relates also to a method and device for decoding a bitstream representing a residual frame calculated by dividing a frame by a backlight frame.
Description
METHOD AND DEVICE FOR ENCODING A HIGH-DYNAMIC RANGE FRAME AND/OR DECODING A BITSTREAM.
1. Field of disclosure.
The present disclosure generally relates to frame/video encoding and decoding. In particular, the technical field of the present disclosure is related to encoding/decoding of a frame whose pixels values belong to a high- dynamic range.
2. Technical background.
The present section is intended to introduce the reader to various aspects of art, which may be related to various aspects of the present disclosure that are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present disclosure. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art.
Low-Dynamic-Range frames (LDR frames) are frames whose luminance values are represented with a limited number of bits (most often 8 or 10). This limited representation does not allow correct rendering of small signal variations, in particular in dark and bright luminance ranges. In high- dynamic range frames (HDR frames), the signal representation is extended in order to maintain a high accuracy of the signal over its entire range. In HDR frames, pixel values are usually represented in floating-point format (either 32-bit or 1 6-bit for each component, namely float or half-float), the most popular format being openEXR half-float format (16-bit per RGB component, i.e. 48 bits per pixel) or in integers with a long representation, typically at least 1 6 bits.
A typical approach for encoding an HDR frame is to reduce the dynamic range of the frame in order to encode the frame by means of a legacy encoding scheme (initially configured to encode LDR frames).
According to a well-known approach, a backlight frame is determined from the luminance component of the input HDR frame. A residual frame is then obtained by dividing the input HDR frame by the backlight frame and both the backlight frame and the residual frame are encoded by a legacy encoder such as H.264/AVC (("Advanced video coding for generic audiovisual Services", SERIES H: AUDIOVISUAL AND MULTIMEDIA SYSTEMS, Recommendation ITU-T H.264, Telecommunication Standardization Sector of ITU, January 2012).) or HEVC ("High Efficiency Video Coding", SERIES H: AUDIOVISUAL AND MULTIMEDIA SYSTEMS, Recommendation ITU-T H.265, Telecommunication Standardization Sector of ITU, April 2013).
The capability of compression depends strongly on the temporal correlation between frames and the spatial smoothness of each frame. Actually, video encoders use temporal and intra-frame prediction as a very efficient tool to compress video data.
The capability of providing automatically a viewable LDR frame from the encoded residual frame is a key advantage as it allows distributing the HDR frame (video) to customers equipped with standard LDR TV sets and reception device not dedicated to post-process both the backlight frame and the residual frame to decode an HDR frame (video).
Unfortunately, if one is not careful with the backlight frame building at the encoder side, this backlight frame may lack uniformity and this non uniformity creates some smooth gradient in the residual frame (LDR viewable frame). The human visual system is very sensitive to such artifacts. Thus, a dedicated backlight frame building capable of providing a uniform backlight frame is wishful.
3. Summary of the disclosure.
The disclosure sets out to remedy some of the drawbacks of the prior art with a method for encoding a frame comprising a processor configured for:
- encoding a backlight frame determined from the frame to be encoded;
- obtaining at least one component of a residual frame by dividing each component of the frame by a decoded version of the backlight frame;
- encoding the residual frame;
The method is characterized in that the processor is further configured in order that the backlight frame is represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
According to an embodiment, the 2D shape functions are defined according to constraints of smoothness on the boundaries of their supports.
According to an embodiment, the 2D shape functions of a set have the shape of a model symmetric function.
According to an embodiment, the 2D shape functions are defined over a regular grid.
According to an embodiment, the model symmetric function is the product of two separable 1 D polynomials.
According to an embodiment, the backlight frame is represented by a weighted linear combination of several sets of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
According to an embodiment, coefficients of the weighted linear combination and/or the 2D shape functions are encoded in a bitstream.
The disclosure further relates to a method for decoding a frame from at least one bitstream representing a backlight frame obtained from the frame
and a residual frame calculated by dividing the frame by the backlight frame. The decoding method comprising a processor configured for:
- obtaining a decoded backlight frame by a at least partially decoding of a bitstream;
- obtaining a decoded residual frame by a at least partially decoding of a bitstream; and
- obtaining a decoded frame by multiplying the decoded residual frame by the decoded backlight frame;
the method is characterized in that the processor is further configured in order that the decoded backlight frame is represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
According to an embodiment, the backlight frame is represented by a weighted linear combination of several sets of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
According to an embodiment, the processor is further configured for obtaining coefficients of the weighted linear combination of 2D shape functions by a at least partially decoding of a bitstream and/or the 2D shape functions of a set of 2D shape functions by a at least partially decoding of a bitstream.
According to other of its aspects, the disclosure relates to an encoding and decoding device, a computer program product, a processor readable medium and non-transitory storage medium.
The specific nature of the disclosure as well as other objects, advantages, features and uses of the disclosure will become evident from the following description of a preferred embodiment taken in conjunction with the accompanying drawings.
4. List of figures.
The embodiments will be described with reference to the following figures:
- Fig. 1 shows a block diagram of the steps of a method for encoding a frame I in accordance with an embodiment of the disclosure;
- Fig. 2 shows an example of support for the 2D shape functions;
- Fig. 3 show an example of the model symmetric function φ over a triangular meshing;
- Fig. 4 shows an example of a representation using four sets of 2D shape functions defined over a rectangular grid;- Fig. 5 shows a block diagram of a step of the method in accordance with an embodiment of the disclosure;
- Fig. 6 shows a block diagram of a step of the method in accordance with an embodiment of the disclosure;
- Fig. 7 shows a block diagram of a step of the method in accordance with an embodiment of the disclosure;
- Fig. 8 shows a block diagram of the steps of a method, in accordance with an embodiment of the disclosure, for decoding a bitstream representing a residual frame calculated by dividing an frame by a backlight frame;
- Fig. 9 shows an example of an architecture of a device in accordance with an embodiment of the disclosure; and
- Fig. 10 shows two remote devices communicating over a communication network in accordance with an embodiment of the disclosure.
5. Detailed description of preferred embodiments of the disclosure. The present disclosure will be described more fully hereinafter with reference to the accompanying figures, in which embodiments of the disclosure are shown. This disclosure may, however, be embodied in many
alternate forms and should not be construed as limited to the embodiments set forth herein. Accordingly, while the disclosure is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that there is no intent to limit the disclosure to the particular forms disclosed, but on the contrary, the disclosure is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the disclosure as defined by the claims. Like numbers refer to like elements throughout the description of the figures.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the disclosure. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises", "comprising," "includes" and/or "including" when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. Moreover, when an element is referred to as being "responsive" or "connected" to another element, it can be directly responsive or connected to the other element, or intervening elements may be present. In contrast, when an element is referred to as being "directly responsive" or "directly connected" to other element, there are no intervening elements present. As used herein the term "and/or" includes any and all combinations of one or more of the associated listed items and may be abbreviated as"/".
It will be understood that, although the terms first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first element could be termed a second element, and, similarly, a second element could be termed a first element without departing from the teachings of the disclosure.
Although some of the diagrams include arrows on communication paths to show a primary direction of communication, it is to be understood that communication may occur in the opposite direction to the depicted arrows.
Some embodiments are described with regard to block diagrams and operational flowcharts in which each block represents a circuit element, module, or portion of code which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in other implementations, the function(s) noted in the blocks may occur out of the order noted. For example, two blocks shown in succession may, in fact, be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending on the functionality involved.
Reference herein to "one embodiment" or "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one implementation of the disclosure. The appearances of the phrase "in one embodiment" or "according to an embodiment" in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments necessarily mutually exclusive of other embodiments.
Reference numerals appearing in the claims are by way of illustration only and shall have no limiting effect on the scope of the claims.
While not explicitly described, the present embodiments and variants may be employed in any combination or sub-combination.
The disclosure is described for encoding/decoding a frame but extends to the encoding/decoding of a sequence of frames (video) because each frame of the sequence is sequentially encoded/decoded as described below.
Fig. 1 shows a block diagram of the steps of a method for encoding a frame I in accordance with an embodiment of the disclosure.
In step 10, a module IC obtains the luminance component L and potentially at least one color component C(i) of the frame I to be encoded.
For example, when the frame I belongs to the color space (Χ,Υ,Ζ), the luminance component L is obtained by a transform f(.) of the component Y, e.g. L=f(Y).
When the frame I belongs to the color space (R,G,B), the luminance component L is obtained, for instance in the 709 gamut, by a linear combination which is given by:
L=0.2127.R+0.7152.G+0.0722.B
In step 1 1 , a module BAM determines a backlight frame Bal from the luminance component L of the frame I.
According to an embodiment of the step 1 1 , illustrated in Fig. 5, a module Bl determines a backlight frame Ba as being represented by a weighted linear combination of at least one set of 2D shape functions.
Mathematically speaking, the backlight frame Ba is given by
Ba =∑j∑i aix l (1 )
with a{ being weighting coefficients, the 2D shape functions of a set j of 2D shape functions.
The 2D shape functions of each set j form a partition of unity of a domain Do defined by the union of their supports, i.e, by definition, the 2D shape functions ψ{ have compact supports and their sum over the domain Do equals 1 :
∑i = i
According to an embodiment of the step 1 1 , the 2D shape functions are defined over a regular grid.
As a variant of this embodiment, a rectangular grid is considered by using multiplicative stretching in the two directions of the square grid.
According to an embodiment of the step 1 1 , illustrated in Fig. 2, the 2D shape functions ψ( are obtained as follows.
Assuming the centers (0,0), (0,1 ), (1 ,0) and (1 ,1 ) of the 2D shape functions ψ[, ψ[, ψ{ and ψ{ respectively, are on a regular grid (here
rectangular grid), each of these four 2D shape functions has its support contained in the four grid elements around its centers. The domain Do is thus defined by the union of the four grid elements around the center of each 2D shape functions. The partition of unity condition on the domain Do is then given by:
xp1 + xp2 + xp3 + xp4 = 1
According to an embodiment of the step 1 1 , the 2D shape functions of a set j are the translated functions of a model symmetric function φ, i.e.
0(x,y) = φ +χ, +y). In this case, the partition of unity condition on the domain Do is then given by:
0(x, y) + φ(1 - x, y) + φ{χ, 1 - y) + 0(1 - x, 1 - y) = 1 on any of one grid element E as depicted on Fig. 2.
According to an embodiment of the step 1 1 , the model symmetric function φ is the product of two separable 1 D (one dimensional) polynomials P(x) and P(y), i.e. φ(χ,γ) = P(x)P(y).
When P fulfills the partition of unity condition, i.e. P(x) + P(1 -x) = 1 , the partition of unity condition of the 2D shape function is fulfilled because one gets
P(x)P(y) + P(l - x)P(y) + P(x)P(l - y) + P(l - x)P l - y)
= (P(x) + P(l - x))(P(y) + P(l - y)) = 1
According to an embodiment of the step 1 1 , the 2D shape functions xfjj are defined according to constraints of smoothness on the boundaries of their supports.
This allows continuity of the 2D shape functions over the frame domain and smoothness of the global representation of the backlight frame Ba, because a finite sum of smooth functions is smooth.
According to an embodiment of the step 1 1 , the constraints on the boundaries of the 2D shape functions are:
- a boundary constraint P(0) = 1 , which means that only one 2D shape function contributes at its center;
- a constraint relative to the continuity and the continuity of the derivative at the boundary of the support of the 2D shape function : P(1 ) = P'(1 ) = 0;
- a constraint of the derivative at the center: P'(0) = 0 to ensure smoothness around the center.
One may look for analytical solutions for the 1 D polynomial P with the above four boundary constraints.
For example, a solution for the 1 D function P is a polynomial of degree three given by P(x) = θχ3 + ϋχ2 + κχ + λ. The above constraints lead to the following conditions on the parameters
- I = P(O) = A
- Ο = Ρ(1 ) = 0 + ι9 + κ + Λ
- 0 = P'(0) = K
One easily solves this linear system of equations to obtain Θ = 2, ϋ =
-3, K = Ο, λ = 1, and the model symmetric function φ(χ,γ) is then given by:
<p(x,y) = (2x3-3x2+ 1) (2y3-3y2+ 1).
Other polynomial examples are obtained by relaxing some smoothness. For instance, by requiring the continuity only, one notices that P(x) = 1 -x is a continuous solutions leading to a partition of unity with the following model symmetric function
<t>(x,y) = (1-x) (1-y).
whose derivative is not continuous.
Also, one may look for higher order parametric polynomial. For instance, the four conditions above when applied to a forth order polynomial P{x) = μχ3 + θχ3 + ϋχ2 + κχ + λ lead to an under-constrained linear system whose solutions are parameterized by one parameter a, thus leading to an infinity of polynomial solutions Pa.
One way to decide which solution is the best is to choose the parameter a such that the Laplace operator is minimized, i.e.
P = argminpJlAPjl.
Finally, non-analytical purely numerical solutions can be obtained from the resolutio
where P is understood not necessarily being a polynomial but any smooth enough function, for instance of class C1. The numerical resolution of the minimization problem will be detailed later in the two-dimensional non- symmetric case.
Fig. 3 show an example of the model symmetric function φ when the regular grid is a triangular meshing.
Assuming the centers pt1 =(0,0), pt2=(-1 /3,sqrt(3)/2) and pt3=(1 /3,sqrt(3)/2) of the 2D shape functions Ι/Ί, Ι/>2 and ψ3 respectively, on the triangular meshing, each of the three 2D shape functions has its support contained in the six triangles around its center (Fig. 3). The domain DO is thus defined here by the union of the six triangles around the center of each 2D shape function. The partition of unity condition on Do is then given by:
According to an embodiment of the step 1 1 , the 2D shape functions have the shape of a model symmetric function φ whose the support is a single triangle.
The representation on a support of a 2D shape function (here six triangles around its center) of the backlight frame Ba which respects the partition of unity condition is then obtained when the model symmetric function φ is defined on a single triangle.
According to an embodiment of the step 1 1 , the model symmetric function φ whose the support is a triangle is defined to constraints on its boundaries.
This allows continuity of the 2D shape functions over the frame domain and smoothness of the global representation of the backlight frame Ba.
According to an embodiment of the step 1 1 , the constraints on the boundaries of the model symmetric function φ are:
Finally, the model symmetric function φ is determined by solving the minimization problem
minimize ||Δφ|| such that conditions (2)and (3) are fulfilled in order to maximize the smoothness of the solution.
Numerical solutions are standardly found by discretization of the domain Do and then minimization under constrains on the boundary and partition of unity. One may refer to the book «lntroduction a I'analyse numerique des equations aux derivees partielles» by Raviart & Thomas for the introduction to the involved numerical tools.
In brief, one generates a mesh on the domain Do and the model symmetric function φ takes one discrete value on each mesh. These discrete values can be summed up into an unknown vector x. The Laplace operator is itself discretized under a matrix M such that the operator Δφ become Mx in its discrete form. The conditions (2) and (3) are put a matrix system Cx = 0. Thus, the minimization problem becomes
minimize ||Mx|| such that Cx = 0
This is a classical linear least square minimization problem under constraint. There exist many solver to find the solution x leading to a discrete approximation of the wanted solution φ. See Gill, P.E., W. Murray, and M.H. Wright, Practical Optimization, Academic Press, London, UK, 1981 .
According to an embodiment of the step 1 1 , the backlight frame is represented by a weighted linear combination of several sets of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports.
Fig. 4 shows an example of a representation using four sets of 2D shape functions defined over a rectangular grid. The black circles represent the centers of the 2D shape functions of a first set, the triangles the centers of a second one, the ovals the centers of a third one and the crosses the centers of a forth one.
The sum of all the 2D shape functions of all the sets is constant everywhere on the grid, with value 4 here. So one gets a partition of unity, up to a multiplicative factor. The key point here is that the 2D shape functions overlap much more than on the example using only a single set of 2D shape functions as described in relation with Fig. 2 and 3. By using several set of 2D shape functions, it is possible to get even more overlapping and to use 2D shape functions with closer centers, i.e. a higher density of shape functions to follow the HDR signal smoothly while remaining precise.
The disclosure is not limited to a specific number of sets of 2D shape functions nor to the way to combine them together on a mesh nor the kind of mesh used to locate the centers of the 2D shape functions.
According to one of the embodiments or variants of the step 1 1 , the backlight frame Bal, output from step 1 1 , is the backlight frame Ba given by equation (1 ).
According to an embodiment of the step 1 1 , illustrated in Fig. 6, a module BM modulates the backlight frame Ba (given by equation (1 )) with a mean luminance value Lmean of the frame I obtained by the means of a module HL.
According to this embodiment, the backlight frame Bal, output from step 1 1 , is the modulated backlight frame.
According to an embodiment, the module HL is configured to calculate the mean luminance value Lmean over the whole luminance component L.
According to an embodiment, the module HL is configured to calculate the mean luminance value Lmean by
Lmean = E(^^)^
with β being a coefficient less than 1 and E(X) the mathematical expectation value (mean) of the luminance component L.
This last embodiment is advantageous because it avoids that the mean luminance value Lmean be influenced by a few pixels with extreme high values which usually leads to very annoying temporal mean brightness instability when the frame I belongs to a sequence of frames.
The disclosure is not limited to a specific embodiment for calculating the mean luminance value Lmean.
According to a variant of this embodiment, illustrated in Fig. 7, a module N normalizes the backlight frame Ba (given by equation (1 )) by its mean value E(Ba) such that one gets a mid-gray-at-one backlight frame Bagray for the frame (or for all frames if the frame I belongs to a sequence of frames):
Ba
Bagray = I{Ba)
Then, the module BM is configured to modulate the mid-gray-at-one backlight frame Bagray with the mean luminance value Lmean of the frame L, by using the following relation
Bo-mod ~ cstmod. Lmean . Bagray (4)
with cstmod being a modulation coefficient and a being another modulation coefficient less than 1 , typically 1 /3.
According to this variant, the backlight frame Bal, output from step 1 1 , is the modulated backlight frame Bamod given by equation (4).
It may be noted that the modulation coefficient cstmod is tuned to get a good looking brightness for the residual frame and highly depends on the process to obtain the backlight frame. For example, cstm0d s 1 .7 for a backlight frame obtained by least means squares.
Practically, by linearity, all operations to modulate the backlight frame apply to the backlight coefficients a[ as a correcting factor which transforms the coefficients a{ into new coefficients a{ such that one gets
Bamod = ^ ^ a¾
j i
In step 12 (Fig. 1 ), the data needed to determine the backlight frame Bal, output from step 1 1 , are encoded by means of an encoder ENC1 and added in a bitstream BF which may be stored on a local or remote memory and/or transmitted through a communication interface (e.g. to a bus or over a communication network or a broadcast network).
For example, the data to be encoded are limited to the weighting coefficients a{ or a{ when known non-adaptive shape functions are used, but the 2D shape functions ψ( may also be a priori unknown and then encoded in the bitstream BF, for instance in a case of a somewhat optimal mathematical construction for better fitting. So, all the weighting coefficients a[ or a{ (and potentially shape functions ψ() are encoded in the bitstream BF.
Advantageously, the weighting coefficients a[ or a{ are quantized before encoded in order to reduce the size of the bitstream BF.
In step 1 3, a residual frame Res is calculated by dividing the frame by a decoded version Ba of the backlight frame.
It is advantageous to use the decoded version Ba of the backlight frame to ensure a same backlight frame on both encoder and decoder side, thus leading to a better precision of a final decoded frame /.
More precisely, the luminance component L and potentially each colour component C(i) of the frame I, obtained from the module IC, is divided by the decoded version Ba of the backlight frame. This division is done pixel per pixel.
For example, when the components R, G or B of the frame I are expressed in the color space (R,G,B), the component RRes, GRes and BRes are obtained as follows:
Rres= R/Ba, Gres= G/Ba, Bres= B/Ba,
For example, when the components X, Y or Z of the frame I are expressed in the color space (Υ,Υ,Ζ), the component XRes, YRes and ZRes are obtained as follows:
Xres= XI Ba Yres= Y/Ba Zres= Z/Ba
According to an embodiment, in step 14, the decoded version Ba of the backlight frame is obtained by decoding at least partially the bitstream BF by means of a decoder DEC1 .
As explained before, some data needed to obtain the backlight frame, output of step 1 1 , have been encoded (step 12) and then obtained by at least partially decoding the bitstream BF.
Following the example given above, weighting coefficients a{ (and potentially shape functions ψ() are then obtained as output of step 14.
Then, in step 15, a module BAG generates a decoded version Ba of the backlight frame from the weighting coefficients a{ and either some known non-adaptive shape functions or the shape functions ψ( by:
In step 1 6, a module TMO tone-maps the residual frame Res in order to get a viewable residual frame Resv.
It may appear that the residual frame Res may not be viewable because its dynamic range is too high and because a decoded version of this residual frame Res shows too visible artifacts. Tone-mapping the residual frame remedies to at least one of these drawbacks.
The disclosure is not limited to any specific tone-mapping operator.
This single condition is that the tone-mapping operator shall be reversible.
For example, the tone-mapping operator defined by Reinhard may be used {Reinhard, E., Stark, M., Shirley, P., and Ferwerda, J., \Photographic tone reproduction for digital frames, " ACM Transactions on Graphics 21 (July 2002)), or Boitard, R., Bouatouch, K., Cozot, R., Thoreau, D., & Gruson, A. (2012). Temporal coherency for video tone mapping. In A. M. J. van Eijk, C. C. Davis, S. M. Hammel, & A. K. Majumdar (Eds.), Proc. SPIE 8499, Applications of Digital Frame Processing (p. 84990 D-84990D-10)).
In step 19, the viewable residual frame Resv is encoded by means of an encoder ENC2 in a bitstream F which may be stored on a local or remote memory and/or transmitted through a communication interface (e.g. on a bus or over a communication network or a broadcast network).
According to an embodiment of the step 1 6, tone mapping the residual frame comprises either a gamma correction or a SLog correction according to the pixel values of the residual frame.
The viewable residual frame Resv is then given, for example, by:
Resv = A. ResY
with A being a constant value, γ being a coefficient of a gamma curve equal, for example, to 1 /2.4.
Alternatively, the viewable residual frame Resv is given, for example, by:
Resv = a. \n(Res + b) + c
with a,b and c being coefficients of a SLog curve determined such that 0 and 1 are invariant, and the derivative of the SLog curve is continuous in 1 when prolonged by a gamma curve below 1 . Thus, a,b and c are functions of the parameter y.
According to an embodiment, the parameter y of the gamma-Slog curve is encoded in the bitstream BF.
Applying a gamma correction on the residual frame Res, pulls up the dark regions but does not lower enough high lights to avoid burning of bright pixels.
Applying a SLog correction on the residual frame Res lowers enough high lights but does not pull up the dark regions.
Then, according to a preferred embodiment of the step 16, the module
TMO applies either the gamma correction or the SLog correction according to the pixel values of the residual frame Res.
For example, when the pixel value of the residual frame Res is below a threshold (equal to 1 ), then the gamma correction is applied and otherwise the SLog correction is applied.
By construction, the viewable residual frame Resv usually has a mean value more or less close to 1 depending on the brightness of the frame I, making the use of the above gamma-Slog combination particularly efficient.
According to an embodiment of the method, in step 17, a module SCA scales the viewable residual frame Resv before encoding (step 19) by
multiplying each component of the viewable residual frame Resv by a scaling factor cstscaiing- The resulting residual frame Ress is then given by
ReSs = CStscaling- ReSv
In a specific embodiment, the scaling factor cstscaiing is defined to map the values of the viewable residual frame Resv between from 0 to the maximum value 2N-1 , where N is the number of bits allowed as input for the coding by the encoder ENC2.
This is naturally obtained by mapping the value 1 (which is roughly the mean value of the viewable residual frame Resv) to the mid-gray value 2N"1. Thus, for a viewable residual frame Resv with a standard number of bits N=8, a scaling factor equal to 120 is a very consistent value because very closed to the neutral gray at 27=128.
According to this embodiment of the method, in step 19, the residual frame Ress is encoded by means of an encoder ENC2.
According to an embodiment of the method, in step 18, a module CLI clips the viewable residual frame Resv before encoding to limit its dynamic range to a targeted dynamic range TDR which is defined, for example, according to the capabilities of the encoder ENC2.
According to this last embodiment, the resulting residual frame Resc is given, for example, by:
Resc = max(2N, Resv)
Resc = max(2N, Ress)
according to the embodiments of the method.
The disclosure is not limited to such clipping (max(.)) but extends to any kind of clipping.
According to this embodiment of the method, in step 19, the residual frame Resc is encoded by means of an encoder ENC2.
Combining the scaling and clipping embodiments leads to a residual frame Ressc given by:
Ressc = max(2N, cstScaiing*Resv)
or by Ressc = max(2N, cstSCaiing*Ress) according to the embodiments of the method.
According to this embodiment of the method, in step 19, the residual frame Ressc is encoded by means of an encoder ENC2.
The tone-mapping and scaling of the viewable residual frame Resv is a parametric process. The parameters may be fixed or not and in the latter case they may be encoded in the bitstream BF by means of the encoder ENC1 .
According to an embodiment of the method, the constant value γ of the gamma correction, the scaling factor cstscaiing may be parameters which are encoded in the bitstream BF.
It may be noted that the choice of the parameters , cstmod, cstscaling, Y, β gives room for the choice of the tone-mapping which suits the content the best following the taste of an expert in post-production and color grading.
On the other hand, universal parameters may be defined in order to be acceptable for all of a large variety of frames. Then, no parameters are encoded in the bitstream BF.
Fig. 8 shows a block diagram of the steps of a method, in accordance with an embodiment of the disclosure, for decoding a bitstream representing a residual frame calculated by dividing an frame by a backlight frame.
As explained above, in steps 14 and 15, a backlight frame Ba~ is obtained for example by at least partially decoding a bitstream BF by means of the decoder DEC1 .
The bitstream BF may have been stored locally or received from a communication network.
In step 81 , a decoded residual frame Res ' is obtained by a at least partial decoding of a bitstream F by means of a decoder DEC2.
The bitstream F may have been stored locally or received from a communication network.
As explained below, the decoded residual frame Res ' is viewable by a traditional apparatus.
In step 84, a decoded frame / is obtained by multiplying the decoded residual frame Res~ by the backlight frame B^.
According to an embodiment of step 14, the parameters γ and/or
are a'so obtained either from a local memory or by a at least partial decoding of the bitstream BF by means of the decoder DEC1 .
According to the method, in step 82, a module ISCA applied an inverse scaling to the decoded residual frame Res~ by dividing the decoded residual frame Res~ by the parameter cst^ing.
In step 83, a module ITMO applied an inverse-tone-mapping to the decoded residual frame Res ' , by means of the parameters γ.
For example, the parameter γ defines a gamma curve and the inverse-tone-mapping is just to find, from the gamma curve, the values which correspond to the pixel values of the decoded residual frame Res~ .
The decoders DEC1 , respectively DEC2, is configured to decode data which have been encoded by the encoder ENC1 , respectively DEC2.
The encoders ENC1 and ENC2 (and decoders DEC1 and DEC2) are not limited to a specific encoder (decoder) but when an entropy encoder (decoder) is required, an entropy encoder such as a Huffmann coder, an arithmetic coder or a context adaptive coder like Cabac used in h264/AVC or HEVC is advantageous.
The encoders ENC1 and ENC2 (and decoders DEC1 and DEC2) are not limited to a specific encoder which may be, for example, an frame/video coder with loss like JPEG, JPEG2000, MPEG2, h264/AVC or HEVC.
On Fig. 1-8, the modules are functional units, which may or not be in relation with distinguishable physical units. For example, these modules or some of them may be brought together in a unique component or circuit, or contribute to functionalities of a software. A contrario, some modules may potentially be composed of separate physical entities. The apparatus which are compatible with the disclosure are implemented using either pure hardware, for example using dedicated hardware such ASIC or FPGA or VLSI, respectively « Application Specific Integrated Circuit », « Field- Programmable Gate Array », « Very Large Scale Integration », or from several integrated electronic components embedded in a device or from a blend of hardware and software components.
Fig. 9 represents an exemplary architecture of a device 90 which may be configured to implement a method described in relation with Fig. 1-8.
Device 90 comprises following elements that are linked together by a data and address bus 91 :
- a microprocessor 92 (or CPU), which is, for example, a DSP (or Digital Signal Processor);
- a ROM (or Read Only Memory) 93;
- a RAM (or Random Access Memory) 94;
- an I/O interface 95 for reception of data to transmit, from an application; and
- a battery 96
According to a variant, the battery 96 is external to the device. Each of these elements of Fig. 9 are well-known by those skilled in the art and won't be disclosed further. In each of mentioned memory, the word « register » used in the specification can correspond to area of small capacity (some bits) or to very large area (e.g. a whole program or large amount of received or decoded data). ROM 93 comprises at least a program and parameters. Algorithm of the methods according to the disclosure is stored in the ROM 93. When switched on, the CPU 92 uploads the program in the RAM and executes the corresponding instructions.
RAM 94 comprises, in a register, the program executed by the CPU 92 and uploaded after switch on of the device 90, input data in a register, intermediate data in different states of the method in a register, and other variables used for the execution of the method in a register.
The implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method or a device), the implementation of features discussed may also be implemented in other forms (for example a program). An apparatus may be implemented in, for example, appropriate hardware, software, and firmware. The methods may be implemented in, for example, an apparatus such as, for example, a
processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs"), and other devices that facilitate communication of information between end-users.
According to a specific embodiment of encoding or encoder, the frame I is obtained from a source. For example, the source belongs to a set comprising:
- a local memory (93 or 94), e.g. a video memory or a RAM (or
Random Access Memory), a flash memory, a ROM (or Read Only Memory), a hard disk ;
- a storage interface (95), e.g. an interface with a mass storage, a RAM, a flash memory, a ROM, an optical disc or a magnetic support;
- a communication interface (95), e.g. a wireline interface (for example a bus interface, a wide area network interface, a local area network interface) or a wireless interface (such as a IEEE 802.1 1 interface or a Bluetooth® interface); and
- an frame capturing circuit (e.g. a sensor such as, for example, a
CCD (or Charge-Coupled Device) or CMOS (or Complementary Metal-Oxide-Semiconductor)).
According to different embodiments of the decoding or decoder, the decoded frame Ϊ is sent to a destination; specifically, the destination belongs to a set comprising:
- a local memory (93 or 94), e.g. a video memory or a RAM, a flash memory, a hard disk ;
- a storage interface (95), e.g. an interface with a mass storage, a RAM, a flash memory, a ROM, an optical disc or a magnetic support;
- a communication interface (95), e.g. a wireline interface (for example a bus interface (e.g. USB (or Universal Serial Bus)), a
wide area network interface, a local area network interface, a HDMI (High Definition Multimedia Interface) interface) or a wireless interface (such as a IEEE 802.1 1 interface, WiFi ® or a Bluetooth ® interface); and
- a display.
According to different embodiments of encoding or encoder, the bitstream BF and/or F are sent to a destination. As an example, one of bitstream F and BF or both bitstreams F and BF are stored in a local or remote memory, e.g. a video memory (94) or a RAM (94), a hard disk (93). In a variant, one or both bitstreams are sent to a storage interface (65), e.g. an interface with a mass storage, a flash memory, ROM, an optical disc or a magnetic support and/or transmitted over a communication interface (95), e.g. an interface to a point to point link, a communication bus, a point to multipoint link or a broadcast network.
According to different embodiments of decoding or decoder, the bitstream BF and/or F is obtained from a source. Exemplarily, the bitstream is read from a local memory, e.g. a video memory (94), a RAM (94), a ROM (93), a flash memory (93) or a hard disk (93). In a variant, the bitstream is received from a storage interface (95), e.g. an interface with a mass storage, a RAM, a ROM, a flash memory, an optical disc or a magnetic support and/or received from a communication interface (95), e.g. an interface to a point to point link, a bus, a point to multipoint link or a broadcast network.
According to different embodiments, device 90 being configured to implement an encoding method described in relation with Fig. 1-7, belongs to a set comprising:
- a mobile device ;
- a communication device ;
- a game device ;
- a tablet (or tablet computer) ;
- a laptop ;
- a still frame camera;
- a video camera ;
- an encoding chip;
- a still frame server ; and
- a video server (e.g. a broadcast server, a video-on-demand server or a web server).
According to different embodiments, device 90 being configured to implement a decoding method described in relation with Fig. 8, belongs to a set comprising:
- a mobile device ;
- a communication device ;
- a game device ;
- a set top box;
- a TV set;
- a tablet (or tablet computer) ;
- a laptop ;
- a display and
- a decoding chip.
According to an embodiment illustrated in Fig. 10, in a transmission context between two remote devices A and B over a communication network NET, the device A comprises means which are configured to implement a method for encoding an frame as described in relation with the Fig. 1 and the device B comprises means which are configured to implement a method for decoding as described in relation with Fig. 8.
According to a variant of the disclosure, the network is a broadcast network, adapted to broadcast still frames or video frames from device A to decoding devices including the device B.
Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications, particularly, for example, equipment or applications. Examples of such equipment include an encoder, a decoder, a post-processor processing output from a decoder, a pre-processor providing input to an encoder, a video coder, a video decoder, a video codec, a web server, a set-top box, a laptop, a personal computer, a cell phone, a PDA, and other communication
devices. As should be clear, the equipment may be mobile and even installed in a mobile vehicle.
Additionally, the methods may be implemented by instructions being performed by a processor, and such instructions (and/or data values produced by an implementation) may be stored on a processor-readable medium such as, for example, an integrated circuit, a software carrier or other storage device such as, for example, a hard disk, a compact diskette ("CD"), an optical disc (such as, for example, a DVD, often referred to as a digital versatile disc or a digital video disc), a random access memory ("RAM"), or a read-only memory ("ROM"). The instructions may form an application program tangibly embodied on a processor-readable medium. Instructions may be, for example, in hardware, firmware, software, or a combination. Instructions may be found in, for example, an operating system, a separate application, or a combination of the two. A processor may be characterized, therefore, as, for example, both a device configured to carry out a process and a device that includes a processor-readable medium (such as a storage device) having instructions for carrying out a process. Further, a processor-readable medium may store, in addition to or in lieu of instructions, data values produced by an implementation.
As will be evident to one of skill in the art, implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted. The information may include, for example, instructions for performing a method, or data produced by one of the described implementations. For example, a signal may be formatted to carry as data the rules for writing or reading the syntax of a described embodiment, or to carry as data the actual syntax-values written by a described embodiment. Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal. The formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream. The information that the signal carries may be, for example, analog or digital information. The signal may be transmitted over a variety of different
wired or wireless links, as is known. The signal may be stored on a processor-readable medium.
A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made. For example, elements of different implementations may be combined, supplemented, modified, or removed to produce other implementations. Additionally, one of ordinary skill will understand that other structures and processes may be substituted for those disclosed and the resulting implementations will perform at least substantially the same function(s), in at least substantially the same way(s), to achieve at least substantially the same result(s) as the implementations disclosed. Accordingly, these and other implementations are contemplated by this application.
Claims
1 . A method for encoding a frame, comprising:
- encoding (12) a backlight frame determined (1 1 ) from the frame to be encoded;
- obtaining (13) at least one component of a residual frame by dividing each component of the frame by a decoded version of the backlight frame;
- encoding (19) the residual frame;
the backlight frame being represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports, wherein the 2D shape functions of a set have the shape of a model symmetric function,
wherein the model symmetric function is the product of two separable 1 D polynomials.
2. A device for encoding a frame, comprising a processor configured to:
- encode a backlight frame determined from the frame to be encoded;
- obtain at least one component of a residual frame by dividing each component of the frame by a decoded version of the backlight frame;
- encode the residual frame;
the backlight frame being represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports, wherein the 2D shape functions of a set have the shape of a model symmetric function,
wherein the model symmetric function is the product of two separable 1 D polynomials.
3. A method for decoding a frame from at least one bitstream representing a backlight frame obtained from the frame and a residual frame calculated by
dividing the frame by the backlight frame, the method comprising a processor configured to:
- obtain (14) a decoded backlight frame by a at least partially decoding of a bitstream;
- obtain a decoded residual frame by a at least partially decoding of a bitstream; and
- obtain (51 ) a decoded frame by multiplying the decoded residual frame by the decoded backlight frame;
the decoded backlight frame being represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports,
wherein the 2D shape functions of a set have the shape of a model symmetric function,
wherein the model symmetric function is the product of two separable
1 D polynomials.
4. A device for decoding a frame from at least one bitstream representing a backlight frame obtained from the frame and a residual frame calculated by dividing the frame by the backlight frame, the device comprising a processor configured to:
- obtain a decoded backlight frame by a at least partially decoding of a bitstream;
- obtain a decoded residual frame by a at least partially decoding of a bitstream; and
- obtain a decoded frame by multiplying the decoded residual frame by the decoded backlight frame;
the decoded backlight frame being represented by a weighted linear combination of at least one set of 2D shape functions, the 2D shape functions of each set forming a partition of unity of a domain defined by the union of their supports,
wherein the 2D shape functions of a set have the shape of a model symmetric function,
wherein the model symmetric function is the product of two separable 1 D polynomials.
5. A computer program product comprising program code instructions to execute the steps of the encoding method according to the claim 1 when this program is executed on a computer,
6. A computer program product comprising program code instructions to execute the steps of the decoding method according to the claim 3 when this program is executed on a computer.
7. A processor readable medium having stored therein instructions for causing a processor to perform at least the steps of the encoding method according to the claim 1 .
8. A processor readable medium having stored therein instructions for causing a processor to perform at least the steps of the decoding method according to the claim 3.
9. Non-transitory storage medium carrying instructions of program code for executing steps of the method according to one of claims 1 or 3 when said program is executed on a computing device.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP14305732 | 2014-05-20 | ||
EP14305732.1 | 2014-05-20 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015177139A1 true WO2015177139A1 (en) | 2015-11-26 |
Family
ID=50884314
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2015/060983 WO2015177139A1 (en) | 2014-05-20 | 2015-05-19 | Method and device for encoding a high-dynamic range frame and/or decoding a bitstream |
Country Status (2)
Country | Link |
---|---|
TW (1) | TW201608876A (en) |
WO (1) | WO2015177139A1 (en) |
-
2015
- 2015-05-15 TW TW104115465A patent/TW201608876A/en unknown
- 2015-05-19 WO PCT/EP2015/060983 patent/WO2015177139A1/en active Application Filing
Non-Patent Citations (3)
Title |
---|
DAVID TOUZÉ ET AL: "HDR Video Coding based on Local LDR Quantization", HDRI2014 -SECOND INTERNATIONAL CONFERENCE AND SME WORKSHOP ON HDR IMAGING, 4 March 2014 (2014-03-04), XP055112158, Retrieved from the Internet <URL:http://people.irisa.fr/Ronan.Boitard/articles/2014/HDR Video Coding based on Local LDR Quantization.pdf> [retrieved on 20140404] * |
LASSERRE S ET AL: "High Dynamic Range video coding", 16. JCT-VC MEETING; 9-1-2014 - 17-1-2014; SAN JOSE; (JOINT COLLABORATIVE TEAM ON VIDEO CODING OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ); URL: HTTP://WFTP3.ITU.INT/AV-ARCH/JCTVC-SITE/,, no. JCTVC-P0159, 5 January 2014 (2014-01-05), XP030115677 * |
TAKAO JINNO ET AL: "New local tone mapping and two-layer coding for HDR images", 2012 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2012) : KYOTO, JAPAN, 25 - 30 MARCH 2012 ; [PROCEEDINGS], IEEE, PISCATAWAY, NJ, 25 March 2012 (2012-03-25), pages 765 - 768, XP032227239, ISBN: 978-1-4673-0045-2, DOI: 10.1109/ICASSP.2012.6287996 * |
Also Published As
Publication number | Publication date |
---|---|
TW201608876A (en) | 2016-03-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102367205B1 (en) | Method and device for encoding both a hdr picture and a sdr picture obtained from said hdr picture using color mapping functions | |
US9924178B2 (en) | Method and device for encoding a high-dynamic range image and/or decoding a bitstream | |
US9979973B2 (en) | Encoding and decoding methods for adapting the average luminance of high dynamic range pictures and corresponding encoder and decoder | |
US20160337668A1 (en) | Method and apparatus for encoding image data and method and apparatus for decoding image data | |
EP3087740A1 (en) | Method and device for tone-mapping a high dynamic range image | |
EP3107301A1 (en) | Method and device for encoding both a high-dynamic range frame and an imposed low-dynamic range frame | |
TW201540052A (en) | Method and device for encoding a high-dynamic range image into a bitstream and/or decoding a bitstream representing a high-dynamic range image | |
US10148958B2 (en) | Method and device for encoding and decoding a HDR picture and a LDR picture | |
WO2015177132A1 (en) | Method and device for encoding a high-dynamic range frame and/or decoding a bitstream | |
WO2015097126A1 (en) | Method and device for encoding a high-dynamic range image and/or decoding a bitstream | |
WO2015177139A1 (en) | Method and device for encoding a high-dynamic range frame and/or decoding a bitstream | |
WO2015177133A1 (en) | Method and device for encoding a high-dynamic range frame and/or decoding a bitstream | |
WO2015177136A1 (en) | Method and device for encoding a high-dynamic range frame and/or decoding a bitstream | |
WO2015097135A1 (en) | Method and device for encoding a high-dynamic range image | |
EP3113494A1 (en) | Method and device for encoding a high-dynamic range image | |
WO2015177123A1 (en) | Method and device for encoding a frame and/or decoding a bitstream representing a frame | |
EP3146718B1 (en) | Method and device for scalable encoding of a high dynamic range frame and/or decoding a bitstream representing such a frame | |
EP2887665A1 (en) | Method and device for encoding a high-dynamic range image | |
WO2015097131A1 (en) | Method and device for encoding a high-dynamic range image | |
WO2015097129A1 (en) | Method and device for encoding a high-dynamic range image | |
WO2015097124A1 (en) | Method and device for encoding a high-dynamic range image and/or decoding a bitstream | |
WO2015097134A1 (en) | Method and device for encoding a high-dynamic range image and/or decoding a bitstream | |
EP2938083A1 (en) | Method and device for processing an information data indicating that a frame contains samples of at least two distinct packed constituent frames according to a specific frame-packing-arrangement scheme | |
WO2015177125A1 (en) | Method and device for encoding a frame and/or decoding a bitstream representing a frame | |
WO2015177126A1 (en) | Method and device for encoding a frame and/or decoding a bitstream representing a frame |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15726041 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15726041 Country of ref document: EP Kind code of ref document: A1 |