WO2013001013A1 - Procédé pour décoder un train de bits vidéo extensible, et dispositif de décodage correspondant - Google Patents

Procédé pour décoder un train de bits vidéo extensible, et dispositif de décodage correspondant Download PDF

Info

Publication number
WO2013001013A1
WO2013001013A1 PCT/EP2012/062586 EP2012062586W WO2013001013A1 WO 2013001013 A1 WO2013001013 A1 WO 2013001013A1 EP 2012062586 W EP2012062586 W EP 2012062586W WO 2013001013 A1 WO2013001013 A1 WO 2013001013A1
Authority
WO
WIPO (PCT)
Prior art keywords
coefficients
blocks
transformed
decoding
image
Prior art date
Application number
PCT/EP2012/062586
Other languages
English (en)
Inventor
Fabrice Le Leannec
Sébastien Lasserre
Original Assignee
Canon Kabushiki Kaisha
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Kabushiki Kaisha filed Critical Canon Kabushiki Kaisha
Publication of WO2013001013A1 publication Critical patent/WO2013001013A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • H04N19/126Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • H04N19/139Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/18Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/187Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Definitions

  • the present invention concerns a method for decoding a scalable video bit- stream, and an associated decoding device.
  • Video compression algorithms such as those standardized by the standardization organizations ITU, ISO, and SMPTE, exploit the spatial and temporal redundancies of images in order to generate bitstreams of data of smaller size than original video sequences.
  • These powerful video compression tools known as spatial (or intra) and temporal (or inter) predictions, make the transmission and/or the storage of video sequences more efficient.
  • Video encoders and/or decoders are often embedded in portable devices with limited resources, such as cameras or camcorders.
  • Conventional embedded codecs can process at best high definition (HD) digital videos, i.e 1080x1920 pixel frames.
  • Real time encoding and decoding are however limited by the limited resources of the portable devices, especially regarding slow access to the working memory (e.g. random access memory, or RAM) and regarding the central processing unit (CPU).
  • working memory e.g. random access memory, or RAM
  • CPU central processing unit
  • UHD is typically four times (4k2k pixels) the definition of an HD video which is the current standard definition video. Furthermore, very ultra high definition, which is sixteen times that definition (i.e. 8k4k pixels), is even being considered in a more long- term future.
  • the inventors provide a UHD codec with low complexity based on scalable encoding.
  • the UHD video is encoded into a base layer and one or more enhancement layers.
  • the base layer results from the encoding of a low resolution version of the UHD images, in particular having a HD resolution, with a standard existing codec (e.g. H.264 or HEVC - High Efficiency Video Coding).
  • a standard existing codec e.g. H.264 or HEVC - High Efficiency Video Coding.
  • the compression efficiency of such a codec relies on spatial and temporal predictions.
  • an enhancement image is obtained from subtracting an interpolated (or up-scaled or upsampled) decoded image of the base layer from the corresponding original UHD image.
  • the enhancement images which are residuals or pixel differences with UHD resolution, are then encoded into an enhancement layer.
  • Figure 1 illustrates such approach at the encoder 10.
  • An input raw video 1 in particular a UHD video, is down-sampled 12 to obtain a so-called base layer, for example with HD resolution, which is encoded by a standard base video coder 13, for instance H.264/AVC or HEVC. This results in a base layer bit stream 14.
  • a standard base video coder 13 for instance H.264/AVC or HEVC.
  • the encoded base layer is decoded 15 and up-sampled 16 into the initial resolution (UHD in the example) to obtain the up- sampled decoded base layer.
  • the latter is then subtracted 17, in the pixel domain, from the original raw video to get the residual enhancement layer X.
  • the information contained in X is the error or pixel difference due to the base layer encoding and the up-sampling. It is also known as a "residual".
  • a conventional block division is then applied, for instance a homogenous 8x8 block division (but other divisions with non-constant block size are also possible).
  • DCT blocks forming the DCT image X DCT having the initial UHD resolution are DCT blocks forming the DCT image X DCT having the initial UHD resolution.
  • This DCT image X DCT is encoded in XDCT.Q ⁇ V an enhancement video encoding module 19 into an enhancement layer bit stream 20.
  • the encoded bit-stream EBS resulting from the encoding of the raw video 1 1 is made of: - the base layer bit-stream 14 produced by the base video encoder 13;
  • Figure 2 illustrates the associated processing at the decoder 30 receiving the encoded bit-stream EBS.
  • Part of the processing consists in decoding the base layer bit-stream 14 by the standard base video decoder 31 to produce a decoded base layer.
  • the decoding of the base layer comprises using motion information to temporally predict blocks of a first base image from blocks of a decoded reference base image.
  • reference image is well-known from conventional encoding methods.
  • the decoded base layer is then up-sampled 32 into the initial resolution, i.e. UHD resolution.
  • each enhancement image of the enhancement layer has a high resolution (in the example UHD) and temporally corresponds to a base image of the base layer. Due to the scalability of the video, at a given time, there is a temporal correspondence between an image of the video, a base image and at least one enhancement image.
  • both the enhancement layer bit-stream 20 and the parameters 21 are used by the enhancement video decoding and dequantization module 33 to generate ⁇ ⁇ .
  • the image X ⁇ c is the result of the quantization and then the inverse quantization on the image X DCT .
  • An inverse DCT transform 34 is then applied to each block of the dequantized image to obtain the decoded residual (of UHD resolution) in the pixel domain.
  • Each decoded enhancement image is then added, in the pixel domain, to the corresponding up-sampled decoded base image (block by block), to obtain a decoded video of decoded high resolution images.
  • each decoded residual X ⁇ T Q - ⁇ is added 35 to the corresponding block in the up-sampled decoded base layer to obtain decoded images of the video.
  • Filter post-processing for instance with a deblocking filter 36, is finally applied to obtain the decoded video 37 which is output by the decoder 30.
  • Reducing UHD encoding and decoding complexity relies on simplifying the encoding of the enhancement images at the enhancement video encoding module 19 compared to the conventional encoding scheme.
  • the inventors dispense with the temporal prediction and possibly the spatial prediction when encoding the UHD enhancement images. This is because the temporal prediction is very expensive in terms of memory bandwidth consumption, since it often requires accessing other enhancement images as reference images. Low-complexity codecs may then be designed, in particular at the encoding side.
  • the enhancement video encoding module 19 may model the statistical distribution of the DCT coefficients within the DCT blocks of a current enhancement image X by fitting a parametric probabilistic model.
  • This fitted model becomes the channel model of DCT coefficients and the fitted parameters are output in the parameter bit-stream 21 coded by the enhancement layer encoder.
  • a channel model may be obtained for each DCT coefficient position within a DCT block based on fitting the parametric probabilistic model onto the corresponding collocated DCT coefficients throughout all the DCT blocks of the image X DCT or of part of it.
  • the channel models are reconstructed from the received parameters 21 , enabling retrieval of the selected quantizers and the entropy coding used.
  • conventional quantizers and conventional entropy coding such as Huffman codes, may also be used.
  • the present invention particularly focuses on the decoding of a resulting scalable video bit-stream.
  • decoding a first enhancement image may involve using the motion information of the corresponding first base image in order to obtain residual blocks from a "reference" decoded UHD image (temporally corresponding to the decoded reference base image used for predicting the first base image). Such blocks may then be used to correct the enhancement image data directly obtained from the bit-stream 20.
  • temporal prediction information of the base layer to decode the enhancement layer is known from the standard SVC (standing for "Scalable Video Coding").
  • SVC Scalable Video Coding
  • the present invention intends to improve the efficiency of a decoding method based on predicting the enhancement layer. This aims at improving the quality of reconstructed high resolution (e.g. UHD) images, while keeping low complexity at the encoding and decoding sides.
  • UHD high resolution
  • a first aspect of the invention concerns a method for decoding a scalable video bit-stream, comprising decoding a base layer from the bit-stream, decoding an enhancement layer from the bit-stream and adding the enhancement layer to the base layer to obtain a decoded video of high resolution images, wherein decoding the enhancement layer comprises:
  • the method for decoding a scalable video bit-stream comprises:
  • the decoding of the low resolution version comprising using motion information to temporally predict blocks of a low resolution image from blocks of a decoded reference low resolution image;
  • each enhancement image of the enhancement version having a high resolution and temporally corresponding to a low resolution image of the low resolution video
  • decoding a first enhancement image temporally corresponding to a first low resolution image comprises:
  • the blocks of the encoded enhancement layer obtained from the bit-stream and the residual blocks obtained using the motion information of the base layer are merged together to form parts of the decoded enhancement image.
  • this decoded enhancement image is then added to an up-sampled decoded base image to obtain a decoded high resolution (e.g. UHD) image.
  • the quality of the decoded high resolution image is improved compared to known techniques. This is due to the use of two probabilistic distributions that model both the original transformed coefficients and an error of temporal prediction, when merging the transformed coefficients (e.g. DCT coefficients).
  • the first probabilistic distribution corresponding to the transformed coefficients encoded in the bit-stream may be obtained from the bit-stream itself, for instance from the parameters 21 defined above. These may represent statistical modelling of the original transformed coefficients (i.e. before quantization and encoding).
  • the second probabilistic distributions that correspond to the blocks predicted using the motion information of the base layer provide information about the noise of temporal prediction. In particular they provide modelled information on the difference between those predicted coefficients and the transformed coefficients. Since the original transformed coefficients are not known by the decoder, the decoded and dequantized transformed coefficients known at the decoding side are used in place of the original transformed coefficients. The inventors have observed that using those coefficients rather than the original ones provides modelling that is quite close to reality.
  • probabilities such as the expected value example below, provide good statistical results.
  • the temporally predicted blocks may more often provide relevant information on the original DCT coefficients than the quantization level obtained by the dequantization.
  • the opposite occurs.
  • the invention allows gains up to several dBs in rate-distortion performance at almost no cost of additional complexity at the decoder, and at the cost of zero additional rate when the parameters 21 have already been transmitted.
  • the approach according to the invention does not necessarily have to be performed at the decoding side. For example, it may be switched off in case of very low complexity decoders. Further, the encoding is independent of the switching decision.
  • the invention also relates to a decoding device for decoding a scalable video bit-stream, comprising a base layer decoder configured to decode a base layer from the bit-stream, an enhancement layer decoder configured to decode an enhancement layer from the bit-stream and a video building unit configured to add the enhancement layer to the base layer to obtain a decoded video, wherein the enhancement layer decoder is further configured to:
  • the decoding device for decoding a scalable video bit- stream comprises:
  • a base decoder configured to decode a low resolution version of the video, using motion information to temporally predict blocks of a low resolution image from blocks of a decoded reference low resolution image;
  • an enhancement decoder configured to decode an enhancement version of the video, each enhancement image of the enhancement version having a high resolution and temporally corresponding to a low resolution image of the low resolution video; and - an image building unit configured to add each decoded enhancement image to an up-sampled version of the corresponding low resolution image, to obtain a decoded video of decoded high resolution images;
  • enhancement decoder is further configured to:
  • Another aspect of the invention relates to an information storage means, able to be read by a computer system, comprising instructions for a computer program adapted to implement the decoding method as set out above, when the program is loaded into and executed by the computer system.
  • Yet another aspect of the invention relates to a computer program product able to be read by a microprocessor, comprising portions of software code adapted to implement the decoding method as set out above, when it is loaded into and executed by the microprocessor.
  • the decoding device, the computer program and the information storage means may have features and advantages that are analogous to those set out above and below in relation to the decoding method, in particular that of refining decoded transformed DCT coefficients and of improving the quality of decoded high resolution images.
  • Another aspect of the invention relates to a method for decoding an image substantially as herein described with reference to, and as shown in, Figure 5; Figures 5 and 11 ; Figures 5, 10 and 11 of the accompanying drawings.
  • Another aspect of the invention relates to a decoding device for encoding an image substantially as herein described with reference to, and as shown in, Figure 5; Figures 5 and 10 of the accompanying drawings.
  • the step of merging may merge a dequantized transformed coefficient with a collocated coefficient in the transformed residual blocks (meaning collocated blocks and collocated coefficients within those blocks), using first and second probabilistic distributions associated with these collocated coefficients, on a quantization interval associated with the value of the corresponding quantized transformed coefficient (i.e. the value before the quantized transformed coefficient is dequantized).
  • the first and second probabilistic distributions are integrated using Riemann sums over that quantization interval during the merging step. This provision makes it possible to perform a probabilistic merger of transformed coefficients, on low complexity decoders.
  • the step of merging comprises calculating the expectation of a block coefficient, given the quantization interval associated with the value of the corresponding quantized transformed coefficient and given its corresponding value in the transformed residual blocks, based on the first and second probabilistic distributions.
  • calculating the expectation x ⁇ of a block coefficient / comprises calculating the following value:
  • PDFi is the first probabilistic distribution associated with the block coefficient / '
  • PDF N is the second probabilistic distribution
  • Y 0 is the value of the coefficient collocated with said block coefficient / ' in the transformed residual blocks
  • Q m is the quantization interval associated with the value of the quantized transformed coefficient collocated with said block coefficient /.
  • the probabilistic distributions are generalized Gaussian distributions
  • GGD ⁇ a, , /?,. , x) — exp (-lx / a,
  • This parametric model is well-suited for modelling noise, such as the residuals.
  • the obtaining of the second probabilistic distribution comprises fitting a Generalized Gaussian Distribution model onto the differences between the coefficients in the transformed residual blocks and the dequantized transformed coefficients.
  • the second probabilistic distribution is statistically obtained based on the coefficients that are actually handled by the decoder.
  • the obtaining of the first probabilistic distribution comprises obtaining parameters from the bit-stream and applying these parameters to a probabilistic distribution model.
  • the low resolution or base image temporally corresponding to a first enhancement image to decode is an image bi-directionally predicted from reference low resolution or base images using motion information in each of the two directions
  • the decoding of the first enhancement image comprises obtaining transformed residual blocks for each direction and merging together the transformed residual blocks in both directions with the dequantized blocks of dequantized transformed coefficients.
  • This approach proves to be more precise than an approach which first determines a single transformed residual block based on prediction in both directions. This is because a motion prediction noise estimation in each direction is separately obtained, improving a probabilistic merger.
  • the merging can be based on calculating an expectation.
  • the step of merging may comprise calculating the merger value x. of a block coefficient / ' using the formula:
  • PDF N (x - y 0 ) PDF N ,(x - Y )dx
  • PDF is the first probabilistic distribution associated with the block coefficient / '
  • PDF N and PDF N ' are the second probabilistic distributions for respectively each of the two directions
  • Y 0 and Y' 0 are the value of the coefficient collocated with said block coefficient in the transformed residual blocks in respectively each of the two directions
  • Q m is the quantization interval associated with the value of the quantized transformed coefficient collocated with said block coefficient / ' .
  • obtaining residual blocks comprises:
  • These steps define the temporal prediction of the enhancement layer based on the images already reconstructed. They produce another enhancement layer (since each obtained block is the difference with the base layer) from which a modelling of the temporal prediction noise can be performed.
  • that motion information before using the motion information, that motion information is up-sampled (or interpolated) into high resolution. This is because the reference image on which that information is about to be used is of high resolution.
  • the motion information that is up-sampled comprises, for a given block, a motion vector and a temporal residual block; and the obtaining of the motion predictor blocks comprises
  • the decoding method may further comprise filtering, using a deblocking filter, the obtained decoded high resolution images; wherein parameters (e.g. the filter strength parameter or the quantization- dependent parameter) of the deblocking filter depend on the first and second probabilistic distributions used during the merger.
  • parameters e.g. the filter strength parameter or the quantization- dependent parameter
  • the second probabilistic distributions are obtained for blocks collocated with enhancement image blocks of the corresponding low resolution or base image that are encoded with the same coding mode.
  • the coding mode of the base (low resolution) layer is for example the INTER mode, which may be further subdivided into an INTER P-prediction mode and an INTER B-prediction mode, or the SKIP mode (as defined in H.264).
  • first probabilistic distributions are obtained for respectively each of a plurality of channels, wherein a channel is associated with collocated coefficients having the same block coefficient position in their respective blocks. Furthermore, a channel may be restricted to the blocks collocated with base layer blocks having the same coding mode.
  • a method of decoding video data comprising:
  • the decompression of the video data of the enhancement layer employs a method embodying the aforesaid first aspect of the present invention.
  • the decoding of the base layer video data is in conformity with HEVC.
  • the first resolution is UHD and the second resolution is HD.
  • the compression of the residual data does not involve temporal prediction and/or that the compression of the residual data also does not involve spatial prediction.
  • FIG. 1 schematically shows an encoder for a scalable codec
  • FIG. 3 schematically illustrates the enhancement video encoding module of the encoder of Figure 1 ;
  • FIG. 4 schematically illustrates the enhancement video decoding module of the encoder of Figure 2;
  • FIG. 5 is a more detailed schematic illustration of the decoder of Figure 2 according to the invention.
  • FIG. 6 illustrates a structure of a 4:2:0 macroblock
  • FIG. 7 illustrates an example of a quantizer based on Voronoi cells
  • Figure 9 illustrates an implementation of entry points to allow spatial random access as illustrated in Figure 8.
  • FIG. 13 shows a particular hardware configuration of a device able to implement methods according to the invention.
  • FIG. 3 which illustrates an embodiment of an enhancement video encoding module 19 (or "enhancement layer encoder")
  • a low resolution version of the initial image has been encoded into an encoded low resolution image, referred above as the base layer; and a residual enhancement image has been obtained by subtracting an interpolated high resolution (or up-sampled) decoded version of the encoded low resolution image from said initial image.
  • that residual enhancement image is then transformed from the spatial domain (i.e. pixels) into the (spatial) frequency domain, using for example a block-based DCT transform, to obtain an image of transformed block coefficients.
  • X DCT which comprises a plurality of DCT blocks, each comprising DCT coefficients.
  • the residual enhancement image has been divided into blocks B k , for instance 8x8 blocks but other divisions may be considered, on which the DCT transform is applied.
  • Blocks are grouped into macroblocks MB k .
  • a very common case for so- called 4:2:0 YUV video streams is a macroblock made of 4 blocks of luminance Y, 1 block of chrominance U and 1 block of chrominance V, as illustrated in Figure 6.
  • other configurations may be considered.
  • a macroblock MB k is made of 16x16 pixels of luminance Y and the chrominance has been down-sampled by a factor two both horizontally and vertically to obtain 8*8 pixels of chrominance U and 8*8 pixels of chrominance V.
  • the four blocks within a macroblock MB k are referenced k' a k> ti k> a k ⁇
  • a probabilistic distribution P of each DCT coefficient is determined using a parametric probabilistic model. This step is referenced 190 in the Figure.
  • the image X DCT is a residual image, i.e. information is about a noise residual, it is efficiently modelled by Generalized Gaussian Distributions (GGD) having a zero mean: DCT X » GGD(a,fi) _ where ⁇ , ⁇ are two parameters to be determined and the GGD follows the
  • the DCT coefficients cannot be all modelled by the same parameters and, practically, the two parameters ⁇ , ⁇ may depend on:
  • a DCT channel is thus defined as the DCT coefficients collocated (i.e. having the same index) within a plurality of DCT blocks (possibly all the blocks of the image).
  • a DCT channel can therefore be identified by the corresponding index i ; and/or
  • Intra blocks of the base layer do not behave the same way as Inter blocks.
  • Blocks with a coded residual in the base layer do not behave the same way as blocks without such a residual (i.e. Skipped blocks).
  • blocks coded with a non-nil texture data according to the coded-block-pattern syntax element as defined in H.264/AVC do not behave the same way as those blocks without non-nil texture data.
  • the collocation of blocks should take into account that down-sampling.
  • the four blocks of the n-th macroblock in the residual enhancement layer with UHD resolution are collocated with the n-th block of the base layer having a HD resolution. That is why, generally, all the blocks of a macroblock in an enhancement image have the same base coding mode.
  • the modelling 190 has to determine the parameters of 64 DCT channels for each base coding mode.
  • the luminance component Y and the chrominance components U and V have dramatically different source contents, they must be encoded in different DCT channels. For example, if it is decided to encode the luminance component Y on one channel and the chrominance components UV on another channel, 128 channels are needed for each base coding mode. At least 64 pairs of parameters for each base coding mode may appear as a substantial amount of data to transmit to the decoder (see parameters 21). However, experience proves that this is quite negligible compared to the volume of data needed to encode the residuals of Ultra High Definition (4k2k or more) videos. As a consequence, one may understand that such a technique is preferably implemented on large videos, rather than on very small videos because the parametric data would be too costly.
  • the Generalized Gaussian Distribution model is fitted onto the DCT block coefficients of the DCT channel, i.e. the DCT coefficients collocated within the DCT blocks with the same base coding mode. Since, this fitting is based on the values of the DCT coefficients before quantization (of the DCT blocks having the same base coding mode in the example), the probabilistic distribution is a statistical distribution of the DCT coefficients within a considered channel i.
  • the fitting may be simply and robustly obtained using the moment of order k of the absolute value of a GGD:
  • the value of the parameter ⁇ can thus be estimated by computing the above ratio of the two first and second moments, and then the inverse of the above function of ⁇ ,.
  • the two parameters ⁇ ,, ⁇ , being determined for the DCT coefficients i, the probabilistic distribution P, of each DCT coefficient / in a considered block is defined by
  • a quantization 192 of the DCT coefficients of X DCT is then performed, to obtain quantized DCT coefficients X ' DCT>Q (i.e. symbols or values).
  • the quantization of those coefficients may involve optimal quantizers chosen (step 191) for each DCT channel i based on the corresponding probabilistic distribution P,(x) of the DCT coefficients.
  • the quantizers may be predefined prior to the encoding. Since the quantization is not the core of the present invention, it is here assumed that a quantizer is selected for each DCT channel and each base coding mode as defined above, meaning that various quantizers are generally used for quantizing various DCT coefficients.
  • Figure 7 illustrates an exemplary Voronoi cell based quantizer.
  • a quantizer is made of M Voronoi cells distributed over the values of the
  • Each cell corresponds to an interval [ m ⁇ m+i ] , called quantum Q m .
  • Each cell has a centroid c m , as shown in the Figure.
  • intervals are used for quantization: a DCT coefficient comprised in the interval [ ⁇ , + ⁇ ] is quantized by a symbol a m associated with that interval.
  • centroids are used for de-quantization: a symbol a m associated with an interval is de-quantized into the centroid value c m of that interval.
  • the probabilities ⁇ p i m ⁇ for an alphabet A thus define the probabilistic distribution of the possible symbols or values defined therein.
  • the probabilistic distribution is the same for the alphabets associated with DCT coefficients collocated within a plurality of the blocks of the image.
  • Such probabilities may be computed off-line and stored in memory of the encoder, in order to decrease the complexity of real time encoding. This is for example possible when the parameters ⁇ , ⁇ for modelling the distribution of the DCT coefficients are chosen from a limited number of possible parameters, and when the possible quantizers are know in advance.
  • the probabilities ⁇ pi ,m ⁇ and the quantized symbols or values a m obtained from each DCT coefficient in the DCT image X DCT are then provided for entropy coding 193 as shown in Figure 3.
  • the quantized DCT coefficients are generally processed according to a zigzag scan.
  • the entropy coding may then take into account these probabilities to provide improved encoding.
  • conventional Huffman entropy coding may be implemented in step 193.
  • the entropy coding 193 compresses the DCT image X DCT and generates encoded DCT images X D C C NC T Q which constitute the enhancement layer bit-stream 20.
  • This encoding scheme of the enhancement layer has spatial random access properties due to the absence of inter frame (temporal) and possibly intra block (spatial) predictions.
  • the absence of prediction in the enhancement layer ensures that no dependence between macroblocks exists.
  • an entry point of the generated bit-stream 20 and the index of the associated macroblock are given, it is possible to perform the decoding from that point, without decoding other parts of the encoded video.
  • bit-stream has the random spatial access property because it is possible to decode only a part of the image (a region of interest) once the associated entry points are given.
  • Figure 8 illustrates how the residual enhancement image may be subdivided into spatial zones made of macroblocks, with entry points in order to allow efficient random access compliant coding.
  • the position of the entry points may be encoded in the header of the bit- stream 20 in order to facilitate easy extraction from the server side and allow the reconstruction of a valid stream on the decoder side.
  • Figure 9 shows the meta-organization of a bit-stream header.
  • the slice header shown in the Figure re-uses the slice header of the H.264/AVC video compression standard.
  • coded slice length a new field which indicates the length in bytes of the coded slice. The entry points can therefore be easily computed from the "coded slice length" fields.
  • Another advantage of this independence between macroblocks of the residual enhancement image is the possibility to perform parallel entropy decoding on the decoder side. Each decoding thread starts decoding from one of the entry points as defined above.
  • Figure 4 illustrates the associated enhancement video decoder 33.
  • the channel models are reconstructed, meaning that a probabilistic distribution GGD(ai, i) is known for each encoded DCT coefficient of a channel / ' .
  • Quantizers are chosen 330 from the pool of quantizers, possibly based on these probabilistic distributions.
  • an entropy decoder 331 is applied to the received enhancement layer bit-stream 20 to obtain the quantized DCT image X DEC .
  • conventional Huffman codes can be used, possibly taking into account the probabilistic distributions;
  • dequantization (or inverse quantization) 332 is then performed by using the chosen quantizers for each coefficient, to obtain a dequantized version of the
  • the dequantized version is referenced X c , since it is different from the original version X DCT due to the lossy quantization.
  • the present invention particularly focuses on the rest of the decoding process, from the dequantized DCT coefficients X c of that dequantized image, as described now with reference to Figure 5.
  • the decoding method according to the invention comprises a step of merging 38 the dequantized DCT blocks X ⁇ c of dequantized
  • DCT coefficients with DCT residual blocks (or "predictor blocks") Y.
  • the DCT residual blocks Y are generated by an enhancement prediction module 40 which is further detailed below.
  • the DCT blocks X form a first version of the residual enhancement image currently decoded, while the DCT residual blocks Y form, at least partly, a second version of the same residual enhancement image, that is temporally predicted based on base layer motion information and an already decoded UHD image of the video, as explained below.
  • the merger of the blocks X c with the blocks Y may be a probabilistic merging process that is based on the parameters 21 (i.e. the probabilistic distributions of the DCT coefficients as determined by the encoder) and on a second probabilistic distribution that characterizes the temporal prediction of the enhancement layer by the module 40.
  • the second probabilistic distribution is a probabilistic distribution of the differences between the coefficients of the DCT residual blocks Y and the dequantized DCT coefficients of the dequantized DCT blocks X ⁇ 0 .
  • Figure 10 illustrates the generation of the DCT residual blocks Y, i.e. of transformed residual blocks of the enhancement image associated with a current image I to decode.
  • This prediction successively consists in temporally predicting current enhancement image in the pixel domain (thanks to up-sampled motion information), computing the pixel difference data between temporal predicted image and up-sampled reconstructed base image and then applying a DCT transform on the difference image.
  • This motion information comprises, for each block or macroblock, a base motion field BMF (including a motion vector and a reference image index) and a base residual BR, as well-known by one skilled in the art of video coding.
  • BMF base motion field
  • BR base residual BR
  • an image l B of the base layer may be obtained by interpolating other decoded base images.
  • the available motion information for those other decoded base images may also be interpolated to provide motion information specific to blocks or macroblocks of the interpolated base image l B .
  • the following explanation also applies for such kind of base image l B .
  • the corresponding motion information is up-sampled 400 into high resolution corresponding to the resolution of the enhancement layer (e.g. UHD). It is shown in the Figure by the references UMF (up-sampled motion field) and UR (up- sampled residual).
  • UMF up-sampled motion field
  • UR up- sampled residual
  • this up-sampling comprises for each base macroblock:
  • up-sampling the partition consists in multiplying the width and height of macroblock partitions by a factor of 2;
  • This texture up-sampling process may use an interpolation filter that is identical to that used in inter-layer residual prediction mechanisms of the SVC scalable video compression standard;
  • the generation of a DCT residual macroblock Y comprises a motion compensated prediction step 405 from the decoded UHD image that temporally corresponds to the reference base image l R B used for the decoding of the base layer, and based on the up-sampled motion information UMF and UR. That decoded UHD image is considered, for the temporal prediction, as the reference decoded image l R UHD .
  • This motion compensation 405 in the pixel domain, leads to obtaining, using the motion information, motion predictor blocks from the decoded reference high resolution image l R UHD .
  • the up-sampled prediction information is applied to the reference decoded image l R UHD to determine predicted macroblocks.
  • the motion compensation results in a partially- reconstructed image. This is because the macroblocks reconstructed by prediction are obtained at spatial positions corresponding to INTER macroblocks in the base image l B only (because there is no motion information for other macroblocks). In other words, there is no predicted block that is generated for the macroblocks collocated with INTRA macroblocks in the base layer.
  • residual blocks are obtained by subtracting 410 each motion predictor block from a corresponding (i.e. collocated) up-sampled block in the up-sampled decoded base image (which is obtained by the up-sampling 32 of Figure 2).
  • This step calculates the difference image (or residual) between the temporally predicted image and the up-sampled reconstructed base layer image.
  • This difference image has the same nature as the residual enhancement image.
  • the module 40 ends by applying 415 a block-based transform, e.g. DCT on 8x8 blocks, on the obtained residual blocks to obtain transformed residual blocks that are the DCT residuals Y discussed above.
  • a block-based transform e.g. DCT on 8x8 blocks
  • a plurality of DCT residual macroblocks Y is obtained for the current image to decode, which generally represent a partial predicted enhancement image.
  • next steps of the decoding method according to the invention may be applied to the entirety of that plurality of macroblocks Y, or to a part of it depending for example on the base coding mode (P image Inter prediction, B image Inter prediction, Skip mode) in which case only the DCT predictor macroblocks Y and the dequantized
  • DCT macroblocks X ⁇ c collocated with base macroblocks having the same coding mode are handled together.
  • macroblocks Y collocated with P, B and SKIP base macroblocks are considered separately, as was done at the encoder when determining the probabilistic distribution of each DCT channel.
  • a probabilistic distribution may be obtained for the entire set of coefficients of the considered DCT residual macroblocks, or for each DCT channel / ' in which case the explanation below should be applied for the DCT coefficients of the same channel.
  • Each DCT residual macroblock Y made of DCT coefficients for the current image to decode is considered as a version of the original DCT coefficients that would have been altered through a communication channel. It has been observed that the quantity ⁇ - ⁇ DCT (i.e. the noise of the residual Y compared to the DCT coefficients before encoding) can be well modelled by a generalized Gaussian distribution as introduced above: DCT (Y - X DCT ) « GGD(a N ,fi N )
  • the modelling of the predictor noise thus comprises fitting a Generalized Gaussian Distribution model onto the differences between the coefficients in the transformed residual blocks Y and the dequantized transformed coefficients vDEC
  • the same mechanisms based on the first and second moments as described above can be applied to obtain the two parameters ⁇ ⁇ , ⁇ (either for all the considered macroblocks Y and X ⁇ E S C , or for each DCT channel of coefficients in those macroblocks).
  • the merged value can take the form of a probabilistic estimation of the original DCT coefficients value, given the known quantization interval of this DCT coefficient, and an aside approximation of the coefficient resulting from its motion compensated temporal prediction (blocks Y).
  • a merged value according to the invention denoted x t , may be the expectation (or the "expected value") of the considered coefficient given the quantization interval Q m associated with the value of the corresponding quantized transformed coefficient in X DEC and given its corresponding value Y 0 in the residual blocks Y.
  • the quantization interval Q m is directly retrieved from the quantized DCT coefficient obtained from the bit-stream 20, since its value a m is the index of the quantization interval Q m given the quantizer used.
  • the two distributions are multiplied over the interval Q m to calculate the desired conditional expected value of X DCT .
  • the integrals of those distributions can be computed using Riemann sums over the quantization interval.
  • the values x i calculated for the DCT coefficients of all the considered macroblocks are stored in memory to form, at least partially, the merged enhancement image corresponding to the current image to decode.
  • Figure 12 illustrates the performance of the present invention, in which the rate-distortion curves are plotted when the merging according to the invention is respectively not implemented and implemented.
  • the Figure shows that an improvement the codec rate distortion performance is obtained, especially at low bitrates. This may be understood intuitively, since the quantization intervals get larger as the bitrate decreases, therefore increasing the relevant information brought by the temporal DCT residuals Y compared to the quantization level obtained by the dequantization step 332.
  • the invention also works for zero bitrate (meaning that no enhancement layer bitstream 20 is encoded or received by the decoder).
  • the parameters 21 ( ,, ⁇ for each DCT channel) are received and are used with the parameters ⁇ ⁇ , ⁇ ⁇ calculated with the present invention to obtain an improvement of the decoding quality of the base layer by several dBs.
  • the above performance is obtained with no complexity cost at the encoding side and with no additional bitrate when the parameters 21 are already needed and transmitted (e.g. for selecting the quantizers and/or entropy decoding).
  • the complexity increase due to the merging step remains reasonable at the decoding side.
  • a device implementing the invention is for example a microcomputer 50, a workstation, a personal digital assistant, or a mobile telephone connected to various peripherals.
  • the device is in the form of a photographic apparatus provided with a communication interface for allowing connection to a network.
  • the peripherals connected to the device comprise for example a digital camera 64, or a scanner or any other image acquisition or storage means, connected to an input/output card (not shown) and supplying image data to the device.
  • the device 50 comprises a communication bus 51 to which there are connected:
  • CPU 52 taking for example the form of a microprocessor
  • a read only memory 53 in which may be contained the programs whose execution enables the methods according to the invention. It may be a flash memory or EEPROM;
  • RAM random access type
  • This RAM memory 54 stores in particular the various images and the various blocks of pixels as the processing is carried out (transform, quantization, storage of the reference images) on the video sequences;
  • a hard disk 58 or a storage memory, such as a memory of compact flash type, able to contain the programs of the invention as well as data used or produced on implementation of the invention;
  • an optional diskette drive 59 or another reader for a removable data carrier, adapted to receive a diskette 63 and to read/write thereon data processed or to process in accordance with the invention
  • a communication interface 60 connected to the telecommunications network 61 , the interface 60 being adapted to transmit and receive data.
  • the device 50 is preferably equipped with an input/output card (not shown) which is connected to a microphone 62.
  • the communication bus 51 permits communication and interoperability between the different elements included in the device 50 or connected to it.
  • the representation of the bus 51 is non-limiting and, in particular, the central processing unit 52 unit may communicate instructions to any element of the device 50 directly or by means of another element of the device 50.
  • the diskettes 63 can be replaced by any information carrier such as a compact disc (CD-ROM) rewritable or not, a ZIP disk or a memory card.
  • CD-ROM compact disc
  • an information storage means which can be read by a micro-computer or microprocessor, integrated or not into the device for processing a video sequence, and which may possibly be removable, is adapted to store one or more programs whose execution permits the implementation of the method according to the invention.
  • the executable code enabling the coding device to implement the invention may equally well be stored in read only memory 53, on the hard disk 58 or on a removable digital medium such as a diskette 63 as described earlier.
  • the executable code of the programs is received by the intermediary of the telecommunications network 61 , via the interface 60, to be stored in one of the storage means of the device 50 (such as the hard disk 58) before being executed.
  • the central processing unit 52 controls and directs the execution of the instructions or portions of software code of the program or programs of the invention, the instructions or portions of software code being stored in one of the aforementioned storage means.
  • the program or programs which are stored in a non-volatile memory for example the hard disk 58 or the read only memory 53, are transferred into the random-access memory 54, which then contains the executable code of the program or programs of the invention, as well as registers for storing the variables and parameters necessary for implementation of the invention.
  • the device implementing the invention or incorporating it may be implemented in the form of a programmed apparatus.
  • a device may then contain the code of the computer program(s) in a fixed form in an application specific integrated circuit (ASIC).
  • ASIC application specific integrated circuit
  • the device described here and, particularly, the central processing unit 52 may implement all or part of the processing operations described in relation with Figures 1 to 19, to implement methods according to the present invention and constitute devices according to the present invention.
  • the above examples are merely embodiments of the invention, which is not limited thereby.
  • the calculation of the DCT residual macroblocks Y is adapted, meaning that they must be predicted from the correct decoded UHD image temporally corresponding to the selected reference image for the base layer.
  • the base image is an image bi-directionally predicted from decoded reference base images using motion information in each of the two directions
  • it may be decided to calculate separately transformed residual blocks Y, Y' for each direction, in which case the merging consists in merging together these two transformed residual blocks with the dequantized DCT blocks X ⁇ c .
  • the parameters (e.g. the filter strength parameter or the quantization-dependent parameter) of the deblocking filter 36 depends on the first and second probabilistic distributions ⁇ ( ⁇ ,, ⁇ ) and GGD(a N ,p N ) used during the merging.
  • the merging according to the invention may increase or modify the conventional blocking artefact due to the quantization.
  • adjusting the deblocking filter based on the probabilistic distributions that drives the merging permits to optimise this filtering.
  • the difference between the two blocks is high in the DCT level, then it may be decided to strengthen the deblocking filter process between the two blocks. This may be done by specifying a higher filter strength parameter, according to the deblocking filtering process of the H.264/AVC standard.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

La présente invention se rapporte à la compression vidéo d'une vidéo extensible 11. Lorsque la couche de base 14 est décodée par prédiction temporelle et que la couche d'amélioration 20 n'a pas de prédiction temporelle, le décodage de la couche d'amélioration consiste : à décoder 331, à partir du train de bits EBS, et à déquantifier 332 une formule de coefficients DCT (I) de la couche d'amélioration ; à utiliser les données de mouvement Ml de la couche de base dans le but de prédire 405 des coefficients de la couche d'amélioration ; et à transformer les coefficients prédits en coefficients DCT prédits Y ; à utiliser des paramètres α,β du train de bits dans le but d'obtenir une distribution probabilistique GGD(αii) des coefficients DCT ; à obtenir une distribution probabilistique GGD(αnn) des différences entre la formule de coefficients DCT déquantifiée (I) et les coefficients DCT prédits Y ; et à fusionner 38 ces coefficients, sur la base des distributions probabilistiques obtenues. Il est possible, de cette manière, d'améliorer le rapport entre distorsion et rendement de la vidéo extensible codée.
PCT/EP2012/062586 2011-06-30 2012-06-28 Procédé pour décoder un train de bits vidéo extensible, et dispositif de décodage correspondant WO2013001013A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB1111198.6 2011-06-30
GB201111198A GB2492396A (en) 2011-06-30 2011-06-30 Decoding a Scalable Video Bit-Stream

Publications (1)

Publication Number Publication Date
WO2013001013A1 true WO2013001013A1 (fr) 2013-01-03

Family

ID=44511903

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2012/062586 WO2013001013A1 (fr) 2011-06-30 2012-06-28 Procédé pour décoder un train de bits vidéo extensible, et dispositif de décodage correspondant

Country Status (2)

Country Link
GB (1) GB2492396A (fr)
WO (1) WO2013001013A1 (fr)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104661034A (zh) * 2015-02-11 2015-05-27 北京大学 基于视频的局部特征描述子的压缩方法及系统
US10368078B2 (en) 2013-07-15 2019-07-30 Sony Corporation Extensions of motion-constrained tile sets SEI message for interactivity
CN110855994A (zh) * 2013-04-05 2020-02-28 Vid拓展公司 针对多层视频编码的层间参考图片增强的设备
CN111641832A (zh) * 2019-03-01 2020-09-08 杭州海康威视数字技术股份有限公司 编码方法、解码方法、装置、电子设备及存储介质
CN112053285A (zh) * 2020-09-01 2020-12-08 平安科技(深圳)有限公司 图像处理方法、装置、计算机设备及存储介质
CN113096019A (zh) * 2021-04-28 2021-07-09 中国第一汽车股份有限公司 图像重建方法、装置、图像处理设备及存储介质

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6150134B2 (ja) * 2014-03-24 2017-06-21 ソニー株式会社 画像符号化装置および方法、画像復号装置および方法、プログラム、並びに記録媒体
JP6642427B2 (ja) * 2014-06-30 2020-02-05 ソニー株式会社 情報処理装置および方法
WO2021001030A1 (fr) * 2019-07-02 2021-01-07 Telefonaktiebolaget Lm Ericsson (Publ) Traitement d'inférence de données

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6700933B1 (en) * 2000-02-15 2004-03-02 Microsoft Corporation System and method with advance predicted bit-plane coding for progressive fine-granularity scalable (PFGS) video coding

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1911290A4 (fr) * 2005-07-08 2010-04-28 Lg Electronics Inc Procede permettant de modeliser les informations codees d'un signal video pour la compression/decompression des informations codees
GB2440004A (en) * 2006-07-10 2008-01-16 Mitsubishi Electric Inf Tech Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6700933B1 (en) * 2000-02-15 2004-03-02 Microsoft Corporation System and method with advance predicted bit-plane coding for progressive fine-granularity scalable (PFGS) video coding

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
BRUNO MACCHIAVELLO ET AL: "A STATISTICAL MODEL FOR A MIXED RESOLUTION WYNER-ZIV FRAMEWORK", 26. PICTURE CODING SYMPOSIUM; LISBON, 7 November 2007 (2007-11-07), XP030080372 *
DEBARGHA MUKHERJEE ET AL: "A simple reversed-complexity Wyner-Ziv video coding mode based on a spatial reduction framework", VISUAL COMMUNICATIONS AND IMAGE PROCESSING; SAN JOSE, 30 January 2007 (2007-01-30), XP030081145 *
LASSERRE S ET AL: "Low Complexity Scalable Extension of HEVC intra pictures based on content statistics", 9. JCT-VC MEETING; GENEVA; (JOINT COLLABORATIVE TEAM ON VIDEO CODING OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16), no. JCTVC-I0190, 26 April 2012 (2012-04-26), XP030052774 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110855994A (zh) * 2013-04-05 2020-02-28 Vid拓展公司 针对多层视频编码的层间参考图片增强的设备
CN110855994B (zh) * 2013-04-05 2023-09-19 交互数字麦迪逊专利控股公司 针对多层视频编码的层间参考图片增强的设备
US11553190B2 (en) 2013-07-15 2023-01-10 Sony Corporation Extensions of motion-constrained tile sets SEI message for interactivity
US10368078B2 (en) 2013-07-15 2019-07-30 Sony Corporation Extensions of motion-constrained tile sets SEI message for interactivity
US10841592B2 (en) 2013-07-15 2020-11-17 Sony Corporation Extensions of motion-constrained tile sets sei message for interactivity
CN104661034A (zh) * 2015-02-11 2015-05-27 北京大学 基于视频的局部特征描述子的压缩方法及系统
CN104661034B (zh) * 2015-02-11 2018-03-02 北京大学 基于视频的局部特征描述子的压缩方法及系统
CN111641832A (zh) * 2019-03-01 2020-09-08 杭州海康威视数字技术股份有限公司 编码方法、解码方法、装置、电子设备及存储介质
CN111641832B (zh) * 2019-03-01 2022-03-25 杭州海康威视数字技术股份有限公司 编码方法、解码方法、装置、电子设备及存储介质
CN112053285A (zh) * 2020-09-01 2020-12-08 平安科技(深圳)有限公司 图像处理方法、装置、计算机设备及存储介质
CN112053285B (zh) * 2020-09-01 2024-05-31 平安科技(深圳)有限公司 图像处理方法、装置、计算机设备及存储介质
CN113096019A (zh) * 2021-04-28 2021-07-09 中国第一汽车股份有限公司 图像重建方法、装置、图像处理设备及存储介质
CN113096019B (zh) * 2021-04-28 2023-04-18 中国第一汽车股份有限公司 图像重建方法、装置、图像处理设备及存储介质

Also Published As

Publication number Publication date
GB2492396A (en) 2013-01-02
GB201111198D0 (en) 2011-08-17

Similar Documents

Publication Publication Date Title
US10666938B2 (en) Deriving reference mode values and encoding and decoding information representing prediction modes
US10791333B2 (en) Video encoding using hierarchical algorithms
US8553769B2 (en) Method and device for improved multi-layer data compression
EP2428042B1 (fr) Procédé de codage vidéo évolutif, codeur et programme informatique
WO2013001013A1 (fr) Procédé pour décoder un train de bits vidéo extensible, et dispositif de décodage correspondant
EP3700213B1 (fr) Procédé et appareil pour coder ou décoder une image avec une prédiction d'informations de mouvement inter-couche selon un système de compression d'informations de mouvement
US10931945B2 (en) Method and device for processing prediction information for encoding or decoding an image
US9521412B2 (en) Method and device for determining residual data for encoding or decoding at least part of an image
KR102518627B1 (ko) 이미지 인코딩 방법, 디코딩방법, 인코더 및 디코더
CA2763489C (fr) Procede et dispositif de compression de donnees multicouches ameliore
WO2021136375A1 (fr) Procédés et appareils de codage ou de décodage vidéo liés à la signalisation d'information de haut niveau
EP2901690A1 (fr) Appareil, procédé et programme informatique pour codage et décodage vidéo
US20150063436A1 (en) Method for encoding and decoding an image, and corresponding devices
WO2013000575A1 (fr) Procédés et dispositifs pour un codage vidéo extensible
US20130177075A1 (en) Weighted Prediction Method and Apparatus in Quantization Matrix Coding
GB2492394A (en) Image block encoding and decoding methods using symbol alphabet probabilistic distributions
US20130230102A1 (en) Methods for encoding and decoding an image, and corresponding devices
EP3672241A1 (fr) Procédé, appareil et produit programme informatique pour codage et décodage vidéo
US11736731B2 (en) Encoding and decoding a sequence of pictures
WO2023237809A1 (fr) Procédé, appareil et produit-programme informatique de codage et de décodage vidéo
WO2023194651A1 (fr) Procédé, appareil et produit-programme informatique de codage vidéo et de décodage vidéo
WO2023126568A1 (fr) Procédé, appareil et produit programme informatique de codage et de décodage vidéo
GB2492395A (en) Entropy encoding and decoding methods using quantized coefficient alphabets restricted based on flag magnitude
CN113132734A (zh) 一种编码、解码方法、装置及其设备
WO2013000973A2 (fr) Procédé de codage et de décodage d'une image, et dispositifs correspondants

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12729663

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12729663

Country of ref document: EP

Kind code of ref document: A1