WO2015007604A1 - Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device - Google Patents

Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device Download PDF

Info

Publication number
WO2015007604A1
WO2015007604A1 PCT/EP2014/064817 EP2014064817W WO2015007604A1 WO 2015007604 A1 WO2015007604 A1 WO 2015007604A1 EP 2014064817 W EP2014064817 W EP 2014064817W WO 2015007604 A1 WO2015007604 A1 WO 2015007604A1
Authority
WO
WIPO (PCT)
Prior art keywords
blocks
group
information
region
image
Prior art date
Application number
PCT/EP2014/064817
Other languages
French (fr)
Inventor
Philippe Bordes
Patrick Lopez
Mehmet TURKAN
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing filed Critical Thomson Licensing
Priority to KR1020167001286A priority Critical patent/KR20160034900A/en
Priority to EP14744790.8A priority patent/EP3022922A1/en
Priority to CN201480040499.7A priority patent/CN105659600A/en
Priority to US14/905,960 priority patent/US9961353B2/en
Priority to JP2016526536A priority patent/JP6393323B2/en
Publication of WO2015007604A1 publication Critical patent/WO2015007604A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/34Scalability techniques involving progressive bit-plane based encoding of the enhancement layer, e.g. fine granular scalability [FGS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/36Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/96Tree coding, e.g. quad-tree coding

Definitions

  • the invention relates to the general domain of the scalable coding of an image sequence. More particularly, the invention also relates to a method for decoding a scalable stream representative of an image sequence comprising a base layer and at least one enhancement layer. The invention also relates to a coding method and the corresponding devices.
  • Scalable coding methods are known that enable the information to be coded in a hierarchical manner such that this information can be decoded at different resolutions (spatial scalability), different qualities (SNR scalability, for "Signal to Noise Ratio”), with different gamuts or colour spaces (CGS scalability, for "Colour Gamut Scalability”).
  • a stream generated by a scalable coder is divided into different layers: a base layer and one or more enhancement layers.
  • Such devices can generate a stream of data that adapts to different transmission conditions (bandwidth, error rate, etc.) and also to the capacities of the reception devices.
  • a scalable coder spatially codes a first part of data relating to low resolution images in a base layer and another part of data relating to high resolution images (i.e. having a size in number of horizontal and/or vertical pixels greater than the low resolution image) in at least one enhancement layer.
  • the data coded in the enhancement layer is generally done so by taking into account data from the base layer so as to reduce the redundancy between the low resolution images and high resolution images and thus to reduce the coding cost.
  • the I BL images re- sampled in this manner are used to predict the I EL images with a view to their coding 14.
  • the different layers can possibly be multiplexed 1 6 into a single scalable stream S.
  • the scalable coding is not limited to spatially scalable coding.
  • scalable coders are known that code, in a base layer, data relating to images represented in a first colour space and that code, in at least one enhancement layer, data relating to images represented in a second colour space different from the first colour space.
  • the images of the base layer reconstructed at least partially are then processed so that their characteristics in terms of colour are close to those of the images of the enhancement layer.
  • the I BL images processed in this manner are used to predict the I EL images with a view to their coding.
  • an image of the base layer is fully processed before being stored in the reference image buffer or DPB (decoding picture buffer) of the enhancement layer.
  • the processing of an image of the base layer is, for example, carried out just before the decoding of an image of the enhancement layer that it is used to predict.
  • Such a solution requires the processing of the entire image of the reconstructed base layer which can be costly in terms of calculations.
  • the coding mode of the block indicates it. It is therefore possible, when decoding such a block of the enhancement layer to check the coding mode of this block and only process the region of the image of the reconstructed base layer necessary for reconstructing the relevant block. In this case, the processing of the image of the reconstructed base layer necessary for reconstructing the relevant block is carried out just after the decoding of the coding mode of the block.
  • This solution is economic in terms of calculations as only the regions of reconstructed images of the base layer necessary for reconstructing the images of the enhancement layer are processed.
  • this solution requires many accesses to the buffer storing the images of the reconstructed base layer and many calls to the processing function. Indeed, the processing function is called as many times as a block whose coding mode indicates that it is predicted in relation to a base layer image (e.g. inter-layer mode) is decoded.
  • the purpose of the invention is to overcome at least one of the disadvantages of the prior art.
  • a method for decoding a scalable stream comprising a base layer and at least one enhancement layer is described. The method comprises:
  • This decoding method advantageously reduces the number of accesses to the processing function but also processes only the useful regions of the image of the reconstructed base layer.
  • the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the reconstructed image corresponding to a block of the group of blocks if the block must be processed or not.
  • the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the group of blocks if the block uses for its reconstruction a processed block of the reconstructed image.
  • the base layer is another enhancement layer of a lower level than the at least one enhancement layer.
  • the method comprises a decoding step from the enhancement layer of an additional item of information for at least one group of blocks indicating whether the information for identifying at least one region of the reconstructed image to process is coded or not for the group of blocks.
  • the group of blocks is an image slice or an image.
  • the decoding method comprises a decoding step from the enhancement layer of a first additional item of information indicating whether the information for identifying at least one region of the reconstructed image to process can be coded or not for the group of blocks.
  • the method comprises, when the first item of additional information indicates that an item of information for identifying at least one region of the reconstructed image to process can be coded, a decoding step from the enhancement layer of a second item of additional information indicating for a group of blocks whether the information for identifying at least one region of the reconstructed image is coded or not for the group of blocks.
  • a coding method of a sequence of images in the form of a scalable stream comprising a base layer and at least one enhancement layer is also described.
  • the coding method comprises:
  • the enhancement data comprising an item of information for identifying at least one region of the reconstructed image to process
  • the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the reconstructed image corresponding to a block of the group of blocks if the block must be processed or not.
  • the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the group of blocks if the block uses for its reconstruction a processed block of the reconstructed image.
  • the base layer is another enhancement layer of a lower level than the at least one enhancement layer.
  • the method comprises a coding step in the enhancement layer of an additional item of information for at least one group of blocks indicating whether the information for identifying at least one region of the reconstructed image to process is coded or not for the group of blocks.
  • the coding method comprises a coding step in the enhancement layer of a first additional item of information indicating whether the information for identifying at least one region of the reconstructed image to process can be coded or not for a group of blocks.
  • the method comprises, when the first item of additional information indicates that an item of information for identifying at least one region of the reconstructed image to process can be coded, a coding step in the enhancement layer of a second item of additional information indicating for a group of blocks whether the information for identifying at least one region of the reconstructed image is coded or not for the group of blocks.
  • a decoder configured for decoding a scalable stream comprising a base layer and at least one enhancement layer is also described. The decoder comprises:
  • the decoder is configured to execute the steps of the decoding method according to one of the embodiments.
  • a coder configured to code a sequence of images in the form of a scalable stream comprising a base layer and at least one enhancement layer is described which comprises:
  • enhancement data in a first part of the enhancement layer relating to a group of blocks, the enhancement data comprising an item of information for identifying at least one region of the reconstructed image to process
  • the coder is configured to execute the steps of the coding method according to one of the embodiments described.
  • a coded data stream representative of a sequence of images comprising a base layer and at least one enhancement layer, the base layer comprising data relative to at least one image of the base layer and the enhancement layer comprising in a first part relating to a group of blocks enhancement data which comprise an item of information for identifying at least one region of the image of the base layer to process after reconstruction and in a second part data representative of residues which are obtained from at least one block of the group and from the at least one region processed.
  • a computer program product comprises program code instructions for executing the steps of the decoding method according to one of the embodiments, when the program is executed on a computer.
  • a computer program product is described that comprises program code instructions for executing the steps of the coding method according to one of the embodiments, when the program is executed on a computer.
  • FIG. 2 shows a decoding method according to a particular and non- restrictive embodiment of the invention
  • FIGS. 3 and 4 show an item of information info in a specific map form to identify at least one region of an image of the reconstructed base layer to process
  • FIG. 8 shows a coding method according to a particular and non- restrictive embodiment of the invention
  • the invention relates to a device for decoding a scalable stream representative of an image sequence comprising a base layer and at least one enhancement layer.
  • An image sequence is a series of several images. Each image comprises pixels or image points, with each of which is associated at least one item of image data.
  • An item of image data is for example an item of luminance data or an item of chrominance data.
  • the coding and decoding methods are described with reference to blocks of pixels. It is clear that these methods can be applied to several blocks of an image and to several images of a sequence with a view to coding respectively the reconstruction of one or more images.
  • a scalable stream represents an image sequence in different versions, for example, different resolutions, different gamuts/colour spaces, etc.
  • motion data is to be understood in the widest sense. It signifies the motion vectors and possibly the reference image indexes enabling a reference image to be identified in the image sequence. It can also comprise an item of information indicating the interpolation type used to determine the prediction block. In fact, in the case where the motion vector associated with a block Be does not have integer coordinates, the image data must be interpolated in the reference image Iref to determine the prediction block.
  • the motion data associated with a block is generally calculated by a motion estimation method, for example by block pairing. However, the invention is in no way limited by the method enabling a motion vector to be associated with a block.
  • residual data designates data obtained after extraction of other data.
  • the extraction is generally a subtraction pixel by pixel of prediction data from source data. However, the extraction is more general and comprises notably a weighted subtraction.
  • residual data is synonymous with the term “residue”.
  • transformed residual data designates residual data to which a transform has been applied.
  • a DCT Discrete Cosine Transform
  • Such transforms "transform” a block of image data, e.g. luminance and/or chrominance residual data, into a "block of transformed data” also called “transformed block", "frequency data block” or "block of coefficients".
  • the block of coefficients generally comprises a low-frequency coefficient known under the name of direct current coefficient or DC coefficient and high- frequency coefficients known under the name of AC coefficients.
  • prediction data designates data used to predict other data.
  • a prediction block or predictor is a block of pixels with which prediction data is associated.
  • a prediction block is obtained from a block or several blocks of the same image as the image to which belongs the block that it predicts (spatial prediction or intra-image prediction) or from one (mono-directional prediction) or several blocks (bi-directional or bi-predicted prediction) of a different image (temporal prediction or inter-image prediction) of the image to which the block that it predicts belongs.
  • a prediction block can also be obtained for a block of an enhancement layer from at least one block of an image of the base layer, e.g. the co-located block (inter-layer prediction).
  • prediction mode designates the way in which a block is predicted.
  • coding mode designates the way in which a block is coded in the stream. It includes the prediction mode if it is predicted.
  • the prediction modes there is the INTRA mode that corresponds to a spatial prediction, the INTER mode that corresponds to a temporal prediction, the INTER-LAYER mode that corresponds to an inter-layer prediction.
  • the INTER-LAYER mode can be implemented like an INTER mode if the reconstructed image of the base layer is stored in the reference buffer of the enhancement layer.
  • the standard INTER modes are thus distinguished from an INTER mode corresponding to an INTER-LAYER mode by the value of the reference index.
  • the prediction mode can specify the way in which the block is partitioned to be coded.
  • the 8x8 INTER prediction mode associated with a block of size 16x16 signifies that the 16x16 block is partitioned into 4 8x8 blocks and predicted by temporal prediction.
  • the expression "reconstructed data” designates data (for example pixels, blocks) obtained after merging residues with prediction data.
  • the merge is generally a sum of prediction data with residues. However, merging is more general and comprises notably the weighted sum.
  • a reconstructed block is a block of reconstructed pixels.
  • a "causal neighbouring area" of a current block designates a neighbouring area of this block that comprises pixels codes/reconstructed before those of the current block.
  • the term coding is to be taken in the widest sense.
  • the coding can possibly comprise the transformation and/or the quantization of image data. It can also designate the entropy coding only.
  • Figure 2 shows a decoding method of a scalable stream S comprising a base layer and at least one enhancement layer according to a particular and non- restrictive embodiment of the invention.
  • the invention described for a base layer and an enhancement layer can be applied to two enhancement layers of different levels, the lower enhancement layer playing the role of a base layer with respect to the upper level layer.
  • At least one image l B i_ of the base layer is reconstructed from the base layer.
  • the base layer is decoded at least partially by entropy decoding, for example by CABAC decoding ("Context- adaptive binary arithmetic coding") or CAVLC (“Context-adaptive variable- length coding”) decoding.
  • the decoded data notably comprises data representative of residues, for example quantized DCT coefficients, coding modes, possibly motion data.
  • the reconstruction of the image lBi_ for a current block of the image comprises the obtaining of a spatial or temporal predictor according to the coding mode of the block.
  • the predictor is obtained from a part of an image previously reconstructed either in the same image as the current block to reconstruct (spatial prediction) or in another image (temporal prediction) called reference image.
  • the predictor is added to residues to reconstruct the current block.
  • the residues are, for example, obtained by inverse quantization and inverse transformation of data relating to residues obtained during the entropy decoding of the base layer.
  • the term inverse is used to signify that the quantization and transformation operations on the decoder are inverse from those used on the coder.
  • no predictor is added to the residues in the specific case where the current block is coded without prediction (e.g. according to the I PCM Intra mode of H.264).
  • the residues correspond in the latter case directly to the reconstructed current block.
  • the residues are null. This is particularly the case for a current block coded according to a particular coding mode called skip mode.
  • the reconstructed current block is the prediction block in this case.
  • the reconstruction process described for a current block is reiterated to reconstruct the image I BL of the base layer. Naturally, this process can also be reiterated to reconstruct all the images of the base layer.
  • a first part EL1 of the enhancement layer relating to a group of blocks is decoded into decoded enhancement data, the group of blocks belonging to an image of the enhancement layer to reconstruct.
  • the part EL1 is decoded by entropy decoding, e.g. by CABAC decoding.
  • the decoded enhancement data thus comprise information relating to the group of blocks of the image of the enhancement layer to reconstruct.
  • a group of blocks is for example an image or an image slice.
  • this enhancement data is obtained by decoding an image slice header, a PPS (Picture Parameter Set) or even a message SEI (Supplemental Enhancement Information).
  • PPS and SEI are notably defined in paragraph 3.107 (definition) and the appendix D respective of the ISO/I EC document 14496-10 describing the video coding standard H.264. Indeed, such elements contain information relating to a group of blocks. However, note that an image slice can only contain a single block. However, coding images with image slices of a single block is ineffective in terms of bitrate. Slice coding, each slice containing several blocks, e.g. 2 blocks, groups common information together and thus prevents duplicating the same information in the headers of each block.
  • this decoded enhancement data comprises an item of information info for identifying at least one region of the image l B i_ reconstructed in step 20 to process.
  • the decoder From the decoding of the first part of enhancement layer EL1 relating to the group of blocks, the decoder has knowledge of the region of the reconstructed image l B i_ to process, i.e. the one that will effectively be used during the reconstruction of an image of the enhancement layer. This information can thus distinguish between the regions of the reconstructed image I BL that will be used during the reconstruction of an image of the enhancement layer and hence that must be processed and those that will not be used and hence that must not be processed.
  • This information info is presented for example in the form of a map indicating for each block of the group of blocks of the enhancement layer whether the corresponding region, e.g. the block, of the base layer must be processed or not.
  • a region of the reconstructed image l B i_ corresponds to a block of an image of the enhancement layer if it is co-located, i.e. if it occupies in the image the same spatial position possibly after processing (spatial scalability case).
  • the information info is presented for example in the form of a map indicating for each region, e.g. block, of the reconstructed image I BL corresponding to a block of the group of blocks if this region must be processed or not.
  • This information irrespective of whether it relates to the base layer or to the enhancement layer is coded in the form of a regular map, e.g. in the case of the HEVC standard for each CTU (Coding Tree Unit) or for example for each macroblock as shown in figure 3.
  • this information is coded in the form of an irregular map, e.g. in the case of the HEVC standard for each CU (Coding Unit) as shown in figure 4.
  • a value 0 indicates that the reconstructed region l B i_ is not to be processed and a value 1 indicates that it is to be processed.
  • the inverse can be chosen by convention.
  • This information is decoded by entropy decoding, for example CAVLC or CABAC.
  • the region or regions of the image of the reconstructed base layer identified by the information info decoded in step 22 is processed.
  • the processing is for example a re-sampling of the region in the case of spatial scalability, a colour conversion in the case of CGS scalability, a noise reduction filtering in the case of SNR scalability.
  • This solution advantageously enables only the useful regions of the reconstructed image l B i_ to be processed, i.e. used in the reconstruction of blocks of an image l E i_ of the enhancement layer. Moreover, by proceeding thus the useful regions of the base layer are processed before the reconstruction of the blocks of the group of blocks and are thus all available during the reconstruction of the group of blocks which avoids too frequent accesses to the image buffer of the reconstructed base layer and too many calls on the processing function.
  • residues are obtained for at least one block of the group of blocks from a second part of the enhancement layer EL2.
  • residues are obtained for each block of the group of blocks.
  • This second part comprises data representative of residues for example quantized DCT coefficients from which residues are calculated or else a particular coding mode (skip mode) indicating that these residues are null.
  • This step possibly comprises the entropy decoding, the inverse quantization and the inverse transformation of data of the second part, e.g. quantized DCT coefficients, to obtain the residues res.
  • This second part possibly comprises motion data.
  • the blocks of the group are reconstructed from residues res and from the region processed.
  • a predictor is obtained either from an image part previously reconstructed or from the same image as the current block of the group to reconstruct (spatial prediction) or from another image of the enhancement layer (temporal prediction) or even from the region of the reconstructed image IBL processed (inter-layer prediction).
  • the predictor is added to the residues of the current block to reconstruct the current block. Note that according to a variant no predictor is added to the residues of the current block, in the specific case where the current block is coded without prediction (e.g. according to the I PCM coding mode of H.264).
  • the residues correspond in the latter case directly to the reconstructed current block.
  • the coding modes can be varied, some blocks being coded according to an I PCM mode, by spatial prediction, others by temporal prediction, others by inter-layer prediction and even others according to a skip mode.
  • an additional item of information is decoded from the scalable stream, more precisely, from the enhancement layer for example of a PPS or of an image slice header that indicates for a group of blocks whether the item of information for identifying at least one region of the reconstructed image I BL to process is coded or not for this group of blocks.
  • the enhancement layer for example of a PPS or of an image slice header that indicates for a group of blocks whether the item of information for identifying at least one region of the reconstructed image I BL to process is coded or not for this group of blocks.
  • the additional information is for example a flag, i.e. a binary data.
  • a first PPS1 is decoded with the additional value equal to Flg1 , e.g.
  • each image of the enhancement layer can according to the case refer to one or other of the PPS. If an image of the enhancement layer I E L refers to the first PPS1 , then an item of information for identifying at least one region of the reconstructed image I BL to process is decoded, e.g. in each slice header of the image of the enhancement layer as shown in figure 5.
  • this flag can be coded in each image slice header, in which case if the value indicates that the item of information for identifying at least one region of the reconstructed image I B L to process is coded, this item of information for identifying at least one region of the reconstructed image I BL to process follows in the additional information stream Fig as shown in figure 6.
  • a first item of additional information Flg1 is decoded from the enhancement layer that indicates whether the information info for identifying at least one region of the reconstructed image I BL to process can be decoded or not for a group of blocks.
  • This first item of additional information Flg1 is for example decoded from the PPS. For example, if its value is 0 this signifies that no item of information for identifying at least one region of the reconstructed image I BL to process can be decoded from the stream. If its value is 1 , this signifies that an item of information for identifying at least one region of the reconstructed image I B L to process can be decoded.
  • a second item of additional information Flg2 is decoded for example in the header of each image slice that indicates for a group of blocks whether the item of information info for identifying at least one region of the reconstructed image IBLJO process is coded or not for this group of blocks.
  • this item of information info follows the second item of information in the stream.
  • Figure 8 shows a coding method of an image sequence in the form of a scalable stream comprising a base layer and at least one enhancement layer according to a particular and non-restrictive embodiment of the invention. The same variants as the ones described in relation to the decoding method apply to the coding method even in the case where they may not be described explicitly in the remainder of the description.
  • the coding generally comprises, for blocks of the image, the spatial or temporal prediction of the blocks, the obtaining of residues by subtracting a predictor from the block to code, the transformation and quantization of residues before the entropy coding.
  • the same coding modes described in relation to the decoding method can be used, particularly the skip mode and I PCM coding mode of AVC.
  • Other data is also coded such as coding modes, possibly motion data.
  • the coding comprises the reconstruction of the image of the base layer. Indeed, the latter can then be used for the prediction of other data.
  • enhancement data is coded in a first part of the enhancement layer EL1 relating to a group of blocks of the image of the enhancement layer to reconstruct.
  • a group of blocks is for example an image or a slice of images.
  • this enhancement data is coded in an image slice header, in PPS or even in an SEI message. Indeed, such elements contain information relating to a group of blocks.
  • an image slice can only contain a single block.
  • coding images with image slices of a single block is ineffective in terms of bitrate.
  • Slice coding each slice containing several blocks, e.g. 2 blocks, groups common information together and thus prevents duplicating the same information in the headers of each block.
  • this coded enhancement data comprises an item of information info for identifying at least one region of the reconstructed image I BL to process.
  • this information is obtained during a previous step of deciding on the coding mode.
  • this map can be constructed. Indeed, all the blocks of the map that correspond to an inter-layer prediction mode are predicted in relation to a block of the image of the reconstructed base layer processed. This block must then be identified in the map as belonging to the region to process.
  • the decoder has knowledge of the region or regions of the reconstructed image I BL to process, i.e. the one(s) that will effectively be used during the reconstruction of an image of the enhancement layer.
  • This information can thus distinguish between the regions of the reconstructed image I BL that will be used during the reconstruction of an image of the enhancement layer and hence that must be processed and those that will not be used and hence that must not be processed.
  • This information info is presented for example in the form of a map indicating for each block of the group of blocks of the enhancement layer whether the corresponding region, e.g. the block, of the base layer must be processed or not.
  • a region of the reconstructed image I B L corresponds to a block of an image of the enhancement layer if it is co-located, i.e. if it occupies in the image the same spatial position possibly after processing (spatial scalability case).
  • the information info is presented for example in the form of a map indicating for each region, e.g. block, of the reconstructed image I B L corresponding to a block of the group of blocks if said region must be processed or not.
  • This information irrespective of whether it relates to the base layer or to the enhancement layer is coded in the form of a regular map, e.g. in the case of the HEVC standard for each CTU (Coding Tree Unit) or for example for each macroblock as shown in figure 3.
  • this information is coded in the form of an irregular map, e.g. in the case of the HEVC standard for each CU (Coding Unit) as shown in figure 4.
  • a value 0 indicates that the reconstructed region l B i_ is not to be processed and a value 1 indicates that it is to be processed.
  • the inverse can be chosen by convention.
  • This information is coded by entropy coding, for example CAVLC or CABAC.
  • step 84 the region or regions of the image of the reconstructed base layer identified by the information info coded in step 82 is/are processed.
  • the processing is for example a re-sampling of the region in the case of spatial scalability, a colour conversion in the case of CGS scalability, a noise reduction filtering in the case of SNR scalability.
  • This solution advantageously enables only the useful regions of the reconstructed image l B i_ to be processed, i.e. used in the coding of blocks of an image I EL of the enhancement layer. Moreover, by proceeding thus the useful regions of the base layer are processed before the coding of the blocks of the group of blocks and are thus all available during the coding of the group of blocks which avoids too frequent accesses to the image buffer of the reconstructed base layer and too many calls on the processing function.
  • residues are obtained for at least one block of the group from blocks of the group and from the region processed.
  • residues are obtained for each block of the group of blocks.
  • a predictor is obtained either from an image part previously reconstructed or from the same image as the current block to code (spatial prediction) or from another image of the enhancement layer (temporal prediction) or even from the region of said reconstructed image I BL processed (inter-layer prediction).
  • the predictor is subtracted from the current block to obtain residues. Note that according to a variant no predictor is subtracted from the current block, in the case where the block was coded without prediction (e.g. I PCM coding mode of H.264).
  • the residues block corresponds in the latter case directly to the block to code.
  • the residues are null.
  • the coding mode of the block coded in the stream indicates it.
  • the coding modes can be varied, some blocks being coded according to an I PCM mode, by spatial prediction, others by temporal prediction, others by inter-layer prediction and even others according to a skip mode.
  • the residues relating to at least one block of the group of blocks are coded in a second part of the enhancement layer.
  • This step possibly comprises the transformation and quantization of the residues before their entropy coding.
  • the additional information shown in figures 5 to 7 and described in relation to the decoding method is coded in the scalable stream S. All the embodiments described in relation to the decoding method apply symmetrically to the coding method.
  • the invention also relates to a scalable stream representative of an image sequence.
  • This stream is for example produced by the coding method described previously.
  • This scalable stream comprises a base layer and at least one enhancement layer.
  • the base layer comprises data relating to at least one image of the base layer.
  • the enhancement layer comprises in a first part EL1 relating to a group of blocks enhancement data which comprise an item of information for identifying at least one region of the image of the base layer to process after its reconstruction.
  • the enhancement layer comprises in a second part EL2 data representative of residues for at least one block of the group of blocks, the residues being obtained from at least one block of the group and from the region processed.
  • the enhancement layer comprises in a second part EL2 data representative of residues for each of the blocks of the group of blocks.
  • This data comprises for example a coding mode, possibly motion data, of data relating to residues (e.g. quantized DCT coefficients) or of data enabling such residues to be determined, e.g. the skip coding mode.
  • the scalable stream comprises the additional information shown in figures 5 to 7 and described in relation to the decoding method.
  • Figure 9 shows a decoding device 9 according to a non-restrictive embodiment of the invention.
  • the decoding device 9 comprises a decoding module 90 able to decode a base layer BL of a scalable stream S and to reconstruct images of the base layer.
  • the decoding module 90 is notably configured to implement step 20 of the decoding method.
  • the decoding device 9 further comprises a processing module 92 able to process at least one region of an image of the reconstructed base layer.
  • the processing module 92 is notably configured to implement step 24 of the decoding method.
  • the decoding device 9 also comprises another decoding module 94 able to decode an enhancement layer EL of a scalable stream S and to reconstruct images of the enhancement layer.
  • the decoding module 94 is also able to supply the processing module 92 with the information info.
  • the decoding module 94 is notably configured to implement steps 22, 26 and 28 of the decoding method.
  • the decoding device 9 optionally comprises a demultiplexing module 96 configured to demultiplex the scalable stream S into a base layer BL and one or more enhancement layers EL.
  • FIG. 10 shows a particular implementation of such a decoding device 9.
  • the decoding device 9 comprises hardware components such as at least a processor 900 (e.g. CPU/GPU), a memory 902 (e.g. RAM), a user interface 904 (e.g. keyboards, mouse, etc.) and one or more input/output interfaces 906 able to be connected to other devices. These components are for example interconnected by an address and data bus 908. According to an implementation example, when powered up, the processor 900 loads the instructions corresponding to the steps of the decoding method into the memory 902 and executes them.
  • a processor 900 e.g. CPU/GPU
  • memory 902 e.g. RAM
  • user interface 904 e.g. keyboards, mouse, etc.
  • input/output interfaces 906 able to be connected to other devices.
  • these components are for example interconnected by an address and data bus 908.
  • the processor 900 when powered up, loads the instructions corresponding to the steps of the decoding method into
  • the decoding device 9 advantageously comprises an interface to read a computer program stored on a digital data support 910 that stores the program code instructions, that when they are executed by the processor 900, carry out the steps of the decoding method according to any one of the embodiments of the invention described previously.
  • the user interface is external to the decoding device 9.
  • FIG 11 shows a coding device 1 1 according to a non-restrictive embodiment of the invention.
  • the coding device 1 1 comprises a coding module 1 10 able to code and reconstruct images l B i_ in a base layer BL of a scalable stream.
  • the coding module 1 10 is notably configured to implement step 80 of the coding method.
  • the coding device 1 1 further comprises a processing module 1 12 able to process at least one part of an image of the reconstructed base layer.
  • the processing module 1 12 is notably configured to implement step 84 of the coding method.
  • the coding device 1 1 also comprises another coding module 1 14 able to code images I EL in an enhancement layer EL of the scalable stream S.
  • the coding module 1 14 is also able to supply the processing module 1 12 with the information info.
  • the coding module 1 14 is notably configured to implement steps 82, 86 and 88 of the coding method.
  • the coding device 1 1 optionally comprises a multiplexing module 1 16 configured to multiplex the base layer and enhancement layer or layers into a single scalable stream S.
  • FIG 12 shows a particular implementation of such a coding device 1 1 .
  • the coding device 1 1 comprises hardware components such as at least a processor 1 100 (e.g. CPU/GPU), a memory 1 102 (e.g. RAM), a user interface 1 104 and one or more input/output interfaces 1 106 able to be connected to other devices. These components are for example interconnected by an address and data bus 1 108.
  • the processor 1 100 when powered up, loads the instructions corresponding to the steps of the coding method into the memory 1 102 and executes them.
  • the coding device 1 1 advantageously comprises an interface to read a computer program stored on a digital data support 1 1 10 that stores the program code instructions, that when they are executed by the processor 1 100, carry out the steps of the coding method according to any one of the embodiments of the invention described previously.
  • the user interface is external to the coding device 1 1 .
  • decoding 9 and coding 1 1 devices described are simplified for reasons of clarity.
  • Real devices may comprise other elements such as for example an internal or external battery, non-volatile memory (e.g. flash memory, hard disk drive, etc.), etc.
  • decoding 9 and coding 1 1 devices compatible with the invention are implemented according to a purely hardware embodiment, for example in the form of a dedicated component (for example in an ASIC (Application Specific Integrated Circuit) or FPGA (Field-Programmable Gate Array) or VLSI (Very Large Scale Integration) or of several electronic components integrated into a device or even in a form of a combination of hardware elements and software elements.
  • ASIC Application Specific Integrated Circuit
  • FPGA Field-Programmable Gate Array
  • VLSI Very Large Scale Integration
  • the invention described for a base layer and an enhancement layer can be applied to two enhancement layers of different levels, the lower enhancement layer playing the role of a base layer with respect to the upper level layer.
  • the invention is not restricted by the type of quantizing, transform or entropy coding used.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A method for decoding a scalable stream comprising a base layer and at least one enhancement layer is described. The method comprises the following steps: - reconstructing (20) at least one image from the base layer, - decoding (22) a first part of the enhancement layer relating to a group of blocks into decoded enhancement data which comprise at least one item of information for identifying at least one region of the reconstructed image to process, - processing (24) the region of the reconstructed image identified by the information, - obtaining (26) residues for at least one block of the group of blocks from a second part of the enhancement layer, and - reconstructing (28) the blocks of the group from residues and from the region processed.

Description

METHOD AND DEVICE FOR DECODING A SCALABLE STREAM REPRESENTATIVE OF AN IMAGE SEQUENCE AND CORRESPONDING
CODING METHOD AND DEVICE 1 . Domain of the invention
The invention relates to the general domain of the scalable coding of an image sequence. More particularly, the invention also relates to a method for decoding a scalable stream representative of an image sequence comprising a base layer and at least one enhancement layer. The invention also relates to a coding method and the corresponding devices.
2. Prior art
Scalable coding methods are known that enable the information to be coded in a hierarchical manner such that this information can be decoded at different resolutions (spatial scalability), different qualities (SNR scalability, for "Signal to Noise Ratio"), with different gamuts or colour spaces (CGS scalability, for "Colour Gamut Scalability"). A stream generated by a scalable coder is divided into different layers: a base layer and one or more enhancement layers. Such devices can generate a stream of data that adapts to different transmission conditions (bandwidth, error rate, etc.) and also to the capacities of the reception devices. A scalable coder spatially codes a first part of data relating to low resolution images in a base layer and another part of data relating to high resolution images (i.e. having a size in number of horizontal and/or vertical pixels greater than the low resolution image) in at least one enhancement layer. The data coded in the enhancement layer is generally done so by taking into account data from the base layer so as to reduce the redundancy between the low resolution images and high resolution images and thus to reduce the coding cost. In reference to figure 1 , it is known to code 1 0 the low resolution images, called lBi_ images, reconstruct them at least partially and process 1 2 them, that is to say re-sample them at the size of the high resolution images, called I EL images. The I BL images re- sampled in this manner are used to predict the I EL images with a view to their coding 14. The different layers can possibly be multiplexed 1 6 into a single scalable stream S. The scalable coding is not limited to spatially scalable coding. Hence, scalable coders are known that code, in a base layer, data relating to images represented in a first colour space and that code, in at least one enhancement layer, data relating to images represented in a second colour space different from the first colour space. The images of the base layer reconstructed at least partially are then processed so that their characteristics in terms of colour are close to those of the images of the enhancement layer. The I BL images processed in this manner are used to predict the I EL images with a view to their coding.
Generally, an image of the base layer is fully processed before being stored in the reference image buffer or DPB (decoding picture buffer) of the enhancement layer. The processing of an image of the base layer is, for example, carried out just before the decoding of an image of the enhancement layer that it is used to predict. Such a solution requires the processing of the entire image of the reconstructed base layer which can be costly in terms of calculations.
According to a variant, when a block of the enhancement layer is reconstructed from an image of the base layer, the coding mode of the block indicates it. It is therefore possible, when decoding such a block of the enhancement layer to check the coding mode of this block and only process the region of the image of the reconstructed base layer necessary for reconstructing the relevant block. In this case, the processing of the image of the reconstructed base layer necessary for reconstructing the relevant block is carried out just after the decoding of the coding mode of the block. This solution is economic in terms of calculations as only the regions of reconstructed images of the base layer necessary for reconstructing the images of the enhancement layer are processed. However, this solution requires many accesses to the buffer storing the images of the reconstructed base layer and many calls to the processing function. Indeed, the processing function is called as many times as a block whose coding mode indicates that it is predicted in relation to a base layer image (e.g. inter-layer mode) is decoded.
3. Summary of the invention The purpose of the invention is to overcome at least one of the disadvantages of the prior art. For this purpose, a method for decoding a scalable stream comprising a base layer and at least one enhancement layer is described. The method comprises:
- reconstructing at least one image from the base layer,
- decoding a first part of the at least one enhancement layer relating to a group of blocks into decoded enhancement data which comprise at least one item of information for identifying at least one region of the reconstructed image to process,
- processing the at least one region of the reconstructed image identified by the information,
- obtaining residues for at least one block of the group of blocks from a second part of the at least one enhancement layer, and
- reconstructing the blocks of the group from residues and from the at least one region processed.
This decoding method advantageously reduces the number of accesses to the processing function but also processes only the useful regions of the image of the reconstructed base layer.
According to a particular embodiment, the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the reconstructed image corresponding to a block of the group of blocks if the block must be processed or not.
According to a variant, the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the group of blocks if the block uses for its reconstruction a processed block of the reconstructed image.
According to a particular characteristic of the invention, the base layer is another enhancement layer of a lower level than the at least one enhancement layer.
Advantageously, the method comprises a decoding step from the enhancement layer of an additional item of information for at least one group of blocks indicating whether the information for identifying at least one region of the reconstructed image to process is coded or not for the group of blocks. According to a particular characteristic of the invention, the group of blocks is an image slice or an image.
According to another aspect of the invention, the decoding method comprises a decoding step from the enhancement layer of a first additional item of information indicating whether the information for identifying at least one region of the reconstructed image to process can be coded or not for the group of blocks.
According to a particular characteristic of the invention, the first item of additional information being decoded for at least one image, the method comprises, when the first item of additional information indicates that an item of information for identifying at least one region of the reconstructed image to process can be coded, a decoding step from the enhancement layer of a second item of additional information indicating for a group of blocks whether the information for identifying at least one region of the reconstructed image is coded or not for the group of blocks.
A coding method of a sequence of images in the form of a scalable stream comprising a base layer and at least one enhancement layer is also described. The coding method comprises:
- coding in the base layer and reconstructing at least one image,
- coding enhancement data in a first part of the enhancement layer relating to a group of blocks, the enhancement data comprising an item of information for identifying at least one region of the reconstructed image to process,
- processing the at least one region of the reconstructed image identified by the information,
- obtaining residues for at least one block of the group of blocks from the blocks of the group and from the at least one region processed, and
- coding the residues in a second part of the enhancement layer.
According to a particular embodiment, the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the reconstructed image corresponding to a block of the group of blocks if the block must be processed or not. According to a variant, the information for identifying at least one region of the reconstructed image to process is a map indicating for each block of the group of blocks if the block uses for its reconstruction a processed block of the reconstructed image.
According to a particular characteristic of the invention, wherein the base layer is another enhancement layer of a lower level than the at least one enhancement layer.
Advantageously, the method comprises a coding step in the enhancement layer of an additional item of information for at least one group of blocks indicating whether the information for identifying at least one region of the reconstructed image to process is coded or not for the group of blocks.
As a variant, the coding method comprises a coding step in the enhancement layer of a first additional item of information indicating whether the information for identifying at least one region of the reconstructed image to process can be coded or not for a group of blocks.
According to a particular characteristic of the invention, the first item of additional information being coded for at least one image, the method comprises, when the first item of additional information indicates that an item of information for identifying at least one region of the reconstructed image to process can be coded, a coding step in the enhancement layer of a second item of additional information indicating for a group of blocks whether the information for identifying at least one region of the reconstructed image is coded or not for the group of blocks. A decoder configured for decoding a scalable stream comprising a base layer and at least one enhancement layer is also described. The decoder comprises:
- means for reconstructing at least one image from the base layer,
- means for decoding a first part of the at least one enhancement layer relating to a group of blocks into decoded enhancement data which comprise at least one item of information for identifying at least one region of the reconstructed image to process, - means for processing the at least one region of the reconstructed image identified by the information,
- means for obtaining residues for at least one block of the group of blocks from a second part of the at least one enhancement layer, and
- means for reconstructing the blocks of the group from residues and from the at least one region processed.
The decoder is configured to execute the steps of the decoding method according to one of the embodiments.
A coder configured to code a sequence of images in the form of a scalable stream comprising a base layer and at least one enhancement layer is described which comprises:
means for coding in the base layer and for reconstructing at least one image,
- means for coding enhancement data in a first part of the enhancement layer relating to a group of blocks, the enhancement data comprising an item of information for identifying at least one region of the reconstructed image to process,
- means for processing the at least one region of the reconstructed image identified by the information,
- means for obtaining residues for at least one block of the group of blocks from the blocks of the group and from the at least one region processed, and
- means for coding the residues in a second part of the enhancement layer.
The coder is configured to execute the steps of the coding method according to one of the embodiments described.
A coded data stream representative of a sequence of images comprising a base layer and at least one enhancement layer, the base layer comprising data relative to at least one image of the base layer and the enhancement layer comprising in a first part relating to a group of blocks enhancement data which comprise an item of information for identifying at least one region of the image of the base layer to process after reconstruction and in a second part data representative of residues which are obtained from at least one block of the group and from the at least one region processed.
A computer program product is described that comprises program code instructions for executing the steps of the decoding method according to one of the embodiments, when the program is executed on a computer. A computer program product is described that comprises program code instructions for executing the steps of the coding method according to one of the embodiments, when the program is executed on a computer.
4. List of figures
The invention will be better understood and illustrated by means of non- restrictive embodiments and advantageous implementations, with reference to the accompanying drawings, wherein:
- Figure 1 shows a coding method according to the prior art,
- Figure 2 shows a decoding method according to a particular and non- restrictive embodiment of the invention,
- Figures 3 and 4 show an item of information info in a specific map form to identify at least one region of an image of the reconstructed base layer to process,
- Figures 5 to 7 show parts of an enhancement layer according to different embodiments of the invention,
- Figure 8 shows a coding method according to a particular and non- restrictive embodiment of the invention,
- Figures 9 and 10 show a decoding device according to particular and non- restrictive embodiments of the invention, and
- Figures 1 1 and 12 show a coding device according to particular and non- restrictive embodiments of the invention. 5. Detailed description of the invention
The invention relates to a device for decoding a scalable stream representative of an image sequence comprising a base layer and at least one enhancement layer.
An image sequence is a series of several images. Each image comprises pixels or image points, with each of which is associated at least one item of image data. An item of image data is for example an item of luminance data or an item of chrominance data. In the remainder of the description, the coding and decoding methods are described with reference to blocks of pixels. It is clear that these methods can be applied to several blocks of an image and to several images of a sequence with a view to coding respectively the reconstruction of one or more images. In the specific case of scalability, a scalable stream represents an image sequence in different versions, for example, different resolutions, different gamuts/colour spaces, etc.
The expression "motion data" is to be understood in the widest sense. It signifies the motion vectors and possibly the reference image indexes enabling a reference image to be identified in the image sequence. It can also comprise an item of information indicating the interpolation type used to determine the prediction block. In fact, in the case where the motion vector associated with a block Be does not have integer coordinates, the image data must be interpolated in the reference image Iref to determine the prediction block. The motion data associated with a block is generally calculated by a motion estimation method, for example by block pairing. However, the invention is in no way limited by the method enabling a motion vector to be associated with a block.
The expression "residual data" designates data obtained after extraction of other data. The extraction is generally a subtraction pixel by pixel of prediction data from source data. However, the extraction is more general and comprises notably a weighted subtraction. The expression "residual data" is synonymous with the term "residue".
The expression "transformed residual data" designates residual data to which a transform has been applied. A DCT (Discrete Cosine Transform) is an example of such a transform described in chapter 3.4.2.2 of the book by I. E. Richardson entitled Ή.264 and MPEG-4 video compression" published by J. Wiley & Sons in September 2003. The wavelet transform described in chapter 3.4.2.3 of the book by I. E. Richardson and the Hadamard transform are other examples. Such transforms "transform" a block of image data, e.g. luminance and/or chrominance residual data, into a "block of transformed data" also called "transformed block", "frequency data block" or "block of coefficients". The block of coefficients generally comprises a low-frequency coefficient known under the name of direct current coefficient or DC coefficient and high- frequency coefficients known under the name of AC coefficients.
The expression "prediction data" designates data used to predict other data. A prediction block or predictor is a block of pixels with which prediction data is associated.
A prediction block is obtained from a block or several blocks of the same image as the image to which belongs the block that it predicts (spatial prediction or intra-image prediction) or from one (mono-directional prediction) or several blocks (bi-directional or bi-predicted prediction) of a different image (temporal prediction or inter-image prediction) of the image to which the block that it predicts belongs. In the scalable case, a prediction block can also be obtained for a block of an enhancement layer from at least one block of an image of the base layer, e.g. the co-located block (inter-layer prediction).
The expression "prediction mode" designates the way in which a block is predicted. The expression "coding mode" designates the way in which a block is coded in the stream. It includes the prediction mode if it is predicted. Among the prediction modes, there is the INTRA mode that corresponds to a spatial prediction, the INTER mode that corresponds to a temporal prediction, the INTER-LAYER mode that corresponds to an inter-layer prediction. Note that the INTER-LAYER mode can be implemented like an INTER mode if the reconstructed image of the base layer is stored in the reference buffer of the enhancement layer. The standard INTER modes are thus distinguished from an INTER mode corresponding to an INTER-LAYER mode by the value of the reference index. The prediction mode can specify the way in which the block is partitioned to be coded. Thus, the 8x8 INTER prediction mode associated with a block of size 16x16 signifies that the 16x16 block is partitioned into 4 8x8 blocks and predicted by temporal prediction. The expression "reconstructed data" designates data (for example pixels, blocks) obtained after merging residues with prediction data. The merge is generally a sum of prediction data with residues. However, merging is more general and comprises notably the weighted sum. A reconstructed block is a block of reconstructed pixels.
A "causal neighbouring area" of a current block designates a neighbouring area of this block that comprises pixels codes/reconstructed before those of the current block.
In reference to image decoding, the terms "reconstruction" and "decoding" are very often used as being synonymous. Thus, a "reconstructed block" is also designated under the terminology of "decoded block".
The term coding is to be taken in the widest sense. The coding can possibly comprise the transformation and/or the quantization of image data. It can also designate the entropy coding only.
Figure 2 shows a decoding method of a scalable stream S comprising a base layer and at least one enhancement layer according to a particular and non- restrictive embodiment of the invention. Naturally, the invention described for a base layer and an enhancement layer can be applied to two enhancement layers of different levels, the lower enhancement layer playing the role of a base layer with respect to the upper level layer.
During a step 20, at least one image lBi_ of the base layer is reconstructed from the base layer. For this purpose, the base layer is decoded at least partially by entropy decoding, for example by CABAC decoding ("Context- adaptive binary arithmetic coding") or CAVLC ("Context-adaptive variable- length coding") decoding. The decoded data notably comprises data representative of residues, for example quantized DCT coefficients, coding modes, possibly motion data. The reconstruction of the image lBi_ for a current block of the image comprises the obtaining of a spatial or temporal predictor according to the coding mode of the block. The predictor is obtained from a part of an image previously reconstructed either in the same image as the current block to reconstruct (spatial prediction) or in another image (temporal prediction) called reference image. The predictor is added to residues to reconstruct the current block. The residues are, for example, obtained by inverse quantization and inverse transformation of data relating to residues obtained during the entropy decoding of the base layer. The term inverse is used to signify that the quantization and transformation operations on the decoder are inverse from those used on the coder. According to a variant, no predictor is added to the residues in the specific case where the current block is coded without prediction (e.g. according to the I PCM Intra mode of H.264). The residues correspond in the latter case directly to the reconstructed current block. According to another variant, the residues are null. This is particularly the case for a current block coded according to a particular coding mode called skip mode. The reconstructed current block is the prediction block in this case. The reconstruction process described for a current block is reiterated to reconstruct the image I BL of the base layer. Naturally, this process can also be reiterated to reconstruct all the images of the base layer.
During a step 22, a first part EL1 of the enhancement layer relating to a group of blocks is decoded into decoded enhancement data, the group of blocks belonging to an image of the enhancement layer to reconstruct. The part EL1 is decoded by entropy decoding, e.g. by CABAC decoding. The decoded enhancement data thus comprise information relating to the group of blocks of the image of the enhancement layer to reconstruct. A group of blocks is for example an image or an image slice. For example, this enhancement data is obtained by decoding an image slice header, a PPS (Picture Parameter Set) or even a message SEI (Supplemental Enhancement Information). The terms PPS and SEI are notably defined in paragraph 3.107 (definition) and the appendix D respective of the ISO/I EC document 14496-10 describing the video coding standard H.264. Indeed, such elements contain information relating to a group of blocks. However, note that an image slice can only contain a single block. However, coding images with image slices of a single block is ineffective in terms of bitrate. Slice coding, each slice containing several blocks, e.g. 2 blocks, groups common information together and thus prevents duplicating the same information in the headers of each block. Advantageously, this decoded enhancement data comprises an item of information info for identifying at least one region of the image lBi_ reconstructed in step 20 to process. Hence, from the decoding of the first part of enhancement layer EL1 relating to the group of blocks, the decoder has knowledge of the region of the reconstructed image lBi_ to process, i.e. the one that will effectively be used during the reconstruction of an image of the enhancement layer. This information can thus distinguish between the regions of the reconstructed image I BL that will be used during the reconstruction of an image of the enhancement layer and hence that must be processed and those that will not be used and hence that must not be processed.
This information info is presented for example in the form of a map indicating for each block of the group of blocks of the enhancement layer whether the corresponding region, e.g. the block, of the base layer must be processed or not. A region of the reconstructed image lBi_ corresponds to a block of an image of the enhancement layer if it is co-located, i.e. if it occupies in the image the same spatial position possibly after processing (spatial scalability case).
As a variant, the information info is presented for example in the form of a map indicating for each region, e.g. block, of the reconstructed image I BL corresponding to a block of the group of blocks if this region must be processed or not.
This information irrespective of whether it relates to the base layer or to the enhancement layer is coded in the form of a regular map, e.g. in the case of the HEVC standard for each CTU (Coding Tree Unit) or for example for each macroblock as shown in figure 3. According to an embodiment variant, this information is coded in the form of an irregular map, e.g. in the case of the HEVC standard for each CU (Coding Unit) as shown in figure 4. In these figures, a value 0 indicates that the reconstructed region lBi_ is not to be processed and a value 1 indicates that it is to be processed. Naturally, the inverse can be chosen by convention. This information is decoded by entropy decoding, for example CAVLC or CABAC.
During a step 24, the region or regions of the image of the reconstructed base layer identified by the information info decoded in step 22 is processed. The processing is for example a re-sampling of the region in the case of spatial scalability, a colour conversion in the case of CGS scalability, a noise reduction filtering in the case of SNR scalability.
This solution advantageously enables only the useful regions of the reconstructed image lBi_ to be processed, i.e. used in the reconstruction of blocks of an image lEi_ of the enhancement layer. Moreover, by proceeding thus the useful regions of the base layer are processed before the reconstruction of the blocks of the group of blocks and are thus all available during the reconstruction of the group of blocks which avoids too frequent accesses to the image buffer of the reconstructed base layer and too many calls on the processing function.
During a step 26, residues are obtained for at least one block of the group of blocks from a second part of the enhancement layer EL2. According to a variant, residues are obtained for each block of the group of blocks. This second part comprises data representative of residues for example quantized DCT coefficients from which residues are calculated or else a particular coding mode (skip mode) indicating that these residues are null. This step possibly comprises the entropy decoding, the inverse quantization and the inverse transformation of data of the second part, e.g. quantized DCT coefficients, to obtain the residues res. This second part possibly comprises motion data.
During a step 28, the blocks of the group are reconstructed from residues res and from the region processed. For this purpose, for each block of the group a predictor is obtained either from an image part previously reconstructed or from the same image as the current block of the group to reconstruct (spatial prediction) or from another image of the enhancement layer (temporal prediction) or even from the region of the reconstructed image IBL processed (inter-layer prediction). The predictor is added to the residues of the current block to reconstruct the current block. Note that according to a variant no predictor is added to the residues of the current block, in the specific case where the current block is coded without prediction (e.g. according to the I PCM coding mode of H.264). The residues correspond in the latter case directly to the reconstructed current block. In the group of blocks, the coding modes can be varied, some blocks being coded according to an I PCM mode, by spatial prediction, others by temporal prediction, others by inter-layer prediction and even others according to a skip mode.
According to a particular embodiment, an additional item of information is decoded from the scalable stream, more precisely, from the enhancement layer for example of a PPS or of an image slice header that indicates for a group of blocks whether the item of information for identifying at least one region of the reconstructed image I BL to process is coded or not for this group of blocks. Using this particular information, it is possible to specify for each image or even for each image slice if the item of information for identifying at least one region of the reconstructed image I BL to process is coded or not. The additional information is for example a flag, i.e. a binary data. As a simple illustrative example, a first PPS1 is decoded with the additional value equal to Flg1 , e.g. 1 , indicating that the item of information for identifying the region of the reconstructed image lBi_ to process is coded. A second PPS1 is decoded with the additional value equal to Flg2, for example 0, indicating that the information for identifying at least one region of the reconstructed image I BL to process is not coded. Hence, each image of the enhancement layer can according to the case refer to one or other of the PPS. If an image of the enhancement layer I EL refers to the first PPS1 , then an item of information for identifying at least one region of the reconstructed image I BL to process is decoded, e.g. in each slice header of the image of the enhancement layer as shown in figure 5. In the case where an image J EL refers to the PPS2 no item of information info is present. According to a variant, this flag can be coded in each image slice header, in which case if the value indicates that the item of information for identifying at least one region of the reconstructed image I BL to process is coded, this item of information for identifying at least one region of the reconstructed image I BL to process follows in the additional information stream Fig as shown in figure 6.
According to another embodiment illustrated by figure 7, a first item of additional information Flg1 is decoded from the enhancement layer that indicates whether the information info for identifying at least one region of the reconstructed image I BL to process can be decoded or not for a group of blocks. This first item of additional information Flg1 is for example decoded from the PPS. For example, if its value is 0 this signifies that no item of information for identifying at least one region of the reconstructed image I BL to process can be decoded from the stream. If its value is 1 , this signifies that an item of information for identifying at least one region of the reconstructed image I BL to process can be decoded. In this case, a second item of additional information Flg2 is decoded for example in the header of each image slice that indicates for a group of blocks whether the item of information info for identifying at least one region of the reconstructed image IBLJO process is coded or not for this group of blocks. In the case where the second item of information indicates for a group of blocks that said item of information for identifying at least one region of the reconstructed image IBL to process is coded, this item of information info follows the second item of information in the stream. This solution enables greater flexibility. Figure 8 shows a coding method of an image sequence in the form of a scalable stream comprising a base layer and at least one enhancement layer according to a particular and non-restrictive embodiment of the invention. The same variants as the ones described in relation to the decoding method apply to the coding method even in the case where they may not be described explicitly in the remainder of the description.
During a step 80, at least one image IBL of the base layer is coded in the base layer. The coding generally comprises, for blocks of the image, the spatial or temporal prediction of the blocks, the obtaining of residues by subtracting a predictor from the block to code, the transformation and quantization of residues before the entropy coding. The same coding modes described in relation to the decoding method can be used, particularly the skip mode and I PCM coding mode of AVC. Other data is also coded such as coding modes, possibly motion data. The coding comprises the reconstruction of the image of the base layer. Indeed, the latter can then be used for the prediction of other data. These steps are well known to those skilled in the specialised art of video coding and will not be described in further detail. The process can be reiterated for all the images of the base layer so as to code the entire sequence.
During a step 82, enhancement data is coded in a first part of the enhancement layer EL1 relating to a group of blocks of the image of the enhancement layer to reconstruct. A group of blocks is for example an image or a slice of images. For example, this enhancement data is coded in an image slice header, in PPS or even in an SEI message. Indeed, such elements contain information relating to a group of blocks. However, note that an image slice can only contain a single block. However, coding images with image slices of a single block is ineffective in terms of bitrate. Slice coding, each slice containing several blocks, e.g. 2 blocks, groups common information together and thus prevents duplicating the same information in the headers of each block. Advantageously, this coded enhancement data comprises an item of information info for identifying at least one region of the reconstructed image I BL to process. For example, this information is obtained during a previous step of deciding on the coding mode. Indeed, as soon as the coding modes of the blocks of the group are selected, for example by minimising a bitrate-distortion type criterion, this map can be constructed. Indeed, all the blocks of the map that correspond to an inter-layer prediction mode are predicted in relation to a block of the image of the reconstructed base layer processed. This block must then be identified in the map as belonging to the region to process. Hence, from the decoding of the first part of enhancement layer EL1 relating to the group of blocks, the decoder has knowledge of the region or regions of the reconstructed image I BL to process, i.e. the one(s) that will effectively be used during the reconstruction of an image of the enhancement layer. This information can thus distinguish between the regions of the reconstructed image I BL that will be used during the reconstruction of an image of the enhancement layer and hence that must be processed and those that will not be used and hence that must not be processed. This information info is presented for example in the form of a map indicating for each block of the group of blocks of the enhancement layer whether the corresponding region, e.g. the block, of the base layer must be processed or not. A region of the reconstructed image I BL corresponds to a block of an image of the enhancement layer if it is co-located, i.e. if it occupies in the image the same spatial position possibly after processing (spatial scalability case). As a variant, the information info is presented for example in the form of a map indicating for each region, e.g. block, of the reconstructed image I BL corresponding to a block of the group of blocks if said region must be processed or not. This information irrespective of whether it relates to the base layer or to the enhancement layer is coded in the form of a regular map, e.g. in the case of the HEVC standard for each CTU (Coding Tree Unit) or for example for each macroblock as shown in figure 3. According to an embodiment variant, this information is coded in the form of an irregular map, e.g. in the case of the HEVC standard for each CU (Coding Unit) as shown in figure 4. In these figures, a value 0 indicates that the reconstructed region lBi_ is not to be processed and a value 1 indicates that it is to be processed. Naturally, the inverse can be chosen by convention. This information is coded by entropy coding, for example CAVLC or CABAC.
During a step 84, the region or regions of the image of the reconstructed base layer identified by the information info coded in step 82 is/are processed.
The processing is for example a re-sampling of the region in the case of spatial scalability, a colour conversion in the case of CGS scalability, a noise reduction filtering in the case of SNR scalability.
This solution advantageously enables only the useful regions of the reconstructed image lBi_ to be processed, i.e. used in the coding of blocks of an image I EL of the enhancement layer. Moreover, by proceeding thus the useful regions of the base layer are processed before the coding of the blocks of the group of blocks and are thus all available during the coding of the group of blocks which avoids too frequent accesses to the image buffer of the reconstructed base layer and too many calls on the processing function.
During a step 86, residues are obtained for at least one block of the group from blocks of the group and from the region processed. According to a variant, residues are obtained for each block of the group of blocks. For this purpose, for each block of the group a predictor is obtained either from an image part previously reconstructed or from the same image as the current block to code (spatial prediction) or from another image of the enhancement layer (temporal prediction) or even from the region of said reconstructed image I BL processed (inter-layer prediction). The predictor is subtracted from the current block to obtain residues. Note that according to a variant no predictor is subtracted from the current block, in the case where the block was coded without prediction (e.g. I PCM coding mode of H.264). The residues block corresponds in the latter case directly to the block to code. In the case of the skip mode, the residues are null. It is the coding mode of the block coded in the stream that indicates it In the group of blocks, the coding modes can be varied, some blocks being coded according to an I PCM mode, by spatial prediction, others by temporal prediction, others by inter-layer prediction and even others according to a skip mode.
During a step 88, the residues relating to at least one block of the group of blocks are coded in a second part of the enhancement layer. This step possibly comprises the transformation and quantization of the residues before their entropy coding.
According to a particular embodiment, the additional information shown in figures 5 to 7 and described in relation to the decoding method is coded in the scalable stream S. All the embodiments described in relation to the decoding method apply symmetrically to the coding method.
The invention also relates to a scalable stream representative of an image sequence. This stream is for example produced by the coding method described previously.
This scalable stream comprises a base layer and at least one enhancement layer. The base layer comprises data relating to at least one image of the base layer.
The enhancement layer comprises in a first part EL1 relating to a group of blocks enhancement data which comprise an item of information for identifying at least one region of the image of the base layer to process after its reconstruction.
The enhancement layer comprises in a second part EL2 data representative of residues for at least one block of the group of blocks, the residues being obtained from at least one block of the group and from the region processed. According to a variant, the enhancement layer comprises in a second part EL2 data representative of residues for each of the blocks of the group of blocks. This data comprises for example a coding mode, possibly motion data, of data relating to residues (e.g. quantized DCT coefficients) or of data enabling such residues to be determined, e.g. the skip coding mode.
Advantageously, the scalable stream comprises the additional information shown in figures 5 to 7 and described in relation to the decoding method.
Figure 9 shows a decoding device 9 according to a non-restrictive embodiment of the invention. The decoding device 9 comprises a decoding module 90 able to decode a base layer BL of a scalable stream S and to reconstruct images of the base layer. The decoding module 90 is notably configured to implement step 20 of the decoding method. The decoding device 9 further comprises a processing module 92 able to process at least one region of an image of the reconstructed base layer. The processing module 92 is notably configured to implement step 24 of the decoding method. The decoding device 9 also comprises another decoding module 94 able to decode an enhancement layer EL of a scalable stream S and to reconstruct images of the enhancement layer. The decoding module 94 is also able to supply the processing module 92 with the information info. The decoding module 94 is notably configured to implement steps 22, 26 and 28 of the decoding method. The decoding device 9 optionally comprises a demultiplexing module 96 configured to demultiplex the scalable stream S into a base layer BL and one or more enhancement layers EL.
Figure 10 shows a particular implementation of such a decoding device 9. The decoding device 9 comprises hardware components such as at least a processor 900 (e.g. CPU/GPU), a memory 902 (e.g. RAM), a user interface 904 (e.g. keyboards, mouse, etc.) and one or more input/output interfaces 906 able to be connected to other devices. These components are for example interconnected by an address and data bus 908. According to an implementation example, when powered up, the processor 900 loads the instructions corresponding to the steps of the decoding method into the memory 902 and executes them.
According to a particular embodiment, the decoding device 9 advantageously comprises an interface to read a computer program stored on a digital data support 910 that stores the program code instructions, that when they are executed by the processor 900, carry out the steps of the decoding method according to any one of the embodiments of the invention described previously. According to a variant, the user interface is external to the decoding device 9.
Figure 11 shows a coding device 1 1 according to a non-restrictive embodiment of the invention. The coding device 1 1 comprises a coding module 1 10 able to code and reconstruct images lBi_ in a base layer BL of a scalable stream. The coding module 1 10 is notably configured to implement step 80 of the coding method. The coding device 1 1 further comprises a processing module 1 12 able to process at least one part of an image of the reconstructed base layer. The processing module 1 12 is notably configured to implement step 84 of the coding method. The coding device 1 1 also comprises another coding module 1 14 able to code images I EL in an enhancement layer EL of the scalable stream S. The coding module 1 14 is also able to supply the processing module 1 12 with the information info. The coding module 1 14 is notably configured to implement steps 82, 86 and 88 of the coding method. The coding device 1 1 optionally comprises a multiplexing module 1 16 configured to multiplex the base layer and enhancement layer or layers into a single scalable stream S.
Figure 12 shows a particular implementation of such a coding device 1 1 . The coding device 1 1 comprises hardware components such as at least a processor 1 100 (e.g. CPU/GPU), a memory 1 102 (e.g. RAM), a user interface 1 104 and one or more input/output interfaces 1 106 able to be connected to other devices. These components are for example interconnected by an address and data bus 1 108. According to an implementation example, when powered up, the processor 1 100 loads the instructions corresponding to the steps of the coding method into the memory 1 102 and executes them.
The coding device 1 1 advantageously comprises an interface to read a computer program stored on a digital data support 1 1 10 that stores the program code instructions, that when they are executed by the processor 1 100, carry out the steps of the coding method according to any one of the embodiments of the invention described previously. According to a variant, the user interface is external to the coding device 1 1 .
The decoding 9 and coding 1 1 devices described are simplified for reasons of clarity. Real devices may comprise other elements such as for example an internal or external battery, non-volatile memory (e.g. flash memory, hard disk drive, etc.), etc. According to variants, decoding 9 and coding 1 1 devices compatible with the invention are implemented according to a purely hardware embodiment, for example in the form of a dedicated component (for example in an ASIC (Application Specific Integrated Circuit) or FPGA (Field-Programmable Gate Array) or VLSI (Very Large Scale Integration) or of several electronic components integrated into a device or even in a form of a combination of hardware elements and software elements. Each embodiment can be used independently from the other embodiments or else the embodiments can be combined with each other.
Naturally, the invention described for a base layer and an enhancement layer can be applied to two enhancement layers of different levels, the lower enhancement layer playing the role of a base layer with respect to the upper level layer. Furthermore, the invention is not restricted by the type of quantizing, transform or entropy coding used.

Claims

Claims
1 . A method for decoding a scalable stream comprising a base layer and at least one enhancement layer, the method comprising:
- reconstructing (20) at least one image from said base layer,
- decoding (22) a first part of said at least one enhancement layer relating to a group of blocks into decoded enhancement data which comprise at least one item of information for identifying at least one region of said reconstructed image to process,
- processing (24) said at least one region of said reconstructed image identified by said information,
- obtaining (26) residues for at least one block of the group of blocks from a second part of said at least one enhancement layer, and
- reconstructing (28) the blocks of the group from said residues and from said at least one region processed.
2. Method according to claim 1 , wherein said information for identifying at least one region of said reconstructed image to process is a map indicating for each block of said reconstructed image corresponding to a block of the group of blocks if the block must be processed or not.
3. Method according to claim 1 , wherein said information for identifying at least one region of said reconstructed image to process is a map indicating for each block of the group of blocks if the block uses for its reconstruction a processed block of said reconstructed image.
4. Method according to one of claims 1 to 3, wherein said base layer is another enhancement layer of a lower level than said at least one enhancement layer.
5. Method according to one of claims 1 to 4, comprising a decoding step from said enhancement layer of an additional item of information for at least one group of blocks indicating whether said information for identifying at least one region of said reconstructed image to process is coded or not for the group of blocks.
Method according to claim 5, wherein the group of blocks is an image slice or an image.
Method according to one of claims 1 to 4, comprising decoding, from said enhancement layer, a first additional item of information indicating whether said information for identifying at least one region of said reconstructed image to process can be coded or not for a group of blocks.
Method according to claim 7, wherein said first item of additional information being decoded for at least one image, the method comprises, when said first item of additional information indicates that an item of information for identifying at least one region of said reconstructed image to process can be coded, decoding, from said enhancement layer, a second item of additional information indicating for a group of blocks whether said information for identifying at least one region of said reconstructed image to process is coded or not for the group of blocks.
A method for coding a sequence of images in the form of a scalable stream comprising a base layer and at least one enhancement layer, the method comprising:
- coding (80) at least one image in said base layer and reconstructing (80) said at least one image,
- coding (82) enhancement data in a first part of said enhancement layer relating to a group of blocks, said enhancement data comprising an item of information for identifying at least one region of said reconstructed image to process,
- processing (84) said at least one region of said reconstructed image identified by said information, - obtaining (86) residues for at least one block of the group of blocks from said blocks of the group and from said at least one region processed, and
- coding (88) the residues in a second part of said enhancement layer.
1 0. Method according to claim 9, wherein said information for identifying at least one region of said reconstructed image to process is a map indicating for each block of said reconstructed image corresponding to a block of the group of blocks if the block must be processed or not.
1 1 . Method according to claim 9, wherein said information for identifying at least one region of said reconstructed image to process is a map indicating for each block of the group of blocks if the block uses for its reconstruction a processed block of said reconstructed image.
1 2. Method according to one of claims 9 to 1 1 , wherein said base layer is another enhancement layer of a lower level than said at least one enhancement layer.
1 3. Method according to one of claims 9 to 1 2, comprising coding, in said enhancement layer, an additional item of information for at least one group of blocks indicating whether said information for identifying at least one region of said reconstructed image to process is coded or not for the group of blocks.
14. Method according to one of claims 9 to 1 2, comprising coding, in said enhancement layer, a first additional item of information indicating whether said information for identifying at least one region of said reconstructed image to process can be coded or not for a group of blocks.
1 5. Method according to claim 14, wherein said first item of additional information being coded for at least one image, the method comprises, when said first item of additional information indicates that an item of information for identifying at least one region of said reconstructed image to process can be coded, coding, in said enhancement layer, a second item of additional information indicating for a group of blocks whether said information for identifying at least one region of said reconstructed image to process is coded or not for the group of blocks.
16. A decoder configured for decoding a scalable stream comprising a base layer and at least one enhancement layer, the decoder comprising:
- means (90, 900, 902, 904, 906, 908) for reconstructing at least one image from said base layer,
- means (94, 900, 902, 904, 906, 908) for decoding a first part of said at least one enhancement layer relating to a group of blocks into decoded enhancement data which comprise at least one item of information for identifying at least one region of said reconstructed image to process,
- means (92, 900, 902, 904, 906, 908) for processing said at least one region of said reconstructed image identified by said information,
- means (94, 900, 902, 904, 906, 908) for obtaining residues for at least one block of the group of blocks from a second part of said at least one enhancement layer, and
- means (94, 900, 902, 904, 906, 908) for reconstructing said blocks of the group from said residues and from said at least one region processed.
17. The decoder according to claim 16, the decoder being configured to execute the steps of the decoding method according to one of claims 1 to 8.
18. A coder configured to code a sequence of images in the form of a scalable stream comprising a base layer and at least one enhancement layer, the coder comprising: means for coding in said base layer and for reconstructing (1 1 0, 1 1 00, 1 1 02, 1 1 04, 1 1 06, 1 1 08) at least one image,
- means for coding (1 14, 1 100, 1 1 02, 1 1 04, 1 1 06, 1 1 08) enhancement data in a first part of said enhancement layer relating to a group of blocks, said enhancement data comprising an item of information for identifying at least one region of said reconstructed image to process,
- means for processing (1 1 2, 1 1 00, 1 1 02, 1 104, 1 1 06, 1 1 08) said at least one region of said reconstructed image identified by said information,
- means for obtaining (1 14, 1 1 00, 1 1 02, 1 1 04, 1 106, 1 1 08) residues for at least one block of the group of blocks from said blocks of the group and from said at least one region processed, and
- means for coding (1 14, 1 1 00, 1 1 02, 1 1 04, 1 1 06, 1 1 08) said residues in a second part of said enhancement layer.
19. Coder according to claim 1 8, the coder being configured to execute the steps of the coding method according to one of claims 9 to 1 5.
20. A data stream representative of an image sequence comprising a base layer and at least one enhancement layer, the base layer comprising data relative to at least one image of the base layer and
the enhancement layer comprising in a first part relative to a group of blocks of enhancement data which comprise an item of information for identifying at least one region of said image of the base layer to process after reconstruction and in a second part of data representative of residues which are obtained from at least one block of the group and from said at least one region processed.
21 . A computer program product, characterised in that it comprises program code instructions for executing the steps of the method according to any one of claims 1 to 8 when the program is executed on a computer.
22. A computer program product, characterised in that it comprises program code instructions for executing the steps of the method according to any one of claims 9 to 15 when the program is executed on a computer.
PCT/EP2014/064817 2013-07-17 2014-07-10 Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device WO2015007604A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
KR1020167001286A KR20160034900A (en) 2013-07-17 2014-07-10 Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device
EP14744790.8A EP3022922A1 (en) 2013-07-17 2014-07-10 Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device
CN201480040499.7A CN105659600A (en) 2013-07-17 2014-07-10 Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device
US14/905,960 US9961353B2 (en) 2013-07-17 2014-07-10 Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device
JP2016526536A JP6393323B2 (en) 2013-07-17 2014-07-10 Method and device for decoding a scalable stream representing an image sequence and corresponding encoding method and device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
FR1357034 2013-07-17
FR1357034A FR3008840A1 (en) 2013-07-17 2013-07-17 METHOD AND DEVICE FOR DECODING A SCALABLE TRAIN REPRESENTATIVE OF AN IMAGE SEQUENCE AND CORRESPONDING ENCODING METHOD AND DEVICE

Publications (1)

Publication Number Publication Date
WO2015007604A1 true WO2015007604A1 (en) 2015-01-22

Family

ID=50097765

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2014/064817 WO2015007604A1 (en) 2013-07-17 2014-07-10 Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device

Country Status (8)

Country Link
US (1) US9961353B2 (en)
EP (1) EP3022922A1 (en)
JP (1) JP6393323B2 (en)
KR (1) KR20160034900A (en)
CN (1) CN105659600A (en)
FR (1) FR3008840A1 (en)
TW (1) TW201505424A (en)
WO (1) WO2015007604A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR3083662A1 (en) * 2018-07-05 2020-01-10 Awecom, Inc. ROBUST COMPRESSION AND DECOMPRESSION OF DIGITAL IMAGES

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017073360A1 (en) * 2015-10-30 2017-05-04 ソニー株式会社 Image processing device and method
KR102601732B1 (en) * 2016-05-31 2023-11-14 삼성디스플레이 주식회사 Method for image encoding and method for image decoding
TWI618029B (en) * 2016-07-26 2018-03-11 聯發科技股份有限公司 Graphics processing device
US10681382B1 (en) * 2016-12-20 2020-06-09 Amazon Technologies, Inc. Enhanced encoding and decoding of video reference frames
FR3072850B1 (en) * 2017-10-19 2021-06-04 Tdf CODING AND DECODING METHODS OF A DATA FLOW REPRESENTATIVE OF AN OMNIDIRECTIONAL VIDEO
US10701376B2 (en) 2018-07-05 2020-06-30 Awecom, Inc. Resilient image compression and decompression

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060133503A1 (en) * 2004-12-06 2006-06-22 Park Seung W Method for scalably encoding and decoding video signal
US20080007438A1 (en) * 2006-07-10 2008-01-10 Sharp Laboratories Of America, Inc. Methods and Systems for Signaling Multi-Layer Bitstream Data
WO2012167712A1 (en) * 2011-06-10 2012-12-13 Mediatek Inc. Method and apparatus of scalable video coding

Family Cites Families (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005142654A (en) * 2003-11-04 2005-06-02 Matsushita Electric Ind Co Ltd Video transmitting apparatus and video receiving apparatus
KR100763196B1 (en) * 2005-10-19 2007-10-04 삼성전자주식회사 Method for coding flags in a layer using inter-layer correlation, method for decoding the coded flags, and apparatus thereof
KR100781524B1 (en) * 2006-04-04 2007-12-03 삼성전자주식회사 Method and apparatus for encoding/decoding using extended macroblock skip mode
JP5134001B2 (en) * 2006-10-18 2013-01-30 アップル インコーポレイテッド Scalable video coding with lower layer filtering
FR2907630A1 (en) * 2006-10-19 2008-04-25 Thomson Licensing Sas DEVICE AND METHOD FOR SCALABLE FORMAT CODING OF IMAGE SEQUENCE AND CORRESPONDING DECODING DEVICE AND METHOD
EP1985121A4 (en) 2006-11-17 2010-01-13 Lg Electronics Inc Method and apparatus for decoding/encoding a video signal
EP1933564A1 (en) * 2006-12-14 2008-06-18 Thomson Licensing Method and apparatus for encoding and/or decoding video data using adaptive prediction order for spatial and bit depth prediction
EP2051527A1 (en) 2007-10-15 2009-04-22 Thomson Licensing Enhancement layer residual prediction for bit depth scalability using hierarchical LUTs
HUE024173T2 (en) 2008-04-16 2016-05-30 Ge Video Compression Llc Bit-depth scalability
US20100278232A1 (en) 2009-05-04 2010-11-04 Sehoon Yea Method Coding Multi-Layered Depth Images
KR101066117B1 (en) 2009-11-12 2011-09-20 전자부품연구원 Method and apparatus for scalable video coding
CN104054338B (en) 2011-03-10 2019-04-05 杜比实验室特许公司 Locating depth and color scalable video
PL2822276T3 (en) * 2012-02-29 2019-04-30 Lg Electronics Inc Inter-layer prediction method and apparatus using same
WO2014052740A1 (en) * 2012-09-28 2014-04-03 Vid Scale, Inc. Adaptive upsampling for multi-layer video coding
WO2014053518A1 (en) * 2012-10-01 2014-04-10 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Scalable video coding using subblock-based coding of transform coefficient blocks in the enhancement layer
CN105165011A (en) * 2013-02-25 2015-12-16 三星电子株式会社 Device and method for scalable video encoding considering memory bandwidth and computational quantity, and device and method for scalable video decoding
EP2979447B1 (en) * 2013-03-28 2018-01-03 Huawei Technologies Co., Ltd. Method for determining predictor blocks for a spatially scalable video codec
KR20140121315A (en) * 2013-04-04 2014-10-15 한국전자통신연구원 Method and apparatus for image encoding and decoding based on multi-layer using reference picture list
KR102290421B1 (en) * 2013-04-05 2021-08-17 삼성전자주식회사 Method and apparatus for multi-layer video encoding for random access, method and apparatus for multi-layer video decoding for random access
EP3410715A1 (en) * 2013-04-05 2018-12-05 Vid Scale, Inc. Inter-layer reference picture enhancement for multiple layer video coding
KR20140122202A (en) * 2013-04-05 2014-10-17 삼성전자주식회사 Method and apparatus for video stream encoding according to layer ID extention, method and apparatus for video stream decoding according to layer ID extention

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060133503A1 (en) * 2004-12-06 2006-06-22 Park Seung W Method for scalably encoding and decoding video signal
US20080007438A1 (en) * 2006-07-10 2008-01-10 Sharp Laboratories Of America, Inc. Methods and Systems for Signaling Multi-Layer Bitstream Data
WO2012167712A1 (en) * 2011-06-10 2012-12-13 Mediatek Inc. Method and apparatus of scalable video coding

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHEN J ET AL: "SHVC Test Model 2", 13. JCT-VC MEETING; 104. MPEG MEETING; 18-4-2013 - 26-4-2013; INCHEON; (JOINT COLLABORATIVE TEAM ON VIDEO CODING OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ); URL: HTTP://WFTP3.ITU.INT/AV-ARCH/JCTVC-SITE/,, no. JCTVC-M1007, 6 June 2013 (2013-06-06), XP030114429 *
EDOUARD FRANCOIS ET AL: "Extended Spatial Scalability: A generalization of spatial scalability for SVC Extension of AVC/H.264", 25. PICTURE CODING SYMPOSIUM;24-4-2006 - 26-4-2006; BEIJING,, 24 April 2006 (2006-04-24), XP030080338 *
KIM C ET AL: "Description of scalable video coding technology proposal by LG Electronics and MediaTek (differential coding mode on)", 11. JCT-VC MEETING; 102. MPEG MEETING; 10-10-2012 - 19-10-2012; SHANGHAI; (JOINT COLLABORATIVE TEAM ON VIDEO CODING OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ); URL: HTTP://WFTP3.ITU.INT/AV-ARCH/JCTVC-SITE/,, no. JCTVC-K0033, 1 October 2012 (2012-10-01), XP030112965 *
SCHWARZ H ET AL: "Overview of the Scalable Video Coding Extension of the H.264/AVC Standard", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 17, no. 9, 1 September 2007 (2007-09-01), pages 1103 - 1120, XP011193019, ISSN: 1051-8215, DOI: 10.1109/TCSVT.2007.905532 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR3083662A1 (en) * 2018-07-05 2020-01-10 Awecom, Inc. ROBUST COMPRESSION AND DECOMPRESSION OF DIGITAL IMAGES

Also Published As

Publication number Publication date
FR3008840A1 (en) 2015-01-23
US20160165244A1 (en) 2016-06-09
US9961353B2 (en) 2018-05-01
JP6393323B2 (en) 2018-09-19
KR20160034900A (en) 2016-03-30
JP2016526865A (en) 2016-09-05
EP3022922A1 (en) 2016-05-25
TW201505424A (en) 2015-02-01
CN105659600A (en) 2016-06-08

Similar Documents

Publication Publication Date Title
US11218694B2 (en) Adaptive multiple transform coding
US11575906B2 (en) Image coding device, image decoding device, image coding method, and image decoding method
US9961353B2 (en) Method and device for decoding a scalable stream representative of an image sequence and corresponding coding method and device
JP6285020B2 (en) Inter-component filtering
US8654860B2 (en) Apparatus and method for high efficiency video coding using flexible slice structure
US9800884B2 (en) Device and method for scalable coding of video information
US9532057B2 (en) Inter-layer prediction using sample-adaptive adjustments for bit depth scalable video coding
US20140198846A1 (en) Device and method for scalable coding of video information
KR102272564B1 (en) Image decoding device, image coding device, image decoding method, image coding method and storage medium
US9584808B2 (en) Device and method for scalable coding of video information
US20140092956A1 (en) Adaptive transform options for scalable extension
KR20180053285A (en) Methods of decoding using skip mode and apparatuses for using the same
JP2015521826A (en) Method and device for encoding or decoding images
JPWO2010137323A1 (en) Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
KR20100133006A (en) Dynamic image encoding/decoding method and device
KR102528365B1 (en) Video Encoding and Decoding using Resolution Enhancement Scheme
KR20170114598A (en) Video coding and decoding methods using adaptive cross component prediction and apparatus
KR102505364B1 (en) Decryption device, program, and decryption method
CN108432251B (en) Bit stream conversion device, bit stream conversion method, distribution system, distribution method, and computer-readable storage medium
KR20130105554A (en) Intra prediction method for multi-layered video and apparatus using the same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14744790

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2016526536

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 20167001286

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 14905960

Country of ref document: US

Ref document number: 2014744790

Country of ref document: EP