WO2015016535A1 - Procédé de codage et de décodage supportant une pluralité de couches et appareil utilisant celui-ci - Google Patents

Procédé de codage et de décodage supportant une pluralité de couches et appareil utilisant celui-ci Download PDF

Info

Publication number
WO2015016535A1
WO2015016535A1 PCT/KR2014/006792 KR2014006792W WO2015016535A1 WO 2015016535 A1 WO2015016535 A1 WO 2015016535A1 KR 2014006792 W KR2014006792 W KR 2014006792W WO 2015016535 A1 WO2015016535 A1 WO 2015016535A1
Authority
WO
WIPO (PCT)
Prior art keywords
reference picture
layer
prediction
current
picture set
Prior art date
Application number
PCT/KR2014/006792
Other languages
English (en)
Korean (ko)
Inventor
이배근
김주영
Original Assignee
주식회사 케이티
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020130138517A external-priority patent/KR102175542B1/ko
Application filed by 주식회사 케이티 filed Critical 주식회사 케이티
Priority to CN201480043587.2A priority Critical patent/CN105453562B/zh
Priority to US14/908,214 priority patent/US9894369B2/en
Publication of WO2015016535A1 publication Critical patent/WO2015016535A1/fr
Priority to US15/859,508 priority patent/US10200702B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/58Motion compensation with long-term prediction, i.e. the reference frame for a current frame not being the temporally closest one
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/573Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction

Definitions

  • the present invention relates to video compression technology, and more particularly, to a method and apparatus for performing video coding that supports multiple layers.
  • High efficiency image compression techniques can be used to solve these problems caused by high resolution and high quality image data.
  • An inter-screen prediction technique for predicting pixel values included in the current picture from a picture before or after the current picture using an image compression technique an intra prediction technique for predicting pixel values included in a current picture using pixel information in the current picture
  • One embodiment of the present invention is to provide a method for generating a reference picture list including pictures of another layer with respect to a layer to be currently encoded and decoded, and an apparatus using the same.
  • Another object of the present invention is to provide a method for generating a reference picture list using information on a reference picture set of a reference layer and an apparatus using the same.
  • Another object of the present invention is to provide a method for generating a reference picture list using information on a position at which a reference picture set is added to an initial reference picture list, and an apparatus using the same.
  • flag information indicating whether reference picture set information of a reference layer referenced by the current picture is used Receiving; Generating the initial reference picture list based on the flag information; And performing prediction on the current picture based on the initial reference picture list.
  • the flag information may indicate whether the short-term reference picture set and the long-term reference picture set of the reference layer are derived to the short-term reference picture set and the long-term reference picture set of the current picture.
  • the flag information may indicate whether a short-term reference picture set of the reference layer is derived to a short-term reference picture set of the current picture.
  • the flag information may indicate whether to derive a long-term reference picture set of the reference layer to a long-term reference picture set of the current picture.
  • the reference layer may be a base layer.
  • the flag information may be a position at which the inter-layer reference picture set of the reference layer is added to the initial reference picture list of the reference layer, and a position at which the inter-layer reference picture set of the current picture is added to the initial reference picture list of the current picture. It may indicate whether or not to use.
  • the flag information may indicate whether to induce an interlayer reference picture set of the reference layer to an interlayer reference picture set of the current picture.
  • the flag information may be included in a sequence parameter set and received.
  • the method may further include generating a final reference picture list by modifying the initial reference picture list.
  • the performing of the prediction on the current picture may use a reference picture included in the interlayer reference picture set as a reference picture of the current picture.
  • an image decoding apparatus supporting a plurality of layers includes an entropy decoder which decodes information for prediction and decoding of an image received through a bitstream, and an initial reference picture list of a current picture. Generate the initial reference picture list based on flag information indicating whether reference picture set information of a reference layer referenced by the current picture is used, and perform prediction for the current picture based on the initial reference picture list. It may include a prediction unit to perform.
  • a method for generating a reference picture list including pictures of another layer with respect to a layer to be currently encoded and decoded, and an apparatus using the same are provided.
  • a method of generating a reference picture list using information on a reference picture set of a reference layer and an apparatus using the same are provided.
  • a method for generating a reference picture list using information on a position at which a reference picture set is added to an initial reference picture list, and an apparatus using the same are provided.
  • FIG. 1 is a block diagram schematically illustrating an encoding apparatus according to an embodiment of the present invention.
  • FIG. 2 is a block diagram schematically illustrating a decoding apparatus according to an embodiment of the present invention.
  • FIG. 3 is a conceptual diagram illustrating an example of a reference picture set including a short-term reference picture.
  • FIG. 4 is a conceptual diagram illustrating a method of deriving a POC of a long term reference picture.
  • FIG. 5 is a diagram illustrating an example of configuring a reference picture list.
  • FIG. 6 is a conceptual diagram illustrating an example of configuring an inter-layer reference picture list according to an embodiment of the present invention.
  • FIG. 7 is a diagram illustrating a configuration of a reference picture list O according to an embodiment of the present invention.
  • FIG. 8 is a diagram illustrating a configuration of a reference picture list O according to another example of the present invention.
  • FIG. 9 is a diagram illustrating a configuration of reference picture list 1 according to an embodiment of the present invention.
  • FIG. 10 is a diagram illustrating a configuration of reference picture list 1 according to another example of the present invention.
  • FIG. 11 illustrates a reference picture list constructed by using a multi-view reference picture according to an embodiment of the present invention.
  • FIG. 12 illustrates a reference picture list constructed using a multi-view reference picture according to another example of the present invention.
  • FIG. 13 illustrates a reference picture list constructed by using an index of a temporal sublayer according to an embodiment of the present invention.
  • FIG. 14 illustrates an example of referring to an additional position of an interlayer reference picture set of a reference layer according to an embodiment of the present invention.
  • 15 is a control flowchart for explaining a method of decoding an image according to the present invention.
  • flag information indicating whether reference picture set information of a reference layer referenced by the current picture is used Receiving; Generating the initial reference picture list based on the flag information; And performing prediction on the current picture based on the initial reference picture list.
  • the flag information may indicate whether the short-term reference picture set and the long-term reference picture set of the reference layer are derived to the short-term reference picture set and the long-term reference picture set of the current picture.
  • the flag information may indicate whether a short-term reference picture set of the reference layer is derived to a short-term reference picture set of the current picture.
  • the flag information may indicate whether to derive a long-term reference picture set of the reference layer to a long-term reference picture set of the current picture.
  • the reference layer may be a base layer.
  • the flag information may be a position at which the inter-layer reference picture set of the reference layer is added to the initial reference picture list of the reference layer, and a position at which the inter-layer reference picture set of the current picture is added to the initial reference picture list of the current picture. It may indicate whether or not to use.
  • the flag information may indicate whether to induce an interlayer reference picture set of the reference layer to an interlayer reference picture set of the current picture.
  • the flag information may be included in a sequence parameter set and received.
  • the method may further include generating a final reference picture list by modifying the initial reference picture list.
  • the performing of the prediction on the current picture may use a reference picture included in the interlayer reference picture set as a reference picture of the current picture.
  • an image decoding apparatus supporting a plurality of layers includes an entropy decoder which decodes information for prediction and decoding of an image received through a bitstream, and an initial reference picture list of a current picture. Generate the initial reference picture list based on flag information indicating whether reference picture set information of a reference layer referenced by the current picture is used, and perform prediction for the current picture based on the initial reference picture list. It may include a prediction unit to perform.
  • each of the components disclosed in the embodiments and the drawings of the present invention are disclosed in an independent configuration to represent different characteristic functions of the image encoding apparatus. This does not mean that each component is necessarily made up of separate hardware or one software component. In other words, each component is included in each component for convenience of description, and at least two of the components may be combined into one component, or one component may be divided into a plurality of components to perform a function. The integrated and separated embodiments of the components are also included in the scope of the present invention, without departing from the spirit of the invention.
  • some of the components disclosed in the present invention may not be essential components for performing essential functions in the present invention but may be optional components for improving performance.
  • the present invention can be implemented including only the components essential for implementing the essentials of the present invention except for the components used for improving performance, and the structure including only the essential components except for the optional components used for improving performance. Also included within the scope of the present invention.
  • Encoding and decoding of a video supporting a plurality of layers in a bitstream is called scalable coding. Since there is a strong correlation between the plurality of layers, the prediction may be performed by using this correlation to remove redundant elements of data and to improve the encoding performance of the image.
  • Performing prediction of the current layer, which is a target of prediction using information of another layer, is referred to as inter-layer prediction in the following.
  • At least one of a resolution, a frame rate, and a color format may be different from each other, and upsampling or downsampling of a layer may be performed to adjust the resolution when inter-layer prediction is performed.
  • FIG. 1 is a block diagram schematically illustrating an encoding apparatus according to an embodiment of the present invention.
  • the encoding apparatus 100 includes an encoder 100a for an upper layer and an encoder 100b for a lower layer.
  • the upper rare layer may be expressed as a current layer or an enhancement layer
  • the lower layer may be expressed as a reference layer or a base layer.
  • the upper layer and the lower layer may have at least one of a resolution, a frame rate, and a color format. When a resolution change is necessary to perform inter-layer prediction, upsampling or downsampling of a layer may be performed.
  • the encoder 100a of the upper layer includes a splitter 110, a predictor 100, an intra prediction unit 121, an inter prediction unit 122, an inter layer prediction unit 123, and a transformer 130. , Quantization unit 140, reordering unit 150, entropy encoding unit 160, inverse quantization unit 170, inverse transform unit 180, filter unit 190 and memory 195, and MUX 197. can do.
  • the encoder 100b of the lower layer includes a splitter 111, a predictor 125, an intra prediction unit 126, an inter prediction unit 127, a transform unit 131, a quantizer 141, and a rearrangement.
  • the unit 151 may include an entropy encoding unit 161, an inverse quantization unit 171, an inverse transform unit 181, a filter unit 191, and a memory 196.
  • the encoder may be implemented by the image encoding method described in the following embodiments of the present invention, but operations in some components may not be performed to reduce the complexity of the encoding apparatus or for fast real time encoding.
  • some limited number of methods are used without selecting the optimal intra intra coding method using all intra prediction modes in order to perform encoding in real time.
  • a method of selecting one intra prediction mode among them as a final intra prediction mode using the intra prediction mode of the image may be used.
  • the unit of a block processed by the encoding apparatus may be a coding unit that performs encoding, a prediction unit that performs prediction, or a transformation unit that performs transformation.
  • a coding unit may be represented by a term such as a coding unit (CU), a prediction unit is a prediction unit (PU), and a transformation unit is a transform unit (TU).
  • the splitters 110 and 111 divide a layer image into a combination of a plurality of coding blocks, prediction blocks, and transform blocks, and one of the coding blocks, prediction blocks, and transform blocks according to a predetermined criterion (for example, a cost function). You can split the layer by selecting the combination of. For example, to split a coding unit in a layer image, a recursive tree structure such as a quad tree structure may be used.
  • a recursive tree structure such as a quad tree structure may be used.
  • the meaning of the coding block may be used not only as a block for encoding but also as a block for decoding.
  • the prediction block may be a unit for performing prediction such as intra prediction or inter prediction.
  • the block for performing intra prediction may be a block having a square shape such as 2N ⁇ 2N or N ⁇ N.
  • Predictive block partitioning is performed by using 2NxN, Nx2N, or asymmetric Asymmetric Motion Partitioning (AMP). There is a way.
  • the transform unit 115 may change a method of performing the transform.
  • the prediction units 120 and 125 of the encoders 100a and 100b may include the intra prediction units 121 and 126 performing intra prediction and the inter prediction unit performing inter prediction. (1122, 126).
  • the predictor 120 of the higher layer encoder 100a further includes an inter-layer predictor 123 that performs prediction on the upper layer by using information of the lower layer.
  • the prediction units 120 and 125 may determine whether to use inter prediction or intra prediction on the prediction block.
  • the processing unit in which the prediction is performed and the processing block in which the prediction method is determined may be different. For example, in performing intra prediction, a prediction mode is determined based on a prediction block, and a process of performing prediction may be performed based on a transform block.
  • the residual value (residual block) between the generated prediction block and the original block may be input to the transformers 130 and 131.
  • prediction mode information and motion vector information used for prediction may be encoded by the entropy encoder 130 together with the residual value and transmitted to the decoding apparatus.
  • the original block may be encoded as it is and transmitted to the decoder without performing prediction through the prediction units 120 and 125.
  • PCM Pulse Coded Modulation
  • the intra prediction units 121 and 126 may generate an intra prediction block based on reference pixels present around the current block (the block to be predicted).
  • the intra prediction mode may include a directional prediction mode using reference pixel information according to a prediction direction and a non-directional mode using no directional information when performing prediction.
  • the mode for predicting luma information and the mode for predicting color difference information may be different.
  • intra prediction mode information or predicted luma signal information may be used. If the reference pixel is not available, the prediction block may be generated by replacing the unavailable reference pixel with another pixel.
  • the prediction block may include a plurality of transform blocks. If the prediction block has the same size as the transform block when the intra prediction is performed, pixels present on the left side of the prediction block, pixels present on the upper left side, and top Intra-prediction of the prediction block may be performed based on the pixels present in the. However, when the prediction block is different from the size of the transform block when the intra prediction is included, and a plurality of transform blocks are included in the prediction block, the intra prediction is performed using a reference pixel determined based on the transform block. can do.
  • the intra prediction method may generate a prediction block after applying a mode dependent intra smoothing (MDIS) filter to a reference pixel according to the intra prediction mode.
  • MDIS mode dependent intra smoothing
  • the type of MDIS filter applied to the reference pixel may be different.
  • the MDIS filter is an additional filter applied to the predicted block in the picture by performing the intra prediction, and may be used to reduce the step present in the predicted block in the picture generated after performing the prediction with the reference pixel.
  • filtering on a reference pixel and some columns included in the predicted block in the screen may perform different filtering according to the direction of the intra prediction mode.
  • the inter prediction units 122 and 127 may perform prediction by referring to information of a block included in at least one of a previous picture or a subsequent picture of the current picture.
  • the inter prediction units 122 and 127 may include a reference picture interpolator, a motion predictor, and a motion compensator.
  • the reference picture interpolation unit may receive reference picture information from the memories 195 and 196 and generate pixel information of an integer pixel or less in the reference picture.
  • a DCT-based 8-tap interpolation filter having different filter coefficients may be used to generate pixel information of integer pixels or less in units of 1/4 pixels.
  • a DCT-based interpolation filter having different filter coefficients may be used to generate pixel information of an integer pixel or less in units of 1/8 pixels.
  • the inter prediction units 122 and 127 may perform motion prediction based on the reference picture interpolated by the reference picture interpolator.
  • various methods such as a full search-based block matching algorithm (FBMA), a three step search (TSS), and a new three-step search algorithm (NTS) may be used.
  • the motion vector may have a motion vector value of 1/2 or 1/4 pixel units based on the interpolated pixels.
  • the inter prediction units 122 and 127 may perform prediction on the current block by applying one inter prediction method among various inter prediction methods.
  • various methods such as a skip method, a merge method, and an MVP (Motion Vector Prediction) method, may be used as the inter prediction method.
  • a skip method a merge method
  • MVP Motion Vector Prediction
  • motion information that is, information such as an index of a reference picture, a motion vector, and a residual signal
  • residuals may not be generated, transformed, quantized, or transmitted.
  • the interlayer prediction unit 123 performs interlayer prediction for predicting an upper layer by using information of a lower layer.
  • the inter-layer prediction unit 123 uses inter-layer texture prediction and inter-layer inter prediction by using textures of lower layers, intra prediction mode information, motion information, and syntax information. ), Inter-layer syntax prediction, and the like.
  • Inter-layer texture prediction means using the texture of the reference block in the lower layer as a prediction sample of the current block of the upper layer.
  • the texture of the reference block may be scaled by upsampling.
  • Inter-layer texture prediction includes the intra BL and upsampled lower layers, which upsample the reconstructed values of the reference blocks in the lower layers and encode the residuals with the current blocks using the upsampled reference blocks as predictions for the current block.
  • Intra-prediction of the upper layer may be performed using intra-prediction mode information of the lower layer.
  • the intra-prediction mode of the lower layer may be referred to as a BL intra mode.
  • Inter-layer motion prediction is also called inter-layer inter prediction, and according to inter-layer motion prediction, prediction of a current block of an upper layer may be performed using motion information of a lower layer.
  • the motion information may include a motion vector and a reference picture index.
  • the inter-layer prediction unit 123 may also perform inter-layer syntax prediction for predicting or generating a texture of the current block by using syntax information of the lower layer.
  • the syntax information of the lower layer used for the prediction of the current block may be information about an intra prediction mode, motion information, and the like.
  • the prediction of the current block is performed by using the difference image generated as a difference value between the reconstructed image of the upper layer and the upsampled reconstructed image of the lower layer. Can be.
  • inter-layer prediction As an example of the inter-layer prediction, the inter-layer texture prediction, the inter-layer motion prediction, the inter-layer syntax prediction, and the inter-layer differential prediction have been described, but the inter-layer prediction applicable to the present invention is not limited thereto.
  • a residual block including residual information which is a difference between the predicted block generated by the predictors 120 and 125 and the reconstructed block of the predicted block, is generated, and the residual block is input to the transformers 130 and 131.
  • the transform units 130 and 131 may transform the residual block using a transform method such as a discrete cosine transform (DCT) or a discrete sine transform (DST). Whether DCT or DST is applied to transform the residual block may be determined based on intra prediction mode information of the prediction block used to generate the residual block and size information of the prediction block. That is, the transformers 130 and 131 may apply the transformation method differently according to the size of the prediction block and the prediction method.
  • a transform method such as a discrete cosine transform (DCT) or a discrete sine transform (DST).
  • DCT discrete cosine transform
  • DST discrete sine transform
  • the quantizers 140 and 141 may quantize the values transformed by the transformers 130 and 131 into the frequency domain.
  • the quantization coefficient may change depending on the block or the importance of the image.
  • the values calculated by the quantizers 140 and 141 may be provided to the dequantizers 170 and 17 and the reordering units 150 and 151.
  • the reordering units 150 and 151 may reorder coefficient values with respect to the quantized residual value.
  • the reordering units 150 and 151 may change the two-dimensional block shape coefficients into a one-dimensional vector form through a coefficient scanning method.
  • the realignment units 150 and 151 may scan DC coefficients to coefficients in the high frequency region by using a Zig-Zag scan method and change them into one-dimensional vectors.
  • a vertical scan method for scanning two-dimensional block shape coefficients in a column direction, not a zig-zag scan method, and a horizontal scan method for scanning two-dimensional block shape coefficients in a row direction Can be used. That is, according to the size of the transform block and the intra prediction mode, it is possible to determine which scan method among zigzag-scan, vertical scan and horizontal scan is used.
  • the entropy encoders 160 and 161 may perform entropy encoding based on the values calculated by the reordering units 150 and 151. Entropy encoding may use various encoding methods such as, for example, Exponential Golomb, Context-Adaptive Variable Length Coding (CAVLC), and Context-Adaptive Binary Arithmetic Coding (CABAC).
  • Exponential Golomb Context-Adaptive Variable Length Coding
  • CABAC Context-Adaptive Binary Arithmetic Coding
  • the entropy encoders 160 and 161 transmit residual value coefficient information, block type information, prediction mode information, partition unit information, prediction block information, and the like of the coding block from the reordering units 150 and 151 and the prediction units 120 and 125. Entropy encoding may be performed based on a predetermined encoding method by receiving various information such as unit information, motion vector information, reference frame information, interpolation information of a block, and filtering information. In addition, the entropy encoder 160 or 161 may entropy-encode coefficient values of coding units input from the reordering unit 150 or 151.
  • the entropy encoders 160 and 161 may encode the intra prediction mode information of the current block by performing binarization on the intra prediction mode information.
  • the entropy encoder 160 or 161 may include a codeword mapping unit for performing such a binarization operation, and may perform different binarization according to the size of a prediction block for performing intra prediction.
  • the codeword mapping unit the codeword mapping table may be adaptively generated or stored in advance through a binarization operation.
  • the entropy encoders 160 and 161 may express prediction mode information in the current screen using a codenum mapping unit for performing codenum mapping and a codeword mapping unit for performing codeword mapping. In the codenum mapping unit and the codeword mapping unit, a codenum mapping table and a codeword mapping table may be generated or stored.
  • the inverse quantizers 170 and 171 and the inverse transformers 180 and 181 inverse quantize the quantized values in the quantizers 140 and 141 and inversely transform the converted values in the transformers 130 and 131.
  • the residual values generated by the inverse quantizers 170 and 171 and the inverse transformers 180 and 181 may be predicted by the motion estimator, the motion compensator, and the intra prediction unit included in the predictors 120 and 125. It may be combined with the prediction block to generate a reconstructed block.
  • the filters 190 and 191 may include at least one of a deblocking filter, an offset corrector, and an adaptive loop filter (ALF).
  • a deblocking filter may include at least one of a deblocking filter, an offset corrector, and an adaptive loop filter (ALF).
  • ALF adaptive loop filter
  • the deblocking filter may remove block distortion caused by boundaries between blocks in the reconstructed picture.
  • it may be determined whether to apply a deblocking filter to the current block based on the pixels included in several columns or rows included in the block.
  • a strong filter or a weak filter may be applied according to the required deblocking filtering strength.
  • horizontal filtering and vertical filtering may be performed in parallel when vertical filtering and horizontal filtering are performed.
  • the offset correction unit may correct the offset with respect to the original image on a pixel-by-pixel basis for the deblocking image.
  • the pixels included in the image are divided into a predetermined number of areas, and then, an area to be offset is determined, an offset is applied to the corresponding area, or offset considering the edge information of each pixel. You can use this method.
  • the adaptive loop filter may perform filtering based on a value obtained by comparing the filtered reconstructed image with the original image. After dividing the pixels included in the image into at least one group, one filter to be applied to the group may be determined and filtering may be performed for each group.
  • the filter units 190 and 191 may apply only the deblocking filter, only the deblocking filter and the ALF, or may apply only the deblocking filter and the offset correction unit without applying all of the deblocking filter, the ALF, and the offset correction unit.
  • the memories 195 and 196 may store reconstructed blocks or pictures calculated by the filters 190 and 191, and the stored reconstructed blocks or pictures may be provided to the predictors 120 and 125 when performing inter prediction. have.
  • the information output from the entropy encoder 100b of the lower layer and the information output from the entropy encoder 100a of the upper layer may be multiplexed by the MUX 197 and output as a bitstream.
  • the MUX 197 may be included in the encoder 100b of the lower layer or may be implemented as an independent device or module separate from the encoder 100.
  • FIG. 2 is a block diagram schematically illustrating a decoding apparatus according to an embodiment of the present invention.
  • the decoding apparatus 200 includes a decoder 200a of an upper layer and a decoder 200b of a lower layer.
  • the decoder 200a of the upper layer includes an entropy decoder 210, a reordering unit 220, an inverse quantization unit 230, an inverse transform unit 245, a prediction unit 250, a filter unit 260, and a memory 240. ) May be included.
  • the lower layer decoding unit 200b includes an entropy decoding unit 211, a reordering unit 221, an inverse quantization unit 231, an inverse transform unit 241, a prediction unit 251, a filter unit 261, and a memory 271. ) May be included.
  • the DEMUX 280 may demultiplex information for each layer and transmit the information to the decoders 200a and 200b for each layer.
  • the input bitstream may be decoded in a procedure opposite to that of the encoding apparatus.
  • the entropy decoders 210 and 211 may perform entropy decoding in a procedure opposite to that of the entropy encoder in the encoding apparatus.
  • Information for generating a prediction block among the information decoded by the entropy decoders 210 and 211 is provided to the predictors 250 and 251, and the residual value of the entropy decoding performed by the entropy decoder is the reordering unit 220 or 221. Can be entered.
  • the entropy decoders 210 and 211 may perform inverse transform using at least one of CABAC and CAVLC.
  • the entropy decoders 210 and 211 may decode information related to intra prediction and inter prediction performed by the encoding apparatus.
  • the entropy decoding unit may include a codeword mapping unit and include a codeword mapping table for generating a received codeword as an intra prediction mode number.
  • the codeword mapping table may be stored in advance or generated adaptively.
  • a codenum mapping unit for performing codenum mapping may be additionally provided.
  • the reordering units 220 and 221 may reorder the bitstreams entropy decoded by the entropy decoding units 210 and 211 based on a method of rearranging the bitstreams by the encoder. Coefficients expressed in the form of a one-dimensional vector can be rearranged by restoring the coefficients in a two-dimensional block form.
  • the reordering unit may be realigned by receiving information related to coefficient scanning performed by the encoder and performing reverse scanning based on the scanning order performed by the encoder.
  • the inverse quantization units 230 and 231 may perform inverse quantization based on quantization parameters provided by the encoding apparatus and coefficient values of the rearranged block.
  • the inverse transformers 240 and 241 may perform inverse DCT and inverse DST on the DCT and DST performed by the transformers 130 and 131 with respect to the quantization result performed by the encoding apparatus.
  • the inverse transform may be performed based on a transmission unit determined by the encoding apparatus.
  • DCT and DST may be selectively performed by the transform unit of the encoding apparatus according to a plurality of pieces of information, such as a prediction method, a size and a prediction direction of the current block, and the inverse transform unit 225 of the decoding apparatus may be Inverse transformation may be performed based on the transformation information.
  • the transform may be performed based on the coding block rather than the transform block.
  • the prediction units 250 and 251 may generate the prediction blocks based on the prediction block generation related information provided by the entropy decoding units 210 and 211 and previously decoded blocks or picture information provided by the memories 270 and 271. .
  • the predictors 250 and 251 may include a prediction unit determiner, an inter prediction unit, and an intra prediction unit.
  • the prediction unit discriminator receives various information such as prediction unit information input from the entropy decoder, prediction mode information of the intra prediction method, and motion prediction related information of the inter prediction method, and distinguishes the prediction block from the current coding block. It is possible to determine whether to perform this inter prediction or intra prediction.
  • the inter prediction unit uses information required for inter prediction of the current prediction block provided by the encoding apparatus to the current prediction block based on information included in at least one of a previous picture or a subsequent picture of the current picture including the current prediction block. Inter prediction can be performed. Whether the motion prediction method of the prediction block included in the coding block is skip mode, merge mode, or AMVP mode to perform inter prediction. Can be determined.
  • the intra prediction unit may generate a prediction block based on pixel information in the current picture.
  • intra prediction may be performed based on intra prediction mode information of the prediction block provided by the encoding apparatus.
  • the intra prediction unit is an MDIS filter that performs filtering on the reference pixels of the current block, a reference pixel interpolator which generates reference pixels in pixel units of an integer value or less by interpolating the reference pixels, and filters when the prediction mode of the current block is DC mode. It may include a DC filter for generating a prediction block through.
  • the predictor 250 of the upper layer decoder 200a may further include an inter-layer predictor that performs inter-layer prediction for predicting an upper layer by using information of the lower layer.
  • the inter-layer prediction unit uses inter-layer texture prediction, inter-layer inter prediction, and inter-layer prediction by using textures of lower layers, intra prediction mode information, motion information, and syntax information. Inter-layer syntax prediction may be performed.
  • prediction may be performed using the texture of the reference block in the lower layer as a prediction value of the current block of the upper layer.
  • the texture of the reference block can be scaled by upsampling.
  • Inter-layer texture prediction includes the intra BL and upsampled base layers that upsample the reconstructed values of the reference blocks in the lower layers and encode residuals with the current blocks using the upsampled reference blocks as predictions for the current block. Is stored in memory and uses the stored base layer as a reference index.
  • Intra-prediction of the upper layer may be performed using intra-prediction mode information of the lower layer, and in this case, the intra-prediction mode of the lower layer may be expressed as a BL intra mode.
  • prediction of a current block of an upper layer may be performed using motion information of a lower layer.
  • the inter-layer prediction unit may also perform inter-layer syntax prediction for predicting or generating a texture of the current block using syntax information of the lower layer.
  • the syntax information of the lower layer used for the prediction of the current block may be information about an intra prediction mode, motion information, and the like.
  • the inter-layer prediction unit may perform differential prediction between layers predicting the current block by using the difference image generated as a difference value between the reconstructed image of the upper layer and the resampled image of the lower layer.
  • inter-layer prediction As an example of the inter-layer prediction, the inter-layer texture prediction, the inter-layer motion prediction, the inter-layer syntax prediction, and the inter-layer differential prediction have been described, but the inter-layer prediction applicable to the present invention is not limited thereto.
  • the reconstructed block or picture may be provided to the filter units 260 and 261.
  • the filter units 260 and 261 may include a deblocking filter, an offset corrector, and an ALF.
  • the deblocking filter of the decoding apparatus may receive the deblocking filter related information provided by the encoding apparatus and perform the deblocking filtering on the corresponding block in the decoding apparatus.
  • the offset correction unit may perform offset correction on the reconstructed image based on the type of offset correction and offset value information applied to the image during encoding.
  • the adaptive loop filter may perform filtering based on a value obtained by comparing the restored image with the original image after performing the filtering.
  • the ALF may be applied to the coding unit based on the ALF application information, the ALF coefficient information, and the like provided from the encoding apparatus. Such ALF information may be provided included in a specific parameter set.
  • the memories 270 and 271 may store the reconstructed picture or block to be used as the reference picture or the reference block, and output the reconstructed picture.
  • the encoding apparatus and the decoding apparatus may encode three or more layers instead of two layers, and in this case, the encoder for the upper layer and the decoder for the upper layer are provided in plural numbers corresponding to the number of upper layers. Can be.
  • an embodiment of the present invention uses a coding unit as a coding block for convenience of description, it may be a block for performing decoding as well as encoding.
  • a method of generating a reference picture list described with reference to the drawings according to an embodiment of the present invention may be implemented according to the functions of the respective modules described above with reference to FIGS. 1 and 2, and such an encoding device and a decoding device are rights of the present invention. It is included in a range.
  • a method of encoding / decoding an enhancement layer using motion information and texture information of the base layer, a coding unit (CU) of the base layer Alternatively, there is a method of encoding / decoding using an up-sampled base layer as a reference picture without changing a decoding method in units of prediction units (PUs).
  • the latter method may be expressed in a reference picture index (RefIdx) mode.
  • a picture already decoded in the enhancement layer may be used as a reference picture, or an image of a base layer or a lower layer may be upsampled to be used as a reference picture.
  • the upsampled base layer or lower layer image is referred to as an inter-layer reference picture.
  • inter prediction may be performed using a layer for a view other than the current layer as a reference picture, and a layer of another view referred to for encoding and decoding of the layer may also be inter It may be expressed as a layer reference picture.
  • an interlayer reference picture when constructing a reference picture set of an enhancement layer for coding an image having a multi-layer structure in which scalability is supported instead of a single layer, an interlayer reference picture should also be considered.
  • inter prediction may use at least one of a previous picture or a next picture of a current picture as a reference picture, and perform prediction on the current block based on the reference picture.
  • An image used for prediction of the current block is called a reference picture or a reference frame.
  • the reference picture is specified by a reference picture index refIdx, and a predetermined region in the reference picture is specified as a reference block through a motion vector.
  • the inter prediction may select a reference picture and a reference block corresponding to the current block in the reference picture to generate a prediction block for the current block.
  • the encoding apparatus and the decoding apparatus may derive the motion information of the current block and then perform the inter prediction and / or motion compensation based on the derived motion information. In this case, the encoding apparatus and the decoding apparatus may move a coll block corresponding to the current block in a neighboring block and / or a collocated picture that has already been restored. By using the information, the encoding / decoding efficiency can be improved.
  • the reconstructed neighboring block is a block in the current picture that is already encoded and / or decoded and reconstructed, and may include a block adjacent to the current block and / or a block located at an outer corner of the current block.
  • the encoding apparatus and the decoding apparatus may determine a predetermined relative position based on a block existing at a position spatially corresponding to the current block in the call picture, and the predetermined relative position (spatially corresponding to the current block)
  • the call block may be derived based on the location of the inside and / or outside of the block existing at the location.
  • the call picture may correspond to one picture among the reference pictures included in the reference picture list.
  • a prediction block may be generated such that a residual signal with the current block is minimized and a motion vector size is also minimized.
  • the motion information derivation scheme may vary depending on the prediction mode of the current block.
  • Prediction modes applied for inter prediction may include Advanced Motion Vector Predictor (AMVP), merge, and the like.
  • AMVP Advanced Motion Vector Predictor
  • the encoding apparatus and the decoding apparatus may generate a motion vector candidate list using the reconstructed motion vector of the neighboring block and / or the motion vector of the call block. That is, the motion vector of the reconstructed neighboring block and / or the motion vector of the call block may be used as a motion vector candidate.
  • the encoding apparatus may transmit a predicted motion vector index indicating the optimal motion vector selected from the motion vector candidates included in the list to the decoding apparatus.
  • the decoding apparatus may select the predicted motion vector of the current block from the motion vector candidates included in the motion vector candidate list using the motion vector index.
  • the encoding apparatus may obtain a motion vector difference (MVD) between the motion vector of the current block and the predictive motion vector, and may encode the same and transmit the encoded motion vector to the decoding apparatus.
  • the decoding apparatus may decode the received motion vector difference, and may derive the motion vector of the current block through the sum of the decoded motion vector difference and the predicted motion vector.
  • the encoding apparatus may also transmit a reference picture index or the like indicating the reference picture to the decoding apparatus.
  • the decoding apparatus may predict the motion vector of the current block using the motion information of the neighboring block, and may derive the motion vector for the current block using the residual received from the encoding apparatus.
  • the decoding apparatus may generate a prediction block for the current block based on the derived motion vector and the reference picture index information received from the encoding apparatus.
  • the encoding apparatus and the decoding apparatus may generate the merge candidate list using the motion information of the reconstructed neighboring block and / or the motion information of the call block. That is, when there is motion information of the reconstructed neighboring block and / or the call block, the encoding apparatus and the decoding apparatus may use this as a merge candidate for the current block.
  • the encoding apparatus may select a merge candidate capable of providing an optimal encoding efficiency among the merge candidates included in the merge candidate list as motion information for the current block.
  • a merge index indicating the selected merge candidate may be included in the bitstream and transmitted to the decoding apparatus.
  • the decoding apparatus may select one of the merge candidates included in the merge candidate list by using the transmitted merge index, and determine the selected merge candidate as motion information of the current block. Therefore, when the merge mode is applied, the motion information of the restored neighboring block and / or the call block may be used as the motion information of the current block.
  • the decoding apparatus may reconstruct the current block by adding the prediction block and the residual transmitted from the encoding apparatus.
  • the motion information of the reconstructed neighboring block and / or the motion information of the call block may be used to derive the motion information of the current block.
  • the encoding apparatus does not transmit syntax information such as residual to the decoding apparatus other than the information indicating which block motion information to use as the motion information of the current block.
  • the encoding apparatus and the decoding apparatus may generate a prediction block of the current block by performing motion compensation on the current block based on the derived motion information.
  • the prediction block may mean a motion compensated block generated as a result of performing motion compensation on the current block.
  • the plurality of motion compensated blocks may constitute one motion compensated image.
  • the decoding apparatus may check and derive motion information necessary for inter prediction of the current block, for example, a skip flag, a merge flag, and the like, received from the encoding apparatus, corresponding to the motion vector, the reference picture index, and the like.
  • the processing unit in which the prediction is performed and the processing unit in which the prediction method and the details are determined may be different.
  • a prediction mode may be determined in units of prediction blocks, and prediction may be performed in units of transform blocks, or a prediction mode may be determined in units of prediction blocks, and intra prediction may be performed in units of transform blocks.
  • Pictures encoded / decoded prior to the current picture may be stored in a memory (eg, a Decoded Picture Buffer (DPB)) and used for prediction of the current block (current picture).
  • a memory eg, a Decoded Picture Buffer (DPB)
  • DPB Decoded Picture Buffer
  • the list of pictures available for inter prediction of the current block is maintained as a reference picture list.
  • P slices are slices that are decoded through intra prediction or inter prediction using at most one motion vector and one reference picture.
  • a B slice is a slice decoded through intra prediction or inter prediction using up to two motion vectors and two reference pictures.
  • the reference picture includes a short term reference picture and a long term reference picture.
  • the picture may be specified as a picture order count (POC) indicating a display order, and the short-term reference pictures may be pictures that do not have a large POC difference from the current picture, and the long-term reference pictures may be pictures having a large POC difference from the current picture. Can be.
  • POC picture order count
  • Reference picture list 0 (referred to as 'L0' for convenience of description) is a reference picture list used for inter prediction of a P slice or a B slice.
  • Reference picture list 1 (referred to as 'L1' for convenience of description) is used for inter prediction of a B slice. Therefore, L0 is used for inter prediction on blocks of P slices that perform unidirectional prediction, and L0 and L1 are used for inter prediction on blocks of B slices which perform bidirectional prediction.
  • the decoding apparatus constructs a reference picture list when decoding P slices and B slices through inter prediction.
  • the reference picture used for inter prediction is specified through a reference picture list.
  • the reference picture index is an index indicating a reference picture on the reference picture list.
  • the reference picture list may be configured based on a reference picture set transmitted from the encoding apparatus.
  • the reference picture set may include a POC of a picture used as a reference picture and a flag (used_by_curr_pic_s0_flag) indicating whether the corresponding picture is directly referenced.
  • Reference pictures constituting the reference picture list may be stored in a memory (eg, DPB). Pictures stored in the memory (pictures encoded / decoded before the current picture) are managed by the encoding apparatus and the decoding apparatus.
  • FIG. 3 is a conceptual diagram illustrating an example of a reference picture set including a short-term reference picture.
  • a reference picture set is a POC of a picture for a short-term reference picture that should be stored in the DPB at this point in time for each picture, and a flag indicating whether the current picture directly refers to a specific picture. It can consist of information.
  • Pictures not shown in the reference picture set in the current picture may be removed from the DPB after an unused for reference indicating that they are not used as the reference picture.
  • FIG. 4 is a conceptual diagram illustrating a method of deriving a POC of a long term reference picture.
  • the long-term reference picture may be represented using the least significant bit (LSB) and the most significant bit (MSB) of the POC.
  • MaxPocLsb means the maximum value that can be represented by the LSB. If MaxPocLsb is 32, the long term reference picture (LTRF) having a POC of 84 may be expressed as 32 * 2 + 20, where the LSB is 20 and the MSB is 2.
  • LTRF long term reference picture
  • the maximum value that can be represented by the LSB is 32, it can be expressed as 32 * 10 + 11, so that 10 is the MSB value and 11 is the LSB value.
  • delta_poc_msb_cycle_lt is a value for determining DeltaPocMsbCycleLt which is the MSB of the POC in the long-term reference picture set of the current picture.
  • DeltaPocMsbCycleLt may correspond to the MSB difference value of the POC of the current picture and the MSB of the POC of the reference picture.
  • the POC of the reference picture can be derived by using the LSB value of the POC of the reference picture and the MSB difference of the POC of the current picture and the MSB of the POC of the reference picture.
  • LTRP [0] long-term reference picture having a POC of 331 and a POC indexed by 0 of 308 having an LSB value of 20, and an MSB difference value between the MSB of the POC of the current picture and the POC of the reference picture is Using 1 can be expressed as 331-1 * 32-11 + 20.
  • LTRP [1] long-term reference picture with a POC indexed by 1 of 170 is 331-5 * 32- using an LSB value of 10, the difference between the MSB of the POC of the current picture and the MSB of the POC of the reference picture. It can be expressed as 11 + 10.
  • an initial reference picture list that can be referenced by the current picture may be configured with a list of pictures existing in the short-term reference picture buffer and the long-term reference picture buffer.
  • FIG. 5 is a diagram illustrating an example of configuring a reference picture list.
  • Reference pictures are the first short-term reference picture set RefPicSetStCurr0 consisting of reference pictures Ref 1 and Ref 2 that are smaller than the POC of the current picture Curr based on the current picture, and reference pictures that are larger than the POC of the current picture. It may be classified into a second short-term reference picture set RefPicSetStCurr1 composed of Ref 3 and Ref 4) and a long-term reference picture set RefPicSetLtCurr composed of long-term reference pictures Ref LT1 and Ref LT2.
  • the initial reference picture list may be composed of a set of reference picture sets having different properties.
  • the reference picture list 0, that is, L0 is configured in the order of the first short-term reference picture set RefPicSetStCurr0, the second short-term reference picture set RefPicSetStCurr1, and the long-term reference picture set RefPicSetLtCurr.
  • the reference picture list 1, that is, L1 is configured in the order of the second short-term reference picture set RefPicSetStCurr1, the first short-term reference picture set RefPicSetStCurr0, and the long-term reference picture set RefPicSetLtCurr.
  • the decoding apparatus may use the number of reference pictures specified by the received num_ref_idx_lX_default_active_minus1 plus 1 as a default value in the current sequence.
  • the decoding apparatus may apply a value specified by adding 1 to the received num_ref_idx_l1_active_minus1 as the number of reference pictures for the current picture or the current slice.
  • motion compensation may be performed by using the specified reference picture in the reference picture list configured as described above.
  • a reference picture of an enhancement layer may be composed of reference pictures on the same layer and inter-layer reference picture.
  • signaling for the inter-layer reference picture may be performed through information for identifying a layer and information for identifying a reference picture.
  • a nuh_layer_id value present in the same access unit as the current picture of the i th layer and transmitted in the NAL unit header (i is greater than j) is a nuh_layer_id value. If it is equal to the RefPiclayerId value for, the picture may be determined to be used as a reference picture for the current picture.
  • the interlayer reference picture may be displayed as a long term reference picture.
  • RefPicLayerId is a value that can be signaled by the syntax element inter_layer_pred_layer_idc included in the slice header, and refers to a layer that the current layer refers to inter-layer prediction.
  • FIG. 6 is a conceptual diagram illustrating an example of configuring an inter-layer reference picture list according to an embodiment of the present invention.
  • the initial reference picture list in the multi-layer structure image is a short-term reference picture set (RefPicSetStCurrBefore [i], hereinafter referred to as a first reference picture set) consisting of short-term reference pictures having a smaller POC than the POC of the current picture, and the current picture.
  • a short-term reference picture set (RefPicSetStCurrBefore [i], hereinafter referred to as a first reference picture set) consisting of short-term reference pictures having a smaller POC than the POC of the current picture, and the current picture.
  • a short-term reference picture set (RefPicSetStCurr After [i], hereinafter referred to as the second reference picture set) consisting of short-term reference pictures with a larger POC than the POC, followed by the long-term reference picture set (RefPicSetLtCurr [i], hereinafter referred to as the third reference picture set), and It may consist of an inter-layer reference picture set (RefPicSetILCurr [i], hereinafter referred to as a fourth reference picture set).
  • the reference picture included in the fourth reference picture set may be a picture of a layer supporting spatial scalability, and a picture or depth scalability or quality scalability included in a layer supporting multiview scalability. It may be a picture of a supporting layer. In this case, spatial scalability and image quality scalability may be coded with the same codec structure.
  • the reference pictures for each scalability may be collectively composed of one reference picture set.
  • information about an order of scalability types aligned with the reference picture set may be transmitted from the encoding apparatus to the decoding apparatus.
  • a reference picture set may be separately configured for each of a plurality of scalabilities. That is, a fourth reference picture set consisting of inter-layer reference pictures, a fifth reference picture set, a sixth reference picture set, etc. are generated for the current picture in addition to the first reference picture set, the second reference picture set, and the third reference picture set. Can be.
  • Such an inter-layer reference picture set may be adaptively added to the reference picture list according to the frequency of occurrence. For example, a high frequency reference picture set may be assigned to a low index of the reference picture list.
  • the initial reference picture list of FIG. 6 is L0 in which the first reference picture set is configured in the lowest order of the list.
  • the fourth reference picture set may be added in the last order of the initial reference picture list L0.
  • the initial reference picture list 0 includes pictures having flag information (used_by_curr_pic_flag) of 1 indicating whether the reference picture is included as a reference picture among the reference pictures included in each reference picture set. As shown, the initial reference picture list 0 may be composed of only the first and second pictures among the four pictures included in each reference picture set.
  • the initial reference picture list may be modified for each picture or slice, and the reference pictures may be modified according to the change.
  • the initial reference picture list may be changed by syntax elements included in the slice hair (eg, ref_pic_list_modification_flag_l0, list_entry_l0, ref_pic_list_modification_flag_l1, list_entry_l1).
  • ref_pic_list_modification_flag_l0 and ref_pic_list_modification_flag_l1 explicitly indicate whether reference picture list 0 and reference picture list 1 are transmitted from the encoding device to the decoding device. If the flag value is 1, the reference picture list is explicitly specified using the transmitted reference picture information. If the flag value is 0, the reference picture list is implicitly derived to the initial reference picture set.
  • list_entry_l0 indicates specific entry information constituting the reference picture list for L0, that is, the index of the reference picture
  • list_entry_l1 indicates specific entry information constituting the reference picture list for L1.
  • the final reference picture list 0 is the first and second reference pictures (0, 1) of the first reference picture set, the first and second reference pictures (0, 1) and the first of the second reference picture set. It may consist of the first reference picture (0) of the 4 reference picture set.
  • the fourth reference picture set for the inter-layer reference picture when configuring the reference picture list, may be arranged in an order other than the last in the list of FIG. 6.
  • the inter-layer reference picture since there is a strong correlation between the enhancement layer and the base layer due to the characteristics of the image, the inter-layer reference picture may be frequently referred to.
  • the encoding performance of the reference picture list may be improved by adding the high frequency interlayer reference picture to a position other than the last of the reference picture list.
  • the inter-layer reference picture may be added at various positions of the reference picture list.
  • 7 to 10 are conceptual diagrams illustrating an example of configuring an interlayer reference picture list according to other examples of the present invention.
  • FIG. 7 illustrates a structure of a reference picture list O according to an embodiment of the present invention.
  • a fourth reference picture set which is an interlayer reference picture set, may be arranged in a second order of the reference picture list.
  • the reference picture list O consists of a first reference picture set, a fourth reference picture set, a second reference picture set, and a third reference picture set.
  • Reference picture list 0 is composed of pictures having flag information (used_by_curr_pic_flag) of 1 indicating whether the reference picture is included as a reference picture among reference pictures included in each reference picture set. As shown, the initial reference picture list 0 may be composed of only the first and second pictures among the four pictures included in each reference picture set.
  • the initial reference picture list configured as described above may be modified for each picture or slice, and the entry of the reference pictures may be modified according to the change.
  • the final reference picture list 0 is the first and second reference pictures (0, 1) of the first reference picture set, the first and second reference pictures (0, 1) of the fourth reference picture set, It may consist of the first reference picture (0) of the second reference picture set.
  • FIG. 8 is a diagram illustrating a configuration of a reference picture list O according to another example of the present invention.
  • the reference picture list O includes a first reference picture set composed of short-term reference pictures having a smaller POC than the POC of the current picture, and a second reference picture including a short-term reference picture having a larger POC than the POC of the current picture.
  • the set may be configured in the order of a fourth reference picture set composed of inter-layer reference pictures and a third reference picture set composed of long-term reference pictures.
  • the reference picture list O may include a first reference picture set including a short-term reference picture, in which a fourth reference picture set consisting of inter-layer reference pictures is aligned first, and having a POC smaller than that of the current picture, as shown in FIG.
  • the second reference picture set consisting of short-term reference pictures having a larger POC than the POC of the current picture and the third reference picture set consisting of a long-term reference picture may be ordered.
  • a first reference picture set consisting of short-term reference pictures having a smaller POC than the POC of the current picture, and a short-term reference picture having a POC larger than the POC of the current picture.
  • a reference picture list can be constructed by adding an inter-layer reference picture from the specified position.
  • Such signaling may be encoded in a slice header, a sequence header, or a video parameter set and transmitted to the decoding apparatus.
  • FIG. 9 illustrates a configuration of reference picture list 1 according to an embodiment of the present invention.
  • a fourth reference picture set which is an interlayer reference picture set, may be arranged in the last order of the reference picture list.
  • Reference picture list 1 includes a second reference picture set, a first reference picture set, a third reference picture set, and a fourth reference picture set.
  • Reference picture list 1 includes pictures having flag information (used_by_curr_pic_flag) of 1 indicating whether the reference picture is included as a reference picture among reference pictures included in each reference picture set. As shown, the initial reference picture list 1 may be composed of only the first and second pictures among the four pictures included in each reference picture set.
  • the initial reference picture list configured as described above may be modified for each picture or slice, and the entry of the reference pictures may be modified according to the change.
  • the final reference picture list 1 is used to determine the first reference picture (0) of the second reference picture set, the first and second reference pictures (0, 1) of the first reference picture set, and the fourth reference picture set. It may consist of the first and second reference pictures (0, 1).
  • FIG. 10 is a diagram illustrating a configuration of reference picture list 1 according to another example of the present invention.
  • reference picture list 1 includes a second reference picture set including short-term reference pictures having a larger POC than the POC of the current picture, and a first reference picture including a short-term reference picture having a POC smaller than the POC of the current picture.
  • the set may be configured in order of a fourth reference picture set composed of inter-layer reference pictures and a third reference picture set composed of long-term reference pictures.
  • reference picture list 1 As shown in ⁇ b> of FIG. 10, a second reference picture set composed of short-term reference pictures having a larger POC than the POC of the current picture is allocated to the first order, and the fourth reference picture composed of the interlayer reference pictures Sets can be assigned in second order. Subsequently, reference picture list 1 may be configured in order of a first reference picture set composed of short-term reference pictures having a smaller POC than the POC of the current picture and a third reference picture set composed of long-term reference pictures.
  • a second reference picture set consisting of short-term reference pictures having a larger POC than the POC of the current picture, and a first reference picture including a short-term reference picture having a POC smaller than the POC of the current picture.
  • the position at which the inter-layer reference picture is added may be specified in the slice or sequence level, that is, the slice header or the sequence parameter header.
  • a reference picture list can be constructed by adding an inter-layer reference picture from the specified position.
  • Such signaling may be encoded in a slice header, a sequence header, or a video parameter set and transmitted to the decoding apparatus.
  • a layer for a different view may be used as a reference picture of the current layer.
  • a fourth reference picture set composed of inter-layer reference pictures may be divided and included in the reference picture list.
  • the reference picture list may be configured in various ways according to the viewpoint order of the current layer and the layer of the reference picture or the degree of proximity of the viewpoint.
  • the reference pictures may be separated and sorted when configuring the reference picture list according to the difference between the current view and the reference view, and the lower view and the higher view than the current view may be separated and sorted based on the current view. It may be.
  • FIG. 11 illustrates a reference picture list constructed by using a multi-view reference picture according to an embodiment of the present invention.
  • view 2 may be assumed to be the current layer.
  • a view 0 and a view 1 having a view order lower than the current view based on the current view may constitute one sub reference picture set (first sub reference picture set) or may be included in the first sub reference picture set, and may be included in the current view.
  • Viewpoints 3 and 4 having higher view order may constitute one sub-reference picture set (second sub-reference picture set) or may be included in the second sub-reference picture set.
  • the first sub-reference layer picture set may be composed of pictures having a viewpoint order that is smaller than the viewpoint of the current picture among the multi-view reference layer pictures
  • the second sub-reference layer picture set may be configured of the current picture among the multi-view reference layer pictures. It may be composed of pictures having a view order larger than a view.
  • the reference picture list 0 includes a first reference picture set consisting of short-term reference pictures having a smaller POC than the POC of the current picture, a first sub-reference picture set, and a short-term reference picture having a larger POC than the POC of the current picture.
  • the third reference picture set may include a second reference picture set, a second sub reference picture set, and a long term reference picture.
  • Reference picture list 1 includes a second reference picture set consisting of short-term reference pictures having a larger POC than the POC of the current picture, a second sub-reference picture set, a first reference picture set consisting of a short-term reference picture having a POC smaller than the POC of the current picture,
  • the first reference picture set may be configured in order of the third reference picture set including the long-term reference picture.
  • reference picture list 0 may be a short-term reference picture having a POC larger than the POC of the current picture, the first reference picture set, the first sub-reference picture set, and the POC of the current picture, each having a shorter POC than the POC of the current picture, as shown in ⁇ b>.
  • the second reference picture set may be configured
  • the third reference picture set may be configured by a long-term reference picture
  • the second sub-reference picture set may be configured in this order. That is, the second sub-reference picture set may be added to the end of the reference picture list.
  • Reference picture list 1 similarly includes a second reference picture set consisting of short-term reference pictures having a larger POC than the POC of the current picture, a second sub-reference picture set, and a first reference picture consisting of a short-term reference picture having a POC smaller than the POC of the current picture.
  • the reference picture list may be configured in the order of the set and the third reference picture set including the long-term reference picture, and the first sub-reference picture set may be added in the last order.
  • the multi-layer reference picture set of FIG. 11 includes a multi-view reference picture
  • the type of scalability is spatial scalability, even when pictures of a plurality of layers are used as the reference picture, the reference picture shown in FIG. 11 is used.
  • the construction of the list can be applied. That is, the reference pictures may be divided and added to the reference picture list according to a difference in the order of identifying the current picture and the spatial layer.
  • FIG. 12 illustrates a reference picture list constructed using a multi-view reference picture according to another example of the present invention.
  • the order in which the reference picture is added to the reference picture list may be adaptively changed according to the frequency of occurrence of the reference picture, that is, the proximity order in consideration of the reference degree between viewpoints.
  • a low index of a reference picture list may be allocated to a reference picture of a layer having a small difference between the current layer and a viewpoint
  • a high index of a reference picture list may be allocated to a reference picture of a layer having a large difference between the current layer and a viewpoint.
  • reference picture 0 among reference picture 0, reference picture 1, reference picture 2, and reference picture 3, which may be included in a multi-view reference picture set, is arranged in front of reference picture list 0.
  • the reference picture 1, the reference picture 2, and the reference picture 3 may be aligned at the end of the reference picture list 0 which is the next order of the third reference picture set.
  • the reference picture 0 may be a reference picture of a layer having a small difference between the current layer and a viewpoint
  • the reference picture 1, the reference picture 2, and the reference picture 3 may be pictures having a large difference between the current layer and the viewpoint.
  • the difference value between the current layer and the start point of the reference layer that may be aligned in the front in the reference picture list may be a default value, and may be variably set in the encoding apparatus and transmitted to the decoding apparatus.
  • the order of the inter-layer reference picture set may be set according to the index (tempralId) of the temporal sub-layer of the current picture.
  • FIG. 13 illustrates a reference picture list constructed by using an index of a temporal sublayer according to an embodiment of the present invention.
  • the frequency of using the inter-layer reference picture may vary according to the temporal sublayer of the current picture belonging to the enhancement layer. As the tempralId which is an index value of the temporal sublayer is larger, it may be more effective to perform inter prediction with reference to pictures in the same layer than to reference pictures of other layers.
  • the tempralId is small, it is more effective to code the interlayer picture than the same layer, and if the tempralId is larger, it is better to code the picture of the same layer as the reference picture than to use the interlayer as the reference picture. It can be efficient.
  • the position at which the interlayer reference picture is added to the initial reference picture list is adaptively changed according to the index tempralId of the temporal sublayer. Can be.
  • FIG. 13 is a diagram illustrating the configuration of a reference picture list 0 when the tempralId value is 0 to 2.
  • the fourth reference picture set composed of the interlayer reference pictures may be added to the beginning of the initial reference picture list.
  • An initial reference picture list consisting of pictures having a flag information (used_by_curr_pic_flag) of 1 indicating whether the reference picture included in each reference picture set is used as a reference picture is four pictures included in the fourth reference picture set (0 , 1, 2, 3), the first reference picture set, the second reference picture set, and the first and second pictures (0, 1) included in the third reference picture set.
  • flag information used_by_curr_pic_flag
  • the reference picture list may be modified and then the final reference picture list may be set.
  • Information for modifying the reference picture list may be signaled in the slice header.
  • the modified last reference picture list 0 is the first and second pictures (0,1) of the first reference picture set, the first and second pictures (0,1) of the fourth reference picture set, and the second reference picture of the second reference picture set. It consists of the first picture (0).
  • the fourth reference picture set including the interlayer reference picture may be added after the first reference picture set in the initial reference picture list 0.
  • the initial reference picture list which consists of pictures with flag information (used_by_curr_pic_flag) equal to 1 indicating whether it is used as a reference picture, is included in the first reference picture set, the fourth reference picture set, the second reference picture set, and the third reference picture set. It consists of the first and second pictures (0, 1).
  • the modified last reference picture list 0 is the first and second pictures (0,1) of the first reference picture set, the first and second pictures (0,1) of the fourth reference picture set, and the second reference picture of the second reference picture set. It consists of the first picture order.
  • the tempralId value is 2
  • An initial reference picture list 0 consisting of pictures with flag information (used_by_curr_pic_flag) of 1 indicating whether it is used as a reference picture is assigned to the first reference picture set, the second reference picture set, and the third reference picture set and the fourth reference picture set. It consists of the first and second pictures (0, 1) included.
  • the modified last reference picture list includes the first and second pictures (0,1) of the first reference picture set, the first and second pictures (0,1) of the second reference picture set, and the first of the fourth reference picture sets. It consists of the first picture (0).
  • the initial reference list L1 may also change the position where the fourth reference picture set is added according to the temporalId value.
  • the initial reference list L1 may be configured in the order of the fourth reference picture set, the second reference picture set, the first reference picture set, and the third reference picture set, and the temporalId value is one.
  • the second reference picture set, the fourth reference picture set, the first reference picture set, and the third reference picture set may be configured in this order.
  • the fourth reference picture set may be added in the last order of the initial reference list L1.
  • the position at which the interlayer reference picture is added to the initial reference picture list may be adaptively changed according to the temporalId value of the temporal sublayer.
  • the order of configuring the reference picture list may be changed according to the difference in the layer index temporalId value of the temporal sublayer between layers in the multi-layer structure.
  • the enhancement layer refers to the base layer, that is, there is an interlayer reference relationship.
  • the temporalId value of the enhancement layer and the baseor temporalId value are the same or the difference between the temporalId values is small, the possibility that the enhancement layer refers to the base layer, that is, the existence of an inter-layer reference relationship increases.
  • the position at which the interlayer reference picture is added to the reference picture list can be adaptively changed according to the difference of the temporalId values between the layers. .
  • the interlayer reference picture set is added to the beginning of the reference picture list, and the picture belonging to the two layers is added. If the temporalId value is different, the inter-layer reference picture set may be added at the end after the long-term reference picture set.
  • an interlayer reference picture set when adding an interlayer reference picture to the reference picture list, if the difference between the temporalId value of the picture belonging to the enhancement layer and the temporalId value of the picture belonging to the base layer or the reference layer is smaller than a specific value, the first part of the reference picture list is displayed.
  • an inter-layer reference picture set may be added at the end after the long-term reference picture set.
  • an interlayer reference picture when added to the front of the reference picture list, it may be added as a second reference picture set other than the first in the reference picture list, or when the interlayer reference picture is added to the back of the reference picture list. It may be added before the last non-long-term reference picture set of the list.
  • the interlayer reference picture may be added to the beginning of the reference picture list.
  • the interlayer reference picture is added to the reference picture list in the order of the long-term reference picture set. Can be.
  • a reference picture list may be adaptively constructed according to the number of maximum time base sublayers that allow interlayer prediction.
  • the bitstream supporting the multi-layer may include information about a maximum temporal sub-layer that allows inter-layer prediction. Such information may be transmitted in syntax elements of max_tid_il_ref_pics_plus1 included in the video parameter set. A value obtained by subtracting 1 from max_tid_il_ref_pics_plus1 represents the maximum number of time base sublayers that allow interlayer prediction.
  • the interlayer reference when constructing the reference picture list, if the temporalId value of the current layer is larger than the value specified by subtracting 1 from max_tid_il_ref_pics_plus1, it is better to use the picture on the same layer as the reference picture than the upsampled base layer picture. It is advantageous in terms of encoding and decoding efficiency. Therefore, when the temporalId value of the current layer is larger than the value specified by max_tid_il_ref_pics_plus1-1, the interlayer reference may not be used. In this case, the interlayer reference picture set may not be added to the reference picture list, or a separate signal may not be signaled even if the interlayer reference picture set is added.
  • the reference picture list may be configured by combining the short-term reference picture set, the long-term reference picture set, and the inter-layer reference picture set for the current layer.
  • the reference picture set may be configured according to the embodiments described with reference to FIGS. 6 to 13 and the embodiments that can be modified in the embodiments.
  • a lower layer whose temporalId value is greater than the value obtained by subtracting 1 from max_tid_il_ref_pics_plus1 is not referenced when predicting the current layer.
  • inter prediction does not refer to a slice or picture whose termporalId value is larger than the current slice or picture
  • the slice or picture of a base layer having a termporalId value larger than the termporalId value of the current slice or picture is not referred to. Can be.
  • the reference picture list may be constructed using only the short-term reference picture set and the long-term reference picture set for the current layer.
  • an image of an enhancement layer and a base layer in a bitstream with spatial scalability, or an image of a neighboring layer near a current layer in a bitstream with multiview scalability is similar to each other may have similar characteristics. Can be. In such a case, there may be a similarity of the reference picture set.
  • prediction of an enhancement layer may be performed using reference picture set information of a base layer as a reference picture set of an enhancement layer.
  • the encoding apparatus and the decoding apparatus may use a reference picture set of a layer adjacent to a base layer, another lower layer, or a current layer to be encoded or decoded (hereinafter, referred to as a reference layer) as it is in the current layer.
  • a reference layer a current layer to be encoded or decoded
  • the encoding apparatus and the decoding apparatus may obtain reference picture set information of the reference layer, that is, information about a reference picture constituting the reference picture set and flag information indicating whether or not to use the encoder from the encoder and the decoder encoding or decoding the reference layer. This information can be used as is for the prediction of the enhancement layer, that is, the current layer.
  • the base layer may use information of the short-term reference picture set and the long-term reference picture set.
  • the enhancement layer uses a reference picture set of the base layer, such signaling may be performed through a sequence parameter set (SPS).
  • SPS sequence parameter set
  • Table 1 below shows the SPS according to an embodiment of the present invention.
  • the inter_layer_copy_flag value when the inter_layer_copy_flag value is 1, it indicates that the short-term reference picture set and the long-term reference picture set of the reference layer referenced by the current layer are used as the short-term reference picture set and the long-term reference picture set of the current layer. If the inter_layer_copy_flag value is 0, the slice header of the current layer rather than the reference layer is decoded to derive the short-term reference picture set and the long-term reference picture set of the current layer.
  • inter_layer_copy_flag information may be included in the SPS and transmitted.
  • flag information may be included in the picture level or the slice level and transmitted.
  • num_short_term_ref_pic_sets represents the number of short-term reference picture sets included in the SPS. As shown in Table 1, num_short_term_ref_pic_sets is received when inter_layer_copy_flag is 0.
  • a long_term_ref_pics_present_flag of 0 indicates that there is no long-term reference picture used for prediction of the current layer picture
  • a long_term_ref_pics_present_flag of 1 indicates that at least one long-term reference picture is used for prediction of the current layer picture.
  • num_long_term_ref_pics_sps represents the number of long-term reference picture sets included in the SPS.
  • lt_ref_pic_poc_lsb_sps [i] represents MaxPocLsb of the i-th candidate long-term reference picture specified in the SPS.
  • used_by_curr_pic_lt_sps_flag [i] When used_by_curr_pic_lt_sps_flag [i] is 0, it indicates that the i-th candidate long-term reference picture specified in the SPS is not used as a reference picture of a picture that includes the i-th candidate long-term reference picture in the corresponding long-term reference picture set.
  • the reference layer referenced by the current layer is a base layer
  • the short-term reference picture set and the long-term reference picture set of the base layer are It may be used as it is.
  • flag information such as inter_layer_copy_flag may be included in the SPS and signaled.
  • the encoding apparatus and the decoding apparatus may use a part of the reference picture set of the reference layer as the reference picture set of the current layer. That is, when at least one of the short-term reference picture set or the long-term reference picture set of the reference layer is used as the reference picture set of the current layer, it may be signaled whether each reference picture set is used.
  • Table 2 shows the SPS according to an embodiment of the present invention.
  • inter_layer_short_copy_flag value 1
  • the inter_layer_short_copy_flag value 0
  • the slice header of the current layer rather than the reference layer is decoded to derive the short-term reference picture set of the current layer.
  • inter_layer_short_copy_flag information may be included in the SPS and transmitted.
  • flag information may be included in the picture level or the slice level and transmitted.
  • num_short_term_ref_pic_sets represents the number of short-term reference picture sets included in the SPS. As shown in Table 2, num_short_term_ref_pic_sets is received when inter_layer_short_copy_flag is 0.
  • a long_term_ref_pics_present_flag of 0 indicates that there is no long-term reference picture used for prediction of the current layer picture
  • a long_term_ref_pics_present_flag of 1 indicates that at least one long-term reference picture is used for prediction of the current layer picture.
  • An inter_layer_long_copy_flag value of 1 indicates that the long-term reference picture set of the reference layer referenced by the current layer is used as the long-term reference picture set of the current layer. If the value of inter_layer_ long_copy_flag is 0, the slice header of the current layer rather than the reference layer is decoded to derive the long-term reference picture set of the current layer.
  • inter_layer_long_copy_flag is 0 and long_term_ref_pics_present_flag is 1, num_long_term_ref_pic_sps indicating the number of long-term reference pictures included in the SPS is signaled.
  • lt_ref_pic_poc_lsb_sps [i] represents MaxPocLsb of the i-th candidate long-term reference picture specified in the SPS.
  • used_by_curr_pic_lt_sps_flag [i] When used_by_curr_pic_lt_sps_flag [i] is 0, it indicates that the i-th candidate long-term reference picture specified in the SPS is not used as a reference picture of a picture that includes the i-th candidate long-term reference picture in the corresponding long-term reference picture set.
  • the reference layer referenced by the current layer is a base layer
  • the short-term reference picture set and the long-term reference picture set of the base layer are It may be used as it is.
  • the SPS may include flag information such as inter_layer_ short_copy_flag or inter_layer_long_copy_flag.
  • the copy information for each of the short-term reference picture set and the long-term reference picture set may be signaled instead of the copy flag for the entire reference picture set of the current layer.
  • the current layer may use the interlayer reference picture set of the reference layer as the interlayer reference picture set of the current layer, and in this case, flag information indicating whether to copy the interlayer reference picture set is signaled. Can be.
  • the flag information described above may be included in a sequence parameter set and signaled as shown in Tables 1 and 2, and may be transmitted in a video parameter set, or included in a picture level or a slice level.
  • position information to which the inter-layer reference picture is added to the reference picture list may be referenced.
  • the reference picture set of the current layer is added to the reference picture list
  • the position information of the reference layer may be used. The method of adding the interlayer reference picture to the reference picture list is the same as described with reference to FIGS. 6 to 13.
  • FIG. 14 illustrates an example of referring to an additional position of an interlayer reference picture set of a reference layer according to an embodiment of the present invention.
  • the current layer is assigned the order in which the inter-layer reference picture for the reference layer is added to the initial reference picture list, or a specific position to add to the reference picture list. Can also be used as it is during encoding or decoding.
  • a fourth reference picture set composed of inter-layer reference pictures may be added after picture 0 and picture 1 constituting the first reference picture set when added to an initial reference picture list. That is, the inter-layer reference pictures may be sorted in order of the second reference picture set of the initial reference picture list.
  • the picture 0 constituting the second reference picture set having used_by_curr_pic_s0_flag equal to 1 the picture 0 constituting the picture 1, and the third reference picture set are aligned.
  • the modified initial reference picture list is finally composed of picture 0 of the first reference picture set, picture 1, picture 0 of the fourth reference picture set, picture 1 and picture 0 of the second reference picture set.
  • inter-layer reference pictures may also be arranged in a second order after the first reference picture set in the initial reference picture list for the current layer.
  • such position information may be signaled as flag information indicating whether the position of the inter-layer reference picture set of the reference layer is merged.
  • the ID of the reference layer into which the location information is merged may be additionally signaled.
  • the encoding apparatus may encode the position information of the position at which the inter-layer reference picture set is added to the initial reference picture list in a slice or sequence level, for example, a slice header or a sequence header.
  • a slice or sequence level for example, a slice header or a sequence header.
  • the temporalId value may be different for each multilayer, or the frequency of interlayer references may vary according to the temporalId value.
  • a position for adding an interlayer reference picture for each layer may be signaled.
  • the encoding apparatus may limit a position at which a reference picture is added to a predetermined number, and determine a position at which an interlayer reference picture may be added as many as the number. After setting, the bit distortion rate can be calculated only for the set position. An optimal position at which the interlayer reference picture may be added may be derived based on the calculated bit distortion rate, and the encoding apparatus may transmit the position to the decoding apparatus.
  • an inter-layer reference picture may be set at three positions as follows, after the first and short-term reference picture sets in the initial reference picture list and after the long-term reference picture sets.
  • the encoding apparatus may signal by encoding an identification index of the position where the interlayer reference picture is added among the three positions.
  • 15 is a control flowchart illustrating a method of generating a reference picture list according to the present invention.
  • the decoding apparatus receives flag information indicating whether a current picture uses reference picture set information of a reference layer (S1510).
  • Such flag information may be flag information indicating whether the short-term reference picture set and the long-term reference picture set of the reference layer are derived to the short-term reference picture set and the long-term reference picture set of the current picture.
  • the flag information may indicate whether only the short-term reference picture set of the reference layer is derived to the short-term reference picture set of the current picture, or the flag information indicates whether to derive only the long-term reference picture set of the reference layer to the long-term reference picture set of the current picture. It may be information.
  • the reference The layer may be a base layer that does not refer to an inter-layer reference picture.
  • the flag information may indicate whether the interlayer reference picture set of the reference layer is derived to the interlayer reference picture set of the current picture. That is, the current picture may refer to a picture of another layer referenced by the reference layer.
  • the flag information is used as a position for adding the inter-layer reference picture set of the reference layer to the initial reference picture list as the position for adding the inter-layer reference picture set of the current picture to the initial reference picture list as described in FIG. 13. It can also indicate whether or not.
  • the information on the position at which the interlayer reference picture set of the reference layer is added to the initial reference picture list is not included in the flag information but at the position index at which the interlayer reference picture set is added to the initial reference picture list. It may be signaled as difference value information, that is, a prediction value. For example, if the position where the inter-layer reference picture set of the reference layer is added to the initial reference picture list is the first order, and the position at which the inter-layer reference picture set of the current layer is added to the initial reference picture list is the second order, encoding is performed.
  • the device may encode and transmit difference value information for the first position and the second position.
  • the decoding apparatus may derive the position where the inter-layer reference picture set of the current layer is added to the initial reference picture list using the signaled difference value information.
  • Such flag information may be included and signaled in a video level, a sequence level, a picture level, or a slice level.
  • the decoding apparatus generates an initial reference picture list based on the flag information (S1520).
  • the reference picture set of the reference layer may be used as it is.
  • the initial reference picture list of the current picture is independently configured without information of the reference layer.
  • the prediction unit may modify the initial reference picture list based on the change information for the additional reference picture list.
  • the prediction unit performs prediction on the current picture based on the initial reference picture list or the modified last reference picture list (S1530).
  • a reference picture included in the interlayer reference picture set is represented as a long-term reference picture, and the decoding apparatus may use a reference picture included in the interlayer reference picture set as a reference picture of the current picture.
  • the present invention can be used to code a video signal of a multilayer structure.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

La présente invention concerne un procédé de décodage d'image supportant une pluralité de couches, pouvant comprendre les étapes consistant à : quand une liste initiale d'images de référence d'une image en cours est configurée, recevoir des informations d'indicateur indiquant si des informations d'ensemble d'images de référence d'une couche de référence à laquelle l'image en cours fait référence sont utilisées; générer la liste initiale d'images de référence en fonction des informations d'indicateur; et prédire l'image en cours en fonction de la liste initiale d'images de référence. Par conséquent, la présente invention concerne un procédé permettant de générer une liste d'images de référence comprenant une image d'une couche, qui est différente d'une couche devant être actuellement codée et décodée, et concerne un appareil utilisant celui-ci.
PCT/KR2014/006792 2013-07-30 2014-07-25 Procédé de codage et de décodage supportant une pluralité de couches et appareil utilisant celui-ci WO2015016535A1 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN201480043587.2A CN105453562B (zh) 2013-07-30 2014-07-25 支持多个层的图像编码和解码方法以及使用该方法的装置
US14/908,214 US9894369B2 (en) 2013-07-30 2014-07-25 Image encoding and decoding method supporting plurality of layers and apparatus using same
US15/859,508 US10200702B2 (en) 2013-07-30 2017-12-30 Image encoding and decoding method supporting plurality of layers and apparatus using same

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
KR20130090228 2013-07-30
KR20130090227 2013-07-30
KR10-2013-0090228 2013-07-30
KR10-2013-0090227 2013-07-30
KR1020130138517A KR102175542B1 (ko) 2013-07-15 2013-11-14 복수의 레이어를 지원하는 영상의 부호화 및 복호화 방법 및 이를 이용하는 장치
KR10-2013-0138517 2013-11-14

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US14/908,214 A-371-Of-International US9894369B2 (en) 2013-07-30 2014-07-25 Image encoding and decoding method supporting plurality of layers and apparatus using same
US15/859,508 Continuation US10200702B2 (en) 2013-07-30 2017-12-30 Image encoding and decoding method supporting plurality of layers and apparatus using same

Publications (1)

Publication Number Publication Date
WO2015016535A1 true WO2015016535A1 (fr) 2015-02-05

Family

ID=52432014

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2014/006792 WO2015016535A1 (fr) 2013-07-30 2014-07-25 Procédé de codage et de décodage supportant une pluralité de couches et appareil utilisant celui-ci

Country Status (1)

Country Link
WO (1) WO2015016535A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070286283A1 (en) * 2004-10-13 2007-12-13 Peng Yin Method And Apparatus For Complexity Scalable Video Encoding And Decoding
KR20080037593A (ko) * 2006-10-25 2008-04-30 한국전자통신연구원 다시점 비디오의 스케일러블 코딩 및 디코딩 방법과, 코딩및 디코딩 장치
JP2009522888A (ja) * 2006-01-10 2009-06-11 トムソン ライセンシング スケーラブル・ビデオ用参照ピクチャ・リストを構築するための方法及び装置
US20120050475A1 (en) * 2009-05-01 2012-03-01 Dong Tian Reference picture lists for 3dv
KR20130000334A (ko) * 2011-06-15 2013-01-02 광운대학교 산학협력단 스케일러블 비디오 코딩 및 디코딩 방법과 이를 이용한 장치

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070286283A1 (en) * 2004-10-13 2007-12-13 Peng Yin Method And Apparatus For Complexity Scalable Video Encoding And Decoding
JP2009522888A (ja) * 2006-01-10 2009-06-11 トムソン ライセンシング スケーラブル・ビデオ用参照ピクチャ・リストを構築するための方法及び装置
KR20080037593A (ko) * 2006-10-25 2008-04-30 한국전자통신연구원 다시점 비디오의 스케일러블 코딩 및 디코딩 방법과, 코딩및 디코딩 장치
US20120050475A1 (en) * 2009-05-01 2012-03-01 Dong Tian Reference picture lists for 3dv
KR20130000334A (ko) * 2011-06-15 2013-01-02 광운대학교 산학협력단 스케일러블 비디오 코딩 및 디코딩 방법과 이를 이용한 장치

Similar Documents

Publication Publication Date Title
WO2020171444A1 (fr) Procédé et dispositif d'inter-prédiction basés sur un dmvr
WO2020004990A1 (fr) Procédé de traitement d'image sur la base d'un mode de prédiction inter et dispositif correspondant
WO2017069505A1 (fr) Procédé de codage/décodage d'image et dispositif correspondant
WO2019194514A1 (fr) Procédé de traitement d'image fondé sur un mode de prédiction inter et dispositif associé
WO2015133838A1 (fr) Procédé de codage/décodage d'image basé sur une unité polygonale et appareil associé
WO2015034215A1 (fr) Appareil et procédé de codage/décodage de signal vidéo évolutif
WO2016190627A1 (fr) Procédé et dispositif pour traiter un signal vidéo
WO2015060614A1 (fr) Procédé et dispositif pour coder/décoder un signal vidéo multi-couches
WO2019194653A1 (fr) Procédé de traitement d'image de fourniture de processus de mode de fusion complexe d'informations de mouvement, procédé de codage et de décodage d'image l'utilisant, et appareil associé
WO2021172914A1 (fr) Procédé de décodage d'image pour un codage résiduel et dispositif associé
WO2015099400A1 (fr) Procédé et appareil de codage/décodage d'un signal vidéo multicouche
WO2015064990A1 (fr) Procédé et dispositif d'encodage/décodage de signal vidéo multicouche
WO2015037920A1 (fr) Procédé et appareil de codage/décodage d'un signal vidéo évolutif
WO2015147426A1 (fr) Procédé et dispositif de codage/décodage de signal vidéo multicouche
WO2021182816A1 (fr) Appareil et procédé d'encodage/de décodage d'image pour encoder sélectivement les informations de taille de tranche rectangulaire, et procédé de transmission de flux binaire
WO2021145669A1 (fr) Procédé et appareil d'inter-prédiction dans un système de codage vidéo
WO2020262929A1 (fr) Procédé et dispositif de signalisation de syntaxe dans un système de codage d'image/de vidéo
WO2021006698A1 (fr) Procédé et dispositif de codage d'image dans un système de codage d'image
WO2020256393A1 (fr) Codage de vidéo ou d'image basé sur un mappage de luminance
WO2020017786A1 (fr) Procédé de décodage d'image sur la base d'informations d'image contenant des informations de prédiction intra dans un système de codage d'image, et appareil associé
WO2018117686A2 (fr) Procédé et dispositif de codage ou de décodage de signal vidéo
WO2021256785A1 (fr) Procédé et appareil de codage/décodage d'image sur la base d'informations de msb de poc, et support d'enregistrement pour stocker un flux binaire
WO2021201628A1 (fr) Procédé et appareil de codage/décodage pour déterminer s'il faut ou non se référer à un ensemble de paramètres d'après un ensemble de couches de sortie, et procédé de transmission de flux binaire
WO2021172916A1 (fr) Procédé de décodage d'image et appareil associé
WO2021172907A1 (fr) Procédé de décodage d'image et appareil associé

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201480043587.2

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14831453

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 14908214

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 13/05/2016)

122 Ep: pct application non-entry in european phase

Ref document number: 14831453

Country of ref document: EP

Kind code of ref document: A1