WO2015057036A1 - 멀티-뷰 비디오를 디코딩하는 비디오 디코딩 방법 및 장치 - Google Patents
멀티-뷰 비디오를 디코딩하는 비디오 디코딩 방법 및 장치 Download PDFInfo
- Publication number
- WO2015057036A1 WO2015057036A1 PCT/KR2014/009858 KR2014009858W WO2015057036A1 WO 2015057036 A1 WO2015057036 A1 WO 2015057036A1 KR 2014009858 W KR2014009858 W KR 2014009858W WO 2015057036 A1 WO2015057036 A1 WO 2015057036A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- block
- prediction
- current block
- view
- current
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/117—Filters, e.g. for pre-processing or post-processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
- H04N19/463—Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/91—Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/80—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
- H04N19/82—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
Definitions
- the present invention relates to video coding, and more particularly, to coding of 3D video images.
- High-efficiency image compression technology can be used to effectively transmit, store, and reproduce high-resolution, high-quality video information.
- 3D video can provide realism and immersion using a plurality of view channels.
- 3D video can be used in a variety of areas such as free viewpoint video (FVV), free viewpoint TV (FTV), 3DTV, social security and home entertainment.
- FVV free viewpoint video
- FTV free viewpoint TV
- 3DTV social security and home entertainment.
- 3D video using multi-view has a high correlation between views of the same picture order count (POC). Since the multi-view image captures the same scene at the same time by using several adjacent cameras, that is, multiple views, the correlation between the different views is high because it contains almost the same information except for parallax and slight lighting differences.
- POC picture order count
- the decoding target block of the current view may be predicted or decoded with reference to the block of another view.
- An object of the present invention is to provide a method and apparatus for effectively deriving a predictive sample of a current block using information of another view.
- An object of the present invention is to provide a method and apparatus for effectively performing entropy coding using information of neighboring blocks.
- An object of the present invention is to provide a method and apparatus for entropy coding information for applying residual prediction using information of another view, in consideration of whether the same scheme is applied to neighboring blocks.
- One embodiment of the invention is a video decoding apparatus for decoding multi-view video
- the video apparatus entropy decodes a bitstream to decode information about residual prediction by referring to neighboring blocks of a current block, a memory to store pictures referred to in decoding the current block, and information about the residual. Reconstructs the prediction unit and the prediction sample to derive a prediction sample for the current block based on the difference in the sample values between the first block and the second block corresponding to the current block in the pictures stored in the memory. And a filtering unit configured to apply filtering to the current picture, wherein at least one of the first block and the second block belongs to a different view from the view of the current block, and the first block is a motion vector or disc of the current block. Specified by a parity vector, the second block is a motion vector and a disparity of the current block It may be specified by the vector.
- Another embodiment of the present invention is a video decoding method for decoding multi-view video, the video decoding method comprising entropy decoding information on residual prediction with reference to a neighboring block of a current block, the information on the residual Deriving a predictive sample for the current block based on a difference in sample values between the first block and the second block corresponding to the current block by using and applying filtering to the reconstructed current picture using the predictive sample And at least one of the first block and the second block belongs to a different view than the view of the current block, wherein the first block is specified by a motion vector or disparity vector of the current block, and The second block may be specified by the motion vector and the disparity vector of the current block.
- entropy coding can be effectively performed using information of neighboring blocks.
- 1 is a diagram schematically illustrating a process of encoding and decoding 3D video.
- FIG. 2 is a diagram schematically illustrating a configuration of a video encoding apparatus.
- FIG. 3 is a diagram schematically illustrating a configuration of a video decoding apparatus.
- FIG. 4 is a diagram schematically illustrating inter view coding.
- FIG. 5 schematically illustrates a multi-view coding method using a depth map.
- FIG. 6 is a diagram schematically illustrating residual prediction according to the present invention.
- FIG. 9 is a flowchart schematically illustrating an operation of a video encoding apparatus according to the present invention.
- FIG. 10 is a flowchart schematically illustrating an operation of a video decoding apparatus according to the present invention.
- a pixel or a pel may mean a minimum unit constituting one image.
- the term 'sample' may be used as a term indicating a value of a specific pixel.
- the sample generally indicates the value of the pixel, but may indicate only the pixel value of the Luma component or only the pixel value of the Chroma component.
- a unit may mean a basic unit of image processing or a specific position of an image.
- the unit may be used interchangeably with terms such as 'block' or 'area' as the case may be.
- an M ⁇ N block may represent a set of samples or transform coefficients composed of M columns and N rows.
- 1 is a diagram schematically illustrating a process of encoding and decoding 3D video.
- the 3D video encoder may encode a video picture, a depth map, and a camera parameter to output a bitstream.
- the depth map may be composed of distance information (depth information) between a camera and a subject with respect to pixels of a corresponding video picture (texture picture).
- the depth map may be an image in which depth information is normalized according to bit depth.
- the depth map may be composed of recorded depth information without color difference representation.
- disparity information indicating the correlation between views may be derived from depth information of the depth map using camera parameters.
- a general color image that is, a bitstream including a depth map and camera information together with a video picture (texture picture) may be transmitted to a decoder through a network or a storage medium.
- the decoder side can receive the bitstream and reconstruct the video.
- the 3D video decoder may decode the video picture and the depth map and the camera parameters from the bitstream. Based on the decoded video picture, the depth map and the camera parameters, the views required for the multi view display can be synthesized. In this case, when the display used is a stereo display, a 3D image may be displayed using two pictures from the reconstructed multi views.
- the stereo video decoder can reconstruct two pictures that will each be incident in both from the bitstream.
- a stereoscopic image may be displayed by using a view difference or disparity between a left image incident to the left eye and a right image incident to the right eye.
- the multi view display is used together with the stereo video decoder, different views may be generated based on the two reconstructed pictures to display the multi view.
- the 2D image may be restored and the image may be output to the 2D display.
- the decoder may output one of the reconstructed images to the 2D display when using a 3D video decoder or a stereo video decoder.
- view synthesis may be performed at the decoder side or may be performed at the display side.
- the decoder and the display may be one device or separate devices.
- the 3D video decoder, the stereo video decoder, and the 2D video decoder are described as separate decoders.
- one decoding apparatus may perform 3D video decoding, stereo video decoding, and 2D video decoding.
- the 3D video decoding apparatus may perform 3D video decoding
- the stereo video decoding apparatus may perform stereo video decoding
- the 2D video decoding apparatus may perform the 2D video decoding apparatus.
- the multi view display may output 2D video or output stereo video.
- the video encoding apparatus 200 may include a picture splitter 205, a predictor 210, a subtractor 215, a transformer 220, a quantizer 225, a reorderer 230, An entropy encoding unit 235, an inverse quantization unit 240, an inverse transform unit 245, an adder 250, a filter unit 255, and a memory 260 are included.
- the picture dividing unit 205 may divide the input picture into at least one processing unit block.
- the processing unit block may be a coding unit block, a prediction unit block, or a transform unit block.
- the coding unit block may be divided along the quad tree structure from the largest coding unit block as a unit block of coding.
- the prediction unit block is a block partitioned from the coding unit block and may be a unit block of sample prediction. In this case, the prediction unit block may be divided into sub blocks.
- the transform unit block may be divided from the coding unit block along a quad tree structure, and may be a unit block for deriving a transform coefficient or a unit block for deriving a residual signal from the transform coefficient.
- a coding unit block is called a coding block or a coding unit (CU)
- a prediction unit block is called a prediction block or a prediction unit (PU)
- a transform unit block is a transform block.
- a transform unit (TU) transform unit
- a prediction block or prediction unit may mean a specific area in the form of a block within a picture or may mean an array of prediction samples.
- a transform block or a transform unit may mean a specific area in a block form within a picture, or may mean an array of transform coefficients or residual samples.
- the prediction unit 210 may perform a prediction on a block to be processed (hereinafter, referred to as a current block) and generate a prediction block including prediction samples of the current block.
- the unit of prediction performed by the prediction unit 210 may be a coding block, a transform block, or a prediction block.
- the prediction unit 210 may determine whether intra prediction or inter prediction is applied to the current block.
- the prediction unit 210 may derive a prediction sample for the current block based on neighboring block pixels in a picture to which the current block belongs (hereinafter, referred to as the current picture). In this case, the prediction unit 210 may (i) derive a prediction sample based on the average or interpolation of neighbor reference samples of the current block, and (ii) a specific direction with respect to the prediction target pixel among the neighboring blocks of the current block. A prediction sample may be derived based on a reference sample present at. For convenience of explanation, the case of (i) is referred to as non-directional mode and the case of (ii) is referred to as directional mode. The prediction unit 210 may determine the prediction mode applied to the current block by using the prediction mode applied to the neighboring block.
- the prediction unit 210 may derive a prediction sample for the current block based on the samples specified by the motion vector on the reference picture.
- the predictor 210 may induce a prediction sample for the current block by applying any one of a skip mode, a merge mode, and an MVP mode.
- the prediction unit 210 may use the motion information of the neighboring block as the motion information of the current block.
- the skip mode unlike the merge mode, the difference (residual) between the prediction sample and the original sample is not transmitted.
- the motion vector of the neighboring block may be used as a motion vector predictor (MVP) to derive the motion vector of the current block.
- MVP motion vector predictor
- the neighboring block includes a spatial neighboring block present in the current picture and a temporal neighboring block present in the collocated picture.
- the motion information includes a motion vector and a reference picture.
- motion information of a temporal neighboring block is used in the skip mode and the merge mode, the highest picture on the reference picture list may be used as the reference picture.
- the prediction unit 210 may perform inter view prediction.
- the predictor 210 may construct a reference picture list by including pictures of other views. For inter view prediction, the predictor 210 may derive a disparity vector. Unlike a motion vector that specifies a block corresponding to the current block in another picture in the current view, the disparity vector may specify a block corresponding to the current block in another view of the same access unit (AU) as the current picture.
- AU access unit
- the prediction unit 210 may specify a depth block in a depth view based on the disparity vector, configure the merge list, inter view motion prediction, and residual. Prediction, illumination compensation (IC), view synthesis, and the like can be performed.
- the disparity vector for the current block can be derived from the depth value using the camera parameter or from the motion vector or disparity vector of the neighboring block in the current or other view.
- the prediction unit 210 may include an inter-view merging candidate (IvMC) corresponding to temporal motion information of a reference view and an inter-view disparity vector candidate corresponding to the disparity vector.
- view disparity vector candidate (IvDC) shifted IvMC derived by shifting the disparity vector
- texture merge candidate derived from the texture corresponding to when the current block is a block on the depth map texture merging candidate (T)
- D disparity derived merging candidate
- VSP view synthesis prediction merge candidate derived based on view synthesis : VSP
- the number of candidates included in the merge candidate list applied to the dependent view may be limited to a predetermined value.
- the prediction unit 210 may apply the inter-view motion vector prediction to predict the motion vector of the current block based on the disparator vector.
- the prediction unit 210 may derive the disparity vector based on the conversion of the maximum depth value in the corresponding depth block.
- a block including the reference sample may be used as the reference block.
- the prediction unit 210 may use the motion vector of the reference block as a candidate motion parameter or motion vector predictor candidate of the current block, and use the disparity vector as a candidate disparity vector for DCP.
- the subtraction unit 215 generates a residual sample which is a difference between the original sample and the prediction sample.
- residual samples may not be generated as described above.
- the transform unit 220 generates a transform coefficient by transforming the residual sample in units of transform blocks.
- the quantization unit 225 may quantize the transform coefficients to generate quantized transform coefficients.
- the reordering unit 230 rearranges the quantized transform coefficients.
- the reordering unit 230 may reorder the quantized transform coefficients in the form of a block into a one-dimensional vector form by scanning the coefficients.
- the entropy encoding unit 235 may perform entropy encoding on the quantized transform coefficients.
- Entropy encoding may include, for example, encoding methods such as Exponential Golomb, Context-Adaptive Variable Length Coding (CAVLC), and Context-Adaptive Binary Arithmetic Coding (CABAC).
- CABAC Context-Adaptive Binary Arithmetic Coding
- the entropy encoding unit 235 may encode information necessary for video reconstruction other than the quantized transform coefficients (eg, a value of a syntax element) together or separately.
- Entropy-encoded information may be transmitted or stored in units of NAL units in the form of a bitstream.
- the dequantization unit 240 inversely quantizes the quantized transform coefficients to generate transform coefficients.
- the inverse transform unit 245 inverse transforms the transform coefficients to generate residual samples.
- the adder 250 reconstructs the picture by combining the residual sample and the predictive sample.
- the residual sample and the predictive sample may be added in units of blocks to generate a reconstructed block.
- the adder 250 has been described in a separate configuration, the adder 250 may be part of the predictor 210.
- the filter unit 255 may apply a deblocking filter and / or an offset to the reconstructed picture. Through the deblocking filtering mill / or offset, the artifacts at the block boundaries in the reconstructed picture or the distortion in the quantization process can be corrected.
- the offset may be applied on a sample basis or may be applied after the process of deblocking filtering is completed.
- the memory 260 may store information necessary for reconstructed pictures or encoding / decoding.
- the memory 260 may store pictures used for inter prediction / inter-view prediction.
- pictures used for inter prediction / inter-view prediction may be designated by a reference picture set or a reference picture list.
- one encoding device has been described as encoding the independent view and the dependent view, this is for convenience of description, and a separate encoding device is configured for each view or a separate internal module (for example, prediction for each view). B) may be configured.
- the video decoding apparatus 300 includes an entropy decoding unit 310, a reordering unit 320, an inverse quantization unit 330, an inverse transform unit 340, a predictor 350, and an adder 360.
- the filter unit 370 and the memory 380 are included.
- the video decoding apparatus 300 may reconstruct the video in response to a process in which the video information is processed in the video encoding apparatus.
- the video decoding apparatus 300 may perform video decoding using a processing unit applied in the video encoding apparatus.
- the processing unit block of video decoding may be a coding unit block, a prediction unit block, or a transform unit block.
- the coding unit block may be divided along the quad tree structure from the largest coding unit block as a unit block of decoding.
- the prediction unit block is a block partitioned from the coding unit block and may be a unit block of sample prediction. In this case, the prediction unit block may be divided into sub blocks.
- the transform unit block may be divided from the coding unit block along a quad tree structure, and may be a unit block for deriving a transform coefficient or a unit block for deriving a residual signal from the transform coefficient.
- the entropy decoding unit 310 may parse the bitstream and output information necessary for video reconstruction or picture reconstruction. For example, the entropy decoding unit 310 may decode the information in the bitstream based on the exponential Golomb, CAVLC, CABAC, etc., and output a syntax element value required for video reconstruction, a quantized value of transform coefficients related to the residual, and the like. have. In this case, the entropy decoding unit 310 may entropy decode information about the current block by referring to neighboring blocks of the current block.
- the bitstream may be input for each view.
- information about each view may be multiplexed in the bitstream.
- the entropy decoding unit 310 may de-multiplex the bitstream and parse for each view.
- the reordering unit 320 may rearrange the quantized transform coefficients in the form of a two-dimensional block.
- the reordering unit 320 may perform reordering in response to coefficient scanning performed by the encoding apparatus.
- the inverse quantization unit 330 may dequantize the quantized transform coefficients based on the (inverse) quantization parameter and output the transform coefficients.
- information for deriving a quantization parameter may be signaled from the encoding apparatus.
- the inverse transform unit 340 may inverse residual transform coefficients to derive residual samples.
- the prediction unit 350 may perform prediction on the current block and generate a prediction block including prediction samples for the current block.
- the unit of prediction performed by the prediction unit 350 may be a coding block, a transform block, or a prediction block.
- the prediction unit 350 may determine whether to apply intra prediction or inter prediction.
- a unit for determining which of intra prediction and inter prediction is to be applied and a unit for generating a prediction sample may be different.
- the unit for generating the prediction sample in inter prediction and intra prediction may also be different.
- the prediction unit 350 may derive the prediction sample for the current block based on the neighboring block pixels in the current picture.
- the prediction unit 350 may derive the prediction sample for the current block by applying the directional mode or the non-directional mode based on the peripheral reference samples of the current block.
- the prediction mode to be applied to the current block may be determined using the intra prediction mode of the neighboring block.
- the prediction unit 350 may derive the prediction sample for the current block based on the samples specified by the motion vector on the reference picture.
- the prediction unit 350 may induce a prediction sample for the current block by applying any one of a skip mode, a merge mode, and an MVP mode.
- the motion information of the neighboring block may be used as the motion information of the current block.
- the neighboring block may include a spatial neighboring block and a temporal neighboring block.
- the predictor 350 may construct a merge candidate list using motion information of available neighboring blocks, and use information indicated by the merge index on the merge candidate list as a motion vector of the current block.
- the merge index may be signaled from the encoding device.
- the motion information includes a motion vector and a reference picture. When motion information of a temporal neighboring block is used in the skip mode and the merge mode, the highest picture on the reference picture list may be used as the reference picture.
- the difference (residual) between the prediction sample and the original sample is not transmitted.
- the motion vector of the current block may be derived using the motion vector of the neighboring block as a motion vector predictor (MVP).
- the neighboring block may include a spatial neighboring block and a temporal neighboring block.
- the prediction unit 350 may perform inter view prediction.
- the prediction unit 350 may configure a reference picture list including pictures of other views.
- the predictor 210 may derive a disparity vector.
- the prediction unit 350 may specify a depth block in a depth view based on the disparity vector, configure the merge list, inter view motion prediction, and residual. Prediction, illumination compensation (IC), view synthesis, and the like can be performed.
- the disparity vector for the current block can be derived from the depth value using the camera parameter or from the motion vector or disparity vector of the neighboring block in the current or other view.
- Camera parameters may be signaled from the encoding device.
- the prediction unit 350 shifts the IvMC corresponding to the temporal motion information of the reference view, the IvDC corresponding to the disparity vector, and the disparity vector. Shifted IvMC derived by a subfield, a texture merge candidate (T) derived from a texture corresponding to a case in which the current block is a block on a depth map, and a disparity derivation merge candidate (D) derived using disparity from a texture merge candidate. ), A view synthesis prediction merge candidate (VSP) derived based on view synthesis may be added to the merge candidate list.
- VSP view synthesis prediction merge candidate
- the number of candidates included in the merge candidate list applied to the dependent view may be limited to a predetermined value.
- the prediction unit 350 may apply inter-view motion vector prediction to predict the motion vector of the current block based on the disparator vector.
- the prediction unit 350 may use a block in the reference view specified by the disparity vector as the reference block.
- the prediction unit 350 may use the motion vector of the reference block as a candidate motion parameter or motion vector predictor candidate of the current block, and use the disparity vector as a candidate disparity vector for DCP.
- the adder 360 may reconstruct the current block or the current picture by adding the residual sample and the predictive sample.
- the adder 360 may reconstruct the current picture by adding the residual sample and the predictive sample in block units. Since the residual is not transmitted when the skip mode is applied, the prediction sample may be a reconstruction sample.
- the adder 360 has been described in a separate configuration, the adder 360 may be part of the predictor 350.
- the filter unit 370 may apply deblocking filtering and / or offset to the reconstructed picture.
- the offset may be adaptively applied as an offset in a sample unit.
- the memory 380 may store information necessary for reconstruction picture or decoding.
- the memory 380 may store pictures used for inter prediction / inter-view prediction.
- pictures used for inter prediction / inter-view prediction may be designated by a reference picture set or a reference picture list.
- the reconstructed picture can be used as a reference picture.
- the memory 380 may output the reconstructed picture in the output order.
- the output unit may display a plurality of different views.
- each decoding apparatus may operate for each view, and an operation unit (eg, a prediction unit) corresponding to each view may be provided in one decoding apparatus.
- an operation unit eg, a prediction unit
- the encoding apparatus and the decoding apparatus may improve the efficiency of video coding for the current view using the coded data of another view belonging to the same access unit (AU) as the current picture.
- pictures having the same POC may be referred to as one AU.
- the POC corresponds to the display order of the pictures.
- the encoding apparatus and the decoding apparatus may code views in units of AUs, and may code pictures in units of views. Coding proceeds between views according to a predetermined order.
- the first coded view may be referred to as a base view or an independent view.
- a view that can be coded by referencing another view after the independent view is coded can be called a dependent view.
- another view referred to in coding (encoding / decoding) of the current view may be referred to as a reference view.
- FIG. 4 is a diagram schematically illustrating inter view coding.
- coding is performed in units of AU, where V0 is an independent view and V1 is a dependent view.
- inter-picture prediction that refers to another picture 430 of the same view using the motion vector 440 may be referred to as motion-compensated prediction (MCP).
- MCP motion-compensated prediction
- disparity-compensated prediction is performed by using the disparity vector 450 for inter-picture prediction that refers to the picture 420 of another view in the same access unit, that is, the same POC.
- DCP Compensated Picture
- a depth map may be used in addition to a method of using pictures of other views.
- FIG. 5 schematically illustrates a multi-view coding method using a depth map.
- a block (current block) 505 of the current picture 500 in the current view may be coded (encoded / decoded) using the depth map 510.
- the depth value d of the position (x, y) of the sample 520 in the depth map 510 corresponding to the position (x, y) of the sample 515 in the current block 505 is the disparity vector. 525.
- the depth value d can be derived based on the distance between the sample (pixel) and the camera.
- the encoding apparatus and the decoding apparatus may add the disparity vector 525 to the sample 530 position (x, y) to determine the position of the reference sample 535 in the current picture 540 in the reference view.
- the disparity vector may have only x-axis components. Accordingly, the value of the disparity vector may be (disp, 0), and the position (xr, y) of the reference sample 540 may be determined as (x + disp, y).
- the encoding apparatus and the decoding apparatus may use the motion parameter of the reference block 545 including the reference pixel 535 as a candidate of the motion parameter of the current block. For example, if the reference picture 550 in the reference view is a reference picture for the reference block 545, the motion vector 555 of the reference block 545 may be derived to the motion vector 560 of the current block 505. It may be. At this time, the picture 565 is a reference picture in the current view.
- the encoding apparatus and the decoding apparatus may perform residual prediction on the MCP block of the dependent view.
- the residual prediction is performed by the prediction units of the encoding apparatus and the decoding apparatus.
- the prediction unit may predict the residual of the current block by using the motion vector predicted in the block of the reference view corresponding to the current block.
- FIG. 6 is a diagram schematically illustrating residual prediction according to the present invention.
- the prediction unit may determine the position of the block 610 in the reference view Vref corresponding to the current block 605 by using the disparity vector 615. Can be specified.
- block A 620 is a block that can be referred to in the L0 direction
- block B 625 is a block that can be referenced in the L1 direction.
- Block A 620 may be specified by motion vector mv0 630 and block B 625 may be specified by motion vector mv1 635.
- the prediction unit may specify a reference block C 640 of the corresponding block 610 using mv0 and specify a reference block D 645 of the corresponding block 620 using mv1.
- the POC of the picture to which the reference block C 640 belongs and the POC of the picture to which the reference block A 620 belongs may be the same.
- the POC of the picture to which the reference block D 645 belongs and the POC of the picture to which the reference block B 625 belongs may be the same.
- Inter-view residual prediction may be applied only for a specific prediction direction in consideration of complexity.
- FIG. 6 illustrates that both the motion vector mv0 in the L0 direction and the motion vector mv1 in the L1 direction may be considered, only mv0 in the L0 direction may be used.
- (1) prediction using the same view reference block 620 is performed on the current block 605 and (2) reference of another view to the current block 605. Prediction may be performed using block 610.
- the prediction unit is formed between the sample value of the corresponding block 610 and the sample value of the reference block 640 in the reference view vref.
- the difference value may be used to modify the predictive sample of the current block 605.
- the sample value of the used block may be a sample value of the prediction sample.
- the predictor is the sample value of the reference block A 620 in the current view and the reference block C 640 in the reference view.
- the difference between the sample values of may be used to modify the predictive sample of the current block 605.
- the sample value of the used block may be a sample value of the prediction sample.
- the reference picture of the same view as the current block and the reference picture of the corresponding picture may be used for residual prediction. Therefore, in this case, the first block in the reference picture of the same view may be specified as a motion vector, and the second block in the reference picture of the corresponding picture may be specified as a disparity vector and a motion vector, and the first block and the second block may be specified.
- the difference of the block samples may be used to modify the prediction samples of the current block.
- the corresponding picture of the reference view and the reference picture of the corresponding picture may be used for residual prediction. Accordingly, in this case, the first block in the corresponding picture of the reference view may be specified as a disparity vector, and the second block in the reference picture of the corresponding picture may be specified as a disparity vector and a motion vector. The difference between the two block samples may be used to modify the prediction sample of the current block.
- the prediction sample of the current block 605 may be modified by adding a value by adding a predetermined weight to the difference value.
- the difference value may be referred to as an inter-view residual in the sense of a difference value obtained by using samples of another view.
- the predictor or adder may induce a reconstruction sample of the current block based on the modified prediction sample.
- the prediction unit or the adder may derive the reconstruction sample for the current block by adding the residual sample for the current block to the modified prediction sample.
- the weighting factor applied to the derived inter-view residual sample may be one of 0, 0.5, or 1.
- the weight 0 may be indicative that residual prediction is not applied.
- Index information indicating which weight to apply may be transmitted in units of blocks.
- Inter-view residual prediction may not be applied when there is no motion vector, such as a DCP block or a block of intra mode.
- VSP View Synthesis Prediction
- IC illumination compensation
- the current block 720 may be a prediction block.
- prediction may be performed using the reference block 740 in the inter-view reference picture 730.
- the current picture 710 and the reference picture 730 belonging to different views are pictures captured by different cameras. Therefore, as described above, there may be a mismatch between views due to different shooting conditions, shooting settings, etc. for each camera.
- the inter-view prediction may include a case where a prediction is performed by synthesizing a virtual view based on a depth map.
- the prediction of the current PU and the reference compensation may be compensated by reflecting characteristics of neighboring pixels of the current block (eg, CU) and the inter-view reference block.
- Equation 1 the sample position (i, j) is the position of a sample in the current PU, and R is a reference sample of the interview reference block specified by the disparity vector dv. a and b are some linear coefficients for compensation.
- the encoder generates a histogram representing a distribution of pixel values of the current picture and the interview reference picture. This histogram can indicate how many times a corresponding pixel value appears in each picture.
- the encoder calculates and adds a difference value for each pixel value of two histograms. Find the SAD of the histogram.
- the encoder sets a value of a flag indicating whether to perform compensation to 1 and compensates for the difference in pixel value distribution. If the SAD value is less than or equal to the threshold value, set the flag value to 0 and disable compensation.
- the encoder stores the flag information in an adaptive parameter set (APS) for efficient signaling, and transmits the value only when there is a change so that it can be applied in units of slices. Can be used.
- APS adaptive parameter set
- the decoder may compensate for the difference in pixel value distribution as indicated by flag information transmitted from the encoder.
- related information may be transmitted from the encoder side to the decoder side so as to effectively control the residual prediction or the IVMC.
- Signaling of the video information may be delivered over a network, or may be delivered through a recording medium or the like.
- Entropy coding (entropy encoding / entropy decoding) can be used to code the information transferred from the encoder to the decoder.
- the video information may be entropy encoded and transmitted to the decoder in a bitstream, and the decoding apparatus may obtain information by entropy decoding the bitstream.
- Context-based Adaptive Binary Arithmetic Code (CABAC) may be used for entropy coding of video information.
- the entropy decoding unit may binarize the input value and output the binned value.
- the empty string may mean a binary sequence or binary code composed of one or more bins.
- a bin means a value (0 or 1) of each digit that makes up a binary sequence (or binary code) when the value of a symbol and / or syntax element is represented as a binary sequence (or binary code) through binarization. Can be.
- binarization scheme can be determined differently depending on the syntax element. Binarization types that can be applied to syntax elements include, for example, unary binarization, truncated unary binarization, truncated rice binarization, exp-Golomb binarization and fixed length. -length) binarization and the like.
- the binarized signal (empty string) may be determined whether to be arithmetic coded or bypassed for each bin.
- a context that reflects a probability value is assigned to the bin to be decoded, and the bin may be coded (entropy encoded / entropy decoded) based on the assigned context. After coding for each bean, the context can be updated.
- each of the bins may be indexed by a bin index binIdx.
- the entropy decoding unit may derive a context index (ctxIdx) for each bin.
- the probability to apply to the bin that is, the context index (ctxIdc) indicating the context model, can be derived using the context index offset value (ctxIdxOffset) and the context index increments (ctxIdc). have.
- the entropy decoding unit may set an offset (ctxIdxOffset) value to apply a different context according to the slice type (I slice, P slice, or B slice).
- the entropy decoding unit may determine a context index increment (ctxIdc) for each bin index. For example, the entropy decoding unit may determine the context index increment based on a preset 'context table'.
- the entropy decoding unit may add the context index increment to the context index offset to derive the context index for the corresponding bin and to derive the random variable pStateIdx based on the context index.
- the random variable may be determined using an initial value indicated by the context index on the context index table set for each syntax element.
- the entropy decoding unit may decode the target bin by determining a state value valMps having a high probability of occurrence based on the random variable.
- bypass coding a procedure of allocating and updating a context for a bean may be bypassed.
- coding may be performed by applying a uniform probability distribution.
- the entropy decoding unit may determine the binary code as the value of the syntax element when the binary code of the decoded bins is mapped to a specific syntax element.
- inter-view residual prediction information about whether residual prediction can be applied at the slice header, picture, sequence, or video level, i.e., residual, to control the residual prediction at a higher level than block units. Information about whether the prediction is applied may be transmitted.
- inter-view residual prediction is sometimes called ARP (Advanced Residual Prediction).
- information indicating whether residual prediction is applied, information indicating whether a weighted index is used, information about an index indicating which weight is applied, and the like may be transmitted in units of blocks.
- the block that is a unit of transmission may be a prediction block.
- residual prediction is configured to be applied to the current slice
- a flag indicating whether residual prediction is applied on a block-by-block basis to indicate whether the current block is predicted based on the residual prediction or not is applied.
- An indicating flag and / or weight index may be transmitted.
- the encoding apparatus may not transmit an ARP flag and a weight index indicating whether residual prediction is applied. In this case, the decoding apparatus may omit parsing about the ARP flag and the weight index for the current block.
- the encoding apparatus transmits the ARP flag only when the highest candidate is used on the merge candidate list (that is, when the merge index is 0). May parse the ARP flag. If the merge mode is applied to the current block but the merge index is greater than 0, the ARP flag and the weighted index may be derived from the merge candidate block.
- the encoding apparatus may encode two syntax element values and transmit the encoded syntax element values to the decoding apparatus.
- transmitting the information to the decoding apparatus includes not only transmitting the information to the decoding apparatus over a network, but also storing the information in a storage medium and transmitting the information to the decoding apparatus.
- Two syntax elements for performing residual prediction that is, an information element, include an ARP flag (iv_res_pred_flag) indicating whether or not residual prediction can be applied to the current block, and if the residual prediction can be applied to the current block.
- the encoding device may encode the values of iv_res_pred_flag and iv_res_pred_weight_idx and transmit the encoded values to the decoding device.
- iv_res_pred_flag indicates whether inter-view residual prediction is used in the decoding process. When the value of iv_res_pred_flag is 0, this indicates that inter-view residual prediction is not applied. When the value of iv_res_pred_flag is 1, it indicates that inter-view residual prediction can be applied.
- iv_res_pred_weight_idx may indicate an index of a weighting factor used for residual prediction. If the value of iv_res_pred_weight_idx is 0, this may indicate that residual prediction is not applied to the current block. If the value of iv_res_pred_weight_idx is not 0, it may indicate that residual prediction is used for the current block.
- the encoding apparatus may code (encode / decode) the ARP flag for the current block in consideration of the correlation with the ARP flag of the neighboring block in consideration of coding efficiency.
- a context model to be used when encoding / decoding an ARP flag of a current block may be determined using ARP flag information of A1 and B1 blocks among neighboring blocks of the current block.
- the context model specifies a context to use when coding a syntax element, and may be a context index indicating a context to apply to the syntax element.
- the A1 block is the left block of the current block
- the B1 block is the upper block of the current block.
- the encoding apparatus and the decoding apparatus may code the ARP flag of the current block in consideration of both the A1 block and the B1 block.
- the encoding apparatus and the decoding apparatus may code the ARP flag of the current block in consideration of the A1 block.
- the encoding apparatus and the decoding apparatus may code the ARP flag of the current block in consideration of the B1 block.
- both A1 and B1 blocks are coded based on residual prediction (that is, considering both the left neighboring block and the upper neighboring block, both the upper neighboring block and the left neighboring block are coded based on the residual prediction. Case), the value of the context model used when coding the ARP flag of the current block is 2.
- the context model value of the current block is one.
- the context model value to be applied to the coding of the ARP flag for the current block is 1 Can be.
- the context model value to be applied to the coding of the ARP flag for the current block may be 1.
- the context model value to be applied to the coding of the ARP flag for the current block may be 1.
- the context model value for the ARP flag of the current block is zero.
- the context model value to be applied to the coding of the ARP flag for the current block may be zero. have.
- the context model value to be applied to the coding of the ARP flag for the current block may be zero.
- the context model value to be applied to the coding of the ARP flag for the current block may be zero.
- the context modeling of the current block can be variably set using the context modeling as described above.
- the entropy decoding unit is configured to the context models (context indexes) applicable to the ARP flag. By setting all the initial values equal, the context can be derived without considering correlation with neighboring blocks.
- the entropy decoding unit may set an initial value differently for each context model in consideration of the correlation between the current block and the neighboring block. In this case, the probability may be more accurately reflected in the current block in consideration of the context of the neighboring block.
- Table 1 briefly illustrates an example in which different initial values are applied to each context model of the ARP flag.
- the encoding apparatus may encode and transmit the ARP flag, and then encode and transmit the ARP weighting index.
- ARP inter-view residual prediction
- the weighting index indicates whether the weight is actually applied. Can be used.
- a context to be applied to the current block may be applied without considering neighboring blocks.
- the initial value for the context model of the weighting index (weighting index) without considering the frequency of occurrence of the weight of 0.5 and the weight of 1 (initial value) can be set.
- the context to be applied to the current block may be determined in consideration of the neighboring block.
- the weight to be applied to the current block may be determined in consideration of the occurrence frequency of each weight. In general, since the value of the weight index is 1, it occurs more frequently than the value of the weight index is 0.5. Therefore, the initial value may be set in consideration of this. That is, the encoding apparatus may indicate that ARP is applied to the current block by coding after weighting iv_res_pred_weight_idx in consideration of the context of the neighboring block.
- iv_res_pred_weight_idx when iv_res_pred_weight_idx is not a value indicating that ARP is not applied, ARP may be applied through an iv_res_pred_weight_idx value and may indicate how much a weight is.
- the decoding apparatus may receive a value of iv_res_pred_weight_idx, which is weight information, as a bitstream and decode in consideration of neighboring blocks. If the value of the decoded iv_res_pred_weight_idx is 0, the decoding apparatus may not apply inter-layer residual prediction (ARP). When the value of the decoded weight information iv_res_pred_weight_idx is not 0, the decoding apparatus may apply the ARP to the current block by applying the weight indicated by the value of the weight information.
- ARP inter-layer residual prediction
- a method may be used that takes into account how the same syntax element is decoded in that block in a manner that considers the neighboring block. For example, when decoding the weight information of the current block in consideration of the left block, the decoding apparatus may decode the weight information of the current block based on how much the value of the weight information decoded in the left block is.
- Table 2 shows an example of an initial value table used for coding a syntax element regarding weights.
- the initial value may be a value determined by considering neighboring blocks.
- the context model may be a value determined in consideration of neighboring blocks.
- neighboring blocks may be considered in determining a context index increment or context index offset.
- the context model may be determined in consideration of at least one of the neighboring blocks A0, A1, B0, B1, and B2 illustrated in FIG. 8. .
- the number of context models may be determined according to the number of neighboring blocks that may be considered for decoding the corresponding syntax element.
- information (syntax element) necessary for the application of IVMC may be encoded and transmitted in the bitstream.
- a flag signaling whether to compensate for an inter-view miss match at a block level may be required. This flag is most likely related to the statistics of neighboring blocks. That is, if a block neighboring the current block is a block that has compensated for the inter-view mismatch, it is likely that the current block is also a target of mismatch compensation.
- the mismatch flag can be coded more efficiently if the correlation with the neighboring block is considered.
- the conventional mismatch flag coding method performs coding without considering the correlation between the mismatch flag for the current block and the mismatch flag for the neighboring block, so that the mismatch flag is coded using one context model.
- the mismatch flag of the current block is coded in consideration of the correlation with the mismatch flag of the neighboring block, the efficiency can be further increased.
- the correlation with the mismatch flag of the neighboring block is considered, several context models may be used.
- coding may be performed by selecting a context model suitable for a mismatch flag of the current block from a total of three context models in consideration of correlation with a context model of a neighboring block.
- the neighboring blocks of the current block shown in FIG. 8 may also be used for coding a mismatch flag.
- the encoding apparatus and the decoding apparatus may consider at least one of the left block A1 of the current block and the upper block B1 of the current block in coding the mismatch flag.
- Equation 2 shows a method of determining the mismatch flag value of the current block by using the mismatch flag ic_flag (A1) of the neighboring block A1 and the mismatch flag ic_flag (B1) of the neighboring block B2.
- Equation 2 if IVMC is applied to the A1 block, the value of ic_flag (A1) is 1, and if no IVMC is applied to the A1 block, the value of ic_flag (A1) is 0. Similarly for the B1 block, ic_flag (B1) becomes 1 if IVMC is applied to the B1 block, and ic_flag (B1) becomes 0 if IVMC is not applied to the B1 block.
- an initial value suitable for each context model for ic_flag of the current block may be allocated using the value of ctx_current considering the neighboring block.
- Table 3 shows an initial value allocated to ic_flag according to each ctx_current.
- a context model for coding a mismatch flag ic_flag of the current block may be determined through various combinations of the neighboring blocks A0, A1, B0, B1, and B2 shown in FIG. 8.
- a context model for coding (encoding / decoding) a mismatch flag for the current block may be determined based on the A0, B0, and B2 blocks.
- the ctx_current of the current block may be calculated by applying the mismatch flag value ic_flag of each block as in Equation 3.
- the encoding apparatus and the decoding apparatus may allocate an initial value for coding the ic_flag of the current block according to the ctx_current value of the current block determined through Equation 3.
- a context model for coding a mismatch flag may be determined using all of the neighboring blocks illustrated in FIG. 8. For example, as shown in Equation 4, the ctx_current value of the current block may be determined by reflecting a value of the mismatch flag for each block.
- the encoding apparatus and the decoding apparatus may allocate an initial value for coding the value of ic_flag for the current block according to the value of ctx_current.
- FIG. 9 is a flowchart schematically illustrating an operation of a video encoding apparatus according to the present invention.
- the encoding apparatus derives a prediction sample for the current block (S910).
- the encoding apparatus may apply inter-view residual prediction on the current block to derive the prediction sample.
- the encoding apparatus may perform mismatch compensation on the current block to derive the predictive sample. Details of the inter-view residual prediction and mismatch compensation are as described above.
- the encoding apparatus may derive the residual sample for the current block (S920).
- the encoding apparatus may derive the difference between the prediction sample derived by applying inter-view residual prediction and / or mismatch compensation and the original sample of the current block as the residual sample.
- the encoding apparatus may entropy decode video information for video decoding (S930).
- the encoding device may entropy decode the video information and output the video information in a bitstream.
- the output bitstream may be transmitted through a network or stored in a storage medium.
- the video information may include information specifying a residual sample value or a residual sample value for the current block.
- the video information may include information for performing inter-view residual prediction (eg, ARP flag and / or weight index), and may include information for mismatch correction (eg, IC flag).
- FIG. 10 is a flowchart schematically illustrating an operation of a video decoding apparatus according to the present invention.
- the decoding apparatus entropy decodes video information included in a bitstream (S1010).
- the video information that the decoding apparatus entropy decodes and derives includes not only residual information for the current block but also values of syntax elements for reconstructing the current block.
- the decoding apparatus may entropy decode information indicating whether inter-view residual prediction may be applied to the current block.
- the decoding apparatus may entropy decode information indicating which weight is applied if inter-view radial prediction is applied.
- the decoding apparatus may entropy decode information indicating whether a mismatch correction is applied to the current block.
- the decoding apparatus may derive the prediction sample of the current block based on the entropy decoded information (S1020).
- the decoding apparatus may apply inter-view residual prediction to the current block to derive the prediction sample.
- the decoding apparatus may perform mismatch compensation on the current block to derive the predictive sample. Details of the inter-view residual prediction and mismatch compensation are as described above.
- the decoding apparatus may reconstruct the current picture by deriving a reconstructed sample based on the derived prediction sample and the residual sample.
- the reconstruction of the sample may be performed in units of blocks or pictures.
- the decoding apparatus may apply filtering to the reconstructed picture (S1030).
- the decoding apparatus may modify the reconstructed picture closer to the original picture by applying a deblocking filter to the reconstructed picture or applying an offset of a sample unit.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
Claims (14)
- 멀티-뷰 비디오를 디코딩하는 비디오 디코딩 장치로서,
비트스트림을 엔트로피 디코딩하여 레지듀얼 예측에 관한 정보를 현재 블록의 주변 블록을 참조하여 디코딩하는 엔트로피 디코딩부;
현재 블록의 디코딩에 참조되는 픽처들을 저장하는 메모리;
상기 레지듀얼에 관한 정보를 이용하여 상기 메모리에 저장된 픽처들에서 상기 현재 블록에 대응하는 제1 블록 및 제2 블록 사이의 샘플 값 차이를 기반으로 현재 블록에 대한 예측 샘플을 유도하는 예측부; 및
상기 예측 샘플을 이용하여 복원된 현재 픽처에 필터링을 적용하는 필터링부를 포함하며,
상기 제1 블록 및 제2 블록 중 적어도 하나는 상기 현재 블록의 뷰와는 상이한 뷰에 속하며,
상기 제1 블록은 상기 현재 블록의 움직임 벡터 또는 디스패리티 벡터에 의해 특정되고,
상기 제2 블록은 상기 현재 블록의 움직임 벡터 및 디스패리티 벡터에 의해 특정되는 것을 특징으로 하는 비디오 디코딩 장치. - 제1항에 있어서, 상기 레지듀얼 예측에 관한 정보는 상기 현재 블록에 레지듀얼 예측을 적용하는지를 지시하는 지시 정보를 포함하며,
상기 엔트로피 디코딩부는 상기 현재 블록의 좌측 블록 또는 상측 블록의 정보를 기반으로 상기 지시 정보를 디코딩하는 것을 특징으로 하는 비디오 디코딩 장치. - 제2항에 있어서, 상기 엔트로피 디코딩부는 상기 현재 블록의 좌측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여 상기 지시 정보를 디코딩하는 것을 특징으로 하는 비디오 디코딩 장치.
- 제3항에 있어서, 상기 엔트로피 디코딩부는 상기 현재 블록의 좌측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여, 상기 지시 정보의 디코딩에 적용할 콘택스트를 결정하는 것을 특징으로 하는 비디오 디코딩 장치.
- 제2항에 있어서, 상기 엔트로피 디코딩부는 상기 현재 블록의 상측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여 상기 지시 정보를 디코딩하는 것을 특징으로 하는 비디오 디코딩 장치.
- 제5항에 있어서, 상기 엔트로피 디코딩부는 상기 현재 블록의 상측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여, 상기 지시 정보의 디코딩에 적용할 콘택스트를 결정하는 것을 특징으로 하는 비디오 디코딩 장치.
- 제2항에 있어서, 상기 엔트로피 디코딩부는 상기 현재 블록의 좌측 블록 또는 상측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여 상기 지시 정보를 디코딩하는 것을 특징으로 하는 비디오 디코딩 장치.
- 제1항에 있어서, 상기 제1 블록의 샘플 및 상기 제2 블록의 샘플은 예측 샘플인 것을 특징으로 하는 비디오 디코딩 장치.
- 멀티-뷰 비디오를 디코딩하는 비디오 디코딩 방법으로서,
현재 블록의 주변 블록을 참조하여 레지듀얼 예측에 관한 정보를 엔트로피 디코딩하는 단계;
상기 레지듀얼에 관한 정보를 이용하여 상기 현재 블록에 대응하는 제1 블록 및 제2 블록 사이의 샘플 값 차이를 기반으로 현재 블록에 대한 예측 샘플을 유도하는 단계; 및
상기 예측 샘플을 이용하여 복원된 현재 픽처에 필터링을 적용하는 단계를 포함하며,
상기 제1 블록 및 제2 블록 중 적어도 하나는 상기 현재 블록의 뷰와는 상이한 뷰에 속하며,
상기 제1 블록은 상기 현재 블록의 움직임 벡터 또는 디스패리티 벡터에 의해 특정되고,
상기 제2 블록은 상기 현재 블록의 움직임 벡터 및 디스패리티 벡터에 의해 특정되는 것을 특징으로 하는 비디오 디코딩 방법. - 제9항에 있어서, 상기 레지듀얼 예측에 관한 정보는 상기 현재 블록에 레지듀얼 예측을 적용하는지를 지시하는 지시 정보를 포함하며,
상기 엔트로피 디코딩 단계는 상기 현재 블록의 좌측 블록 또는 상측 블록의 정보를 기반으로 상기 지시 정보를 디코딩하는 것을 포함하는 것을 특징으로 하는 비디오 디코딩 방법. - 제10항에 있어서, 상기 엔트로피 디코딩 단계에서는 상기 현재 블록의 좌측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여 상기 지시 정보를 디코딩하는 것을 특징으로 하는 비디오 디코딩 방법.
- 제10항에 있어서, 상기 엔트로피 디코딩 단계에서는 상기 현재 블록의 상측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여 상기 지시 정보를 디코딩하는 것을 특징으로 하는 비디오 디코딩 방법.
- 제10항에 있어서, 상기 엔트로피 디코딩 단계에서는 상기 현재 블록의 좌측 블록 또는 상측 블록에 레지듀얼 예측이 적용되었는지 여부에 기반하여 상기 지시 정보를 디코딩하는 것을 특징으로 하는 비디오 디코딩 방법.
- 제9항에 있어서, 상기 제1 블록의 샘플 및 상기 제2 블록의 샘플은 예측 샘플인 것을 특징으로 하는 비디오 디코딩 방법.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/026,152 US10321157B2 (en) | 2013-10-18 | 2014-10-20 | Video decoding method and apparatus for decoding multi-view video |
KR1020167008558A KR102269506B1 (ko) | 2013-10-18 | 2014-10-20 | 멀티-뷰 비디오를 디코딩하는 비디오 디코딩 방법 및 장치 |
EP14854485.1A EP3059967B1 (en) | 2013-10-18 | 2014-10-20 | Video decoding method and apparatus for decoding multi-view video |
CN201480057178.8A CN105659605B (zh) | 2013-10-18 | 2014-10-20 | 用于解码多视图视频的视频解码方法和装置 |
JP2016522059A JP6542206B2 (ja) | 2013-10-18 | 2014-10-20 | マルチビュービデオをデコードするビデオデコード方法及び装置 |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361892452P | 2013-10-18 | 2013-10-18 | |
US201361892466P | 2013-10-18 | 2013-10-18 | |
US61/892,452 | 2013-10-18 | ||
US61/892,466 | 2013-10-18 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015057036A1 true WO2015057036A1 (ko) | 2015-04-23 |
Family
ID=52828404
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2014/009858 WO2015057036A1 (ko) | 2013-10-18 | 2014-10-20 | 멀티-뷰 비디오를 디코딩하는 비디오 디코딩 방법 및 장치 |
Country Status (6)
Country | Link |
---|---|
US (1) | US10321157B2 (ko) |
EP (1) | EP3059967B1 (ko) |
JP (1) | JP6542206B2 (ko) |
KR (1) | KR102269506B1 (ko) |
CN (1) | CN105659605B (ko) |
WO (1) | WO2015057036A1 (ko) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017086765A3 (ko) * | 2015-11-22 | 2017-07-27 | 엘지전자(주) | 비디오 신호를 엔트로피 인코딩, 디코딩하는 방법 및 장치 |
KR20170132682A (ko) * | 2016-05-24 | 2017-12-04 | 한국전자통신연구원 | 영상 부호화/복호화 방법 및 이를 위한 기록 매체 |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015137783A1 (ko) * | 2014-03-14 | 2015-09-17 | 삼성전자 주식회사 | 인터 레이어 비디오의 복호화 및 부호화를 위한 머지 후보 리스트 구성 방법 및 장치 |
ES2699723B2 (es) * | 2016-06-24 | 2020-10-16 | Kt Corp | Metodo y aparato para tratar una senal de video |
FR3058858A1 (fr) * | 2016-11-15 | 2018-05-18 | Orange | Procede et dispositif de codage et de decodage d'une sequence multi-vues |
KR102640142B1 (ko) * | 2019-01-12 | 2024-02-22 | 엘지전자 주식회사 | 영상 코딩 시스템에서 레지듀얼 정보를 사용하는 영상 디코딩 방법 및 그 장치 |
AU2020236358A1 (en) * | 2019-03-11 | 2021-11-11 | Interdigital Vc Holdings, Inc. | Reducing the number of regular coded bins |
WO2020189971A1 (ko) * | 2019-03-15 | 2020-09-24 | 엘지전자 주식회사 | 영상 코딩 시스템에서 변환 스킵 플래그를 이용한 영상 디코딩 방법 및 그 장치 |
CN113906742A (zh) * | 2019-06-03 | 2022-01-07 | Lg电子株式会社 | 基于矩阵的帧内预测装置和方法 |
US11516489B2 (en) * | 2019-06-24 | 2022-11-29 | Tencent America LLC | Method and apparatus for video coding |
CN112135135A (zh) * | 2019-06-24 | 2020-12-25 | 腾讯美国有限责任公司 | 视频解码方法、装置、计算机设备及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20120011428A (ko) * | 2010-07-29 | 2012-02-08 | 에스케이 텔레콤주식회사 | 블록 분할예측을 이용한 영상 부호화/복호화 방법 및 장치 |
KR20120083209A (ko) * | 2011-01-17 | 2012-07-25 | 삼성전자주식회사 | 깊이 영상 부호화/복호화 장치 및 방법 |
WO2012138032A1 (ko) * | 2011-04-07 | 2012-10-11 | 엘지전자 주식회사 | 영상 정보 부호화 방법 및 복호화 방법 |
WO2013039348A1 (ko) * | 2011-09-16 | 2013-03-21 | 엘지전자 주식회사 | 영상 정보 시그널링 방법 및 이를 이용한 비디오 디코딩 방법 |
WO2013058583A1 (ko) * | 2011-10-18 | 2013-04-25 | 주식회사 케이티 | 영상 부호화 방법, 영상 복호화 방법, 영상 부호화기 및 영상 복호화기 |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU2006201490B2 (en) | 2005-04-19 | 2008-05-22 | Samsung Electronics Co., Ltd. | Method and apparatus for adaptively selecting context model for entropy coding |
US9712819B2 (en) * | 2011-10-12 | 2017-07-18 | Lg Electronics Inc. | Image encoding method and image decoding method |
US20130258052A1 (en) * | 2012-03-28 | 2013-10-03 | Qualcomm Incorporated | Inter-view residual prediction in 3d video coding |
US9357212B2 (en) | 2012-12-07 | 2016-05-31 | Qualcomm Incorporated | Advanced residual prediction in scalable and multi-view video coding |
JP6118199B2 (ja) | 2013-07-09 | 2017-04-19 | シャープ株式会社 | 画像復号装置、画像符号化装置、画像復号方法、画像符号化方法及びコンピュータ読み取り可能な記録媒体。 |
-
2014
- 2014-10-20 EP EP14854485.1A patent/EP3059967B1/en active Active
- 2014-10-20 WO PCT/KR2014/009858 patent/WO2015057036A1/ko active Application Filing
- 2014-10-20 CN CN201480057178.8A patent/CN105659605B/zh active Active
- 2014-10-20 KR KR1020167008558A patent/KR102269506B1/ko active IP Right Grant
- 2014-10-20 JP JP2016522059A patent/JP6542206B2/ja active Active
- 2014-10-20 US US15/026,152 patent/US10321157B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20120011428A (ko) * | 2010-07-29 | 2012-02-08 | 에스케이 텔레콤주식회사 | 블록 분할예측을 이용한 영상 부호화/복호화 방법 및 장치 |
KR20120083209A (ko) * | 2011-01-17 | 2012-07-25 | 삼성전자주식회사 | 깊이 영상 부호화/복호화 장치 및 방법 |
WO2012138032A1 (ko) * | 2011-04-07 | 2012-10-11 | 엘지전자 주식회사 | 영상 정보 부호화 방법 및 복호화 방법 |
WO2013039348A1 (ko) * | 2011-09-16 | 2013-03-21 | 엘지전자 주식회사 | 영상 정보 시그널링 방법 및 이를 이용한 비디오 디코딩 방법 |
WO2013058583A1 (ko) * | 2011-10-18 | 2013-04-25 | 주식회사 케이티 | 영상 부호화 방법, 영상 복호화 방법, 영상 부호화기 및 영상 복호화기 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017086765A3 (ko) * | 2015-11-22 | 2017-07-27 | 엘지전자(주) | 비디오 신호를 엔트로피 인코딩, 디코딩하는 방법 및 장치 |
US10609374B2 (en) | 2015-11-22 | 2020-03-31 | Lg Electronics Inc. | Method and apparatus for entropy-encoding and entropy-decoding video signal |
KR20170132682A (ko) * | 2016-05-24 | 2017-12-04 | 한국전자통신연구원 | 영상 부호화/복호화 방법 및 이를 위한 기록 매체 |
KR102338356B1 (ko) | 2016-05-24 | 2021-12-13 | 한국전자통신연구원 | 영상 부호화/복호화 방법 및 이를 위한 기록 매체 |
KR20220066243A (ko) * | 2016-05-24 | 2022-05-24 | 한국전자통신연구원 | 영상 부호화/복호화 방법 및 이를 위한 기록 매체 |
US11743470B2 (en) | 2016-05-24 | 2023-08-29 | Electronics And Telecommunications Research Institute | Image encoding/decoding method and recording medium for same |
KR102618642B1 (ko) | 2016-05-24 | 2023-12-27 | 한국전자통신연구원 | 영상 부호화/복호화 방법 및 이를 위한 기록 매체 |
Also Published As
Publication number | Publication date |
---|---|
KR102269506B1 (ko) | 2021-06-25 |
JP6542206B2 (ja) | 2019-07-10 |
KR20160072100A (ko) | 2016-06-22 |
EP3059967A4 (en) | 2017-06-07 |
US20160249066A1 (en) | 2016-08-25 |
JP2016537858A (ja) | 2016-12-01 |
EP3059967B1 (en) | 2020-04-22 |
CN105659605B (zh) | 2019-05-21 |
US10321157B2 (en) | 2019-06-11 |
EP3059967A1 (en) | 2016-08-24 |
CN105659605A (zh) | 2016-06-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102269506B1 (ko) | 멀티-뷰 비디오를 디코딩하는 비디오 디코딩 방법 및 장치 | |
JP6698351B2 (ja) | ビュー内でのおよびビューにわたる深度ルックアップテーブルの予測コーディング | |
WO2015057039A1 (ko) | 멀티-뷰 비디오 코딩에 있어서, 뷰 합성 예측 방법 및 이를 이용한 머지 후보 리스트 구성 방법 | |
WO2015057037A1 (ko) | 멀티-뷰 비디오를 디코딩하는 비디오 디코딩 장치 및 방법 | |
CN106664423B (zh) | 视频编译中的深度图片编译方法 | |
US20160255371A1 (en) | Method and apparatus for coding/decoding 3d video | |
US10587894B2 (en) | Method and device for encoding/decoding 3D video | |
US20170310993A1 (en) | Movement information compression method and device for 3d video coding | |
US10419779B2 (en) | Method and device for processing camera parameter in 3D video coding | |
KR102343817B1 (ko) | 멀티-뷰 비디오의 디코딩 방법 및 장치 | |
US20160255368A1 (en) | Method and apparatus for coding/decoding video comprising multi-view | |
US10397611B2 (en) | Method and device for encoding/decoding 3D video | |
KR20210158391A (ko) | 루마 샘플들의 맵핑 및 크로마 샘플들의 스케일링 기반 비디오 또는 영상 코딩 | |
KR20220017425A (ko) | 루마 샘플들의 맵핑 및 크로마 샘플들의 스케일링 기반 비디오 또는 영상 코딩 | |
WO2015141977A1 (ko) | 3d 비디오 부호화/복호화 방법 및 장치 | |
CN114731442B (zh) | 使用加权预测的图像编码/解码方法和装置以及发送比特流的方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14854485 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15026152 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 20167008558 Country of ref document: KR Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2014854485 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014854485 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2016522059 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |