WO2016003074A1 - 움직임 병합 모드에서 시점 합성 예측 후보의 중복성 제거 장치 및 방법 - Google Patents
움직임 병합 모드에서 시점 합성 예측 후보의 중복성 제거 장치 및 방법 Download PDFInfo
- Publication number
- WO2016003074A1 WO2016003074A1 PCT/KR2015/005412 KR2015005412W WO2016003074A1 WO 2016003074 A1 WO2016003074 A1 WO 2016003074A1 KR 2015005412 W KR2015005412 W KR 2015005412W WO 2016003074 A1 WO2016003074 A1 WO 2016003074A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- candidate
- block
- vsp
- merge
- list
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
Definitions
- the present invention relates to an image encoding / decoding apparatus and a method, and more particularly, to an encoding and decoding apparatus and method for removing redundancy of a view synthesis prediction candidate in a motion merge mode.
- the 3D image requires depth map information as well as actual texture information.
- the 3D image requires much more information than the conventional 2D image, and is sufficient when encoding / decoding the 3D image using an image encoding / decoding apparatus and method for the 2D image. There was a problem that the encoding / decoding efficiency did not come out.
- An object of the present invention is to provide an image encoding / decoding apparatus and method for removing redundancy of a merge candidate inserted into a merge list.
- Another object of the present invention is to provide an apparatus and method for removing redundancy of merge candidates in 3D image encoding / decoding.
- Another object of the present invention is to provide an apparatus and method for removing redundancy of a view synthesis prediction candidate in 3D image encoding / decoding.
- the present invention when the first candidate block is inserted into the merge candidate list, when the view synthesis prediction (VSP) is used in the first candidate block, information indicating that the VSP is used is provided.
- the present invention provides a 3D image decoding method including generating and not inserting a VSP candidate of a current block into a merge candidate list when information indicating that the VSP is used is present.
- the first candidate block may be a spatial neighboring block.
- the first candidate block may be the spatial neighboring block located to the left of the current block.
- the first candidate block may be an A1 block.
- the information indicating that the VSP is used may include information indicating whether the first candidate block is available or information indicating whether the VSP is used in the first candidate block.
- the method further includes inserting the first candidate block into the merge candidate list, and then inserting a second candidate block into the merge candidate list, and generating information indicating that the VSP is used.
- information indicating that the VSP is used may be generated.
- the first candidate block may be a spatial neighboring block located to the left of the current block
- the second candidate block may be a spatial neighboring block located above the current block
- the first candidate block may be an A1 block
- the second candidate block may be a B1 block.
- the current block may be a sub prediction block.
- a spatial merge candidate insertion module for inserting a first candidate block into a merge candidate list and a VSP is used when a view synthesis prediction (VSP) is used in the first candidate block.
- VSP view synthesis prediction
- a VSP redundancy determination module for generating information indicating that the VSP is used, and a VSP candidate insertion module for not inserting the VSP candidate of the current block in a merge candidate list when information indicating that the VSP is used is provided.
- the first candidate block may be a spatial neighboring block.
- the first candidate block may be the spatial neighboring block located to the left of the current block.
- the first candidate block may be an A1 block.
- the information indicating that the VSP is used may include information indicating whether the first candidate block is available or information indicating whether the VSP is used in the first candidate block.
- the spatial merge candidate insertion module inserts the first candidate block into a merge candidate list, and then inserts a second candidate block into a merge candidate list
- the VSP redundancy determination module is configured to insert the first candidate block or the second candidate block.
- the first candidate block may be a spatial neighboring block located to the left of the current block
- the second candidate block may be a spatial neighboring block located above the current block
- the first candidate block may be an A1 block
- the second candidate block may be a B1 block.
- the current block may be a sub prediction block.
- the first candidate block when the first candidate block is inserted into the merge candidate list, when the view synthesis prediction (VSP) is used in the first candidate block, information indicating that the VSP is used And generating information indicating that the VSP has been used, and not inserting the VSP candidate of the current block into a merge candidate list.
- VSP view synthesis prediction
- the first candidate block may be a spatial neighboring block.
- the first candidate block may be the spatial neighboring block located to the left of the current block.
- the first candidate block may be an A1 block.
- the information indicating that the VSP is used may include information indicating whether the first candidate block is available or information indicating whether the VSP is used in the first candidate block.
- the method further includes inserting the first candidate block into the merge candidate list, and then inserting a second candidate block into the merge candidate list, and generating information indicating that the VSP is used.
- information indicating that the VSP is used may be generated.
- the first candidate block may be a spatial neighboring block located to the left of the current block
- the second candidate block may be a spatial neighboring block located above the current block
- the first candidate block may be an A1 block
- the second candidate block may be a B1 block.
- the current block may be a sub prediction block.
- a spatial merge candidate insertion module for inserting a first candidate block into a merge candidate list, and when a view synthesis prediction (VSP) is used in the first candidate block, the VSP is A 3D image encoding apparatus including a VSP redundancy determination module for generating information indicating that the VSP candidate is used and a VSP candidate insertion module for inserting the VSP candidate of the current block into a merge candidate list when information indicating that the VSP is used exists; to provide.
- VSP view synthesis prediction
- the first candidate block may be a spatial neighboring block.
- the first candidate block may be the spatial neighboring block located to the left of the current block.
- the first candidate block may be an A1 block.
- the information indicating that the VSP is used may include information indicating whether the first candidate block is available or information indicating whether the VSP is used in the first candidate block.
- the spatial merge candidate insertion module inserts the first candidate block into a merge candidate list, and then inserts a second candidate block into a merge candidate list
- the VSP redundancy determination module is configured to insert the first candidate block or the second candidate block.
- the first candidate block may be a spatial neighboring block located to the left of the current block
- the second candidate block may be a spatial neighboring block located above the current block
- the first candidate block may be an A1 block
- the second candidate block may be a B1 block.
- the current block may be a sub prediction block.
- the method when the first candidate block is inserted into the merge candidate list, when the view synthesis prediction (VSP) is used in the first candidate block, information indicating that the VSP is used And a method for executing the 3D image decoding method on the computer, the method comprising the step of generating a CSP and if the information indicating that the VSP has been used does not insert the VSP candidate of the current block into the merge candidate list.
- VSP view synthesis prediction
- the first candidate block may be a spatial neighboring block.
- the first candidate block may be the spatial neighboring block located to the left of the current block.
- the first candidate block may be an A1 block.
- the information indicating that the VSP is used may include information indicating whether the first candidate block is available or information indicating whether the VSP is used in the first candidate block.
- the method further includes inserting the first candidate block into the merge candidate list, and then inserting a second candidate block into the merge candidate list, and generating information indicating that the VSP is used.
- information indicating that the VSP is used may be generated.
- the first candidate block may be a spatial neighboring block located to the left of the current block
- the second candidate block may be a spatial neighboring block located above the current block
- the first candidate block may be an A1 block
- the second candidate block may be a B1 block.
- the current block may be a sub prediction block.
- VSP view synthesis prediction
- a computer program for executing a 3D image encoding method including a step of generating a data suffix and a step of not inserting a VSP candidate of a current block into a merge candidate list when information indicating that the VSP is used exists.
- the first candidate block may be a spatial neighboring block.
- the first candidate block may be the spatial neighboring block located to the left of the current block.
- the first candidate block may be an A1 block.
- the information indicating that the VSP is used may include information indicating whether the first candidate block is available or information indicating whether the VSP is used in the first candidate block.
- the method further includes inserting the first candidate block into the merge candidate list, and then inserting a second candidate block into the merge candidate list, and generating information indicating that the VSP is used.
- information indicating that the VSP is used may be generated.
- the first candidate block may be a spatial neighboring block located to the left of the current block
- the second candidate block may be a spatial neighboring block located above the current block
- the first candidate block may be an A1 block
- the second candidate block may be a B1 block.
- the current block may be a sub prediction block.
- redundancy of a merge candidate inserted into a merge list can be eliminated in image encoding / decoding.
- redundancy of merge candidates can be eliminated in 3D image encoding / decoding.
- redundancy of a view synthesis prediction candidate can be eliminated in 3D image encoding / decoding.
- Figure 1 schematically shows the basic structure of a three-dimensional video system.
- FIG. 2 is a diagram illustrating an example of an actual image and a depth information map image of a “balloons” image.
- FIG. 3 is a diagram schematically illustrating a segmentation structure of an image when encoding and decoding an image.
- FIG. 4 illustrates a form of a prediction unit PU that a coding unit CU may include.
- FIG. 5 schematically illustrates an example of a quadtree splitting structure of a transform unit (TU).
- FIG. 6 illustrates an example of a structure of inter view prediction in a 3D video codec.
- FIG. 7 illustrates an example of encoding and / or decoding a texture view and a depth view map in a 3D video encoder and / or a decoder.
- FIG. 8 is a block diagram illustrating a configuration of a video encoder according to an embodiment.
- FIG. 9 is a block diagram illustrating a configuration of a video decoder according to an embodiment.
- FIG. 10 is a diagram illustrating an example of a prediction structure for a 3D video codec.
- FIG. 11 shows an example of neighboring blocks used to construct a merge candidate list for a current block.
- FIG. 12 is a diagram illustrating an example of a process of deriving motion information of a current block by using motion information of an adjacent view.
- FIG. 13 is a diagram illustrating an example in which one prediction unit (PU) is divided into several sub-prediction units.
- FIG. 14 is a diagram illustrating an example of a process of deriving motion information of a current block by using a reference block.
- FIG. 15 schematically illustrates the concept of View Synthesis Prediction (VSP).
- VSP View Synthesis Prediction
- FIG. 16 illustrates an example of neighboring blocks inherited as VSP candidates in 3D image encoding / decoding.
- 17 illustrates an example of a merge candidate list in 3D image encoding / decoding.
- FIG. 18 is a flowchart of a method of restricting addition of a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to an embodiment of the present invention.
- 19 is a flowchart illustrating a method of restricting adding a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to another embodiment of the present invention in 3D-HEVC.
- FIG. 20 is a block diagram of an apparatus for restricting adding a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to an embodiment of the present invention.
- 21 is a block diagram of an apparatus for restricting adding a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to another embodiment of the present invention.
- FIG. 22 is a view illustrating a merge candidate derivation method and a merge candidate derivation method according to another embodiment of the present invention.
- FIG. 23 is a flowchart illustrating a method of restricting insertion of a VSP candidate into a merge list by referring to two directions among spatial merge candidates according to another embodiment of the present invention.
- 24A to 24B are flowcharts illustrating a method of restricting insertion of a VSP candidate into a merge list by referring to two directions among spatial merge candidates according to another embodiment of the present invention, in 3D HEVC.
- 25 is a view illustrating a merge candidate derivation method and a merge candidate derivation method according to another embodiment of the present invention.
- FIG. 26 is a flowchart illustrating a method of restricting insertion of a VSP candidate into a merge list with reference to one direction among spatial merge candidates according to another embodiment of the present invention.
- 27A to 27B are flowcharts illustrating a method of restricting insertion of a VSP candidate into a merge list with reference to one direction among spatial merge candidates according to another embodiment of the present invention, in 3D HEVC.
- FIG. 28 is a block diagram of an apparatus for restricting inserting a VSP candidate into a merge list according to another embodiment of the present invention.
- 29 illustrates an example in which redundancy between VSP candidates is removed through the above-described 3D image encoding / decoding apparatus and method.
- first and second may be used to describe various components, but the components should not be limited by the terms. The terms are used only for the purpose of distinguishing one component from another.
- the first component may be referred to as the second component, and similarly, the second component may also be referred to as the first component.
- each component shown in the embodiments of the present invention are shown independently to represent different characteristic functions, and do not mean that each component is made of separate hardware or one software unit.
- each component is included in each component for convenience of description, and at least two of the components may be combined into one component, or one component may be divided into a plurality of components to perform a function.
- Integrated and separate embodiments of the components are also included within the scope of the present invention without departing from the spirit of the invention.
- the components may not be essential components for performing essential functions in the present invention, but may be optional components for improving performance.
- the present invention can be implemented including only the components necessary to implement the essentials of the present invention except for the components used for improving performance, the structure including only the essential components except for the optional components used for improving performance Also included in the scope of the present invention.
- Three-dimensional video provides a three-dimensional effect as seen and felt in the real world through a three-dimensional stereoscopic display device.
- JCT-3V Joint Collaborative Team on 3D Video Coding Extension Development
- MPEG Moving Picture Experts Group
- VCEG Video Coding Experts Group
- Figure 1 schematically shows the basic structure of a three-dimensional video system.
- a 3D video (3D video) system may include a sender and a receiver.
- the 3D video system of FIG. 1 may be a basic 3D video system considered in the 3D video standard, and the 3D video standard uses a stereoscopic image by using a depth map corresponding to a real image and a real image.
- the 3D video standard can include standards for advanced data formats and related technologies that can support the playback of autostereoscopic images.
- the sender may generate multi-view video content.
- the transmitter may generate video information using a stereo camera and a multi-view camera, and generate a depth map or a depth view using the depth information camera.
- the transmitter may convert a 2D image into a 3D image using a converter.
- the transmitter may generate video content of an N (N ⁇ 2) view (that is, a multiview) using the generated video information, the depth information map, and the like.
- the video content of the N view may include video information of the N view, depth-map information thereof, and additional information related to a camera.
- the video content of N views may be compressed using a multi-view video encoding method in a 3D video encoder, and the compressed video content (bitstream) may be transmitted to a terminal of a receiving side through a network.
- the receiver may provide a multi-view image by decoding the image content received from the transmitter.
- the receiving side may reconstruct an image of N time point by decoding a bitstream received using a multiview video decoding method in a video decoder (for example, a 3D video decoder, a stereo video decoder, a 2D video decoder, etc.).
- a video decoder for example, a 3D video decoder, a stereo video decoder, a 2D video decoder, etc.
- virtual view images of more than N views may be generated using a reconstructed N-view image and a depth-image-based rendering (DIBR) process.
- DIBR depth-image-based rendering
- the generated virtual viewpoint images of N or more viewpoints are reproduced for various stereoscopic display apparatuses (eg, N-view display, stereo display, two-dimensional display, etc.) to provide a user with a three-dimensional image.
- FIG. 2 is a diagram illustrating an example of an actual image and a depth information map image of a “balloons” image.
- FIG. 2 (a) shows “balloons” images being used in the 3D video coding standard of MPEG, which is an international standardization organization.
- FIG. 2B illustrates a depth information map image of the “balloons” image illustrated in FIG. 2A.
- depth information displayed on the screen is expressed by 8 bits per pixel.
- the depth map is used to generate a virtual view image.
- the depth map is a constant bit of the distance between the camera and the real object (depth information corresponding to each pixel at the same resolution as the real image) in the real world. It is expressed as a number.
- the depth information map may be acquired by using a depth information map camera or by using an actual general image.
- the depth map obtained using the depth map map mainly provides reliable depth information in a stationary object or scene, but has a problem in that the depth map camera operates only within a certain distance.
- the depth information map camera may use a laser, a structured light technique, or a depth measurement technique based on time-of-flight of light (TFL).
- the depth information map may be generated using an actual general image and a disparity vector.
- the disparity vector refers to information representing a viewpoint difference between two general images.
- the disparity vector compares an arbitrary pixel at the present time point with those at another time point to find the most similar pixel. It can be obtained through the distance between one pixel and the most similar pixel).
- the actual image and its depth information map may be images obtained from not only one camera but also several cameras. Images obtained from several cameras may be independently encoded or encoded / decoded using a general two-dimensional video encoding / decoding codec. In addition, since images obtained by multiple cameras have correlations between viewpoints, images obtained by multiple cameras may be encoded using prediction between different viewpoints in order to increase encoding efficiency.
- the 'depth information map' may mean a 'depth map' or a 'depth picture'.
- the 'depth information map' is referred to as a 'depth map' or 'depth picture' for convenience of description. depth picture) '.
- FIG. 3 is a diagram schematically illustrating a segmentation structure of an image when encoding and decoding an image.
- encoding and decoding may be performed for each coding unit (CU).
- a unit is a combination of a syntax element and a block including image samples.
- Splitting a unit may mean splitting a block corresponding to the unit.
- the image 300 is sequentially divided into units of a largest coding unit (LCU) (hereinafter referred to as an LCU), and then a division structure is determined for each LCU.
- LCU largest coding unit
- the LCU may be used in the same meaning as a coding tree unit (CTU).
- the partition structure refers to a distribution of a coding unit (hereinafter referred to as a CU) for efficiently encoding an image in the LCU 310, and this distribution decreases one CU to half of its horizontal and vertical sizes. It may be determined according to whether to split into CUs.
- the partitioned CU may be recursively divided into four CUs whose horizontal size and vertical size are reduced by half with respect to the CU partitioned in the same manner.
- the CU may be recursively divided up to a predefined depth.
- Depth information is information indicating the size of a CU, it may be stored for each CU.
- the depth of the LCU may be 0, and the depth of the smallest coding unit (SCU) may be a predefined maximum depth.
- the LCU is a coding unit having a maximum size
- the smallest coding unit (SCU) is a coding unit having a minimum size.
- each time the division from the LCU 310 to half of the horizontal and vertical sizes increases the depth of the CU by one. For example, if the size of the CU is 2N ⁇ 2N at a certain depth L, the size of the CU is still 2N ⁇ 2N when no splitting is performed, and the size of the CU is N ⁇ N when splitting is performed. At this time, the depth of the NxN size CU is a depth L + 1. That is, the size of N corresponding to the size of the CU decreases in half each time the depth increases by one.
- an LCU having a minimum depth of 0 may be 64x64 pixels, and an SCU having a maximum depth of 3 may be 8x8 pixels.
- the depth of the CU (LCU) of 64x64 pixels may be represented by 0, the depth of the CU of 32x32 pixels is 1, the depth of the CU of 16x16 pixels is 2, and the depth of the CU (SCU) of 8x8 pixels is 3.
- information on whether to partition a specific CU may be expressed through partition information of 1 bit per CU.
- This partitioning information may be included in all CUs except the SCU. For example, when partitioning a CU, 0 may be stored in partitioning information, and when partitioning a CU, 1 may be stored in partitioning information.
- the above-mentioned 'coding unit (CU)' may mean a 'coding block', 'coding unit' or 'coding block', and the like.
- the 'coding unit' will be described.
- the term 'coding block', 'coding unit' or 'coding block' may be used interchangeably.
- FIG. 4 illustrates a form of a prediction unit PU that a coding unit CU may include.
- CUs that are no longer split among the CUs partitioned from the LCU may be partitioned or partitioned into one or more prediction units.
- a prediction unit (hereinafter, referred to as a PU) is a basic unit for performing prediction, and is encoded and decoded in any one of a skip mode, an inter mode, and an intra mode, and the prediction unit is according to each mode. It can be partitioned in various forms.
- the 2N ⁇ 2N mode 410 having the same size as the CU may be supported without a partition of the CU.
- nRx2N mode 445 For inter mode, eight partitioned forms for the CU, such as 2Nx2N mode 410, 2NxN mode 415, Nx2N mode 420, NxN mode 425, 2NxnU mode 430, 2NxnD mode 435 nLx2N mode 440 and nRx2N mode 445 may be supported.
- the 2Nx2N mode 410 and the NxN mode 425 may be supported for the CU.
- PU 'prediction unit
- PU 'prediction unit
- 'prediction unit' may be used interchangeably as 'prediction block'.
- FIG. 5 schematically illustrates an example of a quadtree splitting structure of a transform unit (TU).
- Transform Unit is a basic unit used in a spatial transform and quantization process in a CU.
- the TU may have a square shape or a rectangular shape, and the TU may be included in the CU.
- the size of the TU may be the same as the size of the CU or smaller than the size of the CU, and one CU may include a plurality of TUs having different sizes.
- the TU may have the same size as the CU, and the TU may be divided from the CU using quadtree partitioning.
- the TU may be split recursively up to twice from the CU, but in this specification, as shown in FIG. 5, the TU is recursively from the CU. The division of three or more times is not excluded from the scope of rights.
- the TU when a CU has a size of 2N ⁇ 2N, the TU may have a size of 2N ⁇ 2N, which is the same size as the CU, and when the quad tree is split once from the CU, the TU may have a size of N ⁇ N.
- the TU when a quadtree is divided twice from the CU, the TU may have a size of N / 2 ⁇ N / 2, and when the quadtree is divided three times from the CU, the TU may have a size of N / 4 ⁇ N / 4.
- Transform Unit may mean a "Transform Block.”
- Transform Unit may be used interchangeably as a “Transform Block.”
- FIG. 6 illustrates an example of a structure of inter view prediction in a 3D video codec.
- View 1 and view 2 may perform inter-view prediction using view 0 as a reference image, and the encoding order is view 1 and view 2 View 0 should be coded before.
- view 0 is called an independent view because it may be independently encoded regardless of other views.
- view 1 and view 2 are referred to as dependent views because they are encoded using view 0 as a reference image.
- Independent viewpoint images may be encoded using a general two-dimensional video codec.
- the dependent view image since the dependent view image needs to perform inter-view prediction, it may be encoded using a 3D video codec including an inter-view prediction process.
- the view 1 and the view 2 may be encoded using the depth information map.
- the real image and the depth information map may be encoded and / or decoded independently of each other.
- the real image and the depth information map may be encoded and / or decoded depending on each other as shown in FIG. 7.
- FIG. 7 illustrates an example of encoding and / or decoding a texture view and a depth view map in a 3D video encoder and / or a decoder.
- the 3D video encoder may include a real image encoder encoding a texture view and a depth map encoder encoding a depth view map. have.
- the real image encoder may encode the real image using the depth information map encoded by the depth information map encoder.
- the depth information map encoder may encode the depth information map by using the real image encoded by the real image encoder.
- the 3D video decoder may include a real image decoder that decodes an actual image and a depth information decoder that decodes a depth information map.
- the real image decoder may decode the real image using the depth information map decoded by the depth information map decoder.
- the depth information map decoder may decode the depth information map by using the real image decoded by the real image decoder.
- FIG. 8 is a block diagram illustrating a configuration of a video encoder according to an embodiment.
- FIG. 8 illustrates an embodiment of a video encoder applicable to a multi-view structure, wherein the video encoder for the multi-view structure may be implemented by extending a video encoder for a single view structure.
- the video encoder of FIG. 8 may be used in the real image encoder and / or the depth information map encoder of FIG. 7, and the encoder may mean an encoding apparatus.
- the video encoder 800 includes an inter predictor 810, an intra predictor 820, a switch 815, a subtractor 825, a transformer 830, a quantizer 840, and entropy encoding.
- the unit 850 includes an inverse quantization unit 860, an inverse transform unit 870, an adder 875, a filter unit 880, and a reference picture buffer 890.
- the video encoder 800 may encode the input video in an intra mode or an inter mode and output a bitstream.
- Intra prediction means intra picture prediction
- inter prediction means inter picture prediction or inter-view prediction.
- the switch 815 is switched to the intra mode
- the switch 815 is switched to the inter mode.
- the video encoder 800 may generate a prediction block for a block (current block) of an input picture and then encode a difference between the current block and the prediction block.
- the intra predictor 820 may use the pixel value of an already encoded block around the current block as a reference pixel.
- the intra predictor 820 may generate prediction samples for the current block by using the reference pixel.
- the inter prediction unit 810 may obtain a motion vector specifying a reference block corresponding to an input block (current block) from a reference picture stored in the reference picture buffer 890.
- the inter prediction unit 810 may generate a prediction block for the current block by performing motion compensation using the motion vector and the reference picture stored in the reference picture buffer 890.
- inter prediction applied in inter mode may include inter view prediction.
- the inter prediction unit 810 may configure an inter view reference picture by sampling a picture of the reference view.
- the inter prediction unit 810 may perform inter view prediction by using a reference picture list including an inter view reference picture. Reference relationships between views may be signaled via information specifying dependencies between views.
- sampling applied to the reference view picture may mean generation of a reference sample by copying or interpolating a sample from the reference view picture.
- sampling applied to the reference view picture may mean upsampling or downsampling.
- the inter view reference picture may be configured by upsampling the reconstructed picture of the reference view.
- Which view picture is used to configure the inter-view reference picture may be determined in consideration of a coding cost or the like.
- the encoder may transmit information specifying a view to which the picture to be used as the inter-view reference picture belongs to the decoding apparatus.
- a view used in inter-view prediction that is, a picture used for prediction of the current block in the reference view may be a picture of the same access unit (AU) as the current picture (picture to be predicted in the current view).
- AU access unit
- the subtractor 825 may generate a residual block (residual signal) by the difference between the current block and the prediction block.
- the transform unit 830 may transform the residual block to output transform coefficients, and the transform unit 830 may perform transform using the transform block. When the transform skip mode is applied, the transform unit 830 may omit the transform for the residual block.
- the quantization unit 840 may output the quantized coefficients by quantizing the transform coefficients according to the quantization parameter.
- the entropy encoder 850 may entropy-encode the values calculated by the quantizer 840 or the encoding parameter values calculated in the encoding process according to a probability distribution to output a bitstream.
- the entropy encoder 850 may entropy encode information (eg, syntax elements) for video decoding in addition to the pixel information of the video.
- the encoding parameter is information necessary for encoding and decoding, and may include information that may be inferred in the encoding or decoding process, as well as information encoded by an encoder and transmitted to the decoding apparatus, such as a syntax element.
- the residual signal may mean a difference between the original signal and the prediction signal, and a signal in which the difference between the original signal and the prediction signal is transformed or a signal in which the difference between the original signal and the prediction signal is converted and quantized It may mean.
- the residual signal may be referred to as a residual block.
- an encoding method such as exponential Golomb, context-adaptive variable length coding (CAVLC), or context-adaptive binary arithmetic coding (CABAC) may be used.
- the entropy encoder 850 may perform entropy encoding using a variable length coding (VLC) table.
- VLC variable length coding
- the entropy encoder 850 derives a binarization method of the target symbol and a probability model of the target symbol / bin, and then performs entropy encoding using the derived binarization method or the probability model. You may.
- the quantized coefficient may be inversely quantized by the inverse quantizer 860 and inversely transformed by the inverse transformer 870.
- the inverse quantized and inverse transformed coefficients are added to the prediction block through the adder 875 and a reconstruction block can be generated.
- the reconstruction block passes through the filter unit 880, and the filter unit 880 applies at least one or more of a deblocking filter, a sample adaptive offset (SAO), and an adaptive loop filter (ALF) to the reconstructed block or reconstructed picture. can do.
- the reconstructed block that has passed through the filter unit 880 may be stored in the reference image buffer 890.
- FIG. 9 is a block diagram illustrating a configuration of a video decoder according to an embodiment.
- FIG. 9 illustrates an embodiment of a video decoder applicable to a multi-view structure, wherein the video decoder for the multi-view structure may be implemented by extending a video decoder for a single view structure.
- the video decoder of FIG. 9 may be used in the real image decoder and / or the depth information map decoder of FIG. 7.
- “decoding” and “decoding” may be mixed, or “decoding device” and “decoder” may be mixed.
- the video decoder 900 includes an entropy decoder 910, an inverse quantizer 920, an inverse transformer 930, an intra predictor 940, an inter predictor 950, and a filter ( 960 and reference picture buffer 970.
- the video decoder 900 may receive the bitstream output from the encoder and perform decoding in an intra mode or an inter mode, and output a reconstructed image, that is, a reconstructed image.
- the switch In the intra mode, the switch may be switched for intra prediction, and in the inter mode, the switch may be switched for inter prediction.
- the video decoder 900 may obtain a residual block reconstructed from the received bitstream, generate a prediction block, and then add the reconstructed residual block and the prediction block to generate a reconstructed block, that is, a reconstruction block. .
- the entropy decoder 910 may entropy decode the input bitstream according to a probability distribution, and output information such as quantized coefficients and syntax elements.
- the quantized coefficients are inversely quantized by the inverse quantizer 920 and inversely transformed by the inverse transformer 930. Inverse quantization / inverse transformation of the quantized coefficients may produce a reconstructed residual block.
- the intra predictor 940 may generate a prediction block for the current block by using pixel values of an already encoded block around the current block.
- the inter prediction unit 950 may generate a prediction block for the current block by performing motion compensation using the reference picture stored in the motion vector and the reference picture buffer 970.
- inter prediction applied in inter mode may include inter view prediction.
- the inter prediction unit 950 may configure an inter view reference picture by sampling a picture of the reference view.
- the inter prediction unit 950 may perform inter view prediction by using a reference picture list including an inter view reference picture. Reference relationships between views may be signaled via information specifying dependencies between views.
- sampling applied to the reference view picture may mean generation of a reference sample by copying or interpolating a sample from the reference view picture.
- sampling applied to the reference view picture may mean upsampling or downsampling.
- the inter-view reference picture may be constructed by upsampling the reconstructed picture of the reference view.
- information specifying a view to which a picture to be used as the inter-view reference picture belongs may be transmitted from the encoder to the decoder.
- a view used in inter-view prediction that is, a picture used for prediction of the current block in the reference view may be a picture of the same access unit (AU) as the current picture (picture to be predicted in the current view).
- AU access unit
- the reconstructed residual block and the predictive block are added at the adder 955 to generate a reconstructed block.
- the residual sample and the predictive sample are added to generate a reconstructed sample or a reconstructed picture.
- the reconstructed picture is filtered by the filter unit 960.
- the filter unit 960 may apply at least one or more of the deblocking filter, SAO, and ALF to the reconstructed block or the reconstructed picture.
- the filter unit 960 outputs a modified or filtered reconstructed picture.
- the reconstructed picture may be stored in the reference picture buffer 970 and used for inter prediction.
- each module performs a different function
- the present invention is not limited thereto and may perform two or more functions in one module.
- operations of the intra predictor and the inter predictor may be performed in one module (prediction unit) in FIGS. 8 and 9.
- 8 and 9 illustrate that one encoder / decoder processes all encoding / decoding for the multi-view, this is for convenience of description, and the encoder / decoder may be configured for each view.
- the encoder / decoder of the current view may perform encoding / decoding of the current view using information of another view.
- the prediction unit (inter prediction unit) of the current view may perform intra prediction or inter prediction on the current block by using pixel information or reconstructed picture information of another view.
- the encoder / decoder may perform encoding / decoding on the current layer using information of another view regardless of whether it is configured for each view or one device processes multiple views. have.
- the description of the view in the present invention can be equally applied to a layer supporting scalability.
- the view may be a layer.
- FIG. 10 is a diagram illustrating an example of a prediction structure for a 3D video codec.
- FIG. 10 shows a prediction structure for encoding a real image acquired by three cameras and a depth information map corresponding to the real image.
- FIG. 10 three real images acquired from three cameras are represented as T0, T1, and T2 according to a view, and three depth maps corresponding to the actual image are respectively D0 and D1 according to a view. , Represented by D2.
- T0 and D0 are images acquired at View
- T1 and D1 are images acquired at View 1
- T2 and D2 are images acquired at View 2.
- the rectangle illustrated in FIG. 10 represents an image (picture).
- Each picture is divided into an I picture (Intra Picture), a P picture (Uni-prediction Picture), and a B picture (Bi-prediction Picture) according to an encoding / decoding type, and each picture is an encoding / decoding type of each picture.
- Can be encoded / decoded according to The picture itself may be encoded without inter prediction in the I picture, the inter prediction may be performed using only the reference picture existing in the unidirectional direction in the P picture, and the inter prediction may be performed using the reference picture existing in both directions in the B picture.
- the arrow of FIG. 10 indicates the prediction direction. That is, the actual image and its depth information map may be encoded / decoded depending on the prediction direction.
- motion information of the current block is required.
- a method of inferring current motion information there is a method of using motion information of a block adjacent to the current block, a method of using temporal correlation within the same viewpoint, or a method of using inter-view correlation at adjacent viewpoints.
- the method can be used interchangeably in one picture.
- the current block refers to a block on which prediction is performed.
- the motion information may mean a motion vector, a reference picture number, and / or a prediction direction (eg, unidirectional prediction or bidirectional prediction, whether to use temporal correlation or inter-view correlation).
- the prediction direction may be largely divided into unidirectional prediction and bidirectional prediction according to the use of a reference picture list (RefPicList).
- Unidirectional prediction is divided into forward prediction (Pred_L0: Prediction L0) using the forward reference picture list (LIST 0, L0) and backward prediction (Pred_L1: Prediction L1) using the reverse reference picture list (LIST 1, L1). do.
- bidirectional prediction (Pred_BI: Prediction BI) may use both the forward reference picture list (LIST 0) and the reverse reference picture list (LIST 1) to say that both forward and backward prediction exist.
- the list LIST 0 may be copied to the backward reference picture list LIST 1 to be included in the bidirectional prediction even when there are two forward predictions.
- Prediction direction can be defined using predFlagL0 and predFlagL1.
- predFlagL0 is an indicator indicating whether to use the forward reference picture list (List 0)
- predFlagL1 corresponds to an indicator indicating whether to use the backward reference picture list (List 1).
- predFlagL0 may be '1' for unidirectional prediction and forward prediction
- predFlagL1 may be '0' for unidirectional prediction
- predFlagL0 may be '0' for unidirectional prediction and reverse prediction
- predFlagL1 may be '1'.
- predFlagL0 may be '1' and predFlagL1 may be '1'.
- the motion prediction method includes 1) a motion vector predictor and a motion vector difference, and 2) a flag (merge_flag) indicating whether to use motion merging and a motion inheritance at a certain position.
- merge motion derivation that is, a merge method using an index (merge_idx) which is information on whether to receive.
- FIG. 11 shows an example of neighboring blocks used to construct a merge candidate list for a current block.
- Merge mode is one of methods of performing inter prediction, and in merge mode, neighboring blocks of the current block as motion information (eg, at least one of a motion vector, a reference picture list, and a reference picture index) of the current block.
- Motion information can be used.
- using motion information of the neighboring block as motion information of the current block is called merging, motion merging, or merging motion.
- a merging candidate list that is, a merge motion candidate list may be configured.
- the merge candidate list represents a list of motion information and may be generated before the merge mode is performed.
- the motion information of the merge candidate list may be motion information of neighboring blocks adjacent to the current block or new motion information created by combining motion information already present in the merge candidate list.
- the motion information (eg, the motion vector and / or the reference picture index) of the neighboring block may be motion information specified by the neighboring block or stored in the neighboring block (used for decoding the neighboring block).
- neighboring blocks A, B, C, D, and E that are spatially adjacent to the current block and the current block are temporal.
- a corresponding co-located block H or M may be included.
- a candidate block at the same position refers to a block at a corresponding position in a co-located picture corresponding to the current picture including the current block in time. If the H block in the picture at the same position is available, the H block may be determined as a candidate block at the same position. If the H block is not available, the M block in the picture at the same position may be determined as the candidate block at the same position.
- merge candidates in which motion information of neighboring blocks A, B, C, D, and E and candidate blocks H or M at the same position form a merge candidate list of the current block It can be determined whether or not can be used as. That is, motion information of blocks available for inter prediction of the current block may be added to the merge candidate list as merge candidates.
- a method of constructing a merge candidate list for an X block 1) First, when neighboring block A is available, the neighboring block A is included in the merge candidate list. 2) Then, the neighboring block B is included in the merge candidate list only when the motion information of the neighboring block B is not the same as the motion information of the neighboring block A. 3) In the same manner, the neighboring block C is included in the merge candidate list only when the motion information of the neighboring block C is different from the motion information of the neighboring block B. 4) The motion information of the neighboring block D is identical to the motion information of the neighboring block C. The neighbor block D is included in the merge candidate list only when different.
- the merge candidate list includes the neighboring block H (or M). Let's do it. That is, each neighboring block may be added to the merge candidate list in the order of A ⁇ B ⁇ C ⁇ D ⁇ E ⁇ H (or M) blocks.
- the same motion information may mean using the same motion vector, the same reference picture, and the same prediction direction (unidirectional or bidirectional).
- the expression of adding the neighboring block to the merge candidate list as the merge candidate and the expression of adding the motion information of the neighboring block to the merge candidate list as the merge candidate are used for convenience of explanation, and the two expressions are substantially different.
- the neighboring block as the merge candidate may mean motion information of the corresponding block.
- FIG. 12 is a diagram illustrating an example of a process of deriving motion information of a current block by using motion information of an adjacent view.
- FIG. 12 for convenience of description, a process of deriving motion information of the current block using only one adjacent viewpoint is illustrated as an example, but two or more adjacent viewpoints may be provided.
- motion information of adjacent viewpoints may be used to efficiently encode / decode motion information.
- the current block (block for the current position X) of FIG. 12 finds a target block (reference position XR) located at an adjacent viewpoint in order to derive motion information for the current block.
- the target block located in the adjacent view means a block corresponding to the current block, and since the current picture at the current view and the current picture at the reference view only have a difference in camera position, the variation vector ( Disparity Vector (DV) may be used to derive a target block located at an adjacent viewpoint.
- Disparity Vector DV
- FIG. 13 is a diagram illustrating an example in which one prediction unit (PU) is divided into several sub-prediction units.
- an example of a prediction unit (PU) having a 64x64 size is divided into sub prediction units having a size of 8x8.
- the prediction unit has a size of 64x64, but the prediction unit may have a size of 32x32, 16x16, 8x8, 4x4, etc., as well as the size of 64x64.
- one prediction unit may be divided into several sub-prediction units. In this case, the derivation of the motion information using the disparity vector is performed in units of sub prediction units.
- the size of the sub prediction unit may have a preset size (eg, 4x4, 8x8, 16x16, etc.), and the size of the sub prediction unit may be designated at the time of encoding.
- Information about the size of the sub prediction unit may be signaled by being included in a video parameter set extension syntax (VPS Extension syntax).
- VPS Extension syntax video parameter set extension syntax
- FIG. 14 is a diagram illustrating an example of a process of deriving motion information of a current block by using a reference block.
- the derivation of motion information for the current block means setting motion information existing in the reference block as motion information of the current block.
- derivation of motion information is performed in units of prediction blocks.
- motion is performed not only in the prediction block unit but also in each sub-prediction unit for the current block X located in the current picture at the current view. Information can be derived.
- motion information existing in the sub prediction unit of the reference block XR may be set as motion information for the sub prediction unit of the current block X.
- the reference block XR may refer to the reference block XR located in the current picture at the reference time point, and a detailed motion information derivation process will be described later.
- the inter prediction unit may induce motion information not only in the prediction unit but also in the sub prediction unit in the derived reference block of the reference view.
- the inter prediction unit obtains motion information in units of sub prediction units within the derived reference block of the reference view, not all sub prediction units have motion information. That is, there is a part in the reference block where motion information cannot be obtained.
- the inter prediction unit checks whether motion information available in the sub prediction unit located in the center of the reference block exists.
- the inter prediction unit stores the motion information present in the sub prediction unit located in the center of the reference block in a storage space
- the storage space is video encoded / decoded. It may mean a separate storage device other than the device, and may be located inside the video encoding / decoding device.
- the storage space may mean a memory, a processor, or the like, and the storage space may mean a virtual, logical storage space, not a physical space. In this case, the motion information stored in the storage space may mean initial motion information.
- the inter prediction unit derives the motion information for the sub prediction unit of the current block, if the motion information does not exist in the sub prediction unit of the reference block corresponding to the sub prediction unit of the current block, the above-described initial motion information is present. It can be used as motion information for the sub-prediction unit of the block. If there is no motion information available in the sub prediction unit located in the center of the reference block, the inter prediction unit may not use the temporal inter-view motion information prediction. In this case, the motion information may be used as a motion merge candidate, that is, a merge candidate.
- a merge motion candidate list that is, a merge candidate list, in a 3D image
- the merge candidate in the 3D image encoding / decoding (eg, 3D-HEVC) may be derived to form a list.
- the merge candidate in the 3D image encoding / decoding is derived, but in this specification, the derivation of the merge candidate is limited to the above-described order. It doesn't happen.
- the merge candidates are arranged in the list according to a predetermined order, but in the present specification, the merge candidates are adaptively rearranged according to the video encoding / decoding situation from the scope of rights. It is not.
- the existing 2D image encoding / decoding merge candidate may mean motion information of a block that is spatially and temporally adjacent to the current block, and the 3D image encoding / decoding merge candidate is a candidate of the image of the 3D image. It may mean merge candidates added according to characteristics.
- the motion information of a block that is spatially and temporally adjacent to the current block may include A0, A1, B0, B1, B2, and Col (H or M) candidates, and A1 is the left side of the prediction target block as described above.
- B1 may mean motion information of a block located above a prediction target block as described above.
- B0 means motion information of a block located on the upper right side of the prediction target block
- A0 means motion information of a block located on the lower left side of the prediction target block as described above
- B2 is described above. It may mean motion information of a block located on the upper left side of the prediction target block.
- Col (H or M) may refer to motion information derived through a block of a corresponding position in a co-located picture corresponding to the current picture including the current block in time. have.
- merge candidates added to the characteristics of the 3D image may include motion parameter inheritance (MPI), inter-view merging candidate (IVMC), and inter-view disparity vector. candidate, IvDC), view synthesis prediction (VSP), shift inter view (ShiftIV), and Bi and Zero candidates.
- MPI means motion information inheriting the motion information of the above-described real image
- IvMC means motion information using merge between views
- IvDC is derived by using the above-described disparity vector (ie, parallax). It may mean the movement information.
- VSP may mean motion information derived by synthesizing a viewpoint
- ShiftIV may mean motion information derived using a corrected parallax.
- Bi may mean a motion candidate derived using both directions
- Zero may mean a zero vector.
- FIG. 15 schematically illustrates the concept of View Synthesis Prediction (VSP).
- VSP View Synthesis Prediction
- the view synthesis prediction may use a texture picture, a reference texture picture and a reference depth picture, and the texture picture includes a current block and a current block. Neighboring blocks may be included.
- the reference depth picture may include a depth block used for the VSP and a block adjacent to the depth block used for the VSP.
- the texture picture is a picture to which the merge candidate is derived, and the texture picture may mean an actual image.
- the current block exists in the texture picture, and means a block in which prediction is made.
- the current block may mean a coding unit, and in the sense that prediction is performed in the current block, the current block may mean a prediction unit or the aforementioned sub prediction unit.
- the 'texture picture' may mean 'texture slice', 'texture view', 'texture image' and / or 'real picture', etc.
- the 'texture picture' may be represented as a 'texture slice', a 'texture view', a 'texture image' and / or a 'real picture'.
- the reference texture picture refers to a texture picture at a different point of time than the texture picture, and the reference texture pictures for the texture picture may be a plurality of pictures.
- the reference texture picture refers to a texture picture at a different viewpoint than the texture picture, but the reference texture picture does not exclude a case in which the reference texture picture is the same viewpoint as the texture picture.
- the 'reference texture picture' may include 'reference texture slice', 'reference texture view', 'reference texture image', and / or 'reference actual picture'.
- 'reference texture picture' means 'reference texture slice', 'reference texture view', 'reference texture image' and / Or as a “reference actual picture”.
- the reference depth picture refers to a depth information map that exists at a different point of time than the above-described texture picture, and the depth information map may be obtained through the above-described process.
- the reference depth picture refers to a depth picture at a different viewpoint than the texture picture, but the reference depth picture does not exclude a case in which the reference depth picture is the same viewpoint as the texture picture.
- the 'reference depth picture' may include 'reference depth slice', 'reference depth view', 'reference depth image' and / or 'reference depth picture', and the like.
- 'reference depth picture' may be referred to as 'Reference depth slice', 'Reference depth view', 'Reference depth image' and / Or as a “reference depth picture”.
- the VSP is generated using a disparity vector (DV), a referenced depth information map, and general image information of a corresponding view.
- DV disparity vector
- a method of generating a VSP will be described in more detail with reference to FIG. 15. 1.
- a 3D image encoding / decoding apparatus derives a DV present in a block neighboring a current block. Thereafter, the 3D image encoding / decoding apparatus 2. copies the DV from the neighboring block, and 3. generates a virtual newly generated block through backward warping using the reference texture picture. Thereafter, the current block uses the virtual newly generated block as VSP prediction information, that is, a VSP merge candidate.
- FIG. 16 illustrates an example of neighboring blocks inherited as VSP candidates in 3D image encoding / decoding.
- each neighboring block may be a prediction block / sub prediction block unit, and neighboring blocks may mean reconstructed blocks.
- each size of the neighboring blocks may have the same size as the current block or a different size from the current block, and having a size different from the current block may mean that the size is smaller or larger than the current block.
- Neighboring blocks may be square or non-square.
- neighboring blocks of the current block there may be a spatial neighboring block using the VSP.
- neighboring blocks of the current block that is, A1 and B0 of neighboring blocks may use the VSP. have.
- the spatial neighboring blocks (eg, A0, A1, B0, B1, B2) use the VSP
- the spatial neighboring blocks using the VSP are called Inherited VSP candidates because they are inherited by the VSP.
- the inherited VSP candidate uses the same disparity vector as the VSP candidate of the current block.
- 17 illustrates an example of a merge candidate list in 3D image encoding / decoding.
- FIG. 17 illustrates an example of a method of generating a merge candidate list when A1 and B0 use VSP candidates among neighboring blocks of the current block, that is, when A1 and B0 are inherited VSP candidates. It is shown schematically. In FIG. 17, the case where A1 and B0 are inherited VSP candidates is described, but each of the spatial neighboring blocks A0, A1, B0, B1, and B2 may have inherited VSP candidates.
- each block located on the lower left side A0, the left side A1, the upper side B1, the upper right side B0, and / or the upper left side B2 can be inherited as a VSP candidate, respectively, in which other peripherals Regardless of whether a block is an inherited VSP candidate, each neighboring block may be inherited as a VSP candidate.
- an apparatus and method for restricting the insertion of a VSP candidate in 3D image encoding / decoding are proposed.
- the efficiency of using a VSP candidate in a merge candidate list and the possibility of using another candidate may be secured. Can be.
- only one VSP candidate may be used in constructing the merge list.
- redundancy that may occur between VSP candidates in 3D image encoding / decoding may be removed, and another candidate may be inserted into the merge list, thereby increasing encoding / decoding efficiency.
- FIG. 18 is a flowchart of a method of restricting addition of a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to an embodiment of the present invention.
- the 3D image encoding / decoding apparatus derives a merge motion candidate, that is, a merge candidate (S1810).
- the derived merge candidates are A0, A1, B0, B1, B2, Col (H or M), Motion Parameter Inheritance (MPI), inter-view merging candidate (IVMC) as described above.
- MPI Motion Parameter Inheritance
- IVMC inter-view merging candidate
- IvDC Inter-view disparity vector candidate
- VSP view synthesis prediction
- ShiftIV shift inter-view
- Bi / or Zero candidates
- the 3D image encoding / decoding apparatus may insert the derived merge candidate into the merge candidate list.
- the derived merge candidate in step S1810 may mean A1 and / or B1
- the 3D image encoding / decoding apparatus may insert A1 and / or B1 into the merge candidate list.
- the 3D image encoding / decoding apparatus determines whether there is a block using a VSP among spatial candidates (S1820).
- the spatial candidate may mean a spatial neighboring block, and the spatial neighboring block may include A0, A1, B0, B1, B2 and / or Col (H or M) candidates as described above, for example, in step S1820.
- the determined spatial candidate block may be A1 and / or B1.
- a block using a VSP among spatial candidates may mean an inherited VSP candidate, and a detailed description of the inherited VSP candidate is as described above.
- the 3D image encoding / decoding apparatus inserts the VSP candidates into the merge candidate list (S1830).
- the VSP candidate inserted into the merge candidate list means the VSP candidate of the current block, and the details of the VSP candidate are as described above.
- the 3D image encoding / decoding apparatus derives the remaining merge motion candidates (S1840).
- the remaining merge motion candidates that is, the remaining merge candidates are A0, A1, B0, B1, B2, Col (H or M), Motion Parameter Inheritance (MPI), and inter-view merging candidates.
- IvMC inter-view disparity vector candidate
- IvDC inter-view disparity vector candidate
- VSP view synthesis prediction
- ShitIV shift interview
- Bi Bi and / or Zero candidate
- steps S1810 to S1830 A0, A1, B0, B1, B2, Col (H or M)
- Motion Parameter Inheritance MPI
- Interview merge candidate inter-view merging candidate (IvMC), inter-view disparity vector candidate (IvDC), view synthesis prediction (VSP), shift interview (ShiftIV), Bi and / or Zero candidates It may mean candidates except candidates inserted into the merge list.
- step S1840 the 3D image encoding / decoding apparatus merges the remaining merge candidates except for the A1, B1, and VSP candidates into the merge candidate list. You can derive whether or not to insert it.
- the 3D image encoding / decoding apparatus does not insert the VSP candidate into the merge candidate list and derives the remaining merge motion candidates (S1840). ).
- the details of the 3D image encoding / decoding apparatus deriving the remaining merge motion candidates are as described above.
- 19 is a flowchart illustrating a method of restricting adding a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to another embodiment of the present invention in 3D-HEVC.
- the 3D image encoding / decoding apparatus derives a merge motion candidate, that is, a merge candidate (S1910).
- a merge candidate S1910
- the details of the 3D image encoding / decoding apparatus inducing a merge candidate are as described above.
- the 3D image encoding / decoding apparatus determines whether there is a block using a VSP among spatial candidates (S1920). In this case, as described above, the 3D image encoding / decoding apparatus determines whether there is a block using the VSP among the spatial candidates.
- the 3D image encoding / decoding apparatus sets skipVSP to true (S1930).
- skipVSP means information for determining whether to insert the VSP candidate into the merge candidate list or not. If skipVSP is true, the VSP candidate is not inserted into the merge candidate list. In addition, when skipVSP is false, the VSP candidate is inserted into the merge candidate list.
- the 3D image encoding / decoding apparatus sets skipVSP to false (S1940).
- the 3D image encoding / decoding apparatus determines whether skipVSP is true (S1950).
- the 3D image encoding / decoding apparatus inserts the VSP candidate into the merge candidate list (S1960).
- the details of inserting the VSP candidate into the merge candidate list by the 3D image encoding / decoding apparatus are as described above.
- the 3D image encoding / decoding apparatus derives the remaining merge motion candidates (S1970).
- the details of deriving the remaining merge motion candidates by the 3D image encoding / decoding apparatus are as described above.
- step S1950 the 3D image encoding apparatus derives the remaining merge motion candidate without inserting the VSP candidate into the merge candidate list (S1970).
- step S1970 the 3D video encoding apparatus does not insert the VSP candidate into the merge candidate list and derives the remaining merge motion candidates as described above.
- the 3D image encoding / decoding apparatus may not add the VSP candidate to the merge candidate list. That is, through the above-described method, when there is an inherited VSP among neighboring blocks, that is, neighboring blocks, the 3D image encoding / decoding apparatus does not add the VSP candidate for the current block to the merge candidate list. Redundancy of existing merge candidates can be reduced, and additional merge candidates other than VSP candidates can be inserted into the merge candidate list, thereby improving coding efficiency.
- the apparatus for restricting the addition of the VSP candidate to the merge candidate list according to whether the inherited VSP candidate exists may be a configuration in which a module for removing redundancy of the VSP is added to the existing inter prediction unit.
- a module for removing redundancy of the VSP is added to the existing inter prediction unit.
- FIG. 20 is a block diagram of an apparatus for restricting adding a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to an embodiment of the present invention.
- the inter prediction unit 2000 may include an existing motion candidate generation module, a VSP redundancy removal module, and a 3D motion candidate generation module.
- the existing motion candidate generation module derives a merge motion candidate, that is, a merge candidate.
- the derived merge candidates are as described above, and details of each merge candidate are as described above.
- the existing motion candidate generation module may insert the derived merge candidate into the merge candidate list.
- the derived merge candidate may mean A1 and / or B1
- the 3D image encoding / decoding apparatus may insert A1 and / or B1 into the merge candidate list.
- the 'existing motion candidate generation module' may mean 'spatial merge candidate insertion module'.
- the 'existing motion candidate generation module' may be mixed with the 'spatial merge candidate insertion module'. have.
- the VSP redundancy removal module may determine whether any of the spatial candidates use a VSP block.
- a spatial candidate may mean a spatial neighboring block, and the spatial neighboring block may include A0, A1, B0, B1, B2 and / or Col (H or M) candidates as described above, eg, eliminating VSP redundancy.
- the spatial candidate blocks determined in the module may be A1 and / or B1.
- the block using the VSP among the spatial candidates may mean an inherited VSP candidate, and a detailed description of the inherited VSP candidate is as described above.
- the VSP redundancy removal module may set the skipVSP value to true or false based on the above determination result. In this case, the VSP redundancy removal module sets the skipVSP value to true or false based on the above determination result, as described above.
- the 'VSP redundancy removal module' may mean 'VSP redundancy determination module'.
- the 'VSP redundancy removal module' and the 'VSP redundancy determination module' may be used interchangeably.
- the 3D motion candidate generation module inserts the VSP candidate into the merge candidate list and then derives the remaining merge motion candidates. Details of the 3D motion candidate generation module deriving the remaining merge motion candidates are the same as described above, and details of the remaining merge motion candidates are the same as described above.
- the 3D image encoding / decoding apparatus is a 3D image encoding / decoding apparatus except for the A1, B1, and VSP candidates. It may be derived whether to insert the remaining merge candidates into the merge candidate list.
- the 3D motion candidate generation module does not insert the VSP candidates into the merge candidate list. Deriving a merge motion candidate. The details of the 3D image encoding / decoding apparatus deriving the remaining merge motion candidates are as described above.
- the 'three-dimensional motion candidate generation module' may mean 'three-dimensional merge candidate insertion module' or 'VSP candidate insertion module'.
- the 'three-dimensional motion candidate generation module' and '3' The dimension merge candidate insertion module 'and the' VSP candidate insertion module 'may be used interchangeably.
- the apparatus for restricting the addition of the VSP candidate to the merge candidate list according to whether the inherited VSP candidate exists may be a configuration in which the VSP redundancy removal module is extended / added to the 3D motion candidate generation module. In this case, it may be determined whether the VSP candidate is used according to whether the spatial neighbor candidate determined by the existing motion candidate generation module is to be used, and the determination of whether the VSP candidate is to be used may be made before the decision on deriving the VSP candidate.
- an apparatus for restricting adding a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to another embodiment of the present invention will be described in more detail with reference to FIG. 21.
- 21 is a block diagram of an apparatus for restricting adding a VSP candidate to a merge candidate list according to whether an inherited VSP candidate exists according to another embodiment of the present invention.
- the inter prediction unit 2100 may include an existing motion candidate generation module and a 3D motion candidate generation module, and the 3D motion candidate generation module may include a VSP redundancy removal module.
- the existing motion candidate generation module is as described above, and details of the existing motion candidate generation module are as described above.
- the 3D motion candidate generation module may include a VSP redundancy removal module, and may determine whether a VSP candidate is to be used or not, according to whether a spatial neighbor candidate is used in the VSP candidate, which is determined in the existing motion candidate generation module. May be made prior to the determination of VSP candidate derivation. Details of the 3D motion candidate generation module and the VSP redundancy removal module are as described above.
- the 3D encoding / decoding apparatus may refer to only two or one direction among the spatial merge candidates, By referring to only two or one direction among spatial merge candidates, the decoding apparatus may reduce the number of times of referencing a flag for determining whether to use a VSP. In this case, decreasing the number of times of referring to a flag for determining whether to use a VSP may reduce the number of times of memory access by the 3D encoding / decoding apparatus.
- FIG. 22 is a view illustrating a merge candidate derivation method and a merge candidate derivation method according to another embodiment of the present invention.
- each of the spatial neighboring blocks A0, A1, B0, B1, and / or B2 may be encoded / decoded in a VSP mode, that is, an inherited VSP in derivation of an existing merge candidate.
- VSP mode that is, an inherited VSP in derivation of an existing merge candidate.
- FIG. 22B illustrates an example of derivation of a merge candidate according to another embodiment of the present invention.
- the 3D encoding / decoding apparatus constructs a merge candidate list for the current block, whether the VSP is used in A1 and / or B1 among neighboring blocks, that is, whether the VSP is inherited To judge. Thereafter, when the VSP is used in A1 and / or B1 among neighboring blocks, the 3D image encoding / decoding apparatus restricts the insertion of the VSP candidate into the merge list.
- FIG. 23 is a flowchart illustrating a method of restricting insertion of a VSP candidate into a merge list by referring to two directions among spatial merge candidates according to another embodiment of the present invention.
- the 3D image encoding / decoding apparatus inserts a candidate located on the left side of the current block, that is, a candidate located on the left side (S2300).
- a candidate located on the left side of the current block is only an example of the present invention, and the 3D image encoding / decoding apparatus is performed in step S2300.
- the candidate to insert into the list may be any one of spatial candidates (eg, candidates located on the left, bottom left, top left, top, and top right sides of the current block).
- the candidate located on the left side of the current block may be A1, the candidate located on the lower left side may be A0, and the candidate located on the upper left side may be B2 candidate.
- the candidate located on the upper side of the current block may be B1, and the candidate located on the upper right side may be B0.
- the 3D image encoding / decoding apparatus determines whether the left candidate uses a VSP, that is, whether the left candidate is an inherited VSP (S2310), and the specific method of determining whether the left candidate uses the VSP is as described above.
- the 3D image encoding / decoding apparatus When the left candidate uses the VSP, the 3D image encoding / decoding apparatus generates information indicating that the VSP is used (S2320), and the details of the information indicating that the VSP is used have been described above.
- the 3D image encoding / decoding apparatus inserts a candidate located above the current block into the merge list (S2330).
- the candidate located above the current block is only an example of the present invention, and the 3D image encoding / decoding apparatus is performed in step S2330.
- the candidate to insert into the list may be any one of spatial candidates (eg, candidates located on the left, bottom left, top left, top, and top right sides of the current block).
- the 3D image encoding / decoding apparatus does not generate information indicating that the VSP is used, and inserts an upper candidate into the list (S2330).
- the 3D image encoding / decoding apparatus determines whether information indicating that the upper candidate uses the VSP and that the VSP is used does not exist (S2340).
- the 3D image encoding / decoding apparatus If the upper candidate used the VSP and there is no information indicating that the VSP is used, the 3D image encoding / decoding apparatus generates information indicating that the VSP is used (S2350). In this case, the information indicating that the VSP is used in step S2350 is generated only when the upper candidate uses the VSP and there is no information indicating that the VSP is used.
- the 3D image encoding / decoding apparatus determines whether information indicating that the VSP is used exists (S2360), and details of determining whether information indicating that the VSP is used exist as described above.
- the 3D image encoding / decoding apparatus when there is information indicating that the VSP is used, the 3D image encoding / decoding apparatus does not insert the VSP candidate into the merge candidate list.
- the 3D image encoding / decoding apparatus If there is no information indicating that the VSP is used in step S2360, the 3D image encoding / decoding apparatus generates the VSP candidate to the merge candidate list, that is, inserts the VSP candidate into the merge candidate list (S2370).
- the above-described method can be applied to each spatial neighbor candidate, and if VSP inheritance is made from the previous spatial candidate, the current spatial candidate can be used as a candidate for predicting using only DV.
- the above-described method may be implemented in the above-described three-dimensional image encoding / decoding apparatus.
- 24A to 24B are flowcharts illustrating a method of restricting insertion of a VSP candidate into a merge list by referring to two directions among spatial merge candidates according to another embodiment of the present invention, in 3D HEVC.
- the 3D image encoding / decoding apparatus derives an existing HEVC merge candidate and sets usedVSP to false (S2400). Detailed description of deriving an existing HEVC merge candidate and setting usedVSP to false is as described above.
- the 3D image encoding / decoding apparatus determines whether the left candidate uses the VSP (S2410). In this case, the details of whether the 3D image encoding / decoding apparatus uses the VSP for the left candidate are as described above.
- the 3D image encoding / decoding apparatus inserts the left candidate into the list as the inherited VSP candidate and sets usedVSP to true (S2430).
- the details of inserting the left candidate into the list as the inherited VSP candidate and setting the usedVSP to true are as described above.
- the 3D image encoding / decoding apparatus inserts the left candidate into the merge candidate list (S2420).
- the details of inserting the left candidate into the merge candidate list by the 3D image encoding / decoding apparatus are as described above.
- the 3D image encoding / decoding apparatus determines whether the upper candidate uses the VSP and whether the usedVSP is false (S2440). In this case, as described above, the image encoding / decoding apparatus determines whether the VSP is used for the upper candidate and whether the usedVSP is false.
- the 3D video encoding / decoding apparatus inserts the upper candidate into the list as the inherited VSP candidate and sets usedVSP to true.
- the 3D image encoding / decoding apparatus inserts the upper candidate into the merge list (S2460).
- the details of inserting the upper candidate into the merge list by the 3D image encoding / decoding apparatus are as described above.
- the 3D image encoding / decoding apparatus derives the remaining spatial merging candidates (S2470).
- the details of deriving the spatial merge candidate in which the 3D image encoding / decoding device remains are as described above.
- the 3D image encoding / decoding apparatus derives the merge candidate before the VSP candidate (S2475).
- the merge candidates after the step S2470 and before the VSP candidates are A0, A1, B0, B1, B2, Col (H or M), Motion Parameter Inheritance (MPI), and inter-view merging candidate.
- MPI Motion Parameter Inheritance
- IvMC inter-view disparity vector candidate
- IvDC inter-view disparity vector candidate
- VSP view synthesis prediction
- ShiftIV shift interview
- Bi Bi and / or Zero candidate
- the 3D image encoding / decoding apparatus determines whether usedVSP is true (S2480). In this case, the details of determining whether the 3D image encoding / decoding device is true for the usedVSP are as described above.
- the 3D image encoding / decoding apparatus inserts the VSP candidate into the merge list (S2485). In this case, the 3D image encoding / decoding apparatus inserts the VSP candidate into the merge list, and the details of the VSP candidate are as described above.
- the 3D image encoding / decoding apparatus inserts the remaining merge candidates into the list (S2490).
- the details of inserting the remaining merge candidates are as described above.
- the 3D image encoding / decoding apparatus inserts the remaining merge candidates into the list without inserting the VSP candidates into the list (S2490). At this time, the specific details of inserting the remaining merge candidates into the list without inserting the VSP candidates into the list are as described above.
- 25 is a view illustrating a merge candidate derivation method and a merge candidate derivation method according to another embodiment of the present invention.
- 25A illustrates an example of derivation for an existing merge candidate.
- specific details regarding the derivation of the existing merge candidate are as described above.
- FIG. 22B illustrates an example of derivation of a merge candidate according to another embodiment of the present invention.
- the 3D encoding / decoding apparatus determines whether a VSP is used in A1 among neighboring blocks, that is, whether it is an inherited VSP. Subsequently, when the VSP is used in A1 among neighboring blocks, the 3D image encoding / decoding apparatus restricts the insertion of the VSP candidate into the merge list.
- FIG. 26 is a flowchart illustrating a method of restricting insertion of a VSP candidate into a merge list with reference to one direction among spatial merge candidates according to another embodiment of the present invention.
- the 3D image encoding / decoding apparatus inserts a candidate located on the left side of the current block, that is, a candidate located on the left side (S2600).
- a candidate located on the left side of the current block is only an example of the present invention, and the 3D image encoding / decoding apparatus is performed in step S2600.
- the candidate to insert into the list may be any one of spatial candidates (eg, candidates located on the left, bottom left, top left, top, and top right sides of the current block).
- the 3D image encoding / decoding apparatus determines whether the left candidate uses a VSP, that is, whether the left candidate is an inherited VSP (S2610), and the specific method of determining whether the left candidate uses the VSP is as described above.
- the 3D image encoding / decoding apparatus When the left candidate uses the VSP, the 3D image encoding / decoding apparatus generates information indicating that the VSP is used (S2620), and the details of the information indicating that the VSP is used have been described above.
- the 3D image encoding / decoding apparatus does not generate information indicating that the VSP is used.
- the 3D image encoding / decoding apparatus determines whether information indicating that the VSP is used exists (S2630), and details of determining whether information indicating that the VSP has been used exist as described above.
- the 3D image encoding / decoding apparatus when there is information indicating that the VSP is used, the 3D image encoding / decoding apparatus does not insert the VSP candidate into the merge candidate list.
- the 3D image encoding / decoding apparatus If there is no information indicating that the VSP is used in step S2630, the 3D image encoding / decoding apparatus generates the VSP candidate to the merge candidate list, that is, inserts the VSP candidate into the merge candidate list (S2640).
- the above-described method can be applied to each spatial neighbor candidate, and if VSP inheritance is made from the previous spatial candidate, the current spatial candidate can be used as a candidate for predicting using only DV.
- the above-described method may be implemented in the above-described three-dimensional image encoding / decoding apparatus.
- 27A to 27B are flowcharts illustrating a method of restricting insertion of a VSP candidate into a merge list with reference to one direction among spatial merge candidates according to another embodiment of the present invention, in 3D HEVC.
- the 3D image encoding / decoding apparatus derives an existing HEVC merge candidate and sets usedVSP to false (S2700). Detailed description of deriving an existing HEVC merge candidate and setting usedVSP to false is as described above.
- the 3D image encoding / decoding apparatus determines whether the left candidate uses a VSP (S2710). In this case, the details of whether the 3D image encoding / decoding apparatus uses the VSP for the left candidate are as described above.
- the 3D image encoding / decoding apparatus inserts the left candidate into the list as the inherited VSP candidate, and sets usedVSP to true (S2720).
- the details of inserting the left candidate into the list as the inherited VSP candidate and setting the usedVSP to true are as described above.
- the 3D image encoding / decoding apparatus inserts the left candidate into the merge candidate list (S2730).
- the details of inserting the left candidate into the merge candidate list by the 3D image encoding / decoding apparatus are as described above.
- the 3D image encoding / decoding apparatus derives the remaining spatial merging candidates (S2740).
- the details of deriving the spatial merge candidate in which the 3D image encoding / decoding device remains are as described above.
- the 3D image encoding / decoding apparatus derives a merge candidate before the VSP candidate (S2750).
- the merge candidates after the step S2740 and before the VSP candidates are A0, A1, B0, B1, B2, Col (H or M), Motion Parameter Inheritance (MPI), and inter-view merging candidates.
- MPI Motion Parameter Inheritance
- IvMC inter-view disparity vector candidate
- IvDC inter-view disparity vector candidate
- VSP view synthesis prediction
- ShiftIV shift interview
- Bi Bi and / or Zero candidate
- the 3D image encoding / decoding apparatus determines whether usedVSP is true (S2760). In this case, the details of determining whether the 3D image encoding / decoding device is true for the usedVSP are as described above.
- the 3D image encoding / decoding apparatus inserts the VSP candidate into the merge list (S2770). In this case, the 3D image encoding / decoding apparatus inserts the VSP candidate into the merge list, and the details of the VSP candidate are as described above.
- the 3D image encoding / decoding apparatus inserts the remaining merge candidates into the list (S2780).
- the details of inserting the remaining merge candidates are as described above.
- the 3D image encoding / decoding apparatus inserts the remaining merge candidates into the list without inserting the VSP candidates into the list (S2780). At this time, the specific details of inserting the remaining merge candidates into the list without inserting the VSP candidates into the list are as described above.
- FIG. 28 is a block diagram of an apparatus for restricting inserting a VSP candidate into a merge list according to another embodiment of the present invention.
- an apparatus for restricting insertion of a VSP candidate into a merge list may include a 3D HEVC candidate generation module, such as an existing HEVC candidate generation module and a VSPO usage determination module for a left block.
- the existing HEVC candidate generation module may mean the existing motion candidate generation module described above, and details of the existing HEVC candidate generation module are as described above.
- the VSP usage determination module for the left block and the like may each include a VSP deactivation module of an existing candidate, The VSP determination module may deactivate the use of the VSP of another spatial neighbor candidate according to whether the spatial neighbor candidate uses the VSP. As a result, only the candidates in a predetermined direction are used as the VSP candidates, thereby reducing the number of accesses to the VSP flags stored for each block. As a result, in three-dimensional image encoding / decoding, as a result, memory access can be reduced.
- the 3D HEVC candidate generation module may mean the above-described 3D motion candidate generation module, and the 3D HEVC candidate generation module may include a VSP redundancy removal module.
- the above-described 3D video encoding / decoding apparatus and method may be used in 3D-HEVC (High Efficiency Video Coding), which is currently being jointly standardized by a Moving Picture Experts Group (MPEG) and a Video Coding Experts Group (VCEG). Accordingly, the apparatus and method described above may vary in application range according to a block size, a coding unit depth (CU) depth, a transform unit (TU) depth, and the like as shown in Table 1.
- a variable (ie, size or depth information) for determining an application range may be set to use a predetermined value in an encoding / decoding apparatus and a method, or a value determined according to a profile or level may be used.
- variable value may be obtained from the bitstream in the decoding apparatus and the method.
- Table 1 when varying the application range according to the CU depth, A) may be applied only to a depth above a given depth, B) only to be applied below a given depth, and C) may be applied only to a given depth. Can be.
- any flag may indicate that the method described above does not apply, and the CU depth value indicative of coverage is one more than the maximum value of the CU depth. It may be represented by signaling a large value.
- 29 illustrates an example in which redundancy between VSP candidates is removed through the above-described 3D image encoding / decoding apparatus and method.
- VSP candidates are present in the merge candidate list by the conventional method, the VSP candidates are duplicated by the 3D encoding / decoding apparatus proposed herein. You can see that.
- the methods are described based on a flowchart as a series of steps or units, but the present invention is not limited to the order of steps, and certain steps may occur in a different order or simultaneously from other steps as described above. Can be. Also, one of ordinary skill in the art appreciates that the steps shown in the flowcharts are not exclusive, that other steps may be included, or that one or more steps in the flowcharts may be deleted without affecting the scope of the present invention. I can understand.
- the method according to the present invention described above may be stored in a computer-readable recording medium that is produced as a program for execution on a computer, and examples of the computer-readable recording medium include ROM, RAM, CD-ROM, magnetic tape , Floppy disks, optical data storage devices, and the like, and also include those implemented in the form of carrier waves (eg, transmission over the Internet).
- the computer readable recording medium can be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion.
- functional programs, codes, and code segments for implementing the method can be easily inferred by programmers in the art to which the present invention belongs.
Abstract
Description
CU(또는 PU, 또는 TU)의 깊이 | A | B | C |
0 | X | O | X |
1 | X | O | X |
2 | O | O | O |
3 | O | X | X |
4 이상 | O | X | X |
Claims (20)
- 제1 후보 블록을 머지 후보 리스트에 삽입하는 단계;상기 제1 후보 블록에서 시점 합성 예측(View Synthesis Prediction, VSP)이 사용된 경우, VSP가 사용되었음을 나타내는 정보를 생성하는 단계; 및상기 VSP가 사용되었음을 나타내는 정보가 존재하는 경우, 머지 후보 리스트에 현재 블록의 VSP 후보를 삽입하지 않는 단계를 포함하는 3차원 영상 복호화 방법.
- 제1항에 있어서,상기 제1 후보 블록은 공간적 이웃 블록인 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제2항에 있어서,상기 제1 후보 블록은 현재 블록의 좌측에 위치한 상기 공간적 이웃 블록인 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제1항에 있어서,상기 제1 후보 블록은 A1 블록인 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제1항에 있어서,상기 VSP가 사용되었음을 나타내는 정보는 제1 후보 블록이 사용 가능한지 여부에 대해 지시하는 정보 또는 제1 후보 블록에서 VSP가 사용되었는지 여부에 대해 지시하는 정보를 포함하는 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제1항에 있어서,상기 제1 후보 블록을 머지 후보 리스트에 삽입한 후, 제2 후보 블록을 머지 후보 리스트에 삽입하는 단계를 더 포함하고,상기 VSP가 사용되었음을 나타내는 정보를 생성하는 단계에서는, 상기 제1 후보 블록 또는 상기 제2 후보 블록에서 VSP가 사용된 경우, 상기 VSP가 사용되었음을 나타내는 정보를 생성하는 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제6항에 있어서,상기 제1 후보 블록은 현재 블록의 좌측에 위치한 공간적 이웃 블록이고, 상기 제2 후보 블록은 현재 블록의 상측에 위치한 공간적 이웃 블록인 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제6항에 있어서,상기 제1 후보 블록은 A1 블록이고, 상기 제2 후보 블록은 B1 블록인 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제1항에 있어서,상기 현재 블록은 서브 예측 블록인 것을 특징으로 하는 3차원 영상 복호화 방법.
- 제1 후보 블록을 머지 후보 리스트에 삽입하는 공간적 머지 후보 삽입 모듈;상기 제1 후보 블록에서 시점 합성 예측(View Synthesis Prediction, VSP)이 사용된 경우, VSP가 사용되었음을 나타내는 정보를 생성하는 VSP 중복성 판단 모듈; 및상기 VSP가 사용되었음을 나타내는 정보가 존재하는 경우, 머지 후보 리스트에 현재 블록의 VSP 후보를 삽입하지 않는 VSP 후보 삽입 모듈을 포함하는 3차원 영상 복호화 장치.
- 제10항에 있어서,상기 제1 후보 블록은 공간적 이웃 블록인 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제11항에 있어서,상기 제1 후보 블록은 현재 블록의 좌측에 위치한 상기 공간적 이웃 블록인 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제10항에 있어서,상기 제1 후보 블록은 A1 블록인 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제10항에 있어서,상기 VSP가 사용되었음을 나타내는 정보는 제1 후보 블록이 사용 가능한지 여부에 대해 지시하는 정보 또는 제1 후보 블록에서 VSP가 사용되었는지 여부에 대해 지시하는 정보를 포함하는 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제10항에 있어서,상기 공간적 머지 후보 삽입 모듈은 상기 제1 후보 블록을 머지 후보 리스트에 삽입한 후, 제2 후보 블록을 머지 후보 리스트에 삽입하고,상기 VSP 중복성 판단 모듈은 상기 제1 후보 블록 또는 상기 제2 후보 블록에서 VSP가 사용된 경우, 상기 VSP가 사용되었음을 나타내는 정보를 생성하는 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제15항에 있어서,상기 제1 후보 블록은 현재 블록의 좌측에 위치한 공간적 이웃 블록이고, 상기 제2 후보 블록은 현재 블록의 상측에 위치한 공간적 이웃 블록인 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제15항에 있어서,상기 제1 후보 블록은 A1 블록이고, 상기 제2 후보 블록은 B1 블록인 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제10항에 있어서,상기 현재 블록은 서브 예측 블록인 것을 특징으로 하는 3차원 영상 복호화 장치.
- 제1 후보 블록을 머지 후보 리스트에 삽입하는 단계;상기 제1 후보 블록에서 시점 합성 예측(View Synthesis Prediction, VSP)이 사용된 경우, VSP가 사용되었음을 나타내는 정보를 생성하는 단계; 및상기 VSP가 사용되었음을 나타내는 정보가 존재하는 경우, 머지 후보 리스트에 현재 블록의 VSP 후보를 삽입하지 않는 단계를 포함하는 3차원 영상 부호화 방법.
- 제1 후보 블록을 머지 후보 리스트에 삽입하는 공간적 머지 후보 삽입 모듈;상기 제1 후보 블록에서 시점 합성 예측(View Synthesis Prediction, VSP)이 사용된 경우, VSP가 사용되었음을 나타내는 정보를 생성하는 VSP 중복성 판단 모듈; 및상기 VSP가 사용되었음을 나타내는 정보가 존재하는 경우, 머지 후보 리스트에 현재 블록의 VSP 후보를 삽입하지 않는 VSP 후보 삽입 모듈을 포함하는 3차원 영상 부호화 장치.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016572762A JP6673854B2 (ja) | 2014-06-30 | 2015-05-29 | 動き併合モードで視点合成予測候補の重複性除去装置および方法 |
EP15814370.1A EP3163880A4 (en) | 2014-06-30 | 2015-05-29 | Device and method for eliminating redundancy of view synthesis prediction candidate in motion merge mode |
CN201580031649.2A CN106464907B (zh) | 2014-06-30 | 2015-05-29 | 运动合并模式下消除视点合成预测候选冗余的装置和方法 |
US15/317,782 US10194133B2 (en) | 2014-06-30 | 2015-05-29 | Device and method for eliminating redundancy of view synthesis prediction candidate in motion merge mode |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR20140081191 | 2014-06-30 | ||
KR10-2014-0081191 | 2014-06-30 | ||
KR1020150074951A KR102378459B1 (ko) | 2014-06-30 | 2015-05-28 | 움직임 병합 모드에서 시점 합성 예측 후보의 중복성 제거 장치 및 방법 |
KR10-2015-0074951 | 2015-05-28 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016003074A1 true WO2016003074A1 (ko) | 2016-01-07 |
Family
ID=55019561
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2015/005412 WO2016003074A1 (ko) | 2014-06-30 | 2015-05-29 | 움직임 병합 모드에서 시점 합성 예측 후보의 중복성 제거 장치 및 방법 |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2016003074A1 (ko) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017531396A (ja) * | 2014-10-07 | 2017-10-19 | サムスン エレクトロニクス カンパニー リミテッド | ビュー併合予測を利用して、映像を符号化または復号する方法及びその装置 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013067938A1 (en) * | 2011-11-07 | 2013-05-16 | LI, Yingjin | Method of constructing merge list |
WO2013106336A2 (en) * | 2012-01-09 | 2013-07-18 | Qualcomm Incorporated | Device and methods for merge list reordering in video coding |
WO2014053086A1 (en) * | 2012-10-05 | 2014-04-10 | Mediatek Singapore Pte. Ltd. | Method and apparatus of motion vector derivation 3d video coding |
WO2014058280A1 (ko) * | 2012-10-12 | 2014-04-17 | 한국전자통신연구원 | 영상의 부호화/복호화 방법 및 이를 이용하는 장치 |
-
2015
- 2015-05-29 WO PCT/KR2015/005412 patent/WO2016003074A1/ko active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013067938A1 (en) * | 2011-11-07 | 2013-05-16 | LI, Yingjin | Method of constructing merge list |
WO2013106336A2 (en) * | 2012-01-09 | 2013-07-18 | Qualcomm Incorporated | Device and methods for merge list reordering in video coding |
WO2014053086A1 (en) * | 2012-10-05 | 2014-04-10 | Mediatek Singapore Pte. Ltd. | Method and apparatus of motion vector derivation 3d video coding |
WO2014058280A1 (ko) * | 2012-10-12 | 2014-04-17 | 한국전자통신연구원 | 영상의 부호화/복호화 방법 및 이를 이용하는 장치 |
Non-Patent Citations (1)
Title |
---|
Y. ZHANG ET AL.: "Simplification for VSP-related Merge Candidate List Construction.", JCT3V DOCUMENT JCT3V-I0112V1, 26 June 2014 (2014-06-26), pages 1 - 3, XP030132421 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017531396A (ja) * | 2014-10-07 | 2017-10-19 | サムスン エレクトロニクス カンパニー リミテッド | ビュー併合予測を利用して、映像を符号化または復号する方法及びその装置 |
US10321145B2 (en) | 2014-10-07 | 2019-06-11 | Samsung Electronics Co., Ltd. | Method and device for encoding or decoding image by means of view merging prediction |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2018155986A2 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2021025451A1 (ko) | 움직임 정보 후보를 이용한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020180155A1 (ko) | 비디오 신호를 처리하기 위한 방법 및 장치 | |
WO2021040480A1 (ko) | 인루프 필터링 기반 영상 코딩 장치 및 방법 | |
WO2021034123A1 (ko) | 가중 예측을 수행하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020180129A1 (ko) | 인터 예측을 위한 비디오 신호의 처리 방법 및 장치 | |
WO2020251330A1 (ko) | 단순화된 mpm 리스트 생성 방법을 활용하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2019194514A1 (ko) | 인터 예측 모드 기반 영상 처리 방법 및 이를 위한 장치 | |
WO2020184952A1 (ko) | 비디오 신호에서 인터 예측을 위한 움직임 벡터 차분 정보를 처리하기 위한 비디오 신호의 처리 방법 및 장치 | |
WO2020251329A1 (ko) | Mip 모드 매핑이 단순화된 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2019216714A1 (ko) | 인터 예측 모드 기반 영상 처리 방법 및 이를 위한 장치 | |
WO2020184964A1 (ko) | 인터 예측을 위한 비디오 신호의 처리 방법 및 장치 | |
WO2021006579A1 (ko) | 머지 후보의 양방향 예측을 위한 가중치 인덱스를 유도하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020251328A1 (ko) | 인트라 예측 모드 변환에 기반한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020256400A1 (ko) | 루마 맵핑 및 크로마 스케일링 기반 비디오 또는 영상 코딩 | |
WO2021172914A1 (ko) | 레지듀얼 코딩에 대한 영상 디코딩 방법 및 그 장치 | |
WO2020256482A1 (ko) | 변환에 기반한 영상 코딩 방법 및 그 장치 | |
WO2020262962A1 (ko) | 크로마 변환 블록의 최대 크기 제한을 이용한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020184958A1 (ko) | 인터 예측을 위한 비디오 신호의 처리 방법 및 장치 | |
WO2021172907A1 (ko) | 영상 디코딩 방법 및 그 장치 | |
WO2020180153A1 (ko) | 인터 예측을 위한 비디오 신호의 처리 방법 및 장치 | |
WO2020185045A1 (ko) | 크로마 블록에 대한 분할 제한을 이용한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020256485A1 (ko) | 크로마 블록의 적응적 크기 제한을 이용한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2021040406A1 (ko) | 비디오 또는 영상 코딩 방법 및 그 장치 | |
WO2020256488A1 (ko) | 크로마 블록의 크기를 제한하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15814370 Country of ref document: EP Kind code of ref document: A1 |
|
REEP | Request for entry into the european phase |
Ref document number: 2015814370 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15317782 Country of ref document: US Ref document number: 2015814370 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2016572762 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |