WO2014107083A1 - 비디오 신호 처리 방법 및 장치 - Google Patents
비디오 신호 처리 방법 및 장치 Download PDFInfo
- Publication number
- WO2014107083A1 WO2014107083A1 PCT/KR2014/000135 KR2014000135W WO2014107083A1 WO 2014107083 A1 WO2014107083 A1 WO 2014107083A1 KR 2014000135 W KR2014000135 W KR 2014000135W WO 2014107083 A1 WO2014107083 A1 WO 2014107083A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- intra prediction
- mode
- block
- list
- prediction mode
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
- H04N19/159—Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/196—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
- H04N19/197—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters including determination of the initial value of an encoding parameter
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/11—Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
- H04N19/463—Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
Definitions
- the present invention relates to a method and apparatus for coding a video signal.
- Compression coding refers to a series of signal processing techniques that transmit digitized information through a communication line or store the data in a form suitable for a storage medium.
- the object of compression encoding includes objects such as voice, video, text, and the like.
- a technique of performing compression encoding on an image is called video image compression.
- a general feature of a multiview video image is that it has spatial redundancy, temporal redundancy and inter-view redundancy.
- An object of the present invention is to improve the coding efficiency of a video signal.
- the video signal processing method generates a candidate mode list of a current block based on an intra prediction mode from a spatial neighboring block of the current block, and the candidate mode list based on an intra prediction mode from a neighboring viewpoint of the current block. And induce an intra prediction mode of the current block based on the modified candidate mode list, and predict the current block by using the derived intra prediction mode.
- the spatial neighboring block according to the present invention includes at least one of a left neighboring block and an upper neighboring block adjacent to the current block.
- the modifying of the candidate mode list according to the present invention may include obtaining a disparity vector of the current block, obtaining a reference block of a neighbor view based on the disparity vector, and whether the reference block is a block coded in an intra mode. If it is determined whether the reference block is a block coded in an intra mode, the candidate mode list is modified using an intra prediction mode of the reference block.
- the disparity vector of the current block according to the present invention is obtained by using the disparity vector of at least one neighboring block among neighboring blocks located at the left, the top, the bottom left, the top right and the top left of the current block. .
- the disparity vector of the current block according to the present invention may be derived from depth data corresponding to the current block.
- the candidate mode list defines a plurality of intra prediction modes available for the current block and list index information corresponding to each of the plurality of intra prediction modes, wherein the list index information is sequentially included in the candidate mode list. It is characterized by specifying the position of each of the arranged plurality of intra prediction modes.
- the modifying of the candidate mode list according to the present invention may correspond to the same intra prediction mode as the intra prediction mode of the neighboring view when the same intra prediction mode exists in the candidate mode list from the neighboring viewpoint.
- the candidate mode list is rearranged by changing list index information.
- the modifying of the candidate mode list according to the present invention may include replacing any one of a plurality of intra prediction modes in the candidate mode list with an intra prediction mode from the neighboring viewpoint.
- the modifying of the candidate mode list according to the present invention may include adding an intra prediction mode from the neighboring viewpoint to the candidate mode list generated based on the intra prediction mode of the spatial neighboring block.
- the present invention can obtain a more accurate intra prediction mode by generating a candidate mode list from an intra prediction mode of a neighboring block spatially adjacent to the current block and using the same to derive the intra prediction mode of the current block. Furthermore, by deriving the correct intra prediction mode, it is possible to reduce the number of bits required for coding the residual of the current block.
- the present invention generates a candidate mode list using intra prediction modes from reference blocks located in neighboring viewpoints in consideration of similarity between viewpoints as well as using spatially adjacent neighboring blocks, thereby improving coding efficiency of intra prediction of the current block. Can be improved.
- FIG. 1 is a schematic block diagram of a video decoder according to an embodiment to which the present invention is applied.
- FIG. 2 is a block diagram illustrating a broadcast receiver to which a video decoder is applied according to an embodiment to which the present invention is applied.
- FIG 3 illustrates a type of intra prediction mode according to an embodiment to which the present invention is applied.
- FIG. 4 is a flowchart illustrating a method of generating a candidate mode list of a current block according to an embodiment to which the present invention is applied.
- FIG. 5 is a flowchart illustrating a method of generating a candidate mode list including an intra prediction mode of a reference block of a neighbor view according to an embodiment to which the present invention is applied.
- FIG. 6 illustrates a method for generating a modified candidate mode list by using an intra prediction mode of a reference block of a neighboring view as an embodiment to which the present invention is applied.
- FIG. 7 is a flowchart illustrating a method of deriving an intra prediction mode of a current block using a candidate mode list according to an embodiment to which the present invention is applied.
- the video signal processing method generates a candidate mode list of a current block based on an intra prediction mode from a spatial neighboring block of the current block, and intra prediction from a neighboring view of the current block.
- the candidate mode list is modified based on a mode
- the intra prediction mode of the current block is derived based on the modified candidate mode list
- the current block is predicted using the derived intra prediction mode.
- the spatial neighboring block according to the present invention includes at least one of a left neighboring block and an upper neighboring block adjacent to the current block.
- the modifying of the candidate mode list according to the present invention may include obtaining a disparity vector of the current block, obtaining a reference block of a neighbor view based on the disparity vector, and whether the reference block is a block coded in an intra mode. If it is determined whether the reference block is a block coded in an intra mode, the candidate mode list is modified using an intra prediction mode of the reference block.
- the disparity vector of the current block according to the present invention is obtained by using the disparity vector of at least one neighboring block among neighboring blocks located at the left, the top, the bottom left, the top right and the top left of the current block. .
- the disparity vector of the current block according to the present invention may be derived from depth data corresponding to the current block.
- the candidate mode list defines a plurality of intra prediction modes available for the current block and list index information corresponding to each of the plurality of intra prediction modes, wherein the list index information is sequentially included in the candidate mode list. It is characterized by specifying the position of each of the arranged plurality of intra prediction modes.
- the modifying of the candidate mode list according to the present invention may correspond to the same intra prediction mode as the intra prediction mode of the neighboring view when the same intra prediction mode exists in the candidate mode list from the neighboring viewpoint.
- the candidate mode list is rearranged by changing list index information.
- the modifying of the candidate mode list according to the present invention may include replacing any one of a plurality of intra prediction modes in the candidate mode list with an intra prediction mode from the neighboring viewpoint.
- the modifying of the candidate mode list according to the present invention may include adding an intra prediction mode from the neighboring viewpoint to the candidate mode list generated based on the intra prediction mode of the spatial neighboring block.
- FIG. 1 is a schematic block diagram of a video decoder according to an embodiment to which the present invention is applied.
- a video decoder includes a NAL parser 100, an entropy decoder 110, an inverse quantization / inverse transform unit 120, an intra prediction unit 130, an in-loop filter unit 140, and a decoded picture.
- the buffer unit 150 and the inter prediction unit 160 may be included.
- the NAL parser 100 may receive a bitstream including multi-view texture data.
- the bitstream including the encoded depth data may be further received.
- the input texture data and the depth data may be transmitted in one bitstream or may be transmitted in separate bitstreams.
- the NAL parser 100 may parse the NAL unit to decode the input bitstream.
- the input bitstream may further include a camera parameter.
- Camera parameters can have intrinsic camera parameters and extrinsic camera parameters, and inherent camera parameters include focal length, aspect ratio, and principal. point) and the like, and the non-unique camera parameter may include location information of the camera in the world coordinate system.
- the entropy decoding unit 110 may extract quantized transform coefficients, coding information for prediction of a texture picture, and the like through entropy decoding.
- the inverse quantization / inverse transform unit 120 may apply a quantization parameter to the quantized transform coefficients to obtain transform coefficients, and inversely transform the transform coefficients to decode texture data or depth data.
- the decoded texture data or depth data may include residual data according to prediction coding.
- the quantization parameter for the depth block may be set in consideration of the complexity of the texture data. For example, when the texture block corresponding to the depth block is a region of high complexity, a low quantization parameter may be set, and in the case of a region of low complexity, a high quantization parameter may be set.
- the complexity of the texture block may be determined based on a difference value between pixels adjacent to each other in the reconstructed texture picture as shown in Equation 1 below.
- Equation 1 E denotes the complexity of the texture data, C denotes the restored texture data, and N denotes the number of pixels in the texture data area to which the complexity is to be calculated.
- the complexity of the texture data corresponds to the difference value between the texture data corresponding to the (x, y) position and the texture data corresponding to the (x-1, y) position and the (x, y) position. It may be calculated using a difference value between the texture data and the texture data corresponding to the position (x + 1, y).
- the complexity may be calculated for the texture picture and the texture block, respectively, and the quantization parameter may be derived using Equation 2 below.
- the quantization parameter for the depth block may be determined based on a ratio of the complexity of the texture picture and the complexity of the texture block.
- ⁇ and ⁇ may be variable integers derived at the decoder, or may be predetermined integers in the decoder.
- the intra predictor 130 may perform intra prediction using the reconstructed texture data in the current texture picture.
- the sample of the current block may be predicted using the intra prediction mode of the current block and the neighboring samples adjacent to the current block.
- the intra prediction mode of the current block may be derived using the candidate mode list, which will be described in detail with reference to FIGS. 4 to 7.
- the intra prediction may be performed on the depth picture in the same manner as the texture picture.
- coding information used for intra prediction of a texture picture may be similarly used in a step picture.
- the coding information used for intra prediction may include intra prediction mode and partition information of intra prediction.
- the in-loop filter unit 140 may apply an in-loop filter to each coded block to reduce block distortion.
- the filter can smooth the edges of the block to improve the quality of the decoded picture.
- Filtered texture pictures or depth pictures may be output or stored in the decoded picture buffer unit 150 for use as a reference picture.
- the coding efficiency may be reduced.
- a separate in-loop filter for depth data may be defined.
- an in-loop filtering method capable of efficiently coding depth data a region-based adaptive loop filter and a trilateral loop filter will be described.
- the region-based adaptive loop filter it may be determined whether to apply the region-based adaptive loop filter based on the variation of the depth block.
- the variation amount of the depth block may be defined as the difference between the maximum pixel value and the minimum pixel value in the depth block.
- Whether to apply the filter may be determined by comparing the change amount of the depth block with a predetermined threshold. For example, when the amount of change in the depth block is greater than or equal to the predetermined threshold value, since the difference between the maximum pixel value and the minimum pixel value in the depth block is large, it may be determined to apply an area-based adaptive loop filter. . In contrast, when the depth change amount is smaller than the predetermined threshold, it may be determined that the region-based adaptive loop filter is not applied.
- the pixel value of the filtered depth block may be derived by applying a predetermined weight to the neighboring pixel value.
- the predetermined weight may be determined based on a position difference between the pixel currently being filtered and the neighboring pixel and / or a difference value between the pixel value currently being filtered and the neighboring pixel value.
- the neighbor pixel value may mean any one of the pixel values included in the depth block except for the pixel value currently being filtered.
- the trilateral loop filter according to the present invention is similar to the region-based adaptive loop filter except that it additionally considers texture data.
- the trilateral loop filter compares the following three conditions and extracts depth data of neighboring pixels satisfying the following three conditions.
- Condition 1 is to compare the positional difference between the current pixel p and the neighboring pixel q in the depth block with a predetermined parameter
- condition 2 is the depth data of the current pixel p and the depth data of the neighboring pixel q. The difference is compared with the predetermined parameter
- Condition 3 is to compare the difference between the texture data of the current pixel p and the texture data of the neighboring pixel q with the predetermined parameter.
- the neighboring pixels satisfying the three conditions may be extracted, and the current pixel p may be filtered by the median or average value of the depth data.
- the decoded picture buffer unit 600 stores or opens a previously coded texture picture or a depth picture in order to perform inter prediction.
- frame_num and POC Picture Order Count
- POC Picture Order Count
- some of the previously coded pictures may have depth pictures that are different from the current depth picture, and thus, view identification information identifying a view point of the depth picture may be used to use these pictures as reference pictures. have.
- the decoded picture buffer unit 150 may manage the reference picture by using an adaptive memory management operation method and a sliding window method in order to more flexibly implement inter prediction. This is to manage the memory of the reference picture and the non-reference picture into one memory and manage them efficiently with less memory.
- the depth pictures may be marked with a separate mark to distinguish them from texture pictures in the decoded picture buffer unit, and information for identifying each depth picture may be used in the marking process.
- the inter prediction unit 160 may perform motion compensation on the current block by using the reference picture and the motion information stored in the decoded picture buffer unit 150.
- the motion information may be understood as a broad concept including a motion vector and reference index information.
- the inter prediction unit 160 may perform temporal inter prediction to perform motion compensation.
- Temporal inter prediction may refer to inter prediction using motion information of a reference picture and a current texture block located at the same time point and different time zone as the current texture block.
- inter-view inter prediction may be further performed as well as temporal inter prediction.
- Inter-view inter prediction may refer to inter prediction using a reference picture located at a different point in time from the current texture block and motion information of the current texture block.
- interview motion vector and interview reference index information motion information used for inter-view prediction will be referred to as interview motion vector and interview reference index information. Therefore, in the present specification, motion information may be flexibly interpreted as a concept including interview motion vector and interview reference index information.
- FIG. 2 is a block diagram illustrating a broadcast receiver to which a video decoder is applied according to an embodiment to which the present invention is applied.
- the broadcast receiver is for receiving an over-the-air broadcast signal and playing back an image.
- the broadcast receiver may generate 3D content using the received depth related information.
- the broadcast receiver includes a tuner 200, a demodulation / channel decoder 202, a transport demultiplexer 204, a packet release unit 206, an audio decoder 208, a video decoder 210, a PSI / PSIP processor ( 214, a 3D renderer 216, a formatter 220, and a display 222.
- the tuner 200 selects and outputs a broadcast signal of one channel selected by a user from among a plurality of broadcast signals input through an antenna (not shown).
- the demodulation / channel decoder 202 demodulates the broadcast signal from the tuner 200 and performs error correction decoding on the demodulated signal to output the transport stream TS.
- the transport demultiplexer 204 demultiplexes the transport stream, separates the video PES and the audio PES, and extracts PSI / PSIP information.
- the packet release unit 206 restores the video ES and the audio ES by releasing packets for the video PES and the audio PES.
- the audio decoder 208 decodes the audio ES and outputs an audio bitstream.
- the audio bitstream is converted into an analog voice signal by a digital-to-analog converter (not shown), amplified by an amplifier (not shown), and then output through a speaker (not shown).
- the video decoder 210 decodes the video ES to reconstruct the original image.
- the decoding process of the audio decoder 208 and the video decoder 210 may be performed based on the packet ID (PID) confirmed by the PSI / PSIP processing unit 214. In the decoding process, the video decoder 210 may extract depth information.
- additional information required to generate an image of a virtual camera viewpoint for example, camera information, or information for estimating occlusion covered by a relatively forward object (eg, geometric information such as an object outline, an object Transparency information, color information), and the like, may be extracted and provided to the 3D renderer 216.
- the depth information and / or additional information may be separated by the transport demultiplexer 204.
- the PSI / PSIP processing unit 214 receives the PSI / PSIP information from the transport demultiplexing unit 204, parses it, and stores it in a memory (not shown) or a register so that the broadcast is reproduced based on the stored information.
- the 3D renderer 216 may generate color information at the virtual camera position, depth information, etc. using the reconstructed image, depth information, additional information, and camera parameters.
- the 3D renderer 216 generates a virtual image at a virtual camera position by performing 3D warping using the reconstructed image and depth information of the reconstructed image.
- the 3D rendering unit 216 is described as being composed of a separate block from the video decoder 210, but this is only an example, and the 3D rendering unit 216 is the video decoder 210. It may be included in the).
- the formatter 220 formats the image reconstructed in the decoding process, that is, the image captured by the actual camera and the virtual image generated by the 3D renderer 216 according to the display method of the corresponding receiver, and displays the display unit 222. ) Will display the 3D image.
- the synthesis of the depth information and the virtual image at the virtual camera position by the 3D rendering unit 216 and the image formatting by the formatter 220 may be selectively performed in response to a user's command. That is, the viewer may operate the remote controller (not shown) so that the synthesized image is not displayed or may specify a time point at which the image synthesis is to be performed.
- the depth information is used by the 3D rendering unit 216 to generate a 3D image, but may be used by the video decoder 210 as another embodiment.
- FIG 3 illustrates a type of intra prediction mode according to an embodiment to which the present invention is applied.
- the intra prediction mode includes a planar mode (intra_Planar), a DC mode (intra_DC), and an angular mode (intra_Angular).
- the planar mode is a mode for predicting samples of the current block by applying weights to neighboring samples located at the top, left, bottom left, and top right of the current block. Specifically, a value according to bi-linear interpolation of neighboring samples located at the top and bottom left and a value according to double linear interpolation of neighboring samples located at the top left and right top are respectively calculated, and the average value between the calculated values is calculated. Can be used as the predicted sample of the current block.
- the DC mode is a mode for predicting a sample of the current block by using an average value of neighboring samples located at the top, left, and top left of the current block.
- the angular mode is a mode for predicting a sample of the current block using neighboring samples according to the angle of the prediction direction. As illustrated in FIG. 3, the angular mode may be defined as a total of 33 intra prediction modes including a vertical mode and a horizontal mode.
- FIG. 4 is a flowchart illustrating a method of generating a candidate mode list of a current block according to an embodiment to which the present invention is applied.
- the candidate mode list defines a plurality of intra prediction modes available for the current block and list index information corresponding to each intra prediction mode.
- the plurality of intra prediction modes may be arranged in an order in which the current block is likely to be used.
- the list index information may specify the positions of the intra prediction modes sequentially arranged in the candidate mode list.
- an intra prediction mode of a neighboring block may be obtained (S400).
- the neighboring block may include a left neighboring block and an upper neighboring block adjacent to the current block.
- the intra prediction mode of the neighboring block may be obtained by assigning a predefined intra prediction mode (eg, the DC mode) to the neighboring block.
- step S410 when both intra prediction modes are the same, it may be determined whether the intra prediction mode of the left neighboring block is the planar mode or the DC mode (S420).
- the candidate mode list may include the planar mode, the DC mode, and the vertical mode (S430).
- the planar mode may be allocated to a position where list index information is 0, the DC mode to a position where list index information is 1, and the vertical mode may be allocated to a position where list index information is 2 in the candidate mode list.
- the candidate mode list may include the intra prediction mode of the left neighboring block and the intra prediction mode derived therefrom (S440). ).
- the intra prediction mode of the left neighboring block is allocated to the position where the list index information is zero. Then, the intra prediction mode corresponding to the value obtained by subtracting 1 from the value of the intra prediction mode of the left neighboring block is assigned to the position where the list index information is 1, and the value obtained by adding 1 to the value of the intra prediction mode of the left neighboring block is then added.
- the corresponding intra prediction mode may be assigned to the position at which the list index information is two. For example, when the value of the intra prediction mode of the left neighboring block is 10, the intra prediction modes having the values of the intra prediction modes 9 and 11 may be included in the candidate mode list.
- the candidate mode list may include an intra prediction mode of a left neighboring block and an intra prediction mode of an upper neighboring block (S450).
- the intra prediction mode of the left neighboring block may be allocated to the position where the list index information is 0, and the intra prediction mode of the upper neighboring block may be allocated to the position where the list index information is 1.
- the intra prediction mode defined in consideration of the intra prediction mode of the left / top neighboring block may be allocated to the position at which the list index information is 2.
- a mode that is not the same as the intra prediction mode from the left / top neighboring block from the planar mode, the DC mode, and the vertical mode may be selected and assigned to the position where the list index information is two.
- it may be determined whether or not it is the same as the intra prediction mode from the left / top neighboring block in consideration of the priority between the planar mode, the DC mode, and the vertical mode. This checking process may be performed until a mode is found that is not the same as the intra prediction mode from the left / top neighboring block.
- an intra prediction mode of a neighboring neighboring block not only an intra prediction mode of a neighboring neighboring block but also an intra prediction mode of a reference block of a neighboring view may be added to the candidate mode list, which will be described below with reference to FIGS. 5 and 6.
- FIG. 5 is a flowchart illustrating a method of generating a candidate mode list including an intra prediction mode of a reference block of a neighbor view according to an embodiment to which the present invention is applied.
- a disparity vector of a current block at a current time may be obtained (S500).
- the disparity vector may mean a motion vector specifying a reference block corresponding to the current block in the view direction.
- the disparity vector of the current block may be derived from the disparity vector of the neighboring block.
- the neighboring block may include at least one of neighboring blocks located at the left, the top, the bottom left, the top right and the top left of the current block.
- the disparity vector of the neighboring block may be set as the disparity vector of the current block.
- the disparity vector of the current block may be derived from depth data corresponding to the current block. A detailed derivation method will be described based on Equations 3 and 4.
- Z means a distance from the camera of the pixel
- D is a value quantized Z, corresponding to the depth data.
- Znear and Zfar mean the minimum and maximum values of Z defined for the current time point, respectively.
- Znear and Zfar may be extracted from the bitstream through a sequence parameter set, a slice header, and the like, and may be information predefined in the decoder. Therefore, when the distance Z from the camera of the pixel is quantized to 256 levels, Z can be restored using depth data, Znear, and Zfar as shown in Equation (3). Then, the reconstructed Z may be used to derive the disparity vector for the current block as shown in Equation 4.
- Equation 4 d denotes a disparity vector, f denotes a focal length of a camera, and B denotes a distance between cameras. f and B may be assumed to be the same for all cameras, and thus may be information predefined in the decoder.
- a reference block of a neighbor view may be obtained based on the disparity vector of the current block (S510).
- the neighbor view may mean a view other than the current view.
- the neighbor view may mean a base view that is coded independently of another view without performing inter-view prediction.
- the reference block may mean a block specified by the disparity vector based on the position of the current block.
- the modified candidate mode list may be generated using the intra prediction mode of the reference block (S530). A detailed generation method will be described below with reference to FIG. 6.
- FIG. 6 illustrates a method for generating a modified candidate mode list by using an intra prediction mode of a reference block of a neighboring view as an embodiment to which the present invention is applied.
- the candidate mode list may be rearranged by swapping positions between the same intra prediction mode as the intra prediction mode from the neighboring viewpoint and the intra prediction mode of the position where the list index information is zero.
- the same intra prediction mode as the intra prediction mode from the neighboring view already exists at the position where the list index information is 0, the rearrangement process will be skipped.
- the list index information is The positions between the intra prediction mode (MPM2) of the position 2 and the intra prediction mode (MPM0) of the position where the list index information is 0 may be interchanged.
- any one of the plurality of intra prediction modes MPM0, MPM1, and MPM2 in the candidate mode list may be replaced with an intra prediction mode MPMview from a neighboring viewpoint.
- the intra prediction mode MPM2 at the position where the list index information is 2 among the plurality of intra prediction modes may be replaced with the intra prediction mode MPMview from the neighboring viewpoint. This is because the intra prediction mode MPM2 at the position where the list index information is 2 has the lowest selectivity among the intra prediction modes in the candidate mode list.
- the intra prediction mode from the neighboring viewpoint may be added to the candidate mode list derived from the intra prediction mode of the spatial neighboring block.
- the intra prediction mode from the neighboring viewpoint may be located at the bottom of the candidate mode list. That is, the list index information may be added at the position 3. As such, even if the intra prediction mode is added at the position where the list index information is 3 in the candidate mode list, the MPM0 to the MPM3 can still be coded in 2 bits.
- the intra prediction mode from the neighbor view may be located at the top of the candidate mode list. That is, the list index information may be added at a position of zero.
- the intra prediction mode at the position where the list index information is 0 is the position where the list index information is 1
- the intra prediction mode at the position where the list index information is 1 is the position where the list index information is 2
- the intra prediction modes of the in positions may be rearranged to the positions where the list index information is three.
- the candidate mode is positioned at the top of the candidate mode list, otherwise the candidate mode is present. It can also be set at the bottom of the list.
- a base mode use flag may be defined.
- the base mode usage flag may mean information indicating whether to use the intra prediction mode from the neighboring viewpoint in generating the candidate mode list for the current block. Accordingly, the intra prediction mode may be obtained from the reference block of the neighbor view only when the base mode use flag indicates to use the intra prediction mode from the neighbor view.
- the intra prediction mode from the neighboring view is the same as the intra prediction mode in the previously generated candidate mode list, adding the above-described rearrangement method of the candidate mode list, alternative method of intra prediction mode, and intra prediction mode. Any of the schemes can be specified. If the same intra prediction mode as the intra prediction mode from the neighboring viewpoint is included in the previously generated candidate mode list, the replacement of the intra prediction mode or the addition of the intra prediction mode will not be profitable. Therefore, when the intra prediction mode identical to the intra prediction mode from the neighboring viewpoint is included in the previously generated candidate mode list, the rearrangement method of the candidate mode list is used. Otherwise, the alternative prediction method or intra prediction mode of the intra prediction mode is used. It can be set by using an additional method of prediction mode.
- FIG. 7 is a flowchart illustrating a method of deriving an intra prediction mode of a current block using a candidate mode list according to an embodiment to which the present invention is applied.
- the candidate mode list may not only mean a candidate mode list derived using spatial neighboring blocks, but may also be understood as a concept including all candidate mode lists modified using intra prediction modes from neighboring viewpoints.
- a candidate mode flag may be obtained from a bitstream (S700).
- the candidate mode flag may mean information indicating whether the current block uses the intra prediction mode from the candidate mode list.
- intra mode index information may be obtained from the bitstream (S710).
- the intra mode index information may mean information for specifying an intra prediction mode used for intra prediction of the current block in the candidate mode list.
- an intra prediction mode corresponding to the intra mode index information may be obtained from the candidate mode list (S720).
- the intra prediction mode may be extracted from the bitstream (S730).
- the intra prediction mode extracted from the bitstream may specify any one of the intra prediction modes except the intra prediction mode included in the candidate mode list among the 35 intra prediction modes. Accordingly, the intra prediction mode of the current block may be derived based on the extracted intra prediction mode (S740).
- each component or feature is to be considered optional unless stated otherwise.
- Each component or feature may be embodied in a form that is not combined with other components or features. It is also possible to combine some of the components and / or features to form an embodiment of the invention.
- the order of the operations described in the embodiments of the present invention may be changed. Some components or features of one embodiment may be included in another embodiment or may be replaced with corresponding components or features of another embodiment.
- the present invention can be used to code a video signal.
Abstract
Description
Claims (9)
- .현재 블록의 공간적 이웃 블록으로부터의 인트라 예측 모드에 기초하여 현재 블록의 후보자 모드 리스트를 생성하는 단계;현재 블록의 이웃 시점으로부터의 인트라 예측 모드에 기초하여 상기 후보자 모드 리스트를 수정하는 단계;상기 수정된 후보자 모드 리스트에 기초하여 현재 블록의 인트라 예측 모드를 유도하는 단계; 및상기 유도된 인트라 예측 모드를 이용하여 현재 블록을 예측하는 단계를 포함하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 1항에 있어서, 상기 공간적 이웃 블록은 상기 현재 블록에 인접한 좌측 이웃 블록 및 상단 이웃 블록 중 적어도 하나를 포함하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 1항에 있어서, 상기 후보자 모드 리스트를 수정하는 단계는,상기 현재 블록의 변이 벡터를 획득하는 단계;상기 변이 벡터에 기초하여 이웃 시점의 참조 블록을 획득하는 단계;상기 참조 블록이 인트라 모드로 코딩된 블록인지 여부를 확인하는 단계; 및상기 참조 블록이 인트라 모드로 코딩된 블록인 경우, 상기 참조 블록의 인트라 예측 모드를 이용하여 상기 후보자 모드 리스트를 수정하는 단계를 포함하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 3항에 있어서, 상기 현재 블록의 변이 벡터는 상기 현재 블록의 좌측, 상단, 좌측하단, 우측상단 및 좌측상단에 위치한 이웃 블록들 중 적어도 하나의 이웃 블록의 변이 벡터를 이용하여 획득하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 3항에 있어서, 상기 현재 블록의 변이 벡터는 상기 현재 블록에 대응하는 뎁스 데이터로부터 유도되는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 1항에 있어서, 상기 후보자 모드 리스트는 상기 현재 블록이 이용 가능한 복수 개의 인트라 예측 모드 및 상기 복수 개의 인트라 예측 모드 각각에 대응하는 리스트 인덱스 정보를 정의하되,상기 리스트 인덱스 정보는 상기 후보자 모드 리스트 내에 순차적으로 배열된 상기 복수 개의 인트라 예측 모드 각각의 위치를 특정하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 6항에 있어서, 상기 후보자 모드 리스트를 수정하는 단계는,상기 후보자 모드 리스트 내에 상기 이웃 시점으로부터의 인트라 예측 모드와 동일한 인트라 예측 모드가 존재할 경우, 상기 이웃 시점의 인트라 예측 모드와 동일한 인트라 예측 모드에 대응하는 리스트 인덱스 정보를 변경하여 상기 후보자 모드 리스트를 재배열하는 단계를 포함하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 6항에 있어서, 상기 후보자 모드 리스트를 수정하는 단계는,상기 후보자 모드 리스트 내의 복수 개의 인트라 예측 모드 중 어느 하나를 상기 이웃 시점으로부터의 인트라 예측 모드로 대체하는 단계를 포함하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제 6항에 있어서, 상기 후보자 모드 리스트를 수정하는 단계는,상기 공간적 이웃 블록의 인트라 예측 모드에 기초하여 생성된 후보자 모드 리스트에 상기 이웃 시점으로부터의 인트라 예측 모드를 추가하는 단계를 포함하는 것을 특징으로 하는 비디오 신호 처리 방법.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020157021095A KR20160048025A (ko) | 2013-01-07 | 2014-01-07 | 비디오 신호 처리 방법 및 장치 |
CN201480004166.9A CN104904209B (zh) | 2013-01-07 | 2014-01-07 | 视频信号处理方法 |
EP14735407.0A EP2942955A4 (en) | 2013-01-07 | 2014-01-07 | Video signal processing method and apparatus |
JP2015551613A JP6306050B2 (ja) | 2013-01-07 | 2014-01-07 | ビデオ信号処理方法および装置 |
US14/759,580 US9826239B2 (en) | 2013-01-07 | 2014-01-07 | Video signal processing method and device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361749395P | 2013-01-07 | 2013-01-07 | |
US61/749,395 | 2013-01-07 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014107083A1 true WO2014107083A1 (ko) | 2014-07-10 |
Family
ID=51062354
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2014/000135 WO2014107083A1 (ko) | 2013-01-07 | 2014-01-07 | 비디오 신호 처리 방법 및 장치 |
Country Status (6)
Country | Link |
---|---|
US (1) | US9826239B2 (ko) |
EP (1) | EP2942955A4 (ko) |
JP (1) | JP6306050B2 (ko) |
KR (1) | KR20160048025A (ko) |
CN (1) | CN104904209B (ko) |
WO (1) | WO2014107083A1 (ko) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016195460A1 (ko) * | 2015-06-05 | 2016-12-08 | 한양대학교 산학협력단 | 화면 내 예측에 대한 부호화/복호화 방법 및 장치 |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103891283B (zh) | 2011-12-05 | 2018-03-16 | Lg电子株式会社 | 帧内预测的方法和设备 |
CN105556970B (zh) * | 2013-07-12 | 2019-01-11 | 三星电子株式会社 | 使用修正矢量推导的视频编码方法及其设备、视频解码方法及其设备 |
WO2016137166A1 (ko) * | 2015-02-23 | 2016-09-01 | 엘지전자(주) | 인트라 예측 모드 기반 영상 처리 방법 및 이를 위한 장치 |
KR20170058837A (ko) * | 2015-11-19 | 2017-05-29 | 한국전자통신연구원 | 화면내 예측모드 부호화/복호화 방법 및 장치 |
KR20170058871A (ko) * | 2015-11-19 | 2017-05-29 | 한국전자통신연구원 | 화면내 예측 모드 부호화/복호화 방법 및 장치 |
KR20190052128A (ko) * | 2016-10-04 | 2019-05-15 | 김기백 | 영상 데이터 부호화/복호화 방법 및 장치 |
KR20180041833A (ko) * | 2016-10-17 | 2018-04-25 | 디지털인사이트 주식회사 | 화면 내 예측에 대한 참조 모드 리스트를 사용하는 비디오 코딩 방법 및 장치 |
CN109120926B (zh) | 2017-06-23 | 2019-08-13 | 腾讯科技(深圳)有限公司 | 预测模式选择方法、装置及介质 |
US10708626B2 (en) * | 2018-11-19 | 2020-07-07 | Google Llc | Iterative IDCT with adaptive non-linear filtering |
US20220239906A1 (en) * | 2021-01-26 | 2022-07-28 | Beijing Dajia Internet Information Technology Co., Ltd. | System and method for applying adaptive loop filter in video coding |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012090491A1 (ja) * | 2010-12-28 | 2012-07-05 | パナソニック株式会社 | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、および、画像符号化復号装置 |
WO2012164908A1 (ja) * | 2011-05-31 | 2012-12-06 | パナソニック株式会社 | 動画像符号化方法、動画像符号化装置、動画像復号化方法、動画像復号化装置、および動画像符号化復号化装置 |
WO2012164886A1 (ja) * | 2011-05-27 | 2012-12-06 | パナソニック株式会社 | 動画像符号化方法、動画像符号化装置、動画像復号化方法、動画像復号化装置、および動画像符号化復号化装置 |
WO2012171442A1 (en) * | 2011-06-15 | 2012-12-20 | Mediatek Inc. | Method and apparatus of motion and disparity vector prediction and compensation for 3d video coding |
JP2012253460A (ja) * | 2011-05-31 | 2012-12-20 | Jvc Kenwood Corp | 動画像復号装置、動画像復号方法及び動画像復号プログラム |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8804839B2 (en) * | 2007-06-27 | 2014-08-12 | Korea Electronics Technology Institute | Method for image prediction of multi-view video codec and computer-readable recording medium thereof |
US9654785B2 (en) * | 2011-06-09 | 2017-05-16 | Qualcomm Incorporated | Enhanced intra-prediction mode signaling for video coding using neighboring mode |
US9154796B2 (en) * | 2011-11-04 | 2015-10-06 | Qualcomm Incorporated | Intra-mode video coding |
-
2014
- 2014-01-07 KR KR1020157021095A patent/KR20160048025A/ko not_active Application Discontinuation
- 2014-01-07 JP JP2015551613A patent/JP6306050B2/ja active Active
- 2014-01-07 CN CN201480004166.9A patent/CN104904209B/zh active Active
- 2014-01-07 EP EP14735407.0A patent/EP2942955A4/en not_active Withdrawn
- 2014-01-07 WO PCT/KR2014/000135 patent/WO2014107083A1/ko active Application Filing
- 2014-01-07 US US14/759,580 patent/US9826239B2/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012090491A1 (ja) * | 2010-12-28 | 2012-07-05 | パナソニック株式会社 | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、および、画像符号化復号装置 |
WO2012164886A1 (ja) * | 2011-05-27 | 2012-12-06 | パナソニック株式会社 | 動画像符号化方法、動画像符号化装置、動画像復号化方法、動画像復号化装置、および動画像符号化復号化装置 |
WO2012164908A1 (ja) * | 2011-05-31 | 2012-12-06 | パナソニック株式会社 | 動画像符号化方法、動画像符号化装置、動画像復号化方法、動画像復号化装置、および動画像符号化復号化装置 |
JP2012253460A (ja) * | 2011-05-31 | 2012-12-20 | Jvc Kenwood Corp | 動画像復号装置、動画像復号方法及び動画像復号プログラム |
WO2012171442A1 (en) * | 2011-06-15 | 2012-12-20 | Mediatek Inc. | Method and apparatus of motion and disparity vector prediction and compensation for 3d video coding |
Non-Patent Citations (1)
Title |
---|
See also references of EP2942955A4 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016195460A1 (ko) * | 2015-06-05 | 2016-12-08 | 한양대학교 산학협력단 | 화면 내 예측에 대한 부호화/복호화 방법 및 장치 |
US11368675B2 (en) | 2015-06-05 | 2022-06-21 | Dolby Laboratories Licensing Corporation | Method and device for encoding and decoding intra-frame prediction |
Also Published As
Publication number | Publication date |
---|---|
EP2942955A1 (en) | 2015-11-11 |
JP2016506689A (ja) | 2016-03-03 |
JP6306050B2 (ja) | 2018-04-04 |
KR20160048025A (ko) | 2016-05-03 |
CN104904209B (zh) | 2018-07-24 |
US20150350657A1 (en) | 2015-12-03 |
EP2942955A4 (en) | 2016-07-20 |
US9826239B2 (en) | 2017-11-21 |
CN104904209A (zh) | 2015-09-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014107083A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2010087589A2 (ko) | 경계 인트라 코딩을 이용한 비디오 신호 처리 방법 및 장치 | |
WO2015142054A1 (ko) | 다시점 비디오 신호 처리 방법 및 장치 | |
WO2010068020A9 (ko) | 다시점 영상 부호화, 복호화 방법 및 그 장치 | |
US20110222602A1 (en) | Method and device for processing a video signal using inter-view prediction | |
WO2013169031A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2013162273A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2010123198A2 (ko) | 다시점 비디오 신호 처리 방법 및 장치 | |
WO2014010935A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2014168443A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2013191436A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2016056822A1 (ko) | 3d 비디오 코딩 방법 및 장치 | |
WO2013176485A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2014107029A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2013133627A1 (ko) | 비디오 신호 처리 방법 | |
WO2014010918A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2015009098A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2013133587A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2016003209A1 (ko) | 다시점 비디오 신호 처리 방법 및 장치 | |
WO2014109547A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2014054897A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2014058207A1 (ko) | 다시점 비디오 신호의 인코딩 방법, 디코딩 방법 및 이에 대한 장치 | |
WO2015009091A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2014054896A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2014073873A1 (ko) | 비디오 신호 처리 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14735407 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015551613 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014735407 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14759580 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 20157021095 Country of ref document: KR Kind code of ref document: A |