WO2015142054A1 - 다시점 비디오 신호 처리 방법 및 장치 - Google Patents
다시점 비디오 신호 처리 방법 및 장치 Download PDFInfo
- Publication number
- WO2015142054A1 WO2015142054A1 PCT/KR2015/002629 KR2015002629W WO2015142054A1 WO 2015142054 A1 WO2015142054 A1 WO 2015142054A1 KR 2015002629 W KR2015002629 W KR 2015002629W WO 2015142054 A1 WO2015142054 A1 WO 2015142054A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- current block
- block
- merge
- candidate
- merge candidate
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/137—Motion inside a coding unit, e.g. average field, frame or block difference
- H04N19/139—Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0085—Motion estimation from stereoscopic image signals
Definitions
- the present invention relates to a method and apparatus for coding a video signal.
- High efficiency image compression techniques can be used to solve these problems caused by high resolution and high quality image data.
- An inter-screen prediction technique for predicting pixel values included in the current picture from a picture before or after the current picture using an image compression technique an intra prediction technique for predicting pixel values included in a current picture using pixel information in the current picture
- An object of the present invention is to provide a method and apparatus for performing inter-view prediction using a disparity vector in encoding / decoding a multiview video signal.
- An object of the present invention is to provide a method and apparatus for deriving a disparity vector of a texture block using depth data of a depth block in encoding / decoding a multiview video signal.
- An object of the present invention is to provide a method and apparatus for deriving a disparity vector from a neighboring block of a current block in encoding / decoding a multiview video signal.
- An object of the present invention is to provide a method and apparatus for deriving an interview merge candidate using a disparity vector in encoding / decoding a multiview video signal.
- An object of the present invention is to provide a method and apparatus for constructing a merge candidate list for a merge mode in encoding / decoding a multiview video signal.
- An object of the present invention is to provide a method and apparatus for selectively using interview motion candidates in consideration of lighting compensation in encoding / decoding a multiview video signal.
- An object of the present invention is to provide a method and apparatus for determining an arrangement order of merge candidates in a merge candidate list in consideration of lighting compensation in encoding / decoding a multiview video signal.
- a multi-view video signal decoding method and apparatus generate a merge candidate list for a current block, derive a motion vector of the current block based on a merge index for the current block obtained from a bitstream, The prediction block of the current block is obtained using the derived motion vector, and the obtained block is reconstructed by adding the obtained prediction value and a residual value of the current block.
- the merge candidate list is composed of at least one merge candidate, and the merge candidate includes at least one of a spatial neighboring block, a temporal neighboring block, or an interview motion candidate (IvMC). It is characterized by including.
- the merge index is characterized in that it specifies a merge candidate used for decoding the current block in the merge mode.
- the interview motion candidate has a temporal motion vector of a reference block specified by the disparity vector of the current block, wherein the reference block is a symbol of the current block. It is characterized by belonging to a reference view.
- the interview motion candidate is limited to the merge candidate list in consideration of whether illumination compensation is performed on the current block. It is characterized by.
- the interview motion candidate (IvMC) when no illumination compensation is performed on the current block, the interview motion candidate (IvMC) is included in the merge candidate list, and the interview motion candidate ( IvMC), the spatial neighboring blocks and the temporal neighboring blocks are arranged in the merge candidate list in order of priority.
- a multi-view video signal encoding method and apparatus generates a merge candidate list for a current block, derives a motion vector of the current block based on a merge index for the current block, and derives the derived motion vector. Obtain a prediction value of the current block by using, and reconstruct the current block by adding the obtained prediction value and the residual value for the current block.
- a merge candidate list is composed of at least one merge candidate, and the merge candidate includes at least one of a spatial neighboring block, a temporal neighboring block, or an interview motion candidate (IvMC). It is characterized by including.
- the merge index is characterized in that it specifies a merge candidate used to encode the current block in merge mode.
- the interview motion candidate has a temporal motion vector of a reference block specified by the disparity vector of the current block, the reference block of the current block It is characterized by belonging to a reference view.
- the interview motion candidate is limited to the merge candidate list in consideration of whether illumination compensation is performed on the current block. It is characterized by.
- the interview motion candidate (IvMC) when no illumination compensation is performed on the current block, the interview motion candidate (IvMC) is included in the merge candidate list, and the interview motion candidate ( IvMC), the spatial neighboring blocks and the temporal neighboring blocks are arranged in the merge candidate list in order of priority.
- inter-view prediction can be efficiently performed using the disparity vector.
- the variation vector of the current block can be effectively derived from the depth data of the current depth block or the variation vector of the neighboring texture block.
- encoding / decoding performance can be improved by selectively using interview motion candidates in consideration of lighting compensation.
- encoding / decoding performance can be improved by setting priorities of a plurality of merge candidates constituting the merge candidate list.
- FIG. 1 is a schematic block diagram of a video decoder according to an embodiment to which the present invention is applied.
- FIG. 2 illustrates a method of decoding a current block in a merge mode as an embodiment to which the present invention is applied.
- FIG 3 illustrates a method of deriving a motion vector of an interview motion candidate based on an inter-view motion prediction technique as an embodiment to which the present invention is applied.
- VSP candidate view synthesis prediction candidate
- FIG. 5 is a diagram illustrating a method of deriving a disparity vector of a current block using depth data of a depth image as an embodiment to which the present invention is applied.
- FIG. 6 illustrates candidates of spatial / temporal neighboring blocks of the current block as an embodiment to which the present invention is applied.
- FIG. 7 illustrates a method for adaptively using an interview motion candidate IvMC based on an illumination compensation flag ic_flag as an embodiment to which the present invention is applied.
- FIG. 8 illustrates priority of merge candidates based on an illumination compensation flag ic_flag according to an embodiment to which the present invention is applied.
- a multi-view video signal decoding method and apparatus generate a merge candidate list for a current block, derive a motion vector of the current block based on a merge index for the current block obtained from a bitstream, The prediction block of the current block is obtained using the derived motion vector, and the obtained block is reconstructed by adding the obtained prediction value and a residual value of the current block.
- the merge candidate list is composed of at least one merge candidate, and the merge candidate includes at least one of a spatial neighboring block, a temporal neighboring block, or an interview motion candidate (IvMC). It is characterized by including.
- the merge index is characterized in that it specifies a merge candidate used for decoding the current block in the merge mode.
- the interview motion candidate has a temporal motion vector of a reference block specified by the disparity vector of the current block, wherein the reference block is a symbol of the current block. It is characterized by belonging to a reference view.
- the interview motion candidate is limited to the merge candidate list in consideration of whether illumination compensation is performed on the current block. It is characterized by.
- the interview motion candidate (IvMC) when no illumination compensation is performed on the current block, the interview motion candidate (IvMC) is included in the merge candidate list, and the interview motion candidate ( IvMC), the spatial neighboring blocks and the temporal neighboring blocks are arranged in the merge candidate list in order of priority.
- a multi-view video signal encoding method and apparatus generates a merge candidate list for a current block, derives a motion vector of the current block based on a merge index for the current block, and derives the derived motion vector. Obtain a prediction value of the current block by using, and reconstruct the current block by adding the obtained prediction value and the residual value for the current block.
- a merge candidate list is composed of at least one merge candidate, and the merge candidate includes at least one of a spatial neighboring block, a temporal neighboring block, or an interview motion candidate (IvMC). It is characterized by including.
- the merge index is characterized in that it specifies a merge candidate used to encode the current block in merge mode.
- the interview motion candidate has a temporal motion vector of a reference block specified by the disparity vector of the current block, the reference block of the current block It is characterized by belonging to a reference view.
- the interview motion candidate is limited to the merge candidate list in consideration of whether illumination compensation is performed on the current block. It is characterized by.
- the interview motion candidate (IvMC) when no illumination compensation is performed on the current block, the interview motion candidate (IvMC) is included in the merge candidate list, and the interview motion candidate ( IvMC), the spatial neighboring blocks and the temporal neighboring blocks are arranged in the merge candidate list in order of priority.
- Techniques for compression encoding or decoding multi-view video signal data take into account spatial redundancy, temporal redundancy, and redundancy existing between views.
- a multiview texture image photographed from two or more viewpoints may be coded to implement a 3D image.
- depth data corresponding to a multiview texture image may be further coded as necessary.
- compression coding may be performed in consideration of spatial redundancy, temporal redundancy, or inter-view redundancy.
- Depth data represents distance information between a camera and a corresponding pixel
- depth data may be flexibly interpreted as information related to depth, such as a depth value, a depth information, a depth image, a depth picture, a depth sequence, and a depth bitstream.
- coding in this specification may include both the concepts of encoding and decoding, and may be flexibly interpreted according to the technical spirit and technical scope of the present invention.
- FIG. 1 is a schematic block diagram of a video decoder according to an embodiment to which the present invention is applied.
- a video decoder includes a NAL parser 100, an entropy decoder 200, an inverse quantization / inverse transform unit 300, an intra predictor 400, an in-loop filter unit 500, and a decoded picture.
- the buffer unit 600 and the inter prediction unit 700 may be included.
- the NAL parser 100 may receive a bitstream including multi-view texture data.
- the bitstream including the encoded depth data may be further received.
- the input texture data and the depth data may be transmitted in one bitstream or may be transmitted in separate bitstreams.
- the NAL parser 100 may parse the NAL unit to decode the input bitstream.
- the input bitstream is multi-view related data (eg, 3-Dimensional Video)
- the input bitstream may further include a camera parameter.
- Camera parameters can have intrinsic camera parameters and extrinsic camera parameters, and inherent camera parameters include focal length, aspect ratio, and principal. point) and the like, and the non-unique camera parameter may include location information of the camera in the world coordinate system.
- the entropy decoding unit 200 may extract quantized transform coefficients, coding information for prediction of a texture picture, and the like through entropy decoding.
- the inverse quantization / inverse transform unit 300 may apply a quantization parameter to the quantized transform coefficients to obtain transform coefficients, and inversely transform the transform coefficients to decode texture data or depth data.
- the decoded texture data or depth data may mean residual data according to a prediction process.
- the quantization parameter for the depth block may be set in consideration of the complexity of the texture data. For example, when the texture block corresponding to the depth block is a region of high complexity, a low quantization parameter may be set, and in the case of a region of low complexity, a high quantization parameter may be set.
- the complexity of the texture block may be determined based on a difference value between pixels adjacent to each other in the reconstructed texture picture as shown in Equation 1 below.
- Equation 1 E denotes the complexity of the texture data, C denotes the restored texture data, and N denotes the number of pixels in the texture data area to which the complexity is to be calculated.
- the complexity of the texture data corresponds to the difference value between the texture data corresponding to the (x, y) position and the texture data corresponding to the (x-1, y) position and the (x, y) position. It may be calculated using a difference value between the texture data and the texture data corresponding to the position (x + 1, y).
- the complexity may be calculated for the texture picture and the texture block, respectively, and the quantization parameter may be derived using Equation 2 below.
- the quantization parameter for the depth block may be determined based on a ratio of the complexity of the texture picture and the complexity of the texture block.
- ⁇ and ⁇ may be variable integers derived at the decoder, or may be predetermined integers in the decoder.
- the intra predictor 400 may perform intra prediction using the reconstructed texture data in the current texture picture. Intra-prediction may be performed on the depth picture in the same manner as the texture picture.
- coding information used for intra prediction of a texture picture may be similarly used in a step picture.
- the coding information used for intra prediction may include intra prediction mode and partition information of intra prediction.
- the in-loop filter unit 500 may apply an in-loop filter to each coded block to reduce block distortion.
- the filter can smooth the edges of the block to improve the quality of the decoded picture.
- Filtered texture pictures or depth pictures may be output or stored in the decoded picture buffer unit 600 for use as a reference picture.
- the coding efficiency may be reduced.
- a separate in-loop filter for depth data may be defined.
- an in-loop filtering method for efficiently coding depth data a region-based adaptive loop filter and a trilateral loop filter will be described.
- the region-based adaptive loop filter it may be determined whether to apply the region-based adaptive loop filter based on the variation of the depth block.
- the variation amount of the depth block may be defined as the difference between the maximum pixel value and the minimum pixel value in the depth block.
- Whether to apply the filter may be determined by comparing the change amount of the depth block with a predetermined threshold. For example, when the amount of change in the depth block is greater than or equal to the predetermined threshold value, since the difference between the maximum pixel value and the minimum pixel value in the depth block is large, it may be determined to apply an area-based adaptive loop filter. . In contrast, when the depth change amount is smaller than the predetermined threshold, it may be determined that the region-based adaptive loop filter is not applied.
- the pixel value of the filtered depth block may be derived by applying a predetermined weight to the neighboring pixel value.
- the predetermined weight may be determined based on a position difference between the pixel currently being filtered and the neighboring pixel and / or a difference value between the pixel value currently being filtered and the neighboring pixel value.
- the neighbor pixel value may mean any one of the pixel values included in the depth block except for the pixel value currently being filtered.
- the trilateral loop filter according to the present invention is similar to the region-based adaptive loop filter except that it additionally considers texture data.
- the trilateral loop filter compares the following three conditions and extracts depth data of neighboring pixels satisfying the following three conditions.
- Condition 1 is to compare the positional difference between the current pixel p and the neighboring pixel q in the depth block with a predetermined parameter sigma 1
- condition 2 is the depth data of the current pixel p and the depth of the neighboring pixel q.
- the difference between the data is compared with the predetermined parameter? 2
- condition 3 is comparing the difference between the texture data of the current pixel p and the texture data of the neighboring pixel q with the predetermined parameter? 3.
- the neighboring pixels satisfying the three conditions may be extracted, and the current pixel p may be filtered by the median or average value of the depth data.
- the decoded picture buffer unit 600 stores or opens a previously coded texture picture or a depth picture in order to perform inter prediction.
- the frame_num and the POC (Picture Order Count) of each picture may be used.
- some of the previously coded pictures may have depth pictures that are different from the current depth picture, and thus, view identification information identifying a view point of the depth picture may be used to use these pictures as reference pictures. have.
- the decoded picture buffer unit 600 may manage the reference picture using an adaptive memory management control method and a sliding window method in order to more flexibly implement inter prediction.
- the depth pictures may be marked with a separate mark to distinguish them from texture pictures in the decoded picture buffer unit, and information for identifying each depth picture may be used in the marking process.
- the inter prediction unit 700 may perform motion compensation of the current block by using the reference picture and the motion information stored in the decoded picture buffer unit 600.
- the motion information may be understood as a broad concept including a motion vector and reference index information.
- the inter prediction unit 700 may perform temporal inter prediction to perform motion compensation.
- Temporal inter prediction may refer to inter prediction using motion information of a reference picture and a current texture block located at the same time point and different time zone as the current texture block.
- temporal inter prediction may refer to inter prediction using motion information of a reference picture and a current texture block located at the same time point and different time zone as the current texture block.
- the motion information used for the inter-view prediction may include a disparity vector or an inter-view motion vector. A method of performing inter-view prediction using the disparity vector will be described below with reference to FIG. 2.
- FIG. 2 illustrates a method of decoding a current block in a merge mode as an embodiment to which the present invention is applied.
- a merge candidate list for the current block may be generated (S200).
- the merge candidate list of the present invention may include at least one merge candidate available for decoding the current block in merge mode.
- the spatial / temporal neighboring block of the current block may be used as an example of the merge candidate.
- the spatial neighboring block may include at least one of a left neighboring block, an upper neighboring block, an upper right neighboring block, a lower left neighboring block, and an upper left neighboring block of the current block.
- a temporal neighboring block is a block included in a call picture having a temporal order different from that of the current block, and can be defined as a block having the same position as the current block.
- a merge candidate (hereinafter, referred to as an interview merge candidate) based on the correlation between the viewpoints or the correlation between the texture and the depth may be included in the merge candidate list.
- the interview merge candidates include texture merge candidates, inter-view motion candidates, inter-view disparity candidates, and view synthesis prediction candidates (VSP candidates). The method of deriving the motion vector of each interview merge candidate and configuring the merge candidate in the merge candidate list will be described later with reference to FIGS. 3 to 6.
- the motion vector of the current block may be derived based on the merge candidate list and the merge index merge_idx generated in operation S200 (S210).
- a merge candidate corresponding to the merge index of the current block may be selected from the merge candidate list.
- the merge index may be extracted from the bitstream as a syntax for specifying any one of a plurality of merge candidates included in the merge candidate list. That is, the merge index is information for specifying a merge candidate used to derive the motion vector of the current block.
- the motion vector assigned to the selected merge candidate may be set as the motion vector of the current block.
- a predicted value of the current block may be obtained using the motion vector derived in step S210 (S200).
- the current block may perform temporal inter prediction using the motion vector.
- the reference picture of the current block belongs to a point of view different from the current block, the current block may perform inter-view inter prediction using the motion vector.
- Whether the reference picture of the current block belongs to the same view as the current block is determined by using a reference index of the current block to specify a reference picture in the reference picture list, and the view index of the specified reference picture is It may be determined based on whether it is equal to a view index of the current block.
- the current block may be reconstructed by adding a prediction value of the current block obtained in operation S220 and a residual value of the current block (S230).
- the residual value refers to the difference between the reconstruction value and the prediction value of the current block, and may be obtained by performing inverse quantization and / or inverse transformation on a transform coefficient extracted from the bitstream.
- the texture data and the depth data of the video image represent images of the same viewpoint and the same time and have high correlation. Therefore, when the depth data is encoded / decoded by using the same motion vector used for encoding / decoding the texture data, the encoding / decoding performance of the video image may be improved.
- the motion vector of the texture block corresponding to the depth block may be allocated to the texture merge candidate.
- the texture block may be determined as a block having the same position as the depth block.
- the motion vector of the interview motion candidate may be derived based on the inter-view motion prediction technique, which will be described with reference to FIG. 3.
- the current block belonging to the current view view 0 may specify a reference block belonging to the reference view view 1 using the disparity vector.
- the reference block may be specified as a block of a position shifted by the disparity vector at a block position of a reference view corresponding to a position of a current block.
- the temporal motion vector of the reference block may be assigned to the interview motion candidate.
- the current block may perform inter-view motion prediction on a sub-block basis.
- the current block may be divided into sub-block units (for example, 8x8), and the temporal motion vector of the reference block may be obtained in sub-block units and allocated to the interview motion candidates.
- the disparity vector of the current block may be derived from a depth image corresponding to the current block, which will be described in detail with reference to FIG. 5.
- the disparity vector may be derived from a neighboring block spatially adjacent to the current block, or may be derived from a temporal neighboring block located at a different time zone from the current block. A method of deriving a disparity vector from a spatial / temporal neighboring block of the current block will be described with reference to FIG. 6.
- the disparity vector of the interview motion candidate IvMC described above may be shifted by a specific value, and the reference block belonging to the reference view view 0 may be specified using the shifted disparity vector.
- the shifted disparity vector may be derived by shifting the disparity vector of the interview motion candidate IvMC in consideration of the width nPbW / height nPbH of the current block.
- the shifted disparity vector may be derived by shifting the disparity vector of the interview motion candidate IvMC by (nPbW * 2, nPbH * 2).
- the temporal motion vector of the reference block can be assigned to the interview motion shift candidate.
- a disparity vector may be derived from a depth image or a spatial / temporal neighboring block corresponding to the current block.
- a vector having zero vertical component (y component) in the derived disparity vector may be assigned to the interview variation candidate. For example, when the derived disparity vector of the current block is (mvDisp [0], mvDisp [1]), the vector (mvDisp [0], 0) may be assigned to the interview disparity candidate.
- the disparity vector may be derived from a depth image or a spatial / temporal neighboring block corresponding to the current block.
- a vector shifting the horizontal component (x component) in the derived disparity vector by a predetermined value may be assigned to the interview disparity shift candidate. For example, if the motion vector of the interview variation candidate is (mvDisp [0], mvDisp [1]), a vector obtained by moving the horizontal component mvDisp [0] by 4, that is, (mvDisp [0] +4, mvDisp [ 1)) can be assigned to interview variation shift candidates.
- a horizontal component (x component) may be shifted by a predetermined value in the derived disparity vector, and a vector in which the vertical component (y component) is set to 0 may be assigned to the interview variation shift candidate.
- the vector shifts the horizontal component mvDisp [0] by 4 and sets the vertical component mvDisp [1] to 0. That is, (mvDisp [0] +4, 0) may be assigned to the interview variation shift candidate.
- the motion vector of the VSP candidate may also be derived based on the above-described disparity vector of the current block, which will be described with reference to FIG. 4.
- a disparity vector (first disparity vector) may be derived from a depth image of a current block or a spatial / temporal neighboring block (S400), and a method of deriving the disparity vector is described with reference to FIGS. 5 to 6. It will be described later.
- the depth block of the reference view may be specified using the disparity vector derived in S400 (S410).
- the depth block may be included in the reference depth picture.
- the reference depth picture and the reference texture picture belong to the same access unit, and the reference texture picture may correspond to an inter-view reference picture of the current block.
- the modified disparity vector may be derived using at least one depth sample of a pre-defined position in the depth block (S420). For example, depth samples located at four corners belonging to the depth block may be used.
- the second disparity vector may be derived from a maximum value of depth samples located at four corners, may be derived from an average value of depth samples located at four corners, and may be derived from a depth sample of any one of four corners. It may be.
- FIG. 5 is a diagram illustrating a method of deriving a disparity vector of a current block using depth data of a depth image as an embodiment to which the present invention is applied.
- location information of a depth block in a depth picture corresponding to the current block may be acquired based on the location information of the current block (S500).
- the position of the depth block may be determined in consideration of the spatial resolution between the depth picture and the current picture.
- the position of the depth block may be determined as a block having the same position as the current block of the current picture.
- the current picture and the depth picture may be coded at different spatial resolutions. This is because the coding efficiency may not be significantly reduced even if the spatial resolution is coded at a lower level due to the characteristics of depth information representing distance information between the camera and the object. Therefore, when the spatial resolution of the depth picture is coded lower than that of the current picture, the decoder may involve an upsampling process for the depth picture before acquiring position information of the depth block.
- offset information may be additionally considered when acquiring position information of the current depth block in the upsampled depth picture.
- the offset information may include at least one of top offset information, left offset information, right offset information, and bottom offset information.
- the top offset information may indicate a position difference between at least one pixel located at the top of the upsampled depth picture and at least one pixel located at the top of the current picture.
- Left, right, and bottom offset information may also be defined in the same manner.
- depth data corresponding to position information of a depth block may be obtained (S510).
- depth data corresponding to corner pixels of the depth block may be used.
- depth data corresponding to the center pixel of the depth block may be used.
- any one of a maximum value, a minimum value, and a mode value may be selectively used among the plurality of depth data corresponding to the plurality of pixels, or an average value of the plurality of depth data may be used.
- the disparity vector of the current block may be derived by using the depth data obtained in operation S510 (S520).
- the disparity vector of the current block may be derived as in Equation 3 below.
- v denotes depth data
- a denotes a scaling factor
- f denotes an offset used to derive a disparity vector.
- the scaling factor a and offset f may be signaled in a video parameter set or slice header, or may be a value pre-set in the decoder.
- n is a variable representing the value of the bit shift, which may be variably determined according to the accuracy of the disparity vector.
- FIG. 6 illustrates candidates of spatial / temporal neighboring blocks of the current block as an embodiment to which the present invention is applied.
- the spatial neighboring block includes a left neighboring block A1, an upper neighboring block B1, a lower left neighboring block A0, an upper right neighboring block B0, or an upper left neighboring block of the current block. It may include at least one of (B2).
- a temporal neighboring block may mean a block at the same position as the current block.
- the temporal neighboring block is a block belonging to a picture located at a different time zone from the current block, and includes a block BR corresponding to the lower right pixel of the current block, a block CT corresponding to the center pixel of the current block, or a current block. At least one of the blocks TL corresponding to the upper left pixel may be included.
- the disparity vector of the current block may be derived from a disparity-compensated prediction block (hereinafter, referred to as a DCP block) among the spatial / temporal neighboring blocks.
- the DCP block may mean a block encoded through inter-view texture prediction using a disparity vector.
- the DCP block may perform inter-view prediction using texture data of the reference block specified by the disparity vector.
- the disparity vector of the current block may be predicted or reconstructed using the disparity vector used by the DCP block for inter-view texture prediction.
- the disparity vector of the current block may be derived from a disparity vector based-motion compensation prediction block (hereinafter, referred to as a DV-MCP block) among the spatial neighboring blocks.
- the DV-MCP block may mean a block encoded through inter-view motion prediction using a disparity vector.
- the DV-MCP block may perform temporal inter prediction using the temporal motion vector of the reference block specified by the disparity vector.
- the disparity vector of the current block may be predicted or reconstructed using the disparity vector used by the DV-MCP block to obtain the temporal motion vector of the reference block.
- the current block may search whether a spatial / temporal neighboring block corresponds to a DCP block according to a pre-defined priority, and derive a disparity vector from the first found DCP block.
- the search may be performed with the priority of the spatial neighboring block-> temporal neighboring block, and among the spatial neighboring blocks with the priority of A1-> B1-> B0-> A0-> B2. It may be found whether it corresponds to the DCP block.
- this is only an embodiment of the priority, and may be determined differently within the scope apparent to those skilled in the art.
- the spatial / temporal neighboring blocks corresponds to the DCP block, it can additionally search whether the spatial neighboring block corresponds to the DV-MCP block, and likewise derive the disparity vector from the first searched DV-MCP block. .
- the maximum number of merge candidates MaxNumMergeCand constituting the merge candidate list may be variably determined. However, the maximum number of merge candidates MaxNumMergeCand may be limited within a preset range (eg, 1 to 6). Coding performance may be improved by adaptively adjusting the maximum number of merge candidates (MaxNumMergeCand) for each slice.
- the maximum number of merge candidates MaxNumMergeCand may be derived as in Equation 4 below.
- MaxNumMergeCand 5-five_minus_max_num_merge_cand + NumExtraMergeCand (Equation 4)
- five_minus_max_num_merge_cand is a syntax of a slice segment level, and the difference between the maximum number of merge candidates except the number of interview merge candidates (for example, 5) and the maximum number of merge candidates per slice except the number of interview merge candidates is calculated. Can mean.
- the variable NumExtraMergeCand may be derived as in Equation 5 below.
- the variable NumExtraMergeCand may be derived based on iv_mv_pred_flag [nuh_layer_id], mpi_flag [nuh_layer_id], or ViewSynthesisPredFlag.
- the mpi_flag is a syntax indicating whether motion parameter inheritance is performed.
- a motion parameter inheritance method may be used to decode a depth block using a motion vector of a texture block corresponding to the depth block, or to derive a motion vector of a texture block from a reference block of a neighboring view. It is called). Therefore, when motion parameter inheritance is performed according to mpi_flag, the above-described texture merge candidate or interview motion candidate IvMC may be used as the merge candidate of the current block, and the variable NumExtraMergeCand may be set to 1.
- ViewSynthesisPredFlag is a flag indicating whether a VSP candidate is used. Therefore, when the value of ViewSynthesisPredFlag is 1, the current block may add VSP candidates to the merge candidate list, and the variable NumExtraMergeCand may be set to 1.
- the merge candidates described above that is, the spatial / temporal neighboring blocks and the interview merge candidates of the current block may be included in the merge candidate list for the current block, but may be included as much as the maximum number of derived merge candidates MaxNumMergeCand.
- a priority (or an arrangement order) needs to be defined among merge candidates added to the merge candidate list.
- the merge candidates are interview motion candidate (IvMC), left neighboring block (A1), top neighboring block (B1), top right neighboring block (B0), interview variation candidate (IvDC), VSP candidate, lower left neighboring block. (A0), the upper left neighboring block B2, the interview motion shift candidate IvMCShift, and the interview shift shift candidate IvDCShift may have priority.
- the merge candidates may include an interview motion candidate (IvMC), a left neighbor block (A1), an upper neighbor block (B1), a VSP candidate, an upper right neighbor block (B0), an interview variation candidate (IvDC), and a lower left neighbor block (A0).
- IvMC interview motion candidate
- A1 left neighbor block
- B1 upper neighbor block
- VSP lower neighbor block
- IvDC interview variation candidate
- A0 lower left neighbor block
- FIG. 7 illustrates a method for adaptively using an interview motion candidate IvMC based on an illumination compensation flag ic_flag as an embodiment to which the present invention is applied.
- an illumination compensation flag may be obtained from a bitstream (S700).
- the illumination compensation flag may mean information indicating whether illumination compensation is performed on a current block (eg, a coding unit or a prediction unit).
- Lighting compensation of the present invention means compensating for the lighting difference between viewpoints.
- IvDCs interview variation candidates
- IvMCs interview motion candidates
- step S700 it may be checked whether the value of the illumination compensation flag acquired in step S700 is 1 (S710).
- the interview motion candidate IvMC may be derived (S720).
- a merge candidate list including the interview motion candidate IvMC derived may be generated (S740). That is, the interview motion candidate IvMC may be added to the merge candidate list according to the above-described priority.
- the interview motion shift candidate IvMCShift associated with the interview motion candidate IvMC may also be excluded from the merge candidate list.
- Merge candidates other than the interview motion candidate IvMC and / or the interview motion shift candidate IvMCShift may be added to the merge candidate list according to the above-described priority. For example, within the maximum number of merge candidates (MaxNumMergeCand), the left neighbor block A1, the top neighbor block B1, the interview variation candidate IvDC, the top right neighbor block B0, the VSP candidate, and the bottom left neighbor.
- the priority may be added to the merge candidate list with the priority of the block A0, the upper left neighboring block B2, and the interview shift shift candidate IvDCShift.
- the priority of the upper left neighboring block B2 may be added to the merge candidate list.
- the interview variation shift candidate (IvDCShift) may be added to the merge candidate list in the order of priority.
- FIG. 8 illustrates priority of merge candidates based on an illumination compensation flag ic_flag according to an embodiment to which the present invention is applied.
- the encoding performance may be improved by changing the priority so that the interview variation candidate IvDC has a higher priority than the interview motion candidate IvMC.
- the priority may be changed so that the interview shift shift candidate IvDCShift has a higher priority than the interview motion shift candidate IvMCShift.
- the interview variation candidate IvDC when the value of the illumination compensation flag is 1, the interview variation candidate IvDC, the left neighbor block A1, the upper neighbor block B1, the interview motion candidate IvMC, and the upper right neighbor block B0.
- the VSP candidate, the interview shift shift candidate IvDCShift, the interview motion shift candidate IvMCShift, the lower left neighboring block A0 and the upper left neighboring block B2 may be added to the merge candidate list.
- the priority of the merge candidate shown in FIG. 8 means that the smaller the value, the higher the priority, and the larger the value, the lower the priority.
- the interview motion candidate (IvMC), the left neighboring block (A1), the top neighboring block (B1), the interview variation candidate (IvDC), the top right neighboring block (B0), the VSP candidate The priority of the interview motion shift candidate IvMCShift, the interview shift shift candidate IvDCShift, the lower left neighboring block A0, and the upper left neighboring block B2 may be added to the merge candidate list.
- the present invention can be used to code a multiview video signal.
Abstract
Description
Claims (15)
- 현재 블록에 관한 머지 후보 리스트를 생성하는 단계; 여기서, 상기 머지 후보 리스트는 적어도 하나의 머지 후보자로 구성되고, 상기 머지 후보자는 공간적 이웃 블록, 시간적 이웃 블록 또는 인터뷰 모션 후보자(IvMC) 중 적어도 하나를 포함함,비트스트림으로부터 획득된 상기 현재 블록에 대한 머지 인덱스에 기반하여 상기 현재 블록의 모션 벡터를 유도하는 단계; 여기서, 상기 머지 인덱스는 상기 현재 블록을 머지 모드로 복호화하기 위해 이용되는 머지 후보자를 특정함,상기 유도된 모션 벡터를 이용하여 상기 현재 블록의 예측값을 획득하는 단계; 및상기 획득된 예측값과 상기 현재 블록에 관한 잔차값을 더하여 상기 현재 블록을 복원하는 단계를 포함하는 것을 특징으로 하는 다시점 비디오 신호 디코딩 방법.
- 제1항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록의 변이 벡터에 의해 특정되는 참조 블록의 시간적 모션 벡터를 가지되,상기 참조 블록은 상기 현재 블록의 참조 시점(reference view)에 속하는 것을 특징으로 하는 다시점 비디오 신호 디코딩 방법.
- 제2항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록에 대해 조명 보상(illumination compensation)이 수행되는지 여부를 고려하여 상기 머지 후보 리스트에 제한적으로 포함되는 것을 특징으로 하는 다시점 비디오 신호 디코딩 방법.
- 제3항에 있어서,상기 현재 블록에 대해 조명 보상이 수행되지 아니하는 경우, 상기 인터뷰 모션 후보자(IvMC)는 상기 머지 후보 리스트에 포함되되,상기 인터뷰 모션 후보자(IvMC), 상기 공간적 이웃 블록, 상기 시간적 이웃 블록의 우선 순위로 상기 머지 후보 리스트에 배열되는 것을 특징으로 하는 다시점 비디오 신호 디코딩 방법.
- 현재 블록에 관한 머지 후보 리스트를 생성하고, 비트스트림으로부터 획득된 상기 현재 블록에 대한 머지 인덱스에 기반하여 상기 현재 블록의 모션 벡터를 유도하며, 상기 유도된 모션 벡터를 이용하여 상기 현재 블록의 예측값을 획득하는 인터 예측부; 및상기 획득된 예측값과 상기 현재 블록에 관한 잔차값을 더하여 상기 현재 블록을 복원하는 복원부를 포함하되,상기 머지 후보 리스트는 적어도 하나의 머지 후보자로 구성되고, 상기 머지 후보자는 공간적 이웃 블록, 시간적 이웃 블록 또는 인터뷰 모션 후보자(IvMC) 중 적어도 하나를 포함하며, 상기 머지 인덱스는 상기 현재 블록을 머지 모드로 복호화하기 위해 이용되는 머지 후보자를 특정하는 것을 특징으로 하는 다시점 비디오 신호 디코딩 장치.
- 제5항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록의 변이 벡터에 의해 특정되는 참조 블록의 시간적 모션 벡터를 가지되,상기 참조 블록은 상기 현재 블록의 참조 시점(reference view)에 속하는 것을 특징으로 하는 다시점 비디오 신호 디코딩 장치.
- 제6항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록에 대해 조명 보상(illumination compensation)이 수행되는지 여부를 고려하여 상기 머지 후보 리스트에 제한적으로 포함되는 것을 특징으로 하는 다시점 비디오 신호 디코딩 장치.
- 제7항에 있어서, 상기 인터 예측부는,상기 현재 블록에 대해 조명 보상이 수행되지 아니하는 경우에 상기 인터뷰 모션 후보자(IvMC)를 상기 머지 후보 리스트에 포함하되,상기 인터뷰 모션 후보자(IvMC), 상기 공간적 이웃 블록, 상기 시간적 이웃 블록의 우선 순위로 상기 머지 후보 리스트에 배열하는 것을 특징으로 하는 다시점 비디오 신호 디코딩 장치.
- 현재 블록에 관한 머지 후보 리스트를 생성하는 단계; 여기서, 상기 머지 후보 리스트는 적어도 하나의 머지 후보자로 구성되고, 상기 머지 후보자는 공간적 이웃 블록, 시간적 이웃 블록 또는 인터뷰 모션 후보자(IvMC) 중 적어도 하나를 포함함,상기 현재 블록에 대한 머지 인덱스에 기반하여 상기 현재 블록의 모션 벡터를 유도하는 단계; 여기서, 상기 머지 인덱스는 상기 현재 블록을 머지 모드로 부호화하기 위해 이용되는 머지 후보자를 특정함,상기 유도된 모션 벡터를 이용하여 상기 현재 블록의 예측값을 획득하는 단계; 및상기 획득된 예측값과 상기 현재 블록에 관한 잔차값을 더하여 상기 현재 블록을 복원하는 단계를 포함하는 것을 특징으로 하는 다시점 비디오 신호 인코딩 방법.
- 제9항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록의 변이 벡터에 의해 특정되는 참조 블록의 시간적 모션 벡터를 가지되,상기 참조 블록은 상기 현재 블록의 참조 시점(reference view)에 속하는 것을 특징으로 하는 다시점 비디오 신호 인코딩 방법.
- 제10항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록에 대해 조명 보상(illumination compensation)이 수행되는지 여부를 고려하여 상기 머지 후보 리스트에 제한적으로 포함되는 것을 특징으로 하는 다시점 비디오 신호 인코딩 방법.
- 제11항에 있어서,상기 현재 블록에 대해 조명 보상이 수행되지 아니하는 경우, 상기 인터뷰 모션 후보자(IvMC)는 상기 머지 후보 리스트에 포함되되,상기 인터뷰 모션 후보자(IvMC), 상기 공간적 이웃 블록, 상기 시간적 이웃 블록의 우선 순위로 상기 머지 후보 리스트에 배열되는 것을 특징으로 하는 다시점 비디오 신호 인코딩 방법.
- 현재 블록에 관한 머지 후보 리스트를 생성하고, 상기 현재 블록에 대한 머지 인덱스에 기반하여 상기 현재 블록의 모션 벡터를 유도하며, 상기 유도된 모션 벡터를 이용하여 상기 현재 블록의 예측값을 획득하는 인터 예측부; 및상기 획득된 예측값과 상기 현재 블록에 관한 잔차값을 더하여 상기 현재 블록을 복원하는 복원부를 포함하되,상기 머지 후보 리스트는 적어도 하나의 머지 후보자로 구성되고, 상기 머지 후보자는 공간적 이웃 블록, 시간적 이웃 블록 또는 인터뷰 모션 후보자(IvMC) 중 적어도 하나를 포함하며, 상기 머지 인덱스는 상기 현재 블록을 머지 모드로 부호화하기 위해 이용되는 머지 후보자를 특정하는 것을 특징으로 하는 다시점 비디오 신호 인코딩 장치.
- 제13항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록의 변이 벡터에 의해 특정되는 참조 블록의 시간적 모션 벡터를 가지되,상기 참조 블록은 상기 현재 블록의 참조 시점(reference view)에 속하는 것을 특징으로 하는 다시점 비디오 신호 인코딩 장치.
- 제14항에 있어서,상기 인터뷰 모션 후보자(IvMC)는 상기 현재 블록에 대해 조명 보상(illumination compensation)이 수행되는지 여부를 고려하여 상기 머지 후보 리스트에 제한적으로 포함되는 것을 특징으로 하는 다시점 비디오 신호 인코딩 장치.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201580014887.2A CN106105191B (zh) | 2014-03-19 | 2015-03-18 | 用于处理多视图视频信号的方法和装置 |
US15/125,689 US10257531B2 (en) | 2014-03-19 | 2015-03-18 | Method and apparatus for processing multiview video signals based on illumination compensation and inter-view motion candidate |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR10-2014-0032101 | 2014-03-19 | ||
KR20140032101 | 2014-03-19 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015142054A1 true WO2015142054A1 (ko) | 2015-09-24 |
Family
ID=54144943
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2015/002629 WO2015142054A1 (ko) | 2014-03-19 | 2015-03-18 | 다시점 비디오 신호 처리 방법 및 장치 |
Country Status (4)
Country | Link |
---|---|
US (1) | US10257531B2 (ko) |
KR (1) | KR20150109282A (ko) |
CN (5) | CN108965888B (ko) |
WO (1) | WO2015142054A1 (ko) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018012886A1 (ko) * | 2016-07-12 | 2018-01-18 | 한국전자통신연구원 | 영상 부호화/복호화 방법 및 이를 위한 기록 매체 |
CN113873243A (zh) * | 2016-08-31 | 2021-12-31 | 株式会社Kt | 用于对视频进行解码的方法和用于对视频进行编码的方法 |
Families Citing this family (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20170065502A (ko) * | 2014-10-08 | 2017-06-13 | 엘지전자 주식회사 | 3d 비디오 부호화/복호화 방법 및 장치 |
KR20180019092A (ko) * | 2015-06-16 | 2018-02-23 | 엘지전자 주식회사 | 영상 코딩 시스템에서 조도 보상에 기반한 블록 예측 방법 및 장치 |
KR102569476B1 (ko) * | 2016-10-11 | 2023-08-24 | 엘지전자 주식회사 | 영상 코딩 시스템에서 인트라 예측에 따른 영상 디코딩 방법 및 장치 |
US10523934B2 (en) * | 2017-05-31 | 2019-12-31 | Mediatek Inc. | Split based motion vector operation reduction |
CN107396083B (zh) * | 2017-07-27 | 2020-01-14 | 青岛海信电器股份有限公司 | 一种全息图像生成方法及装置 |
EP3468194A1 (en) * | 2017-10-05 | 2019-04-10 | Thomson Licensing | Decoupled mode inference and prediction |
EP3758374A4 (en) * | 2018-02-22 | 2021-09-01 | LG Electronics Inc. | IMAGE DECODING METHOD AND DEVICE ACCORDING TO A BLOCK DIVISION STRUCTURE IN AN IMAGE ENCODING SYSTEM |
CN110365987B (zh) * | 2018-04-09 | 2022-03-25 | 杭州海康威视数字技术股份有限公司 | 一种运动矢量确定方法、装置及其设备 |
WO2019234598A1 (en) | 2018-06-05 | 2019-12-12 | Beijing Bytedance Network Technology Co., Ltd. | Interaction between ibc and stmvp |
WO2019245712A1 (en) * | 2018-06-18 | 2019-12-26 | Interdigital Vc Holdings, Inc. | Illumination compensation in video coding |
GB2589223B (en) | 2018-06-21 | 2023-01-25 | Beijing Bytedance Network Tech Co Ltd | Component-dependent sub-block dividing |
WO2019244117A1 (en) | 2018-06-21 | 2019-12-26 | Beijing Bytedance Network Technology Co., Ltd. | Unified constrains for the merge affine mode and the non-merge affine mode |
TWI744661B (zh) | 2018-06-29 | 2021-11-01 | 大陸商北京字節跳動網絡技術有限公司 | 要根據模式檢查的查找表中的運動候選的數量 |
KR20240005239A (ko) | 2018-06-29 | 2024-01-11 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | Lut와 amvp 사이의 상호작용 |
CN110662056B (zh) | 2018-06-29 | 2022-06-07 | 北京字节跳动网络技术有限公司 | 哪个查找表需要更新或不更新 |
BR112020024142A2 (pt) | 2018-06-29 | 2021-03-02 | Beijing Bytedance Network Technology Co., Ltd. | método para processamento de vídeo, aparelho para codificação de dados de vídeo, meio de armazenamento e meio de gravação legíveis por computador não transitório |
CN114845108A (zh) | 2018-06-29 | 2022-08-02 | 抖音视界(北京)有限公司 | 查找表的更新:fifo、约束的fifo |
KR20210025537A (ko) | 2018-06-29 | 2021-03-09 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | 하나 또는 다수의 룩업 테이블들을 사용하여 이전에 코딩된 모션 정보를 순서대로 저장하고 이를 사용하여 후속 블록들을 코딩하는 개념 |
TWI731364B (zh) | 2018-07-02 | 2021-06-21 | 大陸商北京字節跳動網絡技術有限公司 | Hmvp+非相鄰運動 |
CN110876058B (zh) * | 2018-08-30 | 2021-09-21 | 华为技术有限公司 | 一种历史候选列表更新方法与装置 |
GB2590310B (en) | 2018-09-12 | 2023-03-22 | Beijing Bytedance Network Tech Co Ltd | Conditions for starting checking HMVP candidates depend on total number minus K |
CN110944196B (zh) | 2018-09-24 | 2023-05-30 | 北京字节跳动网络技术有限公司 | 简化的基于历史的运动矢量预测 |
CN111010571B (zh) * | 2018-10-08 | 2023-05-16 | 北京字节跳动网络技术有限公司 | 组合仿射Merge候选的生成和使用 |
CN116634170A (zh) * | 2018-10-10 | 2023-08-22 | 华为技术有限公司 | 帧间预测方法及装置 |
CN116939198A (zh) | 2018-10-23 | 2023-10-24 | 北京字节跳动网络技术有限公司 | 并置的局部照明补偿和修改的帧间预测编解码 |
WO2020084509A1 (en) | 2018-10-23 | 2020-04-30 | Beijing Bytedance Network Technology Co., Ltd. | Harmonized local illumination compensation and modified inter coding tools |
CN116760998A (zh) * | 2018-11-08 | 2023-09-15 | Oppo广东移动通信有限公司 | 图像信号编码/解码方法及其设备 |
CN112970262B (zh) * | 2018-11-10 | 2024-02-20 | 北京字节跳动网络技术有限公司 | 三角预测模式中的取整 |
CN113056916A (zh) * | 2018-11-22 | 2021-06-29 | 北京字节跳动网络技术有限公司 | 基于子块的运动候选的选择和信令 |
CN113273189A (zh) * | 2018-12-31 | 2021-08-17 | 北京字节跳动网络技术有限公司 | 具有MVD的Merge和AMVR之间的交互 |
WO2020141849A1 (ko) * | 2019-01-01 | 2020-07-09 | 엘지전자 주식회사 | 화면간 예측을 사용하여 비디오 신호를 처리하기 위한 방법 및 장치 |
KR20240010576A (ko) | 2019-01-10 | 2024-01-23 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | Lut 업데이트의 호출 |
WO2020143824A1 (en) | 2019-01-13 | 2020-07-16 | Beijing Bytedance Network Technology Co., Ltd. | Interaction between lut and shared merge list |
WO2020147747A1 (en) | 2019-01-15 | 2020-07-23 | Beijing Bytedance Network Technology Co., Ltd. | Weighted prediction in video coding |
WO2020147773A1 (en) | 2019-01-16 | 2020-07-23 | Beijing Bytedance Network Technology Co., Ltd. | Inserting order of motion candidates in lut |
CN113302919A (zh) | 2019-01-17 | 2021-08-24 | 北京字节跳动网络技术有限公司 | 在视频处理中使用虚拟候选预测和加权预测 |
EP3878181A4 (en) * | 2019-03-04 | 2022-03-09 | Huawei Technologies Co., Ltd. | ENCODERS, DECODERS AND RELATED METHODS USING IBC MERGE LIST |
CN111698506B (zh) * | 2019-03-11 | 2022-04-26 | 杭州海康威视数字技术股份有限公司 | 运动信息候选者列表构建方法、三角预测解码方法及装置 |
CN113557735A (zh) * | 2019-03-11 | 2021-10-26 | 北京字节跳动网络技术有限公司 | 运动候选列表构造的改进 |
CN113615193A (zh) | 2019-03-22 | 2021-11-05 | 北京字节跳动网络技术有限公司 | Merge列表构建和其他工具之间的交互 |
WO2020233662A1 (en) * | 2019-05-21 | 2020-11-26 | Beijing Bytedance Network Technology Co., Ltd. | Syntax signaling for optical-flow based inter coding |
EP4032290A4 (en) | 2019-10-18 | 2022-11-30 | Beijing Bytedance Network Technology Co., Ltd. | SYNTAX CONSTRAINTS IN REPORTING SUBPICTURE PARAMETER SETS |
WO2024077611A1 (zh) * | 2022-10-14 | 2024-04-18 | Oppo广东移动通信有限公司 | 解码方法、编码方法、解码器以及编码器 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20080034752A (ko) * | 2006-10-17 | 2008-04-22 | 경희대학교 산학협력단 | 다시점 영상의 부호화 및 복호화 방법과 그를 위한 장치 |
KR20090037888A (ko) * | 2006-07-11 | 2009-04-16 | 톰슨 라이센싱 | 가상 참조 픽처를 사용하는 방법 및 장치 |
JP2009530960A (ja) * | 2006-03-22 | 2009-08-27 | 韓國電子通信研究院 | 照明変化補償動き予測符号化および復号化方法とその装置 |
KR20130028633A (ko) * | 2011-09-09 | 2013-03-19 | 주식회사 케이티 | 화면 간 예측 수행시 후보 블록 결정 방법 및 이러한 방법을 사용하는 장치 |
WO2013069933A1 (ko) * | 2011-11-07 | 2013-05-16 | 엘지전자 주식회사 | 영상 부호화/복호화 방법 및 그 장치 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101946376B1 (ko) * | 2007-10-16 | 2019-02-11 | 엘지전자 주식회사 | 비디오 신호 처리 방법 및 장치 |
PL3739883T3 (pl) | 2010-05-04 | 2023-04-11 | Lg Electronics Inc. | Sposób i aparat do kodowania i dekodowania sygnału wideo |
JP5872676B2 (ja) * | 2011-06-15 | 2016-03-01 | メディアテック インコーポレイテッド | 3dビデオコーディングにおけるテクスチャーイメージ圧縮方法および装置 |
CN107295347B (zh) * | 2011-08-29 | 2020-06-26 | 苗太平洋控股有限公司 | 用于解码合并模式下的运动信息的装置 |
KR20130050407A (ko) | 2011-11-07 | 2013-05-16 | 오수미 | 인터 모드에서의 움직임 정보 생성 방법 |
EP4009640A1 (en) | 2011-11-08 | 2022-06-08 | Electronics and Telecommunications Research Institute | Method and device for sharing a candidate list |
US9596466B2 (en) | 2011-12-23 | 2017-03-14 | Electronics And Telecommunications Research Institute | Method and apparatus for setting reference picture index of temporal merging candidate |
US9615090B2 (en) * | 2012-12-28 | 2017-04-04 | Qualcomm Incorporated | Parsing syntax elements in three-dimensional video coding |
-
2015
- 2015-03-18 KR KR1020150037359A patent/KR20150109282A/ko not_active Application Discontinuation
- 2015-03-18 CN CN201810764534.XA patent/CN108965888B/zh active Active
- 2015-03-18 WO PCT/KR2015/002629 patent/WO2015142054A1/ko active Application Filing
- 2015-03-18 CN CN201810763413.3A patent/CN108632629B9/zh active Active
- 2015-03-18 CN CN201810764047.3A patent/CN108683922B/zh active Active
- 2015-03-18 US US15/125,689 patent/US10257531B2/en active Active
- 2015-03-18 CN CN201810763415.2A patent/CN108989815B9/zh active Active
- 2015-03-18 CN CN201580014887.2A patent/CN106105191B/zh active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009530960A (ja) * | 2006-03-22 | 2009-08-27 | 韓國電子通信研究院 | 照明変化補償動き予測符号化および復号化方法とその装置 |
KR20090037888A (ko) * | 2006-07-11 | 2009-04-16 | 톰슨 라이센싱 | 가상 참조 픽처를 사용하는 방법 및 장치 |
KR20080034752A (ko) * | 2006-10-17 | 2008-04-22 | 경희대학교 산학협력단 | 다시점 영상의 부호화 및 복호화 방법과 그를 위한 장치 |
KR20130028633A (ko) * | 2011-09-09 | 2013-03-19 | 주식회사 케이티 | 화면 간 예측 수행시 후보 블록 결정 방법 및 이러한 방법을 사용하는 장치 |
WO2013069933A1 (ko) * | 2011-11-07 | 2013-05-16 | 엘지전자 주식회사 | 영상 부호화/복호화 방법 및 그 장치 |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018012886A1 (ko) * | 2016-07-12 | 2018-01-18 | 한국전자통신연구원 | 영상 부호화/복호화 방법 및 이를 위한 기록 매체 |
US11509930B2 (en) | 2016-07-12 | 2022-11-22 | Electronics And Telecommunications Research Institute | Image encoding/decoding method and recording medium therefor |
US11895329B2 (en) | 2016-07-12 | 2024-02-06 | Electronics And Telecommunications Research Institute | Image encoding/decoding method and recording medium therefor |
CN113873243A (zh) * | 2016-08-31 | 2021-12-31 | 株式会社Kt | 用于对视频进行解码的方法和用于对视频进行编码的方法 |
CN113873243B (zh) * | 2016-08-31 | 2024-02-20 | 株式会社Kt | 用于对视频进行解码的方法和用于对视频进行编码的方法 |
US11962774B2 (en) | 2016-08-31 | 2024-04-16 | Kt Corporation | Method and apparatus for processing video signal |
Also Published As
Publication number | Publication date |
---|---|
CN108989815A (zh) | 2018-12-11 |
CN108632629B (zh) | 2021-05-04 |
CN106105191B (zh) | 2018-08-10 |
CN106105191A (zh) | 2016-11-09 |
CN108989815B (zh) | 2021-05-04 |
US20170006302A1 (en) | 2017-01-05 |
CN108632629B9 (zh) | 2021-06-15 |
CN108632629A (zh) | 2018-10-09 |
CN108989815B9 (zh) | 2021-06-04 |
KR20150109282A (ko) | 2015-10-01 |
CN108965888B (zh) | 2021-05-04 |
US10257531B2 (en) | 2019-04-09 |
CN108683922A (zh) | 2018-10-19 |
CN108683922B (zh) | 2021-05-04 |
CN108965888A (zh) | 2018-12-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015142054A1 (ko) | 다시점 비디오 신호 처리 방법 및 장치 | |
WO2015142057A1 (ko) | 다시점 비디오 신호 처리 방법 및 장치 | |
WO2013169031A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2018174402A1 (ko) | 영상 코딩 시스템에서 변환 방법 및 그 장치 | |
WO2018062921A1 (ko) | 영상 코딩 시스템에서 블록 분할 및 인트라 예측 방법 및 장치 | |
WO2013162273A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2018056603A1 (ko) | 영상 코딩 시스템에서 조도 보상 기반 인터 예측 방법 및 장치 | |
WO2013032074A1 (ko) | 머지 모드 움직임 정보 복호화 장치 | |
WO2016200043A1 (ko) | 비디오 코딩 시스템에서 가상 참조 픽처 기반 인터 예측 방법 및 장치 | |
WO2013165143A1 (ko) | 멀티뷰 영상의 부호화 방법, 부호화 장치, 복호화 방법, 복호화 장치 | |
WO2016056821A1 (ko) | 3d 비디오 코딩을 위한 움직임 정보 압축 방법 및 장치 | |
WO2016056822A1 (ko) | 3d 비디오 코딩 방법 및 장치 | |
WO2019198997A1 (ko) | 인트라 예측에 기반한 영상 코딩 방법 및 그 장치 | |
WO2015057033A1 (ko) | 3d 비디오 부호화/복호화 방법 및 장치 | |
WO2019112071A1 (ko) | 영상 코딩 시스템에서 크로마 성분의 효율적 변환에 기반한 영상 디코딩 방법 및 장치 | |
WO2014168443A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2018056602A1 (ko) | 영상 코딩 시스템에서 인터 예측 방법 및 장치 | |
WO2016056782A1 (ko) | 비디오 코딩에서 뎁스 픽처 코딩 방법 및 장치 | |
WO2019194500A1 (ko) | 인트라 예측에 기반한 영상 코딩 방법 및 그 장치 | |
WO2014010918A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2013133627A1 (ko) | 비디오 신호 처리 방법 | |
WO2016056754A1 (ko) | 3d 비디오 부호화/복호화 방법 및 장치 | |
WO2016003210A1 (ko) | 다시점 비디오 신호 처리 방법 및 장치 | |
WO2016003209A1 (ko) | 다시점 비디오 신호 처리 방법 및 장치 | |
WO2015182927A1 (ko) | 다시점 비디오 신호 처리 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15764242 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15125689 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 13/01/2017) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15764242 Country of ref document: EP Kind code of ref document: A1 |