WO2020262918A1 - 움직임 벡터를 이용한 영상 코딩 방법 및 장치 - Google Patents
움직임 벡터를 이용한 영상 코딩 방법 및 장치 Download PDFInfo
- Publication number
- WO2020262918A1 WO2020262918A1 PCT/KR2020/008137 KR2020008137W WO2020262918A1 WO 2020262918 A1 WO2020262918 A1 WO 2020262918A1 KR 2020008137 W KR2020008137 W KR 2020008137W WO 2020262918 A1 WO2020262918 A1 WO 2020262918A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- prediction
- information
- motion vector
- reference picture
- picture
- Prior art date
Links
- 239000013598 vector Substances 0.000 title claims abstract description 221
- 238000000034 method Methods 0.000 title claims description 214
- 230000001965 increasing effect Effects 0.000 abstract description 6
- 239000000523 sample Substances 0.000 description 69
- 230000007774 longterm Effects 0.000 description 44
- 230000002123 temporal effect Effects 0.000 description 41
- 238000001914 filtration Methods 0.000 description 33
- 230000008569 process Effects 0.000 description 28
- 238000013139 quantization Methods 0.000 description 24
- 238000012545 processing Methods 0.000 description 20
- 238000009795 derivation Methods 0.000 description 19
- 230000001939 inductive effect Effects 0.000 description 16
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 description 11
- 238000010586 diagram Methods 0.000 description 10
- 239000013074 reference sample Substances 0.000 description 10
- 230000011664 signaling Effects 0.000 description 10
- 230000009466 transformation Effects 0.000 description 10
- 230000002457 bidirectional effect Effects 0.000 description 9
- 230000005540 biological transmission Effects 0.000 description 9
- 230000000694 effects Effects 0.000 description 7
- 230000002093 peripheral effect Effects 0.000 description 7
- 241000023320 Luma <angiosperm> Species 0.000 description 6
- OSWPMRLSEDHDFF-UHFFFAOYSA-N methyl salicylate Chemical compound COC(=O)C1=CC=CC=C1O OSWPMRLSEDHDFF-UHFFFAOYSA-N 0.000 description 6
- 230000003044 adaptive effect Effects 0.000 description 5
- 238000004891 communication Methods 0.000 description 5
- 238000007906 compression Methods 0.000 description 5
- 230000006835 compression Effects 0.000 description 5
- 230000001419 dependent effect Effects 0.000 description 4
- 230000014509 gene expression Effects 0.000 description 4
- 238000013144 data compression Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 2
- 230000002146 bilateral effect Effects 0.000 description 2
- 239000000470 constituent Substances 0.000 description 2
- 238000003709 image segmentation Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000005192 partition Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000006698 induction Effects 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000008707 rearrangement Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 239000010454 slate Substances 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/109—Selection of coding mode or of prediction mode among a plurality of temporal predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/132—Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/137—Motion inside a coding unit, e.g. average field, frame or block difference
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/137—Motion inside a coding unit, e.g. average field, frame or block difference
- H04N19/139—Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
- H04N19/159—Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/573—Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/577—Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
Definitions
- a method and apparatus for efficiently performing inter prediction in an image/video coding system are provided.
- a method and apparatus for signaling an SMVD flag is provided.
- efficient inter prediction may be performed by using short-term reference pictures to derive a reference picture index for SMVD.
- FIG. 2 is a diagram schematically illustrating a configuration of a video/video encoding apparatus applicable to embodiments of the present document.
- 10 is a diagram for describing a method of deriving a motion vector in inter prediction.
- 15 shows a process of inducing MVD of MMVD according to an embodiment of the present document.
- the video source may acquire a video/image through a process of capturing, synthesizing, or generating a video/image.
- the video source may include a video/image capturing device and/or a video/image generating device.
- the video/image capture device may include, for example, one or more cameras, a video/image archive including previously captured video/images, and the like.
- the video/image generating device may include, for example, a computer, a tablet and a smartphone, and may (electronically) generate a video/image.
- a virtual video/image may be generated through a computer or the like, and in this case, a video/image capturing process may be substituted as a process of generating related data.
- the tile column is a rectangular region of CTUs, the rectangular region has a height equal to the height of the picture, and the width may be specified by syntax elements in a picture parameter set (The tile column is a rectangular region of CTUs having a height equal to the height of the picture and a width specified by syntax elements in the picture parameter set).
- the tile row is a rectangular region of CTUs, the rectangular region has a width specified by syntax elements in a picture parameter set, and a height may be the same as the height of the picture (The tile row is a rectangular region of CTUs having a height specified by syntax elements in the picture parameter set and a width equal to the width of the picture).
- At least one of A, B and C means “only A”, “only B”, “only C", or "A, B and C May mean any combination of A, B and C”.
- at least one of A, B or C (at least one of A, B or C) or “at least one of A, B and/or C (at least one of A, B and/or C)” It can mean “at least one of A, B and C”.
- the reference picture including the reference block and the reference picture including the temporal neighboring block may be the same or different.
- the temporal neighboring block may be called a collocated reference block, a co-located CU (colCU), and the like, and a reference picture including the temporal neighboring block may be referred to as a collocated picture (colPic).
- the inter prediction unit 221 constructs a motion information candidate list based on neighboring blocks, and provides information indicating which candidate is used to derive a motion vector and/or a reference picture index of the current block. Can be generated. Inter prediction may be performed based on various prediction modes.
- the inverse quantization unit 321 may inverse quantize the quantized transform coefficients and output transform coefficients.
- the inverse quantization unit 321 may rearrange the quantized transform coefficients in a two-dimensional block shape. In this case, the rearrangement may be performed based on the coefficient scan order performed by the encoding device.
- the inverse quantization unit 321 may perform inverse quantization on quantized transform coefficients by using a quantization parameter (for example, quantization step size information) and obtain transform coefficients.
- a quantization parameter for example, quantization step size information
- the prediction unit may perform prediction on the current block and generate a predicted block including prediction samples for the current block.
- the prediction unit may determine whether intra prediction or inter prediction is applied to the current block based on the information about the prediction output from the entropy decoding unit 310, and may determine a specific intra/inter prediction mode.
- the inter prediction unit 332 may derive a predicted block for the current block based on a reference block (reference sample array) specified by a motion vector on the reference picture.
- motion information may be predicted in units of blocks, subblocks, or samples based on correlation between motion information between neighboring blocks and the current block.
- the motion information may include a motion vector and a reference picture index.
- the motion information may further include inter prediction direction (L0 prediction, L1 prediction, Bi prediction, etc.) information.
- the neighboring block may include a spatial neighboring block existing in the current picture and a temporal neighboring block existing in the reference picture.
- the (modified) reconstructed picture stored in the DPB of the memory 360 may be used as a reference picture in the inter prediction unit 332.
- the memory 360 may store motion information of a block from which motion information in a current picture is derived (or decoded) and/or motion information of blocks in a picture that have already been reconstructed.
- the stored motion information may be transmitted to the inter prediction unit 260 to be used as motion information of a spatial neighboring block or motion information of a temporal neighboring block.
- the memory 360 may store reconstructed samples of reconstructed blocks in the current picture, and may be transmitted to the intra prediction unit 331.
- the intra prediction procedure may include determining an intra prediction mode/type, deriving a neighboring reference sample, and deriving an intra prediction mode/type based prediction sample. Also, a post-filtering step may be performed on the derived prediction samples as necessary.
- three types of modes can be considered: default intra modes, neighbor intra modes, and derived intra modes.
- motion information of the current block may be predicted in units of blocks, subblocks, or samples based on correlation between motion information between neighboring blocks and the current block.
- the motion information may include a motion vector and a reference picture index.
- the motion information may further include inter prediction type (L0 prediction, L1 prediction, Bi prediction, etc.) information.
- the neighboring block may include a spatial neighboring block existing in the current picture and a temporal neighboring block existing in the reference picture.
- the reference picture including the reference block and the reference picture including the temporal neighboring block may be the same or different.
- the motion vector L0 may represent a motion vector associated with the reference picture list L0 (L0), and the motion vector L1 may represent a motion vector associated with the reference picture list L1 (L1).
- the reference picture list L0 may include pictures prior to the current picture in output order as reference pictures, and the reference picture list L1 may include pictures after the current picture in output order.
- the previous pictures may be referred to as forward (reference) pictures, and the subsequent pictures may be referred to as reverse (reference) pictures.
- the reference picture list L0 may further include pictures later in output order than the current picture as reference pictures. In this case, the previous pictures in the reference picture list L0 may be indexed first, and the subsequent pictures may be indexed next.
- the encoding apparatus may derive residual samples based on the prediction samples (S410).
- the encoding apparatus may derive the residual samples by comparing the original samples of the current block with the prediction samples.
- the encoding apparatus may generate a reconstructed picture (including reconstructed samples and a reconstructed block) based on the reference samples and the residual samples. This is because the encoding device derives the same prediction result as that performed in the decoding device, and coding efficiency can be improved through this. Accordingly, the encoding apparatus may store a reconstructed picture (or reconstructed samples, and a reconstructed block) in a memory and use it as a reference picture for inter prediction. As described above, an in-loop filtering procedure or the like may be further applied to the reconstructed picture.
- motion information of the current block may be derived without configuring a candidate list.
- motion information of the current block may be derived according to a procedure disclosed in a prediction mode to be described later.
- the configuration of the candidate list as described above may be omitted.
- the MVP mode may also be called an advanced motion vector prediction (AMVP) mode.
- AMVP advanced motion vector prediction
- some modes and/or motion information candidates derived by some modes may be included as one of motion information related candidates of other modes.
- the HMVP candidate may be added as a merge candidate of the merge/skip mode, or may be added as an mvp candidate of the MVP mode.
- the HMVP candidate may be referred to as an HMVP merge candidate.
- the following table shows syntax for a coding unit according to an embodiment of this document.
- the encoder In order to perform the merge mode, the encoder must search for a merge candidate block used to derive motion information of the current prediction block. For example, up to five merge candidate blocks may be used, but embodiments of this document are not limited thereto.
- the maximum number of merge candidate blocks may be transmitted in a slice header or a tile group header. After finding the merge candidate blocks, the encoder may generate a merge candidate list and select a merge candidate block having the lowest cost among them as a final merge candidate block.
- FIG. 7 schematically shows a method of constructing a merge candidate list according to this document.
- the temporal merge candidate may be derived based on motion information of a covered prediction block.
- the predetermined storage unit is a 2nx2n sample unit
- the modified positions ((xTnb>>n) ⁇ n), (yTnb>>) Motion information of the prediction block located at n) ⁇ n)
- the predetermined storage unit is a 16x16 sample unit
- the modified positions ((xTnb>>4) ⁇ 4), (yTnb
- the motion information of the prediction block located at >>4) ⁇ 4) may be used for the temporal merge candidate.
- the predetermined storage unit is an 8x8 sample unit
- the coordinates of the temporal neighboring block are (xTnb, yTnb)
- the modified positions ((xTnb>>3) ⁇ 3), (yTnb> Motion information of the prediction block located at >3) ⁇ 3)) may be used for the temporal merge candidate.
- motion information of the current block may be derived in the same manner as in the case where the merge mode is applied previously.
- the skip mode when the skip mode is applied, the residual signal for the corresponding block is omitted, and thus prediction samples can be directly used as reconstructed samples.
- a spatial candidate block for motion vector prediction may be first searched and inserted into a prediction candidate list (S800). Thereafter, according to an embodiment, it may be determined whether the number of spatial candidate blocks is less than 2 (S810). For example, in an embodiment, when the number of spatial candidate blocks is less than 2, a temporal candidate block may be searched for and additionally inserted into the prediction candidate list (S820), and if a temporal candidate block is not available, a zero motion vector is used. Can be (S830). That is, a zero motion vector may be additionally inserted into the prediction candidate list. Thereafter, according to an embodiment, the configuration of the preliminary candidate list may be terminated (S840). Alternatively, according to an embodiment, when the number of spatial candidate blocks is not less than 2, the configuration of the preliminary candidate list may be terminated (S840).
- the preliminary candidate list may represent an MVP candidate list.
- the abs_mvd_greater0_flag syntax element may indicate information about whether the difference (MVD) is greater than 0, and the abs_mvd_greater1_flag syntax element may indicate information about whether the difference (MVD) is greater than 1.
- the abs_mvd_minus2 syntax element may indicate information about a value equal to -2 to the difference (MVD)
- the mvd_sign_flag syntax element may indicate information about the sign of the difference (MVD).
- [0] of each syntax element may indicate information about L0, and [1] may indicate information about L1.
- a bi-prediction signal ie, bi-prediction samples
- L0 prediction samples L0 prediction samples
- L1 prediction samples L1 prediction samples
- the bi-prediction samples were derived as an average of L0 prediction samples based on the L0 reference picture and MVL0, and L1 prediction samples based on the L1 reference picture and MVL1.
- the pair prediction it is possible to derive the pair prediction signal (the pair prediction samples) through the weighted average of the L0 prediction signal and the L1 prediction signal as follows.
- an MV derivation method considering a long-term reference picture in a motion vector scaling process of a temporal motion candidate (temporal merge candidate, or temporal mvp candidate) Is being used.
- the temporal motion candidate may correspond to mvCol (mvLXCol).
- the temporal motion candidate may be called TMVP.
- FIG. 12 shows a process of inducing MVD of MMVD according to another embodiment of the present document.
- the method illustrated in FIG. 12 may be for a block to which bidirectional prediction is applied.
- the MVD derivation method may vary depending on whether a reference picture referenced by a current picture (or a current slice or a current block) is a long-term reference picture (LTRP) or a short-term reference picture (STRP). have.
- LTRP long-term reference picture
- STP short-term reference picture
- a part of the standard document according to the present embodiment may be described as shown in the following table.
- the motion vector uses the value of the collocated motion vector as it is, but in this case, MV is 0 in MMVD and SMVD. Can be set to At this time, TMVP also sets MV to 0 without further induction.
- SMVD may be limited when referring to a long term reference picture (LTRP).
- LTRP long term reference picture
- a long-term reference picture may be excluded in a reference picture checking process. Accordingly, other reference pictures (eg, not a long-term reference picture) may be considered for SMVD.
- SMVD when the reference picture closest to the current picture is a long-term reference picture, SMVD may not be performed. For example, even if short-term reference pictures are included in the reference picture list, if the reference picture closest to the current picture is a long-term reference picture, SMVD may not be performed.
- FIG. 18 is a flowchart illustrating a method of deriving an MMVD according to an embodiment of the present document.
- the flowchart of FIG. 18 may show a method of deriving an MMVD according to Embodiment Y described above.
- a reference picture type is a long-term reference picture
- conditions for comparing POC differences may be removed, and an anchor MVD used for a mirroring procedure may be fixed to L0 MVD.
- the following table describes a method of deriving MMVDs according to Example Y of Table 30 in a standard document format.
- the MMVD derivation procedure may be limited in all cases. Conditions comparing POC differences for all cases can be removed and the anchor MVD used for the mirroring or scaling procedure can be fixed to L0 MVD.
- the encoding apparatus derives an inter prediction mode for a current block in a current picture (S2000).
- the inter prediction mode may include the above-described merge mode, AMVP mode (mode using motion vector predictor candidates), MMVD, and SMVD.
- the encoding apparatus derives reference pictures for the inter prediction mode (S2010).
- the reference pictures may be included in reference picture list 0 (or L0, reference picture list L0) or reference picture list 1 (or L1, reference picture list L1).
- the encoding apparatus may configure reference picture lists for each slice included in the current picture.
- the encoding apparatus may construct a motion vector predictor candidate list and derive a motion vector predictor based on the list.
- the encoding apparatus may derive motion vectors based on symmetric MVDs and the motion vector predictors.
- the encoding device generates prediction related information including the inter prediction mode (S2040).
- the prediction related information may include information about MMVD, information about SMVD, and the like.
- the encoding apparatus encodes the video/video information including the prediction related information and the residual information (S2060).
- the encoded image/video information may be output in the form of a bitstream.
- the bitstream may be transmitted to a decoding device through a network or a (digital) storage medium.
- the prediction related information may include inter prediction type information indicating whether bi-prediction is applied to the current block in the current picture.
- the prediction related information may include symmetric motion vector difference reference flag information indicating whether to apply a symmetric motion vector difference reference.
- the reference pictures may include short-term reference pictures. Symmetric motion vector difference reference indices may be derived from reference indices indicating short-term reference pictures based on the symmetric motion vector difference reference flag information.
- the motion information may include motion vectors for the current block and the symmetric motion vector difference reference indices.
- the prediction samples may be generated based on the motion vectors and the symmetric motion vector difference reference indices.
- the symmetric motion vector difference reference indices may be derived based on POC differences between each of the short-term reference pictures and the current picture.
- a POC difference between a current picture and a previous reference picture from the current picture may be greater than zero.
- a POC difference between a current picture and a next reference picture from the current picture may be less than zero.
- the encoding apparatus may configure a reference picture list L0 (or reference picture list 0) for L0 prediction and a reference picture list L1 (or reference picture list 0) for L1 prediction.
- the short term reference pictures may include a short term reference picture L0 included in the reference picture list L0 and a short term reference picture L1 included in the reference picture list L1.
- the POC differences may include a first POC difference between the short term reference picture L0 and the current picture, and a second POC difference between the short term reference picture L1 and the current picture.
- the symmetric motion vector difference reference indices may include a symmetric motion vector difference reference index L0 and a symmetric motion vector difference reference index L1.
- the symmetric motion vector difference reference index L0 may be derived based on the first POC difference.
- the symmetric motion vector difference reference index L1 may be derived based on the second POC difference.
- the encoding apparatus may configure a reference picture list L0 for L0 prediction.
- the short term reference pictures may include a first short term reference picture L0 and a second short term reference picture L0 included in the reference picture list L0.
- the POC differences may include a third POC difference between the first short-term reference picture L0 and the current picture, and a fourth POC difference between the second short-term reference picture L0 and the current picture.
- the symmetric motion vector difference reference indices may include a symmetric motion vector difference reference index L0.
- a reference picture index indicating the first short-term reference picture L0 based on the comparison between the third and fourth POC differences may be used as the symmetric motion vector difference reference index L0.
- the reference picture index indicating the first short term reference picture L0 may be used as the symmetric motion vector difference reference index L0.
- the video information may include information on motion vector differences (MVD).
- the motion information may include motion vectors (MVs).
- MVDL0 for L0 prediction may be derived based on the information on the MVD.
- the MVs may be derived based on the MVDL0 and the MVDL1.
- the size of MVDL1 may be the same as the size of MVDL0.
- the code of MVDL1 may be opposite to that of MVDL0.
- FIG. 22 and 23 schematically illustrate an example of a video/video decoding method and related components according to the embodiment(s) of this document.
- the method disclosed in FIG. 22 may be performed by the decoding apparatus disclosed in FIG. 3. Specifically, for example, S2200 of FIG. 22 may be performed by the entropy decoding unit 310 of the decoding device, S2210 to S2230 may be performed by the prediction unit 330 of the decoding device, and S2240 It may be performed by the residual processing unit 320 of the decoding device, and S2250 may be performed by the adding unit 340 of the decoding device.
- the method disclosed in FIG. 22 may include the embodiments described above in this document.
- the decoding apparatus receives/acquires image/video information (S2200).
- the decoding apparatus may receive/acquire the image/video information through a bitstream.
- the image/video information may include prediction related information (including prediction mode information) and residual information.
- the prediction related information may include information about MMVD, information about SMVD, and the like.
- the image/video information may include various information according to an embodiment of the present document.
- the image/video information may include information described with reference to FIGS. 1 to 19 and/or information disclosed in at least one of Tables 1 to 33 described above.
- the decoding apparatus derives motion information for prediction of the current block based on the inter prediction mode (S2220).
- the motion information may include reference picture indexes and motion vectors.
- the decoding apparatus may derive reference indices for SMVD.
- Reference indexes for SMVD may indicate reference pictures for SMVD application.
- Reference indexes for SMVD may include a reference index L0 (RefIdxSumL0) and a reference index L1 (RefIdxSumL1).
- the decoding apparatus may construct a motion vector predictor candidate list and derive a motion vector predictor based on the list.
- the decoding apparatus may derive motion vectors based on symmetric MVDs and the motion vector predictors.
- the decoding apparatus generates prediction samples based on the motion information (S2230).
- the decoding apparatus may generate the prediction samples based on motion vectors and reference picture indices included in the motion information.
- the prediction samples may be generated based on blocks (or samples) indicated by the motion vector among blocks (or samples) in the reference pictures indicated by the reference picture indices.
- the prediction related information may include inter prediction type information indicating whether bi-prediction is applied to the current block in the current picture.
- the prediction related information may include symmetric motion vector difference reference flag information indicating whether to apply a symmetric motion vector difference reference.
- the reference pictures may include short-term reference pictures. Symmetric motion vector difference reference indices may be derived from reference indices indicating short-term reference pictures based on the symmetric motion vector difference reference flag information.
- the motion information may include motion vectors for the current block and the symmetric motion vector difference reference indices.
- the prediction samples may be generated based on the motion vectors and the symmetric motion vector difference reference indices.
- the reference picture index indicating the first short term reference picture L0 may be used as the symmetric motion vector difference reference index L0.
- the video information may include information on motion vector differences (MVD).
- the motion information may include motion vectors (MVs).
- MVDL0 for L0 prediction may be derived based on the information on the MVD.
- the MVs may be derived based on the MVDL0 and the MVDL1.
- the size of MVDL1 may be the same as the size of MVDL0.
- the code of MVDL1 may be opposite to that of MVDL0.
- the above-described method may be implemented as a module (process, function, etc.) performing the above-described functions.
- the modules are stored in memory and can be executed by the processor.
- the memory may be inside or outside the processor, and may be connected to the processor by various well-known means.
- the processor may include an application-specific integrated circuit (ASIC), another chipset, a logic circuit, and/or a data processing device.
- the memory may include read-only memory (ROM), random access memory (RAM), flash memory, memory card, storage medium, and/or other storage device. That is, the embodiments described in this document may be implemented and performed on a processor, microprocessor, controller, or chip.
- the functional units illustrated in each drawing may be implemented and executed on a computer, processor, microprocessor, controller, or chip. In this case, information for implementation (ex. information on instructions) or an algorithm may be stored in a digital storage medium.
- embodiment(s) of this document may be implemented as a computer program product by program code, and the program code may be executed in a computer according to the embodiment(s) of this document.
- the program code may be stored on a carrier readable by a computer.
- the encoding server serves to generate a bitstream by compressing content input from multimedia input devices such as smartphones, cameras, camcorders, etc. into digital data, and transmits it to the streaming server.
- multimedia input devices such as smartphones, cameras, camcorders, etc. directly generate bitstreams
- the encoding server may be omitted.
- Each server in the content streaming system may be operated as a distributed server, and in this case, data received from each server may be distributedly processed.
- the claims set forth herein may be combined in a variety of ways.
- the technical features of the method claims of the present specification may be combined to be implemented as a device, and the technical features of the device claims of the present specification may be combined to be implemented by a method.
- the technical characteristics of the method claim of the present specification and the technical characteristics of the device claim may be combined to be implemented as a device, and the technical characteristics of the method claim of the present specification and the technical characteristics of the device claim may be combined to be implemented by a method.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
Claims (17)
- 디코딩 장치에 의하여 수행되는 영상 디코딩 방법에 있어서,비트스트림으로부터 예측 관련 정보 및 레지듀얼 정보를 포함하는 영상 정보를 수신하는 단계;상기 예측 관련 정보를 기반으로 현재 픽처 내 현재 블록을 위해 인터 예측 모드를 도출하는 단계;상기 인터 예측 모드를 기반으로 상기 현재 블록에 대한 움직임 정보를 도출하는 단계;상기 움직임 정보를 기반으로 예측 샘플들을 생성하는 단계;상기 레지듀얼 정보를 기반으로 레지듀얼 샘플들을 생성하는 단계; 및상기 예측 샘플들 및 상기 레지듀얼 샘플들을 기반으로 현재 픽처의 복원 샘플들을 생성하는 단계를 포함하되,상기 예측 관련 정보는 상기 현재 블록에 쌍예측이 적용되는지 여부를 나타내는 인터 예측 타입 정보를 포함하고,상기 인터 예측 타입 정보를 기반으로 상기 예측 관련 정보는 대칭 움직임 벡터 차분 참조의 적용 여부를 나타내는 대칭 움직임 벡터 차분 참조 플래그 정보를 더 포함하고,상기 대칭 움직임 벡터 차분 참조 플래그 정보를 기반으로, 숏텀 참조 픽처들을 가리키는 참조 인덱스들로부터 대칭 움직임 벡터 차분 참조 인덱스들이 도출되고,상기 움직임 정보는 상기 현재 블록에 대한 움직임 벡터들 및 상기 대칭 움직임 벡터 차분 참조 인덱스들을 포함하고,상기 움직임 벡터들 및 상기 대칭 움직임 벡터 차분 참조 인덱스들을 기반으로 상기 예측 샘플들이 생성되는 것을 특징으로 하는, 영상 디코딩 방법.
- 제1항에 있어서,상기 숏텀 참조 픽처들 각각과 상기 현재 픽처 간의 POC(picture order count) 차분들을 기반으로 상기 대칭 움직임 벡터 차분 참조 인덱스들이 도출되는 것을 특징으로 하는, 영상 디코딩 방법.
- 제2항에 있어서,L0 예측을 위한 참조 픽처 리스트 L0 및 L1 예측을 위한 참조 픽처 리스트 L1을 구성하는 단계를 더 포함하되,상기 숏텀 참조 픽처들은 상기 참조 픽처 리스트 L0에 포함된 숏텀 참조 픽처 L0 및 상기 참조 픽처 리스트 L1에 포함된 숏텀 참조 픽처 L1을 포함하고,상기 POC 차분들은 상기 숏텀 참조 픽처 L0와 상기 현재 픽처 간의 제1 POC 차분, 및 상기 숏텀 참조 픽처 L1와 상기 현재 픽처 간의 제2 POC 차분을 포함하고,상기 대칭 움직임 벡터 차분 참조 인덱스들은 대칭 움직임 벡터 차분 참조 인덱스 L0 및 대칭 움직임 벡터 차분 참조 인덱스 L1을 포함하고,상기 제1 POC 차분을 기반으로 상기 대칭 움직임 벡터 차분 참조 인덱스 L0이 도출되고,상기 제2 POC 차분을 기반으로 상기 대칭 움직임 벡터 차분 참조 인덱스 L1이 도출되는 것을 특징으로 하는, 영상 디코딩 방법.
- 제3항에 있어서,상기 제1 POC 차분은 상기 제2 POC 차분과 동일한 것을 특징으로 하는, 영상 디코딩 방법.
- 제2항에 있어서,L0 예측을 위한 참조 픽처 리스트 L0를 구성하는 단계를 더 포함하되,상기 숏텀 참조 픽처들은 상기 참조 픽처 리스트 L0에 포함된 제1 숏텀 참조 픽처 L0 및 제2 숏텀 참조 픽처 L0를 포함하고,상기 POC 차분들은 상기 제1 숏텀 참조 픽처 L0와 상기 현재 픽처 간의 제3 POC 차분, 및 상기 제2 숏텀 참조 픽처 L0와 상기 현재 픽처 간의 제4 POC 차분을 포함하고,상기 대칭 움직임 벡터 차분 참조 인덱스들은 대칭 움직임 벡터 차분 참조 인덱스 L0를 포함하고,상기 제3 및 제 4 POC 차분들 간의 비교를 기반으로 상기 제1 숏텀 참조 픽처 L0를 가리키는 참조 픽처 인덱스가 상기 대칭 움직임 벡터 차분 참조 인덱스 L0로서 사용되는 것을 특징으로 하는, 영상 디코딩 방법.
- 제5항에 있어서,상기 제3 POC 차분이 상기 제4 POC 차분보다 더 작은 경우, 상기 제1 숏텀 참조 픽처 L0를 가리키는 상기 참조 픽처 인덱스가 상기 대칭 움직임 벡터 차분 참조 인덱스 L0로서 사용되는 것을 특징으로 하는, 영상 디코딩 방법.
- 제1항에 있어서,상기 영상 정보는 MVD(motion vector differences)에 관한 정보를 포함하고,상기 움직임 정보는 MV(motion vector)들을 포함하고,상기 MVD에 관한 정보를 기반으로 L0 예측을 위한 MVDL0가 도출되고,상기 MVDL0를 기반으로 L1 예측을 위한 MVDL1이 도출되고,상기 MV들은 상기 MVDL0 및 상기 MVDL1을 기반으로 도출되는 것을 특징으로 하는, 영상 디코딩 방법.
- 제7항에 있어서,상기 MVDL1의 크기는 상기 MVDL0의 크기와 동일하고,상기 MVDL1의 부호는 상기 MVDL0의 부호와 반대인 것을 특징으로 하는, 영상 디코딩 방법.
- 인코딩 장치에 의하여 수행되는 영상 인코딩 방법에 있어서,현재 픽처 내 현재 블록을 위해 인터 예측 모드를 도출하는 단계;상기 인터 예측 모드를 위한 참조 픽처들을 도출하는 단계;상기 인터 예측 모드를 기반으로 상기 현재 블록의 예측을 위한 움직임 정보를 도출하는 단계;상기 움직임 정보를 기반으로 예측 샘플들을 생성하는 단계;상기 인터 예측 모드를 포함하는 예측 관련 정보를 생성하는 단계;상기 예측 샘플들을 기반으로 레지듀얼 정보를 생성하는 단계;상기 예측 관련 정보 및 상기 레지듀얼 정보를 포함하는 영상 정보를 인코딩하는 단계를 포함하되,상기 예측 관련 정보는 상기 현재 블록에 쌍예측이 적용되는지 여부를 나타내는 인터 예측 타입 정보를 포함하고,상기 인터 예측 타입 정보를 기반으로 상기 예측 관련 정보는 대칭 움직임 벡터 차분 참조의 적용 여부를 나타내는 대칭 움직임 벡터 차분 참조 플래그 정보를 포함하고,상기 참조 픽처들은 숏텀 참조 픽처들을 포함하고,상기 대칭 움직임 벡터 차분 참조 플래그 정보를 기반으로, 숏텀 참조 픽처들을 가리키는 참조 인덱스들로부터 대칭 움직임 벡터 차분 참조 인덱스들이 도출되고,상기 움직임 정보는 상기 현재 블록에 대한 움직임 벡터들 및 상기 대칭 움직임 벡터 차분 참조 인덱스들을 포함하고,상기 움직임 벡터들 및 상기 대칭 움직임 벡터 차분 참조 인덱스들을 기반으로 상기 예측 샘플들이 생성되는 것을 특징으로 하는, 영상 인코딩 방법.
- 제9항에 있어서,상기 숏텀 참조 픽처들 각각과 상기 현재 픽처 간의 POC(picture order count) 차분들을 기반으로 상기 대칭 움직임 벡터 차분 참조 인덱스들이 도출되는 것을 특징으로 하는, 영상 인코딩 방법.
- 제10항에 있어서,L0 예측을 위한 참조 픽처 리스트 L0 및 L1 예측을 위한 참조 픽처 리스트 L1을 구성하는 단계를 더 포함하되,상기 숏텀 참조 픽처들은 상기 참조 픽처 리스트 L0에 포함된 숏텀 참조 픽처 L0 및 상기 참조 픽처 리스트 L1에 포함된 숏텀 참조 픽처 L1을 포함하고,상기 POC 차분들은 상기 숏텀 참조 픽처 L0와 상기 현재 픽처 간의 제1 POC 차분, 및 상기 숏텀 참조 픽처 L1와 상기 현재 픽처 간의 제2 POC 차분을 포함하고,상기 대칭 움직임 벡터 차분 참조 인덱스들은 대칭 움직임 벡터 차분 참조 인덱스 L0 및 대칭 움직임 벡터 차분 참조 인덱스 L1을 포함하고,상기 제1 POC 차분을 기반으로 상기 대칭 움직임 벡터 차분 참조 인덱스 L0이 도출되고,상기 제2 POC 차분을 기반으로 상기 대칭 움직임 벡터 차분 참조 인덱스 L1이 도출되는 것을 특징으로 하는, 영상 인코딩 방법.
- 제11항에 있어서,상기 제1 POC 차분은 상기 제2 POC 차분과 동일한 것을 특징으로 하는, 영상 인코딩 방법.
- 제10항에 있어서,L0 예측을 위한 참조 픽처 리스트 L0를 구성하는 단계를 더 포함하되,상기 숏텀 참조 픽처들은 상기 참조 픽처 리스트 L0에 포함된 제1 숏텀 참조 픽처 L0 및 제2 숏텀 참조 픽처 L0를 포함하고,상기 POC 차분들은 상기 제1 숏텀 참조 픽처 L0와 상기 현재 픽처 간의 제3 POC 차분, 및 상기 제2 숏텀 참조 픽처 L0와 상기 현재 픽처 간의 제4 POC 차분을 포함하고,상기 대칭 움직임 벡터 차분 참조 인덱스들은 대칭 움직임 벡터 차분 참조 인덱스 L0를 포함하고,상기 제3 및 제 4 POC 차분들 간의 비교를 기반으로 상기 제1 숏텀 참조 픽처 L0를 가리키는 참조 픽처 인덱스가 상기 대칭 움직임 벡터 차분 참조 인덱스 L0로서 사용되는 것을 특징으로 하는, 영상 인코딩 방법.
- 제13항에 있어서,상기 제3 POC 차분이 상기 제4 POC 차분보다 더 작은 경우, 상기 제1 숏텀 참조 픽처 L0를 가리키는 상기 참조 픽처 인덱스가 상기 대칭 움직임 벡터 차분 참조 인덱스 L0로서 사용되는 것을 특징으로 하는, 영상 인코딩 방법.
- 제9항에 있어서,상기 영상 정보는 MVD(motion vector differences)에 관한 정보를 포함하고,상기 움직임 정보는 MV(motion vector)들을 포함하고,상기 MVD에 관한 정보를 기반으로 L0 예측을 위한 MVDL0가 도출되고,상기 MVDL0를 기반으로 L1 예측을 위한 MVDL1이 도출되고,상기 MV들은 상기 MVDL0 및 상기 MVDL1을 기반으로 도출되는 것을 특징으로 하는, 영상 영상 인코딩 방법.
- 제15항에 있어서,상기 MVDL1의 크기는 상기 MVDL0의 크기와 동일하고,상기 MVDL1의 부호는 상기 MVDL0의 부호와 반대인 것을 특징으로 하는, 영상 인코딩 방법.
- 영상 디코딩 장치가 영상 디코딩 방법을 수행하도록 야기하는 인코딩된 정보를 저장하는 컴퓨터 판독 가능한 저장 매체에 있어서, 상기 영상 디코딩 방법은:비트스트림으로부터 예측 관련 정보 및 레지듀얼 정보를 포함하는 영상 정보를 수신하는 단계;상기 예측 관련 정보를 기반으로 현재 픽처 내 현재 블록을 위해 인터 예측 모드를 도출하는 단계;상기 인터 예측 모드를 기반으로 상기 현재 블록에 대한 움직임 정보를 도출하는 단계;상기 움직임 정보를 기반으로 예측 샘플들을 생성하는 단계;상기 레지듀얼 정보를 기반으로 레지듀얼 샘플들을 생성하는 단계; 및상기 예측 샘플들 및 상기 레지듀얼 샘플들을 기반으로 현재 픽처의 복원 샘플들을 생성하는 단계를 포함하되,상기 예측 관련 정보는 상기 현재 블록에 쌍예측이 적용되는지 여부를 나타내는 인터 예측 타입 정보를 포함하고,상기 인터 예측 타입 정보를 기반으로 상기 예측 관련 정보는 대칭 움직임 벡터 차분 참조의 적용 여부를 나타내는 대칭 움직임 벡터 차분 참조 플래그 정보를 더 포함하고,상기 대칭 움직임 벡터 차분 참조 플래그 정보를 기반으로, 숏텀 참조 픽처들을 가리키는 참조 인덱스들로부터 대칭 움직임 벡터 차분 참조 인덱스들이 도출되고,상기 움직임 정보는 상기 현재 블록에 대한 움직임 벡터들 및 상기 대칭 움직임 벡터 차분 참조 인덱스들을 포함하고,상기 움직임 벡터들 및 상기 대칭 움직임 벡터 차분 참조 인덱스들을 기반으로 상기 예측 샘플들이 생성되는 것을 특징으로 하는, 컴퓨터 판독 가능한 저장 매체.
Priority Applications (9)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021576744A JP7342153B2 (ja) | 2019-06-24 | 2020-06-24 | 動きベクトルを用いた画像コーディング方法および装置 |
KR1020247005775A KR20240029111A (ko) | 2019-06-24 | 2020-06-24 | 움직임 벡터를 이용한 영상 코딩 방법 및 장치 |
KR1020217041544A KR102640263B1 (ko) | 2019-06-24 | 2020-06-24 | 움직임 벡터를 이용한 영상 코딩 방법 및 장치 |
US17/559,417 US11451821B2 (en) | 2019-06-24 | 2021-12-22 | Image coding method and apparatus using motion vector |
US17/846,499 US11659199B2 (en) | 2019-06-24 | 2022-06-22 | Image coding method and apparatus using motion vector |
US18/132,225 US11991385B2 (en) | 2019-06-24 | 2023-04-07 | Image coding method and apparatus using motion vector |
JP2023139806A JP7513818B2 (ja) | 2019-06-24 | 2023-08-30 | 動きベクトルを用いた画像コーディング方法および装置 |
US18/633,276 US20240357154A1 (en) | 2019-06-24 | 2024-04-11 | Image coding method and apparatus using motion vector |
JP2024103785A JP2024114873A (ja) | 2019-06-24 | 2024-06-27 | 動きベクトルを用いた画像コーディング方法および装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201962865963P | 2019-06-24 | 2019-06-24 | |
US62/865,963 | 2019-06-24 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/559,417 Continuation US11451821B2 (en) | 2019-06-24 | 2021-12-22 | Image coding method and apparatus using motion vector |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020262918A1 true WO2020262918A1 (ko) | 2020-12-30 |
Family
ID=74061809
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2020/008137 WO2020262918A1 (ko) | 2019-06-24 | 2020-06-24 | 움직임 벡터를 이용한 영상 코딩 방법 및 장치 |
Country Status (4)
Country | Link |
---|---|
US (4) | US11451821B2 (ko) |
JP (3) | JP7342153B2 (ko) |
KR (2) | KR102640263B1 (ko) |
WO (1) | WO2020262918A1 (ko) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150078456A1 (en) * | 2013-07-31 | 2015-03-19 | Nokia Corporation | Method and apparatus for video coding and decoding |
KR20170021302A (ko) * | 2014-06-19 | 2017-02-27 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | 통합된 인트라 블록 카피 및 인터 예측 모드 |
KR101941304B1 (ko) * | 2011-10-27 | 2019-01-22 | 선 페이턴트 트러스트 | 화상 부호화 방법, 화상 복호 방법, 화상 부호화 장치 및 화상 복호 장치 |
KR20190032334A (ko) * | 2011-10-28 | 2019-03-27 | 삼성전자주식회사 | 인터 예측 방법 및 그 장치, 움직임 보상 방법 및 그 장치 |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10595035B2 (en) * | 2017-03-22 | 2020-03-17 | Qualcomm Incorporated | Constraining motion vector information derived by decoder-side motion vector derivation |
WO2019027286A1 (ko) * | 2017-08-03 | 2019-02-07 | 엘지전자 주식회사 | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 |
CN118264800A (zh) * | 2017-12-14 | 2024-06-28 | Lg电子株式会社 | 解码和编码设备及发送用于图像的数据的设备 |
EP3797516A1 (en) * | 2018-06-29 | 2021-03-31 | Beijing Bytedance Network Technology Co. Ltd. | Interaction between lut and amvp |
CN118433391A (zh) | 2018-12-21 | 2024-08-02 | Vid拓展公司 | 对称运动矢量差译码 |
CN117119181B (zh) | 2019-02-20 | 2024-03-26 | 北京达佳互联信息技术有限公司 | 用于视频编码和视频解码的方法、计算设备和存储介质 |
KR102612802B1 (ko) * | 2019-03-05 | 2023-12-13 | 엘지전자 주식회사 | 인터 예측을 위한 비디오 신호의 처리 방법 및 장치 |
JP6867611B2 (ja) * | 2019-03-11 | 2021-04-28 | Kddi株式会社 | 画像復号装置、画像復号方法及びプログラム |
US12047582B2 (en) * | 2019-03-24 | 2024-07-23 | Lg Electronics Inc. | Image encoding/decoding method and device using symmetric motion vector difference (SMVD), and method for transmitting bitstream |
WO2020221256A1 (en) | 2019-04-28 | 2020-11-05 | Beijing Bytedance Network Technology Co., Ltd. | Symmetric motion vector difference coding |
CN113767637B (zh) | 2019-04-28 | 2023-09-22 | 北京字节跳动网络技术有限公司 | 对称运动矢量差编解码 |
JP7335365B2 (ja) * | 2019-06-24 | 2023-08-29 | エルジー エレクトロニクス インコーポレイティド | 動きベクトル差分を利用した映像コーディング方法及び装置 |
US11190789B2 (en) * | 2019-06-30 | 2021-11-30 | Tencent America LLC | Method and apparatus for video coding using inter-prediction mode signaling to dertermine motion vectors |
-
2020
- 2020-06-24 JP JP2021576744A patent/JP7342153B2/ja active Active
- 2020-06-24 WO PCT/KR2020/008137 patent/WO2020262918A1/ko active Application Filing
- 2020-06-24 KR KR1020217041544A patent/KR102640263B1/ko active IP Right Grant
- 2020-06-24 KR KR1020247005775A patent/KR20240029111A/ko not_active Application Discontinuation
-
2021
- 2021-12-22 US US17/559,417 patent/US11451821B2/en active Active
-
2022
- 2022-06-22 US US17/846,499 patent/US11659199B2/en active Active
-
2023
- 2023-04-07 US US18/132,225 patent/US11991385B2/en active Active
- 2023-08-30 JP JP2023139806A patent/JP7513818B2/ja active Active
-
2024
- 2024-04-11 US US18/633,276 patent/US20240357154A1/en active Pending
- 2024-06-27 JP JP2024103785A patent/JP2024114873A/ja active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101941304B1 (ko) * | 2011-10-27 | 2019-01-22 | 선 페이턴트 트러스트 | 화상 부호화 방법, 화상 복호 방법, 화상 부호화 장치 및 화상 복호 장치 |
KR20190032334A (ko) * | 2011-10-28 | 2019-03-27 | 삼성전자주식회사 | 인터 예측 방법 및 그 장치, 움직임 보상 방법 및 그 장치 |
US20150078456A1 (en) * | 2013-07-31 | 2015-03-19 | Nokia Corporation | Method and apparatus for video coding and decoding |
KR20170021302A (ko) * | 2014-06-19 | 2017-02-27 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | 통합된 인트라 블록 카피 및 인터 예측 모드 |
Non-Patent Citations (1)
Title |
---|
BENJAMIN BROSS: "Versatile Video Coding (Draft 5", JVET-N1001-V8, JOINT VIDEO EXPERTS TEAM (JVET) OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, 14TH MEETING: GENEVA, CH, vol. 55, 11 June 2019 (2019-06-11), pages 1 - 385, XP030205561 * |
Also Published As
Publication number | Publication date |
---|---|
US20220385936A1 (en) | 2022-12-01 |
US11451821B2 (en) | 2022-09-20 |
KR102640263B1 (ko) | 2024-02-23 |
JP2022538841A (ja) | 2022-09-06 |
KR20240029111A (ko) | 2024-03-05 |
US20220116651A1 (en) | 2022-04-14 |
US11991385B2 (en) | 2024-05-21 |
JP2024114873A (ja) | 2024-08-23 |
KR20210158398A (ko) | 2021-12-30 |
US11659199B2 (en) | 2023-05-23 |
JP7513818B2 (ja) | 2024-07-09 |
JP2023155412A (ja) | 2023-10-20 |
JP7342153B2 (ja) | 2023-09-11 |
US20240357154A1 (en) | 2024-10-24 |
US20230283801A1 (en) | 2023-09-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020171444A1 (ko) | Dmvr 기반의 인터 예측 방법 및 장치 | |
WO2021040481A1 (ko) | 크로스 컴포넌트 필터링 기반 영상 코딩 장치 및 방법 | |
WO2020189893A1 (ko) | Bdof 기반의 인터 예측 방법 및 장치 | |
WO2020184848A1 (ko) | Dmvr 기반의 인터 예측 방법 및 장치 | |
WO2021040479A1 (ko) | 필터링 기반 영상 코딩 장치 및 방법 | |
WO2020184847A1 (ko) | Dmvr 및 bdof 기반의 인터 예측 방법 및 장치 | |
WO2020256390A1 (ko) | 블록 사이즈를 기반으로 bdpcm 을 수행하는 영상 디코딩 방법 및 그 장치 | |
WO2020262919A1 (ko) | 움직임 벡터 차분들을 이용한 영상 코딩 방법 및 장치 | |
WO2020055161A1 (ko) | 영상 코딩 시스템에서 서브 블록 단위의 움직임 예측에 기반한 영상 디코딩 방법 및 장치 | |
WO2020262931A1 (ko) | 비디오/영상 코딩 시스템에서 머지 데이터 신택스의 시그널링 방법 및 장치 | |
WO2020180129A1 (ko) | 인터 예측을 위한 비디오 신호의 처리 방법 및 장치 | |
WO2020251323A1 (ko) | 인터 예측 기반 영상 코딩 방법 및 장치 | |
WO2020251257A1 (ko) | 예측 샘플을 생성하기 위한 가중치 인덱스 정보를 도출하는 영상 디코딩 방법 및 그 장치 | |
WO2020256454A1 (ko) | 현재 블록에 대하여 최종적으로 예측 모드를 선택하지 못하는 경우 인터 예측을 수행하는 영상 디코딩 방법 및 그 장치 | |
WO2020256455A1 (ko) | 디폴트 머지 모드에 기반하여 예측 샘플을 도출하는 영상 디코딩 방법 및 그 장치 | |
WO2021141227A1 (ko) | 영상 디코딩 방법 및 그 장치 | |
WO2021172881A1 (ko) | 인터 예측에 기반한 영상 인코딩/디코딩 방법 및 장치, 그리고 비트스트림을 저장한 기록 매체 | |
WO2020251324A1 (ko) | 움직임 벡터 차분들을 이용한 영상 코딩 방법 및 장치 | |
WO2020262930A1 (ko) | 머지 데이터 신택스에서 중복적인 신택스의 제거 방법 및 장치 | |
WO2020251258A1 (ko) | 쌍 예측이 적용되는 경우 가중 평균을 위한 가중치 인덱스 정보를 도출하는 영상 디코딩 방법 및 그 장치 | |
WO2020262902A1 (ko) | 쌍 예측을 적용하는 영상 디코딩 방법 및 그 장치 | |
WO2021040410A1 (ko) | 레지듀얼 코딩에 대한 영상 디코딩 방법 및 그 장치 | |
WO2020251259A1 (ko) | 쌍 예측(bi-prediction)을 위한 가중치 인덱스 정보를 도출하는 영상 디코딩 방법 및 그 장치 | |
WO2020251268A1 (ko) | 크로마 성분에 대한 영상 디코딩 방법 및 그 장치 | |
WO2020009447A1 (ko) | 인터 예측 모드 기반 영상 처리 방법 및 이를 위한 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20831021 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20217041544 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2021576744 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20831021 Country of ref document: EP Kind code of ref document: A1 |