WO2019027286A1 - 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 - Google Patents
어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 Download PDFInfo
- Publication number
- WO2019027286A1 WO2019027286A1 PCT/KR2018/008845 KR2018008845W WO2019027286A1 WO 2019027286 A1 WO2019027286 A1 WO 2019027286A1 KR 2018008845 W KR2018008845 W KR 2018008845W WO 2019027286 A1 WO2019027286 A1 WO 2019027286A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- mode
- affine
- current block
- motion vector
- flag
- Prior art date
Links
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 title claims abstract description 159
- 238000000034 method Methods 0.000 title claims abstract description 93
- 238000012545 processing Methods 0.000 title description 30
- 230000033001 locomotion Effects 0.000 claims abstract description 377
- 239000013598 vector Substances 0.000 claims abstract description 258
- 230000008569 process Effects 0.000 description 26
- 238000010586 diagram Methods 0.000 description 20
- 238000013139 quantization Methods 0.000 description 16
- 230000003044 adaptive effect Effects 0.000 description 12
- 238000001914 filtration Methods 0.000 description 12
- 230000002123 temporal effect Effects 0.000 description 9
- 238000006243 chemical reaction Methods 0.000 description 5
- 238000007906 compression Methods 0.000 description 5
- 239000013074 reference sample Substances 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 230000006835 compression Effects 0.000 description 4
- 230000011218 segmentation Effects 0.000 description 4
- 238000000638 solvent extraction Methods 0.000 description 4
- 238000013519 translation Methods 0.000 description 4
- 230000014616 translation Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 230000000903 blocking effect Effects 0.000 description 3
- 230000006866 deterioration Effects 0.000 description 3
- 238000005192 partition Methods 0.000 description 2
- 238000003672 processing method Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 239000013068 control sample Substances 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 239000000523 sample Substances 0.000 description 1
- 239000010454 slate Substances 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/44—Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/109—Selection of coding mode or of prediction mode among a plurality of temporal predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/13—Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
- H04N19/159—Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/174—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/537—Motion estimation other than block-based
- H04N19/54—Motion estimation other than block-based using feature points or meshes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/80—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
- H04N19/82—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
Definitions
- the present invention relates to a method and apparatus for encoding / decoding video signals, and more particularly to a method and apparatus for adaptively performing affine prediction.
- Compressive encoding refers to a series of signal processing techniques for transmitting digitized information over a communication line or for storing it in a form suitable for a storage medium.
- Media such as video, image, and audio can be subject to compression coding.
- a technique for performing compression coding on an image is referred to as video image compression.
- Next-generation video content will feature high spatial resolution, high frame rate, and high dimensionality of scene representation. Processing such content will result in a tremendous increase in terms of memory storage, memory access rate, and processing power.
- the present invention proposes a method of encoding and decoding a video signal more efficiently.
- the present invention can be applied to encoding or decoding in consideration of both the AF4 mode, which is a four parameter affine prediction mode using four parameters, and the AF6 mode, which is a six parameter affine prediction mode using six parameters .
- the present invention also proposes a method for adaptively determining (or selecting) an optimal coding mode based on at least one of an AF4 mode or an AF6 mode based on a block size.
- the present invention also proposes a method for adaptively determining (or selecting) an optimal coding mode based on at least one of an AF4 mode or an AF6 mode based on whether neighboring blocks are coded by affine prediction.
- the present invention provides a method for adaptively performing affine prediction based on a block size.
- the present invention also provides a method for adaptively performing affine prediction based on whether neighboring blocks are coded by affine prediction.
- the present invention also provides a method for adaptively determining (or selecting) an optimal coding mode based on at least one of an AF4 mode or an AF6 mode.
- the present invention provides a method for adaptively performing affine prediction based on whether at least one predetermined condition is satisfied, wherein the predetermined condition is a block size, a number of pixels of a block , The width of the block, the height of the block, and whether the neighboring block is coded by affine prediction.
- the predetermined condition is a block size, a number of pixels of a block , The width of the block, the height of the block, and whether the neighboring block is coded by affine prediction.
- the present invention provides a method for adaptively performing affine prediction, thereby improving the performance of affine prediction and reducing the complexity of affine prediction. Coding can be performed.
- FIG. 1 shows a schematic block diagram of an encoder in which the encoding of a video signal is performed, in which the present invention is applied.
- Fig. 2 shows a schematic block diagram of a decoder in which decoding of a video signal is performed, according to an embodiment to which the present invention is applied.
- FIG. 3 is a diagram for explaining a block division structure of a QT (QuadTree, hereinafter referred to as 'QT') to which the present invention can be applied.
- FIG. 4 is a diagram for explaining a BT (Binary Tree, hereinafter referred to as 'BT') block division structure to which the present invention can be applied.
- BT Binary Tree
- FIG. 5 is a diagram for explaining a block division structure of a TT (Ternary Tree) block according to an embodiment of the present invention.
- FIG. 6 is a diagram for explaining an AT (Asymmetric Tree) block partitioning structure to which the present invention can be applied.
- FIG. 7 is a diagram for explaining an inter prediction mode as an embodiment to which the present invention is applied.
- FIG. 8 is a diagram for explaining an affine motion model as an embodiment to which the present invention is applied.
- FIG. 9 is a diagram for explaining an affine motion prediction method using a control point motion vector according to an embodiment of the present invention. Referring to FIG.
- FIG. 10 is a flowchart illustrating a process of processing a video signal including a current block using an affine prediction mode according to an embodiment of the present invention.
- 11 is a flowchart of a method for adaptively determining an optimal coding mode based on at least one of an AF4 mode and an AF6 mode according to an embodiment (1-1) to which the present invention is applied.
- Fig. 12 shows a flowchart for performing adaptive decoding based on the AF4 mode or the AF6 mode as an embodiment (1-2) to which the present invention is applied.
- FIG. 13 shows an embodiment (1-3) to which the present invention is applied, showing a syntax structure for performing decoding based on the AF4 mode or the AF6 mode.
- FIG. 14 is an embodiment (2-1) to which the present invention is applied, in which adaptively determines an optimal coding mode among the motion vector prediction modes including the AF4 mode or the AF6 mode based on the condition A Fig.
- Fig. 15 shows a flowchart for performing adaptive decoding according to the AF4 mode or the AF6 mode based on the condition A (condition A) as an embodiment (2-2) to which the present invention is applied.
- FIG. 16 shows a syntax structure for performing decoding according to the AF4 mode or the AF6 mode on the basis of the condition A (condition A) according to the embodiment (2-3) to which the present invention is applied.
- FIG. 17 shows an embodiment (3-1) to which the present invention is applied.
- the motion vector prediction modes including the AF4 mode or the AF6 mode, based on at least one of the condition B (condition B) and the condition C
- a coding mode adaptively determining an optimum coding mode among the plurality of coding modes.
- FIG. 18 is an embodiment (3-2) to which the present invention is applied and adaptively decodes according to the AF4 mode or the AF6 mode based on at least one of the condition B (condition B) and the condition C (condition C) Fig.
- Fig. 19 is a schematic block diagram of a syntax structure (or a decoder) according to an embodiment (3-3) to which the present invention is applied and which performs decoding in accordance with at least one of the condition B (condition B) or the condition C .
- FIG. 20 is a flowchart of a method for adaptively determining an optimal coding mode among motion vector prediction modes including an AF4 mode or an AF6 mode based on a coding mode of a neighboring block, according to an embodiment (4-1) to which the present invention is applied .
- FIG. 21 shows a flowchart for performing adaptive decoding according to the AF4 mode or the AF6 mode based on the coding mode of the neighboring block, according to an embodiment (4-2) to which the present invention is applied.
- FIG. 22 shows a syntax structure for performing decoding according to the AF4 mode or the AF6 mode based on a coding mode of a neighboring block, according to an embodiment (4-3) to which the present invention is applied.
- FIG. 23 shows an embodiment (5-1) to which the present invention is applied.
- the AF4 mode or the AF6 mode is set based on at least one of a condition A (condition A), a condition B (condition B)
- FIG. 4 is a flowchart illustrating a method of adaptively determining an optimal coding mode among the motion vector prediction modes included in FIG.
- FIG. 24 shows an embodiment (5-2) to which the present invention is applied.
- the condition A condition A
- the condition B condition B
- FIG. 2 is a flowchart illustrating a decoding process according to the present invention.
- FIG. 25 shows an embodiment (5-3) to which the present invention is applied.
- the condition A condition A
- the condition B based on at least one of the condition A (condition A)
- the condition B And then performs a decoding process.
- FIG. 26 shows an embodiment (6-1) to which the present invention is applied.
- FIG. 27 is a flowchart (6-2) of an embodiment (6-2) to which the present invention is applied and which performs adaptive decoding according to the AF4 mode or the AF6 mode based on at least one of a coding condition of a condition A .
- FIG. 28 shows an embodiment (6-3) to which the present invention is applied, showing a syntax structure for performing decoding according to the AF4 mode or the AF6 mode based on at least one of a coding condition of a condition A (condition A) .
- 29 shows a flowchart for generating a motion vector predictor based on at least one of the AF4 mode and the AF6 mode, to which the present invention is applied.
- FIG. 30 shows a flowchart for generating a motion vector predictor based on AF4_flag and AF6_flag according to an embodiment to which the present invention is applied.
- FIG. 31 shows a flowchart for adaptively decoding according to the AF4 mode or the AF6 mode based on whether or not a neighboring block is coded in the AF mode, according to an embodiment to which the present invention is applied.
- 32 shows an embodiment in which the present invention is applied, and shows a syntax for performing adaptive decoding based on AF4_flag and AF6_flag.
- FIG. 33 shows an embodiment in which the present invention is applied, and shows a syntax for performing adaptive decoding according to the AF4 mode or the AF6 mode based on whether or not a neighboring block is coded in the AF mode.
- 35 shows a content streaming system to which the present invention is applied.
- the present invention provides a method of decoding a video signal including a current block based on an Affine Motion Prediction Mode (AF Mode), the method comprising the steps of: determining whether the AF mode is applied to the current block; Wherein the AF mode represents a motion prediction mode using an affine motion model; Determining whether an AF4 mode is used when the AF mode is applied to the current block, wherein the AF4 mode is a mode in which a motion vector is calculated using four parameters constituting the affine motion model, ≪ / RTI > If the AF4 mode is used, a motion vector predictor is generated using the four parameters. If the AF4 mode is not used, a motion vector predictor is generated using six parameters constituting the affine motion model step; And obtaining a motion vector of the current block based on the motion vector predictor.
- AF Mode Affine Motion Prediction Mode
- the method may further comprise obtaining an affine flag from the video signal, wherein the affine flag indicates whether the AF mode is applied to the current block, Whether or not the AF mode is applied is confirmed based on the affine flag.
- the method further includes obtaining an affine parameter flag from the video signal when the AF mode is applied to the current block according to the affine flag, wherein the affine parameter flag And whether the motion vector predictor is generated using the four parameters or using the six parameters.
- the affine flag and the affine parameter flag are defined in at least one level of a slice, a maximum coding unit, a coding unit or a prediction unit.
- the method further comprises checking whether the size of the current block satisfies a predetermined condition, wherein the predetermined condition is a number of pixels in the current block, a width and / Wherein the step of determining whether the AF mode is applied to the current block is performed if at least one of the current block and the current block is larger than a predetermined threshold and the current block size satisfies a predetermined condition .
- the current block when the size of the current block does not satisfy the predetermined condition, the current block is decoded based on a coding mode other than the AF mode.
- the method may further include, if the AF mode is applied to the current block, checking whether the AF mode is applied to the neighboring block, wherein when the AF mode is applied to the neighboring block , A motion vector predictor is generated using the four parameters, and if the AF mode is not applied to the neighboring block, checking whether the AF4 mode is used is performed.
- the present invention relates to an apparatus for decoding a video signal including a current block based on an affine motion prediction mode (AF mode), comprising: a step of checking whether the AF mode is applied to the current block, A motion vector predictor is generated using four parameters when the AF4 mode is used, and an affine motion model is generated when the AF4 mode is not used. and an inter prediction unit for generating a motion vector predictor using the six parameters constituting the motion vector predictor and obtaining a motion vector of the current block based on the motion vector predictor, (affine motion model), and the AF4 mode indicates the affine motion model (affine m and a motion vector prediction unit for estimating a motion vector using four parameters constituting the motion vector.
- affine motion prediction mode affine motion prediction mode
- the apparatus further includes a parsing unit for parsing an affine flag from the video signal, the affine flag indicating whether or not the AF mode is applied to the current block, Whether or not the AF mode is applied is confirmed based on the affine flag.
- the apparatus includes: a parsing unit for obtaining a affine parameter flag from the video signal when the AF mode is applied to the current block according to the affine flag, wherein the affine parameter flag And whether the motion vector predictor is generated using the four parameters or using the six parameters.
- the apparatus includes the inter prediction unit for checking whether a size of the current block satisfies a predetermined condition, wherein the predetermined condition is a number of pixels in the current block, a width and / Wherein the step of determining whether the AF mode is applied to the current block is performed if at least one of the current block size and the current block size is greater than a preset threshold value, do.
- the apparatus includes the inter-prediction unit for checking whether or not an AF mode is applied to a neighboring block when the AF mode is applied to the current block, A motion vector predictor is generated using the four parameters, and if the AF mode is not applied to the neighboring block, checking whether the AF4 mode is used is performed.
- signals, data, samples, pictures, frames, blocks, etc. may be appropriately replaced in each coding process.
- partitioning, decomposition, splitting, and division may be appropriately substituted for each coding process.
- FIG. 1 shows a schematic block diagram of an encoder in which the encoding of a video signal is performed, in which the present invention is applied.
- the encoder 100 includes an image divider 110, a transform unit 120, a quantization unit 130, an inverse quantization unit 140, an inverse transform unit 150, a filtering unit 160, A picture buffer (DPB) 170, an inter prediction unit 180, an intra prediction unit 185, and an entropy encoding unit 190.
- an image divider 110 a transform unit 120, a quantization unit 130, an inverse quantization unit 140, an inverse transform unit 150, a filtering unit 160, A picture buffer (DPB) 170, an inter prediction unit 180, an intra prediction unit 185, and an entropy encoding unit 190.
- DPB picture buffer
- the image divider 110 may divide an input image (or a picture or a frame) input to the encoder 100 into one or more processing units.
- the processing unit may be a coding tree unit (CTU), a coding unit (CU), a prediction unit (PU), or a transform unit (TU).
- the partitioning may be performed by at least one of QT (Quad Tree), BT (Binary Tree), TT (Ternary Tree) and AT (Asymmetric Tree).
- the terms are used only for convenience of explanation of the present invention, and the present invention is not limited to the definition of the term.
- the term coding unit is used as a unit used in a process of encoding or decoding a video signal, but the present invention is not limited thereto but may be appropriately interpreted according to the contents of the invention.
- the encoder 100 can generate a residual signal by subtracting a prediction signal output from the inter prediction unit 180 or the intra prediction unit 185 from the input image signal, The dual signal is transmitted to the conversion unit 120.
- the conversion unit 120 may apply a conversion technique to the residual signal to generate a transform coefficient.
- the conversion process may be applied to a pixel block having the same size of a square, or to a block having a variable size other than a square.
- the quantization unit 130 quantizes the transform coefficients and transmits the quantized transform coefficients to the entropy encoding unit 190.
- the entropy encoding unit 190 entropy-codes the quantized signals and outputs them as a bitstream.
- the quantized signal output from the quantization unit 130 may be used to generate a prediction signal.
- the quantized signal can be reconstructed by applying inverse quantization and inverse transform through the inverse quantization unit 140 and the inverse transform unit 150 in the loop.
- a reconstructed signal can be generated by adding the restored residual signal to a prediction signal output from the inter prediction unit 180 or the intra prediction unit 185.
- the filtering unit 160 applies filtering to the restored signal and outputs the restored signal to the playback apparatus or the decoded picture buffer 170.
- the filtered signal transmitted to the decoding picture buffer 170 may be used as a reference picture in the inter prediction unit 180. [ As described above, not only the picture quality but also the coding efficiency can be improved by using the filtered picture as a reference picture in the inter picture prediction mode.
- the decoded picture buffer 170 may store the filtered picture for use as a reference picture in the inter-prediction unit 180.
- the inter-prediction unit 180 performs temporal prediction and / or spatial prediction to remove temporal redundancy and / or spatial redundancy with reference to a reconstructed picture.
- the reference picture used for prediction is a transformed signal obtained through quantization and inverse quantization in units of blocks at the time of encoding / decoding in the previous time, blocking artifacts or ringing artifacts may exist have.
- the inter-prediction unit 180 can interpolate signals between pixels by sub-pixel by applying a low-pass filter in order to solve the performance degradation due to discontinuity or quantization of such signals.
- a subpixel means a virtual pixel generated by applying an interpolation filter
- an integer pixel means an actual pixel existing in a reconstructed picture.
- the interpolation method linear interpolation, bi-linear interpolation, wiener filter and the like can be applied.
- the interpolation filter may be applied to a reconstructed picture to improve the accuracy of the prediction.
- the inter-prediction unit 180 generates an interpolation pixel by applying an interpolation filter to an integer pixel, and uses an interpolated block composed of interpolated pixels as a prediction block Prediction can be performed.
- the intra prediction unit 185 can predict a current block by referring to samples in the vicinity of a block to be currently encoded.
- the intraprediction unit 185 may perform the following procedure to perform intraprediction. First, a reference sample necessary for generating a prediction signal can be prepared. Then, a prediction signal can be generated using the prepared reference sample. Thereafter, the prediction mode is encoded. At this time, reference samples can be prepared through reference sample padding and / or reference sample filtering. Since the reference samples have undergone prediction and reconstruction processes, quantization errors may exist. Therefore, a reference sample filtering process can be performed for each prediction mode used for intraprediction to reduce such errors.
- a prediction signal generated through the inter prediction unit 180 or the intra prediction unit 185 may be used to generate a reconstructed signal or may be used to generate a residual signal.
- Fig. 2 shows a schematic block diagram of a decoder in which decoding of a video signal is performed, according to an embodiment to which the present invention is applied.
- the decoder 200 includes a parsing unit (not shown), an entropy decoding unit 210, an inverse quantization unit 220, an inverse transform unit 230, a filtering unit 240, a decoded picture buffer (DPB) A decoded picture buffer unit 250, an inter prediction unit 260, an intra prediction unit 265, and a reconstruction unit (not shown).
- the decoder 200 may receive the signal output from the encoder 100 of FIG. 1 and may parse or obtain the syntax element through a parser (not shown). The parsed or obtained signal may be entropy-decoded through the entropy decoding unit 210.
- the inverse quantization unit 220 obtains a transform coefficient from the entropy-decoded signal using the quantization step size information.
- the inverse transform unit 230 obtains a residual signal by inversely transforming the transform coefficient.
- the restoring unit (not shown) adds the obtained residual signal to the prediction signal output from the inter-prediction unit 260 or the intra-prediction unit 265 to generate a reconstructed signal.
- the filtering unit 240 applies filtering to a reconstructed signal and outputs the reconstructed signal to a playback apparatus or a decoding picture buffer unit 250.
- the filtered signal transmitted to the decoding picture buffer unit 250 may be used as a reference picture in the inter prediction unit 260.
- the embodiments described in the filtering unit 160, the inter-prediction unit 180 and the intra-prediction unit 185 of the encoder 100 respectively include the filtering unit 240 of the decoder, the inter-prediction unit 260, The same can be applied to the intra prediction unit 265.
- the reconstructed video signal output through the decoder 200 may be reproduced through a reproducing apparatus.
- FIG. 3 is a diagram for explaining a block division structure of a QT (QuadTree, hereinafter referred to as 'QT') to which the present invention can be applied.
- One block in video coding can be segmented based on QT (QuadTree).
- QT QualityTree
- one sub-block divided by QT can be further recursively partitioned using QT.
- a leaf block that is not QT-divided can be divided by at least one of BT (Binary Tree), TT (Ternary Tree), or AT (Asymmetric Tree).
- BT can have two types of segmentation: horizontal BT (2NxN, 2NxN) and vertical BT (Nx2N, Nx2N).
- TT can have two types of segmentation: horizontal TT (2Nx1 / 2N, 2NxN, 2Nx1 / 2N) and vertical TT (1 / 2Nx2N, Nx2N, 1 / 2Nx2N).
- AT is a horizontal-up AT (2Nx1 / 2N, 2Nx3 / 2N), a horizontal-down AT (2Nx3 / 2N, 2Nx1 / 2N), a vertical-left AT (1 / 2Nx2N, 3 / 2Nx2N) / 2Nx2N, 1 / 2Nx2N).
- Each BT, TT, and AT can be recursively further partitioned using BT, TT, and AT.
- FIG. 3 shows an example of QT division.
- the block A can be divided into four sub-blocks (A0, A1, A2, A3) by QT.
- the sub-block A1 can be further divided into four sub-blocks (B0, B1, B2, B3) by QT.
- FIG. 4 is a diagram for explaining a BT (Binary Tree, hereinafter referred to as 'BT') block division structure to which the present invention can be applied.
- BT Binary Tree
- FIG. 4 shows an example of BT division.
- Block B3 which is no longer partitioned by QT, can be divided into vertical BT (C0, C1) or horizontal BT (D0, D1).
- each sub-block can be further recursively partitioned, such as in the form of horizontal BT (E0, E1) or vertical BT (F0, F1).
- FIG. 5 is a diagram for explaining a block division structure of a TT (Ternary Tree) block according to an embodiment of the present invention.
- FIG. 5 shows an example of TT division.
- Block B3 which is no longer partitioned by QT, may be divided into vertical TT (C0, C1, C2) or horizontal TT (D0, D1, D2).
- each sub-block can be further recursively divided into a horizontal TT (E0, E1, E2) or a vertical TT (F0, F1, F2).
- FIG. 6 is a diagram for explaining an AT (Asymmetric Tree) block partitioning structure to which the present invention can be applied.
- Block B3 which is no longer partitioned by QT, may be partitioned into vertical AT (C0, C1) or horizontal AT (D0, D1).
- each subblock can be further recursively partitioned, such as in the form of horizontal AT (E0, E1) or vertical TT (F0, F1).
- BT, TT, and AT segmentation can be used together.
- a subblock divided by BT can be divided by TT or AT.
- subblocks divided by TT can be divided by BT or AT.
- a subblock divided by AT can be divided by BT or TT.
- each subblock may be partitioned into a vertical BT, or after a vertical BT partition, each subblock may be partitioned into a horizontal BT.
- the two kinds of division methods have the same shape in the final division although the division order is different.
- searching is performed from left to right and from top to bottom, and searching for a block means a procedure for determining whether or not each divided sub-block is further divided into blocks, or when a block is not further divided, Refers to a coding order of a block, or a search order when referring to information of another neighboring block in a sub-block.
- FIG. 7 is a diagram for explaining an inter prediction mode as an embodiment to which the present invention is applied.
- a merge mode In the inter prediction mode to which the present invention is applied, a merge mode, an AMVP (Advanced Motion Vector Prediction) mode or an affine prediction mode (AFM) is used to reduce the amount of motion information .
- AMVP Advanced Motion Vector Prediction
- AFM affine prediction mode
- the merge mode refers to a method of deriving a motion parameter (or information) from a neighboring block spatially or temporally.
- the set of candidates available in the merge mode consists of spatial neighbor candidates, temporal candidates, and generated candidates.
- each spatial candidate block is available according to the order of ⁇ A1, B1, B0, A0, B2 ⁇ . At this time, if the candidate block is encoded in the intra-prediction mode and motion information does not exist, or if the candidate block is located outside the current picture (or slice), the candidate block can not be used.
- the spatial merge candidate can be constructed by excluding unnecessary candidate blocks from the candidate blocks of the current processing block. For example, if the candidate block of the current prediction block is the first prediction block in the same coding block, the candidate blocks excluding the candidate block and the same motion information may be excluded.
- the temporal merge candidate configuration process proceeds according to the order of ⁇ T0, T1 ⁇ .
- a right bottom block T0 of a collocated block of a reference picture is available, the block is configured as a temporal merge candidate.
- a collocated block refers to a block existing at a position corresponding to a current processing block in a selected reference picture. Otherwise, the block (T1) located at the center of the collocated block is constructed as a temporal merge candidate.
- the maximum number of merge candidates can be specified in the slice header. If the number of merge candidates is greater than the maximum number, the spatial candidates and temporal candidates smaller than the maximum number are retained. Otherwise, additional merge candidates (i.e., combined bi-predictive merging candidates) are generated by combining the candidates added so far until the number of merge candidates reaches the maximum number of candidates .
- the encoder constructs a merge candidate list by performing the above-described method and performs motion estimation (Motion Estimation) to obtain a merge index (for example, merge_idx [x0] [y0] ) To signal the decoder.
- FIG. 7B illustrates a case where the B1 block is selected in the merge candidate list. In this case, "Index 1" can be signaled to the decoder as a merge index.
- the decoder constructs a merge candidate list in the same way as the encoder and derives the motion information for the current block from the motion information of the candidate block corresponding to the merge index received from the encoder in the merge candidate list. Then, the decoder generates a prediction block for the current processing block based on the derived motion information.
- the AMVP mode refers to a method of deriving motion vector prediction values from neighboring blocks.
- the horizontal and vertical motion vector difference (MVD), reference index, and inter prediction mode are signaled to the decoder.
- the horizontal and vertical motion vector values are calculated using the derived motion vector prediction value and the motion vector difference (MVD) provided from the encoder.
- the encoder constructs a motion vector prediction value candidate list and performs motion estimation (motion estimation), thereby selecting a motion reference flag (i.e., candidate block information) (e.g., 'mvp_lX_flag [x0] y0] ') to the decoder.
- the decoder constructs a motion vector prediction value candidate list in the same manner as the encoder and derives the motion vector prediction value of the current processing block using the motion information of the candidate block indicated by the motion reference flag received from the encoder in the motion vector prediction value candidate list.
- the decoder obtains a motion vector value for the current processing block using the derived motion vector prediction value and the motion vector difference value transmitted from the encoder.
- the decoder generates a prediction block for the current processing block based on the derived motion information (i.e., motion compensation).
- the motion vector is scaled.
- the candidate composition is terminated. If the number of selected candidates is less than two, temporal motion candidates are added.
- a decoder decodes motion parameters for a processing block (e.g., a prediction unit).
- the decoder can decode the signaled merge index from the encoder.
- the motion parameter of the current processing block can be derived from the motion parameter of the candidate block indicated by the merge index.
- the decoder can decode the horizontal and vertical motion vector difference (MVD) signaled from the encoder, the reference index and the inter prediction mode.
- the motion vector prediction value is derived from the motion parameter of the candidate block indicated by the motion reference flag, and the motion vector value of the current processing block can be derived using the motion vector prediction value and the received motion vector difference value.
- the decoder performs motion compensation on the prediction unit using the decoded motion parameters (or information).
- the encoder / decoder performs motion compensation for predicting an image of the current unit from a previously decoded picture by using the decoded motion parameters.
- the AF mode refers to a motion prediction mode using an affine motion model, and may include at least one of an affine merge mode and an affine inter mode.
- the affine inter mode may include at least one of an AF4 mode or an AF6 mode.
- the AF4 mode represents a four parameter affine prediction mode using four parameters
- the AF6 mode represents a six parameter affine prediction mode using six parameters.
- the AF4 mode or the AF6 mode does not need to be defined as a separate prediction mode, and the AF4 mode or the AF6 mode uses only four parameters or six parameters It can be understood by distinguishing whether or not it is used.
- FIG. 8 The AF mode will be described in more detail in FIGS. 8 to 10.
- FIG. 8 The AF mode will be described in more detail in FIGS. 8 to 10.
- FIG. 8 is a diagram for explaining an affine motion model as an embodiment to which the present invention is applied.
- a common video coding technique uses a translation motion model to represent the motion of a coding block.
- the translation motion model represents a parallel-shifted block-based prediction method. That is, the motion information of the coding block is expressed using one motion vector.
- the optimal motion vector for each pixel in the actual coding block may be different from each other. If optimal motion vectors can be determined for each pixel or sub-block unit only with a small amount of information, the coding efficiency can be increased.
- the present invention proposes an inter prediction-based image processing method that reflects not only a parallel-shifted block-based prediction method but also various motion of an image to enhance the performance of inter prediction.
- the present invention proposes an affine motion estimation method that performs encoding / decoding using an affine motion model.
- the affine motion model represents a prediction method of deriving a motion vector on a pixel-by-pixel or sub-block-by-block basis using a motion vector of a control point.
- an affine motion prediction mode using an affine motion model is referred to as an AF mode (Affine Mode).
- the present invention also provides a method for adaptively performing affine prediction based on a block size.
- the present invention also provides a method for adaptively performing affine prediction based on whether neighboring blocks are coded by affine prediction.
- the present invention also provides a method for adaptively determining (or selecting) an optimal coding mode based on at least one of an AF4 mode or an AF6 mode.
- the AF4 mode represents a four parameter affine prediction mode using four parameters
- the AF6 mode represents a six parameter affine prediction mode using six parameters.
- the affine motion model can represent the four motions shown in FIG.
- an affine motion model can model arbitrary image distortions caused by image translations, image scaling, image rotations, and image shear. have.
- the affine motion model can be expressed in various ways, but in the present invention, distortion is indicated (or identified) by using motion information at a specific reference point (or reference pixel / sample) of the block, And the like.
- the reference point may be referred to as a control point (CP) (or a control pixel, a control sample), and a motion vector at this reference point may be referred to as a control point motion vector (CPMV).
- CP control point
- CPMV control point motion vector
- the degree of distortion that can be expressed can vary depending on the number of control points.
- the affine motion model can be expressed using six parameters (a, b, c, d, e, f) as shown in the following Equation (1).
- (x, y) represents the position of the upper left pixel of the coding block.
- v x and v y represent motion vectors at (x, y), respectively.
- FIG. 9 is a diagram for explaining an affine motion prediction method using a control point motion vector according to an embodiment of the present invention. Referring to FIG.
- the upper left control point (control point) of the current block (901) (CP 0) ( 902) ( hereinafter referred to as a first control point), upper right control point (control point) (CP 1) (Hereinafter referred to as a second control point) and a lower left control point (CP 2 ) 904 (hereinafter referred to as a third control point) may have independent motion information. It can be expressed as CP 0 , CP 1 , and CP 2 , respectively. However, this is an embodiment of the present invention, and the present invention is not limited thereto. For example, it is possible to define various control points such as a right-side control point, a center control point, and a control point for each sub-block.
- At least one of the first to third control points may be a pixel included in the current block.
- at least one of the first to third control points may be a pixel adjacent to a current block not included in the current block.
- Motion information for each pixel or sub-block of the current block 901 may be derived using motion information of one or more control points of the control points.
- an affine motion model using the motion vectors of the upper left control point 902, the upper right control point 903, and the lower left control point 904 of the current block 901 can be defined as Equation 2 below .
- the motion vector of the upper left control point 902 The motion vector of the upper right control point 903, Is a motion vector of the lower left control point 904, . ≪ / RTI >
- w represents the width of the current block 901
- h represents the height of the current block 901.
- an affine motion model expressing three movements of translation, scale, and rotate, which can be expressed by the affine motion model.
- this is referred to as a simplified affine motion model or a similarity affine motion model.
- the simplified affine motion model can be expressed using four parameters (a, b, c, d) as shown in the following Equation (3).
- ⁇ v x , v y ⁇ represents a motion vector at ⁇ x, y ⁇ , respectively.
- An affine motion model that uses four parameters in this way can be referred to as AF4.
- the present invention is not limited to this, and when six parameters are used, it is referred to as AF6, and the above embodiments can be applied equally.
- a motion model which is an AF4 affine model can be defined as the following Equation (4).
- Equation (4) w represents the width of the current block and h represents the height of the current block. And, Respectively represent motion vectors at ⁇ x, y ⁇ positions.
- the encoder or decoder may determine (or derive) a motion vector of each pixel location using a control point motion vector (e.g., a motion vector of the upper left control point 1001 and the upper right control point 1002).
- a control point motion vector e.g., a motion vector of the upper left control point 1001 and the upper right control point 1002.
- a set of motion vectors determined through affine motion prediction may be defined as an affine motion vector field.
- the affine motion vector field may be determined using at least one of Equations 1 to 4 above.
- a motion vector through affine motion prediction in the encoding / decoding process can be determined in units of pixels or in units of a predefined (or preset) block (or subblock). For example, when the determination is made on a pixel-by-pixel basis, a motion vector may be derived on the basis of each pixel in the block, and if it is determined on a sub-block basis, a motion vector may be derived on the basis of each sub-block unit in the current block. As another example, when the motion vector is determined in units of subblocks, the motion vector of the corresponding subblock may be derived based on the upper left pixel or the center pixel.
- the encoder or decoder can determine a motion vector in units of 4x4 sub-blocks using the motion vectors of the upper left control point 801 and the upper right control point 802 of the current block.
- the motion vector of the corresponding sub-block may be determined based on the center pixel value of each sub-block.
- Affine motion prediction can be used as an affine merge mode (hereinafter referred to as AF merge mode) and an affine inter mode (hereinafter referred to as AF inter mode).
- the AF merge mode is a method of encoding or decoding two motion vector motion vectors without encoding the motion vector difference similar to the skip mode or merge mode.
- the AF inter mode is a method of encoding a control point motion vector predictor and a control point motion vector, and then encoding or decoding a control point motion vector difference (CPMVD) corresponding to the difference.
- CPMVD control point motion vector difference
- the motion vector difference of two control points is transmitted in the case of the AF4 mode, and the difference of motion vectors of three control points is transmitted in the AF6 mode.
- the AF4 mode since the AF4 mode transmits a smaller number of motion vector difference values than the AF6 mode, it has an advantage that the control point motion vector (CPMV) can be expressed with a small number of bits.
- the AF6 mode transmits three CPMVDs Therefore, it is possible to generate an excellent predictor, thereby reducing the bit for residual coding.
- the present invention proposes a method of considering both (or simultaneously) the AF4 mode and the AF6 mode in the AF inter mode.
- FIG. 10 is a flowchart illustrating a process of processing a video signal including a current block using an Arde prediction mode (hereinafter, referred to as 'AF mode') to which the present invention is applied.
- 'AF mode' an Rastere prediction mode
- the present invention provides a method for processing a video signal including a current block using an AF mode.
- the video signal processing apparatus may generate a candidate list of motion vector pairs using motion vectors of pixels or blocks adjacent to at least two control points of the current block (S1010).
- the control point represents a corner pixel of the current block
- the motion vector pair represents a motion vector of the upper left corner pixel and the upper right corner pixel of the current block.
- control point includes at least two of an upper left corner pixel, an upper right corner pixel, a lower left corner pixel, or a lower right corner pixel of the current block
- candidate list includes the upper left corner pixel, An upper right corner pixel, and pixels or blocks adjacent to the lower left corner pixel.
- the candidate list includes motion vectors of a diagonal adjacent pixel (A), an upper adjacent pixel (B), and a left adjacent pixel (C) of the upper left corner pixel, D and the motion vectors of the diagonal adjacent pixel E and the motion vectors of the left adjacent pixel F and the diagonal adjacent pixel G of the lower left corner pixel.
- the method may further comprise adding an AMVP candidate list to the candidate list if the motion vector pair of the candidate list is less than two.
- the control point motion vector of the current block is determined as a motion vector derived based on the center positions of the left subblock and the right subblock in the current block,
- the motion vector of the current block is determined as a motion vector based on a center position of an upper sub-block and a lower sub-block in the current block.
- the control point motion vector of the left subblock in the current block is determined by the average value of the first control point motion vector and the third control point motion vector, and the control point motion Vector is determined by a mean value of a second control point motion vector and a fourth control point motion vector, and when the current block is 4xN size, a control point motion vector of an upper sub-block in the current block is a first control point motion vector, And the control point motion vector of the lower sub-block is determined by an average value of the third control point motion vector and the fourth control point motion vector.
- the method may signal a prediction mode or flag information indicating whether the AF mode is performed.
- the video signal processing apparatus may receive the prediction mode or flag information, perform the AF mode according to the prediction mode or the flag information, and derive a motion vector according to the AF mode.
- the AF mode is a mode for deriving a motion vector in units of pixels or sub-blocks using a control point motion vector of the current block.
- the video signal processing apparatus may determine a final candidate list of a predetermined number of motion vector pairs based on a divergence value of the motion vector pair (S1020).
- the final candidate list is determined in descending order of the divergence value, and the divergence value is a value indicating the similarity of the directions of the motion vectors.
- the video signal processing apparatus may determine a control point motion vector of the current block based on a rate-distortion cost from the final candidate list (S1030).
- the video signal processing apparatus may generate a motion vector predictor of the current block based on the control point motion vector (S1040).
- 11 is a flowchart of a method for adaptively determining an optimal coding mode based on at least one of an AF4 mode and an AF6 mode according to an embodiment (1-1) to which the present invention is applied.
- the video signal processing apparatus may perform prediction based on at least one of a skip mode, a merge mode, and an inter mode (S1110).
- the merge mode may include not only the general merge mode but also the AF merge mode described above
- the inter mode may include the general inter mode as well as the AF inter mode described above.
- the video signal processing apparatus may perform motion vector prediction based on at least one of AF4 mode and AF6 mode (S1120).
- steps S1110 and S1120 are not limited to the order.
- the video signal processing apparatus may compare the results of step S1120 and determine an optimal coding mode among the modes (S1130). At this time, the results of step S1120 may be compared based on a Rate-Distortion Cost.
- the video signal processing apparatus can generate a motion vector predictor of the current block based on an optimal coding mode, and subtract the motion vector predictor from the motion vector of the current block to obtain a motion vector difference value .
- Fig. 12 shows a flowchart for performing adaptive decoding based on the AF4 mode or the AF6 mode as an embodiment (1-2) to which the present invention is applied.
- the decoder can receive the bitstream (S1210).
- the bitstream may include information on the coding mode of the current block in the video signal.
- the decoder can check whether the coding mode of the current block is the AF mode (S1220).
- the AF mode means an affine motion prediction mode using an affine motion model.
- the AF mode may be an affine merge mode or an affine inter mode.
- the affine inter mode may include at least one of AF4 mode and AF6 mode.
- the step S1220 may be confirmed by an affine flag indicating whether or not the AF mode is performed.
- the decoder may perform decoding (i.e., motion vector prediction) according to a coding mode other than the AF mode (S1230). For example, a skip mode, a merge mode, or an inter mode may be used.
- the decoder can check whether the AF4 mode is applied to the current block (S1240).
- the step S1240 may be confirmed by a parameter flag indicating whether AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- the affine parameter flag may include at least one of AF4_flag and AF6_flag.
- the execution of the AF4 mode means that the motion vector prediction is performed using an affine motion model represented by four parameters.
- the execution of the AF6 mode implies performing the motion vector prediction using the affine motion model represented by four parameters.
- the affine flag and the affine parameter flag may be defined at a level of at least one of a slice, a maximum coding unit, a coding unit or a prediction unit.
- AF_flag AF_flag
- AF4_flag AF6_flag
- AF6_flag may be defined at the slice level and again defined at the block level or the prediction unit level.
- FIG. 13 shows an embodiment (1-3) to which the present invention is applied, showing a syntax structure for performing decoding based on the AF4 mode or the AF6 mode.
- the decoder acquires the merge_flag to check whether the merge mode is applied to the current block (S1310).
- the decoder may obtain an affine_flag (S1320).
- affine_flag indicates whether the AF mode is performed or not.
- affine_flag 1, that is, if the AF mode is performed for the current block, the decoder can obtain the affine_param_flag (S1330).
- affine_param_flag indicates whether the AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- the decoder can obtain mvd_CP0 and mvd_CP1 which are two motion vector difference values (S1340).
- mvd_CP0 represents a motion vector difference value for the control point
- mvd_CP1 represents a motion vector difference value for the control point 1.
- the decoder can obtain mvd_CPO, mvd_CP1 and mvd_CP2, which are three motion vector difference values (S1350).
- FIG. 14 is an embodiment (2-1) to which the present invention is applied, in which adaptively determines an optimal coding mode among the motion vector prediction modes including the AF4 mode or the AF6 mode based on the condition A Fig.
- the encoder may perform prediction based on at least one of a skip mode, a merge mode, or an inter mode (S1410).
- the encoder may determine whether the condition A is satisfied for the current block to determine an optimal coding mode for motion vector prediction (S1420).
- condition A may mean a condition for a block size.
- the embodiments of Table 1 below can be applied.
- Example 1 of Table 1 the condition A indicates whether the number of pixels (pixNum) of the current block is larger than a threshold value TH1.
- the threshold value is 64, 128, 256, 512, 1024, ...
- TH1 64 means that the block size is 4x16, 8x8, or 16x4, and
- TH1 128 means that the block size is 32x4, 16x8, 8x16, or 4x32 .
- Example 2 it indicates whether the width and the height of the current block are both greater than the threshold value TH1.
- Example 3 it indicates whether the width of the current block is greater than the threshold value TH1 or whether the height of the current block is greater than the threshold value TH1.
- the encoder may perform motion vector prediction based on at least one of the AF4 mode and the AF6 mode (S1430).
- the encoder may compare the results of steps S1410 and S1430 to determine an optimal coding mode among the motion vector prediction modes including the AF4 mode or the AF6 mode (S1440).
- the encoder can determine an optimal coding mode among the modes other than the AF mode (S1440).
- the encoder can generate a motion vector predictor of the current block based on the optimal coding mode, and subtract the motion vector predictor from the motion vector of the current block to obtain a motion vector difference value.
- Fig. 15 shows a flowchart for performing adaptive decoding according to the AF4 mode or the AF6 mode based on the condition A (condition A) as an embodiment (2-2) to which the present invention is applied.
- the decoder can receive the bit stream (S1510).
- the bitstream may include information on the coding mode of the current block in the video signal.
- the decoder can determine whether the condition A is satisfied for the current block to determine an optimal coding mode for motion vector prediction (S1520).
- the condition A may mean a condition for a block size.
- the embodiments of Table 1 above can be applied.
- the decoder can check whether the coding mode of the current block is the AF mode (S1530).
- the AF mode means an affine motion prediction mode using an affine motion model, and the embodiments described herein can be applied.
- the step S1530 may be confirmed by an affine flag indicating whether or not the AF mode is performed.
- the decoder may perform decoding (i.e., motion vector prediction) according to a coding mode other than the AF mode (S 1540 ). For example, a skip mode, a merge mode, or an inter mode may be used.
- the decoder can check whether the AF4 mode is applied to the current block (S1550).
- the step S1550 may be confirmed by a parameter flag indicating whether AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- FIG. 16 shows a syntax structure for performing decoding according to the AF4 mode or the AF6 mode on the basis of the condition A (condition A) according to the embodiment (2-3) to which the present invention is applied.
- the decoder acquires the merge_flag to check whether the merge mode is applied to the current block (S1610).
- the decoder can check whether the condition A is satisfied (S1620).
- the condition A may mean a condition for a block size.
- the embodiments of Table 1 above can be applied.
- affine_flag indicates whether the AF mode is performed or not.
- affine_flag 1, that is, if the AF mode is performed for the current block, the decoder can obtain the affine_param_flag (S1630).
- affine_param_flag indicates whether the AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- the decoder can obtain two motion vector difference values mvd_CPO and mvd_CP1 (S1640).
- mvd_CP0 represents a motion vector difference value for the control point
- mvd_CP1 represents a motion vector difference value for the control point 1.
- the decoder can obtain mvd_CPO, mvd_CP1 and mvd_CP2, which are three motion vector difference values (S1650).
- FIG. 17 shows an embodiment (3-1) to which the present invention is applied.
- the motion vector prediction modes including the AF4 mode or the AF6 mode, based on at least one of the condition B (condition B) and the condition C
- a coding mode adaptively determining an optimum coding mode among the plurality of coding modes.
- the present invention provides a method for adaptively selecting an AF4 mode and an AF6 mode based on the size of a current block.
- the AF6 mode transmits an additional motion vector difference value in addition to the AF4 mode, which is effective in a relatively large block. Accordingly, if the current block size is smaller than (or equal to) a predetermined size, encoding is performed only considering the AF4 mode, and if the current block is greater than or equal to a predetermined size, Can be performed.
- both the AF4 mode and the AF6 mode can be considered and only the optimal mode can be signaled.
- the encoder may perform prediction based on at least one of a skip mode, a merge mode, and an inter mode (S1710).
- the encoder can check whether the condition B is satisfied for the current block (S1720).
- the condition B may mean a condition for a block size.
- Table 2 the embodiments of Table 2 below can be applied.
- Example 1 of Table 2 the condition B indicates whether the number of pixels (pixNum) of the current block is smaller than a threshold value TH2.
- the threshold value is 64, 128, 256, 512, 1024, ...
- TH2 64 means that the block size is 4x16, 8x8, or 16x4, and
- TH2 128 means that the block size is 32x4, 16x8, 8x16, or 4x32 .
- the condition B indicates whether both the width and the height of the current block are less than the threshold value TH2.
- the condition B indicates whether the width of the current current block is smaller than the threshold value TH2 or the height of the current block is smaller than the threshold value TH2.
- the encoder can perform motion vector prediction based on the AF4 mode (S1730).
- the encoder can check whether the condition C is satisfied with respect to the current block (S1740).
- the condition C may mean a condition for a block size.
- the embodiments of Table 3 below can be applied.
- Example 1 of Table 3 the condition A indicates whether the number of pixels (pixNum) of the current block is greater than or equal to a threshold value TH3.
- the threshold value is 64, 128, 256, 512, 1024, ...
- TH3 64 means that the block size is 4x16, 8x8, or 16x4, and
- TH3 128 means that the block size is 32x4, 16x8, 8x16 or 4x32 .
- Example 2 it indicates whether the width and the height of the current block are both greater than or equal to the threshold value TH3.
- Example 3 it indicates whether the width of the current block is greater than or equal to the threshold value TH1, or whether the height of the current block is greater than or equal to the threshold value TH1.
- the encoder can perform motion vector prediction based on the AF6 mode (S1760).
- the encoder can perform motion vector prediction based on the AF4 mode and the AF6 mode (S1750).
- the threshold value TH2 and the threshold value TH3 may be determined so as to satisfy the following equation (5).
- the encoder may compare the results of steps S1710, S1730, S1750, and S1760 to determine an optimal coding mode among the motion vector prediction modes including the AF4 mode or the AF6 mode (S1770).
- the encoder can generate a motion vector predictor of the current block based on the optimal coding mode, and subtract the motion vector predictor from the motion vector of the current block to obtain a motion vector difference value.
- FIG. 18 is an embodiment (3-2) to which the present invention is applied and adaptively decodes according to the AF4 mode or the AF6 mode based on at least one of the condition B (condition B) and the condition C (condition C) Fig.
- the decoder can check whether the coding mode of the current block is the AF mode (S1810).
- the AF mode is an affine motion prediction mode using an affine motion model, and the embodiments described herein can be applied, and redundant description will be omitted.
- the decoder can check whether the condition B is satisfied for the current block (S1820).
- the condition B may mean a condition for a block size.
- the embodiments of Table 2 above can be applied, and redundant descriptions are omitted.
- the decoder can perform motion vector prediction based on the AF4 mode (S1830).
- the decoder can check whether the condition C is satisfied with respect to the current block (S 1840).
- the condition C may mean a condition for a block size.
- the embodiments of Table 3 above can be applied, and redundant descriptions are omitted.
- the threshold value TH2 and the threshold value TH3 may be determined to satisfy Equation (5).
- the decoder can perform motion vector prediction based on the AF6 mode (S1860).
- the decoder can check whether the AF4 mode is applied to the current block (S1850).
- the step S1850 may be confirmed by a parameter flag which is an affirmative flag indicating whether the AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- the decoder may perform decoding (i.e., motion vector prediction) according to a coding mode other than the AF mode (S 1870).
- a coding mode other than the AF mode For example, a skip mode, a merge mode, or an inter mode may be used.
- Fig. 19 is a schematic block diagram of a syntax structure (or a decoder) according to an embodiment (3-3) to which the present invention is applied and which performs decoding in accordance with at least one of the condition B (condition B) or the condition C .
- the decoder acquires the merge_flag to check whether the merge mode is applied to the current block (S1910).
- the decoder may obtain an affine_flag (S1920).
- affine_flag indicates whether the AF mode is performed or not.
- the decoder can check whether the condition B is satisfied (S1620).
- the condition B may mean a condition for a block size.
- the embodiments of Table 2 above can be applied.
- affine_param_flag indicates whether the AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- affine_param_flag 0 means that motion vector prediction is performed according to the AF4 mode.
- affine_param_flag 1 means that motion vector prediction is performed according to the AF6 mode.
- the decoder can obtain the affine_param_flag (S1950).
- the decoder can obtain mvd_CPO and mvd_CP1 which are two motion vector difference values (S1960).
- the decoder can obtain mvd_CPO, mvd_CP1 and mvd_CP2, which are three motion vector difference values (S1970).
- FIG. 20 is a flowchart of a method for adaptively determining an optimal coding mode among motion vector prediction modes including an AF4 mode or an AF6 mode based on a coding mode of a neighboring block, according to an embodiment (4-1) to which the present invention is applied .
- the encoder may perform prediction based on at least one of a skip mode, a merge mode, or an inter mode (S2010).
- the encoder can check whether the neighboring block is coded in the AF mode (S2020).
- the encoder may perform motion vector prediction based on the AF4 mode (S2030).
- the encoder can perform motion vector prediction based on the AF4 mode and perform motion vector prediction based on the AF6 mode (S2040).
- the encoder may compare the results of steps S2030 and S2040 to determine an optimal coding mode among the motion vector prediction modes including the AF4 mode or the AF6 mode (S2050).
- the encoder can generate a motion vector predictor of the current block based on the optimal coding mode, and subtract the motion vector predictor from the motion vector of the current block to obtain a motion vector difference value.
- FIG. 21 shows a flowchart for performing adaptive decoding according to the AF4 mode or the AF6 mode based on the coding mode of the neighboring block, according to an embodiment (4-2) to which the present invention is applied.
- the decoder may receive the bitstream (S2110).
- the bitstream may include information on the coding mode of the current block in the video signal.
- the decoder can check whether the coding mode of the current block is the AF mode (S2120).
- the decoder may perform decoding (i.e., motion vector prediction) according to a coding mode other than the AF mode (S2170). For example, a skip mode, a merge mode, or an inter mode may be used.
- the decoder can check whether the neighboring block is coded in the AF mode (S2130).
- the decoder may perform motion vector prediction based on the AF4 mode (S2140).
- the decoder can check whether the AF4 mode is applied to the current block (S2150).
- the step S2150 may be confirmed by a parameter flag indicating whether AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- FIG. 22 shows a syntax structure for performing decoding according to the AF4 mode or the AF6 mode based on a coding mode of a neighboring block, according to an embodiment (4-3) to which the present invention is applied.
- the decoder acquires the merge_flag to check whether the merge mode is applied to the current block (S2210).
- the decoder may obtain an affine_flag (S2220).
- affine_flag indicates whether the AF mode is performed or not.
- affine_flag 1, that is, if the AF mode is performed on the current block, the decoder can check whether the neighboring block is coded in the AF mode (S2230).
- affine_param_flag indicates whether the AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- the decoder can set affine_param_flag to 0 (S2240).
- the decoder can obtain mvd_CPO and mvd_CP1, which are two motion vector difference values (S2250).
- the decoder can obtain mvd_CPO, mvd_CP1, and mvd_CP2, which are three motion vector difference values (S2260).
- FIG. 23 shows an embodiment (5-1) to which the present invention is applied.
- the AF4 mode or the AF6 mode is set based on at least one of a condition A (condition A), a condition B (condition B)
- FIG. 4 is a flowchart illustrating a method of adaptively determining an optimal coding mode among the motion vector prediction modes included in FIG.
- the present invention shows an embodiment in which Embodiment 2 and Embodiment 3 are combined.
- FIG. 23 an example of the case in which all the conditions A, B, and C are considered is explained, and the order of the conditions can be applied differently.
- the encoder may perform prediction based on at least one of a skip mode, a merge mode, and an inter mode (S1710).
- the encoder can check whether the condition A is satisfied for the current block (S2320).
- the condition A may mean a condition for a block size, and the embodiments of Table 1 may be applied.
- the encoder can determine an optimal coding mode among modes other than the AF mode (S2380).
- the encoder can check whether the condition B is satisfied with respect to the current block (S2330).
- the condition B may mean a condition for a block size, and the embodiments of Table 2 may be applied.
- the encoder can perform motion vector prediction based on the AF4 mode (S2340).
- the encoder can check whether the condition C is satisfied with respect to the current block (S2350).
- the condition C may mean a condition for a block size, and the embodiments of Table 3 may be applied.
- the encoder can perform motion vector prediction based on the AF6 mode (S2370).
- the encoder can perform the motion vector prediction based on the AF4 mode and the motion vector prediction based on the AF6 mode (S2360).
- the threshold value TH2 and the threshold value TH3 may be determined to satisfy Equation (5).
- the encoder may compare the results of steps S2310, S2340, S2360, and S2370 to determine an optimal coding mode (S2380).
- the encoder can generate a motion vector predictor of the current block based on the optimal coding mode, and subtract the motion vector predictor from the motion vector of the current block to obtain a motion vector difference value.
- FIG. 24 shows an embodiment (5-2) to which the present invention is applied.
- the condition A condition A
- the condition B condition B
- FIG. 2 is a flowchart illustrating a decoding process according to the present invention.
- the decoder can check whether the condition A is satisfied for the current block (S2410).
- the condition A may mean a condition for a block size.
- the embodiments of Table 1 above can be applied.
- the decoder can check whether the coding mode of the current block is the AF mode (S2420).
- the AF mode is an affine motion prediction mode using an affine motion model, and the embodiments described herein can be applied, and redundant description will be omitted.
- the decoder may perform decoding (i.e., motion vector prediction) according to a coding mode other than the AF mode (S2480 ). For example, a skip mode, a merge mode, or an inter mode may be used.
- the decoder can check whether the condition B is satisfied for the current block (S2430).
- the condition B may mean a condition for a block size.
- the embodiments of Table 2 above can be applied, and redundant descriptions are omitted.
- the decoder can perform motion vector prediction based on the AF4 mode (S2440).
- the decoder can check whether the condition C is satisfied with respect to the current block (S2450).
- the condition C may mean a condition for a block size.
- the embodiments of Table 3 above can be applied, and redundant descriptions are omitted.
- the threshold value TH2 and the threshold value TH3 may be determined to satisfy Equation (5).
- the decoder can perform motion vector prediction based on the AF6 mode (S2470).
- the decoder can check whether the AF4 mode is applied to the current block (S2460).
- the step S2460 may be confirmed by a parameter flag indicating whether AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- FIG. 25 shows an embodiment (5-3) to which the present invention is applied.
- the condition A condition A
- the condition B based on at least one of the condition A (condition A)
- the condition B And then performs a decoding process.
- the decoder acquires the merge_flag to check whether the merge mode is applied to the current block (S2510).
- the decoder can check whether the condition A is satisfied (S2520).
- the condition A may mean a condition for a block size.
- the embodiments of Table 1 above can be applied.
- affine_flag indicates whether the AF mode is performed or not.
- the decoder can check whether the condition B is satisfied (S2530).
- the condition B may mean a condition for a block size.
- the embodiments of Table 2 above can be applied.
- affine_param_flag indicates whether the AF4 mode is performed (or whether affine motion prediction is performed by four parameters).
- affine_param_flag 0 means that motion vector prediction is performed according to the AF4 mode.
- affine_param_flag 1 means that motion vector prediction is performed according to the AF6 mode.
- the decoder can obtain the affine_param_flag (S2560).
- the decoder can obtain mvd_CPO and mvd_CP1 which are two motion vector difference values (S2570).
- the decoder can obtain mvd_CPO, mvd_CP1 and mvd_CP2, which are three motion vector difference values (S2580).
- FIG. 26 shows an embodiment (6-1) to which the present invention is applied.
- the encoder may perform prediction based on at least one of a skip mode, a merge mode, or an inter mode (S2610).
- the encoder can check whether the condition A is satisfied for the current block (S2620).
- the condition A may mean a condition for a block size, and the embodiments of Table 1 may be applied.
- the encoder can determine an optimal coding mode among modes other than the AF mode (S2660).
- the encoder can check whether the neighboring block is coded in the AF mode (S2630).
- the encoder may perform motion vector prediction based on the AF4 mode (S2640).
- the encoder may perform motion vector prediction based on the AF4 mode, and may perform motion vector prediction based on the AF6 mode (S2650).
- the encoder may compare the results of steps S2610, S2640, and S2650 to determine an optimal coding mode (S2660).
- the encoder can generate a motion vector predictor of the current block based on the optimal coding mode, and subtract the motion vector predictor from the motion vector of the current block to obtain a motion vector difference value.
- FIG. 27 is a flowchart (6-2) of an embodiment (6-2) to which the present invention is applied and which performs adaptive decoding according to the AF4 mode or the AF6 mode based on at least one of a coding condition of a condition A .
- the decoder can receive the bitstream (S2710).
- the bitstream may include information on the coding mode of the current block in the video signal.
- the decoder can determine whether the condition A is satisfied for the current block to determine an optimal coding mode for motion vector prediction (S2720).
- the condition A may mean a condition for a block size.
- the embodiments of Table 1 above can be applied.
- the decoder can check whether the coding mode of the current block is the AF mode (S2730).
- steps S2730 through S2780 the contents described in steps S2120 through S2170 of FIG. 21 may be applied, and redundant description will be omitted.
- FIG. 28 shows an embodiment (6-3) to which the present invention is applied, showing a syntax structure for performing decoding according to the AF4 mode or the AF6 mode based on at least one of a coding condition of a condition A (condition A) .
- the decoder acquires the merge_flag to check whether the merge mode is applied to the current block (S2810).
- the decoder can check whether the condition A is satisfied (S2820).
- the condition A may mean a condition for a block size.
- the embodiments of Table 1 above can be applied.
- affine_flag indicates whether the AF mode is performed or not.
- steps S2830 to S2860 the contents described in steps S2230 to S2260 of FIG. 22 may be applied, and redundant description will be omitted.
- 29 shows a flowchart for generating a motion vector predictor based on at least one of the AF4 mode and the AF6 mode, to which the present invention is applied.
- the decoder can check whether or not the AF mode is applied to the current block (S2910).
- the AF mode represents a motion prediction mode using an affine motion model.
- the decoder may obtain an affine flag from a video signal, and whether or not the AF mode is applied to the current block may be ascertained based on the affine flag.
- the decoder can check whether the AF4 mode is used (S2920).
- the AF4 mode represents a mode for predicting a motion vector using four parameters constituting the affine motion model.
- the decoder can obtain a affine parameter flag from the video signal, Is generated using the four parameters or using the six parameters.
- the affine flag and the affine parameter flag may be defined in at least one level of a slice, a maximum coding unit, a coding unit or a prediction unit.
- the decoder If the AF4 mode is used, the decoder generates a motion vector predictor using the four parameters. If the AF4 mode is not used, the decoder uses six parameters constituting the affine motion model to generate a motion vector A predictor may be generated (S2930).
- the decoder may obtain a motion vector of the current block based on the motion vector predictor (S2940).
- the decoder can check whether the size of the current block meets predetermined conditions. At this time, the preset condition indicates whether at least one of the number of pixels in the current block, the width and / or height of the current block is greater than a predetermined threshold value.
- the decoder can check whether the AF mode is applied to the current block.
- the current block may be decoded based on a coding mode other than the AF mode.
- the decoder can check whether or not the AF mode is applied to the neighboring block.
- the motion vector predictor is generated using the four parameters. If the AF mode is not applied to the neighboring block, the decoder checks whether the AF4 mode is used . ≪ / RTI >
- FIG. 30 shows a flowchart for generating a motion vector predictor based on AF4_flag and AF6_flag according to an embodiment to which the present invention is applied.
- the decoder can obtain at least one of AF4 flag and AF6 flag from the video signal (S3010).
- AF4_flag indicates whether or not the AF4 mode is performed for the current block
- AF6_flag indicates whether or not the AF6 mode is performed for the current block.
- At this time, at least one of the AF4 flag and the AF6 flag is defined at the slice level, and can be defined again at the block level or the prediction unit level.
- the present invention is not limited to this, and at least one of the AF4 flag and the AF6 flag may be defined in at least one of a slice, a maximum coding unit, a coding unit, and a prediction unit.
- the decoder can check the values of AF4 flag and AF6 flag (S3020).
- the execution of the AF4 mode means that the motion vector prediction is performed using the affine motion model expressed by four parameters.
- AF6_flag 1
- AF4_flag 0
- the AF6 mode is not performed for the current block.
- the execution of the AF6 mode implies performing the motion vector prediction using the affine motion model represented by four parameters.
- the decoder can perform motion vector prediction according to modes other than the AF4 mode and the AF6 mode (S3030).
- the decoder can perform motion vector prediction according to the AF4 mode (S3040).
- the decoder can perform motion vector prediction according to the AF6 mode (S3050).
- the decoder can perform the motion vector prediction according to the AF4 mode or the AF6 mode (S3060).
- FIG. 31 shows a flowchart for adaptively decoding according to the AF4 mode or the AF6 mode based on whether or not a neighboring block is coded in the AF mode, according to an embodiment to which the present invention is applied.
- the decoder can check whether or not the AF mode is applied to the current block (S3110).
- the decoder can check whether the neighboring block is coded in the AF mode (S3120).
- the decoder can acquire at least one of AF4 flag or AF6_flag (S3130).
- the decoder may obtain a motion vector of the current block based on the motion vector predictor (S3150).
- 32 shows an embodiment in which the present invention is applied, and shows a syntax for performing adaptive decoding based on AF4_flag and AF6_flag.
- the decoder can obtain AF4 flag and AF6 flag at the slice level (S3010).
- AF4_flag indicates whether or not the AF4 mode is performed for the current block
- AF6_flag indicates whether or not the AF6 mode is performed for the current block.
- the AF4_flag may be expressed by affine_4_flag
- the AF6_flag may be expressed by affine_6_flag.
- the decoder may adaptively perform decoding based on AF4_flag and AF6_flag at the block level or prediction unit level.
- the decoder can acquire the affine flag (S3220).
- the affine flag may indicate whether or not the AF mode is performed.
- the decoder may adaptively perform decoding according to the values of AF4_flag and AF6_flag.
- FIG. 33 shows an embodiment in which the present invention is applied, and shows a syntax for performing adaptive decoding according to the AF4 mode or the AF6 mode based on whether or not a neighboring block is coded in the AF mode.
- the decoder can parse or obtain the affine_param_flag (S3310). At this time, the decoder can perform decoding according to the AF4 mode or the AF6 mode according to the affine_param_flag value at the block level or the prediction unit level.
- the video coding system may include a source device and a receiving device.
- the source device may deliver the encoded video / image information or data in the form of a file or stream to a receiving device via a digital storage medium or network.
- the source device may include a video source, an encoding apparatus, and a transmitter.
- the receiving device may include a receiver, a decoding apparatus, and a renderer.
- the encoding apparatus may be referred to as a video / image encoding apparatus, and the decoding apparatus may be referred to as a video / image decoding apparatus.
- the transmitter may be included in the encoding device.
- the receiver may be included in the decoding apparatus.
- the renderer may include a display unit, and the display unit may be composed of a separate device or an external component.
- a video source can acquire video / image through capturing, compositing, or generating a video / image.
- the video source may include a video / video capture device and / or a video / video generation device.
- the video / video capture device may include, for example, one or more cameras, video / video archives including previously captured video / images, and the like.
- the video / image generation device may include, for example, a computer, tablet, smart phone, and the like (electronically) to generate video / images.
- a virtual video / image may be generated through a computer or the like. In this case, the video / image capturing process may be replaced in the process of generating related data.
- the encoding device may encode the input video / image.
- the encoding apparatus can perform a series of procedures such as prediction, conversion, and quantization for compression and coding efficiency.
- the encoded data (encoded video / image information) can be output in the form of a bitstream.
- the transmitting unit may transmit the encoded video / image information or data output in the form of a bit stream to a receiving unit of the receiving device through a digital storage medium or a network in the form of a file or a stream.
- the digital storage medium may include various storage media such as USB, SD, CD, DVD, Blu-ray, HDD, SSD and the like.
- the transmission unit may include an element for generating a media file through a predetermined file format, and may include an element for transmission over a broadcast / communication network.
- the receiving unit may extract the bitstream and transmit it to the decoding apparatus.
- the decoding apparatus may perform a series of procedures such as inverse quantization, inverse transformation, and prediction corresponding to the operation of the encoding apparatus to decode the video / image.
- the renderer may render the decoded video / image.
- the rendered video / image can be displayed through the display unit.
- 35 shows a content streaming system to which the present invention is applied.
- the content streaming system to which the present invention is applied may include an encoding server, a streaming server, a web server, a media repository, a user device, and a multimedia input device.
- the encoding server compresses content input from multimedia input devices such as a smart phone, a camera, and a camcorder into digital data to generate a bit stream and transmit the bit stream to the streaming server.
- multimedia input devices such as a smart phone, a camera, a camcorder, or the like directly generates a bitstream
- the encoding server may be omitted.
- the bitstream may be generated by an encoding method or a bitstream generating method to which the present invention is applied, and the streaming server may temporarily store the bitstream in the process of transmitting or receiving the bitstream.
- the streaming server transmits multimedia data to a user device based on a user request through the web server, and the web server serves as a medium for informing the user of what services are available.
- the web server delivers it to the streaming server, and the streaming server transmits the multimedia data to the user.
- the content streaming system may include a separate control server. In this case, the control server controls commands / responses among the devices in the content streaming system.
- the streaming server may receive content from a media repository and / or an encoding server. For example, when receiving the content from the encoding server, the content can be received in real time. In this case, in order to provide a smooth streaming service, the streaming server can store the bit stream for a predetermined time.
- Examples of the user device include a mobile phone, a smart phone, a laptop computer, a digital broadcasting terminal, a personal digital assistant (PDA), a portable multimedia player (PMP), a navigation device, a slate PC, Such as tablet PCs, ultrabooks, wearable devices (e.g., smartwatches, smart glass, HMDs (head mounted displays)), digital TVs, desktops Computers, and digital signage.
- PDA personal digital assistant
- PMP portable multimedia player
- slate PC Such as tablet PCs, ultrabooks, wearable devices (e.g., smartwatches, smart glass, HMDs (head mounted displays)), digital TVs, desktops Computers, and digital signage.
- Each of the servers in the content streaming system can be operated as a distributed server. In this case, data received at each server can be distributed.
- the embodiments described in the present invention can be implemented and executed on a processor, a microprocessor, a controller, or a chip.
- the functional units shown in FIGS. 1, 2, 34, and 35 may be implemented and executed on a computer, a processor, a microprocessor, a controller, or a chip.
- the decoder and encoder to which the present invention is applied can be applied to multimedia communication devices such as a multimedia broadcasting transmitting and receiving device, a mobile communication terminal, a home cinema video device, a digital cinema video device, a surveillance camera, a video chatting device, (3D) video device, a video telephony video device, and a medical video device, and can be used to process video signals and data signals, Lt; / RTI >
- the processing method to which the present invention is applied may be produced in the form of a computer-executed program, and may be stored in a computer-readable recording medium.
- the multimedia data having the data structure according to the present invention can also be stored in a computer-readable recording medium.
- the computer-readable recording medium includes all kinds of storage devices in which computer-readable data is stored.
- the computer readable recording medium includes, for example, a Blu-ray Disc (BD), a universal serial bus (USB), a ROM, a RAM, a CD-ROM, a magnetic tape, a floppy disk and an optical data storage device .
- the computer-readable recording medium includes media implemented in the form of a carrier wave (for example, transmission over the Internet).
- the bit stream generated by the encoding method can be stored in a computer-readable recording medium or transmitted over a wired or wireless communication network.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Color Television Systems (AREA)
Abstract
Description
Claims (14)
- 어파인 움직임 예측 모드(AF 모드, Affine mode)에 기초하여 현재 블록을 포함하는 비디오 신호를 디코딩하는 방법에 있어서,상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부를 확인하는 단계, 여기서 상기 AF 모드는 어파인 움직임 모델(affine motion model)을 이용하는 움직임 예측 모드를 나타냄;상기 현재 블록에 대해 상기 AF 모드가 적용되는 경우, AF4 모드가 이용되는지 여부를 확인하는 단계, 여기서 상기 AF4 모드는 상기 어파인 움직임 모델(affine motion model)을 구성하는 4개의 파라미터를 이용하여 움직임 벡터를 예측하는 모드를 나타냄;AF4 모드가 이용되면 상기 4개 파라미터를 이용하여 움직임 벡터 예측자를 생성하고, AF4 모드가 이용되지 않으면 상기 어파인 움직임 모델(affine motion model)을 구성하는 6개 파라미터를 이용하여 움직임 벡터 예측자를 생성하는 단계; 및상기 움직임 벡터 예측자에 기초하여 상기 현재 블록의 움직임 벡터를 획득하는 단계를 포함하는 것을 특징으로 하는 방법.
- 제1항에 있어서, 상기 방법은,상기 비디오 신호로부터 어파인 플래그를 획득하는 단계를 더 포함하되,상기 어파인 플래그는 상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부를 나타내고,상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부는 상기 어파인 플래그에 기초하여 확인되는 것을 특징으로 하는 방법.
- 제2항에 있어서, 상기 방법은,상기 어파인 플래그에 따라 상기 현재 블록에 대해 상기 AF 모드가 적용되는 경우, 상기 비디오 신호로부터 어파인 파라미터 플래그를 획득하는 단계를 더 포함하되,상기 어파인 파라미터 플래그는 상기 움직임 벡터 예측자가 상기 4개의 파라미터를 이용하여 생성되는지 또는 상기 6개의 파라미터를 이용하여 생성되는지 여부를 나타내는 것을 특징으로 하는 방법.
- 제1항에 있어서,상기 어파인 플래그 및 상기 어파인 파라미터 플래그는 슬라이스, 최대 코딩 유닛, 코딩 유닛 또는 예측 유닛 중 적어도 하나의 레벨에서 정의되는 것을 특징으로 하는 방법.
- 제1항에 있어서, 상기 방법은,상기 현재 블록의 크기가 기설정된 조건을 만족하는지를 확인하는 단계를 더 포함하되,상기 기설정된 조건은 상기 현재 블록 내 픽셀 개수, 상기 현재 블록의 너비 및/또는 높이 중 적어도 하나가 기설정된 임계값보다 큰지 여부를 나타내고,상기 현재 블록의 크기가 기설정된 조건을 만족하는 경우, 상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부를 확인하는 단계가 수행되는 것을 특징으로 하는 방법.
- 제5항에 있어서,상기 현재 블록의 크기가 기설정된 조건을 만족하지 않는 경우, 상기 현재 블록은 상기 AF 모드가 아닌 다른 코딩 모드에 기초하여 디코딩되는 것을 특징으로 하는 방법.
- 제1항에 있어서, 상기 방법은,상기 현재 블록에 대해 상기 AF 모드가 적용되는 경우, 이웃 블록에 대해 AF 모드가 적용되었는지 여부를 확인하는 단계를 더 포함하되,상기 이웃 블록에 대해 AF 모드가 적용된 경우, 움직임 벡터 예측자는 상기 4개 파라미터를 이용하여 생성되고,상기 이웃 블록에 대해 AF 모드가 적용되지 않은 경우, 상기 AF4 모드가 이용되는지 여부를 확인하는 단계를 수행하는 것을 특징으로 하는 방법.
- 어파인 움직임 예측 모드(AF 모드)에 기초하여 현재 블록을 포함하는 비디오 신호를 디코딩하는 장치에 있어서,상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부를 확인하고,상기 현재 블록에 대해 상기 AF 모드가 적용되는 경우, AF4 모드가 이용되는지 여부를 확인하고,AF4 모드가 이용되면 4개 파라미터를 이용하여 움직임 벡터 예측자를 생성하고, AF4 모드가 이용되지 않으면 어파인 움직임 모델(affine motion model)을 구성하는 6개 파라미터를 이용하여 움직임 벡터 예측자를 생성하고,상기 움직임 벡터 예측자에 기초하여 상기 현재 블록의 움직임 벡터를 획득하는 인터 예측부를 포함하되,상기 AF 모드는 상기 어파인 움직임 모델(affine motion model)을 이용하는 움직임 예측 모드를 나타내고, 상기 AF4 모드는 상기 어파인 움직임 모델(affine motion model)을 구성하는 4개의 파라미터를 이용하여 움직임 벡터를 예측하는 모드를 나타내는 것을 특징으로 하는 장치.
- 제8항에 있어서, 상기 장치는,상기 비디오 신호로부터 어파인 플래그를 파싱하는 파싱부를 더 포함하되,상기 어파인 플래그는 상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부를 나타내고,상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부는 상기 어파인 플래그에 기초하여 확인되는 것을 특징으로 하는 장치.
- 제9항에 있어서, 상기 장치는,상기 어파인 플래그에 따라 상기 현재 블록에 대해 상기 AF 모드가 적용되는 경우, 상기 비디오 신호로부터 어파인 파라미터 플래그를 획득하는 상기 파싱부를 포함하되,상기 어파인 파라미터 플래그는 상기 움직임 벡터 예측자가 상기 4개의 파라미터를 이용하여 생성되는지 또는 상기 6개의 파라미터를 이용하여 생성되는지 여부를 나타내는 것을 특징으로 하는 장치.
- 제8항에 있어서,상기 어파인 플래그 및 상기 어파인 파라미터 플래그는 슬라이스, 최대 코딩 유닛, 코딩 유닛 또는 예측 유닛 중 적어도 하나의 레벨에서 정의되는 것을 특징으로 하는 장치.
- 제8항에 있어서, 상기 장치는,상기 현재 블록의 크기가 기설정된 조건을 만족하는지를 확인하는 상기 인터 예측부를 포함하되,상기 기설정된 조건은 상기 현재 블록 내 픽셀 개수, 상기 현재 블록의 너비 및/또는 높이 중 적어도 하나가 기설정된 임계값보다 큰지 여부를 나타내고,상기 현재 블록의 크기가 기설정된 조건을 만족하는 경우, 상기 현재 블록에 대해 상기 AF 모드가 적용되는지 여부를 확인하는 단계가 수행되는 것을 특징으로 하는 장치.
- 제12항에 있어서,상기 현재 블록의 크기가 기설정된 조건을 만족하지 않는 경우, 상기 현재 블록은 상기 AF 모드가 아닌 다른 코딩 모드에 기초하여 디코딩되는 것을 특징으로 하는 장치.
- 제8항에 있어서, 상기 장치는,상기 현재 블록에 대해 상기 AF 모드가 적용되는 경우, 이웃 블록에 대해 AF 모드가 적용되었는지 여부를 확인하는 상기 인터 예측부를 포함하되,상기 이웃 블록에 대해 AF 모드가 적용된 경우, 움직임 벡터 예측자는 상기 4개 파라미터를 이용하여 생성되고,상기 이웃 블록에 대해 AF 모드가 적용되지 않은 경우, 상기 AF4 모드가 이용되는지 여부를 확인하는 단계를 수행하는 것을 특징으로 하는 장치.
Priority Applications (14)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020227032285A KR102549228B1 (ko) | 2017-08-03 | 2018-08-03 | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 |
KR1020207004148A KR102345442B1 (ko) | 2017-08-03 | 2018-08-03 | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 |
EP18840631.8A EP3661209A1 (en) | 2017-08-03 | 2018-08-03 | Method and apparatus for processing video signal using affine prediction |
KR1020247014995A KR20240065329A (ko) | 2017-08-03 | 2018-08-03 | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 |
KR1020237021482A KR102664781B1 (ko) | 2017-08-03 | 2018-08-03 | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 |
CN202410534253.0A CN118214883A (zh) | 2017-08-03 | 2018-08-03 | 利用仿射预测处理视频信号的方法和装置 |
CN202410534881.9A CN118214884A (zh) | 2017-08-03 | 2018-08-03 | 利用仿射预测处理视频信号的方法和装置 |
US16/636,263 US11089317B2 (en) | 2017-08-03 | 2018-08-03 | Method and apparatus for processing video signal using affine prediction |
CN201880057060.3A CN111066324B (zh) | 2017-08-03 | 2018-08-03 | 利用仿射预测处理视频信号的方法和装置 |
JP2020505788A JP7393326B2 (ja) | 2017-08-03 | 2018-08-03 | アフィン予測を用いてビデオ信号を処理する方法及び装置 |
KR1020217042670A KR102472053B1 (ko) | 2017-08-03 | 2018-08-03 | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 |
US17/363,648 US11736707B2 (en) | 2017-08-03 | 2021-06-30 | Method and apparatus for processing video signal using affine prediction |
US18/217,126 US20230353768A1 (en) | 2017-08-03 | 2023-06-30 | Method and apparatus for processing video signal using affine prediction |
JP2023198969A JP2024015036A (ja) | 2017-08-03 | 2023-11-24 | アフィン予測を用いてビデオ信号を処理する方法及び装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201762541083P | 2017-08-03 | 2017-08-03 | |
US62/541,083 | 2017-08-03 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/636,263 A-371-Of-International US11089317B2 (en) | 2017-08-03 | 2018-08-03 | Method and apparatus for processing video signal using affine prediction |
US17/363,648 Continuation US11736707B2 (en) | 2017-08-03 | 2021-06-30 | Method and apparatus for processing video signal using affine prediction |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019027286A1 true WO2019027286A1 (ko) | 2019-02-07 |
Family
ID=65232974
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2018/008845 WO2019027286A1 (ko) | 2017-08-03 | 2018-08-03 | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 |
Country Status (6)
Country | Link |
---|---|
US (3) | US11089317B2 (ko) |
EP (1) | EP3661209A1 (ko) |
JP (2) | JP7393326B2 (ko) |
KR (5) | KR102345442B1 (ko) |
CN (3) | CN118214883A (ko) |
WO (1) | WO2019027286A1 (ko) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020251416A3 (en) * | 2019-09-30 | 2021-03-18 | Huawei Technologies Co., Ltd. | Affine motion model restrictions |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11438577B2 (en) * | 2018-03-25 | 2022-09-06 | B1 Institute Of Image Technology, Inc. | Image encoding/decoding method and device |
CN113597767A (zh) * | 2019-03-08 | 2021-11-02 | Oppo广东移动通信有限公司 | 预测方法、编码器、解码器及计算机存储介质 |
KR102640263B1 (ko) * | 2019-06-24 | 2024-02-23 | 엘지전자 주식회사 | 움직임 벡터를 이용한 영상 코딩 방법 및 장치 |
WO2020242350A2 (en) * | 2019-09-30 | 2020-12-03 | Huawei Technologies Co., Ltd. | Usage of dct based interpolation filter and enhanced bilinear interpolation filter in affine motion compensation |
US11363247B2 (en) * | 2020-02-14 | 2022-06-14 | Valve Corporation | Motion smoothing in a distributed system |
CN112055202B (zh) * | 2020-08-21 | 2021-11-16 | 浙江大华技术股份有限公司 | 一种帧间预测方法、视频编码方法及电子设备和存储介质 |
US20230119972A1 (en) * | 2021-10-01 | 2023-04-20 | Mediatek Inc. | Methods and Apparatuses of High Throughput Video Encoder |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9438910B1 (en) * | 2014-03-11 | 2016-09-06 | Google Inc. | Affine motion prediction in video coding |
WO2017118409A1 (en) * | 2016-01-07 | 2017-07-13 | Mediatek Inc. | Method and apparatus for affine merge mode prediction for video coding system |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101003105B1 (ko) * | 2008-01-29 | 2010-12-21 | 한국전자통신연구원 | 어파인 변환 기반의 움직임 보상을 이용한 비디오 부호화 및 복호화 방법 및 장치 |
WO2011013253A1 (ja) * | 2009-07-31 | 2011-02-03 | 株式会社 東芝 | 幾何変換動き補償予測を用いる予測信号生成装置、動画像符号化装置及び動画像復号化装置 |
EP3703373B1 (en) * | 2010-10-08 | 2024-04-17 | GE Video Compression, LLC | Picture coding supporting block partitioning and block merging |
KR102311815B1 (ko) * | 2014-06-19 | 2021-10-13 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | 통합된 인트라 블록 카피 및 인터 예측 모드 |
WO2016008157A1 (en) * | 2014-07-18 | 2016-01-21 | Mediatek Singapore Pte. Ltd. | Methods for motion compensation using high order motion model |
US11082713B2 (en) | 2015-11-20 | 2021-08-03 | Mediatek Inc. | Method and apparatus for global motion compensation in video coding system |
WO2017130696A1 (ja) * | 2016-01-29 | 2017-08-03 | シャープ株式会社 | 予測画像生成装置、動画像復号装置、および動画像符号化装置 |
US10560712B2 (en) * | 2016-05-16 | 2020-02-11 | Qualcomm Incorporated | Affine motion prediction for video coding |
US10448010B2 (en) * | 2016-10-05 | 2019-10-15 | Qualcomm Incorporated | Motion vector prediction for affine motion models in video coding |
US20200014931A1 (en) * | 2018-07-06 | 2020-01-09 | Mediatek Inc. | Methods and Apparatuses of Generating an Average Candidate for Inter Picture Prediction in Video Coding Systems |
US10462488B1 (en) * | 2018-07-13 | 2019-10-29 | Tencent America LLC | Method and apparatus for video coding |
CN110933427B (zh) * | 2018-09-19 | 2023-05-12 | 北京字节跳动网络技术有限公司 | 仿射模式编码的模式相关自适应运动矢量分辨率 |
EP3881533A4 (en) * | 2018-11-14 | 2022-09-07 | Tencent America LLC | VIDEO CODING METHOD AND APPARATUS |
US11006121B2 (en) * | 2019-06-20 | 2021-05-11 | Tencent America LLC | LIC signaling methods |
-
2018
- 2018-08-03 WO PCT/KR2018/008845 patent/WO2019027286A1/ko unknown
- 2018-08-03 KR KR1020207004148A patent/KR102345442B1/ko active IP Right Grant
- 2018-08-03 KR KR1020217042670A patent/KR102472053B1/ko active IP Right Grant
- 2018-08-03 CN CN202410534253.0A patent/CN118214883A/zh active Pending
- 2018-08-03 KR KR1020237021482A patent/KR102664781B1/ko active IP Right Grant
- 2018-08-03 US US16/636,263 patent/US11089317B2/en active Active
- 2018-08-03 EP EP18840631.8A patent/EP3661209A1/en not_active Ceased
- 2018-08-03 JP JP2020505788A patent/JP7393326B2/ja active Active
- 2018-08-03 CN CN201880057060.3A patent/CN111066324B/zh active Active
- 2018-08-03 KR KR1020247014995A patent/KR20240065329A/ko active Application Filing
- 2018-08-03 KR KR1020227032285A patent/KR102549228B1/ko active IP Right Grant
- 2018-08-03 CN CN202410534881.9A patent/CN118214884A/zh active Pending
-
2021
- 2021-06-30 US US17/363,648 patent/US11736707B2/en active Active
-
2023
- 2023-06-30 US US18/217,126 patent/US20230353768A1/en active Pending
- 2023-11-24 JP JP2023198969A patent/JP2024015036A/ja active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9438910B1 (en) * | 2014-03-11 | 2016-09-06 | Google Inc. | Affine motion prediction in video coding |
WO2017118409A1 (en) * | 2016-01-07 | 2017-07-13 | Mediatek Inc. | Method and apparatus for affine merge mode prediction for video coding system |
Non-Patent Citations (3)
Title |
---|
CHEN, JIANLE ET AL.: "Algorithm Description of Joint Exploration Test Model 3", JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3, 1 June 2016 (2016-06-01), Geneva, CH, pages 1 - 34, XP055577005 * |
LI, LI ET AL.: "An Efficient Four-Parameter Affine Motion Model for Video Coding", ARXIV:1702.06297V1, vol. 28, no. 8, 21 February 2017 (2017-02-21), pages 1934 - 1948, XP080747890 * |
ZOU, FENG ET AL.: "Improved Affine Motion Prediction", JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3, 28 May 2016 (2016-05-28), Geneva, CH, pages 1 - 5, XP030150164 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020251416A3 (en) * | 2019-09-30 | 2021-03-18 | Huawei Technologies Co., Ltd. | Affine motion model restrictions |
US11985354B2 (en) | 2019-09-30 | 2024-05-14 | Huawei Technologies Co., Ltd. | Affine motion model restrictions reducing number of fetched reference lines during processing of one block row with enhanced interpolation filter |
Also Published As
Publication number | Publication date |
---|---|
CN118214884A (zh) | 2024-06-18 |
KR102472053B1 (ko) | 2022-11-28 |
KR102664781B1 (ko) | 2024-05-21 |
US11736707B2 (en) | 2023-08-22 |
JP2024015036A (ja) | 2024-02-01 |
JP2020529778A (ja) | 2020-10-08 |
CN111066324A (zh) | 2020-04-24 |
KR102549228B1 (ko) | 2023-06-28 |
KR20200022506A (ko) | 2020-03-03 |
EP3661209A4 (en) | 2020-06-03 |
JP7393326B2 (ja) | 2023-12-06 |
US20230353768A1 (en) | 2023-11-02 |
US11089317B2 (en) | 2021-08-10 |
KR102345442B1 (ko) | 2021-12-31 |
KR20220130270A (ko) | 2022-09-26 |
US20200169744A1 (en) | 2020-05-28 |
EP3661209A1 (en) | 2020-06-03 |
KR20220000990A (ko) | 2022-01-04 |
KR20240065329A (ko) | 2024-05-14 |
CN118214883A (zh) | 2024-06-18 |
US20210329282A1 (en) | 2021-10-21 |
KR20230098377A (ko) | 2023-07-03 |
CN111066324B (zh) | 2024-05-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019027286A1 (ko) | 어파인 예측을 이용하여 비디오 신호를 처리하는 방법 및 장치 | |
WO2018062892A1 (ko) | 가중치 인덱스에 기초하여 최적의 예측을 수행하는 방법 및 장치 | |
WO2020180155A1 (ko) | 비디오 신호를 처리하기 위한 방법 및 장치 | |
WO2021034123A1 (ko) | 가중 예측을 수행하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2019050385A2 (ko) | 비디오 신호를 엔트로피 인코딩, 디코딩하는 방법 및 장치 | |
WO2020251330A1 (ko) | 단순화된 mpm 리스트 생성 방법을 활용하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020055107A1 (ko) | 영상 코딩 시스템에서 어파인 mvp 후보 리스트를 사용하는 어파인 움직임 예측에 기반한 영상 디코딩 방법 및 장치 | |
WO2020180129A1 (ko) | 인터 예측을 위한 비디오 신호의 처리 방법 및 장치 | |
WO2020055161A1 (ko) | 영상 코딩 시스템에서 서브 블록 단위의 움직임 예측에 기반한 영상 디코딩 방법 및 장치 | |
WO2020184952A1 (ko) | 비디오 신호에서 인터 예측을 위한 움직임 벡터 차분 정보를 처리하기 위한 비디오 신호의 처리 방법 및 장치 | |
WO2020036390A1 (ko) | 영상 신호를 처리하기 위한 방법 및 장치 | |
WO2020251328A1 (ko) | 인트라 예측 모드 변환에 기반한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020013541A1 (ko) | 비디오 신호를 처리하기 위한 방법 및 장치 | |
WO2020213976A1 (ko) | Bdpcm을 이용한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020251329A1 (ko) | Mip 모드 매핑이 단순화된 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020130680A1 (ko) | 인트라 예측을 사용하여 비디오 신호를 처리하기 위한 방법 및 장치 | |
WO2021060844A1 (ko) | 팔레트 모드를 이용한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2021172914A1 (ko) | 레지듀얼 코딩에 대한 영상 디코딩 방법 및 그 장치 | |
WO2021006579A1 (ko) | 머지 후보의 양방향 예측을 위한 가중치 인덱스를 유도하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2021040410A1 (ko) | 레지듀얼 코딩에 대한 영상 디코딩 방법 및 그 장치 | |
WO2021006697A1 (ko) | 레지듀얼 코딩에 대한 영상 디코딩 방법 및 그 장치 | |
WO2021015536A1 (ko) | 팔레트 모드의 적용 여부에 따라 디블로킹 필터링을 수행하는 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2021006698A1 (ko) | 영상 코딩 시스템에서 영상 코딩 방법 및 장치 | |
WO2020231144A1 (ko) | 어파인 tmvp를 이용한 영상 부호화/복호화 방법, 장치 및 비트스트림을 전송하는 방법 | |
WO2020032526A1 (ko) | 영상 코딩 시스템에서 컨스트럭티드 어파인 mvp 후보를 사용하는 어파인 움직임 예측에 기반한 영상 디코딩 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18840631 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2020505788 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20207004148 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2018840631 Country of ref document: EP Effective date: 20200226 |