WO2020184264A1 - Dispositif de décodage d'image, procédé de décodage d'image et programme - Google Patents

Dispositif de décodage d'image, procédé de décodage d'image et programme Download PDF

Info

Publication number
WO2020184264A1
WO2020184264A1 PCT/JP2020/008774 JP2020008774W WO2020184264A1 WO 2020184264 A1 WO2020184264 A1 WO 2020184264A1 JP 2020008774 W JP2020008774 W JP 2020008774W WO 2020184264 A1 WO2020184264 A1 WO 2020184264A1
Authority
WO
WIPO (PCT)
Prior art keywords
subblock
image
prediction
block
sub
Prior art date
Application number
PCT/JP2020/008774
Other languages
English (en)
Japanese (ja)
Inventor
圭 河村
内藤 整
Original Assignee
Kddi株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Kddi株式会社 filed Critical Kddi株式会社
Publication of WO2020184264A1 publication Critical patent/WO2020184264A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/583Motion compensation with overlapping blocks

Definitions

  • the present invention relates to an image decoding device, an image decoding method, and a program.
  • Non-Patent Document 1 An image coding method using intra-prediction or inter-prediction, conversion / quantization of a prediction residual signal, and entropy coding has been proposed (see, for example, Non-Patent Document 1).
  • an affine motion compensation prediction method has been proposed in the next-generation moving image coding method WC (see, for example, Non-Patent Document 2).
  • the affine motion compensation prediction method is a kind of motion compensation prediction method, and is configured to generate a prediction image by a motion vector and a translation model.
  • the block size to which the translation model is applied is a subblock of 4 ⁇ 4 pixels, while a different motion vector is derived for each subblock by affine transformation.
  • the affine transformation parameters are expressed as motion vectors at the vertices of the target block for which affine motion compensation prediction is performed in order to be consistent with the existing image coding method.
  • the motion vector at each vertex is called a "control point”.
  • the motion vector (mv x , mv y ) at an arbitrary pixel position in the case of the 4-parameter model shown in FIG. 9 (a) is derived by (Equation 1), and in the case of the 6-parameter model shown in FIG. 9 (b).
  • the motion vector (mv x , mv y ) at an arbitrary pixel position is derived by (Equation 2).
  • (mv 0x , mv 0y ) is the control point (movement vector) of the upper left vertex of the target block (subblock)
  • (mv 1x , mv 1y ) is the control point of the upper right vertex of the target block.
  • (Mv 2x , mv 2y ) are control points of the lower left vertex of the target block.
  • W is the width of the target block
  • H is the height of the target block.
  • affine motion compensation prediction can be performed only in units of 4 ⁇ 4 pixel subblocks, and the prediction performance is low. There was a problem.
  • the present invention has been made in view of the above-mentioned problems, and provides an image decoding device, an image decoding method, and a program capable of improving prediction performance while keeping the rate of increase in the number of subblocks constant.
  • the purpose is.
  • the first feature of the present invention is an image decoding device, which is a subblock division unit configured to acquire a first subblock and a second subblock by dividing a prediction target block at different division positions. And the motion vector generator configured to output the motion vector obtained by the affine transformation corresponding to the first subblock and the second subblock, and the first sub based on the motion vector.
  • the prediction target block is based on the motion compensation prediction unit configured to generate the prediction image of the block and the second subblock, the prediction image of the first subblock, and the prediction image of the second subblock. It is a gist to have an interpolation unit configured to generate a predicted image of.
  • the second feature of the present invention is an image decoding method, in which a step A of acquiring a first subblock and a second subblock by dividing a prediction target block at different division positions, and the first subblock and the first subblock A step B for outputting a motion vector obtained by an affine transformation corresponding to the second subblock, and a step C for generating a predicted image of the first subblock and the second subblock based on the motion vector. It is a gist to have a step D of generating a predicted image of the prediction target block based on the predicted image of the first subblock and the predicted image of the second subblock.
  • a third feature of the present invention is a program that causes a computer to function as an image decoding device, in which the image decoding device divides the prediction target block at different division positions to form a first subblock and a second subblock.
  • a subblock dividing unit configured to acquire the above, and a motion vector generating unit configured to output a motion vector obtained by an affine transformation corresponding to the first subblock and the second subblock.
  • a motion compensation prediction unit configured to generate a prediction image of the first subblock and the second subblock based on the motion vector, a prediction image of the first subblock, and the second subblock.
  • the gist is to have an interpolation unit configured to generate a predicted image of the predicted target block based on the predicted image of the sub-block.
  • an image decoding device an image decoding method and a program capable of improving the prediction performance while suppressing the increase rate of the number of subblocks to a constant level.
  • FIG. 1 is a diagram showing an example of a functional block of the image processing system 1 according to the first embodiment of the present invention.
  • the image processing system 1 includes an image coding device 100 that encodes a moving image and generates coded data, and an image decoding device 200 that decodes the coded data generated by the image coding device 100.
  • the above-mentioned coded data is transmitted and received between the image coding device 100 and the image decoding device 200, for example, via a transmission line.
  • FIG. 2 is a diagram showing an example of a functional block of the image coding device 100.
  • the image coding device 100 includes an inter prediction unit 101, an intra prediction unit 102, a conversion / quantization unit 103, an entropy coding unit 104, and an inverse conversion / inverse quantization unit 105. , An in-loop filter 106 and a frame buffer 107.
  • the inter-prediction unit 101 is configured to perform inter-prediction using the input image and the filtered locally decoded image (described later) input from the frame buffer 109 to generate and output the inter-prediction image.
  • the intra prediction unit 102 is configured to perform intra prediction using an input image and a locally decoded image before filtering (described later) to generate and output an intra prediction image.
  • the conversion / quantization unit 103 performs orthogonal conversion processing on the residual signal input from the subtraction unit 106, and performs quantization processing on the conversion coefficient obtained by such orthogonal conversion processing. It is configured to output the level value.
  • the entropy encoding unit 104 is configured to entropy-encode the quantization coefficient level value, conversion unit size and conversion size, motion compensation method, etc. input from the conversion / quantization unit 103 and output them as encoded data. There is.
  • the inverse conversion / inverse quantization unit 105 performs an inverse quantization process on the quantization coefficient level value input from the conversion / quantization unit 103, and reverses the conversion coefficient obtained by the inverse quantization process. It is configured to output the residual signal obtained by performing the orthogonal conversion process.
  • the subtraction unit 106 is configured to output a residual signal which is a difference between the input image and the intra prediction image or the inter prediction image.
  • the addition unit 107 is configured to output a pre-filter local decoding image obtained by adding the residual signal input from the inverse transformation / inverse quantization unit 105 and the intra-prediction image or the inter-prediction image.
  • the in-loop filter unit 108 applies in-loop filter processing such as deblocking filter processing to the pre-filter local decoding image input from the addition unit 107 to generate and output the post-filter local decoding image. It is configured.
  • the frame buffer 109 accumulates the filtered locally decoded image and appropriately supplies it to the inter-prediction unit as the filtered locally decoded image.
  • FIG. 3 is a block diagram of the image decoding device 200.
  • the image decoding device 200 includes an entropy decoding unit 201, an inverse conversion / inverse quantization unit 202, an inter prediction unit 203, an intra prediction unit 204, an addition unit 205, and an in-loop filter 206. And a frame buffer 207.
  • the entropy decoding unit 201 is configured to entropy decode the encoded data, derive the quantization coefficient level value, the conversion unit size and the conversion size, the motion compensation method, and the like, and output the coded data.
  • the inverse conversion / inverse quantization unit 202 performs an inverse quantization process on the quantization coefficient level value input from the entropy decoding unit 201, and an inverse orthogonal conversion process on the result obtained by the inverse quantization process. Is configured to be output as a residual signal.
  • the inter-prediction unit 203 is configured to perform inter-prediction using the filtered locally decoded image input from the frame buffer 207 to generate and output the inter-prediction image.
  • the intra prediction unit 204 is configured to perform intra prediction using the pre-filter locally decoded image input from the addition unit 205 to generate and output an intra prediction image.
  • the addition unit 205 combines the residual signal input from the inverse transformation / inverse quantization unit 202 with the prediction image (inter prediction image input from the inter prediction unit 203 or intra prediction image input from the intra prediction unit 204). It is configured to output the pre-filter locally decoded image obtained by addition.
  • the prediction image is a prediction with the highest expected coding performance obtained by entropy decoding among the inter prediction image input from the inter prediction unit 203 and the intra prediction image input from the intra prediction unit 204. It is a predicted image calculated by the method.
  • the in-loop filter 206 is configured to apply in-loop filter processing such as deblocking filter processing to the pre-filter locally decoded image input from the addition unit 205 to generate and output the post-filtered locally decoded image. ing.
  • the frame buffer 207 is configured to accumulate the filtered locally decoded image input from the in-loop filter 206, appropriately supply it to the inter-prediction unit 203 as the filtered locally decoded image, and output it as a decoded image. There is.
  • the inter-prediction unit 101 of the image coding device 100 and the inter-prediction unit 203 of the image decoding device 200 will be described with reference to FIGS. 4 to 6. Since the functions of the inter-prediction unit 101 of the image coding device 100 and the functions of the inter-prediction unit 203 of the image decoding device 200 are basically the same, the functions of the inter-prediction unit 203 of the image decoding device 200 will be taken as an example below. I will explain it by listing it.
  • the inter-prediction unit 203 of the image decoding apparatus 200 includes a normal sub-block division unit 101A, an additional sub-block division unit 101B, a motion vector generation unit 101C, and an affine motion vector generation unit 101D1 / 101D2.
  • the frame buffer 101E, the motion compensation prediction units 101F1 to 101F3, the addition unit 101G, and the selection unit 101H are provided.
  • the sub-block division unit 101A is configured to acquire the first sub-block by dividing the prediction target block at the first division position.
  • the normal sub-block division unit 101A divides the prediction target block based on the input coordinates and block size of the prediction target block to acquire the first sub-block, and obtains the first sub-block, and the coordinates of the first sub-block and It is configured to output the block size.
  • the normal subblock division unit 101A divides the prediction target block at the first division position X, and divides 16 subblocks (blocks composed of 4 ⁇ 4 pixels) B1. It may be configured to acquire.
  • the additional sub-block division unit 101B is configured to acquire the second sub-block by dividing the prediction target block at the second division position, which is a position deviated from the first division position by a predetermined pixel.
  • the additional sub-block division unit 101B grasps the first division position from the input coordinates and block size of the prediction target block, and is a position shifted by a predetermined pixel from the first division position. It is configured to divide the prediction target block at the division position, acquire the second subblock, and output the coordinates and block size of the second subblock.
  • the second division position may be a position shifted in the vertical direction and the horizontal direction by half the pixels of the first subblock from the first division position.
  • the additional sub-block division unit 101B divides the prediction target block at the second division position Y shifted in the vertical and horizontal directions by two pixels from the first division position X.
  • Nine second subblocks (blocks composed of 4 ⁇ 4 pixels) B2 may be acquired.
  • the additional subblock division 101B has the coordinates of nine second subblocks composed of 4 ⁇ 4 pixels and the coordinates and It is configured to output the block size.
  • the normal subblock division unit 101A and the additional subblock division unit 101B are configured to acquire the first subblock and the second subblock by dividing the prediction target block at different division positions. It constitutes a division part.
  • the motion vector generation unit 101C is configured to acquire the input coordinates and control points of the prediction target block and output the control points at the upper right vertices of the prediction target block as motion vectors.
  • the affine motion vector generation unit 101D1 is configured to output the motion vector obtained by the affine transformation corresponding to the first subblock.
  • the affine motion vector generation unit 101D2 is configured to output the motion vector obtained by the affine transformation corresponding to the second subblock.
  • the affine motion vector generation units 101D1 / 101D2 perform affine transformation using the input coordinates, block size and control points of the first subblock and the second subblock, and perform the affine transformation, and the first subblock and the second subblock and the second subblock. It is configured to output the motion vector corresponding to the subblock.
  • the frame buffer 101E is configured to output a fractional part of the reference pixel and motion vector indicated by the reference pixel subscript based on the input reference image subscript and motion vector of the block (or subblock). ..
  • the frame buffer 101E may be substituted by the frame buffer 109/207.
  • the motion compensation prediction unit 101F1 / 101F2 is configured to generate prediction images (interpolated images) of the first subblock and the second subblock based on the above-mentioned motion vector.
  • the motion compensation prediction unit 101F1 is configured to generate a prediction image of the second subblock based on the motion vector corresponding to the second subblock output from the frame buffer 101E.
  • the motion compensation prediction unit 101F1 is configured to generate a prediction image of the second subblock from the reference pixel output from the frame buffer 101E and the fractional part of the motion vector corresponding to the second subblock. Has been done.
  • the motion compensation prediction unit 101F2 is configured to generate a prediction image of the first subblock based on the motion vector corresponding to the first subblock output from the frame buffer 101E.
  • the motion compensation prediction unit 101F2 is configured to generate a prediction image of the first subblock from a reference pixel output from the frame buffer 101E and a fractional part of the motion vector corresponding to the first subblock. Has been done.
  • the motion compensation prediction unit 101F3 is configured to generate a prediction image of the prediction target block from the reference pixel output from the frame buffer 101E and a fractional part of the motion vector corresponding to the prediction target block.
  • the addition unit 101G constitutes an interpolation unit configured to generate a prediction image of the prediction target block based on the prediction image of the first subblock and the prediction image of the second subblock.
  • the addition unit 101G is configured to generate a prediction image of the prediction target block by weighted averaging the prediction image of the first subblock and the prediction image of the second subblock at a constant ratio. ing.
  • the selection unit 101H responds to the inter-prediction mode from the prediction image of the prediction target block output from the addition unit 101G and the prediction image of the prediction target block (when no subblock is used) output from the motion compensation prediction unit 101F3. It is configured to select and output a predicted image. The selection is determined by a control unit (not shown), and is signaled from the image coding device 100 to the image decoding device 200 as a motion compensation method.
  • step S101 the image decoding device 200 acquires the first subblock and the second subblock by dividing the prediction target block.
  • step S102 the image decoding device 200 generates a motion vector corresponding to the first subblock, a motion vector corresponding to the second subblock, and a motion vector corresponding to the prediction target subblock.
  • step S103 the image decoding device 200 generates a predicted image of the first subblock based on the motion vector corresponding to the first subblock, and generates a predicted image of the second subblock based on the motion vector corresponding to the second subblock.
  • a prediction image is generated, and a prediction image (when no subblock is used) of the prediction target block is generated based on the motion vector corresponding to the prediction target block.
  • step S104 the image decoding device 200 generates a predicted image of the prediction target block by weighted averaging the predicted image of the first subblock and the predicted image of the second subblock at a constant ratio.
  • step S105 the image decoding device 200 selects a prediction image according to the inter-prediction mode from the prediction image of the prediction target block generated in step S104 and the prediction image of the prediction target block generated in step S103.
  • the area of the reference pixel of the second subblock is not wider than the area of the reference pixel of the first subblock, so that each subblock does not increase the memory bandwidth. Prediction performance can be improved while maintaining the processing unit of.
  • the additional sub-block division unit 101B is configured to output a part of the second sub-block.
  • the additional subblock division unit 101B outputs the second subblock only to a certain ratio (for example, half, etc.) or less of the number of the first subblocks normally acquired by the subblock division unit 101A. It is configured as follows.
  • 16 first subblocks are normally generated by the subblock dividing unit 101A, and 9 additional subblocks 101B are generated by the additional subblock dividing unit 101B as shown in FIG. 7A.
  • a second subblock is generated.
  • the additional subblock division portion 101B is a shaded portion excluding a part of the second subblocks as shown in FIG. 7A. It is configured to output eight second subblocks.
  • the additional subblock dividing unit 101B outputs 49 second subblocks in the above-described first embodiment. In this embodiment, it may be configured to output 32 or less second subblocks.
  • the additional sub-block dividing unit 101B is configured to output 25 second sub-blocks when the second sub-block is reduced so as to have a checkered pattern.
  • the additional subblock dividing unit 101B is configured to output the first n second subblocks of the arbitrarily ordered blocks.
  • n is an integer, and is "32" when the rate of increase is 50%.
  • the prediction accuracy is predicted even though the additional calculation cost increases by a certain ratio even in the worst case. Can be improved.
  • Image processing system 100 ... Image coding devices 101, 203 ... Inter-prediction unit 101A ... Normal sub-block division unit 101B ... Additional sub-block division 101C ... Motion vector generation units 101D1, 101D2 ... Affine motion vector generation units 101F1 to 101F3 ... Motion compensation prediction unit 102, 204 ... Intra-prediction unit 103 ... Conversion / quantization unit 104 ... Entropy coding unit 105, 202 ... Inverse conversion / inverse quantization unit 106 ... Subtraction unit 101G, 107, 205 ... Addition unit 108, 206 ... In-loop filters 101E, 109, 207 ... Frame buffer 200 ... Entropy decoding unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Un dispositif de décodage d'image (200) comprend : des unités de division de sous-bloc (101A/101B) destinées à acquérir un premier sous-bloc et un second sous-bloc en divisant un bloc cible de prédiction au niveau de différentes positions de division ; des unités de génération de vecteur de mouvement (101D1/101D2) destinées à délivrer en sortie des vecteurs de mouvement obtenus par des transformations affines correspondant au premier sous-bloc et au second sous-bloc ; des unités de prédiction de compensation de mouvement (101F1/101F2) destinées à générer des images prédites du premier sous-bloc et du second sous-bloc sur la base des vecteurs de mouvement ; et une unité d'interpolation (101G) destinée à générer une image prédite du bloc cible de prédiction sur la base de l'image prédite du premier sous-bloc et de l'image prédite du second sous-bloc.
PCT/JP2020/008774 2019-03-11 2020-03-02 Dispositif de décodage d'image, procédé de décodage d'image et programme WO2020184264A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2019-044305 2019-03-11
JP2019044305A JP7026064B2 (ja) 2019-03-11 2019-03-11 画像復号装置、画像復号方法及びプログラム

Publications (1)

Publication Number Publication Date
WO2020184264A1 true WO2020184264A1 (fr) 2020-09-17

Family

ID=72427904

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/008774 WO2020184264A1 (fr) 2019-03-11 2020-03-02 Dispositif de décodage d'image, procédé de décodage d'image et programme

Country Status (2)

Country Link
JP (1) JP7026064B2 (fr)
WO (1) WO2020184264A1 (fr)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180070102A1 (en) * 2015-05-15 2018-03-08 Huawei Technologies Co., Ltd. Adaptive Affine Motion Compensation Unit Determing in Video Picture Coding Method, Video Picture Decoding Method, Coding Device, and Decoding Device
WO2020008334A1 (fr) * 2018-07-01 2020-01-09 Beijing Bytedance Network Technology Co., Ltd. Dérivation efficace de vecteur de mouvement par connexion affine

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180070102A1 (en) * 2015-05-15 2018-03-08 Huawei Technologies Co., Ltd. Adaptive Affine Motion Compensation Unit Determing in Video Picture Coding Method, Video Picture Decoding Method, Coding Device, and Decoding Device
WO2020008334A1 (fr) * 2018-07-01 2020-01-09 Beijing Bytedance Network Technology Co., Ltd. Dérivation efficace de vecteur de mouvement par connexion affine

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KAI ZHANG ET AL.: "CE4-related: Interweaved Prediction for Affine Motion Compensation", JOINT VIDEO EXPERTS TEAM (JVET) OF ITU-T SG 16 WP 3 AND ISO/ IEC JTC 1/SC 29/WG 11, JVET-K0102-V2, 11TH MEETING, July 2018 (2018-07-01), Ljubljana, SI, pages 1 - 4 *
KAI ZHANG ET AL.: "Non-CE2: Interweaved Prediction for Affine Motion Compensation", JOINT VIDEO EXPERTS TEAM (JVET) OF ITU-T SG 16 WP 3 AND ISO/ IEC JTC 1/SC 29/WG 11, JVET-M0268-V3 , 13TH MEETING, January 2019 (2019-01-01), Marrakech, MA, pages 1 - 4 *

Also Published As

Publication number Publication date
JP7026064B2 (ja) 2022-02-25
JP2020150342A (ja) 2020-09-17

Similar Documents

Publication Publication Date Title
KR102668077B1 (ko) 영상 부호화 및 복호화 장치 및 그 방법
JP6660074B2 (ja) 映像復号化方法及び装置
JP7445799B2 (ja) 画像復号装置及び画像復号方法
JP5750579B2 (ja) イントラ予測モード復号化方法及び装置
US8031778B2 (en) Picture decoding device and method
JP6352141B2 (ja) 動画像符号化装置、動画像復号装置、動画像圧縮伝送システム、動画像符号化方法、動画像復号方法、およびプログラム
JPWO2009047917A1 (ja) 映像符号化方法及び映像復号方法
KR20110015625A (ko) 움직임 보상 보간 및 참조 픽쳐 필터링에 대한 국부적 적응형 필터링을 위한 방법 및 장치
JP2007274479A (ja) 画像符号化装置及び画像復号化装置
JP5898452B2 (ja) イントラ予測装置、符号化装置、復号装置、及びプログラム
JP2018107692A (ja) 動画像復号装置、動画像復号方法、動画像符号化装置、動画像符号化方法及びコンピュータ可読記録媒体
JP2024024080A (ja) 画像符号化装置、画像符号化方法、画像復号装置、画像復号方法
JP2023181491A (ja) 画像復号装置、画像復号方法及びプログラム
WO2020184264A1 (fr) Dispositif de décodage d'image, procédé de décodage d'image et programme
JP2007288810A (ja) 画像符号化方法,画像復号方法,画像符号化装置,画像復号装置,画像符号化プログラム,画像復号プログラムおよびコンピュータ読み取り可能な記録媒体
JP2013138502A (ja) 符号化装置及びその制御方法、コンピュータプログラム
JP7034363B2 (ja) 画像復号装置、画像復号方法及びプログラム
JP7134118B2 (ja) 画像復号装置、画像復号方法及びプログラム
CN111903132A (zh) 图像处理装置和方法
WO2021131464A1 (fr) Dispositif et procédé de décodage d'image et programme
WO2021131595A1 (fr) Dispositif de décodage d'images, procédé de décodage d'images et programme
JP6917718B2 (ja) 予測装置、符号化装置、復号装置、及びプログラム
JP2022087335A (ja) 画像復号装置、画像復号方法及びプログラム
JP2022079469A (ja) 画像復号装置、画像復号方法及びプログラム
WO2011142221A1 (fr) Dispositif de codage et dispositif de décodage

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20770002

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20770002

Country of ref document: EP

Kind code of ref document: A1