WO2016044979A1 - Prédiction segmentaire pour codage vidéo - Google Patents

Prédiction segmentaire pour codage vidéo Download PDF

Info

Publication number
WO2016044979A1
WO2016044979A1 PCT/CN2014/087094 CN2014087094W WO2016044979A1 WO 2016044979 A1 WO2016044979 A1 WO 2016044979A1 CN 2014087094 W CN2014087094 W CN 2014087094W WO 2016044979 A1 WO2016044979 A1 WO 2016044979A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
segment
segmental
value
pixel
Prior art date
Application number
PCT/CN2014/087094
Other languages
English (en)
Inventor
Kai Zhang
Jicheng An
Xianguo Zhang
Han HUANG
Original Assignee
Mediatek Singapore Pte. Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mediatek Singapore Pte. Ltd. filed Critical Mediatek Singapore Pte. Ltd.
Priority to PCT/CN2014/087094 priority Critical patent/WO2016044979A1/fr
Priority to PCT/CN2015/082074 priority patent/WO2015196966A1/fr
Priority to US15/032,205 priority patent/US10244258B2/en
Priority to CN201580001847.4A priority patent/CN105556968B/zh
Publication of WO2016044979A1 publication Critical patent/WO2016044979A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/20Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
    • H04N19/21Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding with binary alpha-plane coding for video objects, e.g. context-based arithmetic encoding [CAE]

Definitions

  • the invention relates generally to video/image processing.
  • Prediction takes a critical role in video coding.
  • a prediction block which is generated by intra-prediction or inter-prediction, is obtained first before producing residues at encoder or reconstructing the reconstruction samples at decoder.
  • IBC intra-block copy
  • SCC screen content coding
  • HEVC high efficiency video coding
  • Inter simplified depth coding is adopted into 3D-HEVC as a special prediction mode for depth coding.
  • InterSDC Inter simplified depth coding
  • a normal inter-prediction is performed for the current block first.
  • each pixels in the prediction block is added by a coded offset.
  • P i, , j represents the prediction value at pixel position (i, j) after performing the normal inter-prediction
  • Offset is the offset coded for this block.
  • the final prediction value at pixel position is P i, , j + Offset
  • no residues are coded.
  • the final prediction value will be output as the reconstructed value.
  • DLT Depth lookup table
  • SDC intra simplified depth coding
  • DMM depth map modeling
  • Fig. 2 demonstrates an example of DLT approach.
  • DLT is signaled in picture parameter set (PPS) . And it is left to be an encoder issue how to get the DLT when encoding.
  • Fig. 3 and Fig. 4 demonstrate two examples where there are two and three segments with sharp edges in a block.
  • the prediction block is processed by a segmental process before it is used to get residues at an encoder or get reconstruction at a decoder.
  • the segmental prediction method comprises classifying pixels in the prediction block into different categories, named as ‘segment’ , wherein the pixels in the segment can be adjacent or not, treating the pixels in different segments in different ways, and forming a new prediction block after the treatment.
  • the new prediction block is used to get the residues at the encoder or get the reconstruction at the decoder.
  • Fig. 1 is a diagram illustrating intra block copying
  • Fig. 2 is a diagram illustrating is a diagram illustrating an example of DLT where five valid values appear in depth samples.
  • Fig. 3 is a diagram illustrating a block with two segments with sharp sample value changes
  • Fig. 4 is a diagram illustrating a block with three segments with sharp sample value changes
  • Fig. 5 is a diagram illustrating an exemplary segmental prediction architecture at the decoder
  • Fig. 6 is a diagram illustrating an exemplary segmental process architecture
  • Fig. 7 is a diagram illustrating an exemplary treatment for a segment. After the treatment, pixels in the segment can hold different values;
  • Fig. 8 is a diagram illustrating an exemplary treatment for a segment. After the treatment, pixels in the segment hold only one value.
  • a segmental prediction method is proposed.
  • the prediction block is processed by a segmental process before it is used to get the residues at encoder or get the reconstruction at decoder.
  • the prediction block is processed by a segmental process, and then the modified prediction block is output as the reconstruction without adding to residues.
  • Fig. 5 demonstrates segmental prediction architecture at the decoder.
  • the prediction block can be obtained by intra-prediction, inter-prediction, intra-block copy prediction or any combination of them.
  • a part of the prediction block can be obtained by inter-prediction, and another part of the prediction block can be obtained by intra-block copy prediction.
  • Fig. 6 there are two steps generally as depicted in Fig. 6.
  • the first step is called ‘classification’ , in which the pixels in a prediction block are classified into different categories, named as ‘segment’ . Pixels in a segment can be adjacent or not.
  • the second step is called ‘treatment’ , in which pixels in different segments are treated in different ways. Finally, all the pixels after the treatment form a new prediction block which is output then.
  • the number of segments can be any positive integers such as 1, 2, 3, etc.
  • the prediction values can be classified according to their values.
  • the prediction values can be classified according to their positions.
  • the prediction values can be classified according to their gradients. Gradients can be got by applying operators such as Sobel, Roberts, and Prewitt.
  • classification is not applied if the number of segments is one.
  • the prediction block is classified into two segments in the classification step.
  • a pixel is classified according to its relationship with a threshold number T. In one example, a pixel is classified into segment 0 if its value is larger than T, Otherwise it is classified into segment 1. In another example, a pixel is classified into segment 0 if its value is larger than or equal to T, Otherwise it is classified into segment 1.
  • T is calculated as the average value of all the pixel values in the prediction block.
  • T is calculated as the middle value of all the pixel values in the prediction block.
  • the prediction block is classified into M (M>2) segments in the classification step.
  • M is equal to 3
  • a pixel is classified into segment 0 if its value is smaller than T 1 ; it is classified into segment 2 if its value is larger than T 2 ; otherwise it is classified into segment 1.
  • a pixel is classified into segment 0 if its value is smaller than or equal to T 1 ; it is classified into segment 2 if its value is larger than T 2 ; otherwise it is classified into segment 1.
  • T k is calculated as a function of all the pixel values in the prediction block, where k if from 1 to M-1.
  • T k f k (P) , where P represents all the pixel values in the prediction block.
  • T k is calculated as
  • T 1 (T+Vmin) /2
  • T 2 (Vmax+T) /2
  • Vmax and Vmin are the minimum and the maximum pixel value in the prediction block respectively.
  • an offset Off U is added to a pixel in a segment, denoted as segment U, in the treatment process to get the new prediction value.
  • Vnew Vold+ Off U , where Vold is the pixel value before the treatment and Vnew is the pixel value after the treatment respectively.
  • E U is calculated as the average value of all the pixel prediction values in segment U.
  • E U is calculated as the middle value of all the pixel prediction values in segment U.
  • E U is calculated as the average value of the minimum and the maximum pixel prediction value in segment U.
  • E U (V U max+V U min) /2, where V U max and V U min are the minimum and the maximum pixel prediction value in segment U respectively.
  • E U is calculated as the mode value of the pixel values in segment U.
  • the mode value is defined as the value that appears most often in segment U.
  • a exemplary procedure to get the mode value of the pixel values in segment U is as follows.
  • Min V and Max V are possible the minimal and maximal pixel values respectively.
  • Min V is 0 and Max V is 255.
  • a variable Count [i] is initialized to be 0.
  • Count [v] is initialized to be 0.
  • Count [v] is the largest in Count [i] s with i from Min V to Max V.
  • the offset Off U for segment U can be signaled explicitly by the encoder to decoder, or it can be derived implicitly by the decoder.
  • the offset Off U for segment U is calculated by the encoder according to the pixel original values in segment U and pixel prediction values in segment U. For example, the offset Off U for segment U is calculated by the encoder by subtracting the average value of all the pixel original values in segment U and the average value of all the pixel prediction values in segment U.
  • the offset Off U for segment U is calculated by the encoder by subtracting the average value of all the pixel original values in segment U and E U .
  • OffIdx U instead of Off U is coded.
  • OffIdx U is the DLT index offset.
  • a flag is signaled to indicate whether Off or OffIdx for all segments are zero in the block if segmental prediction is applied. If the condition holds, then no Off or OffIdx for segments in the block is signaled and all Off or OffIdx’s are implied as 0.
  • Off or OffIdx if the flag indicates that at least one Off or OffIdx for a segments is not zero in the block if segmental prediction is applied, and all Off or OffIdx’s for segments before the last segments are signaled to be 0, then the Off or OffIdx for the last segment cannot be 0.
  • Off-1 or OffIdx-1 instead of Off or OffIdx for the last segment should be coded.
  • the decoded value plus 1 is assigned to Off or OffIdx for the last segment.
  • the DLT index offset OffIdx U for segment U is calculated by the encoder by subtracting DLT index of the average value of all the pixel original values in segment U and the DLT index of E U .
  • OffIdx U f (A U ) -f (E U ) , where A U is the average value of all the original pixel values in segment U and f represents a function mapping a depth value to a DLT index.
  • the segmental prediction method can be used or not adaptively.
  • the encoder can send the information of whether to use the segmental prediction method to the decoder explicitly. Or the decoder can derive whether to use the segmental prediction method in the same way as the encoder implicitly.
  • residues of a block are not signaled and implied to be all 0 if segmental prediction is applied in the block.
  • the segmental prediction method can be applied on coding tree unit (CTU) , coding unit (CU) , prediction unit (PU) or transform unit (TU) .
  • CTU coding tree unit
  • CU coding unit
  • PU prediction unit
  • TU transform unit
  • the encoder can send the information of whether to use the segmental prediction method to the decoder in video parameter set (VPS) , sequence parameter set (SPS) , picture parameter set (PPS) , slice header (SH) , CTU, CU, PU, or TU.
  • VPS video parameter set
  • SPS sequence parameter set
  • PPS picture parameter set
  • SH slice header
  • the segmental prediction method can only be applied for CU with some particular sizes. For example, it can only be applied to a CU with size larger than 8x8. In another example, it can only be applied to a CU with size smaller than 64x64.
  • the segmental prediction method can only be applied for CU with some particular PU partition. For example, it can only be applied to a CU with 2Nx2N partition.
  • the segmental prediction method can only be applied for CU with some particular coding mode. For example, it can only be applied to a CU with IBC mode.
  • the segmental prediction method can only be applied for CU with InterSDC mode.
  • the number of segments in the segmental prediction method is adaptively.
  • the encoder can send the information of how many segments to the decoder explicitly when the segmental prediction method is used. Or the decoder can derive the number in the same way as the encoder implicitly.
  • the encoder can send the information of how many segments to the decoder in video parameter set (VPS) , sequence parameter set (SPS) , picture parameter set (PPS) , slice header (SH) , CTU, CU, PU, or TU where the segmental prediction method is used.
  • VPS video parameter set
  • SPS sequence parameter set
  • PPS picture parameter set
  • SH slice header
  • the encoder can send the information of the offsets for each segment to the decoder in video parameter set (VPS) , sequence parameter set (SPS) , picture parameter set (PPS) , slice header (SH) , CTU, CU, PU, or TU where the segmental prediction method is used.
  • VPS video parameter set
  • SPS sequence parameter set
  • PPS picture parameter set
  • SH slice header
  • the encoder can send the information of whether to use the segmental prediction method to the decoder in a CU coded with InterSDC mode.
  • the encoder can send the information of how many segments to the decoder in a CU coded with InterSDC mode.
  • the encoder can send the information of the offsets or DLT index offsets for each segment to the decoder in a CU coded with InterSDC and in which the segmental prediction method is used.
  • the segmental prediction method can be applied to the texture component. It can also be applied to depth components in 3D video coding.
  • the segmental prediction method can be applied to the luma component. It can also be applied to chroma components.
  • the decision of whether to use segmental prediction method can be made separately for each component, with information signaled separately. Or, the decision of whether to use segmental prediction method can be made together for all components, with a single piece of information signaled.
  • the number of segments when segmental prediction method is used can be controlled separately for each component separately, with information signaled separately.
  • the number of segments when segmental prediction method is used can be controlled together for all components, with a single piece of information signaled.
  • the offsets for each segment when segmental prediction method is used can be decided separately for each component separately, with information signaled separately.
  • the offsets for each segment when segmental prediction method is used can be decided together for all components, with a single piece of information signaled.
  • an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein.
  • An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
  • DSP Digital Signal Processor
  • the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA) .
  • processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention.
  • the software code or firmware codes may be developed in different programming languages and different format or style.
  • the software code may also be compiled for different target platform.
  • different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé de codage par prédiction segmentaire. La classification de pixels selon la prédiction segmentaire en différents segments traités de différentes manières permet d'améliorer l'efficacité d'un codage par prédiction.
PCT/CN2014/087094 2014-06-23 2014-09-22 Prédiction segmentaire pour codage vidéo WO2016044979A1 (fr)

Priority Applications (4)

Application Number Priority Date Filing Date Title
PCT/CN2014/087094 WO2016044979A1 (fr) 2014-09-22 2014-09-22 Prédiction segmentaire pour codage vidéo
PCT/CN2015/082074 WO2015196966A1 (fr) 2014-06-23 2015-06-23 Procédé de prédiction segmentaire pour des données de profondeur et de texture dans des systèmes de codage tridimensionnel (3d) et multi-vue
US15/032,205 US10244258B2 (en) 2014-06-23 2015-06-23 Method of segmental prediction for depth and texture data in 3D and multi-view coding systems
CN201580001847.4A CN105556968B (zh) 2014-06-23 2015-06-23 三维或多视图视频编码系统中预测编码的装置及方法

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2014/087094 WO2016044979A1 (fr) 2014-09-22 2014-09-22 Prédiction segmentaire pour codage vidéo

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/089040 Continuation-In-Part WO2016061743A1 (fr) 2014-06-23 2014-10-21 Prédiction par segments pour le codage vidéo

Related Child Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/080515 Continuation-In-Part WO2015196333A1 (fr) 2014-06-23 2014-06-23 Prédiction par segments pour le codage vidéo

Publications (1)

Publication Number Publication Date
WO2016044979A1 true WO2016044979A1 (fr) 2016-03-31

Family

ID=55580028

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/087094 WO2016044979A1 (fr) 2014-06-23 2014-09-22 Prédiction segmentaire pour codage vidéo

Country Status (1)

Country Link
WO (1) WO2016044979A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110381391A (zh) * 2019-07-11 2019-10-25 北京字节跳动网络技术有限公司 视频快速切片方法、装置及电子设备

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6798834B1 (en) * 1996-08-15 2004-09-28 Mitsubishi Denki Kabushiki Kaisha Image coding apparatus with segment classification and segmentation-type motion prediction circuit
CN103004206A (zh) * 2010-07-20 2013-03-27 株式会社Ntt都科摩 图像预测编码装置、图像预测编码方法、图像预测编码程序、图像预测解码装置、图像预测解码方法以及图像预测解码程序
CN103108181A (zh) * 2007-03-23 2013-05-15 三星电子株式会社 用于图像编码和图像解码的方法和设备
CN103650498A (zh) * 2012-06-04 2014-03-19 松下电器产业株式会社 运动图像编码方法、运动图像编码装置、运动图像解码方法及运动图像解码装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6798834B1 (en) * 1996-08-15 2004-09-28 Mitsubishi Denki Kabushiki Kaisha Image coding apparatus with segment classification and segmentation-type motion prediction circuit
CN103108181A (zh) * 2007-03-23 2013-05-15 三星电子株式会社 用于图像编码和图像解码的方法和设备
CN103004206A (zh) * 2010-07-20 2013-03-27 株式会社Ntt都科摩 图像预测编码装置、图像预测编码方法、图像预测编码程序、图像预测解码装置、图像预测解码方法以及图像预测解码程序
CN103650498A (zh) * 2012-06-04 2014-03-19 松下电器产业株式会社 运动图像编码方法、运动图像编码装置、运动图像解码方法及运动图像解码装置

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110381391A (zh) * 2019-07-11 2019-10-25 北京字节跳动网络技术有限公司 视频快速切片方法、装置及电子设备
CN110381391B (zh) * 2019-07-11 2021-11-09 北京字节跳动网络技术有限公司 视频快速切片方法、装置及电子设备

Similar Documents

Publication Publication Date Title
CN108028920B (zh) 视频编解码中高级去块滤波的方法以及装置
US20190068977A1 (en) Method and apparatus of localized luma prediction mode inheritance for chroma prediction in video coding
WO2017041676A1 (fr) Procédé et appareil de modélisation de contexte pour éléments syntaxiques en codage d'images et de vidéos
JP5932332B2 (ja) 画像補正に対する修復技術の使用
CA2934116C (fr) Procede et appareil pour prediction de table de palette
TWI677239B (zh) 結合多個去雜訊化技術及並行分組圖像補塊的非局部適應性環路濾波器
US20130170558A1 (en) Video decoding using block-based mixed-resolution data pruning
BRPI0717936B1 (pt) Método e aparelhos para a filtragem de remoção de artefato em laço
KR102378321B1 (ko) 적어도 하나의 이미지 유닛을 인코딩하는 방법 및 디바이스, 및 적어도 하나의 이미지 유닛을 나타내는 스트림을 디코딩하는 방법 및 디바이스
US10499075B2 (en) Method for coding a depth lookup table
US10244258B2 (en) Method of segmental prediction for depth and texture data in 3D and multi-view coding systems
WO2014139567A1 (fr) Procédé de codage reposant sur des sous-plages d'une table de consultation de profondeur
WO2015135473A1 (fr) Procédé et appareil de mode mono-échantillon pour codage vidéo
JP6084687B2 (ja) ビデオストリームを符号化するための方法および装置
WO2016044979A1 (fr) Prédiction segmentaire pour codage vidéo
US20190124361A1 (en) Devices and methods for 3d video coding
WO2016061743A1 (fr) Prédiction par segments pour le codage vidéo
WO2019002615A1 (fr) Appareil de décodage et de codage prédictif basé sur des blocs et procédés correspondants
WO2015196333A1 (fr) Prédiction par segments pour le codage vidéo
WO2020007747A1 (fr) Déblocage d'échantillons de référence intra
WO2024074131A1 (fr) Procédé et appareil pour hériter des paramètres de modèle inter-composantes dans un système de codage vidéo
WO2016115733A1 (fr) Améliorations de la prédiction de résidus intra-composante
WO2024074129A1 (fr) Procédé et appareil pour hériter de paramètres de modèle voisin temporel dans un système de codage vidéo
JP2024511652A (ja) 組み合わせられたループフィルタ処理
WO2016070363A1 (fr) Fusion à décalage d'interprédiction

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14902871

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14902871

Country of ref document: EP

Kind code of ref document: A1