WO2015139201A1 - Simplified illumination compensation in multi-view and 3d video coding - Google Patents

Simplified illumination compensation in multi-view and 3d video coding Download PDF

Info

Publication number
WO2015139201A1
WO2015139201A1 PCT/CN2014/073608 CN2014073608W WO2015139201A1 WO 2015139201 A1 WO2015139201 A1 WO 2015139201A1 CN 2014073608 W CN2014073608 W CN 2014073608W WO 2015139201 A1 WO2015139201 A1 WO 2015139201A1
Authority
WO
WIPO (PCT)
Prior art keywords
training
samples
training process
component
involved
Prior art date
Application number
PCT/CN2014/073608
Other languages
French (fr)
Inventor
Kai Zhang
Jicheng An
Xianguo Zhang
Han HUANG
Original Assignee
Mediatek Singapore Pte. Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mediatek Singapore Pte. Ltd. filed Critical Mediatek Singapore Pte. Ltd.
Priority to PCT/CN2014/073608 priority Critical patent/WO2015139201A1/en
Publication of WO2015139201A1 publication Critical patent/WO2015139201A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation

Definitions

  • the invention relates generally to Three-Dimensional (3D) video processing.
  • the presented invention relates to illumination compensation (IC).
  • illumination compensation (IC) is adopted to compensate the difference of illumination intensity between views.
  • the two parameters a and b are derived (or named as 'trained') with the neighboring reconstructed samples of the current block and the reference block as depicted Fig. 1.
  • a neighboring sample x, of the reference block and a neighboring sample ⁇ of the current block which possess the same relative position as depicted in Fig.2 are treated as a training pair.
  • Fig. 2, Fig. 3 and Fig. 4 demonstrate the neighboring samples when the block sizes are 8x8, 16x16 and 32x32 respectively.
  • the number of training pairs is proportional to the block size. For example, there are 8 training pairs for a 8x8 block and 64 training pairs for a 64x64 block Thus the training process will be more complex for larger blocks.
  • IC is applied to each component such as Y (Luma), U (Cb), and V (Cr) separately.
  • the training process is also executed separately.
  • the parameters a and b are training independently for each component.
  • FIG. 1 is a diagram illustrating a general IC paradigm in the current 3D- HEVC
  • FIG. 2 is a diagram illustrating training samples from neighboring samples of the reference block (labeled as 'x for left neighboring samples and ' ⁇ ,' for above neighboring samples) and from neighboring samples of the current block (labeled as ' >V for left neighboring samples and 'y for above neighboring samples);
  • Fig. 3 is a diagram illustrating training samples (black) from neighboring samples of the reference block or the current block when the block size is 16 X 16;
  • Fig. 4 is a diagram illustrating training samples (black) from neighboring samples of the reference block or the current block when the block size is 32 X 32;
  • Fig. 5 is a diagram illustrating to select one training samples (black) from each four neighboring samples of the reference block or the current block when the block size is 32 X 32;
  • the number of training pairs P is constrained not to be larger than a largest number L no matter how large the block size is.
  • the largest number L can be 4, 8, 16, or 32.
  • the selected samples in the training process are the same to the original design in HEVC if the number of training pairs required by the original design, noted as C, is equal or smaller to L. Otherwise, the samples in the training process are selected in a different way to the original design.
  • P/2 samples from the above and left neighboring samples are involved in the training process when the number of training pairs is P, no matter C is larger than L or not.
  • Fig. 5 demonstrates an example in which one sample from each four adjacent samples is selected into the training set.
  • N can be 2.
  • IC should be applied in different way for different component, such as Y, U and V.
  • IC is not applied to components U and V.
  • the training process in components U and V is not executed.
  • IC is applied to components U and V with the parameters derived by the component Y.
  • the training process in the component U is not executed.
  • IC is applied to the component U with the parameters derived by the component V.
  • the training process in the component V is not executed.
  • IC is applied to the component V with the parameters derived by the component U.
  • the number of training pairs in the training process for components U and V is smaller that the number of training pairs in the training process for the component Y for the same block size. For example, there are 16 training pairs for a 32 X 32 Y block is but there are 8 training pairs for a 32 X 32 U or V block.
  • an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein.
  • An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
  • DSP Digital Signal Processor
  • the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA).
  • processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention.
  • the software code or firmware codes may be developed in different programming languages and different format or style.
  • the software code may also be compiled for different target platform.
  • different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.

Abstract

It is proposed to further reduce the complexity of IC by simplifying the training process of IC. The number of training pairs is constrained and the training process is different for different components.

Description

SIMPLIFIED ILLUMINATION COMPENSATION IN
MULTI-VIEW AND 3D VIDEO CODING
TECHNICAL FIELD
[0001] The invention relates generally to Three-Dimensional (3D) video processing. In particular, the presented invention relates to illumination compensation (IC).
BACKGROUND
[0002] In the current 3D-HEVC [1], illumination compensation (IC) is adopted to compensate the difference of illumination intensity between views.
[0003] When IC is applied, the prediction value y is calculated as y = a*x+b, where x is a sample value in the reference block in the reference view. The two parameters a and b are derived (or named as 'trained') with the neighboring reconstructed samples of the current block and the reference block as depicted Fig. 1.
[0004] In the training process as specified in HEVC, a neighboring sample x, of the reference block and a neighboring sample^ of the current block which possess the same relative position as depicted in Fig.2 are treated as a training pair. To reduce the number of training pairs, only one of each two adjacent samples are involved in the training set. Fig. 2, Fig. 3 and Fig. 4 demonstrate the neighboring samples when the block sizes are 8x8, 16x16 and 32x32 respectively.
[0005] The number of training pairs is proportional to the block size. For example, there are 8 training pairs for a 8x8 block and 64 training pairs for a 64x64 block Thus the training process will be more complex for larger blocks.
[0006] IC is applied to each component such as Y (Luma), U (Cb), and V (Cr) separately. The training process is also executed separately. The parameters a and b are training independently for each component. SUMMARY
[0007] In light of the previously described problems, methods are proposed to simplify IC.
[0008] Other aspects and features of the invention will become apparent to those with ordinary skill in the art upon review of the following descriptions of specific embodiments.
BRIEF DESCRIPTION OF DRAWINGS
The invention can be more fully understood by reading the subsequent detailed description and examples with references made to the accompanying drawings, wherein:
[0009] Fig. 1 is a diagram illustrating a general IC paradigm in the current 3D- HEVC;
[0010] Fig. 2 is a diagram illustrating training samples from neighboring samples of the reference block (labeled as 'x for left neighboring samples and 'χ^,' for above neighboring samples) and from neighboring samples of the current block (labeled as ' >V for left neighboring samples and 'y for above neighboring samples);
[0011] Fig. 3 is a diagram illustrating training samples (black) from neighboring samples of the reference block or the current block when the block size is 16 X 16;
[0012] Fig. 4 is a diagram illustrating training samples (black) from neighboring samples of the reference block or the current block when the block size is 32 X 32;
[0013] Fig. 5 is a diagram illustrating to select one training samples (black) from each four neighboring samples of the reference block or the current block when the block size is 32 X 32;
[0014] Fig. 6 is a diagram illustrating to involve the first 8 samples as the training samples (black) from the above and left neighboring samples of the reference block or the current block when the block size is 32 X 32 and P=L=\6;
[0015] Fig. 7 is a diagram illustrating to Involve the one sample from each two in the first 8 samples as the training samples (black) from the above and left neighboring samples of the reference block or the current block when the block size is 32x32 and P=L=8.
DETAILED DESCRIPTION [0016] The following description is of the best-contemplated mode of carrying out the invention. This description is made for the purpose of illustrating the general principles of the invention and should not be taken in a limiting sense. The scope of the invention is best determined by reference to the appended claims.
[0017]It is proposed to further reduce the complexity of IC by simplifying the training process of IC.
[0018] In one embodiment, the number of training pairs P is constrained not to be larger than a largest number L no matter how large the block size is. For example, the largest number L can be 4, 8, 16, or 32.
[0019] In another embodiment, the selected samples in the training process are the same to the original design in HEVC if the number of training pairs required by the original design, noted as C, is equal or smaller to L. Otherwise, the samples in the training process are selected in a different way to the original design.
[0020] In another embodiment, P/2 samples from the above and left neighboring samples are involved in the training process when the number of training pairs is P, no matter C is larger than L or not.
[0021]In another embodiment, only one of each N adjacent samples are involved in the training set, where N > 2. Fig. 5 demonstrates an example in which one sample from each four adjacent samples is selected into the training set.
[0022] In still another embodiment, only one of each N adjacent samples are involved in the training set, where N depends on the block size of the current block. For example, if the current block size if XM, then N is decided as N=max(2, /8). In a general example, N is decided as N=max(2, 2*M/L), where L is the largest number of training pairs involved in the training process.
[0023] In still another embodiment, only the first P/2 adjacent samples in the left or above neighboring samples are involved in the training set, where P is the number of training pairs involved in the training process. Fig. 6 demonstrates an example in which the first 8 samples are involved as the training samples (black) from the above and left neighboring samples of the reference block or the current block when the block size is 32 X 32 and P=L=\6.
[0024] In still another embodiment, only one of each N adjacent samples in the first P/2 adjacent samples in the left or above neighboring samples are involved in the training set, where P is the number of training pairs involved in the training process. For example, N can be 2. Fig. 7 demonstrates an example in which the one sample from each two in the first 8 samples as the training are involved as the training samples (black) from the above and left neighboring samples of the reference block or the current block when the block size is 32 X 32 and P=L=\6.
[0025] In still another embodiment, IC should be applied in different way for different component, such as Y, U and V.
[0026] In still another embodiment, IC is not applied to components U and V.
[0027] In still another embodiment, the training process in components U and V, is not executed. IC is applied to components U and V with the parameters derived by the component Y.
[0028] In still another embodiment, the training process in the component U is not executed. IC is applied to the component U with the parameters derived by the component V.
[0029] In still another embodiment, the training process in the component V is not executed. IC is applied to the component V with the parameters derived by the component U.
[0030] In still another embodiment, the number of training pairs in the training process for components U and V is smaller that the number of training pairs in the training process for the component Y for the same block size. For example, there are 16 training pairs for a 32 X 32 Y block is but there are 8 training pairs for a 32 X 32 U or V block.
[0031] The methods described above can be used in a video encoder as well as in a video decoder. Embodiments of disparity vector derivation methods according to the present invention as described above may be implemented in various hardware, software codes, or a combination of both. For example, an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein. An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein. The invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention. The software code or firmware codes may be developed in different programming languages and different format or style. The software code may also be compiled for different target platform. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
[0032] The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. To the contrary, it is intended to cover various modifications and similar arrangements (as would be apparent to those skilled in the art). Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.

Claims

1. A method to reduce the complexity of illumination compensation (IC) by simplifying a training process of IC, comprising,
constraining a number of training pairs P not to be larger than a largest number L no matter the block size;
applying illumination compensation in different way for different components.
2. The method as claimed in claim 1, wherein selected samples in the training process are predetermined if the number of training pairs required by an original design, noted as C, is equal or smaller to L; otherwise, the samples in the training process are selected in a different way to the original design.
3. The method as claimed in claim 1 and claim 2, wherein P/2 samples from the above and left neighboring samples are involved in the training process when the number of training pairs is P, no matter C is larger than L or not.
4. The method as claimed in claim 1, wherein only one of each N adjacent samples are involved in a training set.
5. The method as claimed in claim 4, wherein where N> 2.
6. The method as claimed in claim 4, wherein N depend on a block size of a current block.
7. The method as claimed in claim4, wherein N is decided as N=max(2, MIL) if a current block size if XM, where L is the largest number of training pairs involved in the training process.
8. The method as claimed in claim 1, wherein only first P/2 adjacent samples in left or above neighboring samples are involved in the training set, where P is the number of training pairs involved in the training process.
9. The method as claimed in claim 1, wherein only one of each N adjacent samples in first P/2 adjacent samples in left or above neighboring samples are involved in the training set, where P is the number of training pairs involved in the training process, and N depend on a block size of a current block.
10. The method as claimed in claim 1, wherein the training process in components U and V is not executed, and IC is applied to the components U and V with parameters derived by component Y.
11. The method as claimed in claim 1, wherein the training process in component U is not executed, and IC is applied to the component U with parameters derived by component V.
12. The method as claimed in claim 1, wherein the training process in the component V is not executed, and IC is applied to the component V with parameters derived by component U.
13. The method as claimed in claim 1, wherein the number of training pairs in the training process for components U and V is smaller that the number of training pairs in the training process for component Y for the same block size.
14. The method as claimed in claim 1, wherein different components comprise components R, G and B or any other components.
PCT/CN2014/073608 2014-03-18 2014-03-18 Simplified illumination compensation in multi-view and 3d video coding WO2015139201A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2014/073608 WO2015139201A1 (en) 2014-03-18 2014-03-18 Simplified illumination compensation in multi-view and 3d video coding

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2014/073608 WO2015139201A1 (en) 2014-03-18 2014-03-18 Simplified illumination compensation in multi-view and 3d video coding

Publications (1)

Publication Number Publication Date
WO2015139201A1 true WO2015139201A1 (en) 2015-09-24

Family

ID=54143633

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/073608 WO2015139201A1 (en) 2014-03-18 2014-03-18 Simplified illumination compensation in multi-view and 3d video coding

Country Status (1)

Country Link
WO (1) WO2015139201A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101193302A (en) * 2006-12-01 2008-06-04 三星电子株式会社 Illumination compensation method and apparatus and video encoding and decoding method and apparatus
CN101710985A (en) * 2009-12-11 2010-05-19 哈尔滨工业大学 Image brightness compensation method for image coding
US20120213281A1 (en) * 2011-02-21 2012-08-23 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi view video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101193302A (en) * 2006-12-01 2008-06-04 三星电子株式会社 Illumination compensation method and apparatus and video encoding and decoding method and apparatus
CN101710985A (en) * 2009-12-11 2010-05-19 哈尔滨工业大学 Image brightness compensation method for image coding
US20120213281A1 (en) * 2011-02-21 2012-08-23 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi view video

Similar Documents

Publication Publication Date Title
WO2017143467A1 (en) Localized luma mode prediction inheritance for chroma coding
WO2017139937A1 (en) Advanced linear model prediction for chroma coding
US10334281B2 (en) Method of conditional binary tree block partitioning structure for video and image coding
WO2018049594A1 (en) Methods of encoder decision for quad-tree plus binary tree structure
JP6472877B2 (en) Method for 3D or multi-view video encoding including view synthesis prediction
WO2016090568A1 (en) Binary tree block partitioning structure
WO2016008157A1 (en) Methods for motion compensation using high order motion model
WO2016115981A1 (en) Method of video coding for chroma components
WO2016074147A1 (en) Separated coding tree for luma and chroma
WO2017035831A1 (en) Adaptive inter prediction
WO2013155662A1 (en) Methods and apparatuses of simplification for intra chroma lm mode
US20180249146A1 (en) Methods of Depth Based Block Partitioning
WO2017008255A1 (en) Advanced intra prediction mode signaling in video coding
WO2015062002A1 (en) Methods for sub-pu level prediction
WO2015192372A1 (en) A simplified method for illumination compensation in multi-view and 3d video coding
EP3535975B1 (en) Apparatus and method for 3d video coding
WO2015135175A1 (en) Simplified depth based block partitioning method
WO2015180166A1 (en) Improved intra prediction mode coding
EP2787484A2 (en) Efficient 2D adaptive noise thresholding for video processing
WO2015196966A1 (en) Method of segmental prediction for depth and texture data in 3d and multi-view coding systems
WO2016115736A1 (en) Additional intra prediction modes using cross-chroma-component prediction
WO2015131404A1 (en) Methods for depth map coding
WO2015139201A1 (en) Simplified illumination compensation in multi-view and 3d video coding
WO2017035833A1 (en) Neighboring-derived prediction offset (npo)
WO2015100732A1 (en) A padding method for intra block copying

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14886531

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14886531

Country of ref document: EP

Kind code of ref document: A1