WO2016054774A1 - A method for the co-existence of color-space transform and cross-component prediction - Google Patents
A method for the co-existence of color-space transform and cross-component prediction Download PDFInfo
- Publication number
- WO2016054774A1 WO2016054774A1 PCT/CN2014/088140 CN2014088140W WO2016054774A1 WO 2016054774 A1 WO2016054774 A1 WO 2016054774A1 CN 2014088140 W CN2014088140 W CN 2014088140W WO 2016054774 A1 WO2016054774 A1 WO 2016054774A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- current
- ccp
- cst
- utilized
- tus
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/186—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/12—Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
Definitions
- the invention relates generally to video/image coding/processing. Particularly, it is related to the co-existence of color-space transform (CST) and cross-component prediction (CCP) .
- CST color-space transform
- CCP cross-component prediction
- both CST and CCP are proposed to reduce the redundancy among different color components.
- CCP firstly locates the collocated each luma transform block’s (TB’s) prediction residuals. These luma TB residuals are the reconstructed data from entropy decoding and inverse transforming. Then the luma TB residuals are utilized to predict the two groups of chroma TB residuals. And then, all TB residuals are compensated by inter or intra prediction data. At last, all TB data are decoded and construct the decoded CU data.
- TB luma transform block
- an alpha parameter is transmitted in transform unit of the video stream and the luma TB residual multiplied by the alpha parameter and right shifted by 3 bits are utilized as the predicted residuals for the current component TB residuals.
- the decoding flow for the color-space transform is depicted in Fig. 2.
- the added module namely inverse color transform, is highlighted.
- the inverse color transform is invoked to convert the residual domain back to the original domain.
- a flag is signaled to indicate the usage of color-space transform in a CU.
- intra BC and inter modes such a flag is signaled only when there is at least one non-zero coefficient in the current CU.
- intra modes such a flag is signaled only when the chroma mode of the first PU (i. e., top-left PU within one CU) is coded with DM mode.
- the forward and the inverse color-space transforms for lossy coding use the YCoCg transform matrices, which are defined as follows:
- Fig. 1 is a diagram illustrating the decoding flowchart of CCP
- Fig. 2 is a diagram illustrating the decoding flowchart of CST
- Fig. 3 is a diagram illustrating the decoding flowchart of the suggested decoding process of co-existing CCP and CST.
- CST and CCP are never utilized together.
- CCP cannot be utilized for any TU corresponding to the current CU.
- CCP cannot be utilized for any TU larger than m ⁇ m corresponding to the current CU.
- m is equal to 4 as shown in Fig. 3.
- CCP cannot be utilized for any TU when the current CU is encoded by one subset of all the inter, intra or intra block copy modes. For example, if CST is utilized or the current inter CU, CCP is not utilized.
- CST if CCP is utilized the current CU, CST cannot be utilized if CCP is utilized for more than x percentage pixels. For example, if CCP is utilized for more 10%pixels of the current CU, CST is not utilized.
- CST cannot be utilized when the current CU is encoded by one subset of all the intra, inter or intra block copy modes. For example, if CCP is utilized for the current inter CU, CST is not utilized.
- CST cannot be utilized when the current CU is smaller than m ⁇ m. For example, if CCP is utilized current 8x8 CU (m is set 16) , CST is not utilized.
- any methods combining the some of the above methods can be included.
- an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein.
- An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
- DSP Digital Signal Processor
- the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA) .
- processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention.
- the software code or firmware codes may be developed in different programming languages and different format or style.
- the software code may also be compiled for different target platform.
- different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Discrete Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Methods are proposed to constrain the co-existence of CCP and CST.
Description
The invention relates generally to video/image coding/processing. Particularly, it is related to the co-existence of color-space transform (CST) and cross-component prediction (CCP) .
For HEVC-SCC, both CST and CCP are proposed to reduce the redundancy among different color components.
The basic decoding process the CCP in the current HEVC-SCC is illustrated in Fig. 1. CCP firstly locates the collocated each luma transform block’s (TB’s) prediction residuals. These luma TB residuals are the reconstructed data from entropy decoding and inverse transforming. Then the luma TB residuals are utilized to predict the two groups of chroma TB residuals. And then, all TB residuals are compensated by inter or intra prediction data. At last, all TB data are decoded and construct the decoded CU data.
During each procedure of utilizing 1 st component TB residuals to predict the current second or third TB residuals, an alpha parameter is transmitted in transform unit of the video stream and the luma TB residual multiplied by the alpha parameter and right shifted by 3 bits are utilized as the predicted residuals for the current component TB residuals. In the current IRP design, alpha values among-8, -4, -2, -1, 0, 1, 2, 4 and 8.
Whereas for color-space transform, or namely adaptive color transform [1] [2] , the decoding flow for the color-space transform is depicted in Fig. 2. The added module, namely inverse color transform, is highlighted. When one block is coded with color transform enabled, after the conventional inverse DCT/DST transform and CCP, the inverse color transform is invoked to convert the residual domain back to the original domain. A flag is signaled to indicate the usage of color-space transform in a CU. For intra BC and inter modes, such a flag is signaled only
when there is at least one non-zero coefficient in the current CU. For intra modes, such a flag is signaled only when the chroma mode of the first PU (i. e., top-left PU within one CU) is coded with DM mode.
Two different color-space transforms are applied depending on whether the CU is coded losslessly or in a lossy manner. The forward and the inverse color-space transforms for lossy coding use the YCoCg transform matrices, which are defined as follows:
It is obvious that there must be some performance overlap between CCP and CST, since they can both work by reducing the power of the residuals in the second and third color components. In such case, the co-existence of CCP and CST enlarges the decoding time. Therefore, this invention is proposed to constrain the co-existence of CCP and CST. .
SUMMARY
In light of the previously described problems, methods are proposed to signal the palette table correctly and efficiently.
Other aspects and features of the invention will become apparent to those with ordinary skill in the art upon review of the following descriptions of specific embodiments.
BRIEF DESCRIPTION OF DRAWINGS
The invention can be more fully understood by reading the subsequent detailed description and examples with references made to the accompanying drawings, wherein:
Fig. 1 is a diagram illustrating the decoding flowchart of CCP;
Fig. 2 is a diagram illustrating the decoding flowchart of CST;
Fig. 3 is a diagram illustrating the decoding flowchart of the suggested decoding process of co-existing CCP and CST.
The following description is of the best-contemplated mode of carrying out the invention. This description is made for the purpose of illustrating the general principles of the invention and should not be taken in a limiting sense. The scope of the invention is best determined by reference to the appended claims.
In order to reduce the complexity of doing CCP and CST together in the decoding process, we propose several simplification methods to constrain the co-exsiting of CCP and CST.
In one embodiment, CST and CCP are never utilized together.
In second embodiment, if CST is utilized for the current CU, CCP cannot be utilized for any TU corresponding to the current CU.
In another embodiment, if CST is utilized for the current CU, CCP cannot be utilized for any TU larger than m×m corresponding to the current CU. For example, m is equal to 4 as shown in Fig. 3.
In still another embodiment, if CST is utilized for the current CU, CCP cannot be utilized for any TU when the current CU is encoded by one subset of all the inter, intra or intra block copy modes. For example, if CST is utilized or the current inter CU, CCP is not utilized.
In still another embodiment, if CCP is utilized the current CU, CST cannot be utilized if CCP is utilized for more than x percentage pixels. For example, if CCP is utilized for more 10%pixels of the current CU, CST is not utilized.
In still another embodiment, if CCP is utilized for the current CU, CST cannot be utilized when the current CU is encoded by one subset of all the intra, inter or intra block copy modes. For example, if CCP is utilized for the current inter CU, CST is not utilized.
In still another embodiment, if CCP is utilized for the current CU, CST cannot be utilized when the current CU is smaller than m×m. For example, if CCP is utilized current 8x8 CU (m is set 16) , CST is not utilized.
In still other embodiments, any methods combining the some of the above methods can be included.
The methods described above can be used in a video encoder as well as in a video decoder. Embodiments of disparity vector derivation methods according to
the present invention as described above may be implemented in various hardware, software codes, or a combination of both. For example, an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein. An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein. The invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA) . These processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention. The software code or firmware codes may be developed in different programming languages and different format or style. The software code may also be compiled for different target platform. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. To the contrary, it is intended to cover various modifications and similar arrangements (as would be apparent to those skilled in the art) . Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.
References:
[1] L. Zhang, J. Chen, J. Sole, M. Karczewicz, “AhG8: In-loop color-space transform, ” JCTVC-Q0112, Apr. 2014.
[2] L. Zhang, J. Chen, J. Sole, M. Karczewicz, X. Xiu, Y. He, Y. Ye, “SCCE5 Test 3.2.1: In-loop color-space transform, ” JCTVC-R0147, Jul. 2014.
[3] R. Joshi, J. Xu, “HEVC Screen Content Coding Draft Text 1” , JCTVC-R1005, Jul. 2014.
Claims (22)
- A method guarateeing that CCP and CST cannot be co-existed in some kinds of CU or TUs is applied.
- The method as claimed in claim 1, wherein if CST is utilized for one CU, CCP cannot be applied to some kinds of TUs corresponding to this CU.
- The method as claimed in claim 1, wherein if CCP is permitted for one CU, CST cannot be applied to this CU at some kinds of conditions.
- The method as claimed in claim 2, wherein if CST is utilized for the current CU, CCP cannot be applied to the inside TUs if the current CU size is smaller than m×m. For example, m is equal to 16.
- The method as claimed in claim 2, wherein if CST is utilized for the current CU, CCP cannot be applied to the inside TUs which are smaller than m×m when the current CU is predicted by modes in one subset of all the intra, inter or intra block copy modes. When m is set larger than the largest TU size, it means no limitation on the TU size.
- The method as claimed in claim 5, wherein if CST is utilized for the current CU, CCP cannot be applied to the inside TUs.
- The method as claimed in claim 5, wherein if CST is utilized for the current CU, CCP cannot be applied to the inside TUs which are larger than m×m no matter what mode the current CU is coded by. When m is set larger than the largest TU size, it means no limitation on the TU size.
- The method as claimed in claim 5, wherein if CST is utilized for the current CU, CCP cannot be applied to the inside TUs which are larger than m×m when the current CU is coded by inter modes. When m is set larger than the largest TU size, it means no limitation on the TU size.
- The method as claimed in claim 5, wherein if CST is utilized for the current CU, CCP cannot be applied to the inside TUs which are larger than m×m when the current CU is coded by intra block copy modes. When m is set larger than the largest TU size, it means no limitation on the TU size.
- The method as claimed in claim 5, wherein if CST is utilized for the current CU, CCP cannot be applied to the inside TUs which are larger than m×m when the current CU is coded by intra modes. When m is set larger than the largest TU size, it means no limitation on the TU size.
- The method as claimed in claim 3, wherein if CCP is utilized for some TUs of the current CU, CST cannot be applied to the current CU when the current CU is predicted by modes in one subset of all the intra, inter or intra block copy modes.
- The method as claimed in claim 11, if CCP is utilized for more than x percentage of the pixels (covered by the TUs utilized CCP) of the current CU, CST cannot be applied to the current CU when the current CU is predicted by modes in one subset of all the intra, inter or intra block copy modes.
- The method as claimed in claim 11, if CCP is utilized for more than x TUs of the current CU, CST cannot be applied to the current CU when the current CU is predicted by modes in one subset of all the intra, inter or intra block copy modes.
- The method as claimed in claim 11 to claim 13, wherein the related subset indentifies all the intra modes.
- The method as claimed in claim 11 to claim 13, wherein the related subset indentifies all the inter modes.
- The method as claimed in claim 11 to claim 13, wherein the related subset indentifies all the intra block copy modes.
- The method as claimed in claim 11, wherein if CCP is utilized for any TU of the current CU, CST cannot be applied.
- The method as claimed in claim 11, wherein if CCP is allowed for the current CU, CST cannot be applied.
- The method as claimed in claim 5, wherein if CST is applied to the current CU, CCP cannot be applied.
- The method as claimed in claim 5, wherein if CST is applied to the current CU, CCP can be only applied to its inside 4×4 TUs.
- The method as claimed in claim 5, wherein if CST is not applied to the current CU, CCP can be applied to all its inside TUs.
- The method as claimed in claim 5, wherein if CST is allowed to the current CU, CCP cannot be applied.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2014/088140 WO2016054774A1 (en) | 2014-10-08 | 2014-10-08 | A method for the co-existence of color-space transform and cross-component prediction |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2014/088140 WO2016054774A1 (en) | 2014-10-08 | 2014-10-08 | A method for the co-existence of color-space transform and cross-component prediction |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016054774A1 true WO2016054774A1 (en) | 2016-04-14 |
Family
ID=55652463
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2014/088140 WO2016054774A1 (en) | 2014-10-08 | 2014-10-08 | A method for the co-existence of color-space transform and cross-component prediction |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2016054774A1 (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130343464A1 (en) * | 2012-06-22 | 2013-12-26 | Qualcomm Incorporated | Transform skip mode |
CN103503461A (en) * | 2011-05-05 | 2014-01-08 | 三菱电机株式会社 | Residual quadtree structure for transform units in non-square prediction units |
WO2014047693A1 (en) * | 2012-09-28 | 2014-04-03 | Canon Kabushiki Kaisha | Method, apparatus and system for encoding and decoding the transform units of a coding unit |
-
2014
- 2014-10-08 WO PCT/CN2014/088140 patent/WO2016054774A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103503461A (en) * | 2011-05-05 | 2014-01-08 | 三菱电机株式会社 | Residual quadtree structure for transform units in non-square prediction units |
US20130343464A1 (en) * | 2012-06-22 | 2013-12-26 | Qualcomm Incorporated | Transform skip mode |
WO2014047693A1 (en) * | 2012-09-28 | 2014-04-03 | Canon Kabushiki Kaisha | Method, apparatus and system for encoding and decoding the transform units of a coding unit |
Non-Patent Citations (1)
Title |
---|
ZHANG, LI ET AL.: "SCCE5 Test 3.2.1", IN-LOOP COLOR-SPACE TRANSFORM, VOL. ISO/IEC JTC 1/SC 29/WG 11 18TH MEETING, NO. JCT-VC OF ITU-T SG 16 WP3, 9 July 2014 (2014-07-09) * |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10531119B2 (en) | Method and apparatus for palette coding with cross block prediction | |
WO2015176685A1 (en) | Methods for palette size signaling and conditional palette escape flag signaling | |
WO2015100522A1 (en) | Methods for inter-component residual prediction | |
US10448049B2 (en) | Method for color index coding using a generalized copy previous mode | |
TWI741584B (en) | Methods and apparatuses of syntax signaling and referencing constraint in video coding system | |
WO2015090217A1 (en) | Method and apparatus for palette table prediction | |
EP3033878A1 (en) | Method of residue differential pulse-code modulation for hevc range extension | |
WO2016168051A1 (en) | Techniques for advanced chroma processing | |
JP2017513305A (en) | Method and decoder for decoding pictures | |
US9491466B2 (en) | Video coding apparatus and method | |
CN108141601B (en) | Method, apparatus and readable medium for encoding and decoding a sequence of pictures | |
US10021411B2 (en) | Techniques in backwards compatible multi-layer compression of HDR video | |
JP7395497B2 (en) | Data dependencies in encoding/decoding | |
WO2015192372A1 (en) | A simplified method for illumination compensation in multi-view and 3d video coding | |
WO2016049894A1 (en) | Scaling in color transform | |
CN114731438A (en) | Coding and decoding method and device for adaptive color transformation and video coding and decoding equipment | |
WO2016054774A1 (en) | A method for the co-existence of color-space transform and cross-component prediction | |
WO2016065538A1 (en) | Guided cross-component prediction | |
WO2016044974A1 (en) | Palette table signalling | |
JP2023528733A (en) | Method, apparatus and program for boundary processing in video coding | |
TWI839923B (en) | Method and apparatus for prediction based on cross component linear model in video coding system | |
WO2016115733A1 (en) | Improvements for inter-component residual prediction | |
WO2015100732A1 (en) | A padding method for intra block copying | |
WO2024088058A1 (en) | Method and apparatus of regression-based intra prediction in video coding system | |
WO2016070363A1 (en) | Merge with inter prediction offset |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14903758 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14903758 Country of ref document: EP Kind code of ref document: A1 |