CN114946187A - Joint coding and decoding of chroma residual and adaptive color transform - Google Patents

Joint coding and decoding of chroma residual and adaptive color transform Download PDF

Info

Publication number
CN114946187A
CN114946187A CN202180008570.3A CN202180008570A CN114946187A CN 114946187 A CN114946187 A CN 114946187A CN 202180008570 A CN202180008570 A CN 202180008570A CN 114946187 A CN114946187 A CN 114946187A
Authority
CN
China
Prior art keywords
video
mode
block
act
flag
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202180008570.3A
Other languages
Chinese (zh)
Inventor
朱维佳
许继征
张莉
张凯
王悦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Douyin Vision Beijing Co Ltd
ByteDance Inc
Original Assignee
Douyin Vision Beijing Co Ltd
ByteDance Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Douyin Vision Beijing Co Ltd, ByteDance Inc filed Critical Douyin Vision Beijing Co Ltd
Publication of CN114946187A publication Critical patent/CN114946187A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/12Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/129Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Discrete Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Methods, systems, and devices are described for implementing chroma residual joint coding (JCCR) and Adaptive Color Transform (ACT) modes during image/video encoding and decoding. An example method of video processing includes determining that a codec mode of a current video block of a video includes an ACT mode, and enabling a JCCR codec tool for coding the current video block, and performing a conversion between a bitstream of the video and the video based on the determination, wherein a quantization parameter offset used for coding the current video block is based on the mode used by the JCCR codec tool.

Description

Joint coding and decoding of chroma residual and adaptive color transform
CROSS-REFERENCE TO RELATED APPLICATIONS
This application claims priority and benefit from international patent application No. PCT/CN2020/070957 filed on 8/1/2020 in time, according to applicable patent laws and/or rules applied to the paris convention. The entire disclosure of the above application is incorporated by reference into a portion of the disclosure of the present application.
Technical Field
This patent document relates to picture encoding and decoding and video encoding and decoding.
Background
Digital video occupies the largest bandwidth usage on the internet and other digital communication networks. As the number of connected user devices capable of receiving and displaying video increases, the bandwidth requirements for digital video usage are expected to continue to grow.
Disclosure of Invention
This document discloses systems, methods, and devices for video encoding and decoding using joint coding (JCCR) and Adaptive Color Transform (ACT) modes of chroma residuals, in addition to using other coding tools.
In one example aspect, a video processing method is disclosed. The method includes determining that a coding mode of a current video block of the video includes an Adaptive Color Transform (ACT) mode, and enabling a chroma residual joint coding (JCCR) coding tool for coding the current video block, and performing a conversion between the video and a bitstream of the video based on the determination, wherein a quantization parameter offset for coding the current video block is based on a mode used by the JCCR coding tool.
In yet another example aspect, a video encoder apparatus is disclosed. The video encoder includes a processor configured to implement the above-described method.
In yet another example aspect, a video decoder apparatus is disclosed. The video decoder comprises a processor configured to implement the above-described method.
In yet another example aspect, a non-transitory computer-readable medium having code stored thereon is disclosed. The code embodies one of the methods described herein in the form of processor executable code.
These and other features are described throughout this document.
Drawings
Fig. 1 shows a Screen Content Codec (SCC) decoder flow for loop Adaptive Color Transform (ACT).
Fig. 2 shows a decoding process using ACT.
Fig. 3 shows an example of a block coded in palette mode.
Fig. 4 shows an example of signaling palette entries using a palette predictor.
Fig. 5 shows examples of horizontal and vertical traversal scans.
Fig. 6 illustrates an example of coding of palette indices.
Fig. 7 is a block diagram illustrating an example video processing system in which various techniques disclosed herein may be implemented.
FIG. 8 is a block diagram of an example hardware platform for video processing.
Fig. 9 is a block diagram illustrating a video codec system according to some embodiments of the present disclosure.
Fig. 10 is a block diagram illustrating an encoder in accordance with some embodiments of the present disclosure.
Fig. 11 is a block diagram illustrating a decoder according to some embodiments of the present disclosure.
Fig. 12 shows a flow diagram of an example method of video processing.
Detailed Description
Section headings are used in this document to facilitate understanding, and do not limit the techniques and embodiments disclosed in each section to apply only to that section. Furthermore, the use of the H.266 term in some descriptions is for ease of understanding only and is not intended to limit the scope of the disclosed technology. Thus, the techniques described herein are also applicable to other video codec protocols and designs.
1. Initial discussion
This patent document relates to picture/video coding and decoding techniques. In particular, it relates to adaptive color transformation in picture/video codecs. It can be applied to standards under development, such as general video codec. It may also be applicable to future video codec standards or video codecs.
2. Video codec introduction
Video codec standards have evolved largely through the development of well-known ITU-T and ISO/IEC standards. ITU-T produces H.261 and H.263, ISO/IEC produces MPEG-1 and MPEG-4Visual, and both organizations jointly produce the H.262/MPEG-2 video and the H.264/MPEG-4 Advanced Video Codec (AVC) and the H.265/HEVC standards. Since h.262, video codec standards have been based on hybrid video codec structures, in which temporal prediction plus transform coding is used. In order to explore future video codec technologies other than HEVC, VCEG and MPEG united in 2015 to form a joint video exploration team (jfet). Since then, jvt has adopted many new approaches and applied them to reference software named Joint Exploration Model (JEM). In month 4 of 2018, a joint video experts group (jfet) between VCEG (Q6/16) and ISO/IEC JTC1 SC29/WG11(MPEG) was established to address the VVC standard that targets 50% bit rate reduction compared to HEVC.
The latest version of the VVC draft, universal video codec (draft 7), can be found at the following website: http:// phenix. it-supplaris. eu/JVET/doc _ end _ user/documents/16_ Geneva/wg 11/JFET-P2001-v 14.zip
The latest reference software VTM for VVC can be found at the following website: https:// vcgit. hhi. fraunhofer. de/jvet/VVCSOFTWOW _ VTM/tags/VTM-7.0
Adaptive Color Transform (ACT) in HEVC-SCC
On the 18 th JCT-VC conference (30/6 to 9/7/2014, japanese sapoma), Adaptive Color Transform (ACT) was adopted into HEVC Screen Content Codec (SCC) test model 2. ACT performs a loop color-space conversion in the prediction residual domain using a color transform matrix based on YCoCg and YCoCg-R color-spaces. ACT is adaptively turned on or off at the CU level using a flag CU residual ACT flag. ACT may be combined with Cross Component Prediction (CCP), which is another inter component decorrelation method already supported in HEVC. When both are enabled, ACT is performed after CCP at decoder, as shown in fig. 1.
Color space conversion in ACT
The color-space conversion in ACT is based on the YCoCg-R transform. Both lossy and lossless codecs (cu _ transquant _ bypass _ flag ═ 0 or 1) use the same inverse transform, but in the case of lossy codecs, an additional 1-bit left shift is applied to the Co and Cg components. Specifically, the following color-space transforms are used for forward and backward transforms for lossy and lossless codecs:
forward transform of lossy codec (non-canonical):
Figure BDA0003734881610000031
forward transform for lossless codec (non-canonical):
Co=R-B
t=B+(Co>>1)
Cg=(G-t)
Y=t+(Cg>>1)
backward transformation (specification):
Figure BDA0003734881610000041
t=Y-(Cg>>1)
G=Cg+t
B=t-(Co>>1)
R=Co+b
the forward color transform is not normalized, where its norm is approximately equal to Y and Cg
Figure BDA0003734881610000042
And is approximately equal to Co
Figure BDA0003734881610000043
To compensate for the non-normalized nature of the forward transform, delta QPs of (-5, -3, -5) are applied to (Y, Co, Cg), respectively. In other words, for a given "normal" QP for a CU, if ACT is turned on, the quantization parameters are set equal to (QP-5, QP-3, QP-5) for (Y, Co, Cg), respectively. The adjusted quantization parameter only affects the quantization and inverse quantization of the residual in the CU. For deblocking, the "normal" QP value is still used. A clip of 0 is applied to the adjusted QP values to ensure that they will not become negative. Note that this QP adjustment is applicable only to lossy codec, since quantization is not performed in lossless codec (cu _ transquant _ bypass _ flag ═ 1). In SCM 4, PPS/slice level signaling with additional QP offset values is introduced. These QP offset values may be used for CUs instead of (-5, -3, -5) when applying the adaptive color transform.
When the input bit depths of the color components are different, an appropriate left shift is applied during ACT to align the sample bit depth with the maximum bit depth, and an appropriate right shift is applied after ACT to restore the original sample bit depth.
2.2. ACT in VVC
Fig. 2 shows a decoding flowchart of VVC to which ACT is applied. As shown in fig. 2, the color space conversion is performed in the residual domain. In particular, an additional decoding module, i.e. inverse ACT, is introduced after the inverse transform to convert the residual from the YCgCo domain back to the original domain.
In VVC, one CU leaf node is also used as a unit of transform processing unless the maximum transform size is smaller than the width or height of one Codec Unit (CU). Thus, in the proposed implementation, the ACT flag is signaled for one CU to select the color space to codec its residual. In addition, after HEVC ACT design, for inter and IBC CUs, ACT is enabled only when there is at least one non-zero coefficient in the CU. For an intra CU, ACT is enabled only when the chroma component selects the same intra prediction mode as the luma component, i.e., DM mode.
The core transform for color space conversion is consistent with the core transform for HEVC. In addition, as in the ACT design in HEVC, QP adjustment of (-5, -5, -3) is applied to the transform residual in order to compensate for the dynamic range change of the residual signal before and after color transform.
On the other hand, forward and inverse color transforms require access to the residuals of all three components. Correspondingly, in the proposed implementation, ACT is disabled in two cases, where the residuals of the three components are not all available.
1. Separating tree partitions: when the separation tree is applied, the luminance samples and the chrominance samples in one CTU are partitioned according to different structures. This results in a CU in the luma tree containing only the luma component, while a CU in the chroma tree containing only two chroma components.
2. Intra sub-partition prediction (ISP): ISP subpartition applies only to luminance, while chrominance signals are coded without partitioning. In current ISP designs, the other sub-partitions, except the last ISP sub-partition, contain only a luminance component.
The text of the codec unit in the VVC draft is as follows.
Figure BDA0003734881610000051
Figure BDA0003734881610000061
Figure BDA0003734881610000071
Figure BDA0003734881610000081
Figure BDA0003734881610000091
Figure BDA0003734881610000101
Figure BDA0003734881610000111
Figure BDA0003734881610000121
cu _ act _ enabled _ flag equal to 1 specifies that the residual of the current codec unit is coded in YCgCo color space. cu _ act _ enabled _ flag equal to 0 specifies that the residual of the current codec unit is codec in the original color space. When cu _ act _ enabled _ flag is not present, it is inferred to be equal to 0.
Transition skip mode in VVC
As in HEVC, the residual of a block may be coded using a transform skip mode that completely skips the transform process of the block. Furthermore, for transform skip blocks, the minimum allowed Quantization Parameter (QP) signaled in the SPS is used, which is set equal to 6 × (internalBitDepth inputBitDepth) +4 in VTM 7.0.
2.4. Block-Based Delta Pulse Code Modulation (BDPCM)
In JVT-M0413, a block-Based Delta Pulse Codec Modulation (BDPCM) is proposed to efficiently codec the screen content, which is then adopted into VVC.
The prediction directions used in the BDPCM may be a vertical prediction mode and a horizontal prediction mode. Intra prediction is performed on the entire block by sample copy in a similar prediction direction (horizontal or vertical prediction) to intra prediction. The residual is quantized and the delta between the quantized residual and its predicted (horizontal or vertical) quantization value is coded. This can be described in the following way: for a block of size M (rows) x N (columns), after performing intra prediction horizontally (copying left neighbor pixel values across the prediction block row by row) or vertically (copying the top neighbor row to each row in the prediction block) using unfiltered samples from the top or left block boundary samples, let r be the same as the original samples, and then let r be the same as the original samples, and the samples, of the block, and the block, and the block i,j I is more than or equal to 0 and less than or equal to M-1, and j is more than or equal to 0 and less than or equal to N-1 to form a prediction residual error. Let Q (r) i,j ) I is more than or equal to 0 and less than or equal to M-1, and j is more than or equal to 0 and less than or equal to N-1 represents residual error r i,j Wherein the residual is the difference between the original block and the predicted block value. Then, the block DPCM is applied to the quantized residual sampling points to obtain the sampling points with elements
Figure BDA0003734881610000135
Modified M × N array of
Figure BDA0003734881610000136
When signaling the vertical BDPCM:
Figure BDA0003734881610000131
for horizontal prediction, a similar rule applies, and residual quantized samples are obtained by:
Figure BDA0003734881610000132
residual quantized sampling points
Figure BDA0003734881610000137
Is sent to the decoder.
On the decoder side, the above calculations are reversed to produce Q (r) i,j ),0≤i≤M-1,0≤j≤N-1。
For the case of a vertical prediction,
Figure BDA0003734881610000133
in the case of the horizontal situation,
Figure BDA0003734881610000134
the inverse quantized residual Q-1(Q (r) i,j ) ) to the intra block prediction value to produce a reconstructed sample value.
The main advantage of this scheme is that the inverse BDPCM can be done dynamically during coefficient parsing, only the prediction values need to be added as coefficients, or it can be performed after parsing.
In VTM7.0, the BDPCM may also be applied to a chroma block, and the chroma BDPCM has a separate flag and BDPCM direction different from the luma BDPCM mode.
2.5. Scaling process for transform coefficients
The text in JFET-P2001-vE related to the scaling process of the transform coefficients is given as follows.
The inputs to this process are:
specifying a luminance position (xTbY, yTbY) of the current luminance transform block relative to an upper-left luminance sample of the current picture,
a variable nTbW specifying the transform block width,
a variable nTbH specifying the transform block height,
a variable predMode specifying the prediction mode of the codec unit,
a variable cIdx specifying the color component of the current block.
The output of this process is an (nTbW) × (nTbH) array d of scaled transform coefficients with the element d [ x ] [ y ].
The quantization parameter qP is derived as follows:
if cIdx is equal to 0, then the following applies:
qP=QP'Y (1129)
otherwise, if tucresimode [ xTbY ] [ yTbY ] equals 2, then the following applies:
qP=QP'CbCr (1130)
otherwise, if cIdx is equal to 1, then the following applies:
qP=QP'Cb (1131)
otherwise (cIdx equals 2), the following applies:
qP=QP'Cr (1132)
the quantization parameter qP is modified and the variables rectnottsflag and bdShift are derived as follows:
if transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] is equal to 0, the following applies:
qP=qP-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1133)
rectNonTsFlag=(((Log2(nTbW)+Log2(nTbH))&1)==1)?1:0 (1134)
bdShift=BitDepth+rectNonTsFlag+ (1135)
((Log2(nTbW)+Log2(nTbH))/2)-5+pic_dep_quant_enabled_flag
otherwise (transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equals 1), the following applies:
qP=Max(QpPrimeTsMin,qP)-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1136)
rectNonTsFlag=0 (1137)
bdShift=10 (1138)
the variable bdOffset is derived as follows:
bdOffset=(1<<bdShift)>>1 (1139)
the list levelScale [ ] [ ] is specified as levelScale [ j ] [ k ] { {40,45,51,57,64,72}, {57,64,72,80,90,102} }, where j ═ 0..1, k ═ 0..5.
The (nTbW) × (nTbH) array dz is set equal to the (nTbW) × (nTbH) array TransCoeffLevel [ xTbY ] [ yTbY ] [ cIdx ].
For the derivation of the scaling transform coefficients d [ x ] [ y ] (where x is 0.. nTbW-1, y is 0.. nTbH-1), the following applies:
the intermediate scaling factor m [ x ] [ y ] is derived as follows:
m [ x ] [ y ] is set equal to 16 if one or more of the following conditions is true:
sps _ scaling _ list _ enabled _ flag is equal to 0.
pic _ scaling _ list _ present _ flag is equal to 0.
transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equals 1.
scaling _ matrix _ for _ lfnst _ disabled _ flag is equal to 1, and lfnst _ idx [ xTbY ] [ yTbY ] is not equal to 0.
Otherwise, the following applies:
the variable id is derived from predMode, cIdx, nTbW and nTbH as specified in table 36, and the variable log2MatrixSize is derived as follows:
log2MatrixSize=(id<2)?1:(id<8)?2:3 (1140)
the scale factor m [ x ] [ y ] is derived as follows:
m[x][y]=ScalingMatrixRec[id][i][j]
wherein i ═ (x < < Log2MatrixSize) > > Log2(nTbW),
j=(y<<log2MatrixSize)>>Log2(nTbH) (1141)
if id is greater than 13 and x and y are both equal to 0, then m [0] [0] is further modified as follows:
m[0][0]=ScalingMatrixDCRec[id-14] (1142)
note that the quantization matrix element m [ x ] [ y ] may be zeroed when any of the following conditions is true
x is greater than 32
y is greater than 32
Tu decoded is not coded by default transform mode (i.e. transform type is not equal to 0) and x is larger than 16
Tu decoded is not coded by default transform mode (i.e. transform type is not equal to 0) and y is larger than 16
The scale factor ls [ x ] [ y ] is derived as follows:
if pic _ dep _ quant _ enabled _ flag is equal to 1 and transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] is equal to 0, then the following applies:
ls[x][y]=(m[x][y]*levelScale[rectNonTsFlag][(qP+1)%6])<<((qP+1)/6) (1143)
otherwise (pic _ dep _ quant _ enabled _ flag equal to 0 or transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equal to 1), the following applies:
ls[x][y]=(m[x][y]*levelScale[rectNonTsFlag][qP%6])<<(qP/6) (1144)
when Bdpcmflag [ xTbY ] [ yYbY ] [ cIdx ] equals 1, dz [ x ] [ y ] is modified as follows:
if BdpcmDir [ xTbY ] [ yYbY ] [ cIdx ] is equal to 0 and x is greater than 0, then the following applies:
dz[x][y]=Clip3(CoeffMin,CoeffMax,dz[x-1][y]+dz[x][y]) (1145)
otherwise, if BdpcmDir [ xTbY ] [ yTbY ] [ cIdx ] is equal to 1 and y is greater than 0, then the following applies:
dz[x][y]=Clip3(CoeffMin,CoeffMax,dz[x][y-1]+dz[x][y]) (1146)
the value dnc [ x ] [ y ] is derived as follows:
dnc[x][y]=(dz[x][y]*ls[x][y]+bdOffset)>>bdShift (1147)
the scaled transform coefficients d [ x ] [ y ] are derived as follows:
d[x][y]=Clip3(CoeffMin,CoeffMax,dnc[x][y]) (1148)
table 36-Prodcode, cIdx, nTbW, and nTbH
Figure BDA0003734881610000161
2.6. Palette mode
2.6.1. Concept of palette mode
The basic idea behind the palette mode is that pixels in a CU are represented by a small set of representative color values. This set is called a palette. And may also indicate samples outside the palette by signaling escape symbols after the (possibly quantized) component values. Such a pixel is called an escape pixel. The palette mode is shown in fig. 3. As shown in fig. 3, for each pixel having three color components (luminance and two chrominance components), an index of the palette is established, and the block may be reconstructed based on the established values in the palette.
2.6.2. Palette entry codec
For the purpose of encoding and decoding the palette entries, the palette predictor is retained. The maximum size of the palette and the palette predictor are signaled in the SPS. In the HEVC-SCC, a palette _ predictor _ initializer _ present _ flag is introduced in the PPS. When this flag is 1, an entry for initializing the palette predictor is signaled in the bitstream. The palette predictor is initialized at the beginning of each CTU row, each stripe, and each slice. Depending on the value of palette _ predictor _ initializer _ present _ flag, the palette predictor is reset to 0 or initialized using the palette predictor initializer entry signaled in the PPS. In HEVC-SCC, a size 0 palette predictor initializer is enabled to allow explicit disabling of palette predictor initialization at PPS level.
For each entry in the palette predictor, a reuse flag is signaled to indicate whether it is part of the current palette. This is shown in fig. 4. The reuse flag is sent using run length coding of zeros. After this, the number of new palette entries is signaled using a 0 th order Exponential Golomb (EG) code (i.e., EG-0). Finally, the new palette entry component value is signaled.
2.6.3. Palette index coding and decoding
The palette indices are encoded and decoded using the horizontal traversal scan and the vertical traversal scan shown in fig. 5. The scanning order is explicitly signaled in the bitstream using a palette _ transpose _ flag. For the rest of the sub-section, the scan is assumed to be horizontal.
Palette indices are coded using two palette example modes: "COPY _ LEFT" and "COPY _ ABOVE". In the "COPY _ LEFT" mode, palette indices are assigned to decoding indices. In "COPY _ ABOVE" mode, the palette indices of the samples in the upper row are copied. For both the "COPY LEFT" and "COPY ABOVE" modes, a run value is signaled that specifies the number of subsequent samples that are also coded using the same mode.
In palette mode, the index value of an escape symbol is the number of palette entries. And, when an escape symbol is part of a run in "COPY LEFT" or "COPY ABOVE" mode, an escape component value is signaled for each escape symbol. The encoding and decoding of the palette index are shown in fig. 6.
This syntax order is done as follows. First, the number of index values of a CU is signaled. The actual index value of the entire CU is then signaled using truncated binary coding. The index number and the index value are both coded and decoded in bypass mode. This groups together index-dependent bypass bits. The palette sampling pattern (if necessary) and run are then signaled in an interleaved manner. Finally, the component escape values corresponding to the escape symbols of the entire CU are grouped together and coded in bypass mode. The binarization of the escape symbol is the EG of the third-order codec, namely EG-3.
The additional syntax element last _ run _ type _ flag is signaled after signaling the index value. This syntax element, in combination with the number of indices, eliminates the need to signal the run value corresponding to the last run in the block.
In HEVC-SCC, the palette mode is also enabled for 4:2:2, 4:2:0 and monochrome chroma formats. The signaling of palette entries and palette indices is almost the same for all chroma formats. In the case of the non-monochrome format, each palette entry consists of 3 components. For the monochrome format, each palette entry consists of a single component. For the sub-sampled chroma direction, chroma samples are associated with a luma sample index that is divisible by 2. After reconstructing the palette index of a CU, only the first component of the palette entry is used if the sample point has only a single component associated with it. The only difference in signaling is the escape component value. The number of signaled escape component values may be different for each escape symbol depending on the number of components associated with the symbol.
2.6.4. Palette in a double tree
In VVC, a dual tree coding structure is used to code an intra slice, so the luminance component and the two chrominance components may have different palettes and palette indices. Furthermore, the two chroma components share the same palette and palette index.
2.6.5. Line-based CG palette mode
A CG palette mode based on line is adopted for VVC. In this method, each CU of the palette mode is divided into a plurality of segments of m samples (m ═ 16 in this test) based on the traverse scan mode. The coding order of the palette run codec in each slice is as follows: for each pixel, a signal signaling 1 context-coded binary bit run _ COPY _ flag 0, indicating whether the pixel has the same pattern as the previous pixel, i.e. if the previously scanned pixel and the current pixel are both run type COPY ABOVE, or if the previously scanned pixel and the current pixel are both run type INDEX and the same INDEX value. Otherwise, run _ copy _ flag is signaled as 1. If the mode of the pixel is different from that of the previous pixel, a context coded binary bit COPY ABOVE palette indices flag is signaled to indicate the run type of the pixel, i.e., INDEX or COPY ABOVE. Like the palette mode in VTM6.0, if the samples are in the first row (horizontal traversal scan) or the first column (vertical traversal scan), the decoder does not have to parse the run type, since the INDEX mode is used by default. Furthermore, if the previously parsed run type is COPY ABOVE, the decoder does not have to parse the run type. After palette run-length coding of pixels in a segment, the INDEX values (for INDEX mode) and quantized escape colors are bypass coded and grouped separately from the coding/parsing of the context coded binary bits to improve throughput within each line CG. Since the index values are now coded/parsed after the run-length coding, rather than processed before the palette run-length coding as in VTM, the encoder does not have to signal the number of index values num _ palette _ indices _ minus1 and the last run type copy _ above _ indices _ for _ final _ run _ flag.
3. Technical problem solved by embodiments and solutions described herein
In the current design, ACT and luma BDPCM modes may be enabled for a block. However, for blocks coded using the ACT mode, the chroma BDPCM mode is always disabled. Therefore, the prediction signal may be derived differently for the luma block and the chroma block in the same codec unit, which may be inefficient.
When ACT is enabled, the Quantization Parameter (QP) for a block may become negative.
The current design of ACT does not support lossless codec.
The signaling of the use of ACT is not block size dependent.
The maximum palette size and the maximum predictor size are fixed numbers, which may limit the flexibility of the palette mode.
The escape sampling points adopt third-order Exponential Golomb (EG) as a binarization method, but the binarization of the escape sampling points does not depend on a Quantization Parameter (QP).
4. Technical solution
The technical solutions described below should be considered as examples to explain the general concept. These technical solutions should not be interpreted narrowly. Furthermore, these technical solutions may be combined in any manner.
In the following description, the term "block" may represent a video region, such as a Coding Unit (CU), a Prediction Unit (PU), a Transform Unit (TU), which may contain samples in three color components. The term "BDPCM" is not limited to the design in VVC, but it may represent a technique of coding and decoding a residual using different prediction signal generation methods.
In the following description, a video block coded using a chroma residual joint coding (JCCR) mode includes signaling only one chroma residual block (e.g., a Cb residual block), while another chroma residual block (e.g., a Cr residual block) is derived based on the signaled chroma residual block and one or more flags indicating a particular JCCR mode (e.g., at a transform unit level). As described above, the JCCR mode improves coding efficiency using correlation between Cb and Cr residuals.
Interaction between ACT and BDPCM (items 1-4)
1. Whether the chroma BDPCM mode is enabled may depend on the use of ACT and/or luma BDPCM modes.
a. In one example, when ACT is enabled on a block, an indication of the use of chroma BDPCM mode (e.g., intra bdpcmp chroma flag) may be inferred as an indication of the use of luma BDPCM mode (e.g., intra bdpcmp luma flag).
i. In one example, the inferred value of chroma BDPCM mode is defined as (ACT enabled and luma BDPCM mode.
1. In one example, when intra _ bdplcm _ luma _ flag is false, intra _ bdplcm _ chroma _ flag may be set equal to false.
a. Alternatively, when intra _ bdpcmm _ luma _ flag is true, intra _ bdpcmm _ chroma _ flag may be set equal to true.
Alternatively, in one example, if the indication of the use of luma BDPCM mode and ACT of a block is true, the indication of the use of chroma BDPCM mode may be inferred as true.
b. Alternatively, it may be conditionally checked whether the use of ACT of the block is signaled, e.g. the same BDPCM prediction direction is used for luminance and chrominance samples in the block.
i. Alternatively, further, an indication of the use of ACT is signaled after using the BDPCM mode.
2. When ACT is enabled on a block, an indication of the prediction direction of the chroma BDPCM mode (e.g., intra _ bdpcmm _ chroma _ dir _ flag) may be inferred as an indication of the prediction direction of the use of the luma BDPCM mode (e.g., intra _ bdpcmm _ luma _ dir _ flag).
a. In one example, the inferred value of intra bdpcmm chroma dir flag is defined as (ACT enabled.
i. In one example, if the indication of the prediction direction of the luma BDPCM mode is horizontal, the indication of the prediction direction of the chroma BDPCM mode may be inferred to be horizontal.
Alternatively, in one example, if the indication of the prediction direction of the luma BDPCM mode is vertical, the indication of the prediction direction of the chroma BDPCM mode may be inferred to be vertical.
ACT and BDPCM modes may be applied mutually exclusive.
a. In one example, when ACT mode is enabled on a block, BDPCM mode may be disabled on the block.
i. Alternatively, further, the indication of the use of the ACT mode may be signaled after signaling the indication of the use of the BDPCM mode.
Alternatively, furthermore, the indication of the use of the BDPCM mode may not be signaled and is inferred as false (0).
b. In one example, when the BDPCM mode is enabled on a block, the ACT mode may be disabled on the block.
i. Alternatively, further, the indication of the use of the ACT mode may be signaled after signaling the indication of the use of the BDPCM mode.
Alternatively, also, the indication of the use of ACT mode may not be signaled and is inferred to be false (0).
c. In one example, the BDPCM mode in the above example may represent a luma BDPCM mode and/or a chroma BDPCM mode.
4. The inverse ACT may be applied before the inverse BDPCM at the decoder.
a. In one example, ACT may be applied even when the luminance BDPCM and the chrominance BDPCM have different prediction modes.
b. Alternatively, at the encoder, the forward ACT may be applied after BDPCM.
QP setting when ACT is enabled (item 5)
5. It is proposed to clip QP when ACT is enabled.
a. In one example, the clipping function may be defined as (l, h, x), where l is the lowest possible value of input x and h is the highest possible value of input x.
i. In one example, l may be set equal to 0.
in one example, h may be set equal to 63.
b. In one example, the QP may be QP given in section 2.5.
c. In one example, clipping may be performed after QP adjustment for ACT mode.
d. In one example, when transform skip is applied, l may be set equal to the minimum allowed QP for transform skip mode.
Related Palette mode (items 6-7)
6. The value of the maximum allowed palette size and/or the maximum allowed predictor size may depend on the codec characteristics. Suppose S 1 Is a maximum palette size (or palette predictor size) associated with the first coding characteristic; and S 2 Is the maximum palette size (or palette predictor size) associated with the second codec characteristic.
a. In one example, the codec characteristic may be a color component.
i. In one example, the maximum allowed palette size and/or the maximum allowed predicted value size for different color components may have different values.
in one example, the value of the maximum allowed palette size and/or maximum allowed predictor size for a first color component (e.g., Y in YCbCr, G in RGB) may be different from the values of the maximum allowed palette size and/or maximum allowed predictor size for the other two color components that do not include the first color component (e.g., Cb and Cr in YCbCr, B and R in RGB).
b. In one example, the codec characteristic may be a Quantization Parameter (QP).
i. In one example, if QP1 is greater than QP 2 Then QP 1 S of 1 And/or S 2 Should be less than QP 2 S of 1 And/or S 2
in one example, the QP may be a slice level QP or a block level QP.
c. In one example, S 2 May be greater than or equal to S 1
d. For the first and second codec characteristics, the indication of the maximum palette size/palette predictor size may be signaled separately or inferred from one to the other.
i. In one example, S may be signaled 1 And may be based on S 1 Derivation of S 2
1. In one example, S 2 Can be inferred as S 1 n。
2. In one example, S 2 Can be inferred as S 1 >>n。
3. In one example, S 2 Can be inferred as floor (S) 1 (n), wherein floor (x) represents the largest integer no greater than x.
e. In one example, S may be signaled at a high level (e.g., SPS/PPS/PH/slice header) 1 And/or S 2 And adjusting S at a low level (e.g., CU/block) 1 And/or S 2
i. How to adjust S 1 And/or S 2 Possibly depending on the codec information.
1. How to adjust S 1 And/or S 2 Possibly depending on the current QP.
a. In one example, if the current QP increases, S should be decreased 1 And/or S 2
2. How to adjust S 1 And/or S 2 May depend onThe block dimension.
a. In one example, if the current block size increases, S 1 And/or S 2 Should be increased.
f.S 1 And/or S 2 May depend on whether the LMCS is used.
7. The parameters associated with the binarization method of the escape samples/pixels may depend on the codec information, such as the Quantization Parameter (QP).
a. In one example, an EG binarization method may be used, and the order of EG binarization, denoted by k, may depend on the codec information.
i. In one example, k may be decreased as the current QP increases.
Signaling of ACT mode (items 8-10)
8. The indication of the maximum and/or minimum ACT size allowed may be signaled at the sequence/video/slice/sub-picture/tile/other video processing unit level or derived based on codec information.
a. In one example, they may be signaled in the SPS/PPS/picture header/slice header.
b. In one example, they may be signaled conditionally, e.g., according to enabled ACT.
c. In one example, N levels of maximum and/or minimum ACT size allowed may be signaled/defined, e.g., N-2.
i. In one example, the allowed maximum and/or minimum ACT size may be set to K0 or K1 (e.g., K0-64, K1-32).
Alternatively, further, an indication of the level may be signaled, e.g. when N ═ 2, a flag may be signaled.
d. In one example, the difference between the allowed maximum and/or minimum ACT size and the allowed maximum and/or minimum transform (or transform skip) size (e.g., for the luma component) may be signaled.
e. In one example, the maximum and/or minimum ACT sizes allowed (e.g., for the luma component) may be derived from the maximum and/or minimum (or transform skip) sizes allowed.
f. Alternatively, whether and/or how the indication of ACT usage and other side information related to ACT is signaled may depend, inter alia, on the maximum and/or minimum allowed.
9. When a block is larger than the allowed maximum ACT size (or allowed maximum transform size), the block may be automatically divided into a number of sub-blocks, where all sub-blocks share the same prediction mode (e.g., all sub-blocks are intra-coded), and ACT may be enabled at the sub-block level instead of the block level.
10. The indication of the use of ACT mode may be conditionally signaled based on block dimensions (e.g., block width and/or block height, block width times height, ratio between block width and block height, maximum/minimum of block width and block height), and/or maximum allowed ACT size.
a. In one example, an indication of the use of ACT mode may be signaled when certain conditions are met (e.g., according to the block dimension).
i. In one example, the condition is whether the current block width is less than or equal to m and/or the current block height is less than or equal to n.
in one example, the condition is whether the current block width multiplied by height is less than or not greater than m.
in one example, the condition is whether the current block width multiplied by height is greater than or not less than m.
b. Alternatively, in one example, the indication of the use of ACT mode may not be signaled when certain conditions are not met (e.g., in terms of block dimensions).
i. In one example, the condition is whether the current block width is greater than m and/or the current block height is greater than n.
in one example, the condition is whether the current block width multiplied by height is less than or not greater than m.
in one example, the condition is whether the current block width multiplied by height is greater than or not less than m.
Alternatively, further, the indication of the use of ACT mode may be inferred to be 0.
c. In the above examples, the variables m, n may be predefined (e.g., 4, 64, 128), or signaled, or derived on the fly.
i. In one example, m and/or n may be derived based on decoded messages in SPS/PPS/APS/CTU rows/CTU groups/CUs/blocks.
1. In one example, m and/or n may be set equal to a maximum allowed transform size (e.g., MaxTbSizeY).
Signaling of constraint flags in the general constraint information syntax (items 11-16)
The following constraint flags may be signaled in video units other than SPS. For example, they may be signaled in the common constraint information syntax specified in JFET-P2001-vE.
11. It is proposed to use a constraint flag to specify whether the SPS ACT enabled flag (e.g., SPS ACT enabled flag) should be equal to 0.
a. In one example, the flag may be denoted as no _ act _ constraint _ flag
i. When this flag is equal to 1, the SPS ACT enabled flag (e.g., SPS _ ACT _ enabled _ flag) should be equal to 0.
When this flag is equal to 0, it does not impose such a constraint.
12. It is proposed to use a constraint flag to specify whether the SPS BDPCM enabled flag (e.g., SPS _ bdplcm _ enabled _ flag) should be equal to 0.
a. In one example, the flag may be denoted as no _ bdplcm _ constraint _ flag.
i. When this flag is equal to 1, the SPS BDPCM enabled flag (e.g., SPS _ bdplcm enabled flag) should be equal to 0.
When this flag is equal to 0, it does not impose such a constraint.
13. It is proposed to use a constraint flag to specify whether the SPS chroma BDPCM enabled flag (e.g., SPS _ bdplcm _ chroma _ enabled _ flag) should be equal to 0.
a. In one example, the flag may be denoted as no _ bdplcm _ chroma _ constraint _ flag.
i. When this flag is equal to 1, the SPS chroma BDPCM enabled flag (e.g., SPS _ bdplcm _ chroma _ enabled _ flag) should be equal to 0.
When this flag is equal to 0, it does not impose such a constraint.
14. It is proposed to use a constraint flag to specify whether the SPS palette enable flag (e.g., SPS _ palette _ enabled _ flag) should be equal to 0.
a. In one example, the flag may be denoted as no _ palette _ constraint _ flag.
i. When this flag is equal to 1, the SPS palette enabled flag (e.g., SPS _ palette _ enabled _ flag) should be equal to 0.
When this flag is equal to 0, it does not impose such a constraint.
15. It is proposed to use a constraint flag to specify whether the SPS RPR enabled flag (e.g., ref _ pic _ resetting _ enabled _ flag) should be equal to 0.
a. In one example, this flag may be denoted as no _ ref _ pic _ sampling _ constraint _ flag.
i. When this flag is equal to 1, the SPS RPR enabled flag (e.g., ref _ pic _ resetting _ enabled _ flag) should be equal to 0.
When this flag is equal to 0, it does not impose such a constraint.
16. In the above examples (bullets 11-15), such constraint flags may be conditionally signaled, e.g., according to a chroma format (e.g., chroma _ format _ idc) and/or a separate plane codec or chroma arraytype.
ACT QP offset (items 17-19)
17. It is proposed that when applying ACT to a block, the ACT offset may be applied after applying other chroma offsets, e.g. PPS and/or chroma offsets in the Picture Header (PH) and/or Slice Header (SH).
18. It is proposed to set PPS and/or PH offset for JCbCr mode 2, except-5, when applying YCgCo color transform on the block.
a. In one example, the offset may be a number other than-5.
b. In one example, the offset may be indicated in the PPS (e.g., as PPS _ act _ cbcr _ qp _ offset _ plus6), and the offset may be set to PPS _ act _ cbcr _ qp _ offset _ plus 6-6.
c. In one example, the offset may be indicated in the PPS (e.g., as PPS _ act _ cbcr _ qp _ offset _ plus7), and the offset may be set to PPS _ act _ cbcr _ qp _ offset _ plus 7-7.
19. It is proposed to set PPS and/or PH offset for JCbCr mode 2, except 1, when YCgCo-R is applied on the block.
a. In one example, the offset may be a number other than-1.
b. In one example, the offset may be indicated in the PPS (e.g., as PPS _ act _ cbcr _ qp _ offset) and the offset may be set to PPS _ act _ cbcr _ qp _ offset.
c. In one example, the offset may be indicated in the PPS (e.g., as PPS _ act _ cbcr _ qp _ offset _ plus1), and the offset may be set to PPS _ act _ cbcr _ qp _ offset _ plus 1-1.
20. It is proposed that when using JCCR, QP offset (denoted as ACT _ QP _ offset) for ACT with YCgCo transform may depend on JCCR mode.
a. In one example, act _ qp _ offset may be-5 when JCCR mode is 1.
i. Alternatively, in one example, when the JCCR mode is 1, act _ qp _ offset may be-6.
b. In one example, when JCCR mode is 2, act _ qp _ offset may be-7.
i. Alternatively, in one example, act _ qp _ offset may be (-7-pps _ join _ cbcr _ qp _ offset-slice _ join _ cbcr _ qp _ offset) when JCCR mode is 2.
Alternatively, in one example, act _ qp _ offset may be (-7+ pps _ join _ cbcr _ qp _ offset + slice _ join _ cbcr _ qp _ offset) when JCCR mode is 2.
c. In one example, when the JCCR mode is 3, the ACT offset may be-4.
i. Alternatively, in one example, when the JCCR mode is 3, the ACT offset may be-5.
21. It is proposed that when using JCCR, the QP offset (denoted as ACT _ QP _ offset) for ACT with YCgCo-R transform may depend on JCCR mode.
a. In one example, act _ qp _ offset may be 1 when JCCR mode is 1.
i. Alternatively, in one example, act _ qp _ offset may be 0 when JCCR mode is 1.
b. In one example, act _ qp _ offset may be-1 when JCCR mode is 2.
i. Alternatively, in one example, act _ qp _ offset may be (-1-pps _ join _ cbcr _ qp _ offset-slice _ join _ cbcr _ qp _ offset) when JCCR mode is 2.
Alternatively, in one example, act _ qp _ offset may be (-1+ pps _ join _ cbcr _ qp _ offset + slice _ join _ cbcr _ qp _ offset) when JCCR mode is 2.
c. In one example, when the JCCR mode is 3, the ACT offset may be 2.
i. Alternatively, in one example, when the JCCR mode is 3, the ACT offset may be 1.
General technique (items 22-23)
22. In the above example, S 1 、S 2 L, h, m, n and/or k are integers and may depend on a. the message signaled in DPS/SPS/VPS/PPS/APS/picture header/slice group header/maximum coding unit (LCU)/Coding Unit (CU)/LCU row/LCU group/TU/PU block/video coding unit
Position of CU/PU/TU/block/video codec Unit
c. Codec mode for blocks containing samples along edges
d. Transformation matrix applied to blocks containing samples along edges
e. Block dimension/block shape of a current block and/or its neighboring blocks
f. Color format indication (e.g., 4:2:0, 4:4:4, RGB or YUV)
g. Coding and decoding tree structure (for example double tree or single tree)
h. Slice/slice group type and/or picture type
i. Color component (e.g., only possible for Cb or Cr)
j. Time domain layer ID
k. Standard grade/level/hierarchy
Alternatively, S 1 、S 2 L, h, m, n and/or k may be signaled to the decoder.
23. The proposed method described above can be applied under certain conditions.
a. In one example, the condition is that the color format is 4:2:0 and/or 4:2: 2.
b. In one example, an indication of the use of the above method may be signaled in the sequence/picture/slice/tile/video region level (e.g., SPS/PPS/picture header/slice header).
c. In one example, the use of the above method may depend on
i. Video content (e.g. screen content or natural content)
Message signaled in DPS/SPS/VPS/PPS/APS/Picture header/slice group header/maximum codec Unit (LCU)/Codec Unit (CU)/LCU line/LCU group/TU/PU Block/video codec Unit
Location of CU/PU/TU/block/video codec Unit
Codec mode for blocks containing samples along an edge
v. transform matrix applied to blocks containing samples along an edge
Block dimension of a current block and/or its neighboring blocks
Block shape of a current block and/or its neighboring blocks
Color format indication (e.g., 4:2:0, 4:4:4, RGB or YUV)
Encoding and decoding tree structure (e.g. double tree or single tree)
x. slice/slice group type and/or picture type
Color component (e.g., possibly applied only to Cb or Cr)
xi time domain layer ID
Standard grade/level/hierarchy
Alternatively, m and/or n may be signaled to the decoder.
5. Examples of the embodiments
The example is based on JFET-P2001-vE. For newly added text
Figure BDA0003734881610000282
And (4) highlighting. The deleted text is marked with italicized text.
5.1 example #1
This embodiment relates to the interaction between ACT and BDPCM modes.
Figure BDA0003734881610000281
Figure BDA0003734881610000291
5.2. Example #2
This embodiment relates to the interaction between ACT and BDPCM modes.
intra _ bdpcmm _ chroma _ flag equal to 1 specifies that BDPCM is applied to the current chroma coded block at position (x0, y0), i.e., the skip transform, and the intra chroma prediction mode is specified by intra _ bdpcmm _ chroma _ dir _ flag. intra _ bdpcmm _ chroma _ flag equal to 0 specifies that BDPCM is not applied to the current chroma coded block at position (x0, y 0).
When intra _ bdpcmm _ chroma _ flag does not exist and
Figure BDA0003734881610000292
in case of false, it is inferred to be equal to 0.
Figure BDA0003734881610000293
For x0., x0+ cbWidth-1, y y0., y0+ cbHeight-1, and cid 1, 2, the variable BdpcmFlag [ x ] [ y ] [ cid ] is set equal to intra _ bdpcmm _ chroma _ flag.
intra _ bdpcmm _ chroma _ dir _ flag equal to 0 specifies the BDPCM prediction direction as horizontal. Intra _ bdpcmm _chroma _ dir _ flag equal to 1 specifies that the BDPCM prediction direction is vertical.
Figure BDA0003734881610000301
Figure BDA0003734881610000302
The variable BdpcmDir [ x ] [ y ] [ cIdx ] is set equal to intra _ bdpcmm _ chroma _ dir _ flag (x ═ x0.. x0+ cbWidth-1, y ═ y0.. y0+ cbHeight-1, and cIdx ═ 1.
5.3. Example #3
This embodiment is related to QP settings.
8.7.3 scaling process for transform coefficients
The inputs to this process are:
specifying a luminance position (xTbY, yTbY) of the current luminance transform block relative to an upper-left sample of an upper-left luminance sample of the current picture,
a variable nTbW specifying the transform block width,
a variable nTbH specifying the transform block height,
a variable predMode specifying the prediction mode of the codec unit,
a variable cIdx specifying the color component of the current block.
The output of this process is an (nTbW) × (nTbH) array d of scaled transform coefficients with the element d [ x ] [ y ].
The quantization parameter qP is modified and the variables rectnottsflag and bdShift are derived as follows:
if transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] is equal to 0, the following applies:
qP=qP-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1133)
Figure BDA0003734881610000303
rectNonTsFlag=(((Log2(nTbW)+Log2(nTbH))&1)==1)?1:0 (1134)
Figure BDA0003734881610000304
otherwise (transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equals 1), the following applies:
qP=Max(QpPrimeTsMin,qP)-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1136)
Figure BDA0003734881610000312
rectNonTsFlag=0 (1137)
bdShift=10 (1138)
5.4. example #4
8.7.1 derivation of quantization parameter
The chroma quantization parameters for the Cb and Cr components, Qp ' Cb and Qp ' Cr, and the joint Cb-Cr codec Qp ' CbCr are derived as follows:
Qp′Cb=Clip3(-QpBdOffset,63,qPCb+pps_cb_qp_offset+slice_cb_qp_offset+CuQpOffsetCb)+QpBdOffset (1122)
Qp′Cr=Clip3(-QpBdOffset,63,qPCr+pps_cr_qp_offset+slice_cr_qp_offset+CuQpOffsetCr)+QpBdOffset (1123)
Qp′CbCr=Clip3(-QpBdOffset,63,qPCbCr+pps_joint_cbcr_qp_offset+slice_joint_cbcr_qp_offset+CuQpOffsetCbCr)+QpBdOffset (1124)
5.5. example #5
7.3.9.5 codec Unit syntax
Figure BDA0003734881610000311
Figure BDA0003734881610000321
K0 and K1 are set equal to 32.
5.6. Example #6
7.3.9.5 codec Unit syntax
Figure BDA0003734881610000322
5.7. Example #7
8.7.3 scaling process for transform coefficients
The inputs to this process are:
-specifying a luminance position (xTbY, yTbY) of the current luminance transformation block relative to an upper left sample of an upper left luminance sample of the current picture,
a variable nTbW specifying the transform block width,
-a variable nTbH specifying the transform block height,
a variable predMode specifying the prediction mode of the codec unit,
-a variable cIdx specifying the color component of the current block.
The output of this process is an (nTbW) × (nTbH) array d of scaled transform coefficients with the element d [ x ] [ y ].
-if cIdx equals 0, the following applies:
Figure BDA0003734881610000331
qP=Qp′ Y (1129)
otherwise, if tucresimode [ xTbY ] [ yTbY ] is equal to 2, the following applies:
Figure BDA0003734881610000332
qP=Qp′ CbCr (1130)
otherwise, if cIdx is equal to 1, the following applies:
ActQPoffset=-5
qP=qP′Cb (1131)
else (cIdx equals 2), the following applies:
Figure BDA0003734881610000333
qP=Qp′ Cb (1131)
the quantization parameter qP is modified and the variables rectnottsflag and bdShift are derived as follows:
-if transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] is equal to 0, the following applies:
qP=qP-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1133)
Figure BDA0003734881610000334
rectNonTsFlag=(((Log2(nTbW)+Log2(nTbH))&1)==1)?1:0 (1134)
Figure BDA0003734881610000335
-otherwise (transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equals 1), the following applies:
qP=Max(QpPrimeTsMin,qP)-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1136)
Figure BDA0003734881610000336
rectNonTsFlag=0 (1137)
bdShift=10 (1138)
the variable bdOffset is derived as follows:
bdOffset=(1<<bdShift)>>1 (1139)
the list levelScale [ ] [ ] is specified as levelScale [ j ] [ k ] { {40,45,51,57,64,72}, {57,64,72,80,90,102} }, where j ═ 0..1, k ═ 0..5.
5.8. Example #8
8.7.3 scaling process for transform coefficients
The inputs to this process include:
-specifying a luminance position (xTbY, yTbY) of the current luminance transformation block relative to an upper left sample of an upper left luminance sample of the current picture,
a variable nTbW specifying the transform block width,
-a variable nTbH specifying the transform block height,
a variable predMode specifying the prediction mode of the codec unit,
-a variable cIdx specifying the color component of the current block.
The output of this process is an (nTbW) × (nTbH) array d of scaled transform coefficients with an element d [ x ] [ y ].
-if cIdx equals 0, the following applies:
qP=Qp′Y (1129)
Figure BDA0003734881610000341
otherwise, if tucresimode [ xTbY ] [ yTbY ] is equal to 2, the following applies:
Figure BDA0003734881610000342
qP=Qp′ CbCr (1130)
otherwise, if cIdx is equal to 1, the following applies:
Figure BDA0003734881610000343
qP=Qp′ Cb (1131)
-otherwise (cIdx equals 2), the following applies:
Figure BDA0003734881610000351
qP=Qp′ Cr (1132)
the quantization parameter qP is modified and the variables rectnottsflag and bdShift are derived as follows:
-if transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] is equal to 0, the following applies:
qP=qP-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1133)
rectNonTsFlag=(((Log2(nTbW)+Log2(nTbH))&1)==1)?1:0 (1134)
bdShift=BitDepth+rectNonTsFlag+ (1135)
((Log2(nTbW)+Log2(nTbH))/2)-5+pic_dep_quant_enabled_flag
-otherwise (transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equals 1), the following applies:
qP=Max(QpPrimeTsMin,qP)-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1136)
Figure BDA0003734881610000352
rectNonTsFlag=0 (1137)
bdShift=10 (1138)
the variable bdOffset is derived as follows:
bdOffset=(1<<bdShift)>>1 (1139)
the list levelScale [ ] [ ] is specified as levelScale [ j ] [ k ] { {40,45,51,57,64,72}, {57,64,72,80,90,102} } where j ═ 0..1, k ═ 0..5.
5.9. Example #9
7.4.3.4 picture parameter set RBSP semantics
Figure BDA0003734881610000361
8.7.3 scaling process for transform coefficients
The inputs to this process are:
-specifying a luminance position (xTbY, yTbY) of the current luminance transformation block relative to an upper left sample of an upper left luminance sample of the current picture,
-a variable nTbW specifying the transform block width,
-a variable nTbH specifying the transform block height,
a variable predMode specifying the prediction mode of the codec unit,
-a variable cIdx specifying the color component of the current block.
The output of this process is an (nTbW) × (nTbH) array d of scaled transform coefficients with the element d [ x ] [ y ].
-if cIdx is equal to 0, the following applies:
Figure BDA0003734881610000362
qP=Qp′ Y (1129)
otherwise, if tucresimode [ xTbY ] [ yTbY ] is equal to 2, the following applies:
Figure BDA0003734881610000363
qP=Qp′ CbCr (1130)
else, if cIdx is equal to 1, the following applies:
Figure BDA0003734881610000364
qP=Qp′ Cb (1131)
-otherwise (cIdx equals 2), the following applies:
Figure BDA0003734881610000371
qP=Qp′ Cr (1132)
the quantization parameter qP is modified and the variables rectnottsflag and bdShift are derived as follows:
-if transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] is equal to 0, the following applies:
qP=qP-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1133)
Figure BDA0003734881610000372
rectNonTsFlag=(((Log2(nTbW)+Log2(nTbH))&1)==1)?1:0 (1134)
Figure BDA0003734881610000374
-otherwise (transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equals 1), the following applies:
qP=Max(QpPrimeTsMin,qP)-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1136)
Figure BDA0003734881610000373
rectNonTsFlag=0 (1137)
bdShift=10 (1138)
the variable bdOffset is derived as follows:
bdOffset=(1<<bdShift)>>1 (1139)
the list levelScale [ ] [ ] is specified as levelScale [ j ] [ k ] { {40,45,51,57,64,72}, {57,64,72,80,90,102} }, where j ═ 0..1, k ═ 0..5.
5.10. Example #10
7.4.3.4 picture parameter set RBSP semantics
Figure BDA0003734881610000381
8.7.3 scaling process for transform coefficients
The inputs to this process are:
-specifying a luminance position (xTbY, yTbY) of the current luminance transformation block relative to an upper left sample of an upper left luminance sample of the current picture,
a variable nTbW specifying the transform block width,
-a variable nTbH specifying the transform block height,
a variable predMode specifying the prediction mode of the codec unit,
-a variable cIdx specifying the color component of the current block.
The output of this process is an (nTbW) × (nTbH) array d of scaled transform coefficients with the element d [ x ] [ y ].
-if cIdx is equal to 0, the following applies:
qP=Qp′Y (1129)
Figure BDA0003734881610000382
otherwise, if tucresimode [ xTbY ] [ yTbY ] is equal to 2, the following applies:
Figure BDA0003734881610000383
Figure BDA0003734881610000391
qP=Qp′ CbCr (1130)
else, if cIdx is equal to 1, the following applies:
-if cu _ act _ enabled _ flag [ xTbY ] [ yTbY ] is equal to 1, then the following applies:
Figure BDA0003734881610000392
qP=Qp′ Cb (1131)
else (cIdx equals 2), the following applies:
Figure BDA0003734881610000393
qP=Qp′ Cr (1132)
the quantization parameter qP is modified and the variables rectnottsflag and bdShift are derived as follows:
-if transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] is equal to 0, the following applies:
qP=qP-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1133)
rectNonTsFlag=(((Log2(nTbW)+Log2(nTbH))&1)==1)?1:0 (1134)
Figure BDA0003734881610000394
-otherwise (transform _ skip _ flag [ xTbY ] [ yTbY ] [ cIdx ] equals 1), the following applies:
qP=Max(QpPrimeTsMin,qP)-(cu_act_enabled_flag[xTbY][yTbY]?5:0) (1136)
Figure BDA0003734881610000401
rectNonTsFlag=0 (1137)
bdShift=10 (1138)
the variable bdOffset is derived as follows:
bdOffset=(1<<bdShift)>>1 (1139)
the list levelScale [ ] [ ] is specified as levelScale [ j ] [ k ] { {40,45,51,57,64,72}, {57,64,72,80,90,102} } where j ═ 0..1, k ═ 0..5.
Fig. 7 is a block diagram illustrating an example video processing system 700 in which various techniques disclosed herein may be implemented. Various implementations may include some or all of the components of system 700. The system 700 can include an input 702 for receiving video content. The video content may be received in a raw or uncompressed format, such as 8 or 10 bit multi-component pixel values, or may be received in a compressed or encoded format. Input 702 may represent a network interface, a peripheral bus interface, or a storage interface. Examples of network interfaces include wired interfaces, such as ethernet, Passive Optical Networks (PONs), etc., and wireless interfaces, such as Wi-Fi or cellular interfaces.
The system 700 may include a codec component 704 that may implement various codecs or encoding methods described in this document. Codec component 704 can reduce the average bit rate of the video from input 702 to the output of codec component 704 to produce a codec representation of the video. Thus, codec techniques are sometimes referred to as video compression or video transcoding techniques. The output of the codec component 704 can be stored or transmitted via communication over a connection represented by component 706. A stored or transmitted bitstream (or codec) representation of video received at input 702 can be used by component 708 to generate pixel values or displayable video that is sent to display interface 710. The process of generating user-viewable video from a bitstream representation is sometimes referred to as video decompression. Furthermore, although certain video processing operations are referred to as "codec" operations or tools, it will be understood that the encoding tools or operations are used at the encoder and the corresponding decoding tools or operations that reverse the encoding results will be performed by the decoder.
Examples of a peripheral bus interface or display interface may include a Universal Serial Bus (USB) or High Definition Multimedia Interface (HDMI) or display port, among others. Examples of storage interfaces include SATA (serial advanced technology attachment), PCI, IDE interfaces, and the like. The techniques described in this document may be embodied in various electronic devices such as mobile phones, laptop computers, smart phones, or other devices capable of performing digital data processing and/or video display.
Fig. 8 is a block diagram of a video processing device 800. The apparatus 800 may be used to implement one or more of the methods described herein. The apparatus 800 may be embodied in a smartphone, tablet, computer, internet of things (IoT) receiver, and the like. The apparatus 800 may include one or more processors 802, one or more memories 804, and video processing hardware 806. The processor 802 may be configured to implement one or more of the methods described in this document (e.g., fig. 5A and 5B). The one or more memories 804 may be used to store data and code for implementing the methods and techniques described herein. Video processing hardware 806 may be used to implement some of the techniques described in this document in hardware circuitry. In some embodiments, the hardware 806 may reside partially or entirely within the processor 802 (e.g., a graphics processor).
Fig. 9 is a block diagram illustrating an example video codec system 100 that may utilize techniques of this disclosure. As shown in fig. 9, the video codec system 100 may include a source device 110 and a destination device 120. Source device 110 generates encoded video data, which may be referred to as a video encoding device. Destination device 120 may decode the encoded video data generated by source device 110, and source device 110 may be referred to as a video decoding device. The source device 110 may include a video source 112, a video encoder 114, and an input/output (I/O) interface 116.
The video source 112 may include a source such as a video capture device, an interface that receives video data from a video content provider and/or a computer graphics system for generating video data, or a combination of these sources. The video data may include one or more pictures. The video encoder 114 encodes video data from the video source 112 to generate a bitstream. The bitstream may comprise a sequence of bits forming a codec representation of the video data. The bitstream may include coded pictures and associated data. A coded picture is a coded representation of a picture. The associated data may include sequence parameter sets, picture parameter sets, and other syntax structures. The I/O interface 116 may include a modulator/demodulator (modem) and/or a transmitter. The encoded video data may be sent directly to the destination device 120 over the network 130a via the I/O interface 116. The encoded video data may also be stored on a storage medium/server 130b for access by the destination device 120.
Destination device 120 may include I/O interface 126, video decoder 124, and display device 122.
I/O interface 126 may include a receiver and/or a modem. I/O interface 126 may retrieve encoded video data from source device 110 or storage medium/server 130 b. The video decoder 124 may decode the encoded video data. The display device 122 may display the decoded video data to a user. The display device 122 may be integrated with the destination device 120 or may be located external to the destination device 120, the destination device 120 being configured to interface with an external display device.
The video encoder 114 and the video decoder 124 may operate in accordance with video compression standards such as the High Efficiency Video Codec (HEVC) standard, the universal video codec (VVM) standard, and other current and/or out-of-the-curtain standards.
Fig. 10 is a block diagram illustrating an example of a video encoder 200, which video encoder 200 may be the video encoder 114 in the system 100 shown in fig. 9.
Video encoder 200 may be configured to perform any or all of the techniques of this disclosure. In the example of fig. 10, the video encoder 200 includes a number of functional components. The techniques described in this disclosure may be shared among various components of video encoder 200. In some examples, the processor may be configured to perform any or all of the techniques described in this disclosure.
The functional components of the video encoder 200 may include a partitioning unit 201, a prediction unit 202, which may include a mode selection unit 203, a motion estimation unit 204, a motion compensation unit 205, and an intra prediction unit 206, a residual generation unit 207, a transform unit 208, a quantization unit 209, an inverse quantization unit 210, an inverse transform unit 211, a reconstruction unit 212, a buffer 213, and an entropy coding unit 214.
In other examples, video encoder 200 may include more, fewer, or different functional components. In one example, the prediction unit 202 may include an Intra Block Copy (IBC) unit. The IBC unit may perform prediction in an IBC mode, where the at least one reference picture is a picture in which the current video block is located.
Furthermore, some components, such as the motion estimation unit 204 and the motion compensation unit 205, may be highly integrated, but are represented separately in the example of fig. 10 for explanation purposes.
The partition unit 201 may partition a picture into one or more video blocks. The video encoder 200 and the video decoder 300 may support various video block sizes.
The mode selection unit 203 may, for example, select one of the coding and decoding modes, intra or inter, based on the error result, and provide the resulting intra or inter coded block to the residual generation unit 207 to generate residual block data and to the reconstruction unit 212 to reconstruct the coded block to be used as a reference picture. In some examples, mode selection unit 203 may select a combination of intra-prediction and inter-prediction (CIIP) modes, where the prediction is based on the inter-prediction signal and the intra-prediction signal. Mode selection unit 203 may also select the resolution (e.g., sub-pixel or integer-pixel precision) of the motion vectors for the block in the case of inter prediction.
To perform inter prediction on the current video block, motion estimation unit 204 may generate motion information for the current video block by comparing one or more reference frames from buffer 213 to the current video block. Motion compensation unit 205 may determine a prediction video block for the current video block based on the motion information and decoded samples of the picture from buffer 213 (excluding the picture associated with the current video block).
The motion estimation unit 204 and the motion compensation unit 205 may perform different operations on the current video block, for example, depending on whether the current video block is in an I-slice, a P-slice, or a B-slice.
In some examples, motion estimation unit 204 may perform uni-directional prediction on the current video block, and motion estimation unit 204 may search for a reference video block of the current video block in a list 0 or list 1 reference picture. Motion estimation unit 204 may then generate a reference index that indicates a reference picture in list 0 or list 1 that includes the reference video block, and a motion vector that indicates spatial displacement between the current video block and the reference video block. Motion estimation unit 204 may output the reference index, the prediction direction indicator, and the motion vector as motion information for the current video block. The motion compensation unit 205 may generate a prediction video block of the current block based on a reference video block indicated by motion information of the current video block.
In other examples, motion estimation unit 204 may perform bi-prediction on the current video block, and motion estimation unit 204 may search a reference video block of the current video block in a reference picture in list 0 and may also search another reference video block of the current video block in a reference picture in list 1. Motion estimation unit 204 may then generate reference indices that indicate reference pictures in list 0 and list 1 that include reference video blocks, and motion vectors that indicate spatial displacements between the reference video blocks and the current video block. Motion estimation unit 204 may output the reference index and the motion vector of the current video block as motion information for the current video block. Motion compensation unit 205 may generate a prediction video block for the current video block based on the reference video block indicated by the motion information for the current video block.
In some examples, motion estimation unit 204 may output the complete set of motion information for the decoding process of the decoder.
In some examples, motion estimation unit 204 may not output the full set of motion information for the current video. Instead, motion estimation unit 204 may signal motion information for the current video block with reference to motion information of another video block. For example, motion estimation unit 204 may determine that the motion information of the current video block is sufficiently similar to the motion information of the neighboring video block.
In one example, motion estimation unit 204 may indicate a value in a syntax structure associated with the current video block that indicates to video decoder 300 that the current video block has the same motion information as another video block.
In another example, motion estimation unit 204 may identify another video block and a Motion Vector Difference (MVD) in a syntax structure associated with the current video block. The motion vector difference indicates the difference between the motion vector of the current video block and the motion vector of the indicated video block. The video decoder 300 may use the motion vector and the motion vector difference for the indicated video block to determine the motion vector for the current video block.
As discussed above, the video encoder 200 may predictively signal motion vectors. Two examples of predictive signaling techniques that may be implemented by video encoder 200 include Advanced Motion Vector Prediction (AMVP) and merge mode signaling.
The intra prediction unit 206 may perform intra prediction on the current video block. When intra prediction unit 206 performs intra prediction on the current video block, intra prediction unit 206 may generate prediction data for the current video block based on decoded samples of other video blocks in the same picture. The prediction data for the current video block may include a prediction video block and various syntax elements.
Residual generation unit 207 may generate residual data for the current video block by subtracting (e.g., indicated by a minus sign) the prediction video block of the current video block from the current video block. The residual data for the current video block may include residual video blocks corresponding to different sample components of samples in the current video block.
In other examples, for example in skip mode, there may be no residual data for the current video block and residual generation unit 207 may not perform the subtraction operation.
Transform processing unit 208 may generate one or more transform coefficient video blocks for the current video block by applying one or more transforms to a residual video block associated with the current video block.
After transform processing unit 208 generates a transform coefficient video block associated with the current video block, quantization unit 209 may quantize the transform coefficient video block associated with the current video block based on one or more Quantization Parameter (QP) values associated with the current video block.
Inverse quantization unit 210 and inverse transform unit 211 may apply inverse quantization and inverse transform, respectively, to the transform coefficient video blocks to reconstruct residual video blocks from the transform coefficient video blocks. Reconstruction unit 212 may add the reconstructed residual video block to corresponding samples from one or more prediction video blocks generated by prediction unit 202 to produce a reconstructed video block associated with the current block for storage in buffer 213.
After reconstruction unit 212 reconstructs the video block, a loop filtering operation may be performed to reduce video block artifacts in the video block.
Entropy encoding unit 214 may receive data from other functional components of video encoder 200. When entropy encoding unit 214 receives the data, entropy encoding unit 214 may perform one or more entropy encoding operations to generate entropy encoded data and output a bitstream that includes the entropy encoded data.
Fig. 11 is a block diagram illustrating an example of a video decoder 300, the video decoder 300 may be the video decoder 114 in the system 100 shown in fig. 9.
Video decoder 300 may be configured to perform any or all of the techniques of this disclosure. In the example of fig. 11, the video decoder 300 includes a number of functional components. The techniques described in this disclosure may be shared among various components of the video decoder 300. In some examples, the processor may be configured to perform any or all of the techniques described in this disclosure.
In the example of fig. 11, the video decoder 300 includes an entropy decoding unit 301, a motion compensation unit 302, an intra prediction unit 303, an inverse quantization unit 304, an inverse transform unit 305, a reconstruction unit 306, and a buffer 307. In some examples, video decoder 300 may perform a decoding process that is generally the inverse of the encoding process described with respect to video encoder 200 (fig. 10).
The entropy decoding unit 301 may retrieve the encoded bitstream. The encoded bitstream may include entropy encoded video data (e.g., encoded blocks of video data). Entropy decoding unit 301 may decode entropy encoded video data, and motion compensation unit 302 may determine motion information from the entropy decoded video data, including motion vectors, motion vector precision, reference picture list indices, and other motion information. For example, the motion compensation unit 302 may determine such information by performing AMVP and merge modes.
The motion compensation unit 302 may generate a motion compensation block, possibly based on an interpolation filter, to perform interpolation. An identifier for the interpolation filter used with sub-pixel precision may be included in the syntax element.
Motion compensation unit 302 may use interpolation filters used by video encoder 20 during video block encoding to calculate an interpolation of sub-integer pixels of a reference block. The motion compensation unit 302 may determine an interpolation filter used by the video encoder 200 according to the received syntax information and generate a prediction block using the interpolation filter.
The motion compensation unit 302 may use some syntax information to determine the size of blocks used to encode frames and/or slices of an encoded video sequence, partition information that describes how to partition each macroblock of a picture of the encoded video sequence, a mode that indicates how to encode each partition, one or more reference frames (and reference frame lists) of each inter-coded block, and other information used to decode the encoded video sequence.
The intra prediction unit 303 may form a prediction block from spatially neighboring blocks using, for example, an intra prediction mode received in a bitstream. The inverse quantization unit 303 inversely quantizes (i.e., dequantizes) the quantized video block coefficients provided in the bit stream and decoded by the entropy decoding unit 301. The inverse transform unit 303 applies inverse transform.
The reconstruction unit 306 may sum the residual block with the corresponding prediction block generated by the motion compensation unit 202 or the intra prediction unit 303 to form a decoded block. A deblocking filter may also be applied to filter the decoded blocks, if desired, to remove blocking artifacts. The decoded video blocks are then stored in a buffer 307, the buffer 307 providing reference blocks for subsequent motion compensation/intra prediction, and also generating decoded video for presentation on a display device.
Fig. 12 illustrates an example method in which the above-described aspects may be implemented, such as the embodiments shown in fig. 7-11.
As shown in fig. 12, a flow diagram of an example method 1200 of video processing includes, at operation 1210, determining that a coding mode of a current video block of a video includes an Adaptive Color Transform (ACT) mode, and enabling a chroma residual joint coding (JCCR) coding tool for coding the current video block.
At operation 1220, method 1200 includes performing a conversion between the video and a bitstream of the video based on the determination, the quantization parameter offset used to codec the current video block being based on a mode used by the JCCR codec tool.
The following provides a list of solutions preferred by some embodiments.
1.A method of video processing, comprising determining that a coding mode of a current video block of a video comprises an Adaptive Color Transform (ACT) mode, and enabling a chroma residual joint coding (JCCR) coding tool for coding the current video block, and performing a transform between the video and a bitstream of the video based on the determination, wherein a quantization parameter offset used for coding the current video block is based on a mode used by the JCCR coding tool.
2. The method of solution 1, wherein the ACT mode uses YCgCo color transform.
3. The method of solution 2, where the QP offset is-5 or-6 since the mode of the JCCR tool is 1.
4. The method of solution 2, where the QP offset is-7 since the mode of the JCCR tool is 2.
5. The method of solution 2, where the QP offset is-4 or-5 since the mode of the JCCR tool is 3.
6. The method of solution 1, wherein the ACT mode uses YCgCo-R color transform.
7. The method of solution 6, where the QP offset is 0 or 1 since the mode of the JCCR tool is 1.
8. The method of solution 6, where the QP offset is-1 since the mode of the JCCR tool is 2.
9. The method of solution 6, where the QP offset is 1 or 2 since the mode of the JCCR tool is 3.
10. The method according to any of solutions 1 to 9, wherein the QP offset is act _ QP _ offset.
11. The method according to any of solutions 1 to 10, wherein converting comprises decoding the video from the bitstream representation.
12. The method according to any of solutions 1 to 10, wherein the converting comprises encoding the video into a bitstream representation.
13. The method according to any of solutions 1 to 10, wherein converting comprises generating a bitstream from a current video block, and wherein the method further comprises storing the bitstream in a non-transitory computer-readable storage medium.
14. A method of storing a bitstream representing a video to a computer readable storage medium, comprising generating the bitstream from the video according to the method of any one or more of claims 1 to 10; and writing the bitstream to a computer readable storage medium.
15. A video processing apparatus comprising a processor configured to implement a method as claimed in any one or more of claims 1 to 14.
16. A computer readable medium having stored thereon instructions that, when executed, cause a processor to implement a method according to any one or more of claims 1 to 14.
17. A computer readable medium storing a bitstream representation generated according to any one or more of claims 1 to 14.
18. A video processing apparatus for storing a bitstream representation, wherein the video processing apparatus is configured to implement the method according to any one or more of claims 1 to 14.
Another list of solutions preferred by some embodiments is provided next.
P 1.a video processing method comprising determining whether a chroma block Based Delta Pulse Code Modulation (BDPCM) mode is enabled for a video block of a video based on whether use of an Adaptive Color Transform (ACT) mode and/or a luma BDPCM mode of the video block is enabled; and performing a conversion between the video block and a bitstream representation of the video based on the determination.
P2. the method according to solution P1, wherein the signaling of the first value of the first flag associated with enabling the chroma BDPCM mode is determined based on the signaling of the ACT mode enabled for the video block and the signaling of the second value of the second flag associated with the use of the luma BDPCM mode.
P3. the method according to solution P2, wherein the first value of the first flag has a false value in response to the ACT mode being enabled and the second value of the second flag has a false value.
P4. the method according to solution P2, wherein the first value of the first flag has a true value in response to the second value of the second flag having a true value.
P5., the method according to solution P1, wherein the signaling of ACT mode of a video block is conditionally based on the same BDPCM prediction direction for luma and chroma samples of the video block.
P6. method according to solution P5, wherein the signaling of the ACT mode is indicated after the signaling of the chroma BDPCM mode and the luma BDPCM mode.
P7. the method according to solution P1, wherein, in response to the use of the ACT mode being enabled, a first value indicative of a first prediction direction of the chroma BDPCM mode is derived from a second value indicative of a second prediction direction of the luma BDPCM mode.
P8. the method according to solution P7, wherein the first value indicating the first prediction direction of the chroma BDPCM mode is the same as the second value indicating the second prediction direction of the luma BDPCM mode.
P9. the method according to solution P8, wherein the first prediction direction of the chroma BDPCM mode and the second prediction direction of the luma BDPCM mode are in the horizontal direction.
P10. the method according to solution P8, wherein the first prediction direction of the chroma BDPCM mode and the second prediction direction of the luma BDPCM mode are in a vertical direction.
P11. the method according to solution P1, wherein the first value indicating the first prediction direction of the chroma BDPCM mode is zero in response to the use of the ACT mode being disabled.
P12. a video processing method comprising determining whether a block-Based Delta Pulse Codec Modulation (BDPCM) mode is enabled for a video block of a video based on whether use of an Adaptive Color Transform (ACT) mode of the video block is enabled; and performing a conversion between the video block and a bitstream representation of the video according to the determination.
P13. the method according to solution P12, wherein, in response to the ACT mode being enabled for a video block, the BDPCM mode is disabled for the video block.
P14. method according to solution P13, wherein the first flag indicating BDPCM mode is signaled after the second flag indicating ACT mode.
P15. method according to solution P13, wherein no flag indicating BDPCM mode is signaled, wherein the flag is determined to be false or zero.
P16. the method according to solution P12, wherein ACT mode is disabled for video blocks in response to BDPCM mode being enabled for the video blocks.
P17. the method according to solution P16, wherein the first flag indicating BDPCM mode is signaled before the second flag indicating ACT mode.
P18. method according to solution P16, wherein no flag indicating ACT mode is signaled, wherein the flag is determined to be false or zero.
P19. the method according to any of solutions P12 to P18, wherein the BDPCM modes comprise luma BDPCM modes and/or chroma BDPCM modes.
P20. the method according to solution P1, wherein the ACT mode is applied when the chroma BDPCM mode and the luma BDPCM mode are associated with different prediction modes.
P21. the method according to solution P20, wherein the forward ACT mode is applied after the chroma BDPCM mode or the luma BDPCM mode.
P22. the method according to any of solutions P1 to P21, wherein the Quantization Parameter (QP) of the video block is clipped in response to ACT mode being enabled.
P23. method according to solution P22, where the clipping function for clipping QP is defined as (l, h, x), where l is the lowest possible value of input x and h is the highest possible value of input x.
P24. method according to solution P23, wherein l equals zero.
P25. the method according to solution P23, wherein h equals 63.
P26. the method according to solution P22, wherein the QP of the video block is clipped after the QP is adjusted for ACT mode.
P27. the method according to solution P23, wherein, in response to a transform skip being applied to the video block, l is equal to the minimum allowed QP of the transform skip mode.
P28. method according to any of solutions P23 to P26, wherein l, h, m, n and/or k are integers that depend on (i) a message signaled in DPS/SPS/VPS/PPS/APS/picture header/slice group header/maximum codec unit (LCU)/Codec Unit (CU)/LCU line/LCU group/TU/PU block/video codec unit, (ii) the location of CU/PU/TU/block/video codec unit, (iii) the codec mode of the block containing samples along the edge, (iv) the transform matrix applied to the block containing samples along the edge, (v) the block size/block shape of the current block and/or its neighboring blocks, (vi) the color format (e.g. 4:2:0, b, c, 4:4:4, RGB or YUV), (vii) a coding tree structure (e.g., dual tree or single tree), (viii) slice/slice group type and/or picture type, (ix) color components (e.g., which may only be applied to Cb or Cr), (x) temporal layer ID, or (xi) the level/hierarchy of the standard.
P29. the method according to any of the solutions P23 to P26, wherein l, h, m, n and/or k are signaled to the decoder.
P30. method according to solution P30, wherein the color format is 4:2:0 or 4:2: 2.
P31. method according to any of the solutions P1 to P30, wherein an indication of ACT mode or BDPCM mode or chroma BDPCM mode or luma BDPCM mode is signaled at sequence, picture, slice, tile or video region level.
P32. a video processing method comprising determining to use a chroma residual joint coding (JCCR) tool for video blocks of an Adaptive Color Transform (ACT) mode enabled video; and performing a conversion between the video block and a bitstream representation of the video based on the determination, wherein a Quantization Parameter (QP) of the ACT mode is based on a mode of the JCCR tool.
P33. the method according to solution P32, wherein when the mode of the JCCR tool is determined to be 1, QP is-5 or-6.
P34. the method according to solution P32, wherein when the pattern of the JCCR tool is determined to be 3, QP is-4 or-5.
P35. the method according to any of solutions P1 to P34, wherein converting comprises parsing and decoding the codec representation to generate video pixels.
P36. the method according to any of solutions P1 to P34, wherein converting comprises generating a codec representation by encoding the video.
P37. a video decoding device comprising a processor configured to implement the method according to one or more of solutions P1 to P36.
P38. a video coding device comprising a processor configured to implement the method of one or more of solutions P1 to P36.
P39. a computer program product having stored thereon computer code which, when executed by a processor, causes the processor to carry out the method according to solutions P1 to P36.
In this document, the term "video processing" may refer to video encoding, video decoding, video compression, or video decompression. For example, a video compression algorithm may be applied during conversion from a pixel representation of a video to a corresponding bitstream representation, and vice versa. For example, the bitstream representation of the current video block may correspond to bits that are co-located or dispersed at different locations within the bitstream, as defined by the syntax. For example, a macroblock may be encoded according to transform and codec error residual values, and may also be encoded using bits in the header and other fields in the bitstream.
The disclosure and other solutions, examples, embodiments, modules, and functional operations described in this document can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware (including the structures disclosed in this document and their structural equivalents), or in combinations of one or more of them. The disclosed embodiments and other embodiments may be implemented as one or more computer program products, i.e., one or more modules of computer program instructions encoded on a computer-readable medium for execution by, or to control the operation of, data processing apparatus. The computer readable medium can be a machine-readable storage device, a machine-readable storage substrate, a memory device, a composition of matter effecting a machine-readable propagated signal, or a combination of one or more of them. The term "data processing apparatus" encompasses all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them. A propagated signal is an artificially generated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus.
A computer program (also known as a program, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program does not necessarily correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network.
The processes and logic flows described in this document can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. The essential elements of a computer are a processor for executing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks. However, a computer need not have such a device. Computer-readable media suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, such as internal hard disks or removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
Although this patent document contains many specifics, these should not be construed as limitations on the scope of any subject matter or of what may be claimed, but rather as descriptions of features that may be specific to particular embodiments of particular technologies. Certain features that are described in this patent document in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features may in some cases be deleted from a claimed combination and a claimed combination may be directed to a subcombination or variation of a subcombination.
Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. Moreover, the separation of various system components in the embodiments described in this patent document should not be understood as requiring such separation in all embodiments.
Only a few implementations and examples are described, and other implementations, enhancements and variations can be made based on what is described and illustrated in this patent document.

Claims (18)

1.A video processing method, comprising:
determining that a coding mode of a current video block of a video comprises an Adaptive Color Transform (ACT) mode, and enabling a chroma residual joint coding (JCCR) coding tool for coding the current video block, an
Performing a conversion between the video and a bitstream of the video based on the determination,
wherein a quantization parameter QP offset used to codec the current video block is based on a mode used by the JCCR codec tool.
2. The method of claim 1, wherein the ACT mode uses YCgCo color transform.
3. The method of claim 2, wherein the QP offset is-5 or-6 due to the pattern of the JCCR tool being 1.
4. The method of claim 2, wherein the QP offset is-7 due to the pattern of the JCCR tool being 2.
5. The method of claim 2, wherein the QP offset is-4 or-5 due to the pattern of the JCCR tool being 3.
6. The method of claim 1, wherein the ACT mode uses YCgCo-R color transform.
7. The method of claim 6, wherein the QP offset is 0 or 1 due to the mode of the JCCR tool being 1.
8. The method of claim 6, wherein the QP offset is-1 due to the mode of the JCCR tool being 2.
9. The method of claim 6, wherein the QP offset is 1 or 2 due to the mode of the JCCR tool being 3.
10. The method of any of claims 1-9, wherein the QP offset is act QP offset.
11. The method of any of claims 1-10, wherein the converting comprises decoding the video from the bitstream.
12. The method of any of claims 1-10, wherein the converting comprises encoding the video into the bitstream.
13. The method of any of claims 1-10, wherein the converting comprises generating the bitstream from the current video block, and wherein the method further comprises:
storing the bitstream in a non-transitory computer-readable storage medium.
14. A method of storing a bitstream representing a video to a computer-readable storage medium, comprising:
generating a bitstream from a video according to the method of any one or more of claims 1 to 10; and
writing the bitstream to the computer-readable storage medium.
15. A video processing apparatus comprising a processor configured to implement the method of any one or more of claims 1 to 14.
16. A computer-readable medium having stored thereon instructions that, when executed, cause a processor to implement a method according to one or more of claims 1 to 14.
17. A computer readable medium storing the bitstream generated according to any one or more of claims 1 to 14.
18. A video processing apparatus storing a bitstream, wherein the video processing apparatus is configured to implement a method according to any one or more of claims 1 to 14.
CN202180008570.3A 2020-01-08 2021-01-07 Joint coding and decoding of chroma residual and adaptive color transform Pending CN114946187A (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CNPCT/CN2020/070957 2020-01-08
CN2020070957 2020-01-08
PCT/CN2021/070580 WO2021139707A1 (en) 2020-01-08 2021-01-07 Joint coding of chroma residuals and adaptive color transforms

Publications (1)

Publication Number Publication Date
CN114946187A true CN114946187A (en) 2022-08-26

Family

ID=76788415

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202180008570.3A Pending CN114946187A (en) 2020-01-08 2021-01-07 Joint coding and decoding of chroma residual and adaptive color transform

Country Status (2)

Country Link
CN (1) CN114946187A (en)
WO (1) WO2021139707A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024156273A1 (en) * 2023-01-28 2024-08-02 Douyin Vision Co., Ltd. Method, apparatus, and medium for video processing

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10271052B2 (en) * 2014-03-14 2019-04-23 Qualcomm Incorporated Universal color-space inverse transform coding
CN110971905B (en) * 2014-03-14 2023-11-17 Vid拓展公司 Method, apparatus and storage medium for encoding and decoding video content
US10158836B2 (en) * 2015-01-30 2018-12-18 Qualcomm Incorporated Clipping for cross-component prediction and adaptive color transform for video coding
WO2019006735A1 (en) * 2017-07-07 2019-01-10 Guangdong Virtual Reality Technology Co., Ltd. Methods, devices, and systems for identifying and tracking an object with multiple cameras

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024156273A1 (en) * 2023-01-28 2024-08-02 Douyin Vision Co., Ltd. Method, apparatus, and medium for video processing

Also Published As

Publication number Publication date
WO2021139707A1 (en) 2021-07-15

Similar Documents

Publication Publication Date Title
US11528497B2 (en) Palette mode with different partition structures
WO2021155834A1 (en) Deblocking parameters for chroma component
US11838523B2 (en) General constraints information for video coding
WO2021143750A1 (en) Overwriting of quantization parameters in video coding
WO2021121418A1 (en) Joint use of adaptive colour transform and differential coding of video
US20230362375A1 (en) Adaptive Colour Transform in Image/Video Coding
JP2024014958A (en) Pallet mode for local dual tree
JP2023523638A (en) entropy coding for split syntax
CN114946180B (en) Deblocking filtering using quantization parameter information
WO2021139707A1 (en) Joint coding of chroma residuals and adaptive color transforms
CN115211118A (en) Cluster-based palette mode for video coding

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination