WO2013155886A1 - Method and apparatus of simplified luma-based chroma intra prediction - Google Patents

Method and apparatus of simplified luma-based chroma intra prediction Download PDF

Info

Publication number
WO2013155886A1
WO2013155886A1 PCT/CN2013/070912 CN2013070912W WO2013155886A1 WO 2013155886 A1 WO2013155886 A1 WO 2013155886A1 CN 2013070912 W CN2013070912 W CN 2013070912W WO 2013155886 A1 WO2013155886 A1 WO 2013155886A1
Authority
WO
WIPO (PCT)
Prior art keywords
chroma
current
pixels
value
bitdepthc
Prior art date
Application number
PCT/CN2013/070912
Other languages
French (fr)
Inventor
Juliana HSU
Mei Guo
Xun Guo
Original Assignee
Mediatek Singapore Pte. Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mediatek Singapore Pte. Ltd. filed Critical Mediatek Singapore Pte. Ltd.
Priority to US14/380,709 priority Critical patent/US9794557B2/en
Priority to CN201380019470.6A priority patent/CN104471940B/en
Priority to EP13778200.9A priority patent/EP2839656B1/en
Priority to JP2015506080A priority patent/JP5977439B2/en
Publication of WO2013155886A1 publication Critical patent/WO2013155886A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques

Definitions

  • the present invention relates to video coding.
  • the present invention relates to coding techniques associated with simplified chroma intra prediction based on reconstructed luma and chroma pixels.
  • Motion compensated inter- frame coding has been widely adopted in various coding standards, such as MPEG-1/2/4 and H.261/H.263/H.264/AVC. While motion-compensated inter-frame coding can effectively reduce bitrate for compressed video, intra coding is required to compress the regions with high motion or scene changes. Besides, intra coding is also used to process an initial picture or to periodically insert I-pictures or I-blocks for random access or for alleviation of error propagation. Intra prediction exploits the spatial correlation within a picture or within a picture region. In practice, a picture or a picture region is divided into blocks and the intra prediction is performed on a block basis. Intra prediction for a current block can rely on pixels in neighboring blocks that have been processed.
  • neighboring blocks on the top and neighboring blocks on the left of the current block can be used to form intra prediction for pixels in the current block. While any pixels in the processed neighboring blocks can be used for intra predictor of pixels in the current block, very often only pixels of the neighboring blocks that are adjacent to the current block boundaries on the top and on the left are used.
  • the intra predictor is usually designed to exploit spatial features in the picture such as smooth area (DC mode), vertical line or edge, horizontal line or edge and diagonal line or edge. Furthermore, correlation often exists between the luminance (luma) and chrominance (chroma) components. Therefore, reconstructed luma pixels can be used to derive the intra chroma prediction.
  • HEVC High Efficiency Video Coding
  • LM prediction LM mode
  • the main concept is to use the reconstructed luma pixels to generate the predictors of corresponding chroma pixels.
  • IB illustrate the prediction procedure. First, the neighboring reconstructed pixels of a co-located luma block in Fig. 1 A and the neighboring reconstructed pixels of a chroma block in Fig. IB are used to derive the correlation parameters between the blocks. Then, the predicted pixels of the chroma block are generated using the parameters and the
  • reconstructed pixel row and the second left reconstructed pixel column of the current luma block are used.
  • the specific row and column of the luma block are used in order to match the 4:2:0 sampling format of the chroma components.
  • the following illustration is based on 4:2:0 sampling format.
  • LM-mode chroma intra prediction for other sampling formats can be derived similarly.
  • the LM mode is applied to predict chroma samples based on a linear model using reconstructed luma samples of the co-located prediction unit (PU).
  • the parameters of the linear model consist of slope (a»k) and y-intercept (b), where "»" corresponds to the right shift operation.
  • the parameters are derived based on the neighboring luma and chroma samples according to a least mean square criterion.
  • variable k3 and the sample array pY' are derived as:
  • BitDepthc denotes the internal chroma bit depth (i.e., the bit depth with which the chroma signal is processed during video coding process), and
  • the characteristics of the neighboring reconstructed luma samples and neighboring reconstructed chroma samples of the current block are used to determine the linear- model parameters a, k, and b for LM-mode chroma intra prediction.
  • the derived co-located luma pixels of the current luma block can be derived from the current reconstructed luma pixels of the current luma block at pixel locations co-located with the chroma pixels of the current chroma block.
  • L corresponds to the sum of reconstructed luma samples in the neighboring area of the current block
  • C corresponds to the sum of reconstructed chroma samples in the neighboring area of the current block
  • LL corresponds to the sum of squared reconstructed luma samples in the neighboring area of the current block
  • LC corresponds to the sum of cross-product of reconstructed luma samples and reconstructed chroma samples in the neighboring area of the current block.
  • L, C, LL, and LC are right shifted by k3 bits to take into account of the internal bit depth with which the chroma signal is processed during video coding process (i.e., bitDepthC) and the block size (i.e., nS).
  • linear-model parameters a, b and k are derived as follows.
  • Parameter al as defined in equation (10) corresponds to a co variance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block.
  • Variables L, C, LL, and LC as shown in equations (5) through 8 are right shifted by k3 bits, where k3 is related to the block size (i.e., nS).
  • k3 is related to the block size (i.e., nS).
  • X uniformly distributed random variable
  • Y second uniformly distributed random variable
  • al has a form similar to a covariance value for uniformly distributed random variables X and Y.
  • LC is left shifted by k2 bits in order to match the scaling of L*C.
  • al as defined in equation (10) has a covariance-like form associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block.
  • the covariance-like value may be scaled to a desired range and the scaling may be performed by left or right shifting the covariance-like value. For example, the scaling of the covariance-like value used in HM-5.0 is shown in equation (13).
  • parameter a2 as defined in equation (11) corresponds to a variance-like value associated with the neighboring reconstructed luma pixels of the current luma block.
  • the variance-like value may be scaled to a desired range and the scaling may be performed by left or right shifting the variance-like value.
  • the scaling of the variance-like value used in HM-5.0 is shown in equation (14).
  • a division factor, ImDiv is then determined by dividing with rounding a first data range by a2s.
  • the division with rounding by a2s is implemented using a look-up table as shown in Table 1.
  • the first data range corresponds to 2 A 15 (i.e., 215).
  • ImDiv (2 A 15+a2s/2)/a2s.
  • An intermediate parameter, a3 is then determined according to als*lmDiv, where the product als*lmDiv is divided with rounding by 2kl and the result is clipped between -215 and 215-1 if a2s > 1. If a2s is less than 1, a3 is set to 0.
  • HM-5.0 derivation of a3 is shown in equation (15), where the division with rounding by 2kl is implemented by right shifting. The right shifting in equation (15) is performed to reverse the shifting operations that are applied to al and a2 in equation (13) and equation (14).
  • Cliplc(w) Clip3(0, (1 « BitDepth c ) ⁇ 1 , w)
  • a method and apparatus for chroma intra prediction based on reconstructed luma pixels and chroma pixels are disclosed.
  • the chroma intra prediction for chroma pixels of the current chroma block is based on a linear model of derived co-located current luma pixels of the current luma block scaled by a scaling factor.
  • the scaling factor comprises a product term of a division factor and a scaled covariance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block.
  • the division factor is related to a first data range divided with rounding by a scaled variance-like value associated with the neighboring reconstructed luma pixels of the current luma block.
  • the scaled covariance-like value, the first data range, or both of the scaled covariance-like value and the first data range are dependent on the internal bit depth with which the chroma signal is processed during video coding process.
  • the first data range corresponds to 2 A ( BitDepthC + n), where n is an integer from -2 to +4.
  • the division factor can be derived using a look-up table with the scaled variance-like value as the table input. Some table entries can be omitted depending on the selected first data range. Besides, the division factor can also be calculated by dividing with rounding by the scaled variance-like value during coding process.
  • Fig. 1A illustrates an example of derived luma samples derived based on neighboring reconstructed luma pixels and current reconstructed luma samples for chroma intra prediction according to HM-5.0.
  • Fig. IB illustrates an example of neighboring reconstructed chroma samples used for chroma intra prediction of current chroma samples.
  • FIG. 2 illustrates an example of a flowchart for chroma intra prediction incorporating an embodiment of the present invention.
  • the existing LM mode derivation is quite complicated. Some operations require higher arithmetic accuracy, such as the als*lmDiv operation. Therefore, in one embodiment of the present invention, the complexity of the als*lmDiv operation is reduced by incorporating the internal bit depth, with which the chroma signal is processed during video coding process, in the als derivation. As an example, the shift operation for als incorporating an embodiment of the present invention becomes:
  • ImDiv is set to (2 A (BitDepthC-2)+a2s/2)/a2s. Therefore, ImDiv is less than (2 A (BitDepthC-l)) when a2s is equal to 1.
  • Table 4 illustrates the values of
  • ImDiv for BitDepthC 8.
  • variable k is defined as follows:
  • variable a3 and k are defined as follows:
  • variable a3 and k are defined as follows:
  • ImDiv is set to (2 A (BitDepthC+4)+a2s/2)/a2s and a3 in equation (15) is set to zero for a2s ⁇ 32.
  • ImDiv ⁇ 2 A ( BitDepthC) for a2s > 32.
  • Table 11 illustrates the values of
  • ImDiv is set to 0 in this case.
  • Neighboring reconstructed luma pixels and current reconstructed luma pixels of a current luma block are received from a media or a processor as shown in step 210.
  • the reconstructed luma pixels may be retrieved from a media such as a computer memory of buffer (RAM or DRAM).
  • the reconstructed luma pixels may also be received from a processor such as a central processing unit or a digital signal processor that reconstructs the luma pixels from residual signals.
  • the residual signals are generated by the encoder.
  • the residual signals may be derived from the received bitstream.
  • the neighboring reconstructed chroma pixels of a current chroma block are received from a media or a processor as shown in step 220, wherein the current chroma block is co-located with the current luma block.
  • the reconstructed luma samples and chroma samples may be retrieved from the same media (such as the same DRAM device) or separate media (such as separate DRAM devices).
  • reconstructed luma samples and chroma samples may be received from the same processor or different processors (such as a processor for luma samples and another processor for chroma samples).
  • the chroma intra prediction for chroma pixels of the current chroma block is then derived based on derived co-located current luma pixels of the current luma block scaled by a scaling factor as shown in step 230.
  • the scaling factor comprises a product term of a division factor and a scaled covariance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block.
  • the division factor is related to a first data range divided with rounding by a scaled variance-like value associated with the neighboring reconstructed luma pixels of the current luma block.
  • at least one of the scaled covariance-like value and the first data range is dependent on the internal bit depth with which the chroma signal is processed during video coding process.
  • the chroma intra prediction is then provided for encoding or decoding of the chroma pixels of the current chroma block as shown in step 240.
  • Embodiment of the present invention as described above may be implemented in various hardware, software codes, or a combination of both.
  • an embodiment of the present invention can be a circuit integrated into a video compression chip or program code integrated into video compression software to perform the processing described herein.
  • An embodiment of the present invention may also be program code to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
  • DSP Digital Signal Processor
  • the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine- readable software code or firmware code that defines the particular methods embodied by the invention.
  • the software code or firmware code may be developed in different programming languages and different formats or styles.
  • the software code may also be compiled for different target platforms.
  • different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.

Abstract

A method and apparatus for chroma intra prediction based on reconstructed luma pixels and chroma pixels are disclosed. The chroma intra prediction is based on a linear model of derived co-located current luma pixels of the current luma block scaled by a scaling factor. The scaling factor comprises a product term of a division factor and a scaled covariance-like value associated with neighboring reconstructed luma and chroma pixels of a current block. The division factor is related to a first data range divided with rounding by a scaled variance-like value associated with the neighboring reconstructed luma pixels of the current block. The scaled covariance-like value, the first data range, or both of the scaled covariance-like value and the first data range are dependent on the internal bit depth, with which the chroma signal is processed during video coding process, according to an embodiment of the present invention.

Description

METHOD AND APPARATUS OF SIMPLIFIED LUMA-BASED CHROMA
INTRA PREDICTION
BACKGROUND OF THE INVENTION
Cross Reference To Related Applications [0001] The present invention claims priority to PCT Patent Application, Serial No. PCT/CN2012/074118, filed April 16, 2012, entitled "Improvements of Luma-based Chroma Intra Prediction". The PCT Patent Application is hereby incorporated by reference in its entirety.
Field of the Invention [0002] The present invention relates to video coding. In particular, the present invention relates to coding techniques associated with simplified chroma intra prediction based on reconstructed luma and chroma pixels.
Description of the Related Art
[0003] Motion compensated inter- frame coding has been widely adopted in various coding standards, such as MPEG-1/2/4 and H.261/H.263/H.264/AVC. While motion-compensated inter-frame coding can effectively reduce bitrate for compressed video, intra coding is required to compress the regions with high motion or scene changes. Besides, intra coding is also used to process an initial picture or to periodically insert I-pictures or I-blocks for random access or for alleviation of error propagation. Intra prediction exploits the spatial correlation within a picture or within a picture region. In practice, a picture or a picture region is divided into blocks and the intra prediction is performed on a block basis. Intra prediction for a current block can rely on pixels in neighboring blocks that have been processed. For example, if blocks in a picture or picture region are processed row by row first from left to right and then from top to bottom, neighboring blocks on the top and neighboring blocks on the left of the current block can be used to form intra prediction for pixels in the current block. While any pixels in the processed neighboring blocks can be used for intra predictor of pixels in the current block, very often only pixels of the neighboring blocks that are adjacent to the current block boundaries on the top and on the left are used.
[0004] The intra predictor is usually designed to exploit spatial features in the picture such as smooth area (DC mode), vertical line or edge, horizontal line or edge and diagonal line or edge. Furthermore, correlation often exists between the luminance (luma) and chrominance (chroma) components. Therefore, reconstructed luma pixels can be used to derive the intra chroma prediction. In recent development of High Efficiency Video Coding (HEVC), a chroma intra prediction method based on co-located reconstructed luma blocks has been disclosed. The type of chroma intra prediction is termed as LM prediction or LM mode. The main concept is to use the reconstructed luma pixels to generate the predictors of corresponding chroma pixels. Fig. 1A and Fig. IB illustrate the prediction procedure. First, the neighboring reconstructed pixels of a co-located luma block in Fig. 1 A and the neighboring reconstructed pixels of a chroma block in Fig. IB are used to derive the correlation parameters between the blocks. Then, the predicted pixels of the chroma block are generated using the parameters and the
reconstructed pixels of the luma block. In the parameters derivation, the first above
reconstructed pixel row and the second left reconstructed pixel column of the current luma block are used. The specific row and column of the luma block are used in order to match the 4:2:0 sampling format of the chroma components. The following illustration is based on 4:2:0 sampling format. LM-mode chroma intra prediction for other sampling formats can be derived similarly.
[0005] In the Test Model Version 5.0 (HM-5.0), the LM mode is applied to predict chroma samples based on a linear model using reconstructed luma samples of the co-located prediction unit (PU). The parameters of the linear model consist of slope (a»k) and y-intercept (b), where "»" corresponds to the right shift operation. The parameters are derived based on the neighboring luma and chroma samples according to a least mean square criterion. The prediction sample, predSamples[x,y] for the chroma sample to be coded in the LM mode is derived as follows, where x,y = 0...nS-l and nS corresponds to the block size.
[0006] First, variable k3 and the sample array pY' are derived as:
k3 = Max( 0, BitDepthc + Log2( nS ) - 14 ) , (1) where BitDepthc denotes the internal chroma bit depth (i.e., the bit depth with which the chroma signal is processed during video coding process), and
pY'[x,-l] = ( Pw[2x-1 -1] + 2*Pm[2x -1] +
PLM[2X+1 -1] + 2 ) » 2, (2) Pr i-hy] = ( Pw[-h2y] + Pw[-h2y+i] ) » h (3) ργ'[χ,γ] = ( recSamplesL,[2x,2y] + recSamplesL[2x, 2y+l] ) » 1, (4) where x = 0...nS-l, Pmix ] denotes the neighboring reconstructed luma samples, and recSamplesL[x,y] denotes the current reconstructed luma samples of the co-located luma block. The sample array ργ are derived from reconstructed luma samples. Accordingly, pY' is also called derived co-located luma sample in this disclosure. In equations (2) through (4), ργ' [x,y] is only evaluated at positions co-located with the chroma samples.
[0007] In HM-5.0, the characteristics of the neighboring reconstructed luma samples and neighboring reconstructed chroma samples of the current block are used to determine the linear- model parameters a, k, and b for LM-mode chroma intra prediction. The derived co-located luma pixels of the current luma block can be derived from the current reconstructed luma pixels of the current luma block at pixel locations co-located with the chroma pixels of the current chroma block.
[0008] To exemplify the linear model relating the chroma intra prediction with the derived co-located luma pixels, a set of variable, including L, C, LL, LC and k2, are defined. Variables L, C, LL, LC and k2 are derived as follows.
Figure imgf000005_0001
Figure imgf000005_0002
nS-l nS-l
LC ∑ PY l-i, y] * p[-i, y] + ∑ Ργ '[χ, - ΐ] * P[x, -i] » k3 (8) y=0 x=0
kl = Log2( (2*nS) » k3 ) (9)
[0009] As shown in equations (5) through (8), L corresponds to the sum of reconstructed luma samples in the neighboring area of the current block, C corresponds to the sum of reconstructed chroma samples in the neighboring area of the current block, LL corresponds to the sum of squared reconstructed luma samples in the neighboring area of the current block, LC corresponds to the sum of cross-product of reconstructed luma samples and reconstructed chroma samples in the neighboring area of the current block. Furthermore, L, C, LL, and LC are right shifted by k3 bits to take into account of the internal bit depth with which the chroma signal is processed during video coding process (i.e., bitDepthC) and the block size (i.e., nS).
[0010] The linear-model parameters a, b and k are derived as follows.
al ( LC « kl ) - L*C, (10) al ( LL « kl ) - L*L, (11) kl Max(0, Log2( abs(a2)) -5) - Max(0, Log2(abs(ai)) - 14) + 2, (12) als al » Max(0, Log2( abs(ai)) - 14), (13) als = abs( al » Max(0, Log2(abs(a2)) - 5)), (14) a3 = als <1?0 : Clip3(-215,215-l,(oi^/mZ) v + (1 « (kl ~ 1))) » kl),(15) a = a3 » Max(0, Log2( abs(aJ)) - 6), (16) k = 13 - Max( 0, Log2( abs(aJ)) - 6), and (17) b = (L - ((a*C) » kl) + (1 « (kl - 1))) » kl, (18) where ImDiv is specified in Table 1 for all als values.
[0011] Parameter al as defined in equation (10) corresponds to a co variance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block. The covariance σ(Χ,Υ) associated with random variables X and Y are defined as σ(Χ,Υ) =E[XY] - E[X]E[Y], where E[.] is the expected value of the underlying random variable. For a random variable with a uniform distribution, the expected value is equivalent to the average value. Variables L, C, LL, and LC as shown in equations (5) through 8 are right shifted by k3 bits, where k3 is related to the block size (i.e., nS). In other words, if the neighboring reconstructed luma pixels of the current luma block are considered as a first uniformly distributed random variable (i.e., X) and the neighboring reconstructed chroma pixels of the current chroma block are considered as a second uniformly distributed random variable (i.e., Y), al has a form similar to a covariance value for uniformly distributed random variables X and Y. In equation (10), LC is left shifted by k2 bits in order to match the scaling of L*C. Accordingly, al as defined in equation (10) has a covariance-like form associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block. The covariance-like value may be scaled to a desired range and the scaling may be performed by left or right shifting the covariance-like value. For example, the scaling of the covariance-like value used in HM-5.0 is shown in equation (13).
[0012] Similarly, parameter a2 as defined in equation (11) corresponds to a variance-like value associated with the neighboring reconstructed luma pixels of the current luma block. The variance-like value may be scaled to a desired range and the scaling may be performed by left or right shifting the variance-like value. For example, the scaling of the variance-like value used in HM-5.0 is shown in equation (14). A division factor, ImDiv is then determined by dividing with rounding a first data range by a2s. In HM-5.0, the division with rounding by a2s is implemented using a look-up table as shown in Table 1. In HM-5.0, the first data range corresponds to 2A15 (i.e., 215). Accordingly, ImDiv =(2A15+a2s/2)/a2s. An intermediate parameter, a3 is then determined according to als*lmDiv, where the product als*lmDiv is divided with rounding by 2kl and the result is clipped between -215 and 215-1 if a2s > 1. If a2s is less than 1, a3 is set to 0. In HM-5.0, derivation of a3 is shown in equation (15), where the division with rounding by 2kl is implemented by right shifting. The right shifting in equation (15) is performed to reverse the shifting operations that are applied to al and a2 in equation (13) and equation (14).
[0013] Parameters a and k for the linear model are then determined based on a3 as shown in equations (16) and (17) respectively. The y-intercept, b is determined according to equation (18). Finally, the value of the prediction samples predSamples[x,y] is derived as:
predSamples[x,y] = Clvplc((( pY'[x,y] * a) » k) + b), (19) where x, y = 0..nS~l and
Cliplc(w) = Clip3(0, (1 « BitDepthc) ~ 1 , w)
clip3(t, u, v) = ((v < t) ? t : ((v > w) ? u : v))
Table 1
Figure imgf000007_0001
[0014] As shown above, the derivation of the prediction sample, predSamples[x,y] is very computationally intensive. Not only it involves a large number of computations, it also requires buffer to store the table. Furthermore, some operations may require higher precision. For example, in equation (15), als is a signed 15-bit integer, therefore, als*lmDiv requires a 15-bit multiplier. This large multiplier introduces higher computational complexity. Therefore, it is desirable to simplify the derivation of the prediction sample. BRIEF SUMMARY OF THE INVENTION
[0015] A method and apparatus for chroma intra prediction based on reconstructed luma pixels and chroma pixels are disclosed. The chroma intra prediction for chroma pixels of the current chroma block is based on a linear model of derived co-located current luma pixels of the current luma block scaled by a scaling factor. The scaling factor comprises a product term of a division factor and a scaled covariance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block. The division factor is related to a first data range divided with rounding by a scaled variance-like value associated with the neighboring reconstructed luma pixels of the current luma block. In a system incorporating an embodiment of the present invention, the scaled covariance-like value, the first data range, or both of the scaled covariance-like value and the first data range are dependent on the internal bit depth with which the chroma signal is processed during video coding process.
[0016] In one embodiment, the scaled covariance-like value, als is derived according to als = al » Max(0, Log2( abs(al) ) - (BitDepthC -2)), where al represents the covariance-like value, and BitDepthC represents the internal bit depth with which the chroma signal is process in video codec. In another embodiment of the present invention, the first data range corresponds to 2A( BitDepthC + n), where n is an integer from -2 to +4. The division factor can be derived using a look-up table with the scaled variance-like value as the table input. Some table entries can be omitted depending on the selected first data range. Besides, the division factor can also be calculated by dividing with rounding by the scaled variance-like value during coding process.
BRIEF DESCRIPTION OF DRAWINGS
[0017] Fig. 1A illustrates an example of derived luma samples derived based on neighboring reconstructed luma pixels and current reconstructed luma samples for chroma intra prediction according to HM-5.0.
[0018] Fig. IB illustrates an example of neighboring reconstructed chroma samples used for chroma intra prediction of current chroma samples.
[0019] Fig. 2 illustrates an example of a flowchart for chroma intra prediction incorporating an embodiment of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
[0020] As mentioned before, the existing LM mode derivation is quite complicated. Some operations require higher arithmetic accuracy, such as the als*lmDiv operation. Therefore, in one embodiment of the present invention, the complexity of the als*lmDiv operation is reduced by incorporating the internal bit depth, with which the chroma signal is processed during video coding process, in the als derivation. As an example, the shift operation for als incorporating an embodiment of the present invention becomes:
als = al » Max(0, Log2( abs(ai) ) - (BitDepthc-2)). (20) [0021] In equation (20), the constant value (i.e. 14) of equation (13) is replaced by
(BitDepthC -2). Accordingly, the number of bits of als is limited to the bit depth of the chroma signal (i.e., BitDepthC) minus one and plus one sign bit. Furthermore, lmDiv is changed from (2Λ 15+a2s/2)/a2s to (2A(BitDepthC- 1 )+a2s/2)/a2s. Therefore, lmDiv will be less than
(2ABitDepthC) when a2s is equal to 1. As mentioned before, the derivation of lmDiv from a2s, i.e., lmDiv=(2A(BitDepthC-l)+a2s/2)/a2s, can be implemented as a look-up table. Table 2 illustrates the values of lmDiv for BitDepthC = 8. Table 3 illustrates the values of lmDiv for BitDepthC = 10. Table 2
Figure imgf000009_0001
Table 3
Figure imgf000009_0002
[0022] Along with the new als as shown in equation (20) and lmDiv (i.e.,
(2A(BitDepthC-l)+a2s/2)/a2s) mentioned above, other variables that require modification are as follows:
kl = Max(0,Log2(abs(a2))-5)-Max(0,Log2(abs(ai))-(fi iZ)ep^c-2)), (21) k = BitDepthc- 1 - Max(0, Log2(abs(aJ)) - 6). (22)
[0023] In another embodiment, ImDiv is set to (2A(BitDepthC-2)+a2s/2)/a2s. Therefore, ImDiv is less than (2A(BitDepthC-l)) when a2s is equal to 1. Table 4 illustrates the values of
ImDiv for BitDepthC = 8. Table 5 illustrates the values of ImDiv for BitDepthC = 10.
Table 4
Figure imgf000010_0001
Table 5
Figure imgf000010_0002
[0024] For the ImDiv (i.e., (2A(BitDepthC-2)+a2s/2)/a2s) mentioned above, variable k is defined as follows:
k = BitDepthc- 2 - Max(0, Log2(abs(aJ)) - 6). (23) [0025] In yet another embodiment, ImDiv is set to (2A(BitDepthC+2)+a2s/2)/a2s. In this case, ImDiv will be less than 2ABitDepthC for a2s >= 7. Accordingly, Table 6 illustrates the values of ImDiv for BitDepthC = 8. Table 7 illustrates the values of ImDiv for BitDepthC = 10. In Tables 6 and 7, there are no entries for a2s < 7 since a2s is impossible to be any value from 1 to 6. Table 6
Figure imgf000011_0001
Table 7
Figure imgf000011_0002
[0026] For the ImDiv (i.e., (2A(BitDepthC+2)+a2s/2)/a2s) mentioned above, variable a3 and k are defined as follows:
a3 = als <7 ? 0: Clip3(-215, 215-l ,als*/mZ)/v+(l« (kl-l ) »kl), (24) k = BitDepthc + 2 - Max(0, Log2(abs(aJ)) - 6). (25) [0027] In yet another embodiment, ImDiv is changed to (2A(BitDepthC+l)+a2s/2)/a2s and this will ensure ImDiv < 2A(BitDepthC-l) for a2s >= 7. Table 8 illustrates the values of ImDiv for BitDepthC = 8. Table 9 illustrates the values of ImDiv for BitDepthC = 10. In Tables 8 and 9, there are no entries for a2s < 7 since a2s is impossible to be any value from 1 to 6.
Table 8 a2s 7 8 9 10 1 1 12 13 14 15 16 17 18 19 20 21
ImDiv 73 64 57 51 47 43 39 37 34 32 30 28 27 26 24 a2s 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36
ImDiv 23 22 21 20 20 19 18 18 17 17 16 16 15 15 14 a2s 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51
ImDiv 14 13 13 13 12 12 12 12 1 1 1 1 1 1 1 1 10 10 10 a2s 52 53 54 55 56 57 58 59 60 61 62 63
ImDiv 10 10 9 9 9 9 9 9 9 8 8 8
Table 9
Figure imgf000012_0001
[0028] For the ImDiv (i.e., (2A(BitDepthC+l)+a2s/2)/a2s) mentioned above, variable a3 and k is defined as follows:
a3 = als <7 ? 0: Clip3(-215, 215-l,als*/mZ)/v+(l« (kl-l ) »kl), (26) k = BitDepthc + 1 - Max(0, Log2(abs(aJ)) - 6). (27)
[0029] In another embodiment, ImDiv is set to (2A(BitDepthC+4)+a2s/2)/a2s and a3 in equation (15) is set to zero for a2s < 32. In this case, ImDiv < 2A( BitDepthC) for a2s >= 32.
Table 10 illustrates the values of ImDiv for BitDepthC = 8. Table 11 illustrates the values of
ImDiv for BitDepthC = 10. There is no need for table entries corresponding to a2s < 32 since
ImDiv is set to 0 in this case.
Table 10
Figure imgf000012_0002
Table 11 a2s 32 33 34 35 36 37 38 39 40 41 42
ImDiv 512 496 482 468 455 443 431 420 410 400 390 a2s 43 44 45 46 47 48 49 50 51 52 53
ImDiv 381 372 364 356 349 341 334 328 321 315 309 a2s 54 55 56 57 58 59 60 61 62 63
ImDiv 303 298 293 287 282 278 273 269 264 260
[0030] For the ImDiv (i.e., (2A(BitDepthC+4)+a2s/2)/a2s) and a3 mentioned above, other variables that require modification are as follows:
a3 = als <32 ? 0: Clip3(-215, 215-l ,als*/mZ)/v+(l« (kl-l ) »kl), (28) k = BitDepthc+ 4 - Max(0, Log2(abs(aJ)) - 6). (29)
[0031] In another embodiment, ImDiv is set to (2A(BitDepthC+3)+a2s/2)/a2s and a3 in equation (15) is set to zero for a2s<32. In this case, ImDiv will be less than 2A( BitDepthC-1) for a2s>=32. Table 12 illustrates the values of ImDiv for BitDepthC = 8. Table 13 illustrates the values of ImDiv for BitDepthC = 10. . There is no need for table entries corresponding to a2s < 32 since ImDiv is set to 0 in this case.
[0032]
Table 12
Figure imgf000013_0001
Table 13
Figure imgf000013_0002
[0033] For the ImDiv (i.e., (2A(BitDepthC+3)+a2s/2)/a2s) and a3 mentioned above, other variables that require modification are as follows: a3 = als < 32 ? 0 : Clip3(-215, 2\5-\ , als*lmDiv+(\« (kl-\))» kl), (30) k = BitDepthc+ 3 - Max(0, Log2(abs(aJ)) - 6). (31) [0034] The LM-mode chroma intra prediction method described above can be used in a video encoder as well as a video decoder. Fig. 2 illustrates an exemplary flowchart of an encoder or a decoder incorporating an embodiment of the present invention. Neighboring reconstructed luma pixels and current reconstructed luma pixels of a current luma block are received from a media or a processor as shown in step 210. The reconstructed luma pixels may be retrieved from a media such as a computer memory of buffer (RAM or DRAM). The reconstructed luma pixels may also be received from a processor such as a central processing unit or a digital signal processor that reconstructs the luma pixels from residual signals. In a video encoder, the residual signals are generated by the encoder. In a video decoder, the residual signals may be derived from the received bitstream. The neighboring reconstructed chroma pixels of a current chroma block are received from a media or a processor as shown in step 220, wherein the current chroma block is co-located with the current luma block. The reconstructed luma samples and chroma samples may be retrieved from the same media (such as the same DRAM device) or separate media (such as separate DRAM devices). The
reconstructed luma samples and chroma samples may be received from the same processor or different processors (such as a processor for luma samples and another processor for chroma samples). The chroma intra prediction for chroma pixels of the current chroma block is then derived based on derived co-located current luma pixels of the current luma block scaled by a scaling factor as shown in step 230. The scaling factor comprises a product term of a division factor and a scaled covariance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block. The division factor is related to a first data range divided with rounding by a scaled variance-like value associated with the neighboring reconstructed luma pixels of the current luma block. According the present invention, at least one of the scaled covariance-like value and the first data range is dependent on the internal bit depth with which the chroma signal is processed during video coding process. The chroma intra prediction is then provided for encoding or decoding of the chroma pixels of the current chroma block as shown in step 240.
[0035] The flowcharts shown above are intended to illustrate examples of a luma-based chroma intra prediction method for a video encoder and a decoder incorporating embodiments of the present invention. A person skilled in the art may modify each step, re-arranges the steps, split a step, or combine the steps to practice the present invention without departing from the spirit of the present invention.
[0036] The above description is presented to enable a person of ordinary skill in the art to practice the present invention as provided in the context of a particular application and its requirement. Various modifications to the described embodiments will be apparent to those with skill in the art, and the general principles defined herein may be applied to other embodiments. Therefore, the present invention is not intended to be limited to the particular embodiments shown and described, but is to be accorded the widest scope consistent with the principles and novel features herein disclosed. In the above detailed description, various specific details are illustrated in order to provide a thorough understanding of the present invention. Nevertheless, it will be understood by those skilled in the art that the present invention may be practiced.
[0037] Embodiment of the present invention as described above may be implemented in various hardware, software codes, or a combination of both. For example, an embodiment of the present invention can be a circuit integrated into a video compression chip or program code integrated into video compression software to perform the processing described herein. An embodiment of the present invention may also be program code to be executed on a Digital Signal Processor (DSP) to perform the processing described herein. The invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine- readable software code or firmware code that defines the particular methods embodied by the invention. The software code or firmware code may be developed in different programming languages and different formats or styles. The software code may also be compiled for different target platforms. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
[0038] The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. The scope of the invention is therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.

Claims

1. A method of chroma intra prediction based on reconstructed luma pixels and chroma pixels, the method comprising:
receiving neighboring reconstructed luma pixels and current reconstructed luma pixels of a current luma block from a media or a processor;
receiving neighboring reconstructed chroma pixels of a current chroma block from the media or the processor, wherein the current chroma block is co-located with the current luma block;
deriving chroma intra prediction for chroma pixels of the current chroma block based on derived co-located current luma pixels of the current luma block scaled by a scaling factor, wherein the scaling factor comprises a product term of a division factor and a scaled covariance- like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block, wherein the division factor is related to a first data range divided with rounding by a scaled variance-like value associated with the neighboring reconstructed luma pixels of the current luma block, and wherein at least one of the scaled covariance-like value and the first data range is dependent on internal bit depth with which chroma signal is processed during video coding process; and
providing the chroma intra prediction for encoding or decoding of the chroma pixels of the current chroma block.
2. The method of Claim 1, wherein the neighboring reconstructed luma pixels comprise a first group of the reconstructed luma pixels above a top boundary of the current luma block and a second group of the reconstructed luma pixels to a left boundary of the current luma block.
3. The method of Claim 1, wherein said derived co-located current luma pixels of the current luma block are derived from the current reconstructed luma pixels of the current luma block at pixel locations co-located with the chroma pixels of the current chroma block.
4. The method of Claim 1, wherein said scaled covariance-like value is derived based on a covariance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block, wherein said scaled covariance-like value corresponds to right-shifting said covariance-like value by a first number of bits related to said covariance-like value and the internal bit depth with which chroma signal is processed during video coding process.
5. The method of Claim 4, wherein als = al » Max(0, Log2( abs(ai) ) - (BitDepthc -2)), wherein als represents said scaled co variance-like value, al represents said co variance-like value, and BitDepthc represents the internal bit depth with which chroma signal is processed during video coding process.
6. The method of Claim 5, wherein the first data range corresponds to 2A( BitDepthc -2) or 2Λ( BitDepthc -1).
7. The method of Claim 1, wherein said scaled variance-like value is derived based on a variance-like value associated with the neighboring reconstructed luma pixels of the current luma block, wherein said scaled variance-like value corresponds to right-shifting said variancelike value by a second number of bits related to said variance-like value.
8. The method of Claim 7, wherein als = abs( al » Max(0, Log2(abs(a2)) - 5)), wherein als represents said scaled variance-like value and al represents said variance-like value.
9. The method of Claim 8, wherein the division factor corresponds to (2A( BitDepthc - 2)+als/2)/als, (2Λ( BitDepthc -\)+alsl2)l als, (2Λ( BitDepthc +\)+als/2)/als, (2Λ( BitDepthc +2)+als/2)/als, (2Λ( BitDepthc +3)+als/2)/als, or (2Λ( BitDepthc +4)+ als 12)1 als.
10. The method of Claim 9, wherein said derived co-located current luma pixels of the current luma block scaled by the scaling factor are right shifted by a shift factor related to the product term during said deriving chroma intra prediction for chroma pixels of the current chroma block, wherein the shift factor corresponds to (BitDepthc-2-Max(0, Log2(abs(<z3))-6)), (BitDepthc -1-Max(0, Log2(abs(<z?))-6)), (BitDepthc+\-Max(0, Log2(abs(<z?))-6)),
(BitDepthc+2 -Max(0, Log2(abs(<z?))-6)), (BitDepthc+3-Max(0, Log2(abs(<z?))-6)), or (BitDepthc+4 -Max(0, Log2(abs(<z3))-6)) for a respective division factor, wherein a3 represents a processed result corresponding to the product term being right shifted with rounding and clipped to (-215,215-1) if said scaled variance-like value is equal to or greater than one, and wherein a3 is 0 otherwise.
11. The method of Claim 9, wherein the division factor is implemented by a look-up table for a selected internal bit depth of chroma signal with the scaled variance-like value as a table input.
12. The method of Claim 11, wherein the look-up table omits table entries for the scaled variance-like value from 1 to 6 in case of the division factor corresponding to either
(2A( BitDepthc +X)+a2sl2)la2s, (2A( BitDepthc +2)+a2s/2)/a2s and for the scaled variance-like value from 1 to 31 in case of the division factor corresponding to either (2A( BitDepthc
+3)+a2s/2)/a2s, or (2Λ( BitDepthc +4)+a2s/2)/a2s .
13. The method of Claim 12, wherein a3 is set to 0 if a2s is less than 7 in case of the division factor corresponding to either (2A( BitDepthc +l)+ci2s/2)/a2s, (2A( BitDepthc
+2)+a2sl2)la2s and wherein a3 is set to 0 if a2s is less than 32 in case of the division factor corresponding to either (2A( BitDepthc +3)+a2s/2)/a2s, or (2A( BitDepthc +4)+a2s/2)/a2s.
14. The method of Claim 12, wherein the division factor is set to a first value of the division factor corresponding to a2s equal to 7 if a2s is less than 7 or a2s is less than 7 and larger than 0, in case of the division factor corresponding to either (2A( BitDepthc
+X)+a2sl2)la2s, (2A( BitDepthc +2)+a2s/2)/a2s and wherein a3 is set to a second value of the division factor corresponding to a2s equal to 32 if a2s is less than 32 or a2s is less than 32 and larger than 0, in case of the division factor corresponding to either (2A( BitDepthc
+3)+a2s/2)/a2s, or (2A( BitDepthc +4)+a2s/2)/a2s.
15. An apparatus of chroma intra prediction based on reconstructed luma pixels and chroma pixels, the apparatus comprising:
means for receiving neighboring reconstructed luma pixels and current reconstructed luma pixels of a current luma block from a media or a processor;
means for receiving neighboring reconstructed chroma pixels of a current chroma block from the media or the processor, wherein the current chroma block is co-located with the current luma block;
means for deriving chroma intra prediction for chroma pixels of the current chroma block based on derived co-located current luma pixels of the current luma block scaled by a scaling factor, wherein the scaling factor comprises a product term of a division factor and a scaled covariance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block, wherein the division factor is related to a first data range divided with rounding by a scaled variance-like value associated with the neighboring reconstructed luma pixels of the current luma block, and wherein at least one of the scaled covariance-like value and the first data range is dependent on bit depth of chroma signal; and
means for providing the chroma intra prediction for encoding or decoding of the chroma pixels of the current chroma block.
16. The apparatus of Claim 15, wherein the neighboring reconstructed luma pixels comprise a first group of the reconstructed luma pixels above a top boundary of the current luma block and a second group of the reconstructed luma pixels to a left boundary of the current luma block.
17. The apparatus of Claim 15, wherein said derived co-located current luma pixels of the current luma block are derived from the current reconstructed luma pixels of the current luma block at pixel locations co-located with the chroma pixels of the current chroma block.
18. The apparatus of Claim 15, wherein said scaled covariance-like value is derived based on a covariance-like value associated with the neighboring reconstructed luma pixels of the current luma block and the neighboring reconstructed chroma pixels of the current chroma block, wherein said scaled covariance-like value corresponds to right-shifting said covariance-like value by a first number of bits related to said covariance-like value and an internal bit depth with which chroma signal is processed during video coding process.
19. The apparatus of Claim 18, wherein als = al » Max(0, Log2( abs(ai) ) - (BitDepthc - 2)), wherein als represents said scaled covariance-like value, al represents said covariance-like value, and BitDepthc represents the internal bit depth with which chroma signal is processed during video coding process.
20. The apparatus of Claim 19, wherein the first data range corresponds to 2A( BitDepthc - 2) or 2Λ( BitDepthc -1).
21. The apparatus of Claim 15, wherein said scaled variance-like value is derived based on a variance-like value associated with the neighboring reconstructed luma pixels of the current luma block, wherein said scaled variance-like value corresponds to right-shifting said variancelike value by a second number of bits related to said variance-like value.
PCT/CN2013/070912 2012-04-16 2013-01-24 Method and apparatus of simplified luma-based chroma intra prediction WO2013155886A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US14/380,709 US9794557B2 (en) 2012-04-16 2013-01-24 Method and apparatus of simplified luma-based chroma intra prediction
CN201380019470.6A CN104471940B (en) 2012-04-16 2013-01-24 Chroma intra prediction method and device
EP13778200.9A EP2839656B1 (en) 2012-04-16 2013-01-24 Method and apparatus of simplified luma-based chroma intra prediction
JP2015506080A JP5977439B2 (en) 2012-04-16 2013-01-24 Method and apparatus for chroma intra prediction based on simplified luma (LUMA)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CNPCT/CN2012/074118 2012-04-16
PCT/CN2012/074118 WO2013155662A1 (en) 2012-04-16 2012-04-16 Methods and apparatuses of simplification for intra chroma lm mode

Publications (1)

Publication Number Publication Date
WO2013155886A1 true WO2013155886A1 (en) 2013-10-24

Family

ID=49382785

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/CN2012/074118 WO2013155662A1 (en) 2012-04-16 2012-04-16 Methods and apparatuses of simplification for intra chroma lm mode
PCT/CN2013/070912 WO2013155886A1 (en) 2012-04-16 2013-01-24 Method and apparatus of simplified luma-based chroma intra prediction

Family Applications Before (1)

Application Number Title Priority Date Filing Date
PCT/CN2012/074118 WO2013155662A1 (en) 2012-04-16 2012-04-16 Methods and apparatuses of simplification for intra chroma lm mode

Country Status (4)

Country Link
US (1) US9794557B2 (en)
EP (1) EP2839656B1 (en)
JP (1) JP5977439B2 (en)
WO (2) WO2013155662A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7485798B2 (en) 2015-07-08 2024-05-16 インターデイジタル マディソン パテント ホールディングス ソシエテ パー アクシオン サンプリフィエ Enhanced chroma coding with cross-plane filtering

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2920964B1 (en) * 2013-03-26 2018-05-09 MediaTek Inc. Method of cross color intra prediction
WO2015008417A1 (en) * 2013-07-19 2015-01-22 日本電気株式会社 Video coding device, video decoding device, video coding method, video decoding method, and program
US9648332B2 (en) * 2013-10-28 2017-05-09 Qualcomm Incorporated Adaptive inter-color component residual prediction
WO2015139203A1 (en) * 2014-03-18 2015-09-24 Mediatek Singapore Pte. Ltd. Dlt signaling in 3d video coding
KR102291835B1 (en) * 2015-07-08 2021-08-23 인터디지털 매디슨 페턴트 홀딩스 에스에이에스 Enhanced chroma coding with cross-plane filtering
WO2017139937A1 (en) * 2016-02-18 2017-08-24 Mediatek Singapore Pte. Ltd. Advanced linear model prediction for chroma coding
US10652575B2 (en) * 2016-09-15 2020-05-12 Qualcomm Incorporated Linear model chroma intra prediction for video coding
EP3316578A1 (en) * 2016-10-25 2018-05-02 Thomson Licensing Method and apparatus for encoding and decoding a picture
WO2020035837A1 (en) * 2018-08-17 2020-02-20 Beijing Bytedance Network Technology Co., Ltd. Simplified cross component prediction
TWI824006B (en) 2018-09-12 2023-12-01 大陸商北京字節跳動網絡技術有限公司 Downsampling in cross-component linear modeling
EP3861736A4 (en) 2018-11-06 2022-03-30 Beijing Bytedance Network Technology Co., Ltd. Position dependent intra prediction
US11197005B2 (en) 2018-11-08 2021-12-07 Qualcomm Incorporated Cross-component prediction for video coding
WO2020108591A1 (en) 2018-12-01 2020-06-04 Beijing Bytedance Network Technology Co., Ltd. Parameter derivation for intra prediction
WO2020114445A1 (en) 2018-12-07 2020-06-11 Beijing Bytedance Network Technology Co., Ltd. Context-based intra prediction
CN111327903B (en) * 2018-12-13 2023-05-16 华为技术有限公司 Method and device for predicting chroma block
AU2020226565C1 (en) * 2019-02-22 2024-01-11 Beijing Bytedance Network Technology Co., Ltd. Neighbouring sample selection for intra prediction
CN115967810A (en) 2019-02-24 2023-04-14 抖音视界有限公司 Method, apparatus and computer readable medium for encoding and decoding video data
WO2020192642A1 (en) 2019-03-24 2020-10-01 Beijing Bytedance Network Technology Co., Ltd. Conditions in parameter derivation for intra prediction
CN113824961B (en) * 2021-11-22 2022-03-29 康达洲际医疗器械有限公司 Interframe image coding method and system applicable to VVC (variable value code) coding standard

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080063047A1 (en) * 2006-09-08 2008-03-13 Jiun-In Guo Intra prediction circuit device applied to the H.264 video coding standard

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6766021B2 (en) * 2001-03-13 2004-07-20 Adaptive Digital Technologies Echo canceller
US7075993B2 (en) * 2001-06-12 2006-07-11 Digital Interactive Streams, Inc. Correction system and method for enhancing digital video
US20060013320A1 (en) * 2004-07-15 2006-01-19 Oguz Seyfullah H Methods and apparatus for spatial error concealment
CN100562116C (en) * 2007-12-12 2009-11-18 浙江万里学院 A kind of bit rate control method towards multi-view point video
KR101692829B1 (en) 2008-06-12 2017-01-05 톰슨 라이센싱 Methods and apparatus for video coding and decoding with reduced bit-depth update mode and reduced chroma sampling update mode
KR102268821B1 (en) * 2010-04-09 2021-06-23 엘지전자 주식회사 Method and apparatus for processing video data
CA2813191A1 (en) * 2010-10-01 2012-04-05 Samsung Electronics Co., Ltd. Image intra prediction method and apparatus
CN102209243B (en) * 2011-05-27 2012-10-24 山东大学 Depth map intra prediction method based on linear model

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080063047A1 (en) * 2006-09-08 2008-03-13 Jiun-In Guo Intra prediction circuit device applied to the H.264 video coding standard

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JULIANA HSU ET AL.: "Simplified Parameter Calculation for Chroma LM Mode", JCTVC-I0166, JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC TTCL/SC29/WG1 19TH MEETING, 7 May 2012 (2012-05-07), GENEVA, XP030111929, Retrieved from the Internet <URL:http://phenix.int-evry.fr/jct/doc_end_user/documents/9Geneva/wgll/JCTVC-I0166-vl.zip> *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7485798B2 (en) 2015-07-08 2024-05-16 インターデイジタル マディソン パテント ホールディングス ソシエテ パー アクシオン サンプリフィエ Enhanced chroma coding with cross-plane filtering

Also Published As

Publication number Publication date
US9794557B2 (en) 2017-10-17
JP2015517270A (en) 2015-06-18
US20150036745A1 (en) 2015-02-05
EP2839656A4 (en) 2016-01-06
EP2839656B1 (en) 2020-09-02
EP2839656A1 (en) 2015-02-25
JP5977439B2 (en) 2016-08-24
WO2013155662A1 (en) 2013-10-24

Similar Documents

Publication Publication Date Title
WO2013155886A1 (en) Method and apparatus of simplified luma-based chroma intra prediction
AU2012364505B2 (en) Method and apparatus of luma-based chroma intra prediction
US20170295366A1 (en) Method of Cross Color Intra Prediction
US8326064B2 (en) Image re-encoding method to decode image data which is orthogonally transformed per first block and encoded by a first encoding method
US11089332B2 (en) Method and apparatus of alternative transform for video coding
CN104471940B (en) Chroma intra prediction method and device
US11102474B2 (en) Devices and methods for intra prediction video coding based on a plurality of reference pixel values
CN109845263B (en) Apparatus and method for video encoding
EP3834416A1 (en) Methods and apparatuses of video processing with bi-direction predicition in video coding systems
WO2012044487A1 (en) Optimized filter selection for reference picture processing
EP2805501A1 (en) Simplification of lm mode
KR20110052203A (en) Method and apparatus for scalable video coding
JP2009177417A (en) Image encoding apparatus and method, image decoding apparatus and method, and program
WO2013155398A2 (en) Lm mode with uniform bit-width multipliers
KR20210099008A (en) Method and apparatus for deblocking an image
US10045032B2 (en) Efficient region of interest detection
WO2019196941A1 (en) Adaptive implicit transform setting
EP3641311A1 (en) Encoding and decoding methods and apparatus
WO2019206115A1 (en) Method and apparatus for restricted linear model parameter derivation in video coding
WO2020214564A1 (en) Method and apparatus for video encoding and decoding with optical flow based on boundary smoothed motion compensation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13778200

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 14380709

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2013778200

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2015506080

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE