EP3345395A1 - Method and apparatus for quantization in video encoding and decoding - Google Patents

Method and apparatus for quantization in video encoding and decoding

Info

Publication number
EP3345395A1
EP3345395A1 EP16759798.8A EP16759798A EP3345395A1 EP 3345395 A1 EP3345395 A1 EP 3345395A1 EP 16759798 A EP16759798 A EP 16759798A EP 3345395 A1 EP3345395 A1 EP 3345395A1
Authority
EP
European Patent Office
Prior art keywords
block
quantization
luminance
quantization parameter
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
EP16759798.8A
Other languages
German (de)
English (en)
French (fr)
Inventor
Franck Galpin
Hadrien GURNEL
Edouard Francois
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
InterDigital VC Holdings Inc
Original Assignee
Thomson Licensing SAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing SAS filed Critical Thomson Licensing SAS
Publication of EP3345395A1 publication Critical patent/EP3345395A1/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • H04N19/126Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • H04N19/463Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission

Definitions

  • This invention relates to a method and an apparatus for video encoding and decoding, and more particularly, to a method and an apparatus for adjusting a quantization parameter based on glare masking effects when encoding and decoding videos.
  • HDR (High Dynamic Range) videos generally represent a greater range of luminance levels than that can be achieved by conventional SDR (Standard Dynamic Range) videos, which usually have a 8- or 10-bit dynamic range.
  • SDR Standard Dynamic Range
  • some existing methods first perform forward conversion (110), which may include a conversion from HDR linear signals to non-linear signals, color space conversion, bit-depth reduction/quantization, and chroma down-conversion.
  • the signals after forward conversion can then be compressed using a video encoder (120), for example, an HEVC (High Efficiency Video Encoding) encoder that supports 8-bit and 10-bit video formats.
  • HEVC High Efficiency Video Encoding
  • the bitstream is decoded using a video decoder (130), for example, an HEVC decoder, and then is converted to HDR video signals using backward conversion (140), which may include color space conversion, bit-depth inverse quantization, chroma up-conversion, and conversion from non-linear signals to HDR linear signals.
  • a video decoder for example, an HEVC decoder
  • backward conversion 140
  • SMPTE 2084 defines a transfer function that takes into account the sensitivity of the HVS (Human Visual System) to luminance, which applies an OETF (Opto-Electronic Transfer Function) curve to each pixel independently.
  • the forward conversion module (110) may use the OETF curve and bit-depth quantization to transform the HDR videos to video signals represented with fewer bits, for example, to 10 or 12-bit signals according to SMPTE 2084, and the backward conversion module (140) may use an inverse OETF curve that corresponds to the OETF curve, for example, the Perceptual Quantizer (PQ) EOTF curve.
  • PQ Perceptual Quantizer
  • a method of encoding a video comprising: accessing a block of an image of said video; encoding said block based on a quantization parameter for said block, said quantization parameter being determined based on luminance of said block and luminance of neighboring blocks of said block, wherein said luminance of said block is determined using at least one transform coefficient in said block; and generating a bitstream responsive to the encoding.
  • said quantization parameter is determined based on a glare masking effect between one or more pixels of said neighboring blocks of said block and one or more pixels of said block.
  • a glare factor related to the glare masking effect can be calculated as Eqs. (4) and (5) of the detailed description.
  • the method further comprises: determining a JND (Just Noticeable Difference) for a pixel of said block responsive to said glare masking effect, wherein said quantization parameter is determined based on said determined JND and said luminance of said block.
  • said JND can be determined using Eqs. (2)-(3) of the detailed description.
  • said luminance of said block is determined using a DC transform coefficient of said block and a predicted block for said block.
  • quantization of AC transform coefficients of said block is based on said determined quantization parameter, and quantization of said DC transform coefficient is based on another quantization parameter.
  • the method further comprises: determining a quantization ratio based on said luminance of said block and said luminance of neighboring blocks of said block, wherein said quantization parameter is determined based on a second quantization parameter and said quantization ratio.
  • said quantization ratio can be determined as described in Eqs. (7)-(9) in the detailed description.
  • the present embodiments also provide an apparatus for encoding a video, comprising a memory and one or more processors configured to perform any of the methods described above.
  • the present embodiments also provide a non-transitory computer readable storage medium having stored thereon a bitstream generated according to any of the methods described above.
  • a method of decoding a video from a bitstream comprising: accessing said bitstream representing said video; decoding said block based on a quantization parameter for a block of an image of said video, said determined quantization parameter being determined based on luminance of said block and luminance of neighboring blocks of said block, wherein said luminance of said block is determined using at least one transform coefficient in said block; and outputting said video to at least one of a display, a storage, and a communication interface.
  • said luminance of said block is determined using a DC transform coefficient of said block and a predicted block for said block.
  • de-quantization of AC transform coefficients of said block is based on said determined quantization parameter, and de-quantization of said DC transform coefficient is based on another quantization parameter.
  • the method further comprises determining a quantization ratio based on said luminance of said block and said luminance of neighboring blocks of said block, wherein said quantization parameter is determined based on a second quantization parameter and said quantization ratio.
  • said quantization ratio can be determined as described in Eqs. (7)-(9) in the detailed descriptions.
  • the present embodiments also provide an apparatus for decoding a bitstream, comprising a memory and one or more processors configured to perform any of the methods described above.
  • the present embodiments also provide a non-transitory computer readable storage medium having stored thereon instructions for performing any of the methods described above.
  • a bitstream is presented, formatted to include: a block of an image of said video, encoded based on a quantization parameter, said quantization parameter being determined based on luminance of said block and luminance of neighboring blocks of said block, wherein said luminance of said block is determined using at least one transform coefficient in said block.
  • FIG. 1 is a block diagram illustrating an exemplary framework of encoding and decoding HDR signals.
  • FIG. 2A is a pictorial example illustrating an exemplary HDR image
  • FIG. 2B is a pictorial example illustrating the glare masking effect in the exemplary HDR image.
  • FIG. 3 illustrates an exemplary framework for using the glare masking effect to calculate the quantization ratio, according to an embodiment of the present principles.
  • FIG. 4 illustrates an exemplary framework for encoding an HDR image considering the glare masking effect, according to an embodiment of the present principles.
  • FIG. 5 illustrates another exemplary framework for encoding an HDR image considering the glare masking effect, according to an embodiment of the present principles.
  • FIG. 6 is a pictorial example illustrating exemplary causal areas of a current block.
  • FIG. 7 is a pictorial example illustrating an exemplary HEVC encoder.
  • FIG. 8 illustrates an exemplary method for adjusting the quantization parameter in a video encoder, according to an embodiment of the present principles.
  • FIG. 9 is a pictorial example illustrating an exemplary HEVC decoder.
  • FIG. 10 illustrates an exemplary method for adjusting the quantization parameter in a video decoder, according to an embodiment of the present principles.
  • FIG. 11 illustrates a block diagram depicting an exemplary system in which various aspects of the exemplary embodiments of the present principles may be implemented.
  • FIG. 12 illustrates a block diagram depicting an example of a video processing system that may be used with one or more implementations.
  • FIG. 13 illustrates a block diagram depicting another example of a video processing system that may be used with one or more implementations.
  • quantization may be used at different stages in representing the videos. Referring back to the example shown in FIG. 1, quantization is used in forward conversion to reduce the bit depth, and also used in the video encoder to quantize the transform coefficients. Similarly, inverse quantization is used in backward conversion to increase the bit depth and in the video decoder to de-quantize the transform coefficients. [34] Because the human eyes may become less sensitive to dark areas around the very bright areas in the videos (known as glare masking or luminance masking), we may use coarser quantization (i.e., preserving less detail or removing more detail) in such dark areas. In one embodiment, we can use an additional quantization process in the forward conversion or adjust the bit-depth quantization within the forward conversion. In another embodiment, we adjust quantization parameters used in the video encoder and decoder.
  • the glare masking is more common in HDR videos and can also be seen in SDR videos, for example, when an SDR or LDR (Low Dynamic Range) video is displayed by a TV set with strong backlight and high contrast.
  • SDR Small Dynamic Range
  • the forward conversion module (110) may include a conversion from the input SDR linear signals to non-linear signals, for example, using a gamma transfer function as described in SMPTE BT709, a color space conversion, bit-depth reduction/quantization, and a chroma down-conversion.
  • the backward conversion module (140), which converts the decoded signal to SDR video signals, may include color space conversion, bit-depth inverse quantization, chroma up-conversion, and conversion from non-linear signals to SDR linear signals, for example, using an inverse gamma transfer function. Note that in most cases, the signal before the inverse gamma processing can be sent to the display. In the following, we may discuss exemplary embodiments using HDR signals, but the present principles can also be applied to SDR signals. [36] FIG. 2A shows an exemplary HDR image, and FIG. 2B shows the glare masking effect in the exemplary HDR image. In FIG.
  • JND Just Noticeable Difference
  • JND difference limen
  • differential threshold differential threshold
  • JNDL JNDL
  • JNDG JNDG
  • other distortion or quality metrics for example, one that takes into account neighboring information, can be used in place of JND measures.
  • JNDL corresponds to the JND of one pixel without considering the glare masking. JNDL depends only on the luminance of the current pixel. Indeed, JNDL describes the variation of luminance that should exist at the current pixel of the image so that the human eye is able to notice the variation at the current pixel, without considering glare masking.
  • the JNDL can be determined experimentally. For example, for a given luminance L, we can find a value dL that represents the minimum delta in luminance for a human observer to see the change. These values are typically given as a mapping table between L and dL.
  • the JNDL can be determined from the PQ OETF curve for HDR signals, or the OETF defined in BT709 and the target peak luminance for SDR signals.
  • the OETF curve aims at mapping a linear signal (L) into a non-linear signal (Y) using a transfer function TF():
  • the transfer function for example, a PQ curve
  • TF(L+0.9*JND(L)) ⁇ Y+l i.e., a step of 0.9*JND in the linear signal L is less or equal to a step of 1 for the non-linear signal Y.
  • the luminance value of a pixel is directly proportional to the physical luminance (for example expressed in nits, or candela per meter square (cd/m2)).
  • the aim is to have the luminance of a pixel be linear with respect to the human visual system, i.e., a same amount of difference of luminance (dL) in this space should be perceived as a same amount of difference of luminance to human eyes, regardless of the pixel luminance L.
  • the perceptual space is defined experimentally and is subjective. Several perceptual spaces are available as known to those skilled in the art.
  • JNDG corresponds to the JND of one pixel considering glare masking, due to bright neighboring pixels.
  • JNDG depends on the luminance of the current pixel and the luminance of neighboring pixels. Indeed, it describes the minimal variation of luminance that should exist at the current pixel so that the human eye is able to notice a difference of luminance at the current pixel considering the effects of glare masking.
  • JNDG can be determined experimentally, for example, given a luminance, a bright spot and a distance to the current pixel, a mapping table can be obtained. Then modeling can be used to obtain an analytical function to best fit the mapping.
  • JNDG JND L and a glare factor (Gf) that considers the glare masking effect.
  • Gf glare factor
  • the Gf factor is greater than 1 (i.e., a pixel cannot have a JNDG that is smaller than the JND L ).
  • the Gf factor depends on the luminance of the current pixel, the luminance of surrounding pixels, and their position (distance) relative to the current pixel.
  • luminance values are expressed in the linear space (for example in nits or cd/m2). If the video input is expressed in the perceptual space (for example after an OETF function), the inverse function can be applied to the intensity value of the pixel, in order to obtain the linear luminance value.
  • the modeling of the glare factor can be complex depending on the HVS model that is used to consider glare masking.
  • the glare factor of pixel pi which is affected by pixel p 2 (with L(p 2 )>L(p ) can be calculated as:
  • Gf(pi,p 2 ) max ⁇ l,[a 0 + (L(p 2 )-L(pi)) al ] . [(L(p 2 )-L(pi))*a 2 *d(pi,p 2 ) + a 3 ] ⁇ (4)
  • L(p) is the luminance of pixel p
  • d(pi,p 2 ) is the Euclidian distance between pixels pi and p 2
  • [ao...a 3 ] are constants, determined empirically throughout subjective tests.
  • a typical set of values for [ao...a 3 ] can be, for example: [6.75, 0.352, -3.74e-008, 3.360e-005] when SMPTE 2084 OETF is used.
  • the values can be adapted if another OETF is used, for example, when the gamma transfer function defined in SMPTE BT709 is used for SDR signals.
  • the glare factor decreases with the distance because a 2 is negative, increases when pixel p 2 is brighter, and increases when pi is darker, consistent with the perception of the glare masking effect.
  • the above computation for pixels can be simplified. For example, we may only compute glare factors for "dark" pixels, namely, pixels with the darkness below a threshold (typically 100 nits), and we may only compute the contribution to masking for "bright" pixels, namely, pixels with brightness above a threshold (typically 1000 nits). For those pixels that the computation is skipped, we set Gf to 1 (i.e.,
  • FIG. 3 illustrates an exemplary method 300 for calculating a quantization ratio based on the glare masking effect, according to an embodiment of the present principles.
  • Method 300 accesses an original video (for example using the OpenEXR file format as defined by the SMPTE 2065-4:2013) in a linear space as input.
  • the original image can be, for example, in a RGB linear format represented by floating points using the BT2020 format for the primaries with 4000 nits as peak luminance.
  • the luminance value (L) can be calculated in order to consider glare masking (310).
  • an input RGB (linear space, floating values in nits) image I rg b can be converted from RGB BT2020 color space to XYZ color space as:
  • JNDG 320
  • Eq. (3) 0.262700 R + 0.677998 G +0.059302B.
  • JNDG 320
  • the color space conversion can be adapted, for example using the one defined in SMPTE BT709.
  • a quantization ratio can be computed for a pixel (330) as follows:
  • the OETF was designed such that a step of JND in the linear space is no less than a step of 1 in the perceptual space.
  • FIG. 4 illustrates an exemplary method 400 for encoding an HDR image considering the glare masking effect, according to an embodiment of the present principles.
  • the glare masking effect is used to adjust the quantization step size.
  • quantization ratios for individual pixels can be calculated (450), for example, using method 300.
  • the input HDR signals are converted to signals that can be accepted by a video encoder using forward conversion (410).
  • the quantization step size is adjusted based on the quantization ratio. Since video encoding usually proceeds in a block basis, we further calculate the quantization ratio for a block using the quantization ratios calculated for individual pixels.
  • the size of a block depends on how we apply the quantization ratio.
  • the block may corresponds to one or more macroblocks in H.264/AVC, or one or more transform units (TUs) in HEVC.
  • TUs transform units
  • the quantization ratio for a block can be calculated using the max function of the quantization ratios of pixels within the block. That is, the maximum quantization ratio for the pixels in the block is used as the quantization ratio for the block. This approach may improve the compression performance at a cost of the visual quality.
  • the quantization ratio for a block can be calculated using the minimum function, which may better preserve visual quality of the block.
  • the median or average of the quantization ratios can be used as the quantization ratio for the block, which may provide a balance between the compression efficiency and visual quality.
  • the quantization step size with the proposed quantization adjustment can be calculated as:
  • Q 2 min(Q max , Qr*Qi) (10) where Q ma x is the upper limit of the quantization step size.
  • a transform coefficient T can be quantized as: [(
  • Other quantization parameters, such as quantization matrix can also be used during quantization.
  • the value of Q 2 might be further adapted. Different codecs have different constraints on the quantization step size that can be set for a block. For example, in VP9, only a limited number of different quantization step sizes (Qps) are available. In this case, an additional process of Qp clustering can be performed.
  • a delta Qp is encoded instead, limiting the possible value of Q 2 .
  • the quantization may not be performed independently as a separate step within the encoder.
  • the quantization may be integrated with the transform.
  • the quantization step size may be processed before being used for quantization.
  • the quantization parameter when they is to be encoded, they may be mapped to a quantization index before being encoded. For ease of notation, we refer to different representations corresponding to the quantization step size as the quantization step size.
  • the quantization ratio adjustment in method 400 can also be viewed as a pre-processing step to improve the video encoding.
  • the bitstream is decoded (430) and then converted to HDR signals through backward conversion (440).
  • FIG. 5 illustrates another exemplary method 500 for encoding an HDR image considering the glare masking effect, according to an embodiment of the present principles.
  • the glare masking effect is used to adjust the quantization step size that does not require the transmission of the quantization ratio, that is, the quantization ratio based on the glare masking can be deduced on the decoder.
  • the quantization adjustment is performed in both the encoder and decoder, and the process of deriving the quantization ratio is the same at the encoder and decoder.
  • FIG. 6 illustrates an exemplary causal area (an area that is already encoded or decoded, and the pixels are available) of a current block 610.
  • the causal area in this example includes the blocks to the left and above of the current block.
  • neighbors can furthermore be restricted to pixels with brightness above a threshold and pixels within a distance.
  • the video encoder (520) encodes the converted signals, using quantization adjustment (550) based on glare masking.
  • the video decoder (530) decodes the bitstream, using quantization adjustment (560) based on glare masking.
  • the decoded signals are then converted to output HDR signals using backward conversion (540).
  • FIG. 7 illustrates an exemplary HEVC encoder 700 wherein the present principles may be applied.
  • the input of encoder 700 includes a video to be encoded.
  • the exemplary encoder 700 when a block is encoded in an intra mode, it performs intra prediction (770). In an inter mode, the block performs motion estimation (760) and motion compensation (765).
  • the encoder decides which one of the intra mode or inter mode to use for encoding the block (775), and prediction residuals are calculated by subtracting the predicted block from the original image block (705).
  • the prediction residuals are transformed (710) and quantized (720).
  • the quantized transform coefficients, as well as motion vectors and other syntax elements, are entropy coded (730) to generate a bitstream.
  • the encoder decodes an encoded block to provide a reference for further predictions.
  • the quantized transform coefficients are de-quantized (740) and inverse transformed (750) to decode prediction residuals.
  • a deblocking filter (780) and SAO (Sample Adaptive Offset) filter (785) are applied to the reconstructed block.
  • the filtered image is stored at a reference memory (790).
  • FIG. 8 illustrates an exemplary method 800 for adjusting the quantization parameter in a video encoder, according to an embodiment of the present principles.
  • Method 800 may be used in the quantization module (720) of encoder 700.
  • Qi can be decided by a rate control algorithm to meet the bitrate constraint, and can vary from block to block and from picture to picture.
  • the DC coefficient is used to estimate the luminance and estimate the glare factor for the current block.
  • the AC coefficients are quantized based on adjusted quantization step size Q 2 .
  • the input to method 800 includes an original block (B) and a corresponding predicted block (P).
  • a predicted block can be from, for example, intra prediction, or motion compensation.
  • [DC] D 1 (D(DC,Qi)
  • [DC] is the reconstructed DC coefficient.
  • A ⁇ i P (.i) + ⁇ OC ⁇ , (12)
  • P(i) is the intensity of each pixel from the predicted block
  • N is the number of pixels in the block
  • A is the estimate average luminance of the block.
  • the estimated luminance value A is used as the current value for the whole block and the glare factor is computed, using only the causal part of the image for the current block. At this point, the luminance value of reconstructed pixels from the causal part can be used.
  • the glare factor may be computed using luminance in the linear space
  • a quantization ratio (Qr) can then be estimated (870) based on the estimated luminance values for the current block (L) and the causal blocks ( ⁇ Li ⁇ ).
  • the glare factor calculation is similar to the one between pixels, but using block centers to compute the distance between the blocks, and using the average luminance of the block instead of pixels' luminance.
  • FIG. 9 depicts a block diagram of an exemplary HEVC video decoder 900 wherein the present principles may be applied.
  • the input of decoder 900 includes a video bitstream, which may be generated by video encoder 700.
  • the bitstream is first entropy decoded (945) to obtain transform coefficients, motion vectors, and other coded information.
  • the transform coefficients are de-quantized (950) and inverse transformed (955) to decode the prediction residuals.
  • the predicted block may be obtained from intra prediction (960) or motion-compensated prediction (970).
  • a deblocking filter (990) and a SAO filter (995) are applied to the reconstructed block or the reconstructed image.
  • the filtered image is stored at a reference memory (980).
  • FIG. 10 illustrates an exemplary method 1000 for adjusting the quantization parameter in a decoder, according to an embodiment of the present principles.
  • a predicted block (P) can be from, for example, intra prediction, or motion compensation.
  • the DC coefficient is de-quantized (1040) using the quantization step size Qi as:
  • ⁇ DC is the reconstructed DC coefficient.
  • A ⁇ i P (.i) + ⁇ DC] (15) where P(i) is the intensity of each pixel from the predicted block, N is the number of pixels in the block, and A is the estimate average luminance of the block. [81]
  • the estimated luminance value A is used as the current value for the whole block and the glare factor is computed, using only the causal part of the image for the current block. At this point, the luminance value of decoded pixels from the causal part can be used.
  • a quantization ratio (Qr) can then be estimated (1070) based on the estimated luminance values for the current block (L) and the causal blocks ( ⁇ Li ⁇ ).
  • the AC coefficients are de-quantized (1080).
  • the de-quantized DC coefficient and AC coefficients can then be inverse transformed (1090). It should be noted that the quantization ratio (Qr) is not received in the bitstream, rather, it is derived at the decoder.
  • the quantization ratio calculation performed in the encoder and decoder should correspond to each other.
  • the steps of 1040- 1075 performed in method 1000 correspond to the steps of 840-875 in method 800, respectively.
  • the present embodiments may improve video coding efficiency.
  • the adaptive quantization of each block also takes into consideration the masking effect and may also improve the visual quality.
  • the quantization step size can be adjusted by taking into account the glare masking effect.
  • Other quantization parameters such as quantization round offset and quantization matrix can also be adjusted according to the present principles.
  • the quantization matrix can be scaled using the quantization ratio, or the quantization rounding offset may be shifted or scaled based on the quantization ratio.
  • FIG. 11 illustrates a block diagram of an exemplary system in which various aspects of the exemplary embodiments of the present principles may be implemented.
  • System 1100 may be embodied as a device including the various components described below and is configured to perform the processes described above. Examples of such devices, include, but are not limited to, personal computers, laptop computers, smartphones, tablet computers, digital multimedia set top boxes, digital television receivers, personal video recording systems, connected home appliances, and servers.
  • System 1100 may be communicatively coupled to other similar systems, and to a display via a communication channel as shown in FIG. 11 and as known by those skilled in the art to implement the exemplary video system described above.
  • the system 1100 may include at least one processor 1110 configured to execute instructions loaded therein for implementing the various processes as discussed above.
  • Processor 1110 may include embedded memory, input output interface and various other circuitries as known in the art.
  • the system 1100 may also include at least one memory 1120 (e.g., a volatile memory device, a non-volatile memory device).
  • System 1100 may additionally include a storage device 1140, which may include non-volatile memory, including, but not limited to, EEPROM, ROM, PROM, RAM, DRAM, SRAM, flash, magnetic disk drive, and/or optical disk drive.
  • the storage device 1140 may comprise an internal storage device, an attached storage device and/or a network accessible storage device, as non-limiting examples.
  • System 1100 may also include an encoder/decoder module 1130 configured to process data to provide an encoded video or decoded video.
  • Encoder/decoder module 1130 represents the module(s) that may be included in a device to perform the encoding and/or decoding functions. As is known, a device may include one or both of the encoding and decoding modules. Additionally, encoder/decoder module 1130 may be implemented as a separate element of system 1100 or may be incorporated within processors 1110 as a combination of hardware and software as known to those skilled in the art.
  • processors 1110 Program code to be loaded onto processors 1110 to perform the various processes described hereinabove may be stored in storage device 1140 and subsequently loaded onto memory 1120 for execution by processors 1110.
  • one or more of the processor(s) 1110, memory 1120, storage device 1140 and encoder/decoder module 1130 may store one or more of the various items during the performance of the processes discussed herein above, including, but not limited to the HDR video, the bitstream, equations, formula, matrices, variables, operations, and operational logic.
  • the system 1100 may also include communication interface 1150 that enables communication with other devices via communication channel 1160.
  • the communication interface 1150 may include, but is not limited to a transceiver configured to transmit and receive data from communication channel 1160.
  • the communication interface may include, but is not limited to, a modem or network card and the communication channel may be implemented within a wired and/or wireless medium.
  • the various components of system 1100 may be connected or communicatively coupled together using various suitable connections, including, but not limited to internal buses, wires, and printed circuit boards.
  • the exemplary embodiments according to the present principles may be carried out by computer software implemented by the processor 1110 or by hardware, or by a combination of hardware and software.
  • the exemplary embodiments according to the present principles may be implemented by one or more integrated circuits.
  • the memory 1120 may be of any type appropriate to the technical environment and may be implemented using any appropriate data storage technology, such as optical memory devices, magnetic memory devices, semiconductor-based memory devices, fixed memory and removable memory, as non-limiting examples.
  • the processor 1110 may be of any type appropriate to the technical environment, and may encompass one or more of microprocessors, general purpose computers, special purpose computers and processors based on a multi-core architecture, as non-limiting examples.
  • the data transmission system 1200 may be, for example, a head-end or transmission system for transmitting a signal using any of a variety of media, such as, satellite, cable, telephone-line, or terrestrial broadcast.
  • the data transmission system 1200 also may be used to provide a signal for storage.
  • the transmission may be provided over the Internet or some other network.
  • the data transmission system 1200 is capable of generating and delivering, for example, video content and other content.
  • the data transmission system 1200 receives processed data and other information from a processor 1201. In one implementation, the processor 1201 performs forward conversion. The processor 1201 may also provide metadata to 1200 indicating, for example, the format of the video. [97]
  • the data transmission system or apparatus 1200 includes an encoder 1202 and a transmitter 1204 capable of transmitting the encoded signal. The encoder 1202 receives data information from the processor 1201. The encoder 1202 generates an encoded signal(s).
  • the encoder 1202 may include sub-modules, including for example an assembly unit for receiving and assembling various pieces of information into a structured format for storage or transmission.
  • the various pieces of information may include, for example, coded or uncoded video, and coded or uncoded elements.
  • the encoder 1202 includes the processor 1201 and therefore performs the operations of the processor 1201.
  • the transmitter 1204 receives the encoded signal(s) from the encoder 1202 and transmits the encoded signal(s) in one or more output signals.
  • the transmitter 1204 may be, for example, adapted to transmit a program signal having one or more bitstreams representing encoded pictures and/or information related thereto.
  • Typical transmitters perform functions such as, for example, one or more of providing error-correction coding, interleaving the data in the signal, randomizing the energy in the signal, and modulating the signal onto one or more carriers using a modulator 1206.
  • the transmitter 1204 may include, or interface with, an antenna (not shown). Further, implementations of the transmitter 1204 may be limited to the modulator 1206.
  • the data transmission system 1200 is also communicatively coupled to a storage unit 1208.
  • the storage unit 1208 is coupled to the encoder 1202, and stores an encoded bitstream from the encoder 1202.
  • the storage unit 1208 is coupled to the transmitter 1204, and stores a bitstream from the transmitter 1204.
  • the bitstream from the transmitter 1204 may include, for example, one or more encoded bitstreams that have been further processed by the transmitter 1204.
  • the storage unit 1208 is, in different implementations, one or more of a standard DVD, a Blu-Ray disc, a hard drive, or some other storage device.
  • the data receiving system 1300 may be configured to receive signals over a variety of media, such as storage device, satellite, cable, telephone-line, or terrestrial broadcast.
  • the signals may be received over the Internet or some other network.
  • the data receiving system 1300 may be, for example, a cell-phone, a computer, a set-top box, a television, or other device that receives encoded video and provides, for example, decoded video signal for display (display to a user, for example), for processing, or for storage.
  • the data receiving system 1300 may provide its output to, for example, a screen of a television, a computer monitor, a computer (for storage, processing, or display), or some other storage, processing, or display device.
  • the data receiving system 1300 is capable of receiving and processing data information.
  • the data receiving system or apparatus 1300 includes a receiver 1302 for receiving an encoded signal, such as, for example, the signals described in the implementations of this application.
  • the receiver 1302 may receive, for example, a signal providing a bitstream, or a signal output from the data transmission system 1200 of FIG. 12.
  • the receiver 1302 may be, for example, adapted to receive a program signal having a plurality of bitstreams representing encoded pictures.
  • Typical receivers perform functions such as, for example, one or more of receiving a modulated and encoded data signal, demodulating the data signal from one or more carriers using a demodulator 1304, de-randomizing the energy in the signal, de-interleaving the data in the signal, and error-correction decoding the signal.
  • the receiver 1302 may include, or interface with, an antenna (not shown). Implementations of the receiver 1302 may be limited to the demodulator 1304.
  • the data receiving system 1300 includes a decoder 1306.
  • the receiver 1302 provides a received signal to the decoder 1306.
  • the signal provided to the decoder 1306 by the receiver 1302 may include one or more encoded bitstreams.
  • the decoder 1306 outputs a decoded signal, such as, for example, decoded video signals including video information.
  • the data receiving system or apparatus 1300 is also communicatively coupled to a storage unit 1307.
  • the storage unit 1307 is coupled to the receiver 1302, and the receiver 1302 accesses a bitstream from the storage unit 1307.
  • the storage unit 1307 is coupled to the decoder 1306, and the decoder 1306 accesses a bitstream from the storage unit 1307.
  • the bitstream accessed from the storage unit 1307 includes, in different implementations, one or more encoded bitstreams.
  • the storage unit 1307 is, in different implementations, one or more of a standard DVD, a Blu-Ray disc, a hard drive, or some other storage device.
  • the output data from the decoder 1306 is provided, in one implementation, to a processor 1308.
  • the processor 1308 is, in one implementation, a processor configured for performing post-processing.
  • the decoder 1306 includes the processor 1308 and therefore performs the operations of the processor 1308.
  • the processor 1308 is part of a downstream device such as, for example, a set-top box or a television.
  • the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed may also be implemented in other forms (for example, an apparatus or program).
  • An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • the appearances of the phrase “in one embodiment” or “in an embodiment” or “in one implementation” or “in an implementation”, as well any other variations, appearing in various places throughout the specification are not necessarily all referring to the same embodiment.
  • Determining the information may include one or more of, for example, estimating the information, calculating the information, predicting the information, or retrieving the information from memory.
  • Accessing the information may include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
  • Receiving is, as with “accessing”, intended to be a broad term.
  • Receiving the information may include one or more of, for example, accessing the information, or retrieving the information (for example, from memory).
  • “receiving” is typically involved, in one way or another, during operations such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
  • implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted.
  • the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
  • a signal may be formatted to carry the bitstream of a described embodiment.
  • Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
  • the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
  • the information that the signal carries may be, for example, analog or digital information.
  • the signal may be transmitted over a variety of different wired or wireless links, as is known.
  • the signal may be stored on a processor-readable medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
EP16759798.8A 2015-09-02 2016-09-02 Method and apparatus for quantization in video encoding and decoding Pending EP3345395A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP15306348 2015-09-02
PCT/EP2016/070711 WO2017037228A1 (en) 2015-09-02 2016-09-02 Method and apparatus for quantization in video encoding and decoding

Publications (1)

Publication Number Publication Date
EP3345395A1 true EP3345395A1 (en) 2018-07-11

Family

ID=54148449

Family Applications (1)

Application Number Title Priority Date Filing Date
EP16759798.8A Pending EP3345395A1 (en) 2015-09-02 2016-09-02 Method and apparatus for quantization in video encoding and decoding

Country Status (10)

Country Link
US (1) US10491899B2 (ru)
EP (1) EP3345395A1 (ru)
JP (1) JP6989491B2 (ru)
KR (1) KR20180048654A (ru)
CN (1) CN108141598B (ru)
CA (1) CA2996349A1 (ru)
MX (1) MX2018002646A (ru)
RU (1) RU2018111430A (ru)
TW (1) TWI727968B (ru)
WO (1) WO2017037228A1 (ru)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10778978B2 (en) * 2017-08-21 2020-09-15 Qualcomm Incorporated System and method of cross-component dynamic range adjustment (CC-DRA) in video coding
WO2020004833A1 (ko) * 2018-06-29 2020-01-02 엘지전자 주식회사 Dc 계수를 적응적으로 결정하는 방법 및 그 장치
US10931950B2 (en) 2018-11-19 2021-02-23 Intel Corporation Content adaptive quantization for video coding
CN110473149B (zh) * 2019-03-13 2020-06-09 北方工业大学 关键大数据模糊化处理系统
US11557264B2 (en) * 2019-04-10 2023-01-17 Hefei Reliance Memory Limited Display driver system with embedded non-volatile memory
CN110020695A (zh) * 2019-04-19 2019-07-16 杭州电子科技大学 用于滤波器组多载波调制光通信系统的K-means非均匀量化算法
EP3949402A4 (en) * 2019-04-26 2022-06-08 Huawei Technologies Co., Ltd. METHOD AND DEVICE FOR SIGNALING THE MAPPING FUNCTION OF A CHROMA QUANTIZATION PARAMETER
WO2020248099A1 (en) * 2019-06-10 2020-12-17 Realnetworks, Inc. Perceptual adaptive quantization and rounding offset with piece-wise mapping function
CN112738515B (zh) * 2020-12-28 2023-03-24 北京百度网讯科技有限公司 用于自适应量化的量化参数调整方法和装置
WO2022220545A1 (ko) * 2021-04-12 2022-10-20 엘지전자 주식회사 저주파 비분리 변환 설계 방법 및 장치
US20230345015A1 (en) * 2022-04-13 2023-10-26 Tencent America LLC Chroma from luma prediction using neighbor luma samples

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3152765B2 (ja) * 1991-10-31 2001-04-03 株式会社東芝 画像符号化装置
JPH07288809A (ja) * 1994-04-18 1995-10-31 Sony Corp ディジタル画像信号の符号化装置
US5629780A (en) 1994-12-19 1997-05-13 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Image data compression having minimum perceptual error
KR100355375B1 (ko) * 1995-11-01 2002-12-26 삼성전자 주식회사 영상부호화장치에있어서양자화간격결정방법및회로
US6363113B1 (en) 1999-06-07 2002-03-26 Lucent Technologies Inc. Methods and apparatus for context-based perceptual quantization
US7280689B2 (en) * 2002-07-05 2007-10-09 Qdesign U.S.A., Inc. Anti-compression techniques for visual images
US8243797B2 (en) * 2007-03-30 2012-08-14 Microsoft Corporation Regions of interest for quality adjustments
JP5024632B2 (ja) 2008-09-19 2012-09-12 ソニー株式会社 画像処理装置および方法、並びにプログラム
US8279924B2 (en) * 2008-10-03 2012-10-02 Qualcomm Incorporated Quantization parameter selections for encoding of chroma and luma video blocks
TW201134223A (en) * 2010-03-29 2011-10-01 Univ Nat Taiwan Perceptual video encoding system and circuit thereof
RU2013149856A (ru) * 2011-04-08 2015-05-20 Конинклейке Филипс Н.В. Кодирование и декодирование видео
WO2013063638A2 (en) * 2011-10-30 2013-05-10 Hd2 Technologies Pty. Ltd. Perceptually lossless and perceptually enhanced image compression system & method
US20130188691A1 (en) * 2012-01-20 2013-07-25 Sony Corporation Quantization matrix design for hevc standard
US9219916B2 (en) 2012-06-12 2015-12-22 Dolby Laboratories Licensing Corporation Joint base layer and enhancement layer quantizer adaptation in EDR video coding
CN103297773B (zh) * 2013-05-07 2016-05-04 福州大学 基于jnd模型的图像编码方法
US20150006346A1 (en) * 2013-06-28 2015-01-01 Quisk, Inc. Account engine with modular services and access channels
US20150063461A1 (en) * 2013-08-27 2015-03-05 Magnum Semiconductor, Inc. Methods and apparatuses for adjusting macroblock quantization parameters to improve visual quality for lossy video encoding
JP2016541140A (ja) 2013-11-13 2016-12-28 エルジー エレクトロニクス インコーポレイティド Hdr放送サービスの提供のための放送信号送受信方法及び装置
EP2887664A1 (en) * 2013-12-19 2015-06-24 Thomson Licensing Method and device for encoding a high-dynamic range image
CN104378636B (zh) * 2014-11-10 2017-07-07 中安消技术有限公司 一种视频图像编码方法及装置

Also Published As

Publication number Publication date
KR20180048654A (ko) 2018-05-10
CN108141598A (zh) 2018-06-08
US20180255302A1 (en) 2018-09-06
JP6989491B2 (ja) 2022-01-05
WO2017037228A1 (en) 2017-03-09
TWI727968B (zh) 2021-05-21
RU2018111430A (ru) 2019-10-07
US10491899B2 (en) 2019-11-26
TW201720157A (zh) 2017-06-01
MX2018002646A (es) 2018-06-20
RU2018111430A3 (ru) 2020-03-10
JP2018533860A (ja) 2018-11-15
CA2996349A1 (en) 2017-03-09
CN108141598B (zh) 2022-10-04

Similar Documents

Publication Publication Date Title
US10491899B2 (en) Method and apparatus for quantization in video encoding and decoding
US11368701B2 (en) Adaptive perceptual mapping and signaling for video coding
CN107209929B (zh) 用于对高动态范围图像进行处理的方法和装置
JP6651005B2 (ja) Qp値の決定
CN110084755B (zh) 用于图像处理的方法、装置及介质
KR102631484B1 (ko) 이미지들 및 비디오의 전기-광학 및 광-전기 변환을 위한 방법들, 시스템들 및 장치
US10701359B2 (en) Real-time content-adaptive perceptual quantizer for high dynamic range images
US20190281295A1 (en) Encoding device, decoding device, encoding method, and decoding method
US20140254689A1 (en) Video coding method using at least evaluated visual quality and related video coding apparatus
EP3323104B1 (en) A method and device for tone-mapping a picture by using a parametric tone-adjustment function
Zhang et al. High dynamic range image & video compression a review
EP3054417A1 (en) Method and apparatus for encoding and decoding high dynamic range images
CN115499632A (zh) 一种图像信号转换处理方法、装置及终端设备
EP3026908A1 (en) Method and device for quantizing and de-quantizing a picture using scaling factors for chrominance based on luminance
Liu et al. An adaptive perceptual quantization method for HDR video coding
Mir et al. Adaptive residual mapping for an efficient extension layer coding in two-layer HDR video coding
EP3272120B1 (en) Adaptive perceptual mapping and signaling for video coding
EP3121787A1 (en) A method and device for tone-mapping a picture by using a parametric tone-adjustment function
Mantiuk et al. Perception-inspired high dynamic range video coding and compression
Prangnell Spatiotemporal Adaptive Quantization for Video Compression Applications

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20180302

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: INTERDIGITAL VC HOLDINGS, INC.

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20191011

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: INTERDIGITAL VC HOLDINGS, INC.

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS