US11039141B2 - Video coding device, video decoding device, video coding method, video decoding method and program - Google Patents

Video coding device, video decoding device, video coding method, video decoding method and program Download PDF

Info

Publication number
US11039141B2
US11039141B2 US15/516,223 US201515516223A US11039141B2 US 11039141 B2 US11039141 B2 US 11039141B2 US 201515516223 A US201515516223 A US 201515516223A US 11039141 B2 US11039141 B2 US 11039141B2
Authority
US
United States
Prior art keywords
offset
chroma quantization
color
chroma
space
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US15/516,223
Other versions
US20170318293A1 (en
Inventor
Keiichi Chono
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp filed Critical NEC Corp
Assigned to NEC CORPORATION reassignment NEC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHONO, KEIICHI
Publication of US20170318293A1 publication Critical patent/US20170318293A1/en
Application granted granted Critical
Publication of US11039141B2 publication Critical patent/US11039141B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • H04N19/126Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • the present invention relates to a video coding device and a video decoding device using an adaptive color transform in residual domain and a chroma (color difference) quantization offset.
  • each frame of a digitized image is divided into coding tree units (CTUs) and the respective CTUs are coded in the order of raster scan.
  • the CTU is divided into coding units (CUs) in a quad-tree structure and then coded.
  • Each CU is divided into prediction units (PUs) and predicted.
  • PUs prediction units
  • a prediction error of each CU is divided into transform units (TUs) in a quad-tree structure and then frequency-transformed.
  • the CU is a coding unit for an intra prediction or inter-frame prediction.
  • the intra prediction is prediction for generating a prediction signal from a reconstructed image of a frame to be encoded.
  • 33 types of angular intra prediction and the like are defined.
  • angular intra prediction a reconstructed pixel around a block to be encoded is extrapolated in any of 33 directions depicted in FIG. 14 , to generate an intra prediction signal.
  • DC prediction In addition to angular intra prediction, DC prediction and planar prediction are specified as intra prediction.
  • DC prediction a mean value of a reference image is used as the prediction values of all pixels in a TU to be predicted.
  • planar prediction a prediction image is generated by linear interpolation from pixels in a reference image.
  • Inter-frame prediction is a prediction based on an image of a reconstructed frame (reference picture) different in display time from a frame to be encoded. Inter-frame prediction is also referred to as inter prediction. In inter prediction, an inter prediction signal is generated based on a reconstructed image block of a reference picture (using pixel interpolation if necessary).
  • a digital color image is composed of RGB digital images.
  • a digital color image is generally transformed to signals in a color space other than the RGB space to increase the compression efficiency (to decrease the data amount) when the color image is transmitted through a transmission line.
  • the image signal is transformed to a signal in a color space (YCoCr space) where a combination of a luma signal (Y) and chroma signals (Cb, Cr) are composed.
  • a quantization parameter (QP) for the chroma signal is generated by transforming the QP for the luma signal by using an offset value “chroma_qp_index_offset.”
  • cb_qp_index_offset (a first chroma quantization offset) is applied to Cb
  • cr_qp_index_offset (a second chroma quantization offset) is applied to Cr.
  • the adaptive color transform in residual domain is a technique of switching a prediction error signal of an image signal of the RGB space to a signal of the YCoCr space adaptively in block units (per block).
  • FIG. 17 illustrates an example that data is compressed in the YCoCr space with respect to the shaded blocks and data is compressed in the RGB space with respect to other blocks.
  • a receiver restores the signal of the YCoCr space to the signal of the RGB space by using a backward color space transform matrix described below and then performs decoding processing.
  • Patent Literature 1 describes a video coding device and a video decoding device which perform signal processing which is different according to whether an input image signal is a signal of the RGB space or a signal of the YCoCr space. Specifically, when performing a weighted prediction based on the H.264/AVC, the video coding device applies the same offset to the R, G, and B signals and to the luma signal (Y signal) and applies different offsets to the chroma signals, regarding an offset added to a prediction signal. However, Patent Literature 1 does not teach any new knowledge regarding the chroma quantization offset.
  • the following describes the configuration and operation of a general video coding device which outputs a bitstream with each CU of each frame of a digitized image as an input image.
  • a video coding device illustrated in FIG. 18 includes a switch 101 , a color space transformer 102 , a switch 103 , a frequency transformer/quantizer 104 , an inverse quantizer/inverse frequency transformer 105 , a switch 106 , an inverse color space transformer 107 , a switch 108 , a buffer 109 , a predictor 110 , a prediction parameter determiner 111 , an entropy encoder 112 , a subtractor 115 , and an adder 116 .
  • the predictor 110 generates a prediction signal for a CU input image signal. Specifically, the predictor 110 generates a prediction signal (an intra-prediction signal) on the basis of intra prediction and generates a prediction signal (an inter-prediction signal) on the basis of inter-prediction.
  • a prediction image supplied from the predictor 110 is subtracted from an image input to the video coding device at the subtractor 115 and then input to the switch 101 as a prediction error image.
  • an input image signal is a signal of the RGB space.
  • the video coding device has a function of an adaptive color transform in residual domain. For example, the video coding device is able to adaptively switch a prediction error signal of an image signal of the RGB space to a signal of the YCoCr space in block units (per block).
  • the switch 101 When a prediction error signal of the RGB space is used, the switch 101 is set so that the prediction error image is input to the switch 103 . When the prediction error signal of the YCoCr space is used, the switch 101 is set so that the prediction error image is input to the color space transformer 102 . Moreover, the switch 101 sets an output destination of a prediction error image according to the control of the prediction parameter determiner 111 , for example.
  • the color space transformer 102 transforms the prediction error signal of the RGB space to a signal of the YCoCr space by using the aforementioned equation (1) (a forward color space transform matrix) and then outputs the signal to the switch 103 .
  • the switch 103 When a prediction error signal of the RGB space is used, the switch 103 outputs a prediction error signal, which has been received from the switch 101 , to the frequency transformer/quantizer 104 . When a prediction error signal of the YCoCr space is used, the switch 103 outputs a prediction error signal, which has been received from the color space transformer 102 , to the frequency transformer/quantizer 104 . Moreover, the switch 103 selects the input source of the prediction error image according to the control of the prediction parameter determiner 111 , for example.
  • the frequency transformer/quantizer 104 frequency-transforms the prediction error image and quantizes the frequency-transformed prediction error image (a coefficient image).
  • the entropy encoder 112 performs entropy coding of a prediction parameter and of a quantization coefficient image and then outputs a bitstream
  • the inverse quantizer/inverse frequency transformer 105 inverse-quantizes the quantization coefficient image. Further, the inverse quantizer/inverse frequency transformer 105 performs inverse frequency transform on the inverse-quantized coefficient image. The inverse frequency transformed reconstructed prediction error image is input to the switch 106 .
  • the switch 106 When the prediction error signal of the RGB space is used, the switch 106 is set so that the reconstructed prediction error image is input to the switch 108 . When the prediction error signal of the YCoCr space is used, the switch 106 is set so that the reconstructed prediction error image is input to the inverse color space transformer 107 . The switch 106 selects an output destination of the reconstructed prediction error image according to the control of the prediction parameter determiner 111 .
  • the inverse color space transformer 107 transforms a reconstructed prediction error signal of the YCoCr space to a signal of the RGB space by using the aforementioned equation (1) (a backward color space transform matrix) and then outputs the to the switch 108 .
  • the switch 108 selects the reconstructed prediction error signal received from the switch 106 when the prediction error signal of the RGB space is used.
  • the switch 108 selects the reconstructed prediction error signal received from the inverse color space transformer 107 .
  • the switch 108 selects any of the reconstructed prediction error images according to the control of the prediction parameter determiner 111 , for example.
  • the reconstructed prediction error image received from the switch 108 is supplied with a prediction signal by the adder 116 and then supplied as a reconstructed image to the buffer 109 .
  • the buffer 109 stores the reconstructed image.
  • the prediction parameter determiner 111 instructs the predictor 110 to decide the prediction parameter which minimizes coding cost by comparing the input image signal with the prediction signal, for example.
  • the prediction parameter determiner 111 supplies the decided prediction parameter to the entropy encoder 112 .
  • the prediction parameter is information related to block prediction such as a prediction mode (intra prediction, inter-prediction), an intra prediction block size, an intra prediction direction, an inter-prediction block size, a motion vector, or the like.
  • the prediction parameter determiner 111 further instructs to decide whether the prediction error signal of the RGB space or the prediction error signal of the YCoCr space is used for each block.
  • FIG. 19 is a block diagram illustrating an example of the configuration of a general video decoding device which decodes the bitstream output from a general video coding device to acquire a decoded image. Referring to FIG. 19 , the configuration and operation of the general video decoding device will be described below.
  • the video decoding device illustrated in FIG. 19 includes an entropy decoder 212 , an inverse quantizer/inverse frequency transformer 205 , a switch 206 , an inverse color space transformer 207 , a switch 208 , a buffer 209 , a predictor 210 , and an adder 216 .
  • the entropy decoder 212 entropy-decodes an input bitstream.
  • the entropy decoder 212 supplies a quantization coefficient image to the inverse quantizer/inverse frequency transformer 205 and supplies a prediction parameter to the predictor 210 .
  • the inverse quantizer/inverse frequency transformer 205 inverse-quantizes the input quantization coefficient image and outputs it as a coefficient image. Furthermore, the inverse quantizer/inverse frequency transformer 205 transforms the coefficient image in the frequency domain to an image in the space domain and outputs it as a prediction error image. The prediction error image is input to the switch 206 .
  • the switch 206 When the prediction error signal of the RGB space is used, the switch 206 is set so that the prediction error image is input to the switch 208 . When the prediction error signal of the YCoCr space is used, the switch 206 is set so that the prediction error image is input to the inverse color space transformer 207 . Moreover, the switch 206 is able to recognize whether the prediction error signal of the RGB space or the prediction error signal of the YCoCr space should be used according to signaling from the video coding device.
  • the inverse color space transformer 207 transforms the prediction error signal of the YCoCr space to a signal of the RGB space by using the aforementioned equation (1) (the backward color space transform matrix) and then outputs the prediction error signal to the switch 208 .
  • the switch 208 selects a prediction error signal received from the switch 206 .
  • the switch 208 selects the prediction error signal received from the inverse color space transformer 207 .
  • the switch 208 is able to recognize whether the prediction error signal of the RGB space or the prediction error signal of the YCoCr space should be used according to signaling from the video coding device.
  • the prediction error image from the switch 208 is added to a prediction signal supplied from the predictor 210 at the adder 216 and then supplied as a reconstructed image to the buffer 209 .
  • the buffer 209 stores the reconstructed image.
  • the reconstructed image stored in the buffer 209 is output as a decoded image (decoded video).
  • the buffer 209 stores images, which have been decoded in the past, as reference images.
  • the predictor 210 When performing the intra prediction, the predictor 210 generates a prediction image by predicting an image to be decoded on the basis of the adjacent reconstructed images, which have been decoded in the past within the image currently under decoding.
  • the predictor 210 When performing the inter-prediction, the predictor 210 generates a prediction image on the basis of reference images supplied from the buffer 209 .
  • a chroma quantization offset (a chroma QP offset) technique for a purpose of subjective image quality improvement.
  • the chroma quantization offset technique is used to adjust a quantization parameter for each color component by signaling chroma quantization offset values for a second color component and a third color component. Specifically, this technique is able to change the quantization strength.
  • Slice unit slice_qp_delta_cb/slice_qp_delta_cr
  • Block unit cu_chroma_qp_offset_idx
  • the subjective image quality is able to be increased by adjusting the quantization strength for each color component by using any of the above syntaxes.
  • the video coding device illustrated in FIG. 18 and the video decoding device illustrated in FIG. 19 also apply a chroma quantization offset. As illustrated in FIG. 18 , a previously determined chroma quantization offset is input to the video coding device.
  • the frequency transformer/quantizer 104 increases or decreases the quantization parameter of the B component according to the first chroma quantization offset and increases or decreases the quantization parameter of the R component according to the second chroma quantization offset as illustrated in FIG. 20 when quantizing a coefficient image when the prediction error signal of the RGB space is used.
  • the inverse quantizer/inverse frequency transformer 105 increases or decreases the inverse quantization parameter of the B component according to the first chroma quantization offset and increases or decreases the inverse quantization parameter of the R component according to the second chroma quantization offset.
  • the frequency transformer/quantizer 104 increases or decreases the quantization parameter of the Co component according to the first chroma quantization offset and increases or decreases the quantization parameter of the Cr component according to the second chroma quantization offset as illustrated in FIG. 20 when quantizing a coefficient image.
  • the inverse quantizer/inverse frequency transformer 105 increases or decreases the inverse quantization parameter of the Co component according to the first chroma quantization offset and increases or decreases the inverse quantization parameter of the Cr component according to the second chroma quantization offset.
  • the inverse quantizer/inverse frequency transformer 205 operates in the same manner as the inverse quantizer/inverse frequency transformer 105 in the video coding device.
  • the chroma quantization offset technique is a technique of signaling the chroma quantization offset values for the second color component and the third color component. Therefore, if the adaptive color transform in residual domain is combined with the chroma quantization offset, the block compressed in the RGB space and the block compressed in the YCoCr space share the quantization strength as illustrated in FIG. 20 . Thereby, the quantization strength cannot be set appropriately according to the color space. Thus, the subjective image quality improvement effect cannot be acquired by the chroma quantization offset technique.
  • a video coding device capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the device including: adaptive chroma quantization offset derivation means for deriving a chroma quantization offset for each color space; and inverse quantization means for inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
  • a video decoding device capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the device including: adaptive chroma quantization offset derivation means for deriving a chroma quantization offset for each color space; and inverse quantization means for inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
  • a video coding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the method including: deriving a chroma quantization offset for each color space; and inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
  • a video decoding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the method including: deriving a chroma quantization offset for each color space; and inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
  • a video coding program for performing a video coding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the program causing a computer to perform: a process of deriving a chroma quantization offset for each color space; and a process of inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
  • a video decoding program for performing a video decoding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the program causing a computer to perform: a process of deriving a chroma quantization offset for each color space; and a process of inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
  • FIG. 1 is a block diagram illustrating an exemplary embodiment of a video coding device.
  • FIG. 2 is a flowchart illustrating processing related to signaling of chroma quantization offsets.
  • FIG. 3 is a block diagram illustrating an exemplary embodiment of a video decoding device.
  • FIG. 4 is a flowchart illustrating processing related to derivation of the chroma quantization offsets.
  • FIG. 5 is an explanatory diagram illustrating an example of a syntax for transmitting alt_pps_cb_qp_offset and alt_pps_cr_qp_offset.
  • FIG. 6 is an explanatory diagram illustrating an example of a syntax for transmitting alt_slice_qp_delta_cb and alt_slice_qp_delta_cr.
  • FIG. 7 is an explanatory diagram illustrating an example of a syntax for transmitting alt_slice_qp_delta_cb and alt_slice_qp_delta_cr.
  • FIG. 8 is an explanatory diagram illustrating an example of a syntax for transmitting cb_qp_offset_list [i] and cr_qp_offset_list [i].
  • FIG. 9 is an explanatory diagram illustrating an example of a syntax for transmitting alt_cb_qp_offset_list [i] and alt_cr_qp_offset_list [i].
  • FIG. 10 is a block diagram illustrating a configuration example of an information processing system capable of implementing the functions of the video coding device and the video decoding device.
  • FIG. 11 is a block diagram illustrating the main part of the video coding device.
  • FIG. 12 is a block diagram illustrating another example of the main part of the video coding device.
  • FIG. 13 is a block diagram illustrating still another example of the main part of the video coding device.
  • FIG. 14 is a block diagram illustrating the main part of the video decoding device.
  • FIG. 15 is a block diagram illustrating another example of the main part of the video decoding device.
  • FIG. 16 is a block diagram illustrating still another example of the main part of the video decoding device.
  • FIG. 17 is an explanatory diagram illustrating an example of an adaptive color transform in residual domain.
  • FIG. 18 is a block diagram illustrating the configuration of a general video coding device.
  • FIG. 19 is a block diagram illustrating the configuration of a general video decoding device.
  • FIG. 20 is an explanatory diagram illustrating an example of the use of chroma quantization offsets.
  • FIG. 1 is a block diagram illustrating a first exemplary embodiment of a video coding device. Referring to FIG. 1 , description will be made on the configuration of a video coding device which outputs a bitstream with each frame of a digitized video as an input image.
  • the video coding device of the first exemplary embodiment includes a switch 101 , a color space transformer 102 , a switch 103 , a frequency transformer/quantizer 104 , an inverse quantizer/inverse frequency transformer 105 , a switch 106 , an inverse color space transformer 107 , a switch 108 , a buffer 109 , a predictor 110 , a prediction parameter determiner 111 , an entropy encoder 112 , a subtractor 115 , and an adder 116 , similarly to the general video coding device illustrated in FIG. 18 .
  • the video coding device further includes an adaptive chroma quantization offset derivation unit 121 and a switch 122 .
  • the switch 101 , the color space transformer 102 , the switch 103 , the frequency transformer/quantizer 104 , the inverse quantizer/inverse frequency transformer 105 , the switch 106 , the inverse color space transformer 107 , the switch 108 , the buffer 109 , the predictor 110 , the subtractor 115 , and the adder 116 operate in the same manner as those illustrated in FIG. 18 , the following mainly describes the operations of the adaptive chroma quantization offset derivation unit 121 and the switch 122 and the operations of the prediction parameter determiner 111 and the entropy encoder 112 related to the signaling of the chroma quantization offsets. Moreover, the adaptive chroma quantization offset derivation unit 121 receives an input of a chroma quantization offset for the RGB space and an input of a chroma quantization offset for the YCoCr space.
  • FIG. 2 is a flowchart illustrating processing related to signaling of chroma quantization offsets.
  • the video coding device signals information indicating whether the adaptive color transform in residual domain is performed or not by adaptive_color_trans_flag. Moreover, when performing the adaptive color transform in residual domain, the video coding device signals information indicating the color space of the block by cu_residual_csc_flag.
  • the entropy encoder 112 transmits the chroma quantization offset for the RGB space derived by the adaptive chroma quantization offset derivation unit 121 by using the following syntax (steps S 103 and S 104 ).
  • the entropy encoder 112 transmits the chroma quantization offset for the YCoCr space derived by the adaptive chroma quantization offset derivation unit 121 by using the following syntax (steps S 103 and S 105 ).
  • the adaptive chroma quantization offset derivation unit 121 outputs the derived chroma quantization offsets (the first chroma quantization offset and the second chroma quantization offset) for the YCoCr space to the switch 122 .
  • the adaptive chroma quantization offset derivation unit 121 outputs the derived chroma quantization offsets (the first chroma quantization offset and the second chroma quantization offset) for the RGB space to the switch 122 .
  • the adaptive chroma quantization offset derivation unit 121 recognizes whether the compression is performed in the RGB space or in the YCoCr space according to cu_residual_csc_flag.
  • the frequency transformer/quantizer 104 adjusts the quantization parameter by using the chroma quantization offset decided by the prediction parameter determiner 111 .
  • the prediction parameter determiner 111 stores the values of the chroma quantization offsets for the RGB space and the values of the chroma quantization offsets for the YCoCr space in advance, for example, and appropriately supplies the values of the chroma quantization offsets for the RGB space or the values of the chroma quantization offsets for the YCoCr space to the frequency transformer/quantizer 104 .
  • the values of the chroma quantization offsets for the RGB space and the values of the chroma quantization offsets for the YCoCr space are included in the prediction parameters supplied to the entropy encoder 112 .
  • the entropy encoder 112 signals the values of the chroma quantization offsets for the RGB space and the values of the chroma quantization offsets for the YCoCr space.
  • the video coding device explicitly signals the chroma quantization offsets. Furthermore, the video coding device signals the values of the chroma quantization offsets.
  • the operations of the video coding device other than the above operations are the same as the operations of the video coding device illustrated in FIG. 18 .
  • FIG. 3 is a block diagram illustrating the configuration of a video decoding device which acquires a decoded image by decoding the bitstream output from the video coding device which signals chroma quantization offsets. Referring to FIG. 3 , the configuration of the video decoding device of the second exemplary embodiment will be described.
  • the video decoding device of this exemplary embodiment includes an entropy decoder 212 , an inverse quantizer/inverse frequency transformer 205 , a switch 206 , an inverse color space transformer 207 , a switch 208 , a buffer 209 , a predictor 210 , and an adder 216 similarly to the general video decoding device illustrated in FIG. 19 .
  • the video decoding device further includes an adaptive chroma quantization offset derivation unit 221 and a switch 222 .
  • the inverse quantizer/inverse frequency transformer 205 Since the inverse quantizer/inverse frequency transformer 205 , the switch 206 , the inverse color space transformer 207 , the switch 208 , the buffer 209 , the predictor 210 , and the adder 216 operate in the same manner as those illustrated in FIG. 19 , the following mainly describes the operations of the adaptive chroma quantization offset derivation unit 221 and the switch 222 and the operations of the entropy decoder 212 related to derivation of the chroma quantization offsets.
  • FIG. 4 is a flowchart illustrating processing related to derivation of the chroma quantization offsets.
  • the adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets for the RGB space (step S 203 ).
  • the adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets for the RGB space (the first chroma quantization offset q Pi Cb and the second chroma quantization offset q Pi Cr ) as follows.
  • qPi Cb Clip3( ⁇ QpBdOffset C ,57,Qp Y +pps_cb_qp_offset+slice_cb_qp_offset+CuQpOffset Cb
  • qPi Cr Clip3( ⁇ QpBdOffset C ,57,Qp Y +pps_cr_qp_offset+slice_cr_qp_offset+CuQpOffset Cr ) (2)
  • Clip3 (x, y, z) is a function of clipping an input z into the [x, y] range.
  • Qp Y is a quantization parameter of a first color component
  • CuQpOffset Cb is a chroma quantization offset for each block of a second color component
  • CuQpOffset Cr is a chroma quantization offset for each block of a third color component.
  • qPi Cb corresponds to a chroma quantization offset for the B component
  • qPi Cr corresponds to a chroma quantization offset for the R component in the case of the RGB space in which the first color component is the G component, the second color component is the B component, and the third color component is the R component.
  • the adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets for the YCoCr space (the first chroma quantization offset qPi Cb and the second chroma quantization offset qPi Cr ) as described in the following equation (3).
  • qPi Cb Clip3( ⁇ QpBdOffset C ,57,Qp Y +alt_pps_cb_qp_offset+alt_slice_cb_qp_offset+CuQpOffset Cb )
  • qPi Cr Clip3( ⁇ QpBdOffset C ,57,Qp Y +alt_pps_cr_qp_offset+alt_slice_cr_qp_offset+CuQpOffset Cr ) (3)
  • quantization parameter (Qp′ Cb , Qp′ Cr ) is calculated as described in the following equation (4).
  • Qp′ Cb qP Cb +QpBdOffset C
  • Qp′ Cr qP Cr +QpBdOffset C (4)
  • ChromaArrayType is equal to 1
  • the variables qP Cb and qP Cr are set equal to the value of Qp C as specified in Predetermined Table based on the index qPi equal to qPi Cb and qPi Cr , respectively.
  • variables qP Cb and qP Cb are set equal to Min (qPi, 51), based on the index qPi equal to qPi Cb and qPi Cr , respectively.
  • the inverse quantizer/inverse frequency transformer 205 increases or decreases the quantization parameter according to the chroma quantization offsets from the adaptive chroma quantization offset derivation unit 221 when inverse-quantizing the input quantization coefficient image and outputting it as a coefficient image.
  • FIG. 5 is an explanatory diagram illustrating an example of a syntax for transmitting alt_pps_cb_qp_offset and alt_pps_cr_qp_offset (an improvement of a syntax described in “7.3.2.3.2 Picture parameter set range extensions syntax” of Non Patent Literature 1).
  • wordings written in italics indicate characteristics of this exemplary embodiment.
  • FIGS. 6 and 7 are explanatory diagrams illustrating examples of syntaxes for transmitting alt_slice_qp_delta_cb and alt_slice_qp_delta_cr (improvements of a syntax described in “7.3.6.1 General slice segment header syntax”).
  • wordings written in italics indicate characteristics of this exemplary embodiment.
  • the configuration of the video coding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 1 .
  • the entropy encoder 112 transmits information by which the chroma quantization offsets for the RGB space are able to be identified (for example, an index specifying a data table in which chroma quantization offsets retained in the video decoding device are set or values of the chroma quantization offsets) to the video decoding device.
  • the entropy encoder 112 When data is compressed in the YCoCr space, the entropy encoder 112 signals the information by which the chroma quantization offsets for the YCoCr space are able to be identified (for example, the values of the chroma quantization offsets themselves) by using the syntaxes illustrated in FIG. 5 and FIGS. 6 and 7 .
  • the video decoding device of this exemplary embodiment corresponds to the video coding device of the third exemplary embodiment. Note that the configuration of the video decoding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 3 .
  • the adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets in the same manner as in the second exemplary embodiment.
  • the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221 .
  • FIG. 8 is an explanatory diagram illustrating an example of a syntax for additionally transmitting cb_qp_offset_list [i] and cr_qp_offset_list [i] for the YCoCr space (an improvement of a syntax described in “7.3.2.3.2 Picture parameter set range extensions syntax” of NPL 1).
  • FIG. 8 is an explanatory diagram illustrating an example of a syntax for additionally transmitting cb_qp_offset_list [i] and cr_qp_offset_list [i] for the YCoCr space (an improvement of a syntax described in “7.3.2.3.2 Picture parameter set range extensions syntax” of NPL 1).
  • wordings written in italics indicate characteristics of this exemplary embodiment (specifically, the parts indicate that the size of cb_qp_offset_list/cr_qp_offset_list [the range of chroma_qp_offset_list_len_minus1] is increased according to the value of adaptive_color_trans_flag).
  • the quantization offsets are able to be switched between those for the RGB space and those for the YCoCr space in block units by adjusting the value of the cu_chroma_qp_offset_idx syntax transmitted in block units according to the value of the cu_residual_csc_flag syntax.
  • the configuration of the video coding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 1 .
  • the entropy encoder 112 transmits information by which the chroma quantization offsets for the RGB space are able to be identified (for example, a cu_chroma_qp_offset_idx syntax which is an index for specifying a data table in which chroma quantization offsets retained in the video decoding device are set) to the video decoding device.
  • the entropy encoder 112 transmits information by which the chroma quantization offsets for the YCoCr space are able to be identified (for example, a cu_chroma_qp_offset_idx syntax which is an index for specifying a data table in which chroma quantization offsets retained in the video decoding device are set) to the video decoding device.
  • information by which the chroma quantization offsets for the YCoCr space are able to be identified for example, a cu_chroma_qp_offset_idx syntax which is an index for specifying a data table in which chroma quantization offsets retained in the video decoding device are set
  • the chroma quantization offsets are able to be switched between those for the RGB space and those for the YCoCr space in block units on the basis of the value of the cu_chroma_qp_offset_idx syntax transmitted in block units according to the value of the cu_residual_csc_flag syntax.
  • wordings written in italics correspond to the chroma quantization offsets for the YCoCr space described above in FIG. 8 .
  • the video decoding device of this exemplary embodiment corresponds to the video coding device of the fifth exemplary embodiment. Moreover, the configuration of the video decoding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 3 .
  • the chroma quantization offsets are read from the data table specified by the index, for example, and the adaptive chroma quantization offset derivation unit 221 calculates the chroma quantization parameter in the same manner as in the second exemplary embodiment.
  • the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221 .
  • FIG. 9 is an explanatory diagram illustrating an example of a syntax for transmitting alt_cb_qp_offset_list [i] and alt_cr_qp_offset_list [i] for the YCoCr space (an improvement of a syntax described in “7.3.2.3.2 Picture parameter set range extensions syntax” described in NPL 1).
  • wordings written in italics indicate characteristics of this exemplary embodiment.
  • the interpretation of the value of the cu_chroma_qp_offset_idx syntax changes according to the value of the cu_residual_csc_flag syntax in comparison with the fifth exemplary embodiment as described later, and therefore the bits of the cu_chroma_qp_offset_idx syntax transmitted for each block are able to be saved.
  • the configuration of the video coding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 1 .
  • the entropy encoder 112 transmits information by which the chroma quantization offsets for the RGB space are able to be identified (for example, an index for specifying a data table in which the chroma quantization offsets retained in the video decoding device are set) to the video decoding device.
  • the entropy encoder 112 transmits information by which the chroma quantization offsets for the YCoCr space are able to be identified (for example, an index for specifying a data table in which the chroma quantization offsets retained in the video decoding device are set) to the video decoding device.
  • the video decoding device of this exemplary embodiment corresponds to the video coding device of the seventh exemplary embodiment.
  • the configuration of the video decoding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 3 .
  • the chroma quantization offsets are read from the data table specified by the index, for example, and the adaptive chroma quantization offset derivation unit 221 calculates the chroma quantization parameter in the same manner as in the second exemplary embodiment.
  • the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221 .
  • cu_chroma_qp_offset_idx when present, specifies the index into the cb_qp_offset_list [ ] and cr_qp_offset_list [ ] or the alt_cb_qp_offset_list [ ] and alt_cr_qp_offset_list [ ] that is used to determine the value of CuQpOffsetCb and CuQpOffsetCr.
  • the value of cu_chroma_qp_offset_idx shall be in the range of 0 to chroma_qp_offset_list_len_minus1, inclusive.
  • the value of cu_chroma_qp_offset_idx is inferred to be equal to 0.
  • cu_chroma_qp_offset_flag when cu_chroma_qp_offset_flag is present, the following applies:
  • variable IsCuChromaQpOffsetCoded is set equal to 1.
  • CuQpOffsetCb and CuQpOffsetCr are derived as follows:
  • CuQpOffsetCb cb_qp_offset_list [cu_chroma_qp_offset_idx]
  • CuQpOffsetCr cr_qp_offset_list [cu_chroma_qp_offset_idx]
  • the video coding device explicitly signals the chroma quantization offsets in the above exemplary embodiments, the signaling of the chroma quantization offsets may be omitted though it is signaled that the color space of the prediction error signal is selected in block units. In the present description, the above is referred to as chroma quantization offsets are implicitly signaled.
  • the adaptive chroma quantization offset derivation unit 221 reads out the values of the chroma quantization offsets for the RGB space previously stored in the video decoding device.
  • the adaptive chroma quantization offset derivation unit 221 calculates the values of the chroma quantization offsets for the YCoCr space from the values of the previously-stored chroma quantization offsets for the RGB space.
  • the adaptive chroma quantization offset derivation unit 221 is able to derive the chroma quantization offsets for the YCoCr space by using the calculation formula.
  • the video decoding device implicitly derives the chroma quantization offsets.
  • the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221 .
  • the video coding device implicitly signals chroma quantization offsets, the amount of data to be transmitted is able to be reduced.
  • the RGB space and the YCoCr space are illustrated as two color spaces in the above exemplary embodiments, the system of the above exemplary embodiments is applicable even if one or both of the two color spaces are other color spaces.
  • the first color component G, the second color component B, and the third color component R are used in the RGB space in the above exemplary embodiments (see FIG. 20 ), the way of allocating color signals to the respective color components is not limited thereto, but arbitrary color signals can be allocated to the respective color components.
  • the video coding device and the video decoding device use two color spaces in the above exemplary embodiments, these devices are also allowed to use three or more color spaces.
  • each of the exemplary embodiments described above may be realized by hardware, but may also be realized by a computer program.
  • the information processing system illustrated in FIG. 10 includes a processor 1001 , a program memory 1002 , a storage medium 1003 for storing video data, and a storage medium 1004 for storing a bitstream.
  • the storage medium 1003 and the storage medium 1004 may be separate storage media, or storage areas included in the same storage medium.
  • a magnetic storage medium such as a hard disk is available as such a storage medium.
  • the program memory 1002 stores a program for realizing the functions of the blocks (except the block of the buffer) illustrated in each of FIGS. 1 and 3 .
  • the processor 1001 realizes the functions of the video coding device or video decoding device illustrated in FIG. 1 or FIG. 3 by executing the processes according to the program stored in the program memory 1002 .
  • FIG. 11 is a block diagram illustrating the main part of the video coding device.
  • the video coding device 301 includes an adaptive chroma quantization offset derivation unit 311 (corresponding to the adaptive chroma quantization offset derivation unit 121 illustrated in FIG. 1 , for example) which derives chroma quantization offsets for each color space and an inverse quantization unit 312 (corresponding to the inverse quantizer/inverse frequency transformer 105 illustrated in FIG. 1 , for example) which inverse-quantizes a quantization coefficient image by using the chroma quantization offsets for each color space.
  • an adaptive chroma quantization offset derivation unit 311 corresponding to the adaptive chroma quantization offset derivation unit 121 illustrated in FIG. 1 , for example
  • an inverse quantization unit 312 corresponding to the inverse quantizer/inverse frequency transformer 105 illustrated in FIG. 1 , for example
  • FIG. 12 is a block diagram illustrating another example of the main part of the video coding device.
  • the video coding device 302 further includes a color space selection notification unit 313 (corresponding to the entropy encoder 112 illustrated in FIG. 1 , for example) which signals that the color space of a prediction error signal is to be selected in block units.
  • the video coding device 302 includes a means for signaling information by which the values of the quantization offsets for each color space are able to be identified in the configuration illustrated in FIG. 12 , the chroma quantization offsets are implicitly derived in the video coding device 302 .
  • FIG. 13 is a block diagram illustrating still another example of the main part of the video coding device.
  • the video coding device 303 further includes a quantization offset information transmission unit 314 (corresponding to the entropy encoder 112 illustrated in FIG. 1 , for example) which signals information by which the values of the chroma quantization offsets for each color space are able to be identified.
  • the information by which the values of chroma quantization offsets are able to be identified is the values themselves of the chroma quantization offsets or an index for specifying the data table in which the chroma quantization offsets retained in the video decoding device are set, for example.
  • FIG. 14 is a block diagram illustrating the main part of the video decoding device.
  • the video decoding device 401 includes an adaptive chroma quantization offset derivation unit 411 (corresponding to the adaptive chroma quantization offset derivation unit 221 illustrated in FIG. 3 , for example) which derives chroma quantization offsets for each color space and an inverse quantization unit 412 (corresponding to the inverse quantizer/inverse frequency transformer 205 illustrated in FIG. 3 , for example) which inverse-quantizes a quantization coefficient image by using the chroma quantization offsets for each color space.
  • an adaptive chroma quantization offset derivation unit 411 corresponding to the adaptive chroma quantization offset derivation unit 221 illustrated in FIG. 3 , for example
  • an inverse quantization unit 412 corresponding to the inverse quantizer/inverse frequency transformer 205 illustrated in FIG. 3 , for example
  • FIG. 15 is a block diagram illustrating another example of the main part of the video decoding device.
  • the video decoding device 402 further includes a color space selection parsing unit 413 (corresponding to the entropy decoder 212 illustrated in FIG. 3 , for example) which parses a bitstream to interpret that the color space of a prediction error signal is to be selected in block units.
  • a color space selection parsing unit 413 corresponding to the entropy decoder 212 illustrated in FIG. 3 , for example
  • the video decoding device 402 includes a means for parsing a bitstream to interpret information by which the values of the chroma quantization offsets for each color space are able to be identified in the configuration illustrated in FIG. 15 , the video decoding device 402 implicitly derives the chroma quantization offset.
  • FIG. 16 is a block diagram illustrating still another example of the main part of the video decoding device.
  • the video decoding device 403 further includes a chroma quantization offset parsing unit 414 (corresponding to the entropy decoder 212 illustrated in FIG. 3 , for example) which identifies the values of the chroma quantization offsets for each color space on the basis of the information interpreted from the received bitstream.
  • a chroma quantization offset parsing unit 414 corresponding to the entropy decoder 212 illustrated in FIG. 3 , for example

Abstract

A video coding device is able to select, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, and includes: an adaptive chroma quantization offset derivation unit for deriving a quantization offset for each color space; and an inverse quantization unit for inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.

Description

CROSS-REFERENCE TO RELATED PATENT APPLICATIONS
This application is a National Stage Entry of International Application No. PCT/JP2015/004037, filed Aug. 12, 2015, which claims priority from Japanese Patent Application No. 2014-204392, filed Oct. 3, 2014. The entire contents of the above-referenced applications are expressly incorporated herein by reference.
TECHNICAL FIELD
The present invention relates to a video coding device and a video decoding device using an adaptive color transform in residual domain and a chroma (color difference) quantization offset.
BACKGROUND ART
In a video coding system based on high efficiency video coding (HEVC)/H.265, each frame of a digitized image is divided into coding tree units (CTUs) and the respective CTUs are coded in the order of raster scan. The CTU is divided into coding units (CUs) in a quad-tree structure and then coded. Each CU is divided into prediction units (PUs) and predicted. Moreover, a prediction error of each CU is divided into transform units (TUs) in a quad-tree structure and then frequency-transformed.
The CU is a coding unit for an intra prediction or inter-frame prediction.
The intra prediction (intra-frame prediction) is prediction for generating a prediction signal from a reconstructed image of a frame to be encoded. In HEVC/H.265, 33 types of angular intra prediction and the like are defined. In angular intra prediction, a reconstructed pixel around a block to be encoded is extrapolated in any of 33 directions depicted in FIG. 14, to generate an intra prediction signal.
In addition to angular intra prediction, DC prediction and planar prediction are specified as intra prediction. In DC prediction, a mean value of a reference image is used as the prediction values of all pixels in a TU to be predicted. In planar prediction, a prediction image is generated by linear interpolation from pixels in a reference image.
Inter-frame prediction is a prediction based on an image of a reconstructed frame (reference picture) different in display time from a frame to be encoded. Inter-frame prediction is also referred to as inter prediction. In inter prediction, an inter prediction signal is generated based on a reconstructed image block of a reference picture (using pixel interpolation if necessary).
A digital color image is composed of RGB digital images. A digital color image is generally transformed to signals in a color space other than the RGB space to increase the compression efficiency (to decrease the data amount) when the color image is transmitted through a transmission line. For example, the image signal is transformed to a signal in a color space (YCoCr space) where a combination of a luma signal (Y) and chroma signals (Cb, Cr) are composed.
A quantization parameter (QP) for the chroma signal is generated by transforming the QP for the luma signal by using an offset value “chroma_qp_index_offset.” In HEVC, cb_qp_index_offset (a first chroma quantization offset) is applied to Cb, and cr_qp_index_offset (a second chroma quantization offset) is applied to Cr.
In RExt (range extension) of HEVC, the standardization of an extension function was carried out (see Non Patent Literature 1).
As a method for further increasing the compression efficiency of the extension function using RExt, there has been proposed a technique called “adaptive color transform in residual domain” in Non Patent Literature 2. As illustrated in FIG. 17, the adaptive color transform in residual domain is a technique of switching a prediction error signal of an image signal of the RGB space to a signal of the YCoCr space adaptively in block units (per block).
Specifically, it is possible to select, in block units (per block), whether to compress the prediction error signal of the RGB space directly or to transform the prediction error signal to a signal of the YCoCr space by using a forward color space transform matrix (see equation (1)) described below before compression. Moreover, FIG. 17 illustrates an example that data is compressed in the YCoCr space with respect to the shaded blocks and data is compressed in the RGB space with respect to other blocks.
Information on the color space used for the data compression of blocks is signaled by a cu_residual_csc_flag syntax. Equation “cu_residual_csc_flag=0” represents that the signal of the RGB space is compressed, while “cu_residual_csc_flag=1” represents that the signal is compressed after it is transformed to the YCoCr space.
If cu_residual_csc_flag=1, a receiver (a video decoding device) restores the signal of the YCoCr space to the signal of the RGB space by using a backward color space transform matrix described below and then performs decoding processing.
[ Math . 1 ] Forward : [ Y Cb Cr ] = [ 1 2 1 2 0 - 2 - 1 2 - 1 ] [ R G B ] / 4 Backward : [ R G B ] = [ 1 1 - 1 1 0 1 1 - 1 - 1 ] [ Y Cb Cr ] / 4 ( 1 )
Moreover, the norm is not constant in the aforementioned color transform matrix and therefore, when cu_residual_csc_flag=1, a different chroma quantization offset is added to the quantization parameter for each YCoCr component in quantization processing and inverse quantization processing of a block prediction error signal.
Moreover, Patent Literature 1 describes a video coding device and a video decoding device which perform signal processing which is different according to whether an input image signal is a signal of the RGB space or a signal of the YCoCr space. Specifically, when performing a weighted prediction based on the H.264/AVC, the video coding device applies the same offset to the R, G, and B signals and to the luma signal (Y signal) and applies different offsets to the chroma signals, regarding an offset added to a prediction signal. However, Patent Literature 1 does not teach any new knowledge regarding the chroma quantization offset.
CITATION LIST Patent Literature
  • Patent Literature 1: Japanese Patent Application Laid-Open No. 2011-151683
Non Patent Literatures
  • Non Patent Literature 1: D. Flynn, et al., “High Efficiency Video Coding (HEVC) Range Extensions text specification: Draft 7”, JCTVC-Q1005, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 17th Meeting: Valencia, ES, 27 March-4 Apr. 2014
  • Non Patent Literature 2: L. Zhang et al., “SCCE5 Test 3.2.1: In-loop color-space transform”, JCTVC-R0147, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 18th Meeting: Sapporo, JP, 30 June-9 Jul. 2014
SUMMARY OF INVENTION Technical Problem
Next, referring to FIG. 18, the following describes the configuration and operation of a general video coding device which outputs a bitstream with each CU of each frame of a digitized image as an input image.
A video coding device illustrated in FIG. 18 includes a switch 101, a color space transformer 102, a switch 103, a frequency transformer/quantizer 104, an inverse quantizer/inverse frequency transformer 105, a switch 106, an inverse color space transformer 107, a switch 108, a buffer 109, a predictor 110, a prediction parameter determiner 111, an entropy encoder 112, a subtractor 115, and an adder 116.
The predictor 110 generates a prediction signal for a CU input image signal. Specifically, the predictor 110 generates a prediction signal (an intra-prediction signal) on the basis of intra prediction and generates a prediction signal (an inter-prediction signal) on the basis of inter-prediction.
A prediction image supplied from the predictor 110 is subtracted from an image input to the video coding device at the subtractor 115 and then input to the switch 101 as a prediction error image. In the example illustrated in FIG. 18, an input image signal is a signal of the RGB space. Moreover, the video coding device has a function of an adaptive color transform in residual domain. For example, the video coding device is able to adaptively switch a prediction error signal of an image signal of the RGB space to a signal of the YCoCr space in block units (per block).
When a prediction error signal of the RGB space is used, the switch 101 is set so that the prediction error image is input to the switch 103. When the prediction error signal of the YCoCr space is used, the switch 101 is set so that the prediction error image is input to the color space transformer 102. Moreover, the switch 101 sets an output destination of a prediction error image according to the control of the prediction parameter determiner 111, for example.
The color space transformer 102 transforms the prediction error signal of the RGB space to a signal of the YCoCr space by using the aforementioned equation (1) (a forward color space transform matrix) and then outputs the signal to the switch 103.
When a prediction error signal of the RGB space is used, the switch 103 outputs a prediction error signal, which has been received from the switch 101, to the frequency transformer/quantizer 104. When a prediction error signal of the YCoCr space is used, the switch 103 outputs a prediction error signal, which has been received from the color space transformer 102, to the frequency transformer/quantizer 104. Moreover, the switch 103 selects the input source of the prediction error image according to the control of the prediction parameter determiner 111, for example.
The frequency transformer/quantizer 104 frequency-transforms the prediction error image and quantizes the frequency-transformed prediction error image (a coefficient image). The entropy encoder 112 performs entropy coding of a prediction parameter and of a quantization coefficient image and then outputs a bitstream
The inverse quantizer/inverse frequency transformer 105 inverse-quantizes the quantization coefficient image. Further, the inverse quantizer/inverse frequency transformer 105 performs inverse frequency transform on the inverse-quantized coefficient image. The inverse frequency transformed reconstructed prediction error image is input to the switch 106.
When the prediction error signal of the RGB space is used, the switch 106 is set so that the reconstructed prediction error image is input to the switch 108. When the prediction error signal of the YCoCr space is used, the switch 106 is set so that the reconstructed prediction error image is input to the inverse color space transformer 107. The switch 106 selects an output destination of the reconstructed prediction error image according to the control of the prediction parameter determiner 111.
The inverse color space transformer 107 transforms a reconstructed prediction error signal of the YCoCr space to a signal of the RGB space by using the aforementioned equation (1) (a backward color space transform matrix) and then outputs the to the switch 108.
The switch 108 selects the reconstructed prediction error signal received from the switch 106 when the prediction error signal of the RGB space is used. When the prediction error signal of the YCoCr space is used, the switch 108 selects the reconstructed prediction error signal received from the inverse color space transformer 107. Moreover, the switch 108 selects any of the reconstructed prediction error images according to the control of the prediction parameter determiner 111, for example.
The reconstructed prediction error image received from the switch 108 is supplied with a prediction signal by the adder 116 and then supplied as a reconstructed image to the buffer 109. The buffer 109 stores the reconstructed image.
The prediction parameter determiner 111 instructs the predictor 110 to decide the prediction parameter which minimizes coding cost by comparing the input image signal with the prediction signal, for example. The prediction parameter determiner 111 supplies the decided prediction parameter to the entropy encoder 112. The prediction parameter is information related to block prediction such as a prediction mode (intra prediction, inter-prediction), an intra prediction block size, an intra prediction direction, an inter-prediction block size, a motion vector, or the like.
The prediction parameter determiner 111 further instructs to decide whether the prediction error signal of the RGB space or the prediction error signal of the YCoCr space is used for each block.
The bitstream output from the video coding device is transmitted to the video decoding device. The video decoding device reconstructs images of a video by performing decoding processing. FIG. 19 is a block diagram illustrating an example of the configuration of a general video decoding device which decodes the bitstream output from a general video coding device to acquire a decoded image. Referring to FIG. 19, the configuration and operation of the general video decoding device will be described below.
The video decoding device illustrated in FIG. 19 includes an entropy decoder 212, an inverse quantizer/inverse frequency transformer 205, a switch 206, an inverse color space transformer 207, a switch 208, a buffer 209, a predictor 210, and an adder 216.
The entropy decoder 212 entropy-decodes an input bitstream. The entropy decoder 212 supplies a quantization coefficient image to the inverse quantizer/inverse frequency transformer 205 and supplies a prediction parameter to the predictor 210.
The inverse quantizer/inverse frequency transformer 205 inverse-quantizes the input quantization coefficient image and outputs it as a coefficient image. Furthermore, the inverse quantizer/inverse frequency transformer 205 transforms the coefficient image in the frequency domain to an image in the space domain and outputs it as a prediction error image. The prediction error image is input to the switch 206.
When the prediction error signal of the RGB space is used, the switch 206 is set so that the prediction error image is input to the switch 208. When the prediction error signal of the YCoCr space is used, the switch 206 is set so that the prediction error image is input to the inverse color space transformer 207. Moreover, the switch 206 is able to recognize whether the prediction error signal of the RGB space or the prediction error signal of the YCoCr space should be used according to signaling from the video coding device.
The inverse color space transformer 207 transforms the prediction error signal of the YCoCr space to a signal of the RGB space by using the aforementioned equation (1) (the backward color space transform matrix) and then outputs the prediction error signal to the switch 208.
When the prediction error signal of the RGB space is used, the switch 208 selects a prediction error signal received from the switch 206. When the prediction error signal of the YCoCr space is used, the switch 208 selects the prediction error signal received from the inverse color space transformer 207. The switch 208 is able to recognize whether the prediction error signal of the RGB space or the prediction error signal of the YCoCr space should be used according to signaling from the video coding device.
The prediction error image from the switch 208 is added to a prediction signal supplied from the predictor 210 at the adder 216 and then supplied as a reconstructed image to the buffer 209. The buffer 209 stores the reconstructed image.
Moreover, the reconstructed image stored in the buffer 209 is output as a decoded image (decoded video).
The buffer 209 stores images, which have been decoded in the past, as reference images. When performing the intra prediction, the predictor 210 generates a prediction image by predicting an image to be decoded on the basis of the adjacent reconstructed images, which have been decoded in the past within the image currently under decoding. When performing the inter-prediction, the predictor 210 generates a prediction image on the basis of reference images supplied from the buffer 209.
In RExt, a chroma quantization offset (a chroma QP offset) technique for a purpose of subjective image quality improvement has been proposed. The chroma quantization offset technique is used to adjust a quantization parameter for each color component by signaling chroma quantization offset values for a second color component and a third color component. Specifically, this technique is able to change the quantization strength.
The following syntaxes are provided for signaling chroma quantization offset values:
Picture unit: pps_cb_qp_offset/pps_cr_qp_offset/slice_qp_delta_cb/slice_qp_delta_cr
Slice unit: slice_qp_delta_cb/slice_qp_delta_cr
Block unit: cu_chroma_qp_offset_idx
The subjective image quality is able to be increased by adjusting the quantization strength for each color component by using any of the above syntaxes.
The video coding device illustrated in FIG. 18 and the video decoding device illustrated in FIG. 19 also apply a chroma quantization offset. As illustrated in FIG. 18, a previously determined chroma quantization offset is input to the video coding device.
In the video coding device, the frequency transformer/quantizer 104 increases or decreases the quantization parameter of the B component according to the first chroma quantization offset and increases or decreases the quantization parameter of the R component according to the second chroma quantization offset as illustrated in FIG. 20 when quantizing a coefficient image when the prediction error signal of the RGB space is used. The inverse quantizer/inverse frequency transformer 105 increases or decreases the inverse quantization parameter of the B component according to the first chroma quantization offset and increases or decreases the inverse quantization parameter of the R component according to the second chroma quantization offset.
When the prediction error signal of the YCoCr space is used, the frequency transformer/quantizer 104 increases or decreases the quantization parameter of the Co component according to the first chroma quantization offset and increases or decreases the quantization parameter of the Cr component according to the second chroma quantization offset as illustrated in FIG. 20 when quantizing a coefficient image. The inverse quantizer/inverse frequency transformer 105 increases or decreases the inverse quantization parameter of the Co component according to the first chroma quantization offset and increases or decreases the inverse quantization parameter of the Cr component according to the second chroma quantization offset.
In the video decoding device, the inverse quantizer/inverse frequency transformer 205 operates in the same manner as the inverse quantizer/inverse frequency transformer 105 in the video coding device.
The chroma quantization offset technique is a technique of signaling the chroma quantization offset values for the second color component and the third color component. Therefore, if the adaptive color transform in residual domain is combined with the chroma quantization offset, the block compressed in the RGB space and the block compressed in the YCoCr space share the quantization strength as illustrated in FIG. 20. Thereby, the quantization strength cannot be set appropriately according to the color space. Thus, the subjective image quality improvement effect cannot be acquired by the chroma quantization offset technique.
It is an object of the present invention to provide a video coding device, a video decoding device, a video coding method, a video decoding method, and a program capable of preventing a deterioration in the subjective image quality improvement effect in the case where the adaptive color transform in residual domain and the chroma quantization offset are used in combination.
Solution to Problem
According to the present invention, there is provided a video coding device capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the device including: adaptive chroma quantization offset derivation means for deriving a chroma quantization offset for each color space; and inverse quantization means for inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
Furthermore, according to the present invention, there is provided a video decoding device capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the device including: adaptive chroma quantization offset derivation means for deriving a chroma quantization offset for each color space; and inverse quantization means for inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
Furthermore, according to the present invention, there is provided a video coding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the method including: deriving a chroma quantization offset for each color space; and inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
Furthermore, according to the present invention, there is provided a video decoding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the method including: deriving a chroma quantization offset for each color space; and inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
Furthermore, according to the present invention, there is provided a video coding program for performing a video coding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the program causing a computer to perform: a process of deriving a chroma quantization offset for each color space; and a process of inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
Furthermore, according to the present invention, there is provided a video decoding program for performing a video decoding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the program causing a computer to perform: a process of deriving a chroma quantization offset for each color space; and a process of inverse-quantizing a quantization coefficient image by using the chroma quantization offset for each color space.
Advantageous Effects of Invention
According to the present invention, it is possible to prevent a deterioration in a subjective image quality improvement effect.
BRIEF DESCRIPTION OF DRAWINGS
FIG. 1 is a block diagram illustrating an exemplary embodiment of a video coding device.
FIG. 2 is a flowchart illustrating processing related to signaling of chroma quantization offsets.
FIG. 3 is a block diagram illustrating an exemplary embodiment of a video decoding device.
FIG. 4 is a flowchart illustrating processing related to derivation of the chroma quantization offsets.
FIG. 5 is an explanatory diagram illustrating an example of a syntax for transmitting alt_pps_cb_qp_offset and alt_pps_cr_qp_offset.
FIG. 6 is an explanatory diagram illustrating an example of a syntax for transmitting alt_slice_qp_delta_cb and alt_slice_qp_delta_cr.
FIG. 7 is an explanatory diagram illustrating an example of a syntax for transmitting alt_slice_qp_delta_cb and alt_slice_qp_delta_cr.
FIG. 8 is an explanatory diagram illustrating an example of a syntax for transmitting cb_qp_offset_list [i] and cr_qp_offset_list [i].
FIG. 9 is an explanatory diagram illustrating an example of a syntax for transmitting alt_cb_qp_offset_list [i] and alt_cr_qp_offset_list [i].
FIG. 10 is a block diagram illustrating a configuration example of an information processing system capable of implementing the functions of the video coding device and the video decoding device.
FIG. 11 is a block diagram illustrating the main part of the video coding device.
FIG. 12 is a block diagram illustrating another example of the main part of the video coding device.
FIG. 13 is a block diagram illustrating still another example of the main part of the video coding device.
FIG. 14 is a block diagram illustrating the main part of the video decoding device.
FIG. 15 is a block diagram illustrating another example of the main part of the video decoding device.
FIG. 16 is a block diagram illustrating still another example of the main part of the video decoding device.
FIG. 17 is an explanatory diagram illustrating an example of an adaptive color transform in residual domain.
FIG. 18 is a block diagram illustrating the configuration of a general video coding device.
FIG. 19 is a block diagram illustrating the configuration of a general video decoding device.
FIG. 20 is an explanatory diagram illustrating an example of the use of chroma quantization offsets.
DESCRIPTION OF EMBODIMENT Exemplary Embodiment 1
FIG. 1 is a block diagram illustrating a first exemplary embodiment of a video coding device. Referring to FIG. 1, description will be made on the configuration of a video coding device which outputs a bitstream with each frame of a digitized video as an input image.
As illustrated in FIG. 1, the video coding device of the first exemplary embodiment includes a switch 101, a color space transformer 102, a switch 103, a frequency transformer/quantizer 104, an inverse quantizer/inverse frequency transformer 105, a switch 106, an inverse color space transformer 107, a switch 108, a buffer 109, a predictor 110, a prediction parameter determiner 111, an entropy encoder 112, a subtractor 115, and an adder 116, similarly to the general video coding device illustrated in FIG. 18.
As illustrated in FIG. 1, the video coding device further includes an adaptive chroma quantization offset derivation unit 121 and a switch 122.
Since the switch 101, the color space transformer 102, the switch 103, the frequency transformer/quantizer 104, the inverse quantizer/inverse frequency transformer 105, the switch 106, the inverse color space transformer 107, the switch 108, the buffer 109, the predictor 110, the subtractor 115, and the adder 116 operate in the same manner as those illustrated in FIG. 18, the following mainly describes the operations of the adaptive chroma quantization offset derivation unit 121 and the switch 122 and the operations of the prediction parameter determiner 111 and the entropy encoder 112 related to the signaling of the chroma quantization offsets. Moreover, the adaptive chroma quantization offset derivation unit 121 receives an input of a chroma quantization offset for the RGB space and an input of a chroma quantization offset for the YCoCr space.
FIG. 2 is a flowchart illustrating processing related to signaling of chroma quantization offsets.
The video coding device signals information indicating whether the adaptive color transform in residual domain is performed or not by adaptive_color_trans_flag. Moreover, when performing the adaptive color transform in residual domain, the video coding device signals information indicating the color space of the block by cu_residual_csc_flag.
Unless the adaptive color transform in residual domain is performed, the entropy encoder 112 signals adaptive_color_trans_flag=0 and transmits the chroma quantization offset for the RGB space derived by the adaptive chroma quantization offset (the chroma quantization offset for the RGB space input by the adaptive chroma quantization offset derivation unit 121) derivation unit 121 by using the following syntax (steps S101 and S102). If the adaptive color transform in residual domain is performed, the entropy encoder 112 sets adaptive_color_trans_flag=1. In addition, when compression is performed in the RGB space, the entropy encoder 112 transmits the chroma quantization offset for the RGB space derived by the adaptive chroma quantization offset derivation unit 121 by using the following syntax (steps S103 and S104).
    • Picture unit: pps_cb_qp_offset/pps_cr_qp_offset/slice_qp_delta_cb/slice_qp_delta_cr
    • Slice unit: slice_qp_delta_cb/slice_qp_delta_cr
When the compression is performed in the YCoCr space, the entropy encoder 112 transmits the chroma quantization offset for the YCoCr space derived by the adaptive chroma quantization offset derivation unit 121 by using the following syntax (steps S103 and S105).
    • Picture unit: alt_pps_cb_qp_offset/alt_pps_cr_qp_offset/alt_slice_qp_delta_cb/alt_slice_qp_delta_cr
    • Slice unit: alt_slice_qp_delta_cb/alt_slice_qp_delta_cr
Moreover, in the case where the adaptive color transform in residual domain is performed, the entropy encoder 112 signals cu_residual_csc_flag=1 when the compression is performed in the YCoCr space (when the compression is not performed in the RGB space). The adaptive chroma quantization offset derivation unit 121 outputs the derived chroma quantization offsets (the first chroma quantization offset and the second chroma quantization offset) for the YCoCr space to the switch 122.
When the compression is performed in the RGB space, the entropy encoder 112 signals cu_residual_csc_flag=0. The adaptive chroma quantization offset derivation unit 121 outputs the derived chroma quantization offsets (the first chroma quantization offset and the second chroma quantization offset) for the RGB space to the switch 122.
Moreover, the adaptive chroma quantization offset derivation unit 121 recognizes whether the compression is performed in the RGB space or in the YCoCr space according to cu_residual_csc_flag.
Moreover, the frequency transformer/quantizer 104 adjusts the quantization parameter by using the chroma quantization offset decided by the prediction parameter determiner 111.
The prediction parameter determiner 111 stores the values of the chroma quantization offsets for the RGB space and the values of the chroma quantization offsets for the YCoCr space in advance, for example, and appropriately supplies the values of the chroma quantization offsets for the RGB space or the values of the chroma quantization offsets for the YCoCr space to the frequency transformer/quantizer 104. In that case, the values of the chroma quantization offsets for the RGB space and the values of the chroma quantization offsets for the YCoCr space are included in the prediction parameters supplied to the entropy encoder 112. The entropy encoder 112 signals the values of the chroma quantization offsets for the RGB space and the values of the chroma quantization offsets for the YCoCr space.
In this case, the video coding device explicitly signals the chroma quantization offsets. Furthermore, the video coding device signals the values of the chroma quantization offsets.
Moreover, the operations of the adaptive chroma quantization offset derivation unit 121 will be described in more detail in the second exemplary embodiment.
The operations of the video coding device other than the above operations are the same as the operations of the video coding device illustrated in FIG. 18.
Exemplary Embodiment 2
FIG. 3 is a block diagram illustrating the configuration of a video decoding device which acquires a decoded image by decoding the bitstream output from the video coding device which signals chroma quantization offsets. Referring to FIG. 3, the configuration of the video decoding device of the second exemplary embodiment will be described.
As illustrated in FIG. 3, the video decoding device of this exemplary embodiment includes an entropy decoder 212, an inverse quantizer/inverse frequency transformer 205, a switch 206, an inverse color space transformer 207, a switch 208, a buffer 209, a predictor 210, and an adder 216 similarly to the general video decoding device illustrated in FIG. 19.
As illustrated in FIG. 3, the video decoding device further includes an adaptive chroma quantization offset derivation unit 221 and a switch 222.
Since the inverse quantizer/inverse frequency transformer 205, the switch 206, the inverse color space transformer 207, the switch 208, the buffer 209, the predictor 210, and the adder 216 operate in the same manner as those illustrated in FIG. 19, the following mainly describes the operations of the adaptive chroma quantization offset derivation unit 221 and the switch 222 and the operations of the entropy decoder 212 related to derivation of the chroma quantization offsets.
FIG. 4 is a flowchart illustrating processing related to derivation of the chroma quantization offsets.
If the entropy decoder 212 parses a bitstream to get adaptive_color_trans_flag=1 (indicating that the adaptive color transform in residual domain is to be performed) (step S201) and gets cu_residual_csc_flag=1 (indicating that data is compressed in the YCoCr space) (step S202), the adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets for the YCoCr space (step S204). If the entropy decoder 212 gets cu_residual_csc_flag=0 (indicating that data is compressed in the RGB space) (step S202), the adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets for the RGB space (step S203).
The adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets for the RGB space (the first chroma quantization offset qPiCb and the second chroma quantization offset qPiCr) as follows.
qPiCb=Clip3(−QpBdOffsetC,57,QpY+pps_cb_qp_offset+slice_cb_qp_offset+CuQpOffsetCb)
qPiCr=Clip3(−QpBdOffsetC,57,QpY+pps_cr_qp_offset+slice_cr_qp_offset+CuQpOffsetCr)  (2)
In equation (2), Clip3 (x, y, z) is a function of clipping an input z into the [x, y] range. QpY is a quantization parameter of a first color component, CuQpOffsetCb is a chroma quantization offset for each block of a second color component, and CuQpOffsetCr is a chroma quantization offset for each block of a third color component. Although descriptions qPiCb and qPiCr are used, qPiCb corresponds to a chroma quantization offset for the B component and qPiCr corresponds to a chroma quantization offset for the R component in the case of the RGB space in which the first color component is the G component, the second color component is the B component, and the third color component is the R component.
The adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets for the YCoCr space (the first chroma quantization offset qPiCb and the second chroma quantization offset qPiCr) as described in the following equation (3).
qPiCb=Clip3(−QpBdOffsetC,57,QpY+alt_pps_cb_qp_offset+alt_slice_cb_qp_offset+CuQpOffsetCb)
qPiCr=Clip3(−QpBdOffsetC,57,QpY+alt_pps_cr_qp_offset+alt_slice_cr_qp_offset+CuQpOffsetCr)  (3)
Moreover, quantization parameter (Qp′Cb, Qp′Cr) is calculated as described in the following equation (4).
Qp′Cb=qPCb+QpBdOffsetC
Qp′Cr=qPCr+QpBdOffsetC  (4)
The following describes a specific example of the procedure for deriving a chroma quantization offset. In the description below, wordings enclosed by quotation marks indicate characteristics in this exemplary embodiment.
    • “If cu_residual_csc_flag is equal to 0,” the variables qPCb and qPCr are derived as follows:
      qPiCb=Clip3(−QpBdOffsetC,57,QpY+pps_cb_qp_offset+slice_cb_qp_offset+CuQpOffsetCb)
      qPiCr=Clip3(−QpBdOffsetC,57,QpY+pps_cr_qp_offset+slice_cr_qp_offset+CuQpOffsetCr)
“Otherwise (cu_residual_csc_flag is equal to 1), the variables qPCb and qPCr are derived as follows:”
“qPiCb=Clip3(−QpBdOffsetC,57,QpY+alt_pps_cb_qp_offset+alt_slice_cb_qp_offset+CuQpOffsetCr)”
“qPiCr=Clip3(−QpBdOffsetC,57,QpY+alt_pps_cr_qp_offset+alt_slice_cr_qp_offset+CuQpOffsetCr)”
If ChromaArrayType is equal to 1, the variables qPCb and qPCr are set equal to the value of QpC as specified in Predetermined Table based on the index qPi equal to qPiCb and qPiCr, respectively.
Otherwise, the variables qPCb and qPCb are set equal to Min (qPi, 51), based on the index qPi equal to qPiCb and qPiCr, respectively.
The chroma quantization parameters for the Cb and Cr components, Qp′Cb and Qp′Cr, are derived as follows:
Qp′Cb=qPCb+QpBdOffsetC
Qp′Cr=qPCr+QpBdOffsetC
The inverse quantizer/inverse frequency transformer 205 increases or decreases the quantization parameter according to the chroma quantization offsets from the adaptive chroma quantization offset derivation unit 221 when inverse-quantizing the input quantization coefficient image and outputting it as a coefficient image.
Exemplary Embodiment 3
Subsequently, a video coding device according to a third exemplary embodiment will be described. FIG. 5 is an explanatory diagram illustrating an example of a syntax for transmitting alt_pps_cb_qp_offset and alt_pps_cr_qp_offset (an improvement of a syntax described in “7.3.2.3.2 Picture parameter set range extensions syntax” of Non Patent Literature 1). In FIG. 5, wordings written in italics indicate characteristics of this exemplary embodiment.
FIGS. 6 and 7 are explanatory diagrams illustrating examples of syntaxes for transmitting alt_slice_qp_delta_cb and alt_slice_qp_delta_cr (improvements of a syntax described in “7.3.6.1 General slice segment header syntax”). In FIGS. 6 and 7, wordings written in italics indicate characteristics of this exemplary embodiment.
Moreover, the configuration of the video coding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 1. In the video coding device, the entropy encoder 112 transmits information by which the chroma quantization offsets for the RGB space are able to be identified (for example, an index specifying a data table in which chroma quantization offsets retained in the video decoding device are set or values of the chroma quantization offsets) to the video decoding device.
When data is compressed in the YCoCr space, the entropy encoder 112 signals the information by which the chroma quantization offsets for the YCoCr space are able to be identified (for example, the values of the chroma quantization offsets themselves) by using the syntaxes illustrated in FIG. 5 and FIGS. 6 and 7.
Exemplary Embodiment 4
Subsequently, a video decoding device according to a fourth exemplary embodiment will be described. The video decoding device of this exemplary embodiment corresponds to the video coding device of the third exemplary embodiment. Note that the configuration of the video decoding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 3.
When the entropy decoder 212 interprets that data is compressed in the YCoCr space by using the syntaxes illustrated in FIG. 5 and FIGS. 6 and 7 in the video decoding device, the adaptive chroma quantization offset derivation unit 221 derives the chroma quantization offsets in the same manner as in the second exemplary embodiment.
Moreover, in the video coding device, the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221.
Exemplary Embodiment 5
Subsequently, a video coding device according to a fifth exemplary embodiment will be described. FIG. 8 is an explanatory diagram illustrating an example of a syntax for additionally transmitting cb_qp_offset_list [i] and cr_qp_offset_list [i] for the YCoCr space (an improvement of a syntax described in “7.3.2.3.2 Picture parameter set range extensions syntax” of NPL 1). In FIG. 8, wordings written in italics indicate characteristics of this exemplary embodiment (specifically, the parts indicate that the size of cb_qp_offset_list/cr_qp_offset_list [the range of chroma_qp_offset_list_len_minus1] is increased according to the value of adaptive_color_trans_flag). In the video coding device according to this exemplary embodiment, the quantization offsets are able to be switched between those for the RGB space and those for the YCoCr space in block units by adjusting the value of the cu_chroma_qp_offset_idx syntax transmitted in block units according to the value of the cu_residual_csc_flag syntax.
Note that the configuration of the video coding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 1. In the video coding device, the entropy encoder 112 transmits information by which the chroma quantization offsets for the RGB space are able to be identified (for example, a cu_chroma_qp_offset_idx syntax which is an index for specifying a data table in which chroma quantization offsets retained in the video decoding device are set) to the video decoding device.
According to this exemplary embodiment, in the video coding device, the entropy encoder 112 transmits information by which the chroma quantization offsets for the YCoCr space are able to be identified (for example, a cu_chroma_qp_offset_idx syntax which is an index for specifying a data table in which chroma quantization offsets retained in the video decoding device are set) to the video decoding device. In the video decoding device of this exemplary embodiment, the chroma quantization offsets are able to be switched between those for the RGB space and those for the YCoCr space in block units on the basis of the value of the cu_chroma_qp_offset_idx syntax transmitted in block units according to the value of the cu_residual_csc_flag syntax.
Moreover, wordings written in italics (cb_qp_offset_list [i] and cr_qp_offset_list [i]) correspond to the chroma quantization offsets for the YCoCr space described above in FIG. 8.
Exemplary Embodiment 6
Next, a video decoding device according to a sixth exemplary embodiment will be described. The video decoding device of this exemplary embodiment corresponds to the video coding device of the fifth exemplary embodiment. Moreover, the configuration of the video decoding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 3.
When the entropy decoder 212 interprets that data is compressed in the YCoCr space by the syntax illustrated in FIG. 8 in the video decoding device, the chroma quantization offsets are read from the data table specified by the index, for example, and the adaptive chroma quantization offset derivation unit 221 calculates the chroma quantization parameter in the same manner as in the second exemplary embodiment.
Moreover, in the video coding device, the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221.
Exemplary Embodiment 7
Next, a video coding device according to a seventh exemplary embodiment will be described. FIG. 9 is an explanatory diagram illustrating an example of a syntax for transmitting alt_cb_qp_offset_list [i] and alt_cr_qp_offset_list [i] for the YCoCr space (an improvement of a syntax described in “7.3.2.3.2 Picture parameter set range extensions syntax” described in NPL 1). In FIG. 9, wordings written in italics indicate characteristics of this exemplary embodiment.
In the seventh exemplary embodiment, the interpretation of the value of the cu_chroma_qp_offset_idx syntax changes according to the value of the cu_residual_csc_flag syntax in comparison with the fifth exemplary embodiment as described later, and therefore the bits of the cu_chroma_qp_offset_idx syntax transmitted for each block are able to be saved. For example, in the seventh exemplary embodiment, even if cu_chroma_qp_offset_idx=0, cb_qp_offset_list [0] and cr_qp_offset_list [0] for RGB are derived when cu_residual_csc_flag=0, and alt_cb_qp_offset_list [0] and alt_cr_qp_offset_list [0] for YCoCr are derived when cu_residual_csc_flag=1. On the other hand, in the fifth exemplary embodiment, when cu_chroma_qp_offset_idx=0, cb_qp_offset_list [0] and cr_qp_offset_list [0] for RGB are derived. Therefore, in the fifth exemplary embodiment, if the list size is 4 (when chroma_qp_offset_list_len_minus1 is 3), it is necessary to transmit cu_chroma_qp_offset_idx=4 in order to derive cb_qp_offset_list [4] and alt_cr_qp_offset_list [4] for YCoCr.
Moreover, the configuration of the video coding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 1. In the video coding device, the entropy encoder 112 transmits information by which the chroma quantization offsets for the RGB space are able to be identified (for example, an index for specifying a data table in which the chroma quantization offsets retained in the video decoding device are set) to the video decoding device.
According to this exemplary embodiment, in the video coding device, the entropy encoder 112 transmits information by which the chroma quantization offsets for the YCoCr space are able to be identified (for example, an index for specifying a data table in which the chroma quantization offsets retained in the video decoding device are set) to the video decoding device.
Exemplary Embodiment 8
Subsequently, a video decoding device according to an eighth exemplary embodiment will be described. The video decoding device of this exemplary embodiment corresponds to the video coding device of the seventh exemplary embodiment. Moreover, the configuration of the video decoding device of this exemplary embodiment is the same as the configuration illustrated in FIG. 3.
When the entropy decoder 212 interprets that data is compressed in the YCoCr space by the syntax illustrated in FIG. 9 in the video decoding device, the chroma quantization offsets are read from the data table specified by the index, for example, and the adaptive chroma quantization offset derivation unit 221 calculates the chroma quantization parameter in the same manner as in the second exemplary embodiment.
Moreover, in the video coding device, the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221.
The following describes a specific example of a procedure for deriving chroma quantization offsets. In the following description, wordings enclosed by quotation marks indicate characteristics in this exemplary embodiment.
cu_chroma_qp_offset_idx, when present, specifies the index into the cb_qp_offset_list [ ] and cr_qp_offset_list [ ] or the alt_cb_qp_offset_list [ ] and alt_cr_qp_offset_list [ ] that is used to determine the value of CuQpOffsetCb and CuQpOffsetCr. When present, the value of cu_chroma_qp_offset_idx shall be in the range of 0 to chroma_qp_offset_list_len_minus1, inclusive. When not present, the value of cu_chroma_qp_offset_idx is inferred to be equal to 0. When cu_chroma_qp_offset_flag is present, the following applies:
The variable IsCuChromaQpOffsetCoded is set equal to 1.
The variables CuQpOffsetCb and CuQpOffsetCr are derived as follows:
If cu_chroma_qp_offset_flag is equal to 1 and “cu_residual_csc_flag is equal to 0”, the following applies:
CuQpOffsetCb=cb_qp_offset_list [cu_chroma_qp_offset_idx]
CuQpOffsetCr=cr_qp_offset_list [cu_chroma_qp_offset_idx]
“Otherwise if cu_chroma_qp_offset_flag is equal to 1 and cu_residual_csc_flag is equal to 1, the following applies:”
“CuQpOffsetCb=alt_cb_qp_offset_list [cu_chroma_qp_offset_idx]”
“CuQpOffsetCr=alt_cr_qp_offset_list [cu_chroma_qp_offset_idx]”
Otherwise (cu_chroma_qp_offset_flag is equal to 0), CuQpOffsetCb and CuQpOffsetCr are both set equal to 0.
Exemplary Embodiment 9
Although the video coding device explicitly signals the chroma quantization offsets in the above exemplary embodiments, the signaling of the chroma quantization offsets may be omitted though it is signaled that the color space of the prediction error signal is selected in block units. In the present description, the above is referred to as chroma quantization offsets are implicitly signaled.
In the case where the video coding device implicitly signals chroma quantization offsets, the entropy encoder signals adaptive_color_trans_flag=1 and then signals cu_residual_csc_flag in block units, for example, while not signaling information by which the values of the chroma quantization offsets are able to be identified.
In case where the entropy decoder parses a bitstream to get adaptive_color_trans_flag=1 in the video decoding device and gets cu_residual_csc_flag=0 (indicating that data is compressed in the RGB space), the adaptive chroma quantization offset derivation unit 221 reads out the values of the chroma quantization offsets for the RGB space previously stored in the video decoding device. In addition, if the entropy decoder gets cu_residual_csc_flag=1 (indicating that data is compressed in the YCoCr space), the adaptive chroma quantization offset derivation unit 221 calculates the values of the chroma quantization offsets for the YCoCr space from the values of the previously-stored chroma quantization offsets for the RGB space.
Since the chroma quantization offsets for the RGB space correlate with the chroma quantization offsets for the YCoCr space to some extent, in other words, since it is possible to define a calculation formula for calculating the chroma quantization offsets for the YCoCr space from the chroma quantization offsets for the RGB space, the adaptive chroma quantization offset derivation unit 221 is able to derive the chroma quantization offsets for the YCoCr space by using the calculation formula.
Specifically, the video decoding device implicitly derives the chroma quantization offsets.
Moreover, in the video coding device, the adaptive chroma quantization offset derivation unit 121 operates in the same manner as the adaptive chroma quantization offset derivation unit 221.
Moreover, if the video coding device implicitly signals chroma quantization offsets, the amount of data to be transmitted is able to be reduced.
Although the RGB space and the YCoCr space are illustrated as two color spaces in the above exemplary embodiments, the system of the above exemplary embodiments is applicable even if one or both of the two color spaces are other color spaces. Furthermore, although the first color component G, the second color component B, and the third color component R are used in the RGB space in the above exemplary embodiments (see FIG. 20), the way of allocating color signals to the respective color components is not limited thereto, but arbitrary color signals can be allocated to the respective color components.
Although the video coding device and the video decoding device use two color spaces in the above exemplary embodiments, these devices are also allowed to use three or more color spaces.
Furthermore, each of the exemplary embodiments described above may be realized by hardware, but may also be realized by a computer program.
The information processing system illustrated in FIG. 10 includes a processor 1001, a program memory 1002, a storage medium 1003 for storing video data, and a storage medium 1004 for storing a bitstream. The storage medium 1003 and the storage medium 1004 may be separate storage media, or storage areas included in the same storage medium. A magnetic storage medium such as a hard disk is available as such a storage medium.
In the information processing system illustrated in FIG. 10, the program memory 1002 stores a program for realizing the functions of the blocks (except the block of the buffer) illustrated in each of FIGS. 1 and 3. The processor 1001 realizes the functions of the video coding device or video decoding device illustrated in FIG. 1 or FIG. 3 by executing the processes according to the program stored in the program memory 1002.
FIG. 11 is a block diagram illustrating the main part of the video coding device. As illustrated in FIG. 11, the video coding device 301 includes an adaptive chroma quantization offset derivation unit 311 (corresponding to the adaptive chroma quantization offset derivation unit 121 illustrated in FIG. 1, for example) which derives chroma quantization offsets for each color space and an inverse quantization unit 312 (corresponding to the inverse quantizer/inverse frequency transformer 105 illustrated in FIG. 1, for example) which inverse-quantizes a quantization coefficient image by using the chroma quantization offsets for each color space.
FIG. 12 is a block diagram illustrating another example of the main part of the video coding device. As illustrated in FIG. 12, the video coding device 302 further includes a color space selection notification unit 313 (corresponding to the entropy encoder 112 illustrated in FIG. 1, for example) which signals that the color space of a prediction error signal is to be selected in block units.
Moreover, unless the video coding device 302 includes a means for signaling information by which the values of the quantization offsets for each color space are able to be identified in the configuration illustrated in FIG. 12, the chroma quantization offsets are implicitly derived in the video coding device 302.
FIG. 13 is a block diagram illustrating still another example of the main part of the video coding device. As illustrated in FIG. 13, the video coding device 303 further includes a quantization offset information transmission unit 314 (corresponding to the entropy encoder 112 illustrated in FIG. 1, for example) which signals information by which the values of the chroma quantization offsets for each color space are able to be identified. The information by which the values of chroma quantization offsets are able to be identified is the values themselves of the chroma quantization offsets or an index for specifying the data table in which the chroma quantization offsets retained in the video decoding device are set, for example.
FIG. 14 is a block diagram illustrating the main part of the video decoding device. As illustrated in FIG. 14, the video decoding device 401 includes an adaptive chroma quantization offset derivation unit 411 (corresponding to the adaptive chroma quantization offset derivation unit 221 illustrated in FIG. 3, for example) which derives chroma quantization offsets for each color space and an inverse quantization unit 412 (corresponding to the inverse quantizer/inverse frequency transformer 205 illustrated in FIG. 3, for example) which inverse-quantizes a quantization coefficient image by using the chroma quantization offsets for each color space.
FIG. 15 is a block diagram illustrating another example of the main part of the video decoding device. As illustrated in FIG. 15, the video decoding device 402 further includes a color space selection parsing unit 413 (corresponding to the entropy decoder 212 illustrated in FIG. 3, for example) which parses a bitstream to interpret that the color space of a prediction error signal is to be selected in block units.
Moreover, unless the video decoding device 402 includes a means for parsing a bitstream to interpret information by which the values of the chroma quantization offsets for each color space are able to be identified in the configuration illustrated in FIG. 15, the video decoding device 402 implicitly derives the chroma quantization offset.
FIG. 16 is a block diagram illustrating still another example of the main part of the video decoding device. As illustrated in FIG. 16, the video decoding device 403 further includes a chroma quantization offset parsing unit 414 (corresponding to the entropy decoder 212 illustrated in FIG. 3, for example) which identifies the values of the chroma quantization offsets for each color space on the basis of the information interpreted from the received bitstream.
Although the present invention has been described with reference to the foregoing exemplary embodiments and examples, the present invention is not limited to the foregoing exemplary embodiments and examples. Various changes understandable by those skilled in the art can be made to the structures and details of the present invention within the scope of the present invention.
This application claims priority to Japanese Patent Application No. 2014-204392 filed on Oct. 3, 2014, and the entire disclosure thereof is hereby incorporated herein by reference.
REFERENCE SIGNS LIST
    • 101 Switch
    • 102 Color space transformer
    • 103 Switch
    • 104 Frequency transformer/quantizer
    • 105 Inverse quantizer/inverse frequency transformer
    • 106 Switch
    • 107 Inverse color space transformer
    • 108 Switch
    • 109 Buffer
    • 110 Predictor
    • 111 Prediction parameter determiner
    • 112 Entropy encoder
    • 115 Subtractor
    • 116 Adder
    • 121 Adaptive chroma quantization offset derivation unit
    • 122 Switch
    • 205 Inverse quantizer/inverse frequency transformer
    • 206 Switch
    • 207 Inverse color space transformer
    • 208 Switch
    • 209 Buffer
    • 210 Predictor
    • 212 Entropy decoder
    • 216 Adder
    • 221 Adaptive chroma quantization offset derivation unit
    • 222 Switch
    • 301, 302, 303 Video coding device
    • 311 Adaptive chroma quantization offset derivation unit
    • 312 Inverse quantization unit
    • 313 Color space selection notification unit
    • 314 Quantization offset information transmission unit
    • 401, 402, 403 Video decoding device
    • 411 Adaptive chroma quantization offset derivation unit
    • 412 Inverse quantization unit
    • 413 Color space selection parsing unit
    • 414 Chrominance quantization offset parsing unit
    • 1001 Processor
    • 1002 Program memory
    • 1003 Storage medium
    • 1004 Storage medium

Claims (5)

The invention claimed is:
1. A video decoding device capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the device comprising:
a hardware including at least a processor,
wherein the hardware implements:
an adaptive chroma quantization offset derivation unit which derives a chroma quantization offset, in coded block units, using a clipping function which clips a value obtained using a quantization parameter of a first color component of the color space selected from the plurality of color spaces and the chroma quantization offset for each block of a color component other than the first color component into a predetermined range, wherein the chroma quantization offset is a variable value; and
an inverse quantization unit which inverse-quantizes a quantization coefficient image, in coded block units, by using the derived chroma quantization offset of the color space.
2. The video decoding device according to claim 1, wherein the hardware further implements a color space selection parsing unit which parses a received bitstream to interpret that the color space of the prediction error signal is to be selected in block units.
3. The video decoding device according to claim 2, wherein the hardware further implements a chroma quantization offset parsing unit which identifies the value of the chroma quantization offset for each color space of the plurality of color spaces based on information obtained from the received bitstream.
4. A video decoding method, implemented by a processor, capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, the method comprising:
deriving a chroma quantization offset, in coded block units, using a clipping function which clips a value obtained using a quantization parameter of a first color component of the color space selected from the plurality of color spaces and the chroma quantization offset for each block of a color component other than the first color component into a predetermined range, wherein the chroma quantization offset is a variable value; and
inverse-quantizing a quantization coefficient image, in coded block units, by using the derived chroma quantization offset of the color space.
5. A non-transitory computer readable information recording medium storing a video decoding program for performing a video decoding method capable of selecting, from among a plurality of color spaces, a color space of a prediction error signal in coded block units, when executed by a processor, the program performs:
deriving a chroma quantization offset, in coded block units, using a clipping function which clips a value obtained using a quantization parameter of a first color component of a color space selected from the plurality of color spaces and a chroma quantization offset for each block of a color component other than the first color component into a predetermined range, wherein the chroma quantization offset is a valuable value; and
inverse-quantizing a quantization coefficient image, in coded block units, by using the derived chroma quantization offset of the color space.
US15/516,223 2014-10-03 2015-08-12 Video coding device, video decoding device, video coding method, video decoding method and program Active 2037-06-15 US11039141B2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2014204392 2014-10-03
JP2014-204392 2014-10-03
JPJP2014-204392 2014-10-03
PCT/JP2015/004037 WO2016051643A1 (en) 2014-10-03 2015-08-12 Video coding device, video decoding device, video coding method, video decoding method and program

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2015/004037 A-371-Of-International WO2016051643A1 (en) 2014-10-03 2015-08-12 Video coding device, video decoding device, video coding method, video decoding method and program

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/316,431 Continuation US11575900B2 (en) 2014-10-03 2021-05-10 Video coding device, video decoding device, video coding method, video decoding method and program

Publications (2)

Publication Number Publication Date
US20170318293A1 US20170318293A1 (en) 2017-11-02
US11039141B2 true US11039141B2 (en) 2021-06-15

Family

ID=55629723

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/516,223 Active 2037-06-15 US11039141B2 (en) 2014-10-03 2015-08-12 Video coding device, video decoding device, video coding method, video decoding method and program
US17/316,431 Active US11575900B2 (en) 2014-10-03 2021-05-10 Video coding device, video decoding device, video coding method, video decoding method and program

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/316,431 Active US11575900B2 (en) 2014-10-03 2021-05-10 Video coding device, video decoding device, video coding method, video decoding method and program

Country Status (19)

Country Link
US (2) US11039141B2 (en)
EP (1) EP3203739A4 (en)
JP (1) JP6540707B2 (en)
KR (1) KR101989710B1 (en)
CN (3) CN107113431B (en)
AU (1) AU2015326293B2 (en)
BR (1) BR112017006461A2 (en)
CA (1) CA2963338A1 (en)
CL (1) CL2017000781A1 (en)
CO (1) CO2017003030A2 (en)
MA (1) MA40120B1 (en)
MX (1) MX2017004088A (en)
PE (1) PE20171159A1 (en)
PH (1) PH12017500600A1 (en)
RU (1) RU2017110786A (en)
SG (1) SG11201702743WA (en)
TW (1) TW201631974A (en)
WO (1) WO2016051643A1 (en)
ZA (1) ZA201702044B (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5828361B2 (en) * 2012-01-20 2015-12-02 ソニー株式会社 Color difference quantization parameter expansion
WO2016056977A1 (en) * 2014-10-06 2016-04-14 Telefonaktiebolaget L M Ericsson (Publ) Coding and deriving quantization parameters
US9883184B2 (en) * 2014-10-07 2018-01-30 Qualcomm Incorporated QP derivation and offset for adaptive color transform in video coding
US10477212B2 (en) * 2016-10-11 2019-11-12 Dolby Laboratories Licensing Corporation Adaptive chroma quantization in video coding for multiple color imaging formats
EP3526968A4 (en) * 2016-10-12 2020-06-03 Telefonaktiebolaget LM Ericsson (publ) Residual refinement of color components
CN113573057B (en) * 2018-02-01 2023-07-18 寰发股份有限公司 Video encoding or decoding method and apparatus with adaptive quantization of video data
KR20230151079A (en) * 2019-06-11 2023-10-31 엘지전자 주식회사 Chroma quantization parameter data based image decoding method and the apparatus therof
WO2021086024A1 (en) * 2019-10-28 2021-05-06 엘지전자 주식회사 Image encoding/decoding method and apparatus performing residual processing by using adaptive color space transformation, and method for transmitting bitstream
KR20220070012A (en) 2019-10-28 2022-05-27 엘지전자 주식회사 Image encoding/decoding method using color space conversion, apparatus and method for transmitting a bitstream
US11336913B2 (en) * 2019-11-22 2022-05-17 Apple Inc. Reference picture re-sampling
WO2021107641A1 (en) * 2019-11-26 2021-06-03 주식회사 윌러스표준기술연구소 Method and device for processing video signal by using adaptive color space transform
JP7277354B2 (en) * 2019-12-26 2023-05-18 Kddi株式会社 Image decoding device, image decoding method and program
CN115002512B (en) * 2022-05-23 2023-11-28 北京市商汤科技开发有限公司 Video transcoding method and device, electronic equipment and storage medium

Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050259730A1 (en) 2004-05-18 2005-11-24 Sharp Laboratories Of America, Inc. Video coding with residual color conversion using reversible YCoCg
US20060018559A1 (en) 2004-07-22 2006-01-26 Samsung Electronics Co., Ltd. Method and apparatus to transform/inverse transform and quantize/dequantize color image, and method and apparatus to encode/decode color image using it
TW200629907A (en) 2004-11-15 2006-08-16 Allume Systems Inc System and method for lossless compression of digital images
RU2316909C2 (en) 2005-03-04 2008-02-10 Самсунг Электроникс Ко., Лтд. Method and device for video encoding and decoding scalable across color space
US20090168894A1 (en) 2006-01-13 2009-07-02 Detlev Marpe Picture coding using adaptive color space transformation
JP2011151683A (en) 2010-01-22 2011-08-04 Sony Corp Image processing apparatus and method
CN102172024A (en) 2008-10-03 2011-08-31 高通股份有限公司 Quantization parameter selections for encoding of chroma and luma video blocks
CN102223525A (en) 2010-04-13 2011-10-19 富士通株式会社 Video decoding method and system
US20120002721A1 (en) 2010-07-01 2012-01-05 Gheorghe Berbecel Method and system for multi-layer rate control for a multi-codec system
CN102625100A (en) 2011-01-25 2012-08-01 Arm有限公司 Image encoding method
US20130142255A1 (en) 2011-06-27 2013-06-06 Toru Matsunobu Image coding method, image decoding method, image coding apparatus, image decoding apparatus, and image coding-decoding apparatus
US20130188693A1 (en) 2012-01-20 2013-07-25 Sony Corporation Chroma quantization parameter extension
US20130321675A1 (en) 2012-05-31 2013-12-05 Apple Inc. Raw scaler with chromatic aberration correction
US20140003497A1 (en) 2012-07-02 2014-01-02 Microsoft Corporation Control and use of chroma quantization parameter values
US20140003498A1 (en) * 2012-07-02 2014-01-02 Microsoft Corporation Use of chroma quantization parameter offsets in deblocking
CN103501437A (en) 2013-09-29 2014-01-08 北京航空航天大学 Fractal and H.264-based hyper-spectral image compression method
US20140314142A1 (en) * 2011-11-04 2014-10-23 Infobridge Pte. Ltd. Apparatus of decoding video data
US20150256840A1 (en) * 2012-12-18 2015-09-10 Sony Corporation Image processing device and image processing method
US20150264374A1 (en) * 2014-03-14 2015-09-17 Vid Scale, Inc. Systems and methods for rgb video coding enhancement
WO2015143671A1 (en) 2014-03-27 2015-10-01 Microsoft Technology Licensing, Llc Adjusting quantization/scaling and inverse quantization/scaling when switching color spaces
WO2015187978A1 (en) 2014-06-04 2015-12-10 Qualcomm Incorporated Block adaptive color-space conversion coding

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE69817460T2 (en) * 1997-06-09 2004-06-09 Hitachi, Ltd. Image sequence decoding method
KR101138392B1 (en) * 2004-12-30 2012-04-26 삼성전자주식회사 Color image encoding and decoding method and apparatus using a correlation between chrominance components
KR101246915B1 (en) * 2005-04-18 2013-03-25 삼성전자주식회사 Method and apparatus for encoding or decoding moving picture
KR101568708B1 (en) * 2007-09-02 2015-11-12 엘지전자 주식회사 A method and an apparatus for processing a video signal
JP5936939B2 (en) * 2011-07-14 2016-06-22 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America Image encoding method and image decoding method

Patent Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050259730A1 (en) 2004-05-18 2005-11-24 Sharp Laboratories Of America, Inc. Video coding with residual color conversion using reversible YCoCg
US20060018559A1 (en) 2004-07-22 2006-01-26 Samsung Electronics Co., Ltd. Method and apparatus to transform/inverse transform and quantize/dequantize color image, and method and apparatus to encode/decode color image using it
TW200629907A (en) 2004-11-15 2006-08-16 Allume Systems Inc System and method for lossless compression of digital images
RU2316909C2 (en) 2005-03-04 2008-02-10 Самсунг Электроникс Ко., Лтд. Method and device for video encoding and decoding scalable across color space
US20090168894A1 (en) 2006-01-13 2009-07-02 Detlev Marpe Picture coding using adaptive color space transformation
CN102172024A (en) 2008-10-03 2011-08-31 高通股份有限公司 Quantization parameter selections for encoding of chroma and luma video blocks
JP2011151683A (en) 2010-01-22 2011-08-04 Sony Corp Image processing apparatus and method
CN102223525A (en) 2010-04-13 2011-10-19 富士通株式会社 Video decoding method and system
US20120002721A1 (en) 2010-07-01 2012-01-05 Gheorghe Berbecel Method and system for multi-layer rate control for a multi-codec system
CN102625100A (en) 2011-01-25 2012-08-01 Arm有限公司 Image encoding method
US20130142255A1 (en) 2011-06-27 2013-06-06 Toru Matsunobu Image coding method, image decoding method, image coding apparatus, image decoding apparatus, and image coding-decoding apparatus
US20140314142A1 (en) * 2011-11-04 2014-10-23 Infobridge Pte. Ltd. Apparatus of decoding video data
US20130188693A1 (en) 2012-01-20 2013-07-25 Sony Corporation Chroma quantization parameter extension
US20130321675A1 (en) 2012-05-31 2013-12-05 Apple Inc. Raw scaler with chromatic aberration correction
US20140003497A1 (en) 2012-07-02 2014-01-02 Microsoft Corporation Control and use of chroma quantization parameter values
US20140003498A1 (en) * 2012-07-02 2014-01-02 Microsoft Corporation Use of chroma quantization parameter offsets in deblocking
US20150256840A1 (en) * 2012-12-18 2015-09-10 Sony Corporation Image processing device and image processing method
CN103501437A (en) 2013-09-29 2014-01-08 北京航空航天大学 Fractal and H.264-based hyper-spectral image compression method
US20150264374A1 (en) * 2014-03-14 2015-09-17 Vid Scale, Inc. Systems and methods for rgb video coding enhancement
WO2015143671A1 (en) 2014-03-27 2015-10-01 Microsoft Technology Licensing, Llc Adjusting quantization/scaling and inverse quantization/scaling when switching color spaces
WO2015187978A1 (en) 2014-06-04 2015-12-10 Qualcomm Incorporated Block adaptive color-space conversion coding

Non-Patent Citations (19)

* Cited by examiner, † Cited by third party
Title
D. Flynn et al., High Efficiency Video Coding (HEVC) Range Extensions text specification: Draft 7, JCTVC-Q1005_v9, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11, Apr. 2014.
English translation of Office Action dated Apr. 10, 2018, from Federal Service of Intellectual Property in counterpart Russian Patent Application No. 201710786/08(019009).
Extended European Search Report dated Mar. 28, 2018, issued in counterpart European Patent Application No. 15847019.5.
International Search Report and Written Opinion of ISA dated Sep. 29, 2015, in corresponding PCT International Application.
Invitation to Respond to Written Opinion, Written Opinion and Search Report, dated Dec. 1, 2017, issued from the Intellectual Property Office of Singapore, in counterpart Singapore Application No. 11201702743W.
Jonatan Samuelsson et al., "QP Offset for Adaptive Color Transform", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11—S0092-v2, Strasbourg, FR, Oct. 17-24, 2014, pp. 1-3.
Keiichi Chono, "Enhanced Chroma QP Signaling for Adaptive Cross-Component Transform in SCC Extensions", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11—S0040, Strasbourg, FR, Oct. 17-24, 2014, pp. 1-7.
Krishna Rapaka et al., "QP Derivation and Offsets signaling for Adaptive Color Transform", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11—S0300, Strasbourg, FR, Oct. 17-24, 2014, pp. 1-5.
Krishna Rapaka et al., "QP Derivation for Adaptive Color Transform", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11—S0144, Strasbourg, FR, Oct. 17-24, 2014, pp. 1-4.
L. Zhang et al., "SCCE5 Test 3.2.1: In-loop color-space transform", JCTVC-R0147(V2), Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11, pp. 1-8, Jun. 2014.
L. Zhang et al., "SCCE5 Test 3.2.1: In-loop color-space transform", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11, [JCTVC-R0147], pp. 1-21, Jun. 2014.
Notification of Reasons for Refusal dated Jul. 17, 2018, issued by the Japanese Patent Office in counterpart Japanese Patent Application No. 2016-551481.
Notification of Reasons for Refusal, dated Oct. 9, 2018 and an English-language Translation, issued by the Japanese Patent Office in counterpart Japanese Patent Application No. 2016-551481.
Office Action dated Apr. 10, 2018, from the Federal Service of Intellectual Property in counterpart Russian Patent Application No. 201710786/08(019009).
Office Action dated Jul. 10, 2019, issued by the Chinese Patent Office in counterpart Chinese Patent Application No. CN 201580053910.9.
Rajan Joshi, "High Efficiency Video Coding (HEVC) Screen Content Coding: Draft 1", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11—R1005-v2, Sappiro, JP Jul. 9, 2014, pp. i-xiv and 1-344.
Series H: Audiovisual nad Multimedia Systems, Infrastructure of Audiovisual Services—Coding of Moving Video/High Efficiency video coding, Recommendation ITU-H. 265, ITU-T Telecommunication Standardization Sector of ITU, Apr. 13, 2013, pp. 69 and 79.
Taiwanese Office Action dated Feb. 27, 2018, issued in counterpart Taiwanese Patent Application No. 104132320.
W.S. Kim et al., "Residue Color Transform", Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG, (ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 Q.6), 12th Meeting, Jul. 2004.

Also Published As

Publication number Publication date
CN107113431A (en) 2017-08-29
MA40120A1 (en) 2018-02-28
TW201631974A (en) 2016-09-01
CA2963338A1 (en) 2016-04-07
EP3203739A4 (en) 2018-04-25
BR112017006461A2 (en) 2017-12-19
CO2017003030A2 (en) 2017-06-20
EP3203739A1 (en) 2017-08-09
MX2017004088A (en) 2017-07-07
WO2016051643A1 (en) 2016-04-07
AU2015326293A1 (en) 2017-04-06
KR101989710B1 (en) 2019-06-14
ZA201702044B (en) 2018-08-29
US11575900B2 (en) 2023-02-07
RU2017110786A3 (en) 2018-11-06
PH12017500600A1 (en) 2017-09-04
PE20171159A1 (en) 2017-08-17
RU2017110786A (en) 2018-11-06
SG11201702743WA (en) 2017-06-29
CL2017000781A1 (en) 2017-11-03
MA40120B1 (en) 2018-11-30
CN107113431B (en) 2020-12-01
JPWO2016051643A1 (en) 2017-07-20
US20210266553A1 (en) 2021-08-26
CN112261410A (en) 2021-01-22
AU2015326293B2 (en) 2018-05-17
US20170318293A1 (en) 2017-11-02
CN112261411A (en) 2021-01-22
KR20170052624A (en) 2017-05-12
JP6540707B2 (en) 2019-07-10

Similar Documents

Publication Publication Date Title
US11575900B2 (en) Video coding device, video decoding device, video coding method, video decoding method and program
US11539956B2 (en) Robust encoding/decoding of escape-coded pixels in palette mode
US9253483B2 (en) Signaling of scaling list
US9210425B2 (en) Signaling of temporal motion vector predictor (MVP) flag for temporal prediction
US10136033B2 (en) Techniques for advanced chroma processing
US9549177B2 (en) Evaluation of signaling of collocated reference picture for temporal prediction
US11019338B2 (en) Methods and apparatuses of video encoding or decoding with adaptive quantization of video data
EP2982110B1 (en) Method and device for determining the value of a quantization parameter
US20140023142A1 (en) Signaling of temporal motion vector predictor (mvp) enable flag
US20170006283A1 (en) Computationally efficient sample adaptive offset filtering during video encoding
US20240107044A1 (en) Moving picture decoding device, moving picture decoding method, and program obtaining chrominance values from corresponding luminance values
KR102644971B1 (en) Image decoding method and device using chroma quantization parameter table
WO2016063440A1 (en) Video image encoding device, video image decoding device, video image encoding method, video image decoding method, and program
WO2016185651A1 (en) Video encoding device, video decoding device, video encoding method, video decoding method, and program
WO2014028631A1 (en) Signaling of temporal motion vector predictor (mvp) enable flag
KR20130070215A (en) Method and apparatus for seletcing the adaptive depth information and processing deblocking filtering

Legal Events

Date Code Title Description
AS Assignment

Owner name: NEC CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHONO, KEIICHI;REEL/FRAME:041812/0233

Effective date: 20170327

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCV Information on status: appeal procedure

Free format text: NOTICE OF APPEAL FILED

STCV Information on status: appeal procedure

Free format text: APPEAL BRIEF (OR SUPPLEMENTAL BRIEF) ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE