WO2012047047A2 - Procédé et appareil destinés à coder/décoder une vidéo en utilisant un filtre à haute précision - Google Patents

Procédé et appareil destinés à coder/décoder une vidéo en utilisant un filtre à haute précision Download PDF

Info

Publication number
WO2012047047A2
WO2012047047A2 PCT/KR2011/007418 KR2011007418W WO2012047047A2 WO 2012047047 A2 WO2012047047 A2 WO 2012047047A2 KR 2011007418 W KR2011007418 W KR 2011007418W WO 2012047047 A2 WO2012047047 A2 WO 2012047047A2
Authority
WO
WIPO (PCT)
Prior art keywords
block
subsample
image
prediction
value
Prior art date
Application number
PCT/KR2011/007418
Other languages
English (en)
Korean (ko)
Other versions
WO2012047047A3 (fr
Inventor
송진한
임정연
이영렬
문주희
김해광
전병우
한종기
김정필
김대연
Original Assignee
에스케이텔레콤 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020110072196A external-priority patent/KR101341993B1/ko
Application filed by 에스케이텔레콤 주식회사 filed Critical 에스케이텔레콤 주식회사
Priority to CN201180054874.XA priority Critical patent/CN103210649B/zh
Publication of WO2012047047A2 publication Critical patent/WO2012047047A2/fr
Publication of WO2012047047A3 publication Critical patent/WO2012047047A3/fr
Priority to US13/857,708 priority patent/US9420281B2/en
Priority to US15/031,366 priority patent/US9706222B2/en
Priority to US15/031,378 priority patent/US9602834B2/en
Priority to US15/371,215 priority patent/US10158880B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/152Data rate or code amount at the encoder output by measuring the fullness of the transmission buffer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding

Definitions

  • An embodiment of the present invention relates to a method and apparatus for image encoding / decoding using a high precision filter. More particularly, the present invention relates to a method and apparatus for generating a prediction signal using a high precision filter to generate a high precision image signal and encoding / decoding an image using the same.
  • Moving Picture Experts Group (MPEG) and Video Coding Experts Group (VCEG) have developed video compression techniques that are superior to the existing MPEG-4 Part 2 and H.263 standards.
  • the new standard is called H.264 / AVC (Advanced video Coding), which was jointly released as MPEG-4 Part 10 AVC and ITU-T Recommendation H.264.
  • H.264 / AVC (hereinafter, referred to as H.264)
  • JCT-VC Joint Collaborative Team on Video Coding
  • JCT-VC Joint Collaborative Team on Video Coding
  • the intra prediction encoding method predicts using a prediction value from a block encoded in a frame currently being encoded, and the inter prediction prediction encoding a block of the current frame by estimating a motion from a previously reconstructed frame. The method is used.
  • Intra picture prediction methods using intra luma signals include intra 4 ⁇ 4 prediction, intra 16 ⁇ 16 prediction, intra 8 ⁇ 8 prediction, and the like depending on the prediction direction and the block size to be encoded.
  • 1 is a diagram illustrating nine conventional 4x4 intra prediction modes.
  • intra 4 ⁇ 4 prediction includes a vertical mode, a horizontal mode, a direct current mode, a diagonal down-left mode, and a diagonal down-right.
  • Intra 16 ⁇ 16 prediction there are four prediction modes in the intra 16 ⁇ 16 prediction including a vertical mode, a horizontal mode, a DC mode, and a plane mode.
  • Intra 8 ⁇ 8 prediction also has four prediction modes similar to intra 16 ⁇ 16 prediction.
  • Inter predictive coding of video having a 4: 2: 0 image type uses motion compensation that predicts a current block by dividing an image frame and estimating a motion from a previously encoded frame. If the block size of motion compensation is small, more accurate prediction can be made, but the amount of code increases because the motion vector information must be encoded for each block. Also, when performing motion compensation, not only the motion vector is searched for integer samples with integer pixels but also for subsample positions with resolutions of 1/4 samples for luma components and 1/8 samples for chrominance components. As a result, a more accurate motion vector search method has been used. However, since the luminance and chrominance samples of the subsample position do not exist in the reference picture, these values should be generated by interpolating adjacent integer samples of the reference picture.
  • an embodiment of the present invention uses a high-precision filter that is more precise than linear interpolation in interpolating an image to improve the compression efficiency of the image and to effectively reconstruct the image to improve the subjective image quality.
  • the value of the subsample component of the color difference component of the reference block referred to by the motion vector of the color difference component of the current block is FIR.
  • a prediction block for a chrominance component is generated from a value interpolated using a filter and linear interpolation, and the residual block is generated by subtracting the prediction block from the chrominance component of the current block, and the residual block is transformed and quantized to quantize frequency transform.
  • An image encoder for generating a block and encoding the bitstream; And generating a quantized frequency transform block from the bitstream, and inversely quantizing and inverse transforming the quantized frequency transform block to subtract the color difference component of the reference block referred to by the motion vector of the color difference component of the current block to reconstruct and reconstruct the residual block.
  • a prediction block for a color difference component is generated from an interpolated value of a sample component using an FIR filter and linear interpolation, and the color block component of the current block to be restored is restored by adding the reconstructed residual block and the generated prediction block.
  • an image encoding / decoding apparatus comprising an image decoder.
  • an embodiment of the present invention in the apparatus for encoding an image, the value of the subsample component of the color difference component of the reference block referred to by the motion vector of the color difference component of the current block referred to by the motion vector of the color difference component of the current block A prediction unit for generating a prediction block for the color difference component from the interpolated values using the FIR filter and linear interpolation; A subtraction unit which subtracts the prediction block from the color difference component of the current block to generate a residual block; A transformer for converting the residual block to generate a frequency transform block; A quantizer configured to quantize the frequency transform block to generate a quantized frequency transform block; And an encoder configured to encode the quantized frequency transform block into a bitstream.
  • an embodiment of the present invention in the apparatus for encoding an image, the integer pixel of the periphery for 1/2 subsample of the reference block referred to by the motion vector of the current block
  • a prediction unit generating a prediction block by obtaining a half-sample high precision applying a filtering coefficient to a value and obtaining values of all subsamples of the reference block by using the half-sample high precision
  • a subtraction unit for generating a residual block by subtracting the prediction block from the current block
  • a transformer for converting the residual block to generate a frequency transform block
  • a quantizer configured to quantize the frequency transform block to generate a quantized frequency transform block
  • an encoder configured to encode the quantized frequency transform block into a bitstream.
  • an embodiment of the present invention to achieve another object of the present invention, an apparatus for decoding an image, the decoding unit for extracting a quantized frequency transform block from the bitstream; An inverse quantizer for restoring a frequency transform block by inversely quantizing the quantized frequency transform block; An inverse transform unit which inversely transforms the frequency transform block to restore a residual block; A prediction unit for generating a prediction block for the chrominance component from a value obtained by interpolating a value of the subsample component of the chrominance component of the reference block referred to by the motion vector of the chrominance component of the current block using an FIR filter and linear interpolation; And an adder configured to add the reconstructed residual block and the prediction block to reconstruct the color difference component of the current block.
  • an embodiment of the present invention to achieve another object of the present invention, an apparatus for decoding an image, the decoding unit for extracting a quantized frequency transform block from the bitstream; An inverse quantizer for restoring a frequency transform block by inversely quantizing the quantized frequency transform block; An inverse transform unit which inversely transforms the frequency transform block to restore a residual block; For a 1/2 subsample of a reference block referred to by the motion vector of the current block, a 1/2 sample precision obtained by applying a filtering coefficient to an integer pixel value of a neighbor is obtained and the 1/2 sample high precision is used for the reference block. A prediction unit generating a prediction block by obtaining values of all subsamples of the subsample; And an adder configured to reconstruct the current block by adding the reconstructed residual block and the prediction block.
  • an embodiment of the present invention in the method for encoding / decoding an image, the subsample component of the color difference component of the reference block referred to by the motion vector of the color difference component of the current block Generate a prediction block for a chrominance component from an interpolated value using a FIR filter and linear interpolation, subtract the prediction block from the chrominance component of the current block to generate a residual block, transform and quantize the residual block Generating an quantized frequency transform block and encoding the bitstream into a bitstream; And generating a quantized frequency transform block from the bitstream, and inversely quantizing and inverse transforming the quantized frequency transform block to subtract the color difference component of the reference block referred to by the motion vector of the color difference component of the current block to reconstruct and reconstruct the residual block.
  • a prediction block for a color difference component is generated from an interpolated value of a sample component using an FIR filter and linear interpolation, and the color block component of the current block to be restored is restored by adding the reconstructed residual block and the generated prediction block. It provides a video encoding / decoding method comprising a video decoding step.
  • the motion vector of the color difference component is obtained by motion compensation of the color difference component of the current block and the subsample component of the color difference component
  • a prediction step of generating a prediction block for a chrominance component from a value obtained by interpolating the value of F using a FIR filter and linear interpolation A subtraction step of generating a residual block by subtracting the prediction block from the color difference component of the current block; Transforming the residual block to generate a frequency transform block;
  • an encoding step of encoding the quantized frequency transform block into a bitstream
  • an embodiment of the present invention provides a method for encoding an image, wherein integer integers of neighboring one-half subsamples of a reference block referred to by a motion vector of the current block
  • a prediction step of generating a prediction block by obtaining a half-sample high precision applying a filtering coefficient to a value and calculating values of all subsamples of the reference block by using the half-sample high precision Subtracting the prediction block from the current block to generate a residual block; Transforming the residual block to generate a frequency transform block; A quantization step of quantizing the frequency transform block to generate a quantized frequency transform block; And an encoding step of encoding the quantized frequency transform block into a bitstream.
  • an embodiment of the present invention to achieve another object of the present invention, a method for decoding an image, the decoding step of generating a quantized frequency transform block from a bitstream; An inverse quantization step of restoring a frequency transform block by inverse quantizing the quantized frequency transform block; An inverse transform step of restoring a residual block by inversely transforming the frequency transform block; Predicting a motion vector of the color difference component by motion compensating the color difference component of the current block and generating a prediction block for the color difference component from values obtained by interpolating values of the subsample components of the color difference component using an FIR filter and linear interpolation; And adding the reconstructed residual block and the prediction block to reconstruct the color difference component of the current block.
  • an embodiment of the present invention to achieve another object of the present invention, a method for decoding an image, the decoding step of extracting a quantized frequency transform block from the bitstream; An inverse quantization step of restoring a frequency transform block by inverse quantizing the quantized frequency transform block; An inverse transform step of restoring a residual block by inversely transforming the frequency transform block; For a 1/2 subsample of a reference block referred to by the motion vector of the current block, a 1/2 sample precision obtained by applying a filtering coefficient to an integer pixel value of a neighbor is obtained and the 1/2 sample high precision is used for the reference block. A prediction step of generating a prediction block by obtaining values of all subsamples of the subsample; And an addition step of reconstructing the current block by adding the reconstructed residual block and the prediction block.
  • the encoding efficiency can be increased by reducing the difference between the actual block and the predicted block, thereby improving the compression efficiency of the current block. It is possible to effectively reconstruct an image by decoding a block converted into a bitstream by improving and considering an encoding method.
  • 1 is a diagram illustrating nine conventional 4x4 intra prediction modes.
  • FIG. 2 is a diagram illustrating four conventional 16x16 intra prediction modes.
  • FIG. 3 is a diagram illustrating motion prediction using a conventional subsample in a luminance (Luma) component.
  • FIR Finite Impulse Response
  • FIG. 6 is a block diagram schematically illustrating a video encoding apparatus according to an embodiment of the present invention.
  • FIG. 7 is an exemplary diagram for describing a process of the prediction unit 610 interpolating a subsample value in a block according to an embodiment of the present invention.
  • FIG. 8 is a block diagram illustrating a configuration of an image decoding apparatus according to an embodiment of the present invention.
  • FIG. 9 is a flowchart illustrating an image encoding method according to an embodiment of the present invention.
  • FIG. 10 is a flowchart illustrating an image decoding method according to an embodiment of the present invention.
  • a video encoding apparatus (Video Encoding Apparatus), a video decoding apparatus (Video Decoding Apparatus) to be described below is a personal computer (PC), notebook computer, personal digital assistant (PDA), portable multimedia player (PMP) It may be a user terminal such as a portable multimedia player (PSP), a PlayStation Portable (PSP), a wireless communication terminal, a smart phone, a TV, or a server terminal such as an application server or a service server.
  • a communication device such as a communication modem for communicating with various devices or a wired / wireless communication network, a memory for storing various programs and data for encoding or decoding an image or inter or intra prediction for encoding or decoding, and executing a program And a microprocessor for controlling and the like. It can mean a variety of devices.
  • the image encoded in the bitstream by the video encoding apparatus is real-time or non-real-time through the wired or wireless communication network, such as the Internet, local area wireless communication network, wireless LAN network, WiBro network, mobile communication network, or the like, or a cable, universal serial bus (USB: Universal) It may be transmitted to an image decoding apparatus through various communication interfaces such as a serial bus, and may be decoded by the image decoding apparatus to restore and reproduce the image.
  • wired or wireless communication network such as the Internet, local area wireless communication network, wireless LAN network, WiBro network, mobile communication network, or the like, or a cable, universal serial bus (USB: Universal) It may be transmitted to an image decoding apparatus through various communication interfaces such as a serial bus, and may be decoded by the image decoding apparatus to restore and reproduce the image.
  • USB universal serial bus
  • a video may be composed of a series of pictures, and each picture may be divided into a predetermined area such as a frame or a block.
  • the divided blocks may be classified into intra blocks and inter blocks according to an encoding method.
  • An intra block refers to a block that is encoded by using an intra prediction coding scheme. Intra prediction coding is performed by using pixels of blocks that have been previously encoded, decoded, and reconstructed in a current picture that performs current encoding. A prediction block is generated by predicting pixels of a block, and a difference value with pixels of the current block is encoded.
  • An inter block refers to a block that is encoded using inter prediction coding.
  • Inter prediction coding generates a prediction block by predicting a current block in a current picture by referring to one or more past pictures or future pictures, and then generates a current block. This is a method of encoding the difference value with.
  • a frame referred to for encoding or decoding the current picture is referred to as a reference frame.
  • FIG. 3 is a diagram illustrating motion prediction using a conventional subsample in a luminance (Luma) component.
  • a motion vector may be obtained from an integer sample unit up to a subsample position of 1/4 samples.
  • FIR Finite Impulse Response
  • pixel values at 1/2 sample positions are interpolated using 6 values having integer pixel values and 6 filter coefficients ⁇ 1, -5, 20, 20, -5, 1 ⁇ and 1
  • the components of the / 4 sample are interpolated with interpolated 1/2 sample pixel values and integer pixel values or linear interpolation using two 1/2 sample pixel values to interpolate pixel values at positions of 1/4 samples as shown in the following equation. do.
  • FIG. 5 is a diagram illustrating linear interpolation for a chroma sample.
  • multiplying weight values by considering bi-linear interpolation using bi-linear interpolation using the values of four integer pixels as in the following example shows that 1 / The pixel values of the 8 subsamples are interpolated.
  • ⁇ a [(6 ⁇ 5 ⁇ A) + (2 ⁇ 5 ⁇ B) + (6 ⁇ 3 ⁇ C) + (2 ⁇ 3 ⁇ D)] / 64 ⁇ .
  • the color difference signal is 1/4 (horizontal 1/2, vertical 1/2) of the luminance signal in the 4: 2: 0 image form.
  • FIG. 6 is a block diagram schematically illustrating a video encoding apparatus according to an embodiment of the present invention.
  • the image encoding apparatus 600 encodes an image by generating a subsample prediction value of a chrominance component using a motion vector value of a luminance component of a current block of the image.
  • the prediction unit 610 may include a predictor 610, a subtractor 620, a transformer 630, a quantizer 640, and an encoder 650.
  • the input image to be encoded is input in units of macro blocks.
  • the macro blocks are in the form of M ⁇ N, M and N each have a size of 2 n , and M and N may be the same or different. have.
  • the prediction unit 610 generates a prediction block for the color difference component from the value obtained by interpolating a value of the subsample component of the color difference component of the reference block referred to by the motion vector of the color difference component of the current block using an FIR filter and linear interpolation. do.
  • the prediction unit 610 may generate a prediction block using another frame to predict the current macro block. That is, the prediction unit 610 may generate a motion vector through motion estimation in a previous frame that has already been encoded and reconstructed, and generate a prediction block in a motion compensation process using the motion vector. In this case, the prediction unit 610 may use the same motion vector value even in the color difference component by using the motion vector value of the luminance component, and use the high-precision FIR to determine the value of the subsample component of the block in the reference frame indicated by the motion vector. A prediction block having prediction values using a filter and linear interpolation may be generated.
  • the subtractor 620 generates a residual signal by calculating a difference value between the original pixel value of each pixel of the current block and the predicted value generated by the predictor 610.
  • the converter 630 converts the residual signal generated by the subtractor 620 into the frequency domain.
  • the transform unit 630 uses various transformation techniques for transforming an image signal of a time axis into a frequency axis, such as Discrete Cosine Transform (DCT) or Wavelet Transform.
  • DCT Discrete Cosine Transform
  • Wavelet Transform Discrete Cosine Transform
  • the quantization unit 640 quantizes the frequency conversion block formed of the residual signal converted into the frequency domain by the conversion unit 630.
  • various quantization techniques such as dead zone uniform threshold quantization (DZUTQ) or quantization weighted matrix (Quantization Weighted Matrix) may be used.
  • the encoder 650 encodes the quantized frequency transform block including the frequency coefficient quantized by the quantizer 640 into a bitstream.
  • an entropy encoding technique may be used, but various encoding techniques may be used without being limited thereto.
  • the encoder 650 may include not only a bit string encoding the quantized frequency coefficients but also various pieces of information necessary to decode the encoded bit string in the encoded data. That is, the coded data includes a field including a coded block pattern (CBP), a delta quantization parameter, and a bit string in which the quantization frequency coefficients are encoded, and information necessary for prediction (eg, intra prediction). In the case of Intra prediction mode or a motion vector in the case of inter prediction, etc.) may include a field that contains a bit.
  • CBP coded block pattern
  • a delta quantization parameter e.g., a bit string in which the quantization frequency coefficients are encoded
  • information necessary for prediction eg, intra prediction
  • Intra prediction mode or a motion vector in the case of inter prediction, etc. may include a field that contains a bit.
  • the inverse quantization unit 660 inverse quantizes the transformed and quantized residual block (ie, the quantized frequency transform block), and the inverse transform unit 670 inverses the inverse quantized transform residual block. To reconstruct the residual block.
  • inverse quantization and inverse transformation may be performed by inversely performing a transformation process performed by the transform unit 630 and a quantization process performed by the quantization unit 640. That is, the inverse quantization unit 660 and the inverse transform unit 670 are information on transform and quantization generated and transmitted from the transform unit 630 and the quantization unit 640 (for example, information on a transform and quantization type). Inverse quantization and inverse transformation may be performed using.
  • the adder 680 adds the prediction block generated by the predictor 610 and the residual block generated by the inverse transform unit 670 to generate a reconstructed block.
  • the frame memory 690 stores a block reconstructed by the adder 680 and is used as a reference block to generate a prediction block when performing intra or inter prediction.
  • FIG. 7 is an exemplary diagram for describing a process of the prediction unit 610 interpolating a subsample value in a block according to an embodiment of the present invention.
  • Interpolation of the subsample of FIG. 7 may be performed using a method such as Equation 1 to Equation 4.
  • the 1/2 subsample value can be obtained by multiplying the integer pixel values by a predetermined value and then using the 1/2 sample amplification value which is the sum of these multiplied values.
  • the 1/2 subsample value can be obtained by using three left integer pixel values A, B, and C and three right integer pixel values D, E, and F.
  • the 1/4 subsample value can be obtained using the nearest integer pixel value C and the half sample amplification value d '.
  • the formula for obtaining b is obtained by dividing the form of linear interpolation between the nearest integer pixel value (C) and the half-sample amplification value (d ') instead of directly linearly interpolating the 1/2 subsample. The error due to the integerization can be eliminated.
  • b ' can also be obtained using the nearest integer pixel value (C) and the half-sample amplification value (d'), so that the 1/8 subsample value is also the nearest integer pixel value (C) and half-sample amplification. Can be obtained using the value d '.
  • Equations 1, 2, 3, 4, and 7 A, B, C, D, E, and F represent integer pixel values of the color difference component, d represents 1/2 subsample of the color difference component, and b represents the color difference component. 1/4 subsample, a, represents 1/8 subsample of the color difference component.
  • the value of the 1/2 subsample of the chrominance component is generated by using a high precision filter (where the high precision filter can use various filters such as FIR filter), and as shown in Equation 2 Similarly, the values of quarter subsamples use high precision filters and linear interpolation.
  • high-precision linear interpolation can be performed using integer pixels and 1/4 subsamples as shown in Equation 3, and also integer pixels and 1/2 subsamples are used as shown in Equation 4. High accuracy linear interpolation can be achieved.
  • w1 and w2 represent weight values multiplied by integer pixels and 1/2 subsample.
  • the 1/2 subsample value (e.g., d) using Equation 1 and the 1/4 subsample value (e.g., b) using Equation 2 are generated using a high-precision FIR filter, and Equation 3 is used. Since a 1/8 subsample value (eg, a) is generated through linear interpolation between two pixel values, all subsample values can be generated more accurately than when interpolation is performed using only linear interpolation.
  • Equation 5 is an equation in which the above rounding operation is added to Equation 3 below.
  • the FIR filter is a type of digital filter and performs filtering with only certain values of the input signal. Therefore, for the FIR filter, the impulse response, which is the characteristic function of the filter, has a finite length. In addition, in the form of the FIR filter, it does not have a feedback component, so the execution time increases due to the higher order when implementing the same characteristic, but the FIR filter is important when phase shift (ie, the shape maintenance of the waveform between the input and the output) is important. Can be used.
  • the high precision FIR filter used here is a 1/2 subsample using an FIR filter, and a 1/4 subsample is 1/2 divided by 32 to the value (d ') using the FIR filter as shown in Equation 2.
  • the subsample value (d in Equation 1) multiply the value before dividing by 32 (that is, the FIR filtering result d 'obtained by obtaining the value of 1/2 subsample) and the integer pixel value by 32.
  • Linear interpolation of the value 32XC prevents the loss of information due to the division operation in the middle and thus interpolates more precise values.
  • the prediction block for the current block may be obtained as sample values.
  • the number of bits per sample of the prediction block may be better than the number of bits per sample of the current block.
  • Equation 4 only the value obtained by interpolating a (1/8 subsample) is described. However, the weight (w1, w2) is set differently from the case of obtaining a even when obtaining b (1/4 subsample). B can be obtained.
  • f can be obtained by using f 'instead of b', f instead of b, and D instead of C in Equation 2, and g 'instead of a', g 'instead of a, and f' instead of b 'in Equation 3
  • 1/8 subsample c can be obtained by interpolating b and d
  • e can be obtained by interpolating d and f, so that the nearest integer pixel value (C) and 1/2 sample are similar to a or g. It can be obtained using the amplification value d '.
  • the surrounding integer pixel value is obtained to obtain 1/2 subsample, and the value of the finer subsample (1/4 subsample, 1/8 subsample, etc.) is smaller than the left and right (or the top and bottom) subtle subsamples.
  • Interpolation i.e., using 1/2 subsample and integer pixel to obtain a 1/4 sample value
  • the case of using the value of the subsample or the integer pixel for interpolation of the subsamples of various positions may be inferred by those skilled in the art even if the case of all the subsamples is not described.
  • FIR filters having 6-tab or more for 1/2 samples, and high-precision functions such as Equation 2 above for 1/4 samples It is important to interpolate to have.
  • the reference blocks in the reference frame may be interpolated using Equations 1 to 3, and in the case of the 4: 4: 4 picture form, the luminance component in the reference block Since the magnitudes of the chrominance components are the same, the chrominance components, like the luminance component, may be interpolated to a quarter sample position. Therefore, if the 1/8 sample position interpolation of Equation 3 is omitted, it can be used in 4: 4: 4 image form.
  • the horizontal color difference component is interpolated to the 1/8 subsample position as shown in Equations 1 to 3 below.
  • the subsample may be generated by interpolating up to the position of the 1/4 subsample using Equations 1 and 2 similarly to the 4: 4: 4 image form.
  • TMuC Test Model under Consideration
  • TMuC Test Model under Consideration
  • the chrominance components can be up to 1/16 subsamples in 4: 2: 0 video format. You can interpolate.
  • up to 1/8 sub-positions can be made as stated in one embodiment of the present invention and the positions of 1/16 subsamples can be interpolated again using linear interpolation.
  • the filtering and interpolation methods have been described with reference to color difference components as an example.
  • the method may be applied to various blocks such as luminance components and blocks of R, G, and B colors as well as color difference components.
  • the same can be applied to various blocks such as luminance components and blocks of R, G, and B colors as well as color difference components in the decoding method described later.
  • FIG. 8 is a block diagram illustrating a configuration of an image decoding apparatus according to an embodiment of the present invention.
  • the image decoding apparatus 800 is an apparatus for generating and decoding a subsample prediction value of a chrominance component using a motion vector value of a luminance component of a current block of an image. And a decoder 810, an inverse quantizer 820, an inverse transformer 830, an adder 840, and a predictor 850.
  • the decoder 810 extracts a quantized frequency transform block by decoding the bitstream.
  • the decoder 810 may decode and extract not only the quantized frequency transform block but also information necessary for decoding by decoding the encoded data.
  • Information necessary for decoding refers to information necessary for decoding the coded bit string in the encoded data (ie, the bitstream). For example, information about a block type, information about a motion vector, information about a transform and quantization type, and the like. It can be, and a variety of other information can be.
  • the decoder 810 decodes a bitstream, which is data encoded by the image encoding apparatus 600, extracts a quantized frequency transform block including pixel information of the current block of the image, and extracts the extracted prediction signal.
  • the necessary information is transmitted to the prediction unit 850.
  • the prediction unit 850 may predict the current block in the same manner as the prediction unit 610 of the image encoding apparatus 600 by using the information necessary for the prediction transmitted from the decoder 810.
  • the prediction unit 850 generates a prediction block for the color difference component from the value obtained by interpolating the values of the subsample components of the color difference component of the reference block referred to by the motion vector of the color difference component of the current block using an FIR filter and linear interpolation. do.
  • the motion vector value of the luminance component reconstructed in the bitstream is used, and the integer pixel value of the chrominance component of the reference block referred to by the motion vector is converted into a high precision FIR filter.
  • the prediction unit 850 of the image decoding apparatus 800 generates a subsample in the same manner as the prediction unit 610 of the image encoding apparatus 600 described above with reference to FIG. 6. Therefore, detailed description thereof will be omitted to avoid redundant description.
  • the inverse quantizer 820 inverse quantizes the quantized frequency transform block extracted from the bitstream by the decoder 810.
  • the inverse transformer 830 inversely transforms the frequency transform block inversely quantized by the inverse quantizer 820 into the time domain.
  • the adder 840 reconstructs the original pixel value of the current block by adding the residual pixel restored by the inverse transform by the inverse transformer 830 and the predicted pixel value by the predictor 850.
  • the current block reconstructed by the adder 840 may be transferred to the frame memory 860 and used by the predictor 850 to predict another block.
  • the frame memory 860 stores the reconstructed image to enable generation of the intra prediction block and the inter prediction block.
  • An image encoding / decoding apparatus may be configured by connecting a bitstream output terminal of the image encoding apparatus 600 of FIG. 6 to a bitstream input terminal of the image decoding apparatus 800 of FIG. 8.
  • a value obtained by interpolating a value of a subsample component of a chrominance component of a reference block referred to by a motion vector of a chrominance component of a current block using an FIR filter and linear interpolation An image of generating a prediction block for a chrominance component from the chrominance component, subtracting the prediction block from the chrominance component of the current block, generating a residual block, transforming and quantizing the residual block, and generating a quantized frequency transform block and encoding the bitstream
  • a quantized frequency transform block is generated from an encoder and a bitstream, and the quantized frequency transform block is inversely quantized and inverse transformed so that the color difference component of the reference block referred to by the motion vector of the color difference component of the current block to reconstruct and reconstruct the residual block.
  • a video decoder for generating a prediction block for the chrominance component from and reconstructing the chrominance component of the current block to be reconstructed by adding the reconstructed residual block and the generated prediction block.
  • the image encoder may be implemented by the image encoding apparatus 600 according to an embodiment of the present invention, and the image decoder may be implemented by the image decoding apparatus 800 according to the embodiment of the present invention.
  • FIG. 9 is a flowchart illustrating an image encoding method according to an embodiment of the present invention.
  • the image encoding apparatus 600 generates a subsample prediction value of the color difference component using the motion vector value of the luminance component of the current block of the image (S910), and the difference between the original pixel value and the prediction pixel value of the current block.
  • a subtraction step (S920) of generating a residual signal by calculating a value a conversion step (S930) of converting the generated residual signal into a frequency domain using a DCT transform or a wavelet transform, and quantizing the residual signal transformed into a frequency domain
  • the image is encoded through a quantization step S940 and an encoding step S950 of encoding the quantized frequency transform residual signal into a bitstream.
  • the prediction step (S910) is a function of the prediction unit 610
  • the subtraction step (S920) is a function of the subtraction unit 620
  • the transform step (S930) is a function of the transformer 630
  • FIG. 10 is a flowchart illustrating an image decoding method according to an embodiment of the present invention.
  • the video decoding apparatus 800 which receives and stores a bitstream of an image through a wired or wireless communication network or a cable, stores the current block of the image in order to reproduce the image according to a user's selection or an algorithm of another program being executed.
  • a subsample prediction value of the chrominance component is generated using the motion vector value, and the image is decoded and reconstructed.
  • the image decoding apparatus 800 decodes the bitstream and extracts a quantized frequency transform residual signal representing information about the pixel value of the current block of the image (S1010).
  • An inverse quantization step of inversely quantizing a quantized frequency transform residual signal, an inverse transform step of inversely transforming an inverse quantized frequency transform residual signal into a time domain, and a current block represented by a residual signal inversely transformed into a time domain and restored A prediction step (S1040) of generating a subsample prediction value of a chrominance component using the motion vector value of the luminance component of the predicted value of, and prediction of each pixel of the current block predicted in step S1040 and the residual signal of the current block restored in step S1040.
  • the bitstream transmitted through the addition step S1050 of restoring the original pixel value of the current block by adding the pixel value is decoded.
  • the decoding step (S1010) corresponds to the operation of the decoding unit 810
  • the inverse quantization step (S1020) corresponds to the operation of the inverse quantization unit 820
  • the inverse transform step (S1030) of the inverse transformer (830) corresponds to the prediction step (S1040)
  • the addition step (S1050) corresponds to the operation of the adding unit 840, and thus, detailed description thereof will be omitted.
  • An image encoding / decoding method may be realized by combining the image encoding method according to an embodiment of the present invention and the image decoding method according to an embodiment of the present invention.
  • a value obtained by interpolating a value of a subsample component of a chrominance component of a reference block referred to by a motion vector of a chrominance component of a current block using an FIR filter and linear interpolation An image of generating a prediction block for a chrominance component from the chrominance component, subtracting the prediction block from the chrominance component of the current block, generating a residual block, transforming and quantizing the residual block, and generating a quantized frequency transform block and encoding the bitstream
  • a color difference component of a reference block referred to by a motion vector of a color difference component of a current block to generate and quantize the quantized frequency transform block from the encoding step and the bitstream, and to inversely quantize and inverse transform the quantized frequency transform block.
  • the subsample components of are interpolated using FIR filter and linear interpolation. And generating a prediction block for the color difference component from the value, and adding the reconstructed residual block and the generated prediction block to reconstruct the color difference component of the current block to be reconstructed.
  • the image encoding step may be implemented as an image encoding step according to an embodiment of the present invention
  • the image decoding step may be implemented as an image decoding step according to an embodiment of the present invention.
  • all the components constituting the embodiment of the present invention may be implemented in one independent hardware, each of some or all of the components are selectively combined to some or all of the functions combined in one or a plurality of hardware. It may be implemented as a computer program having a program module for performing the operation. Codes and code segments constituting the computer program may be easily inferred by those skilled in the art. Such a computer program may be stored in a computer readable storage medium and read and executed by a computer, thereby implementing embodiments of the present invention.
  • the storage medium of the computer program may include a magnetic recording medium, an optical recording medium, a carrier wave medium, and the like.
  • encoding is applied to an interpolation that minimizes a difference between an actual component and a predicted component and encoding / decoding of the image using the encoding.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Un mode de réalisation de la présente invention se rapporte à un procédé et à un appareil destinés à coder/décoder une vidéo en utilisant un filtre à haute précision. Un mode de réalisation de la présente invention fournit le procédé et l'appareil destinés à coder/décoder la vidéo et dans le procédé destiné à coder/décoder la vidéo, la valeur d'un élément de sous-échantillon d'un bloc de référence, que référence le vecteur de mouvement d'un bloc courant, est codée et décodée en générant un bloc prédictif à partir d'une valeur qui est interpolée en utilisant un filtre FIR et une interpolation linéaire, de façon à interpoler de manière plus précise le bloc courant, en augmentant de ce fait l'efficacité de codage en réduisant la différence entre le bloc réel et le bloc prédictif et en améliorant de ce fait l'efficacité de compression.
PCT/KR2011/007418 2010-10-06 2011-10-06 Procédé et appareil destinés à coder/décoder une vidéo en utilisant un filtre à haute précision WO2012047047A2 (fr)

Priority Applications (5)

Application Number Priority Date Filing Date Title
CN201180054874.XA CN103210649B (zh) 2010-10-06 2011-10-06 使用高精度滤波器编码/解码视频的方法和设备
US13/857,708 US9420281B2 (en) 2010-10-06 2013-04-05 Method and apparatus for encoding/decoding video using high-precision filter
US15/031,366 US9706222B2 (en) 2010-10-06 2016-04-22 Method and apparatus for encoding/decoding video using high-precision filter
US15/031,378 US9602834B2 (en) 2010-10-06 2016-04-22 Method and apparatus for encoding/decoding video using high-precision filter
US15/371,215 US10158880B2 (en) 2010-10-06 2016-12-07 Method and apparatus for encoding/decoding video using high-precision filter

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR10-2010-0097547 2010-10-06
KR20100097547 2010-10-06
KR10-2011-0072196 2011-07-21
KR1020110072196A KR101341993B1 (ko) 2010-10-06 2011-07-21 고정밀 필터를 이용한 영상 부호화/복호화 방법 및 장치

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/857,708 Continuation US9420281B2 (en) 2010-10-06 2013-04-05 Method and apparatus for encoding/decoding video using high-precision filter

Publications (2)

Publication Number Publication Date
WO2012047047A2 true WO2012047047A2 (fr) 2012-04-12
WO2012047047A3 WO2012047047A3 (fr) 2012-06-28

Family

ID=45928243

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2011/007418 WO2012047047A2 (fr) 2010-10-06 2011-10-06 Procédé et appareil destinés à coder/décoder une vidéo en utilisant un filtre à haute précision

Country Status (1)

Country Link
WO (1) WO2012047047A2 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107211150A (zh) * 2015-01-16 2017-09-26 微软技术许可有限责任公司 动态更新质量至较高的色度采样率
US10523953B2 (en) 2012-10-01 2019-12-31 Microsoft Technology Licensing, Llc Frame packing and unpacking higher-resolution chroma sampling formats
CN113302934A (zh) * 2019-12-23 2021-08-24 Oppo广东移动通信有限公司 图像预测方法、编码器、解码器以及存储介质

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004015999A1 (fr) * 2002-08-08 2004-02-19 Matsushita Electric Industrial Co., Ltd. Procede de codage et decodage d'une image en mouvement
KR20100030119A (ko) * 2008-09-09 2010-03-18 에스케이 텔레콤주식회사 회전 움직임을 고려한 인터 예측 장치와 그를 이용한 영상 부호화/복호화 방법 및 장치
WO2010044569A2 (fr) * 2008-10-16 2010-04-22 에스케이텔레콤 주식회사 Procédé et appareil permettant de générer une trame de référence et procédé et appareil permettant le coder/décoder une image au moyen de la trame de référence

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004015999A1 (fr) * 2002-08-08 2004-02-19 Matsushita Electric Industrial Co., Ltd. Procede de codage et decodage d'une image en mouvement
KR20100030119A (ko) * 2008-09-09 2010-03-18 에스케이 텔레콤주식회사 회전 움직임을 고려한 인터 예측 장치와 그를 이용한 영상 부호화/복호화 방법 및 장치
WO2010044569A2 (fr) * 2008-10-16 2010-04-22 에스케이텔레콤 주식회사 Procédé et appareil permettant de générer une trame de référence et procédé et appareil permettant le coder/décoder une image au moyen de la trame de référence

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10523953B2 (en) 2012-10-01 2019-12-31 Microsoft Technology Licensing, Llc Frame packing and unpacking higher-resolution chroma sampling formats
CN107211150A (zh) * 2015-01-16 2017-09-26 微软技术许可有限责任公司 动态更新质量至较高的色度采样率
CN107211150B (zh) * 2015-01-16 2020-02-28 微软技术许可有限责任公司 用于动态更新质量至较高的色度采样率的方法、计算机系统以及计算机可读介质
CN113302934A (zh) * 2019-12-23 2021-08-24 Oppo广东移动通信有限公司 图像预测方法、编码器、解码器以及存储介质
CN113302934B (zh) * 2019-12-23 2022-10-18 Oppo广东移动通信有限公司 图像预测方法、编码器、解码器以及存储介质

Also Published As

Publication number Publication date
WO2012047047A3 (fr) 2012-06-28

Similar Documents

Publication Publication Date Title
WO2013070006A1 (fr) Procédé et appareil de codage et de décodage vidéo faisant appel à un mode de saut
WO2013002549A2 (fr) Procédés et appareil de codage/décodage d'une image
WO2012077960A2 (fr) Procédé et dispositif de codage/décodage d'une image par inter-prédiction en utilisant un bloc aléatoire
WO2013005941A2 (fr) Appareil et procédé de codage et de décodage d'une image
WO2009157669A2 (fr) Procédé de prédiction intra-trame et appareil, procédé de codage/décodage d'image et appareil utilisant ce procédé
KR101648058B1 (ko) 고정밀 필터를 이용한 영상 부호화/복호화 방법 및 장치
WO2011145819A2 (fr) Dispositif et procédé de codage/décodage d'image
WO2013002550A2 (fr) Méthode et appareil de codage/décodage par décision de mode de l'unité de codage grande vitesse
WO2010027182A2 (fr) Procédé et dispositif de codage/décodage d'images au moyen de pixels aléatoires dans un sous-bloc
WO2012096550A2 (fr) Procédé et dispositif de codage/décodage d'image utilisant une prédiction intra bidirectionnelle
WO2009157665A2 (fr) Procédé de prédiction intra-trame et appareil utilisant la transformation par bloc, procédé de codage/décodage d'image et appareil utilisant ce procédé
WO2010044569A2 (fr) Procédé et appareil permettant de générer une trame de référence et procédé et appareil permettant le coder/décoder une image au moyen de la trame de référence
WO2012093854A2 (fr) Procédé et dispositif de codage et de décodage utilisant une prédiction interne parallèle par une unité de codage
WO2012015275A2 (fr) Procédé et dispositif de codage/décodage d'image à l'aide d'une prédiction fragmentée de bloc
WO2013109039A1 (fr) Procédé et appareil de codage/décodage d'images utilisant la prédiction de poids
WO2010047499A2 (fr) Appareil de codage/décodage vidéo, filtre de dégroupage et procédé de filtrage basé sur des directions de prédiction intra, et support d'enregistrement
WO2012091461A2 (fr) Procédé et dispositif de codage / décodage d'une image en utilisant des vecteurs caractéristiques de blocs environnants
WO2012033344A2 (fr) Procédé et appareil destinés à coder/décoder des images à l'aide de la sélection efficace d'un groupe de mode de prédiction intra
WO2012011672A2 (fr) Procédé et dispositif de codage/décodage d'images par mode de saut étendu
WO2012077959A2 (fr) Procédé et dispositif pour coder/décoder une image dans un domaine spatial pour une composante de bruit
WO2012021040A2 (fr) Procédé et dispositif de codage/décodage d'image ayant un mode de filtrage amovible
WO2012081917A2 (fr) Procédé et appareil de génération d'informations de mouvement codées/de récupération d'informations de mouvement à l'aide d'une intégration d'informations de mouvement, et procédé et appareil de codage/de décodage d'images qui utilisent ledit procédé et ledit appareil
WO2011052990A2 (fr) Procédé et appareil de codage/décodage d'images sur la base d'une résolution adaptative
WO2018074626A1 (fr) Procédé et appareil de codage vidéo utilisant un filtre d'interpolation adaptatif
WO2012047047A2 (fr) Procédé et appareil destinés à coder/décoder une vidéo en utilisant un filtre à haute précision

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11830932

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase in:

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 17/07/13)

122 Ep: pct application non-entry in european phase

Ref document number: 11830932

Country of ref document: EP

Kind code of ref document: A2