WO2007105900A1 - Method, medium, and system encoding and/or decoding moving pictures by adaptively applying optimal prediction modes - Google Patents
Method, medium, and system encoding and/or decoding moving pictures by adaptively applying optimal prediction modes Download PDFInfo
- Publication number
- WO2007105900A1 WO2007105900A1 PCT/KR2007/001217 KR2007001217W WO2007105900A1 WO 2007105900 A1 WO2007105900 A1 WO 2007105900A1 KR 2007001217 W KR2007001217 W KR 2007001217W WO 2007105900 A1 WO2007105900 A1 WO 2007105900A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- prediction mode
- image
- residues
- generating
- component
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 85
- 230000033001 locomotion Effects 0.000 claims description 31
- 239000013598 vector Substances 0.000 claims description 30
- 230000001131 transforming effect Effects 0.000 claims description 10
- 230000009466 transformation Effects 0.000 description 93
- 238000013139 quantization Methods 0.000 description 20
- 230000002123 temporal effect Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000006866 deterioration Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 238000011426 transformation method Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/007—Transform coding, e.g. discrete cosine transform
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/109—Selection of coding mode or of prediction mode among a plurality of temporal predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/11—Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/147—Data rate or code amount at the encoder output according to rate distortion criteria
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/15—Data rate or code amount at the encoder output by monitoring actual compressed data size at the memory before deciding storage at the transmission buffer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/186—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/19—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding using optimisation based on Lagrange multipliers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
Definitions
- One or more embodiments of the present invention relate to a method, medium, and system encoding and/or decoding moving pictures, and more particularly, to a method, medium, and system encoding and/or decoding moving pictures in the field of the H.264/MPEG-4 AVC (Advanced Video Coding) Fidelity Range Extensions (FRExt) standardization.
- H.264/MPEG-4 AVC Advanced Video Coding
- FRExt Fidelity Range Extensions
- RGB encoding and decoding technologies according to the H.264/MPEG-4 AVC FRExt standard, are not satisfactory for providing high encoding and decoding efficiencies and thus cannot be used in moving picture reproduction. Disclosure of Invention Technical Solution
- One or more embodiments of the present invention provide a method, medium, and system encoding and/or decoding moving pictures using RGB encoding technology, according to the H.264/MPEG-4 AVC (Advanced Video Coding) Fidelity Range Extensions (FRExt) standard, to provide high encoding and decoding efficiencies.
- H.264/MPEG-4 AVC Advanced Video Coding
- FRExt Fidelity Range Extensions
- encoding and/or decoding of moving pictures may be performed by adaptively applying an optimal prediction mode to corresponding macro blocks, which correspond to each other, of the color components of a current image for each macro block, so that encoding and/or decoding efficiencies increase.
- various encoding methods such as a single prediction mode, a multiple prediction mode, inter prediction, intra prediction, residue transformation, RCT, IPP transformation, RCP transformation, etc., may be selectively used to encode each color component of the current image for each macro block, thereby maximizing encoding and decoding efficiencies.
- FlG. 1 illustrates a moving picture encoding system, according to an embodiment of the present invention
- FlG. 2 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to an embodiment of the present invention
- FlG. 3 illustrates a method of dividing macro blocks for inter prediction
- FlG. 4 illustrates prediction directions defined for intra prediction
- FlG. 5 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to another embodiment of the present invention
- FlG. 6 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to still another embodiment of the present invention
- FIGS. 7 A and 7B illustrate a 5-tap filter and a 3-tap filter, respectively, which may be used in embodiments of the present invention
- FlG. 8 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to another embodiment of the present invention
- FlG. 9 illustrates a restored image generation unit, such as that illustrated in FlG. 1, according to an embodiment of the present invention
- FlG. 10 illustrates a moving picture decoding system, according to an embodiment of the present invention
- FlG. 11 illustrates a predicted image generation unit, such as that illustrated in FlG.
- FlG. 12 illustrates a restored image generation unit, such as that illustrated in FlG.
- FlG. 13 illustrates a residue inverse transformation portion, such as that illustrated in
- FIGS. 14A and 14B illustrate a moving picture encoding method, according to an embodiment of the present invention
- FlG. 15 illustrates a moving picture decoding method, according to an embodiment of the present invention.
- embodiments of the present invention include a predicted image generating method, including selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, and generating a predicted image for the current image according to the selected prediction mode.
- embodiments of the present invention include at least one medium including computer readable code to control at least one processing element to implement a predicted image generating method, the method including selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, and generating a predicted image for the current image according to the selected prediction mode.
- embodiments of the present invention include a predicted image generating system, including a selection unit selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, and a generation unit generating a predicted image for the current image according to the selected prediction mode.
- embodiments of the present invention include an encoding method, including selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, generating a predicted image for the current image according to the selected prediction mode, generating residues, of each color component, which correspond to differences between the current image and the predicted image, and generating a bitstream by encoding the generated residues.
- embodiments of the present invention include at least one medium including computer readable code to control at least one processing element to implement embodiments of the present invention.
- embodiments of the present invention include an encoding system, including a selection unit to select a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, a generation unit to generate a predicted image for the current image according to the selected prediction mode and to generate residues, of each color component, which correspond to differences between the current image and the predicted image, and an encoding unit to generate a bitstream by encoding the generated residues.
- embodiments of the present invention include a predicted image generating method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, and generating a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
- embodiments of the present invention include at least one medium including computer readable code to control at least one processing element to implement a predicted image generating method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, and generating a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
- embodiments of the present invention include a predicted image generating system to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the system including a decoding unit to restore prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, and a generation unit to generate a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
- embodiments of the present invention include a decoding method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, generating residues that correspond to differences between the current image and a reference image for the current image, according to the prediction mode identified by the restored prediction mode information, generating a predicted image according to the prediction mode identified by the restored prediction mode information, and generating a restored image that corresponds to a sum of the generated residues and the generated predicted image.
- embodiments of the present invention include at least one medium including computer readable code to control at least one processing element to implement a decoding method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, generating residues that correspond to differences between the current image and a reference image for the current image, according to the prediction mode identified by the restored prediction mode information, generating a predicted image according to the prediction mode identified by the restored prediction mode information, and generating a restored image that corresponds to a sum of the generated residues and the generated predicted image.
- embodiments of the present invention include a decoding system to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the system including a decoding unit to restore prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, a first generation unit to generate a predicted image for the current image according to the prediction mode identified by the restored prediction mode information, and a second generation unit to generate residues that correspond to differences between the current image and the predicted image according to the prediction mode identified by the restored prediction mode information and to generate a restored image corresponding to a sum of the generated residues and the generated predicted image.
- Mode for Invention to restore prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, a first generation unit to generate a predicted image for the current image according to the prediction mode identified by the restored prediction mode information, and a second generation unit to generate residues that correspond to differences between the current image and the predicted image according to the prediction
- a current image should be considered as meaning a target image in a current moving picture that is to be encoded and/or decoded
- a reference image should be considered as meaning an image that is referred to when encoding or decoding the current image.
- the reference image is a previous image that precedes the current image, though the reference image may also be a subsequent image, and/or a plurality of reference images may be used.
- FIG. 1 is a block diagram of a moving picture encoding system, according to an embodiment of the present invention.
- the moving picture encoding system may include an optimal mode selection unit 110, a residue generation unit 120, a frequency domain transformation unit 130, a quantization unit 400, an entropy encoding unit 150, an inverse quantization unit 160, a frequency domain inverse transformation unit 170, and a restored image generation unit 180, for example.
- the optimal mode selection unit 110 selects an optimal prediction mode for the macro blocks, which correspond to each other, of the color components of a current image based on the characteristics of a sample image.
- macro blocks, which correspond to each other, of one color component also correspond to macro blocks of all the other remaining color components.
- the optimal mode selection unit 110 may select one of a single prediction mode, which may be commonly applied to the macro blocks, which correspond to each other, of the color components of the current image, and a multiple prediction mode, which may be independently applied to the macro blocks, which correspond to each other, of the color components of the current image.
- the optimal mode selection unit 110 selects the single prediction mode, a corresponding prediction mode for generating second residues that correspond to differences between first residues of the color components can be selected.
- the prediction mode may be, for example, an 'RCT mode' for performing Residual Color Transformation (RCT), an 'IPP transformation mode' for performing Inter-Plane Prediction (IPP) transformation, or an 'RCP transformation mode' for performing Residual Color Prediction (RCP), noting that alternative embodiments are equally available.
- RCT Residual Color Transformation
- IPP Inter-Plane Prediction
- RCP Residual Color Prediction
- the referenced sample image may be one of the previous images that precede the current image.
- the optimal mode selection unit 110 may, thus, select all possible prediction modes for the sample image one by one, and compare the results of encoding performed according to the selected prediction mode, thereby selecting a prediction mode that is optimal for the macro blocks, which correspond to each other, of the color components of the sample image. The selected prediction mode may then be used as the optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image.
- the optimal mode selection unit 110 may select a prediction mode that leads to a minimum size of the resultant bitstream from the encoding of the sample image, and/or a minimum amount of distortion in quality between the original sample image and a restored sample image, for example.
- the size of the resultant bitstream obtained as the result of the encoding of the sample image corresponds to the size of the bitstream generated by the entropy encoding unit 150, for example, with the restored sample image corresponding to a restored image generated by the restored image generation unit 180, for example.
- the optimal mode selection unit 110 may select the optimal prediction mode using a Lagrangian optimization method, for example. In other words, the optimal mode selection unit 110 may calculate the distortion in the image from the average of the sums of the products of squaring the differences between the original image and the restored image obtained using the below Equation 1, for example.
- D indicates the degree of picture quality distortion
- p indicates a pixel value in the current image
- q indicates a pixel value in the previous image
- i indicates a pixel index in a current macro block of the current image.
- the optimal mode selection unit 110 may multiply the size of the bitstream 'R', e.g., the number of bits in the bitstream, by a constant ' ⁇ ' and add the degree of picture quality distortion 'D' to the product so as to compensate for a difference, in units between the degree of picture quality distortion and the size of the bitstream, thereby calculating a final value 'L' to be used for selecting the optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image.
- R indicates the size of the bitstream
- ⁇ indicates a predetermined constant
- the final value L can be calculated for each prediction method, and the prediction method with the smallest L may be selected from among the prediction methods used for the calculation.
- the optimal mode selection unit 110 may select the single prediction mode when the similarity between the color components of the sample image is high, and select the multiple prediction mode when the similarity between the color components of the sample is low.
- the residue generation unit 120 may generate a predicted image for the macro blocks, which correspond to each other, of the color components of the current image according to the prediction mode selected by the optimal mode selection unit 110, and generate residues that correspond to differences between the current image and the predicted image.
- the residue generation unit 120 may generate the predicted image for the macro blocks, which correspond to each other, of the color components of the current image according to the prediction mode selected by the optimal mode selection unit 110, generate first residues that correspond to differences between the current image and the predicted image, and generate second residues that correspond to differences between the first residues for the macro blocks, which correspond to each other, of the color components, noting that alternative embodiments are also available.
- the frequency domain transformation unit 130 may further transform residues generated by the residue generation unit 120, which are color space residues, into values in the frequency domain.
- the frequency domain transformation unit 130 may transform the second residues generated by the residue generation unit 120, which are color space residues, into values in the frequency domain.
- DHT Discrete Hadamard Transformation
- DCT Discrete Cosine Transformation
- the quantization unit 140 quantizes the values transformed by the frequency domain transformation unit 130, e.g., the quantization unit 140 divides the frequency component values, which are the results of the transformation by the frequency domain transformation unit 130, by a quantization parameter, and approximates the results to integer values.
- the entropy encoding unit 150 further generates a bitstream by entropy-encoding the values quantized by the quantization unit 140.
- the entropy encoding unit 150 generates a bitstream including information representing which prediction mode was used in the moving picture encoding, e.g., as selected by the optimal mode selection unit 110, by entropy- encoding the quantized values together with such prediction mode information.
- CAVLC Context- Adaptive Variable Length Coding
- CABAC Context-Adaptive Binary Arithmetic Coding
- the entropy encoding unit 150 generates the bitstream including information representing the prediction mode selected by the optimal mode selection unit 110 in a macro block header for each color component of the current image.
- a moving picture decoding system illustrated in FIG. 10 receives and decodes this bitstream so that the prediction mode used by the moving picture encoding system is identified.
- the entropy encoding unit 150 may generate a bitstream including information representing only one prediction mode identically selected by the optimal mode selection unit 110 for all the macro blocks making up a single sequence or for all the macro blocks making up a single picture, at a sequence level or a picture level, which are upper levels relative to the macro block level. Accordingly, with all corresponding macro blocks being encoded with the same prediction mode, the information recorded in corresponding macro block headers can be omitted, thereby increasing the encoding efficiency.
- the entropy encoding unit 150 may generate a bitstream including information representing only one prediction mode identically selected by the optimal mode selection unit 110 for those corresponding macro blocks making up a single sequence or a single picture, at a sequence level or a picture level, and including information representing the corresponding prediction modes, which are selected by the optimal mode selection unit 110 for the remaining corresponding macro blocks, in macro block headers, thereby increasing the encoding efficiency.
- the inverse quantization unit 160 may further perform inverse quantization on the values quantized by the quantization unit 140, e.g., inverse quantization unit 160 may restore the frequency component values by multiplying the integer values approximated by the quantization unit 140 by a quantization parameter.
- the frequency domain inverse transformation unit 170 may then restore the residues corresponding to the differences between the current image and the predicted image by transforming the frequency component values, which are restored by the inverse quantization unit 160, which are frequency domain values, into values in the color space.
- the frequency domain inverse transformation unit 170 restores the aforementioned second residues, which are frequency domain values corresponding to the differences between the aforementioned first residues of the color components, into the values in the color space.
- the restored image generation unit 180 may generate a restored image that corresponds to the sum of the predicted image generated by the residue generation unit 120 and the residues restored by the frequency domain inverse transformation unit 170.
- the restored image generation unit 180 may generate first residues corresponding to the sums of the second residues restored by the frequency domain inverse transformation unit 170 and generate the restored image corresponding to the sum of the predicted image generated by the residue generation unit 120 and the generated first residues.
- FlG. 2 illustrates a residue generation unit 120, according to an embodiment of the present invention.
- the residue generation unit 120 may include a single mode residue generation portion 1211 and a multiple mode residue generation portion 1212, for example.
- the single mode residue generation portion 1211 may perform spatial prediction, such as single intra prediction, for removing special redundancy in the current image, according to the single prediction mode, or perform temporal prediction, such as single inter prediction, for removing temporal redundancy between the current image and the reference image, according to the single prediction mode. Subsequently, the single mode residue generation unit 1211 may generate a residue corresponding to a difference between the current image and the predicted image.
- the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into blocks of a predetermined size which is commonly applied to the macro blocks, which correspond to each other, of the color components of the current image, determine motion vectors between the reference image and the current images, which are commonly applied to the divided blocks, which correspond to each other, of the color components of the current image, and generate the predicted image for the current image from the reference image using the determined motion vectors, thereby performing single inter prediction.
- the single mode residue generation unit 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into equal blocks, determine motion vectors that are commonly applied to the divided blocks, which correspond to each other, of the color components, and generate the predicted image for the current image from the reference image using the determined motion vectors, thereby performing single inter prediction.
- FlG. 3 illustrates a method of dividing macro blocks for inter prediction.
- a 16x16 macro block may be divided into blocks of various sizes of 16x16, 16x8, 8x16, 8x16, 8x8, etc., and a motion vector for each divided block can be determined. Furthermore, after an 8x8 block is further divided into blocks of sizes of 8x8, 8x4, 4x8, 4x4, etc., a motion vector for each divided block may be determined. As macro blocks are divided into blocks of smaller size, more refined motions between the current image and the reference image can be included in the residues.
- the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of a Y component, a Co component, and a Cg component equally, for example, into an 8x8 size, and determine a motion vector which is commonly applied to the divided blocks, which correspond to each other, of the Y component, the Co component, and the Cg component.
- the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of each of a R component, a G component, and a B component equally, for example, into an 8x8 size, and determine a motion vector which is commonly applied to the divided blocks, which correspond to each other, of the R component, the G component, and the B component.
- the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into blocks of a predetermined size which is commonly applied to the macro blocks, which correspond to each other, of the color components of the current image, determine a prediction direction that is commonly applied to the divided blocks, which correspond to each other, of the color components, predict blocks making up the current image from adjacent pixels in the restored image, which is generated by the restored image generation unit 180, using the determined prediction direction, and generate a predicted image made up by the predicted blocks, thereby performing single intra prediction.
- the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into equal blocks, determine a prediction direction which is commonly applied to the divided blocks, which correspond to each other, of the color components, predict blocks making up the current image from adjacent pixels in the restored image, which is generated by the restored image generation portion 180, using the determined prediction direction, and generate a predicted image made up by the predicted blocks, thereby performing single intra prediction.
- FlG. 4 illustrates prediction directions defined for intra prediction, according to an embodiment of the present invention.
- a predicted image may be generated using 9 prediction directions for each divided block.
- the predicted image may be generated using 4 prediction directions for a 16x16 macro block.
- the former case in order to predict blocks of a 4x4 size, i.e., Pa, Pb, through Pq, spatially adjacent pixels PO, Pl, through Pl 2 in the restored image are used.
- the blocks Pa, Pb, through Pq are predicted from the adjacent pixels PO, Pl, through P12 using 9 prediction directions from 0 to 9.
- Pl, P2, P3, and P4 are projected in a vertical direction corresponding to the prediction direction 0, so that Pa, Pe, Pi, and Pm are predicted from Pl ; Pb, Pf, Pj, and Pn are predicted from P2; Pc, Pg, Pk, and Pd are predicted from P3; and Pd, Ph, Pl, and Pq are predicted from P4.
- Pa, Pb, through Pq can be predicted using other prediction directions as described above.
- the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of each of a Y component, a Co component, and a Cg component equally, for example, into blocks of 4x4 size, and determine a prediction direction that is commonly applied to the divided blocks, which correspond to each other, of the Y component, the Co component, and the Cg component.
- the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of each of a R component, a G component, and a B component equally, for example, into blocks of 4x4 size, and determine a prediction direction that is commonly applied to the divided blocks, which correspond to each other, of the R component, the G component, and the B component.
- the single mode residue generation portion 1211 applies the same temporal prediction method and the same spatial prediction method to all the color components, the similarity between the residues of the color components become higher.
- a block size, a motion vector, a prediction direction, etc. may be commonly applied to all the color components, there is no need to encode and transmit such information for each color component. In other words, it is sufficient to encode and transmit such information only once for all the color components, thereby increasing the overall encoding efficiency.
- the multiple mode residue generation portion 1212 may generate a prediction mode by performing temporal prediction, such as multiple inter prediction, for removing temporal redundancy between the current image and the reference image according to the multiple prediction mode, or by performing spatial prediction, such as multiple intra prediction, for removing special redundancy in the current image according to a prediction mode that is independently applied to the macro blocks, which correspond to each other, of the color components of the current image.
- temporal prediction such as multiple inter prediction
- spatial prediction such as multiple intra prediction
- the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components into blocks of sizes which may be independently applied to each of the macro blocks, which correspond to each other, of the color components of the current image, determine motion vectors between the reference image and the current image, which are independently applied to the divided blocks of each color component, and generate a predicted image for the current image using the determined motion vectors, thereby performing multiple inter prediction.
- the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components of the current image into blocks of different sizes, determine different motion vectors for the divided blocks of each color component, and generate a predicted image for the current image from the reference image using the determined motion vectors.
- the multiple mode residue generation portion 1212 may perform inter prediction according to prediction modes that are independently applied to the macro blocks, which correspond to each other, of the color components of the current image, the multiple mode residue generation portion 1212 may still divide all the macro blocks of the color components of the current image equally and may determine motion vectors which are the same for the divided blocks of each color component.
- the multiple mode residue generation portion 1212 may divide the macro blocks of the Y component, the Co component, and the Cg component into blocks of different sizes, e.g., with the macro blocks of the Y component divided into blocks of a 4x4 size, the macro blocks of the Co component divided into blocks of an 8x8 size, and the macro blocks of the Cg component divided into blocks of an 8x8 size, and determine different motion vectors for the divided blocks of each color component, i.e., the Y component, the Co component, and Cg component.
- the multiple mode residue generation portion 1212 may divide the macro blocks of the R component, the G component, and the B component into blocks of different sizes, e.g., with the macro blocks of the R component divided into blocks of an 8x8 size, the macro blocks of the G component divided into blocks of a 4x4 size, and the macro blocks of the B component divided into blocks of an 8x8 size, and determine different motion vectors for the divided blocks of each color component, i.e., the R component, the G component, and the B component.
- the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components into blocks of sizes that are independently applied to the macro blocks, which correspond to each other, of the color components of the current image, determine prediction directions that are independently applied to the divided blocks of each color component, predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 180 using the determined prediction directions, and generate a predicted image made up by the predicted blocks, thereby performing multiple intra prediction.
- the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components of the current image into blocks of different sizes, determine different prediction directions for the divided blocks of each color component, predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 180 using the generated prediction directions, and generate the predicted image made up by the predicted blocks, thereby performing multiple intra prediction.
- the multiple mode residue generation portion 1212 may still divide the macro blocks, which correspond to each other, of the color components of the current image equally and may determine motion vectors which are commonly applied to the divided blocks of the color components.
- the multiple mode residue generation portion 1212 may divide the macro blocks of the Y component, the Co component, and the Cg component into blocks of different sizes, e.g., with the macro blocks of the Y component divided into blocks of a 4x4 size, the macro blocks of the Co component divided into blocks of a 16x16 size, and the macro blocks of the Cg component divided into blocks of a 16x16 size, and determine different prediction directions for the divided blocks of each color component, i.e., the Y component, the Co component, and the Cg component.
- the multiple mode residue generation portion 1212 may divide the macro blocks of the R component, the G component, and the B component into blocks of different sizes, e.g., with the macro blocks of the R component divided into blocks of a 16x16 size, the macro blocks of the G component divided into blocks of a 4x4 size, and the macro blocks of the B component divided into blocks of a 16x16 size, and determine different prediction directions for the divided blocks of each color components, i.e., the R component, the G component, and the B component.
- the multiple mode residue generation portion 1212 may perform multiple inter prediction or multiple intra prediction on the macro blocks, which correspond to each other, of the color components of the current image.
- the multiple mode residue generation portion 1212 may perform multiple inter prediction on the macro blocks of one of the color components of the current image and perform multiple intra prediction on the macro blocks of the other color components.
- the multiple mode residue generation portion 1212 can perform temporal prediction and spatial prediction using different temporal prediction and spatial prediction methods on the different color components, if the similarity between the color components is low, independent encoding methods which are most suitable for each color component can be used for effective prediction encoding, thereby increasing the overall encoding efficiency. However, it may be sufficient to encode any of the block sizes, motion vectors, and prediction directions, which are independently applied to each color component, only once for transmission. Thus, the encoding efficiency increases.
- FlG. 5 illustrates a residue generation unit 120, such as that illustrated in FlG. 1, according to another embodiment of the present invention.
- the residue generation unit 120 may include a single mode residue generation portion 1221, a residue transformation portion 1222, and a multiple mode residue generation portion 1223.
- the single mode residue generation portion 1221 and the multiple mode residue generation portion 1223 may perform similar operations, respectively, as the single mode residue generation portion 1211 and the multiple mode residue generation portion 1212 illustrated in FlG. 2.
- the residue transformation portion 1222 may eliminate the redundancy between the color components.
- the former can be referred to as 'first residues', and the latter can be referred to as 'second residues'.
- the residue transformation portion 1222 may generate second residues that correspond to differences between the first residues generated by the single mode residue generation portion 1221.
- FlG. 6 illustrates a residue generation unit 120, such as that illustrated in FlG. 1, ace ording to another embodiment of the present invention.
- the residue generation unit 120 may include a single mode residue generation portion 1231, a residual color transformation (RCT) portion 1232, an inter-plane prediction (IPP) transformation portion 1233, a residual color prediction (RCP) transformation portion 1234, and a multiple mode residue generation portion 1235, for example.
- the single mode residue generation portion 1231 and the multiple mode residue generation portion 1235 may perform similar operations, respectively, as the single mode residue generation portion 1211 and the multiple mode residue generation portion 1212 illustrated in FlG. 2.
- Methods of generating the second residues that correspond to the differences between the first residues generated by the single mode residue generation portion 1231 include, for example, RCT, IPP transformation, and RCP transformation.
- RCT can be used for generating the second residues in the YCoCg color space
- IPP transformation is for generating the second residues in the RGB color space.
- RCP transformation which is similar to IPP transformation, when a residue of the G component is used as a predictor, the same transformation as IPP transformation can be performed after noise is removed by predetermined filtering.
- the RCT portion 1232 may generate second residues that correspond to the differences between the first residues of the Y component, the Co component, and the Cg component in the YCoCg color space using the below Equation 3, for example.
- ⁇ X denotes first residues
- ⁇ 2X denotes second residues
- ' ' denotes a right shift operation, which is subsequently equivalent to division by 2
- variable t is used for the purpose of temporary calculation.
- the IPP transformation portion 1233 may generate second residues that correspond to the differences between the first residues of the R component, the G component, and the B component in the RGB color space according to the IPP transformation mode in view of the below Equation 4, for example.
- Equation 4 is effective when the G component includes a large amount of image information.
- the second residues can be calculated using the R component or the B component as a dominant component.
- the RCP transformation portion 1234 generates second residues that correspond to the differences between the first residues of the R component, the G component, and the B component.
- Equation 5 [100] Equation 5 :
- the RCP transformation portion 1234 After removing noise from each of the R component, the G component and the B component in the RGB color space using a 3-tap filter, expressed in the below Equation 6, for example, the RCP transformation portion 1234 generates second residues that correspond to the differences between the first residues of the R component, the G component, and the B component.
- Equation 6 [103] Equation 6 :
- Equation 7 [107] [Math.7]
- FIGS. 7 A and 7B illustrate a 5-tap filter and a 3-tap filter, respectively, which are used in an embodiment of the present invention.
- a 5-tap filter used in an embodiment of the present invention uses 4 neighboring pixels around a pixel that is used as a predictor
- a 3-tap filter used in an embodiment of the present invention uses two, left and right pixels on either side of a pixel that is used as a predictor.
- weights may be used as in Equations 5 and 6.
- FlG. 8 illustrates a residue generation unit 120, such as that illustrated in FlG. 1, according to an embodiment of the present invention.
- the residue generation unit 120 may include a multiple mode residue generation portion 1241 and a residue transformation portion 1242, for example.
- the multiple mode residue generation portion 1241 may perform similar operations as the multiple mode residue generation portion 1212 illustrated in FlG. 2, for example.
- the residue transformation portion 1242 may eliminate the redundancy between the color components.
- the former residue can be referred to as 'first residues', and the latter can be referred to as 'second residues'.
- the residue transformation portion 1242 may generate second residues that correspond to the differences between the first residues generated by the multiple mode residue generation portion 1241, according to the residue transformation mode.
- the compression efficiency in the residue transformation portion 1242 may be lower than the compression efficiency in the residue transformation portion 1222 illustrated in FlG. 5.
- FlG. 9 illustrates a restored image generation unit 180, such as that illustrated in FlG. 1, according to an embodiment of the present invention.
- the restored image generation unit 180 may include a residue inverse transformation portion 181 and a prediction compensation portion 182, for example.
- the residue inverse transformation portion 181 may generate first residues that correspond to the sums of the second residues restored by the frequency domain inverse transformation portion 170.
- the residue inverse transformation portion 181 may generate the first residues of the Y component, the Co component, and the Cg component, which correspond to the sums of the second residues of the Y component, the Co component, and the Cg component, respectively, using the below Equation 8, for example.
- ⁇ R 1 : ⁇ B ' + ⁇ 2 B
- ⁇ X' denotes restored first residues
- ⁇ 2X' denotes restored second residues
- the residue inverse transformation portion 181 may generate the first residues of the R component, the G component, and the B component, which correspond to the sums of the second residues of the R component, the G component, and the B component, respectively, using the below Equation 9, for example.
- ⁇ X' denotes restored first residues
- ⁇ X' denotes restored second residues
- the prediction compensation portion 182 may generate a restored image that corresponds to the sum of the predicted image generated by the residue generation unit 120 and the first residues generated by the residue inverse transformation portion 181.
- the prediction compensation portion 182 may generate a restored image in the YCoCg color space by calculating the sum of the predicted image generated by the residue generation unit 120 and the first residues generated by the residue inverse transformation portion 181.
- the prediction compensation portion 182 may generate a restored image in the RGB color space by calculating the sum of the predicted image generated by the residue generation unit 120 and the first residues generated by the residue inverse transformation portion 181.
- alternate embodiments are also available.
- FIG. 10 illustrates a moving picture decoding system, according to an embodiment of the present invention.
- the moving picture decoding system may include an entropy decoding unit 210, an inverse quantization unit 220, a frequency domain inverse transformation unit 230, a predicted image generation unit 240, and a restored image generation unit 250, for example.
- the entropy decoding unit 210 may restore integer values corresponding to the current image and information representing the optimal prediction mode for the blocks of each color component of the current image by performing entropy-decoding on a bitstream, such as a bitstream output from the moving picture encoding system illustrated in FIG. 1.
- the optimal prediction mode for the blocks of each color component of the current image information can be the prediction mode used in the moving picture encoding system.
- the inverse quantization unit 220 may restore the frequency component values by performing inverse quantization on the integer values restored by the entropy decoding unit 210. In other words, the inverse quantization unit 220 restores the frequency component values by multiplying the integer values restored by the entropy decoding unit 210 by a quantization parameter.
- the frequency domain inverse transformation unit 230 may generate residues that correspond to differences between the current image and the predicted image by transforming the frequency component values restored by the inverse quantization unit 220, which are frequency domain values, into values in the color space.
- the frequency domain inverse transformation unit 230 restores second residues that correspond to the differences between the first residues of the color components by transforming the frequency component values restored by the inverse quantization unit 220, which are frequency domain values, into values in the color space.
- the predicted image generation unit 240 may generate a predicted image for the current image according to the prediction mode indicated by the information restored for the macro blocks, which correspond to each other, of the color components by the entropy decoding unit 210.
- the restored image generation unit 250 may generate a restored image that corresponds to the sum of the predicted image generated by the predicted image generation unit 240 and the residues restored by the frequency domain inverse transformation unit 230.
- the restored image generation unit 250 may generate first residues that correspond to the sums of the second residues restored by the frequency domain inverse transformation unit 230 and generate a restored image that corresponds to the sum of the predicted image generated by the predicted image generation unit 240 and the generated first residues.
- FlG. 11 illustrates a predicted image generation unit 240, such as that illustrated in FlG. 10, according to an embodiment of the present invention.
- the predicted image generation unit 240 may include a single mode prediction image generation portion 241 and a multiple mode prediction image generation portion 242, for example.
- the single mode prediction image generation portion 241 may generate a predicted image by performing single intra prediction or single inter prediction, for example.
- the single mode prediction image generation portion 241 may generate a predicted image for the current image from a reference image using motion vectors between the reference image and the current image, which are commonly applied to the blocks, which correspond to each other, of the color components.
- the single mode prediction image generation portion 241 may generate the predicted image for the current image from the reference image using the motion vectors that are commonly applied to the blocks, which are divided from the macro blocks, which correspond to each other, of the color components of the current image and correspond to each other.
- the single mode prediction image generation portion 241 may further predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 250 using prediction directions that are commonly applied to the blocks, which correspond to each other, of the color components, which are divided equally from the macro blocks, which correspond to each other, of the color components of the current image, and generate a predicted image constituted by the predicted blocks.
- the multiple mode prediction image generation portion 242 may generate a predicted image by performing multiple inter prediction or multiple intra prediction.
- the multiple mode prediction image generation portion 242 may generate a predicted image for the current image from a reference image using motion vectors between the reference image and the current image, which are independently applied to the blocks of each color component, which are divided into different sizes from the macro blocks, which correspond to each other, of the color components of the current image.
- the multiple mode prediction image generation portion 242 may generate the predicted image for the current image from the reference image using different motion vectors for the blocks of each color component, which are divided into blocks of different sizes from the macro blocks, which correspond to each other, of the color components of the current image.
- the multiple mode prediction image generation portion 242 may predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 250 using prediction directions that are commonly applied to the blocks, which correspond to each other, of the color components, wherein the blocks are divided to different sizes from the macro blocks, which correspond to each other, of the color components of the current image, and generate a predicted image made up by the predicted blocks, thereby performing multiple intra prediction.
- the multiple mode prediction image generation portion 242 may predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 250 using different prediction directions for the blocks of each color component, which are divided to different sizes from the macro blocks, which correspond to each other, of the color components of the current image and generate the predicted image made up by the predicted blocks.
- FlG. 12 illustrates a restored image generation unit 250, such as that shown in FlG. 10, according to an embodiment of the present invention.
- the restored image generation unit 250 may include a residue inverse transformation portion 251 and a prediction compensation portion 252, for example.
- the residue inverse transformation portion 251 may generate first residues that correspond to the sums of the second residues restored by the frequency domain inverse transformation unit 230.
- the residue inverse transformation portion 251 may generate the first residues of the Y component, the Co component, and the Cg component, which correspond to the sums of the second residues of each respective color component, i.e., the Y component, the Co component, and the Cg component, using the above Equation 8, for example.
- the residue inverse transformation portion 251 may generate the first residues of the R component, the G component, and the B component, which correspond to the sums of the second residues of each respective color component, i.e., the R component, the G component, and the B component, using the above Equation 9, for example.
- the prediction compensation portion 252 may generate a restored image that corresponds to the sum of the predicted image generated by the predicted image generation unit 240 and the first residues generated by the residue inverse transformation portion 251.
- the prediction compensation portion 252 may generate a restored image in the YCoCg color space by calculating the sums of the predicted image of each respective color component, i.e., the Y component, the Co component, and the Cg component, and the first residues generated by the residue inverse transformation portion 251.
- the prediction compensation portion 252 may generate a restored image in the RGB color space by calculating the sums of the predicted image for each respective color component, i.e., the R component, the G component, and the B component, which is generated by the predicted image generation unit 240, and the first residues generated by the residue inverse transformation portion 251.
- FIG. 13 illustrates a residue inverse transformation portion 251, such as that shown in FIG. 12, according to an embodiment of the present invention.
- the residue inverse transformation portion 251 may include an RCT portion 2511, an IPP transformation portion 2512, and an RCP transformation portion 2513, for example.
- the RCT portion 2511 may generate the first residues of each of the Y component, the Co component, and the Cg component, which respectively correspond to the sums of the second residues of each of the Y component, Co component, and the Cg component, using the above Equation 8, for example.
- 2512 may generate the first residues of each of the R component, the G component, and the B component, which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, using the above Equation 9, for example.
- 2513 may generate the first residues of each of the R component, the G component, and the B component, which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, using the above Equation 9, for example.
- FIGS. 14A and 14B illustrate a moving picture encoding method, according to an embodiment of the present invention.
- a moving picture encoding method may include operations which may be time-serially performed in the moving picture encoding system shown in FIG. 1, for example. Therefore, for some embodiments of the present invention, the above-descriptions of the moving picture encoding system illustrated in FIG. 1 can also apply to the moving picture encoding method, even if some descriptions thereof are not directly provided below.
- one of the possible prediction modes for a sample image may be selected. As illustrated in FIG. 14 A, especially, when operation 1409 returns to operation 1401, one from among possible prediction modes except for a previously selected prediction mode may be selected.
- a predicted image may be generated for the current image for each macro block of each color component, according to the prediction mode selected in operation 1401, and residues that correspond to the differences between the current image and the predicted image may be generated.
- a predicted image for the current image for each macro block of each color component may be generated according to the prediction mode selected in operation 1401, and second residues that correspond to the differences between the first residues of the color components may be generated.
- the residues generated in operation 1402 may be transformed, with the residues being color space residues, into values in the frequency domain.
- the second residues generated in operation 1402 may be transformed, with the residues being color space residues, into values in the frequency domain.
- the values transformed in operation 1403 may be quantized.
- a bitstream may be generated by entropy-encoding the values quantized in operation 1404 and information representing the prediction mode selected in operation 1401.
- frequency component values may further be restored by inverse- quantizing the values quantized in operation 1404.
- residues that correspond to differences between the current image and the predicted image may be restored by transforming the frequency component values restored in operation 1406, which are frequency domain values, into values in the color space.
- second residues that correspond to differences between the first residues of each color component may be restored by transforming the frequency component values restored in operation 1406, which are frequency domain values, into values in the color space.
- a restored image that corresponds to the sum of the predicted image generated in operation 1402 and the residues restored in operation 1407 may be generated.
- first residues that correspond to the sums of the second residues restored in operation 1407 may be generated, and a restored image that corresponds to the sum of the predicted image generated in operation 1402 and the generated first residues may be generated.
- operation 1409 in one embodiment, whether operations 1401 through 1408 have been performed for all the prediction modes may be determined. In this embodiment, if operations 1401 through 1408 have been performed on all the prediction modes, the method may proceed to operation 1410 illustrated in FlG. 14B. Otherwise, the method may returns to operation 1401.
- an optimal prediction mode may be selected for the macro blocks, which correspond to each other, of the color components of the current image based on the characteristics of the sample images, i.e., the results of operations 1401 through 1409.
- the moving picture encoding system may select a prediction mode that leads to a minimum size of a resultant bitstream and/or a minimum amount of quality distortion between the sample image and the restored image generated in operation 1408, as an optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image.
- a predicted image for the current image for each macro block of each color component according to the prediction mode selected in operation 1410 may be generated, and residues that correspond to the differences between the current image and the predicted image may be generated.
- a predicted image for the current image for each macro block of each color component may be generated according to the prediction mode selected in operation 1410, first residues that correspond to the differences between the current image and the predicted image may be generated, and second residues that correspond to the differences between the first residues of each color component may be generated.
- the residues generated in operation 1411 may be transformed, with the residues being color space residues, into values in the frequency domain.
- the second residues generated in operation 1411 may be transformed, with the residues being color space residues, into values in the frequency domain.
- the values transformed in operation 1412 may be quantized.
- a bitstream may be generated by entropy-encoding the values quantized in operation 1413 and information representing the prediction mode selected in operation 1410.
- the frequency component values may be restored by inverse- quantizing the values quantized in operation 1413.
- residues that correspond to the differences between the current image and the predicted image may be restored by transforming the frequency component values restored in operation 1415, which are frequency domain values, into values in the color space.
- second residues that correspond to differences between the first residues of each color component may be restored by transforming the frequency component values restored in operation 1415, which are frequency domain values, into values in the color space.
- a restored image that corresponds to the sum of the predicted image generated in operation 1411 and the residues restored in operation 1416 may be generated.
- first residues that correspond to the sums of the second residues of each color component restored in operation 1416 may be generated, and a restored image that corresponds to the sum of the predicted image generated in operation 1411 and the generated first residues may be generated.
- FlG. 15 illustrates a moving picture decoding method, according to an embodiment of the present invention.
- the illustrated operations may be time-serially performed, e.g., such as in the moving picture decoding system shown in FlG. 10. Accordingly, for such an embodiment, the above-descriptions of the moving picture decoding system illustrated in FlG. 10 should also be applied to such a moving picture decoding method, even if some descriptions thereof are not directly provided below.
- integer values corresponding to the current image and information representing the optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image may be restored by entropy- decoding a bitstream, such as that output from the moving picture encoding system illustrated in FlG. 1.
- the frequency component values may be restored by inverse- quantizing the integer values restored in operation 1501.
- residues that correspond to the differences between the current image and the predicted image may be generated by transforming the frequency component values restored in operation 1502, which are frequency domain values, into values in the color space.
- second residues that correspond to the differences between the first residues of each color component may be restored by transforming the frequency component values restored in operation 1502, which are frequency domain values, into values in the color space.
- the below operation 1505 may be performed. Otherwise, the below operation 1506 may be performed.
- first residues that correspond to the sums of the second residues of each color component restored in operation 1503 may be generated.
- the information restored in operation 1501 represents that the RCT transformation mode was applied during encoding
- the first residues of each of the Y component, the Co component, and the Cg component which respectively correspond to the sums of the second residues of each of the Y component, the Co component, and the Cg component, may be generated using the above Equation 8, for example.
- the first residues of each of the R component, the G component, and the B component which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, may be generated using the above Equation 9, for example.
- the RCP transformation portion 2511 may generate the first residues of each of the R component, the G component, and the B component, which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, using the above Equation 9, for example.
- operation 1506 if the information restored in operation 1501 is determined to represent the single prediction mode, operation 1507 may be performed. If the information restored in operation 1501 is determined to represent the multiple prediction mode, operation 1508 may be performed.
- a predicted image may be generated according to the single prediction mode, i.e., using motion vectors that are commonly applied to all the divided blocks of the color components of the current image, which are divided equally, or using prediction directions that are independently applied to all the divided blocks of the color components.
- a predicted image may be generated according to the multiple prediction mode, i.e., using motion vectors that are independently applied to the divided blocks of each color component of the current image, which are divided into blocks of different sizes, or using prediction directions that are independently applied to the divided blocks of each color component.
- a restored image that corresponds to the sum of the predicted image generated in operation 1507 or 1508 and the residues restored in operation 1503 may be generated.
- a restored image that corresponds to the sum of the predicted image generated in operation 1507 or 1508 and the first residues generated in operation 1505 may be generated.
- embodiments of the present invention can also be implemented through computer readable code/instructions in/on a medium, e.g., a computer readable medium, to control at least one processing element to implement any above described embodiment.
- a medium e.g., a computer readable medium
- the medium can correspond to any medium/media permitting the storing and/or transmission of the computer readable code.
- the computer readable code can be recorded/transferred on a medium in a variety of ways, with examples of the medium including recording media, such as magnetic storage media (e.g., ROM, floppy disks, hard disks, etc.) and optical recording media (e.g., CD-ROMs, or DVDs), and transmission media such as carrier waves, as well as through the Internet, for example.
- the medium may further be a signal, such as a resultant signal or bitstream, according to embodiments of the present invention.
- the media may also be a distributed network, so that the computer readable code is stored/ transferred and executed in a distributed fashion.
- the processing element could include a processor or a computer processor, and processing elements may be distributed and/or included in a single device.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Discrete Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Color Television Systems (AREA)
Abstract
A method, medium, and system encoding and/or decoding a moving picture. The moving picture encoding method may include selecting a prediction mode that is optimal for the macro blocks, which correspond to each other, of the color components of a current image based on the characteristics of a predetermined image, generating a predicted image for the current image according to the selected prediction mode, and encoding a moving picture using the predicted image. An optimal prediction mode can be adaptively applied to the macro blocks, which correspond to each other, of the color components, thereby increasing the moving picture's encoding and decoding efficiencies.
Description
Description
METHOD, MEDIUM, AND SYSTEM ENCODING AND/OR DECODING MOVING PICTURES BY ADAPTIVELY
APPLYING OPTIMAL PREDICTION MODES
Technical Field
[1] One or more embodiments of the present invention relate to a method, medium, and system encoding and/or decoding moving pictures, and more particularly, to a method, medium, and system encoding and/or decoding moving pictures in the field of the H.264/MPEG-4 AVC (Advanced Video Coding) Fidelity Range Extensions (FRExt) standardization. Background Art
[2] A recently new RGB encoding technology called 'residual color transformation' has been developed during the H.264/MPEG-4 AVC Fidelity Range Extensions (FRExt) standardization process. This technology prevents picture quality deterioration from occurring during the transformation from the RGB color space to the YCbCr color space. However, RGB encoding and decoding technologies, according to the H.264/MPEG-4 AVC FRExt standard, are not satisfactory for providing high encoding and decoding efficiencies and thus cannot be used in moving picture reproduction. Disclosure of Invention Technical Solution
[3] One or more embodiments of the present invention provide a method, medium, and system encoding and/or decoding moving pictures using RGB encoding technology, according to the H.264/MPEG-4 AVC (Advanced Video Coding) Fidelity Range Extensions (FRExt) standard, to provide high encoding and decoding efficiencies.
[4] Additional aspects and/or advantages of the invention will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the invention. Advantageous Effects
[5] According to one or more embodiments of the present invention, encoding and/or decoding of moving pictures may be performed by adaptively applying an optimal prediction mode to corresponding macro blocks, which correspond to each other, of the color components of a current image for each macro block, so that encoding and/or decoding efficiencies increase. In particular, various encoding methods, such as a single prediction mode, a multiple prediction mode, inter prediction, intra prediction, residue transformation, RCT, IPP transformation, RCP transformation, etc., may be selectively used to encode each color component of the current image for each macro
block, thereby maximizing encoding and decoding efficiencies. Description of Drawings
[6] FlG. 1 illustrates a moving picture encoding system, according to an embodiment of the present invention; [7] FlG. 2 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to an embodiment of the present invention;
[8] FlG. 3 illustrates a method of dividing macro blocks for inter prediction;
[9] FlG. 4 illustrates prediction directions defined for intra prediction;
[10] FlG. 5 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to another embodiment of the present invention; [11] FlG. 6 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to still another embodiment of the present invention; [12] FIGS. 7 A and 7B illustrate a 5-tap filter and a 3-tap filter, respectively, which may be used in embodiments of the present invention; [13] FlG. 8 illustrates a residue generation unit, such as that illustrated in FlG. 1, according to another embodiment of the present invention; [14] FlG. 9 illustrates a restored image generation unit, such as that illustrated in FlG. 1, according to an embodiment of the present invention; [15] FlG. 10 illustrates a moving picture decoding system, according to an embodiment of the present invention; [16] FlG. 11 illustrates a predicted image generation unit, such as that illustrated in FlG.
10, according to an embodiment of the present invention; [17] FlG. 12 illustrates a restored image generation unit, such as that illustrated in FlG.
10, according to an embodiment of the present invention; [18] FlG. 13 illustrates a residue inverse transformation portion, such as that illustrated in
FlG. 12, according to an embodiment of the present invention; [19] FIGS. 14A and 14B illustrate a moving picture encoding method, according to an embodiment of the present invention; and [20] FlG. 15 illustrates a moving picture decoding method, according to an embodiment of the present invention.
Best Mode [21] To achieve the above and/or other aspects and advantages, embodiments of the present invention include a predicted image generating method, including selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, and generating a predicted image for the current image according to the selected prediction mode. [22] To achieve the above and/or other aspects and advantages, embodiments of the
present invention include at least one medium including computer readable code to control at least one processing element to implement a predicted image generating method, the method including selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, and generating a predicted image for the current image according to the selected prediction mode.
[23] To achieve the above and/or other aspects and advantages, embodiments of the present invention include a predicted image generating system, including a selection unit selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, and a generation unit generating a predicted image for the current image according to the selected prediction mode.
[24] To achieve the above and/or other aspects and advantages, embodiments of the present invention include an encoding method, including selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, generating a predicted image for the current image according to the selected prediction mode, generating residues, of each color component, which correspond to differences between the current image and the predicted image, and generating a bitstream by encoding the generated residues.
[25] To achieve the above and/or other aspects and advantages, embodiments of the present invention include at least one medium including computer readable code to control at least one processing element to implement embodiments of the present invention.
[26] To achieve the above and/or other aspects and advantages, embodiments of the present invention include an encoding system, including a selection unit to select a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image, a generation unit to generate a predicted image for the current image according to the selected prediction mode and to generate residues, of each color component, which correspond to differences between the current image and the predicted image, and an encoding unit to generate a bitstream by encoding the generated residues.
[27] To achieve the above and/or other aspects and advantages, embodiments of the present invention include a predicted image generating method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, and generating a predicted image for the
current image according to the prediction mode identified by the restored prediction mode information.
[28] To achieve the above and/or other aspects and advantages, embodiments of the present invention include at least one medium including computer readable code to control at least one processing element to implement a predicted image generating method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, and generating a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
[29] To achieve the above and/or other aspects and advantages, embodiments of the present invention include a predicted image generating system to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the system including a decoding unit to restore prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, and a generation unit to generate a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
[30] To achieve the above and/or other aspects and advantages, embodiments of the present invention include a decoding method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, generating residues that correspond to differences between the current image and a reference image for the current image, according to the prediction mode identified by the restored prediction mode information, generating a predicted image according to the prediction mode identified by the restored prediction mode information, and generating a restored image that corresponds to a sum of the generated residues and the generated predicted image.
[31] To achieve the above and/or other aspects and advantages, embodiments of the present invention include at least one medium including computer readable code to control at least one processing element to implement a decoding method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of
each color component of a corresponding current image, the method including restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, generating residues that correspond to differences between the current image and a reference image for the current image, according to the prediction mode identified by the restored prediction mode information, generating a predicted image according to the prediction mode identified by the restored prediction mode information, and generating a restored image that corresponds to a sum of the generated residues and the generated predicted image.
[32] To achieve the above and/or other aspects and advantages, embodiments of the present invention include a decoding system to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the system including a decoding unit to restore prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode, a first generation unit to generate a predicted image for the current image according to the prediction mode identified by the restored prediction mode information, and a second generation unit to generate residues that correspond to differences between the current image and the predicted image according to the prediction mode identified by the restored prediction mode information and to generate a restored image corresponding to a sum of the generated residues and the generated predicted image. Mode for Invention
[33] Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. Embodiments are described below to explain the present invention by referring to the figures.
[34] In describing embodiments herein, a current image should be considered as meaning a target image in a current moving picture that is to be encoded and/or decoded, and a reference image should be considered as meaning an image that is referred to when encoding or decoding the current image. Generally, the reference image is a previous image that precedes the current image, though the reference image may also be a subsequent image, and/or a plurality of reference images may be used.
[35] FIG. 1 is a block diagram of a moving picture encoding system, according to an embodiment of the present invention.
[36] Referring to FIG. 1, the moving picture encoding system may include an optimal mode selection unit 110, a residue generation unit 120, a frequency domain transformation unit 130, a quantization unit 400, an entropy encoding unit 150, an inverse quantization unit 160, a frequency domain inverse transformation unit 170, and a
restored image generation unit 180, for example.
[37] The optimal mode selection unit 110 selects an optimal prediction mode for the macro blocks, which correspond to each other, of the color components of a current image based on the characteristics of a sample image. Here, macro blocks, which correspond to each other, of one color component also correspond to macro blocks of all the other remaining color components.
[38] For example, the optimal mode selection unit 110 may select one of a single prediction mode, which may be commonly applied to the macro blocks, which correspond to each other, of the color components of the current image, and a multiple prediction mode, which may be independently applied to the macro blocks, which correspond to each other, of the color components of the current image.
[39] When the optimal mode selection unit 110 selects the single prediction mode, a corresponding prediction mode for generating second residues that correspond to differences between first residues of the color components can be selected.
[40] For example, when the optimal mode selection unit 110 selects the single prediction mode, the prediction mode may be, for example, an 'RCT mode' for performing Residual Color Transformation (RCT), an 'IPP transformation mode' for performing Inter-Plane Prediction (IPP) transformation, or an 'RCP transformation mode' for performing Residual Color Prediction (RCP), noting that alternative embodiments are equally available. Such RCT, IPP, and RCP transformations will be described in greater detail below.
[41] In an embodiment of the present invention, the referenced sample image may be one of the previous images that precede the current image. In order to select the optimal prediction mode for the macro blocks, the optimal mode selection unit 110 may, thus, select all possible prediction modes for the sample image one by one, and compare the results of encoding performed according to the selected prediction mode, thereby selecting a prediction mode that is optimal for the macro blocks, which correspond to each other, of the color components of the sample image. The selected prediction mode may then be used as the optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image.
[42] In particular, the optimal mode selection unit 110 may select a prediction mode that leads to a minimum size of the resultant bitstream from the encoding of the sample image, and/or a minimum amount of distortion in quality between the original sample image and a restored sample image, for example. As will be described later, the size of the resultant bitstream obtained as the result of the encoding of the sample image corresponds to the size of the bitstream generated by the entropy encoding unit 150, for example, with the restored sample image corresponding to a restored image generated by the restored image generation unit 180, for example.
[43] In an embodiment of the present invention, the optimal mode selection unit 110 may select the optimal prediction mode using a Lagrangian optimization method, for example. In other words, the optimal mode selection unit 110 may calculate the distortion in the image from the average of the sums of the products of squaring the differences between the original image and the restored image obtained using the below Equation 1, for example.
[44] Equation 1 :
[45] [Math.l]
[46] Here, D indicates the degree of picture quality distortion, p indicates a pixel value in the current image, q indicates a pixel value in the previous image, and i indicates a pixel index in a current macro block of the current image.
[47] Thus, in an embodiment, and as described below in Equation 2, the optimal mode selection unit 110 may multiply the size of the bitstream 'R', e.g., the number of bits in the bitstream, by a constant 'λ' and add the degree of picture quality distortion 'D' to the product so as to compensate for a difference, in units between the degree of picture quality distortion and the size of the bitstream, thereby calculating a final value 'L' to be used for selecting the optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image.
[48] Equation 2
[49] [Math.2]
[50] Here, R indicates the size of the bitstream, and λ indicates a predetermined constant.
The final value L can be calculated for each prediction method, and the prediction method with the smallest L may be selected from among the prediction methods used for the calculation.
[51] According to an experiment performed using such a Lagrangian optimization method, the optimal mode selection unit 110 may select the single prediction mode when the similarity between the color components of the sample image is high, and select the multiple prediction mode when the similarity between the color components of the sample is low.
[52] The residue generation unit 120 may generate a predicted image for the macro blocks, which correspond to each other, of the color components of the current image according to the prediction mode selected by the optimal mode selection unit 110, and generate residues that correspond to differences between the current image and the
predicted image. Alternatively, the residue generation unit 120 may generate the predicted image for the macro blocks, which correspond to each other, of the color components of the current image according to the prediction mode selected by the optimal mode selection unit 110, generate first residues that correspond to differences between the current image and the predicted image, and generate second residues that correspond to differences between the first residues for the macro blocks, which correspond to each other, of the color components, noting that alternative embodiments are also available.
[53] The frequency domain transformation unit 130 may further transform residues generated by the residue generation unit 120, which are color space residues, into values in the frequency domain. Alternatively, in an embodiment, the frequency domain transformation unit 130 may transform the second residues generated by the residue generation unit 120, which are color space residues, into values in the frequency domain. According to the H.264/MPEG-4 AVC standard, Discrete Hadamard Transformation (DHT), Discrete Cosine Transformation (DCT)-based integer transformation, etc., are suggested as color space to frequency domain transformation methods, noting that alternative embodiments may equally be available.
[54] The quantization unit 140 quantizes the values transformed by the frequency domain transformation unit 130, e.g., the quantization unit 140 divides the frequency component values, which are the results of the transformation by the frequency domain transformation unit 130, by a quantization parameter, and approximates the results to integer values.
[55] The entropy encoding unit 150 further generates a bitstream by entropy-encoding the values quantized by the quantization unit 140. In particular, in an embodiment of the present invention, the entropy encoding unit 150 generates a bitstream including information representing which prediction mode was used in the moving picture encoding, e.g., as selected by the optimal mode selection unit 110, by entropy- encoding the quantized values together with such prediction mode information. According to the H.264/MPEG-4 AVC standard, for example, Context- Adaptive Variable Length Coding (CAVLC), Context-Adaptive Binary Arithmetic Coding (CABAC), etc., have been suggested as entropy encoding methods.
[56] In particular, the entropy encoding unit 150 generates the bitstream including information representing the prediction mode selected by the optimal mode selection unit 110 in a macro block header for each color component of the current image. A moving picture decoding system illustrated in FIG. 10 receives and decodes this bitstream so that the prediction mode used by the moving picture encoding system is identified.
[57] When the prediction modes selected by the optimal mode selection unit 110 for the macro blocks, which correspond to each other, of the color components of the current
image are identical to each other, the entropy encoding unit 150 may generate a bitstream including information representing only one prediction mode identically selected by the optimal mode selection unit 110 for all the macro blocks making up a single sequence or for all the macro blocks making up a single picture, at a sequence level or a picture level, which are upper levels relative to the macro block level. Accordingly, with all corresponding macro blocks being encoded with the same prediction mode, the information recorded in corresponding macro block headers can be omitted, thereby increasing the encoding efficiency.
[58] Similarly, when some of the prediction modes selected by the optimal mode selection unit 110 for the macro blocks of the color components of the current image are identical, the entropy encoding unit 150 may generate a bitstream including information representing only one prediction mode identically selected by the optimal mode selection unit 110 for those corresponding macro blocks making up a single sequence or a single picture, at a sequence level or a picture level, and including information representing the corresponding prediction modes, which are selected by the optimal mode selection unit 110 for the remaining corresponding macro blocks, in macro block headers, thereby increasing the encoding efficiency.
[59] The inverse quantization unit 160 may further perform inverse quantization on the values quantized by the quantization unit 140, e.g., inverse quantization unit 160 may restore the frequency component values by multiplying the integer values approximated by the quantization unit 140 by a quantization parameter.
[60] The frequency domain inverse transformation unit 170 may then restore the residues corresponding to the differences between the current image and the predicted image by transforming the frequency component values, which are restored by the inverse quantization unit 160, which are frequency domain values, into values in the color space. Alternatively, in an embodiment, the frequency domain inverse transformation unit 170 restores the aforementioned second residues, which are frequency domain values corresponding to the differences between the aforementioned first residues of the color components, into the values in the color space.
[61] The restored image generation unit 180 may generate a restored image that corresponds to the sum of the predicted image generated by the residue generation unit 120 and the residues restored by the frequency domain inverse transformation unit 170. Alternatively, in an embodiment, the restored image generation unit 180 may generate first residues corresponding to the sums of the second residues restored by the frequency domain inverse transformation unit 170 and generate the restored image corresponding to the sum of the predicted image generated by the residue generation unit 120 and the generated first residues.
[62] FlG. 2 illustrates a residue generation unit 120, according to an embodiment of the
present invention.
[63] Referring to FlG. 2, the residue generation unit 120 may include a single mode residue generation portion 1211 and a multiple mode residue generation portion 1212, for example.
[64] When the prediction mode selected by the optimal mode selection unit 110 is the single prediction mode, the single mode residue generation portion 1211 may perform spatial prediction, such as single intra prediction, for removing special redundancy in the current image, according to the single prediction mode, or perform temporal prediction, such as single inter prediction, for removing temporal redundancy between the current image and the reference image, according to the single prediction mode. Subsequently, the single mode residue generation unit 1211 may generate a residue corresponding to a difference between the current image and the predicted image.
[65] In particular, the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into blocks of a predetermined size which is commonly applied to the macro blocks, which correspond to each other, of the color components of the current image, determine motion vectors between the reference image and the current images, which are commonly applied to the divided blocks, which correspond to each other, of the color components of the current image, and generate the predicted image for the current image from the reference image using the determined motion vectors, thereby performing single inter prediction.
[66] For example, the single mode residue generation unit 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into equal blocks, determine motion vectors that are commonly applied to the divided blocks, which correspond to each other, of the color components, and generate the predicted image for the current image from the reference image using the determined motion vectors, thereby performing single inter prediction.
[67] FlG. 3 illustrates a method of dividing macro blocks for inter prediction. Referring to
FlG. 3, for inter prediction, a 16x16 macro block may be divided into blocks of various sizes of 16x16, 16x8, 8x16, 8x16, 8x8, etc., and a motion vector for each divided block can be determined. Furthermore, after an 8x8 block is further divided into blocks of sizes of 8x8, 8x4, 4x8, 4x4, etc., a motion vector for each divided block may be determined. As macro blocks are divided into blocks of smaller size, more refined motions between the current image and the reference image can be included in the residues.
[68] For example, in a YCoCg color space, the single mode residue generation portion
1211 may divide the macro blocks, which correspond to each other, of a Y component, a Co component, and a Cg component equally, for example, into an 8x8 size, and
determine a motion vector which is commonly applied to the divided blocks, which correspond to each other, of the Y component, the Co component, and the Cg component. As another example, in a RGB color space, the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of each of a R component, a G component, and a B component equally, for example, into an 8x8 size, and determine a motion vector which is commonly applied to the divided blocks, which correspond to each other, of the R component, the G component, and the B component.
[69] In addition, the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into blocks of a predetermined size which is commonly applied to the macro blocks, which correspond to each other, of the color components of the current image, determine a prediction direction that is commonly applied to the divided blocks, which correspond to each other, of the color components, predict blocks making up the current image from adjacent pixels in the restored image, which is generated by the restored image generation unit 180, using the determined prediction direction, and generate a predicted image made up by the predicted blocks, thereby performing single intra prediction.
[70] In other words, for example, the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of the color components of the current image into equal blocks, determine a prediction direction which is commonly applied to the divided blocks, which correspond to each other, of the color components, predict blocks making up the current image from adjacent pixels in the restored image, which is generated by the restored image generation portion 180, using the determined prediction direction, and generate a predicted image made up by the predicted blocks, thereby performing single intra prediction.
[71] FlG. 4 illustrates prediction directions defined for intra prediction, according to an embodiment of the present invention.
[72] Referring to FlG. 4, after a 16x16 macro block has been divided into blocks of 4x4 size, a predicted image may be generated using 9 prediction directions for each divided block. Alternatively, in an embodiment, the predicted image may be generated using 4 prediction directions for a 16x16 macro block. In particular, in the former case, in order to predict blocks of a 4x4 size, i.e., Pa, Pb, through Pq, spatially adjacent pixels PO, Pl, through Pl 2 in the restored image are used. The blocks Pa, Pb, through Pq are predicted from the adjacent pixels PO, Pl, through P12 using 9 prediction directions from 0 to 9. For example, in a prediction direction 0, adjacent pixels Pl, P2, P3, and P4 are projected in a vertical direction corresponding to the prediction direction 0, so that Pa, Pe, Pi, and Pm are predicted from Pl ; Pb, Pf, Pj, and Pn are predicted from P2; Pc,
Pg, Pk, and Pd are predicted from P3; and Pd, Ph, Pl, and Pq are predicted from P4. Similarly, Pa, Pb, through Pq can be predicted using other prediction directions as described above.
[73] For example, in a YCoCg color space, for example, the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of each of a Y component, a Co component, and a Cg component equally, for example, into blocks of 4x4 size, and determine a prediction direction that is commonly applied to the divided blocks, which correspond to each other, of the Y component, the Co component, and the Cg component. As another example, in a RGB color space, the single mode residue generation portion 1211 may divide the macro blocks, which correspond to each other, of each of a R component, a G component, and a B component equally, for example, into blocks of 4x4 size, and determine a prediction direction that is commonly applied to the divided blocks, which correspond to each other, of the R component, the G component, and the B component.
[74] As described above, since the single mode residue generation portion 1211 applies the same temporal prediction method and the same spatial prediction method to all the color components, the similarity between the residues of the color components become higher. In addition, since a block size, a motion vector, a prediction direction, etc., may be commonly applied to all the color components, there is no need to encode and transmit such information for each color component. In other words, it is sufficient to encode and transmit such information only once for all the color components, thereby increasing the overall encoding efficiency.
[75] When the prediction mode selected by the optimal mode selection unit 110 is the multiple prediction mode, the multiple mode residue generation portion 1212 may generate a prediction mode by performing temporal prediction, such as multiple inter prediction, for removing temporal redundancy between the current image and the reference image according to the multiple prediction mode, or by performing spatial prediction, such as multiple intra prediction, for removing special redundancy in the current image according to a prediction mode that is independently applied to the macro blocks, which correspond to each other, of the color components of the current image.
[76] In particular, the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components into blocks of sizes which may be independently applied to each of the macro blocks, which correspond to each other, of the color components of the current image, determine motion vectors between the reference image and the current image, which are independently applied to the divided blocks of each color component, and generate a predicted image for the current image using the determined motion vectors, thereby
performing multiple inter prediction.
[77] In other words, the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components of the current image into blocks of different sizes, determine different motion vectors for the divided blocks of each color component, and generate a predicted image for the current image from the reference image using the determined motion vectors. Of course, since the multiple mode residue generation portion 1212 may perform inter prediction according to prediction modes that are independently applied to the macro blocks, which correspond to each other, of the color components of the current image, the multiple mode residue generation portion 1212 may still divide all the macro blocks of the color components of the current image equally and may determine motion vectors which are the same for the divided blocks of each color component.
[78] For example, in the YCoCg color space, the multiple mode residue generation portion 1212 may divide the macro blocks of the Y component, the Co component, and the Cg component into blocks of different sizes, e.g., with the macro blocks of the Y component divided into blocks of a 4x4 size, the macro blocks of the Co component divided into blocks of an 8x8 size, and the macro blocks of the Cg component divided into blocks of an 8x8 size, and determine different motion vectors for the divided blocks of each color component, i.e., the Y component, the Co component, and Cg component. In addition, for example, in the RGB color space, the multiple mode residue generation portion 1212 may divide the macro blocks of the R component, the G component, and the B component into blocks of different sizes, e.g., with the macro blocks of the R component divided into blocks of an 8x8 size, the macro blocks of the G component divided into blocks of a 4x4 size, and the macro blocks of the B component divided into blocks of an 8x8 size, and determine different motion vectors for the divided blocks of each color component, i.e., the R component, the G component, and the B component.
[79] In addition, the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components into blocks of sizes that are independently applied to the macro blocks, which correspond to each other, of the color components of the current image, determine prediction directions that are independently applied to the divided blocks of each color component, predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 180 using the determined prediction directions, and generate a predicted image made up by the predicted blocks, thereby performing multiple intra prediction.
[80] In other words, the multiple mode residue generation portion 1212 may divide the macro blocks, which correspond to each other, of the color components of the current
image into blocks of different sizes, determine different prediction directions for the divided blocks of each color component, predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 180 using the generated prediction directions, and generate the predicted image made up by the predicted blocks, thereby performing multiple intra prediction. Of course, since the multiple mode residue generation portion 1212 performs intra prediction according to prediction modes that are independently applied to the macro blocks, which correspond to each other, of the color components of the current image, the multiple mode residue generation portion 1212 may still divide the macro blocks, which correspond to each other, of the color components of the current image equally and may determine motion vectors which are commonly applied to the divided blocks of the color components.
[81] For example, in the YCoCg color space, the multiple mode residue generation portion 1212 may divide the macro blocks of the Y component, the Co component, and the Cg component into blocks of different sizes, e.g., with the macro blocks of the Y component divided into blocks of a 4x4 size, the macro blocks of the Co component divided into blocks of a 16x16 size, and the macro blocks of the Cg component divided into blocks of a 16x16 size, and determine different prediction directions for the divided blocks of each color component, i.e., the Y component, the Co component, and the Cg component. In addition, for example, in the RGB color space, the multiple mode residue generation portion 1212 may divide the macro blocks of the R component, the G component, and the B component into blocks of different sizes, e.g., with the macro blocks of the R component divided into blocks of a 16x16 size, the macro blocks of the G component divided into blocks of a 4x4 size, and the macro blocks of the B component divided into blocks of a 16x16 size, and determine different prediction directions for the divided blocks of each color components, i.e., the R component, the G component, and the B component.
[82] Furthermore, the multiple mode residue generation portion 1212 may perform multiple inter prediction or multiple intra prediction on the macro blocks, which correspond to each other, of the color components of the current image. For example, the multiple mode residue generation portion 1212 may perform multiple inter prediction on the macro blocks of one of the color components of the current image and perform multiple intra prediction on the macro blocks of the other color components.
[83] As described above, since the multiple mode residue generation portion 1212 can perform temporal prediction and spatial prediction using different temporal prediction and spatial prediction methods on the different color components, if the similarity between the color components is low, independent encoding methods which are most
suitable for each color component can be used for effective prediction encoding, thereby increasing the overall encoding efficiency. However, it may be sufficient to encode any of the block sizes, motion vectors, and prediction directions, which are independently applied to each color component, only once for transmission. Thus, the encoding efficiency increases.
[84] FlG. 5 illustrates a residue generation unit 120, such as that illustrated in FlG. 1, according to another embodiment of the present invention.
[85] Referring to FlG. 5, the residue generation unit 120 may include a single mode residue generation portion 1221, a residue transformation portion 1222, and a multiple mode residue generation portion 1223. The single mode residue generation portion 1221 and the multiple mode residue generation portion 1223 may perform similar operations, respectively, as the single mode residue generation portion 1211 and the multiple mode residue generation portion 1212 illustrated in FlG. 2.
[86] However, even after inter prediction or intra prediction is performed in the singl Ie mode residue generation portion 1221, redundancy between the color components sun remains. The residue transformation portion 1222 may eliminate the redundancy between the color components. In order to distinguish the residues generated by single mode residue generation portion 1221 and the residues generated by the residue transformation portion 1222, the former can be referred to as 'first residues', and the latter can be referred to as 'second residues'.
[87] If the prediction mode selected by the optimal mode selection unit 110 is the residue transformation mode, the residue transformation portion 1222 may generate second residues that correspond to differences between the first residues generated by the single mode residue generation portion 1221.
[88] FlG. 6 illustrates a residue generation unit 120, such as that illustrated in FlG. 1, ace ording to another embodiment of the present invention.
[89] Referring to FlG. 6, the residue generation unit 120 may include a single mode residue generation portion 1231, a residual color transformation (RCT) portion 1232, an inter-plane prediction (IPP) transformation portion 1233, a residual color prediction (RCP) transformation portion 1234, and a multiple mode residue generation portion 1235, for example. The single mode residue generation portion 1231 and the multiple mode residue generation portion 1235 may perform similar operations, respectively, as the single mode residue generation portion 1211 and the multiple mode residue generation portion 1212 illustrated in FlG. 2. As described above, in order to distinguish the residues generated by the single mode residue generation portion 1231 and the residues generated by the RCT portion 1232, the IPP transformation portion 1233, and the RCP transformation portion 1234, the former can be referred to as 'first residues', and the latter can be referred to as 'second residues'.
[90] Methods of generating the second residues that correspond to the differences between the first residues generated by the single mode residue generation portion 1231 include, for example, RCT, IPP transformation, and RCP transformation. Here, RCT can be used for generating the second residues in the YCoCg color space, and IPP transformation is for generating the second residues in the RGB color space. In RCP transformation, which is similar to IPP transformation, when a residue of the G component is used as a predictor, the same transformation as IPP transformation can be performed after noise is removed by predetermined filtering.
[91] If the prediction mode selected by the optimal mode selection unit 110 is the RCT mode, the RCT portion 1232 may generate second residues that correspond to the differences between the first residues of the Y component, the Co component, and the Cg component in the YCoCg color space using the below Equation 3, for example. In particular, the following relationships may be considered: Y = R + 2G + B » 2, Co = R - B » 1, and Cg = -R + 2G - B » 2.
[92] Equation 3 :
[93] [Math.3]
Δ2B = ΔR - ΔB t = ΔB + ( Δ2B » 1)
Δ2R = ΔG - t
Δ2G = t + ( Δ2R » 1)
[94] Here, Δ X denotes first residues, Δ 2X denotes second residues, and '»' denotes a right shift operation, which is subsequently equivalent to division by 2, and variable t is used for the purpose of temporary calculation.
[95] If the prediction mode selected by the optimal mode selection unit 110 is the IPP transformation mode, the IPP transformation portion 1233 may generate second residues that correspond to the differences between the first residues of the R component, the G component, and the B component in the RGB color space according to the IPP transformation mode in view of the below Equation 4, for example.
[96] Equation 4 :
[97] [Math.4]
Δ2G = ΔG 1
Δ2R = ΔR - ΔG 1
Δ2B = ΔB - ΔG 1
[98] Here, Δ X denotes first residues, Δ 2X denotes second residues, and Δ X' denotes restored first residues. In particular, Equation 4 is effective when the G component includes a large amount of image information. The second residues can be calculated using the R component or the B component as a dominant component.
[99] If the prediction mode selected by the optimal mode selection unit 110 is the RCP transformation mode, after removing noise from each of the R component, the G component, and the B component in the RGB color space according to the RCP transformation mode using a 5-tap filter, expressed in the below Equation 5, for example, the RCP transformation portion 1234 generates second residues that correspond to the differences between the first residues of the R component, the G component, and the B component.
[100] Equation 5 :
[101] [Math.5]
^ = 4 - ^(z,7) + ^(z + l,7) + ^(z,7 + l) + ^G(z -l,7) + ^(z,7 -l), rG (z, j) = F(B^ ) = sign^ s ) - [{ab S^R08 ) + 4) » 3]
[102] In addition, after removing noise from each of the R component, the G component and the B component in the RGB color space using a 3-tap filter, expressed in the below Equation 6, for example, the RCP transformation portion 1234 generates second residues that correspond to the differences between the first residues of the R component, the G component, and the B component.
[103] Equation 6 :
[104] [Math.6]
^ = R^(i - Ij) + 2 ■ Rβ(i,j) + R^(ι + IJ),
[105] In the above Equations 5 and 6, abs(x) denotes the absolute value of x, and sign(x) is given by the below Equation 7, for example. [106] Equation 7 :
[107] [Math.7]
[108] FIGS. 7 A and 7B illustrate a 5-tap filter and a 3-tap filter, respectively, which are used in an embodiment of the present invention.
[109] Referring to FIGS. 7A and 7B, a 5-tap filter used in an embodiment of the present invention uses 4 neighboring pixels around a pixel that is used as a predictor, and a 3-tap filter used in an embodiment of the present invention uses two, left and right pixels on either side of a pixel that is used as a predictor. In addition, weights may be used as in Equations 5 and 6.
[110] FlG. 8 illustrates a residue generation unit 120, such as that illustrated in FlG. 1, according to an embodiment of the present invention.
[Ill] Referring to FlG. 8, the residue generation unit 120 may include a multiple mode residue generation portion 1241 and a residue transformation portion 1242, for example. The multiple mode residue generation portion 1241 may perform similar operations as the multiple mode residue generation portion 1212 illustrated in FlG. 2, for example.
[112] However, even after inter prediction or intra prediction is performed in the multiple mode residue generation portion 1241, redundancy between the color components remains. The residue transformation portion 1242 may eliminate the redundancy between the color components. In order to distinguish the residues generated by the single mode residue generation portion 1241 and the residues generated by the residue transformation portion 1242, the former residue can be referred to as 'first residues', and the latter can be referred to as 'second residues'.
[113] If the prediction mode selected by the optimal mode selection unit 110 is the residue transformation mode, the residue transformation portion 1242 may generate second residues that correspond to the differences between the first residues generated by the multiple mode residue generation portion 1241, according to the residue transformation mode. Here, even after inter prediction or intra prediction is performed in the multiple mode residue generation portion 1241, redundancy between the color components may still remain. However, since the similarity between the color components is lower than in the single prediction mode, the compression efficiency in the residue transformation portion 1242 may be lower than the compression efficiency in the residue transformation portion 1222 illustrated in FlG. 5.
[114] FlG. 9 illustrates a restored image generation unit 180, such as that illustrated in FlG. 1, according to an embodiment of the present invention.
[115] Referring to FlG. 9, the restored image generation unit 180 may include a residue inverse transformation portion 181 and a prediction compensation portion 182, for example.
[116] The residue inverse transformation portion 181 may generate first residues that correspond to the sums of the second residues restored by the frequency domain inverse transformation portion 170. For example, the residue inverse transformation portion 181 may generate the first residues of the Y component, the Co component, and the Cg component, which correspond to the sums of the second residues of the Y component, the Co component, and the Cg component, respectively, using the below Equation 8, for example. H-
[117] Equation 8 +
[118] [Math.8] t = Δ 2G ' - ( Δ: 2R ' » : D
ΔG 1 = = Δ 2R t
ΔB 1 = : t — ( Δ 2B 1 » D
ΔR 1 = : Δ B ' + Δ2B
[119] Here, Δ X' denotes restored first residues, and Δ 2X' denotes restored second residues.
[120] Alternatively, the residue inverse transformation portion 181 may generate the first residues of the R component, the G component, and the B component, which correspond to the sums of the second residues of the R component, the G component, and the B component, respectively, using the below Equation 9, for example.
[121] Equation 9 :
[122] [Math.9]
ΔG ' = Δ2G
ΔR ' = Δ2R ' + ΔG 1
ΔB ' = Δ2B ΔG 1
[123] Here, Δ X' denotes restored first residues, and Δ X' denotes restored second residues.
[124] The prediction compensation portion 182 may generate a restored image that corresponds to the sum of the predicted image generated by the residue generation unit 120 and the first residues generated by the residue inverse transformation portion 181.
For example, the prediction compensation portion 182 may generate a restored image in the YCoCg color space by calculating the sum of the predicted image generated by the residue generation unit 120 and the first residues generated by the residue inverse transformation portion 181. Alternatively, the prediction compensation portion 182 may generate a restored image in the RGB color space by calculating the sum of the predicted image generated by the residue generation unit 120 and the first residues generated by the residue inverse transformation portion 181. Here, it is further noted that alternate embodiments are also available.
[125] FIG. 10 illustrates a moving picture decoding system, according to an embodiment of the present invention.
[126] Referring to FIG. 10, the moving picture decoding system may include an entropy decoding unit 210, an inverse quantization unit 220, a frequency domain inverse transformation unit 230, a predicted image generation unit 240, and a restored image generation unit 250, for example.
[127] The entropy decoding unit 210 may restore integer values corresponding to the current image and information representing the optimal prediction mode for the blocks of each color component of the current image by performing entropy-decoding on a bitstream, such as a bitstream output from the moving picture encoding system illustrated in FIG. 1. Here, for example, the optimal prediction mode for the blocks of each color component of the current image information can be the prediction mode used in the moving picture encoding system.
[128] The inverse quantization unit 220 may restore the frequency component values by performing inverse quantization on the integer values restored by the entropy decoding unit 210. In other words, the inverse quantization unit 220 restores the frequency component values by multiplying the integer values restored by the entropy decoding unit 210 by a quantization parameter.
[129] The frequency domain inverse transformation unit 230 may generate residues that correspond to differences between the current image and the predicted image by transforming the frequency component values restored by the inverse quantization unit 220, which are frequency domain values, into values in the color space. Alternatively, in an embodiment, the frequency domain inverse transformation unit 230 restores second residues that correspond to the differences between the first residues of the color components by transforming the frequency component values restored by the inverse quantization unit 220, which are frequency domain values, into values in the color space.
[130] The predicted image generation unit 240 may generate a predicted image for the current image according to the prediction mode indicated by the information restored for the macro blocks, which correspond to each other, of the color components by the
entropy decoding unit 210.
[131] The restored image generation unit 250 may generate a restored image that corresponds to the sum of the predicted image generated by the predicted image generation unit 240 and the residues restored by the frequency domain inverse transformation unit 230. Alternatively, in an embodiment, the restored image generation unit 250 may generate first residues that correspond to the sums of the second residues restored by the frequency domain inverse transformation unit 230 and generate a restored image that corresponds to the sum of the predicted image generated by the predicted image generation unit 240 and the generated first residues.
[132] FlG. 11 illustrates a predicted image generation unit 240, such as that illustrated in FlG. 10, according to an embodiment of the present invention.
[133] Referring to FlG. 11, the predicted image generation unit 240 may include a single mode prediction image generation portion 241 and a multiple mode prediction image generation portion 242, for example.
[134] If the information restored by the entropy decoding unit 210 represents that a single prediction mode was applied in the encoding, the single mode prediction image generation portion 241 may generate a predicted image by performing single intra prediction or single inter prediction, for example. In particular, the single mode prediction image generation portion 241 may generate a predicted image for the current image from a reference image using motion vectors between the reference image and the current image, which are commonly applied to the blocks, which correspond to each other, of the color components. In other words, the single mode prediction image generation portion 241 may generate the predicted image for the current image from the reference image using the motion vectors that are commonly applied to the blocks, which are divided from the macro blocks, which correspond to each other, of the color components of the current image and correspond to each other.
[135] In addition, the single mode prediction image generation portion 241 may further predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 250 using prediction directions that are commonly applied to the blocks, which correspond to each other, of the color components, which are divided equally from the macro blocks, which correspond to each other, of the color components of the current image, and generate a predicted image constituted by the predicted blocks.
[136] If the information restored by the entropy decoding unit 210 represents that a multiple prediction mode was applied in the encoding, the multiple mode prediction image generation portion 242 may generate a predicted image by performing multiple inter prediction or multiple intra prediction. In particular, the multiple mode prediction image generation portion 242 may generate a predicted image for the current image
from a reference image using motion vectors between the reference image and the current image, which are independently applied to the blocks of each color component, which are divided into different sizes from the macro blocks, which correspond to each other, of the color components of the current image. In other words, the multiple mode prediction image generation portion 242 may generate the predicted image for the current image from the reference image using different motion vectors for the blocks of each color component, which are divided into blocks of different sizes from the macro blocks, which correspond to each other, of the color components of the current image.
[137] In addition, the multiple mode prediction image generation portion 242 may predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 250 using prediction directions that are commonly applied to the blocks, which correspond to each other, of the color components, wherein the blocks are divided to different sizes from the macro blocks, which correspond to each other, of the color components of the current image, and generate a predicted image made up by the predicted blocks, thereby performing multiple intra prediction. In other words, the multiple mode prediction image generation portion 242 may predict blocks making up the current image from adjacent pixels in the restored image generated by the restored image generation unit 250 using different prediction directions for the blocks of each color component, which are divided to different sizes from the macro blocks, which correspond to each other, of the color components of the current image and generate the predicted image made up by the predicted blocks.
[138] FlG. 12 illustrates a restored image generation unit 250, such as that shown in FlG. 10, according to an embodiment of the present invention.
[139] Referring to FlG. 12, the restored image generation unit 250 may include a residue inverse transformation portion 251 and a prediction compensation portion 252, for example.
[140] If the information restored by the entropy decoding unit 210 represents the residue transformation mode was applied in the encoding, the residue inverse transformation portion 251 may generate first residues that correspond to the sums of the second residues restored by the frequency domain inverse transformation unit 230. For example, the residue inverse transformation portion 251 may generate the first residues of the Y component, the Co component, and the Cg component, which correspond to the sums of the second residues of each respective color component, i.e., the Y component, the Co component, and the Cg component, using the above Equation 8, for example. Alternatively, for example, the residue inverse transformation portion 251 may generate the first residues of the R component, the G component, and the B
component, which correspond to the sums of the second residues of each respective color component, i.e., the R component, the G component, and the B component, using the above Equation 9, for example.
[141] The prediction compensation portion 252 may generate a restored image that corresponds to the sum of the predicted image generated by the predicted image generation unit 240 and the first residues generated by the residue inverse transformation portion 251. For example, the prediction compensation portion 252 may generate a restored image in the YCoCg color space by calculating the sums of the predicted image of each respective color component, i.e., the Y component, the Co component, and the Cg component, and the first residues generated by the residue inverse transformation portion 251. Alternatively, for example, the prediction compensation portion 252 may generate a restored image in the RGB color space by calculating the sums of the predicted image for each respective color component, i.e., the R component, the G component, and the B component, which is generated by the predicted image generation unit 240, and the first residues generated by the residue inverse transformation portion 251.
[142] FIG. 13 illustrates a residue inverse transformation portion 251, such as that shown in FIG. 12, according to an embodiment of the present invention.
[143] Referring to FIG. 13, the residue inverse transformation portion 251 may include an RCT portion 2511, an IPP transformation portion 2512, and an RCP transformation portion 2513, for example.
[144] If the information restored by the entropy decoding unit 210 represents that the RCT transformation mode was applied during the encoding, the RCT portion 2511 may generate the first residues of each of the Y component, the Co component, and the Cg component, which respectively correspond to the sums of the second residues of each of the Y component, Co component, and the Cg component, using the above Equation 8, for example.
[145] If the information restored by the entropy decoding unit 210 represents that the IPP transformation mode was applied during the encoding, the IPP transformation portion
2512 may generate the first residues of each of the R component, the G component, and the B component, which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, using the above Equation 9, for example.
[146] If the information restored by the entropy decoding unit 210 represents that the RCP transformation mode was applied during the encoding, the RCP transformation portion
2513 may generate the first residues of each of the R component, the G component, and the B component, which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, using
the above Equation 9, for example.
[147] FIGS. 14A and 14B illustrate a moving picture encoding method, according to an embodiment of the present invention.
[148] Referring to FIGS. 14A and 14B, a moving picture encoding method may include operations which may be time-serially performed in the moving picture encoding system shown in FIG. 1, for example. Therefore, for some embodiments of the present invention, the above-descriptions of the moving picture encoding system illustrated in FIG. 1 can also apply to the moving picture encoding method, even if some descriptions thereof are not directly provided below.
[149] In operation 1401, one of the possible prediction modes for a sample image may be selected. As illustrated in FIG. 14 A, especially, when operation 1409 returns to operation 1401, one from among possible prediction modes except for a previously selected prediction mode may be selected.
[150] In operation 1402, a predicted image may be generated for the current image for each macro block of each color component, according to the prediction mode selected in operation 1401, and residues that correspond to the differences between the current image and the predicted image may be generated. Alternatively, in operation 1402, a predicted image for the current image for each macro block of each color component may be generated according to the prediction mode selected in operation 1401, and second residues that correspond to the differences between the first residues of the color components may be generated.
[151] In operation 1403, the residues generated in operation 1402 may be transformed, with the residues being color space residues, into values in the frequency domain. Alternatively, in operation 1403, the second residues generated in operation 1402 may be transformed, with the residues being color space residues, into values in the frequency domain.
[152] In operation 1404, the values transformed in operation 1403 may be quantized.
[153] In operation 1405, a bitstream may be generated by entropy-encoding the values quantized in operation 1404 and information representing the prediction mode selected in operation 1401.
[154] In operation 1406, frequency component values may further be restored by inverse- quantizing the values quantized in operation 1404.
[155] In operation 1407, residues that correspond to differences between the current image and the predicted image may be restored by transforming the frequency component values restored in operation 1406, which are frequency domain values, into values in the color space. Alternatively, in an embodiment, in operation 1407, second residues that correspond to differences between the first residues of each color component may be restored by transforming the frequency component values restored in operation
1406, which are frequency domain values, into values in the color space.
[156] In operation 1408, a restored image that corresponds to the sum of the predicted image generated in operation 1402 and the residues restored in operation 1407 may be generated. Alternatively, in an embodiment, in operation 1408, first residues that correspond to the sums of the second residues restored in operation 1407 may be generated, and a restored image that corresponds to the sum of the predicted image generated in operation 1402 and the generated first residues may be generated.
[157] In operation 1409, in one embodiment, whether operations 1401 through 1408 have been performed for all the prediction modes may be determined. In this embodiment, if operations 1401 through 1408 have been performed on all the prediction modes, the method may proceed to operation 1410 illustrated in FlG. 14B. Otherwise, the method may returns to operation 1401.
[158] In operation 1410, an optimal prediction mode may be selected for the macro blocks, which correspond to each other, of the color components of the current image based on the characteristics of the sample images, i.e., the results of operations 1401 through 1409. For example, in operation 1410, the moving picture encoding system may select a prediction mode that leads to a minimum size of a resultant bitstream and/or a minimum amount of quality distortion between the sample image and the restored image generated in operation 1408, as an optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image.
[159] In operation 1411, a predicted image for the current image for each macro block of each color component according to the prediction mode selected in operation 1410 may be generated, and residues that correspond to the differences between the current image and the predicted image may be generated. Alternatively, in an embodiment, in operation 1411, a predicted image for the current image for each macro block of each color component may be generated according to the prediction mode selected in operation 1410, first residues that correspond to the differences between the current image and the predicted image may be generated, and second residues that correspond to the differences between the first residues of each color component may be generated.
[160] In operation 1412, the residues generated in operation 1411 may be transformed, with the residues being color space residues, into values in the frequency domain. Alternatively, in operation 1412, the second residues generated in operation 1411 may be transformed, with the residues being color space residues, into values in the frequency domain.
[161] In operation 1413, the values transformed in operation 1412 may be quantized.
[162] In operation 1414, a bitstream may be generated by entropy-encoding the values quantized in operation 1413 and information representing the prediction mode selected
in operation 1410.
[163] In operation 1415, the frequency component values may be restored by inverse- quantizing the values quantized in operation 1413.
[164] In operation 1416, residues that correspond to the differences between the current image and the predicted image may be restored by transforming the frequency component values restored in operation 1415, which are frequency domain values, into values in the color space. Alternatively, in an embodiment, in operation 1416, second residues that correspond to differences between the first residues of each color component may be restored by transforming the frequency component values restored in operation 1415, which are frequency domain values, into values in the color space.
[165] In operation 1417, a restored image that corresponds to the sum of the predicted image generated in operation 1411 and the residues restored in operation 1416 may be generated. Alternatively, in an embodiment, in operation 1417, first residues that correspond to the sums of the second residues of each color component restored in operation 1416 may be generated, and a restored image that corresponds to the sum of the predicted image generated in operation 1411 and the generated first residues may be generated.
[166] FlG. 15 illustrates a moving picture decoding method, according to an embodiment of the present invention.
[167] Referring to FlG. 15, the illustrated operations may be time-serially performed, e.g., such as in the moving picture decoding system shown in FlG. 10. Accordingly, for such an embodiment, the above-descriptions of the moving picture decoding system illustrated in FlG. 10 should also be applied to such a moving picture decoding method, even if some descriptions thereof are not directly provided below.
[168] In operation 1501, integer values corresponding to the current image and information representing the optimal prediction mode for the macro blocks, which correspond to each other, of the color components of the current image may be restored by entropy- decoding a bitstream, such as that output from the moving picture encoding system illustrated in FlG. 1.
[169] In operation 1502, the frequency component values may be restored by inverse- quantizing the integer values restored in operation 1501.
[170] In operation 1503, residues that correspond to the differences between the current image and the predicted image may be generated by transforming the frequency component values restored in operation 1502, which are frequency domain values, into values in the color space. Alternatively, in an embodiment, in operation 1503, second residues that correspond to the differences between the first residues of each color component may be restored by transforming the frequency component values restored in operation 1502, which are frequency domain values, into values in the color space.
[171] In operation 1504, if the information restored in operation 1501 represents that the residue transformation mode was applied during encoding, the below operation 1505 may be performed. Otherwise, the below operation 1506 may be performed.
[172] In operation 1505, first residues that correspond to the sums of the second residues of each color component restored in operation 1503 may be generated. In particular, if the information restored in operation 1501 represents that the RCT transformation mode was applied during encoding, in operation 1505, the first residues of each of the Y component, the Co component, and the Cg component, which respectively correspond to the sums of the second residues of each of the Y component, the Co component, and the Cg component, may be generated using the above Equation 8, for example. Alternatively, if the information restored in operation 1501 represents that the IPP or RCP transformation modes were applied in encoding, in operation 1505, the first residues of each of the R component, the G component, and the B component, which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, may be generated using the above Equation 9, for example.
[173] In an embodiment, if the information restored by the entropy decoding unit 210 represents that the RCP transformation mode was applied in encoding, the RCP transformation portion 2511 may generate the first residues of each of the R component, the G component, and the B component, which respectively correspond to the sums of the second residues of each of the R component, the G component, and the B component, using the above Equation 9, for example.
[174] In operation 1506, if the information restored in operation 1501 is determined to represent the single prediction mode, operation 1507 may be performed. If the information restored in operation 1501 is determined to represent the multiple prediction mode, operation 1508 may be performed.
[175] In operation 1507, a predicted image may be generated according to the single prediction mode, i.e., using motion vectors that are commonly applied to all the divided blocks of the color components of the current image, which are divided equally, or using prediction directions that are independently applied to all the divided blocks of the color components.
[176] In operation 1508, a predicted image may be generated according to the multiple prediction mode, i.e., using motion vectors that are independently applied to the divided blocks of each color component of the current image, which are divided into blocks of different sizes, or using prediction directions that are independently applied to the divided blocks of each color component.
[177] In operation 1509, a restored image that corresponds to the sum of the predicted image generated in operation 1507 or 1508 and the residues restored in operation 1503
may be generated. Alternatively, a restored image that corresponds to the sum of the predicted image generated in operation 1507 or 1508 and the first residues generated in operation 1505 may be generated.
[178] In addition to the above described embodiments, embodiments of the present invention can also be implemented through computer readable code/instructions in/on a medium, e.g., a computer readable medium, to control at least one processing element to implement any above described embodiment. The medium can correspond to any medium/media permitting the storing and/or transmission of the computer readable code.
[179] The computer readable code can be recorded/transferred on a medium in a variety of ways, with examples of the medium including recording media, such as magnetic storage media (e.g., ROM, floppy disks, hard disks, etc.) and optical recording media (e.g., CD-ROMs, or DVDs), and transmission media such as carrier waves, as well as through the Internet, for example. Thus, the medium may further be a signal, such as a resultant signal or bitstream, according to embodiments of the present invention. The media may also be a distributed network, so that the computer readable code is stored/ transferred and executed in a distributed fashion. Still further, as only an example, the processing element could include a processor or a computer processor, and processing elements may be distributed and/or included in a single device.
[180] Although a few embodiments of the present invention have been shown and described, it would be appreciated by those skilled in the art that changes may be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.
Claims
[ 1 ] L A predicted image generating method, comprising : selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image; and generating a predicted image for the current image according to the selected prediction mode.
[2] 2. The method of claim 1, wherein the selecting of the prediction mode comprises selecting a prediction mode that is commonly applied to the macro blocks, which correspond to each other, of the color components when a similarity between the color components of the predetermined image is determined to be high.
[3] 3. The method of claim 1, wherein the selecting of the prediction mode comprises selecting a prediction mode that is independently applied to the macro blocks, which correspond to each other, of the color components when a similarity between the color components of the predetermined image is determined to be low.
[4] 4. The method of claim 1, wherein the generating of the predicted image comprises: dividing the macro blocks, which correspond to each other, of the color components into equal sized blocks; determining motion vectors between a reference image and the current image, which are to be commonly applied to the divided blocks, which correspond to each other, of the color components; and generating the predicted image for the current image from the reference image using the determined motion vectors.
[5] 5. The method of claim 1, wherein the generating of the predicted image comprises: dividing the macro blocks, which correspond to each other, of the color components into equal sized blocks; determining a prediction direction to be commonly applied to the divided blocks, which correspond to each other, of the color components; and predicting blocks making up the current image from adjacent pixels in a restored image using the determined prediction directions and generating the predicted image made up by the predicted blocks.
[6] 6. The method of claim 1, wherein the generating of the predicted image comprises:
dividing the macro blocks, which correspond to each other, of the color components into blocks of different sizes; determining motion vectors between a reference image and the current image, which are to be independently applied to the divided blocks of each color component; and generating the predicted image for the current image from the reference image using the determined motion vectors.
[7] 7. The method of claim 1, where the generating of the predicted image comprises: dividing the macro blocks, which correspond to each other, of the color components into blocks of different sizes; determining a prediction direction to be independently applied to the divided blocks of each color component; and predicting blocks making up the current block from adjacent pixels in a restored image using the determined prediction direction and generating the predicted image made up by the predicted blocks.
[8] 8. The method of claim 1, wherein the selecting of the prediction mode comprises selecting a prediction mode that leads to a minimum bitstream size as a result of encoding the predetermined image.
[9] 9. The method of claim 1, wherein the selecting of the prediction mode comprises selecting a prediction mode that leads to a minimum amount of quality distortion between the predetermined image and a restored image of the predetermined image.
[10] 10. At least one medium comprising computer readable code to control at least one processing element to implement a predicted image generating method, the method comprising: selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image; and generating a predicted image for the current image according to the selected prediction mode.
[11] 11. A predicted image generating system, comprising : a selection unit selecting a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image; and a generation unit generating a predicted image for the current image according to the selected prediction mode.
[12] 12. An encoding method, comprising:
selecting a prediction mode that is determined to be optimal for blocks of each- color component of a current image based on characteristics of a predetermined image; generating a predicted image for the current image according to the selected prediction mode; generating residues, of each color component, which correspond to differences between the current image and the predicted image; and generating a bitstream by encoding the generated residues.
[13] 13. The method of claim 12, wherein the generating of the bitstream comprises generating the bitstream to include information representing the selected prediction mode.
[14] 14. The method of claim 12, wherein the generating of the bitstream comprises generating the bitstream to include information representing the selected prediction mode for the macro blocks, which correspond to each other, of the color components.
[15] 15. The method of claim 12, wherein the generating of the bitstream comprises generating the bitstream to include information representing only one selected prediction mode that is commonly applied at a sequence level, as an upper level relative to a macro block level, to all the macro blocks, which correspond to each other, of the color components making up a single sequence.
[16] 16. The method of claim 12, wherein the generating of the bitstream comprises generating the bitstream to include information representing only one selected prediction mode that is commonly applied at a picture level, as an upper level relative to a macro block level, to all the macro blocks, which correspond to each other, of the color components making up a single picture.
[17] 17. The method of claim 12, wherein, in the generating of the residues, the residues are first residues, and the method further comprises generating second residues that correspond to differences between the first residues of the color components, according to the selected prediction mode, and wherein the generating of the bitstream comprises generating the bitstream by selectively encoding the generated first residues or the generated second residues.
[18] 18. The method of claim 17, wherein the generating of the second residues comprises generating the second residues as corresponding to differences between the first residues of each of a Y component, a Co component, and a Cg component, according to the selected prediction mode.
[19] 19. The method of claim 17, wherein the generating of the second residues comprises generating the second residues as corresponding to differences between the first residues of each of an R component, a G component, and a B
component, according to the selected prediction mode.
[20] 20. The method of claim 17, wherein the generating of the second residues comprises removing noise from each of an R component, a G component, and a B component using a predetermined filter and generating the second residues as corresponding to differences between the first residues of each of the R component, the G component, and the B component.
[21] 21. At least one medium comprising computer readable code to control at least one processing element to implement the method of claim 12.
[22] 22. An encoding system, comprising: a selection unit to select a prediction mode that is determined to be optimal for blocks of each color component of a current image based on characteristics of a predetermined image; a generation unit to generate a predicted image for the current image according to the selected prediction mode and to generate residues, of each color component, which correspond to differences between the current image and the predicted image; and an encoding unit to generate a bitstream by encoding the generated residues.
[23] 23. A predicted image generating method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method comprising: restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode; and generating a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
[24] 24. The method of claim 23, wherein the prediction mode is a prediction mode used in a moving picture decoding system to generate the predicted image.
[25] 25. The method of claim 23, wherein, if the restored prediction mode information identifies the prediction mode as being a prediction mode commonly applied to all the blocks of the color components, the generating of the predicted image comprises generating the predicted image for each block, which is divided into blocks of an equal size that is commonly applied to the blocks, which correspond to each other, of the color components.
[26] 26. The method of claim 23, wherein, if the restored prediction mode information identifies the prediction mode as being a prediction mode commonly applied to the blocks, which correspond to each other, of the color components, the generating of the predicted image comprises generating the predicted image using a motion vector that is commonly applied to the blocks, which correspond
to each other, of the color components.
[27] 27. The method of claim 23, wherein, if the restored prediction mode information identifies the prediction mode as being a prediction mode commonly applied to the blocks, which correspond to each other, of the color components, the generating of the predicted image comprises generating the predicted image using a prediction direction that is commonly applied to the blocks, which correspond to each other, of the color components.
[28] 28. The method of claim 23, wherein, if the restored prediction mode information identifies the prediction mode as being a prediction mode independently applied to the blocks of each color component, the generating of the predicted image comprises generating the predicted image for each block, which is divided into blocks of different sizes that are independently applied to each color component.
[29] 29. The method of claim 23, wherein, if the restored prediction mode information identifies the prediction mode as being a prediction mode independently applied to the blocks of each color component, the generating of the predicted image comprises generating the predicted image using motion vectors that are independently applied to the blocks of each color component.
[30] 30. The method of claim 23, wherein, if the restored prediction mode information identifies the prediction mode as being a prediction mode independently applied to the blocks of each color component, the generating of the predicted image comprises generating the predicted image using prediction directions that are independently applied to the blocks of each color component.
[31] 31. At least one medium comprising computer readable code to control at least one processing element to implement a predicted image generating method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method comprising: restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode; and generating a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
[32] 32. A predicted image generating system to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the system comprising: a decoding unit to restore prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode; and
a generation unit to generate a predicted image for the current image according to the prediction mode identified by the restored prediction mode information.
[33] 33. A decoding method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method comprising: restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode; generating residues that correspond to differences between the current image and a reference image for the current image, according to the prediction mode identified by the restored prediction mode information; generating a predicted image according to the prediction mode identified by the restored prediction mode information; and generating a restored image that corresponds to a sum of the generated residues and the generated predicted image.
[34] 34. The method of claim 33, wherein, in the generating of the residues, the residues are first residues, and the method further comprises restoring second residues that correspond to differences between the first residues of color components, and wherein the generating of the residues comprises generating the first residues as corresponding to sums of the restored second residues of each color component, according to the prediction mode identified by the restored prediction mode information.
[35] 35. The method of claim 34, wherein the restoring of the prediction mode information comprises: restoring integer values that correspond to the current image by entropy- decoding the bitstream and corresponding prediction mode information; and restoring frequency component values by inverse-quantizing the restored integer values, and the restoring of the second residues comprises restoring the second residues by transforming the restored frequency component values, which are frequency domain values, into values in the color space.
[36] 36. The method of claim 34, wherein the generating of the residues comprises generating first residues of each of a Y component, a Co component, and a Cg component according to the prediction mode identified by the restored prediction mode information, the first residues respectively corresponding to sums of restored second residues of each of the Y component, the Co component, and the Cg component.
[37] 37. The method of claim 34, wherein the generating of the residues comprises generating first residues of each of an R component, a G component, and a B component according to the prediction mode identified by the restored prediction mode information, the first residues respectively corresponding to sums of restored second residues of each of the R component, the G component, and the B component.
[38] 38. At least one medium comprising computer readable code to control at least one processing element to implement a decoding method to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the method comprising: restoring prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode; generating residues that correspond to differences between the current image and a reference image for the current image, according to the prediction mode identified by the restored prediction mode information; generating a predicted image according to the prediction mode identified by the restored prediction mode information; and generating a restored image that corresponds to a sum of the generated residues and the generated predicted image.
[39] 39. A decoding system to decode image information encoded in a bitstream, with the image information being encoded according to a prediction mode determined to be optimal for corresponding blocks of each color component of a corresponding current image, the system comprising: a decoding unit to restore prediction mode information from the bitstream, the restored prediction mode information identifying the prediction mode; a first generation unit to generate a predicted image for the current image according to the prediction mode identified by the restored prediction mode information; and a second generation unit to generate residues that correspond to differences between the current image and the predicted image according to the prediction mode identified by the restored prediction mode information and to generate a restored image corresponding to a sum of the generated residues and the generated predicted image.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07715613.1A EP1994763B1 (en) | 2006-03-13 | 2007-03-13 | Method, medium, and system encoding and/or decoding moving pictures by adaptively applying optimal prediction modes |
CN2007800091920A CN101401437B (en) | 2006-03-13 | 2007-03-13 | Method and system for encoding and/or decoding moving pictures by adaptively applying optimal prediction modes |
JP2009500288A JP2009529845A (en) | 2006-03-13 | 2007-03-13 | Method and apparatus for encoding moving picture by adaptively applying optimal prediction mode, and method and apparatus for decoding moving picture |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US78137906P | 2006-03-13 | 2006-03-13 | |
US60/781,379 | 2006-03-13 | ||
KR1020060049080A KR101330630B1 (en) | 2006-03-13 | 2006-05-30 | Method and apparatus for encoding moving picture, method and apparatus for decoding moving picture, applying adaptively an optimal prediction mode |
KR10-2006-0049080 | 2006-05-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2007105900A1 true WO2007105900A1 (en) | 2007-09-20 |
Family
ID=38687637
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2007/001217 WO2007105900A1 (en) | 2006-03-13 | 2007-03-13 | Method, medium, and system encoding and/or decoding moving pictures by adaptively applying optimal prediction modes |
Country Status (6)
Country | Link |
---|---|
US (5) | US10034000B2 (en) |
EP (2) | EP2538679A3 (en) |
JP (4) | JP2009529845A (en) |
KR (2) | KR101330630B1 (en) |
CN (1) | CN101401437B (en) |
WO (1) | WO2007105900A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102282852A (en) * | 2008-10-01 | 2011-12-14 | 韩国电子通信研究院 | Image encoder and decoder using unidirectional prediction |
US8867854B2 (en) | 2008-10-01 | 2014-10-21 | Electronics And Telecommunications Research Institute | Image encoder and decoder using undirectional prediction |
Families Citing this family (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101365575B1 (en) * | 2007-02-05 | 2014-02-25 | 삼성전자주식회사 | Method and apparatus for encoding and decoding based on inter prediction |
EP3190791B1 (en) | 2007-06-29 | 2018-11-28 | Velos Media International Limited | Image encoding device and image decoding device |
KR101407719B1 (en) * | 2008-01-14 | 2014-06-16 | 광주과학기술원 | Multi-view image coding method and apparatus using variable GOP prediction structure, multi-view image decoding apparatus and recording medium storing program for performing the method thereof |
ATE524927T1 (en) * | 2008-01-21 | 2011-09-15 | Ericsson Telefon Ab L M | IMAGE PROCESSING BASED ON PREDICTION |
KR101291196B1 (en) * | 2008-01-25 | 2013-07-31 | 삼성전자주식회사 | Video encoding method and apparatus, and video decoding method and apparatus |
KR101590511B1 (en) | 2009-01-23 | 2016-02-02 | 에스케이텔레콤 주식회사 | / / Motion Vector Coding Method and Apparatus |
JP4598162B2 (en) * | 2009-02-05 | 2010-12-15 | パナソニック株式会社 | Imaging processing device |
JP2010258739A (en) * | 2009-04-24 | 2010-11-11 | Sony Corp | Image processing apparatus, method and program |
KR101633459B1 (en) * | 2009-08-10 | 2016-06-24 | 삼성전자주식회사 | Apparatus and method for encoding and decoding image data using correlation between colors |
JP5421757B2 (en) * | 2009-12-11 | 2014-02-19 | 株式会社Kddi研究所 | Image encoding device |
EP3913923A1 (en) | 2010-01-19 | 2021-11-24 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding/decoding images using a motion vector of a previous block as a motion vector for the current block |
KR101607308B1 (en) * | 2010-01-19 | 2016-03-29 | 삼성전자주식회사 | Method and apparatus for encoding/decoding image by using motion vector of previous block as motion vector of current block |
CN105025299B (en) * | 2010-01-19 | 2019-04-16 | 三星电子株式会社 | The method and apparatus that image is encoded/decoded |
AU2015200748B2 (en) * | 2010-04-05 | 2016-01-07 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding video by compensating for pixel value according to pixel groups, and method and apparatus for decoding video by the same |
KR101529992B1 (en) | 2010-04-05 | 2015-06-18 | 삼성전자주식회사 | Method and apparatus for video encoding for compensating pixel value of pixel group, method and apparatus for video decoding for the same |
KR101503269B1 (en) | 2010-04-05 | 2015-03-17 | 삼성전자주식회사 | Method and apparatus for determining intra prediction mode of image coding unit, and method and apparatus for determining intra predion mode of image decoding unit |
CN103141103B (en) | 2010-04-09 | 2016-02-03 | Lg电子株式会社 | The method and apparatus of processing video data |
JP5346884B2 (en) * | 2010-06-23 | 2013-11-20 | 日本放送協会 | Motion compensation device |
US20120218292A1 (en) * | 2011-02-10 | 2012-08-30 | Ncomputing Inc. | System and method for multistage optimized jpeg output |
CN103430548B (en) * | 2011-03-09 | 2017-10-24 | 松下电器(美国)知识产权公司 | Dynamic image encoding device and dynamic image encoding method |
WO2013077660A1 (en) * | 2011-11-24 | 2013-05-30 | 에스케이텔레콤 주식회사 | Method and apparatus for effective encoding/decoding usnig detailed predictive unit |
KR102072124B1 (en) | 2011-11-24 | 2020-02-04 | 에스케이텔레콤 주식회사 | Method and Apparatus for Image Encoding/Decoding using detailed prediction unit |
US10200710B2 (en) * | 2012-07-02 | 2019-02-05 | Samsung Electronics Co., Ltd. | Motion vector prediction method and apparatus for encoding or decoding video |
US9225991B2 (en) * | 2013-05-30 | 2015-12-29 | Apple Inc. | Adaptive color space transform coding |
US9225988B2 (en) | 2013-05-30 | 2015-12-29 | Apple Inc. | Adaptive color space transform coding |
CN103391440A (en) * | 2013-07-19 | 2013-11-13 | 华为技术有限公司 | Binarization encoding processing method and device of syntactic information |
US10271052B2 (en) | 2014-03-14 | 2019-04-23 | Qualcomm Incorporated | Universal color-space inverse transform coding |
JP2016005210A (en) * | 2014-06-19 | 2016-01-12 | 三菱電機株式会社 | Terminal and data management device |
JP2014222936A (en) * | 2014-07-23 | 2014-11-27 | 株式会社Kddi研究所 | Image decoding device, image decoding method, image encoding/decoding method and image decoding program |
US9609362B2 (en) * | 2014-10-06 | 2017-03-28 | Telefonaktiebolaget Lm Ericsson (Publ) | Coding and deriving quantization parameters |
EP4432669A2 (en) | 2016-10-04 | 2024-09-18 | B1 Institute of Image Technology, Inc. | Image data encoding/decoding method and apparatus |
US12022199B2 (en) | 2016-10-06 | 2024-06-25 | B1 Institute Of Image Technology, Inc. | Image data encoding/decoding method and apparatus |
CN113785566B (en) * | 2019-04-27 | 2024-09-20 | 数码士有限公司 | Method and apparatus for processing video signal based on intra prediction |
GB2624122B (en) * | 2019-07-05 | 2024-07-24 | V Nova Int Ltd | Quantization of residuals in video coding |
CN111050166B (en) * | 2019-12-02 | 2023-08-15 | 咪咕视讯科技有限公司 | Prediction mode determination method, apparatus, and computer-readable storage medium |
CN117834863A (en) * | 2020-04-14 | 2024-04-05 | Lg电子株式会社 | Point cloud data transmitting device and method, and point cloud data receiving device and method |
US20230334618A1 (en) * | 2022-04-15 | 2023-10-19 | Meta Platforms Technologies, Llc | Block-Based Random Access Capable Lossless Graphics Asset Compression |
US11882295B2 (en) | 2022-04-15 | 2024-01-23 | Meta Platforms Technologies, Llc | Low-power high throughput hardware decoder with random block access |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05137130A (en) * | 1991-11-08 | 1993-06-01 | Victor Co Of Japan Ltd | Encoder and decoder for moving image |
JPH05308631A (en) * | 1992-03-03 | 1993-11-19 | Toshiba Corp | Moving image encoder |
Family Cites Families (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5038216A (en) * | 1989-04-20 | 1991-08-06 | Eastman Kodak Company | Automatic brightness algorithm in a slide to video transfer unit |
US5418570A (en) | 1992-03-03 | 1995-05-23 | Kabushiki Kaisha Toshiba | Motion picture coding apparatus |
US5667735A (en) * | 1994-05-23 | 1997-09-16 | 2C Optics, Inc. | Opthalmic mold coatings |
US5821986A (en) * | 1994-11-03 | 1998-10-13 | Picturetel Corporation | Method and apparatus for visual communications in a scalable network environment |
US6909749B2 (en) * | 2002-07-15 | 2005-06-21 | Pts Corporation | Hierarchical segment-based motion vector encoding and decoding |
US7266247B2 (en) | 2002-09-30 | 2007-09-04 | Samsung Electronics Co., Ltd. | Image coding method and apparatus using spatial predictive coding of chrominance and image decoding method and apparatus |
JP3940657B2 (en) | 2002-09-30 | 2007-07-04 | 株式会社東芝 | Moving picture encoding method and apparatus and moving picture decoding method and apparatus |
US7227901B2 (en) * | 2002-11-21 | 2007-06-05 | Ub Video Inc. | Low-complexity deblocking filter |
KR100750110B1 (en) * | 2003-04-22 | 2007-08-17 | 삼성전자주식회사 | 4x4 intra luma prediction mode determining method and apparatus |
US7469069B2 (en) * | 2003-05-16 | 2008-12-23 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding/decoding image using image residue prediction |
US7535961B2 (en) * | 2003-07-16 | 2009-05-19 | Samsung Electronics Co., Ltd. | Video encoding/decoding apparatus and method for color image |
JP4815107B2 (en) * | 2003-07-16 | 2011-11-16 | 三星電子株式会社 | Lossless video encoding / decoding method and apparatus using color plane prediction |
US7426308B2 (en) * | 2003-07-18 | 2008-09-16 | Microsoft Corporation | Intraframe and interframe interlace coding and decoding |
US7724827B2 (en) * | 2003-09-07 | 2010-05-25 | Microsoft Corporation | Multi-layer run level encoding and decoding |
US7317839B2 (en) * | 2003-09-07 | 2008-01-08 | Microsoft Corporation | Chroma motion vector derivation for interlaced forward-predicted fields |
JP4213646B2 (en) * | 2003-12-26 | 2009-01-21 | 株式会社エヌ・ティ・ティ・ドコモ | Image encoding device, image encoding method, image encoding program, image decoding device, image decoding method, and image decoding program. |
JP2005212601A (en) | 2004-01-29 | 2005-08-11 | Nissan Motor Co Ltd | Vehicle body floor structure |
US20050276493A1 (en) * | 2004-06-01 | 2005-12-15 | Jun Xin | Selecting macroblock coding modes for video encoding |
KR100657268B1 (en) * | 2004-07-15 | 2006-12-14 | 학교법인 대양학원 | Scalable encoding and decoding method of color video, and apparatus thereof |
CA2573990A1 (en) * | 2004-07-15 | 2006-02-23 | Qualcomm Incorporated | H.264 spatial error concealment based on the intra-prediction direction |
JP2006140758A (en) * | 2004-11-12 | 2006-06-01 | Toshiba Corp | Method, apparatus and program for encoding moving image |
US20060112653A1 (en) * | 2004-11-29 | 2006-06-01 | Swcs Marketing Group Inc. | Drainage apparatus and methods for installing |
US7672378B2 (en) * | 2005-01-21 | 2010-03-02 | Stmicroelectronics, Inc. | Spatio-temporal graph-segmentation encoding for multiple video streams |
KR100723403B1 (en) | 2005-02-28 | 2007-05-30 | 삼성전자주식회사 | A prediction image generating method and apparatus using using single coding mode among color components, and an image and video encoding/decoding method and apparatus using it |
KR101246915B1 (en) * | 2005-04-18 | 2013-03-25 | 삼성전자주식회사 | Method and apparatus for encoding or decoding moving picture |
EP1753242A2 (en) | 2005-07-18 | 2007-02-14 | Matsushita Electric Industrial Co., Ltd. | Switchable mode and prediction information coding |
CA2610276C (en) | 2005-07-22 | 2013-01-29 | Mitsubishi Electric Corporation | Image encoder and image decoder, image encoding method and image decoding method, image encoding program and image decoding program, and computer readable recording medium recorded with image encoding program and computer readable recording medium recorded with image decoding program |
US8488889B2 (en) * | 2005-07-22 | 2013-07-16 | Mitsubishi Electric Corporation | Image encoder and image decoder, image encoding method and image decoding method, image encoding program and image decoding program, and computer readable recording medium recorded with image encoding program and computer readable recording medium recorded with image decoding program |
JP5348881B2 (en) * | 2007-12-25 | 2013-11-20 | セミコンダクター・コンポーネンツ・インダストリーズ・リミテッド・ライアビリティ・カンパニー | Vibration compensation control circuit |
-
2006
- 2006-05-30 KR KR1020060049080A patent/KR101330630B1/en not_active IP Right Cessation
-
2007
- 2007-03-13 US US11/717,208 patent/US10034000B2/en active Active
- 2007-03-13 WO PCT/KR2007/001217 patent/WO2007105900A1/en active Application Filing
- 2007-03-13 CN CN2007800091920A patent/CN101401437B/en not_active Expired - Fee Related
- 2007-03-13 EP EP12184397.3A patent/EP2538679A3/en not_active Ceased
- 2007-03-13 JP JP2009500288A patent/JP2009529845A/en active Pending
- 2007-03-13 EP EP07715613.1A patent/EP1994763B1/en not_active Not-in-force
-
2011
- 2011-10-12 JP JP2011225278A patent/JP2012034410A/en active Pending
- 2011-11-03 KR KR1020110114132A patent/KR101383693B1/en active IP Right Grant
-
2012
- 2012-03-02 US US13/410,601 patent/US9654779B2/en active Active
-
2014
- 2014-05-23 JP JP2014107202A patent/JP2014197865A/en active Pending
- 2014-05-23 JP JP2014107203A patent/JP2014158305A/en active Pending
-
2015
- 2015-04-21 US US14/692,264 patent/US20150229920A1/en not_active Abandoned
- 2015-04-21 US US14/692,289 patent/US20150229922A1/en not_active Abandoned
- 2015-04-21 US US14/692,176 patent/US20150249831A1/en not_active Abandoned
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05137130A (en) * | 1991-11-08 | 1993-06-01 | Victor Co Of Japan Ltd | Encoder and decoder for moving image |
JPH05308631A (en) * | 1992-03-03 | 1993-11-19 | Toshiba Corp | Moving image encoder |
Non-Patent Citations (1)
Title |
---|
See also references of EP1994763A4 * |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102282852A (en) * | 2008-10-01 | 2011-12-14 | 韩国电子通信研究院 | Image encoder and decoder using unidirectional prediction |
CN103384333A (en) * | 2008-10-01 | 2013-11-06 | 韩国电子通信研究院 | Image encoder and decoder using unidirectional prediction |
CN102282852B (en) * | 2008-10-01 | 2014-09-24 | 韩国电子通信研究院 | Image encoder and decoder using unidirectional prediction |
US8867854B2 (en) | 2008-10-01 | 2014-10-21 | Electronics And Telecommunications Research Institute | Image encoder and decoder using undirectional prediction |
US9332282B2 (en) | 2008-10-01 | 2016-05-03 | Electronics And Telecommunications Research Institute | Image encoder and decoder using undirectional prediction |
US9332281B2 (en) | 2008-10-01 | 2016-05-03 | Electronics And Telecommunications Research Institute | Image encoder and decoder using undirectional prediction |
US9369737B2 (en) | 2008-10-01 | 2016-06-14 | Electronics And Telecommunications Research Institute | Image encoder and decoder using undirectional prediction |
US9407937B2 (en) | 2008-10-01 | 2016-08-02 | Electronics And Telecommunications Research Institute | Image encoder and decoder using undirectional prediction |
CN104219523B (en) * | 2008-10-01 | 2018-02-16 | 韩国电子通信研究院 | Use the video encoding/decoding method of predictive mode |
US9942554B2 (en) | 2008-10-01 | 2018-04-10 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US10178393B2 (en) | 2008-10-01 | 2019-01-08 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US10321137B2 (en) | 2008-10-01 | 2019-06-11 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US20190281305A1 (en) | 2008-10-01 | 2019-09-12 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US10742996B2 (en) | 2008-10-01 | 2020-08-11 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US10917647B2 (en) | 2008-10-01 | 2021-02-09 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US11277622B2 (en) | 2008-10-01 | 2022-03-15 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US11683502B2 (en) | 2008-10-01 | 2023-06-20 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
US11882292B2 (en) | 2008-10-01 | 2024-01-23 | Electronics And Telecommunications Research Institute | Image encoder and decoder using unidirectional prediction |
Also Published As
Publication number | Publication date |
---|---|
US20120163467A1 (en) | 2012-06-28 |
US20150229922A1 (en) | 2015-08-13 |
KR20070093305A (en) | 2007-09-18 |
EP2538679A2 (en) | 2012-12-26 |
KR101383693B1 (en) | 2014-04-14 |
US20070211797A1 (en) | 2007-09-13 |
US10034000B2 (en) | 2018-07-24 |
JP2009529845A (en) | 2009-08-20 |
EP1994763A4 (en) | 2010-11-17 |
JP2014197865A (en) | 2014-10-16 |
KR101330630B1 (en) | 2013-11-22 |
JP2012034410A (en) | 2012-02-16 |
US20150229920A1 (en) | 2015-08-13 |
EP1994763A1 (en) | 2008-11-26 |
CN101401437B (en) | 2011-05-25 |
JP2014158305A (en) | 2014-08-28 |
KR20110135843A (en) | 2011-12-19 |
EP2538679A3 (en) | 2013-08-07 |
US9654779B2 (en) | 2017-05-16 |
EP1994763B1 (en) | 2014-05-21 |
US20150249831A1 (en) | 2015-09-03 |
CN101401437A (en) | 2009-04-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1994763B1 (en) | Method, medium, and system encoding and/or decoding moving pictures by adaptively applying optimal prediction modes | |
US8111914B2 (en) | Method and apparatus for encoding and decoding image by using inter color compensation | |
JP5982612B2 (en) | Intra prediction mode decoding method | |
US8014026B2 (en) | Image encoding and/or decoding system, medium, and method | |
US20150124868A1 (en) | Moving picture coding/decoding method and apparatus | |
US8265149B2 (en) | Method and apparatus encoding and/or decoding image by using diffusion properties of the image | |
US20100034265A1 (en) | Apparatus and method for encoding and decoding high fidelity video, and computer-readable storage medium | |
WO2008020687A1 (en) | Image encoding/decoding method and apparatus | |
WO2007108640A2 (en) | Image encoding/decoding method and apparatus | |
WO2008004768A1 (en) | Image encoding/decoding method and apparatus | |
EP1997317A1 (en) | Image encoding/decoding method and apparatus | |
WO2008004769A1 (en) | Image encoding/decoding method and apparatus | |
KR102349435B1 (en) | The method of encoding and decoding of quantization matrix and the apparatus for using the same | |
JP4939273B2 (en) | Image coding apparatus and image coding method | |
KR100727991B1 (en) | Method for intra predictive coding for image data and encoder thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07715613 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009500288 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007715613 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200780009192.0 Country of ref document: CN |