WO2006109985A1 - Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding - Google Patents

Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding Download PDF

Info

Publication number
WO2006109985A1
WO2006109985A1 PCT/KR2006/001338 KR2006001338W WO2006109985A1 WO 2006109985 A1 WO2006109985 A1 WO 2006109985A1 KR 2006001338 W KR2006001338 W KR 2006001338W WO 2006109985 A1 WO2006109985 A1 WO 2006109985A1
Authority
WO
WIPO (PCT)
Prior art keywords
intra
layer frame
residual signals
frame
base layer
Prior art date
Application number
PCT/KR2006/001338
Other languages
French (fr)
Inventor
Sang-Chang Cha
Woo-Jin Han
Original Assignee
Samsung Electronics Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020050053661A external-priority patent/KR100703774B1/en
Application filed by Samsung Electronics Co., Ltd. filed Critical Samsung Electronics Co., Ltd.
Publication of WO2006109985A1 publication Critical patent/WO2006109985A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/587Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal sub-sampling or interpolation, e.g. decimation or subsequent interpolation of pictures in a video sequence
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques

Definitions

  • the present invention relates generally to a method and apparatus for encoding and decoding macroblocks in an intra-base-layer prediction mode by selectively applying intra-coding.
  • Data can be compressed by eliminating spatial redundancy, such as the case where the same color or object is repeated in an image, temporal redundancy, such as the case where there is little change between neighboring frames or the same sound is repeated, or perceptual/visual redundancy, which takes into account human insensitivity to high frequencies.
  • spatial redundancy such as the case where the same color or object is repeated in an image
  • temporal redundancy such as the case where there is little change between neighboring frames or the same sound is repeated
  • perceptual/visual redundancy which takes into account human insensitivity to high frequencies.
  • temporal redundancy is eliminated by temporal filtering based on motion compensation
  • spatial redundancy is eliminated by a spatial transform.
  • transmission media are necessary. Performance differs according to the transmission medium.
  • Currently used transmission media have various transmission speeds ranging from the speed of an ultra high-speed communication network, which can transmit data at a transmission rate of several tens of megabits per second, to the speed of a mobile communication network, which can transmit data at a transmission rate of 384 Kbits per second.
  • a scalable video encoding method which can support transmission media having a variety of speeds or can transmit multimedia at a transmission speed suitable for each transmission environment, is required.
  • the size of a screen such as the aspect ratio (e.g., 4:3 or 16:9) may vary according to the size or characteristics of a reproduction apparatus at the time of reproduction of the multimedia data.
  • Such a scalable video coding method refers to a coding method that allows a video resolution, frame rate, signal-to-noise ratio (SNR), and other parameters to be adjusted by truncating part of an already compressed bitstream in conformity with surrounding conditions, such as the transmission bit rate, transmission error rate, and system source.
  • SNR signal-to-noise ratio
  • MPEG-21 Moving Picture Experts Group-21
  • scalability can be implemented in such a way that multiple layers, including a base layer, a first enhancement layer and a second enhancement layer, are provided, and respective layers are constructed to have different resolutions, such as a Quarter Common Intermediate Format (QCIF), a Common Intermediate Format (CIF) and a 2CIF, or different frame rates.
  • QCIF Quarter Common Intermediate Format
  • CIF Common Intermediate Format
  • 2CIF 2CIF
  • MVs motion vectors
  • the MVs are obtained separately for each layer and are then used, or they are obtained from a single layer and are then used for other layers (without change or after up/down-sampling).
  • the former case has the advantage of finding exact MVs and the disadvantage that the MVs generated for each layer act as overhead.
  • a goal is to more efficiently eliminate redundancy between the MVs for each layer.
  • FIG. 1 is a diagram showing an example of a conventional scalable video codec using a multi-layer structure.
  • a base layer is defined as a layer having a QCIF and a frame rate of 15 Hz
  • a first enhancement layer is defined as a layer having a CIF and a frame rate of 30 Hz
  • a second enhancement layer is defined as a layer having Standard Definition (SD) format and a frame rate of 60 Hz.
  • SD Standard Definition
  • FlG. 2 is a schematic diagram illustrating the three prediction methods; it shows case (1) where intra-prediction is performed on an arbitrary macroblock 14 of a current frame 11, case (2) where inter-prediction is performed using the current frame 11 and a frame 12 existing at a different temporal location different than that of the current frame 11, and case (3) where intra-BL prediction is performed using texture data for region 16 of a base layer frame 13 corresponding to a macroblock 14.
  • an advantageous method is selected from the three prediction methods and is used on a macroblock basis.
  • FlG. 3 is a diagram illustrating an intra-BL prediction method, which is one of the three prediction methods. Since coding is performed with reference to the macroblock 22 of a base layer frame, a macroblock 24, which is constructed from residual signals obtained by calculating the difference between an original macroblock 21 and the macroblock 22 of the base layer frame, is encoded. In this case, the respective residual signals of sub-blocks constituting each macroblock can be obtained. This is similar to an inter-coding method in that residuals between two frames are obtained. That is, in FlG.
  • the residual signals which are obtained by calculating differences between the sub-blocks 25 of the original macroblock 21 and the sub-blocks 26 of the macroblock 22 of the base layer frame, construct the sub-blocks 28 of the macroblock 24 for which intra-BL prediction is used.
  • the present invention has been made keeping in mind the above problems occurring in the prior art, and an aspect of the present invention increases a compression rate using the similarity existing between pieces of information of sub- blocks within a macroblock that is encoded by intra-BL prediction.
  • Another aspect of the present invention increases a compression rate using an intra prediction method at the time of compressing video information in an intra-BL mode.
  • Exemplary embodiments of the present invention provide methods of encoding video signals in intra-BL prediction mode by selectively applying intra coding in a multilayer-based video encoder, the method including: calculating the difference between an input frame and a base layer frame calculated from the input frame and obtaining residual signals; converting the residual signals using an intra coding method; and generating an enhancement layer frame including the converted residual signals.
  • exemplary embodiments of the present invention provide methods of decoding video signals in intra-BL prediction mode by selectively applying intra coding in a multilayer-based video decoder, the method including: receiving a base layer frame and an enhancement layer frame; performing an inverse transform when the residual signals of the enhancement layer frame are encoded using an intra coding method; and performing restoration by adding the inversely transformed residual signals to the image signals of the base layer frame.
  • exemplary embodiments of the present invention provide an encoder, which may include: a base layer encoder for generating a base layer frame from an input frame; and an enhancement layer encoder for generating an enhancement layer frame from the input frame; wherein, at the time of generating the macroblock of the enhancement layer frame, the enhancement layer encoder includes a conversion unit for performing intra coding on residual signals obtained by calculating the difference between a macroblock of the base layer, which corresponds to the macroblock of the enhancement layer frame, and the macroblock of the input frame.
  • exemplary embodiments of the present invention provide a decoder, which may include: a base layer decoder for restoring a base layer frame; and an enhancement layer decoder for restoring an enhancement layer frame; wherein the enhancement layer decoder performs an inverse transform on residual signals and performs restoration by adding inversely transformed residual signals to image signals of the restored base layer frame, thus restoring the image signals when the residual signals are encoded using an intra-coding method.
  • FIG. 1 is a diagram showing a scalable video codec that uses a multi-layer structure
  • FlG. 2 is a schematic diagram illustrating three prediction methods
  • FlG. 3 is a diagram illustrating the intra-BL prediction method
  • FlG. 4 is a conceptual diagram illustrating the encoding of macroblocks by intra-
  • FlG. 5 is a conceptual diagram illustrating the decoding of macroblocks by intra-
  • FlG. 6 is a block diagram showing the construction of an encoder according to an exemplary embodiment of the present invention.
  • FlG. 7 is a block diagram showing the construction of a decoder according to an exemplary embodiment of the present invention.
  • FlG. 8 is a flowchart illustrating a process of encoding a video signal according to an exemplary embodiment of the present invention
  • FlG. 9 is a flowchart illustrating a process of decoding a video signal according to an exemplary embodiment of the present invention.
  • FlG. 10 is an exemplary diagram illustrating a bit set unit for indicating that the method of the present invention is used when intra-BL prediction is performed according to an exemplary embodiment of the present invention.
  • These computer program instructions may also be stored in computer-usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions, which are stored in the computer-usable or computer-readable memory, enables the production of a product that includes an instruction means for implementing the functions specified in the flowchart block or blocks.
  • the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operation steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process so that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
  • each block in the flowchart illustrations may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur in a different order. For example, two blocks shown in succession may in fact be executed concurrently or may sometimes be executed in reverse order, depending upon the desired functionality.
  • FIG. 4 is a conceptual diagram illustrating the case where a method of encoding macroblocks using intra-BL prediction according to an exemplary embodiment of the present invention is employed.
  • the encoding of macroblocks using intra-BL prediction as described in conjunction with FIG. 4, generates the macroblock 105 of an enhancement layer frame based on the difference between the macroblock 101 of an original video frame and the macroblock 102 of a base layer frame.
  • respective sub-blocks are converted in order to compress information.
  • Image signals or residual signals constituting sub-blocks can be compressed and converted using methods, such as the Discrete Cosine Transform (DCT), wavelet transform, Hadamard transform, and Fourier transform.
  • FIG. 4 shows an example of performing the DCT transform on respective sub-blocks.
  • DCT Discrete Cosine Transform
  • DC Direct Current
  • AC Alternating Current
  • the DC component of each sub-block may be regarded as a characteristic of the corresponding sub-block.
  • a macroblock 105 based on intra-BL prediction is generated from the difference between the macroblock 101 of the original video frame and the macroblock 102 of the base layer frame and, as a result, the sub-blocks of the macroblock 105 have similar information values.
  • a similarity also exists between the DC components of sub-blocks 51, 52, 53, — .
  • compression can be performed in such a manner that the DC components are combined as indicated by reference numeral 151, and the similarity therebetween is eliminated, like the intra-coding applied in an intra- mode method.
  • results obtained by compressing the DC components using the Hadamard transform are indicated by reference numeral 152.
  • FIG. 5 is a conceptual diagram illustrating the case where a method of decoding macroblocks using intra-BL prediction according to an exemplary embodiment of the present invention is employed.
  • Data 152 which are obtained by compressing the DC components generated in FlG. 4 using the Hadamard transform, are decompressed using an inverse Hadamard transform, thereby restoring the DC components.
  • a macroblock 205 is generated by combining the restored DC components 155 and AC components 157. Since the macroblock 205 is a macroblock of an intra-BL mode, a macroblock 201 to be output as an image can be restored by adding the macroblock 205 to the macroblock 202 of the base layer.
  • a module means, but is not limited to, a software or ha rdware component, such as a Field Programmable Gate Array (FPGA) or an Application Specific Integrated Circuit (ASIC), which performs certain tasks.
  • a module may advantageously be configured to reside on the addressable storage medium and may be configured to execute on one or more processors.
  • a module may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.
  • the components and modules may be combined into fewer components and modules or further separated into additional components and modules.
  • the components and modules may be implemented to operate one or more central processing units (CPUs) residing in a device or a secure multimedia card.
  • CPUs central processing units
  • FlG. 6 is a block diagram showing the construction of an encoder according to an exemplary embodiment of the present invention.
  • FlG. 6 is a block diagram showing the construction of an encoder according to an exemplary embodiment of the present invention.
  • FlG. 7 which will be given later, the case of using a single BL and a single enhancement layer is described, it should be apparent to those skilled in the art that the present invention can be applied between a lower layer and a current layer even if more layers are used.
  • the video encoder 500 may be classified into an enhancement layer encoder 400 and a BL encoder 300. First, the construction of the base layer encoder 300 is described below.
  • a down-sampler 310 may down-sample the input video to a resolution and frame rate suitable for the base layer, or it performs the down-sampling in accordance with a desired size of a video image. From the point of view of resolution, the down-sampling may be realized using an MPEG down-sampler or a wavelet down-sampler. From the point of view of frame rate, the down-sampling may be performed using a frame skip method, a frame interpolation method or the like. Down-sampling in accordance with a desired size of a video image refers to a process of adjusting the size thereof so that an original input video image having an aspect ratio of 16:9 can be viewed at an aspect ratio of 4:3. For this purpose, a method of eliminating information corresponding to a boundary region from video information, or a method of reducing the video information to conform to the size of a corresponding screen may be used.
  • a motion estimation unit 350 may perform motion estimation on the base layer frame, thus obtaining MVs for partitions constituting the base layer frame.
  • Motion estimation is a process of searching for a region that is most similar to the respective partitions of a current frame Fc; that is, a region of a previous reference frame Fr' stored in a frame buffer 380 where the error is small.
  • Motion estimation may be performed using various methods, such as a fixed size block matching method and a hierarchical variable size block matching method.
  • the previous reference frame Fr' may be provided from the frame buffer 380.
  • a closed-loop encoding scheme may additionally or alternatively adopt an open-loop encoding scheme using the original base layer frame, which may be provided by the down-sampler 310, as a reference frame.
  • the MVs obtained by the motion estimation unit 350 may be transferred to a virtual region frame generation unit 390.
  • the reason for this is to generate virtual region frames to which virtual regions may be added in the case where the MVs of the boundary region blocks of the current frame are headed for the center of the frame.
  • a motion compensation unit 360 may perform motion compensation on the reference frame using the obtained MVs.
  • a subtractor 315 may calculate the difference between the current frame Fc of the base layer and the motion-compensated reference frame, thus generating a residual frame.
  • a conversion unit 320 may perform a spatial transform on the generated residual frame, thus generating transform coefficients.
  • the Discrete Cosine Transform (DCT) or the wavelet transform may be used as the spatial transform method.
  • the transform coefficients are DCT coefficients in the case where the DCT method is employed, and wavelet coefficients in the case where the wavelet transform is employed.
  • a quantization unit 330 may quantize the transform coefficients generated by the conversion unit 320. Quantization refers to a process of representing the conversion coefficients as discrete values by dividing the conversion coefficients, which are expressed as real numbers, at predetermined intervals, and matching the discrete values to predetermined indices. As described above, the quantized result values are called quantized coefficients.
  • the entropy encoding unit 340 may encode the transform coefficients, which have been quantized by the quantization unit 330, and MVs, which may be generated by the motion estimation unit 350, without loss, thus generating a base layer bitstream.
  • Various lossless encoding methods such as an arithmetic encoding method and a variable length encoding method may be used as such a lossless encoding method.
  • an inverse quantization unit 371 may dequantize the quantized coefficients output from the quantization unit 330.
  • Such a dequantization process is the inverse of the quantization process and is a process of restoring matched quantization coefficients based on the indices, which have been generated for the quantization process, using a quantization table used in the quantization process.
  • An inverse conversion unit 372 may perform an inverse spatial transform on the inversely quantized results.
  • the inverse spatial transform is performed in a reverse order relative to the transform process of the conversion unit 320.
  • the Inverse Discrete Cosine Transform (IDCT) or the inverse wavelet transform may be used as such an inverse spatial transform method.
  • An adder 325 may add the output values of the motion compensation unit 360 and the output values of the inverse conversion unit 372 to restore the current frame (Fc'), and provide the restored frame Fc' to the frame buffer 380.
  • the frame buffer 380 may temporarily store the restored frame and provide it as a reference frame for the inter- prediction of other base layer frames.
  • the restored frame Fc' may be provided to the enhancement layer encoder 400 via an up-sampler 395.
  • the up-sampling process of the up-sampler 395 may be omitted if the resolution of the base layer is identical to that of the enhancement layer.
  • a frame which may be provided by the base layer encoder 300, and an input frame may be input to a subtractor 410.
  • the subtracter 210 may calculate the difference between the input frame and the input base layer frame, which may include a virtual region, thus generating a residual frame.
  • the residual frame may be converted into a bitstream via a conversion unit 420, a quantization unit 430, and an entropy encoding unit 440, and may then be output.
  • the conversion unit 420 of the enhancement layer encoder 400 may perform a spatial transform on the residual signals between the macroblocks of the input frame and the macroblocks of the base layer frame.
  • the DCT or the wavelet transform may be used as the spatial transform method. Due to the characteristics of the macroblocks of the enhancement layer, a similarity exists between the DCT coefficients obtained when DCT is used; the same is true of the wavelet coefficients Accordingly, a process of eliminating the similarity existing between these coefficients and, thereby, increasing the compression rate may be performed by the conversion unit 420 of the enhancement layer encoder 400.
  • the Hadamard transform which has been described in conjunction with FIG. 4, may be employed.
  • Macroblocks may be constructed using the difference signals between the macroblocks of the base layer frame and macroblocks of the input frame in a manner similar to the temporal inter-prediction.
  • the enhancement layer encoder 400 shown in FlG. 6 has been described with emphasis on the encoding of the results of intra-BL prediction of the base layer frame.
  • selective encoding may be performed using a temporal inter- prediction method or a directional intra-prediction method.
  • FlG. 7 is a block diagram showing the construction of a decoder according to an exemplary embodiment of the present invention.
  • the video decoder 550 may be divided into an enhancement layer decoder 700 and a base layer decoder 600. First, the construction of the base layer decoder 600 is described below.
  • An entropy decoding unit 610 may decode a base layer bitstream without loss, and extract texture data of a base layer frame and motion data (MVs, partition information, and a reference frame number).
  • MVs texture data of a base layer frame and motion data
  • a inverse quantization unit 620 may dequantize the texture data. Such a de- quantization process may be the inverse of the quantization process performed in the video encoder 500 Dequantization is a process of restoring quantization coefficients based on the indices, which were generated in the quantization process, using a quant ization table used in the quantization process.
  • An inverse conversion unit 630 may perform an inverse spatial transform on the resulting inversely quantized results, thus restoring a residual frame.
  • the inverse spatial transform may be performed in reverse order to the transform process of the conversion unit 320 of the video encoder 500.
  • the inverse spatial transform method (IDCT) or the inverse wavelet transform may be used.
  • An entropy decoding unit 610 may provide motion data, including MVs, to a motion compensation unit 660.
  • the motion compensation unit 660 may perform motion compensation on a previously restored video frame, that is, a reference frame, which may be provided by a frame buffer 650, using the motion data which may be provided by the entropy decoding unit 610, thus generating a motion compensation frame.
  • An adder 615 may add the residual frame, which may be restored by the inverse conversion unit 630, to the motion compensation frame which may be generated by the motion compensation unit 660, thus restoring the base layer video frame.
  • the restored video frame may be temporarily stored in the frame buffer 650, and may be provided to the motion compensation unit 660 as a reference frame to restore subsequent frames.
  • a restored frame Fc' which is restored from a current frame, may be provided to an enhancement layer decoder 700 via an up-sampler 680. Accordingly, the up- sampling process may be omitted if the resolution of the base layer is identical to that of the enhancement layer. Furthermore, the up-sampling process may be omitted if part of the region information is eliminated by the comparison of the video information of the base layer with the video information of the enhancement layer.
  • the entropy decoding unit 710 may decode the input bitstream without loss, thus extracting the texture data of an asynchronous frame.
  • the extracted texture data may be restored to the residual frame via a quantization unit 720 and an inverse conversion unit 730.
  • the function and operation of the inverse quantization unit 720 may be identical to those of the inverse quantization unit 620 of the base layer decoder 550.
  • An adder 715 may add the base layer frame, which is provided by the base layer decoder 600, to the restored residual frame, thus restoring the original frame.
  • the inverse conversion unit 730 of the enhancement layer decoder 700 may perform an inverse transform based on the method by which the enhanced bitstream of a received macroblock was encoded.
  • the encoding method as described in conjunction with FIG. 6, may determine whether the step of eliminating the similarity between transform coefficients, such as DCT coefficients or wavelet coefficients, which exist in the sub-blocks of each macroblock, was performed in the process of obtaining the difference using the macroblocks of the base layer frame.
  • the inverse process thereof may be performed.
  • the transform coefficients such as DCT coefficients or wavelet coefficients
  • a macroblock constituted by residual signals may be restored based on the restored coefficients. This process has been described in conjunction with FIG. 5.
  • the enhancement layer decoder 700 shown in FIG. 7 has been described based on the operation of performing decoding on the base layer frame using intra-BL prediction. In addition, as described in conjunction with FIG. 2, it should be appreciated by those skilled in the art that selective decoding may be performed using an inter-prediction method or an intra-prediction method.
  • FIG. 8 is a flowchart illustrating a process of encoding a video signal according to an exemplary embodiment of the present invention.
  • An input frame is received and a base layer frame is generated in S 101.
  • the prediction mode varies on a macroblock basis, it is determined which prediction mode (temporal inter-prediction mode, directional intra-prediction mode, and intra-BL prediction mode) provides the highest compression rate for respective macroblocks. If, as a result, the intra-BL prediction mode is selected in S 105, residuals between the corresponding macroblock of the base layer frame and the macroblock of the input frame is obtained in Sl 10. Thereafter, conversion is performed on residual signals in Sl 11. In this case, DCT transform or wavelet transform may be performed. The extent of similarity between transform coefficients obtained by the conversion is determined in S 120.
  • the similarity between the transform coefficients is determined to be high. If the resolution of the base layer frame is different from that of the enhancement layer frame, the similarity therebetween is determined to be low. This is only one embodiment.
  • S 130 the actual correlation between the transform coefficients is obtained, and it is determined that the similarity between the transform coefficient is high when the obtained correlation exceeds a predetermined level. When a similarity exists between the transform coefficients, the similarity is eliminated in S 130.
  • the above-described Hadamard transform may be employed, and the DCT, wavelet transform and Fourier transform are also employed. With respect to operational speed, the Hadamard transform may be faster than the other methods due to the use of addition and subtraction.
  • S 131 is directly performed without performing S 130. In order to notify a decoding stage of whether the similarity has been eliminated, one bit may be set.
  • the intra-BL prediction mode is not used in S 105, the temporal inter-prediction mode or spatial intra-prediction mode is used in S 108.
  • FIG. 9 is a flowchart illustrating a process of decoding a video signal according to an exemplary embodiment of the present invention.
  • a base layer frame and an enhancement layer frame are extracted from a received bitstream in S201. It is determined whether intra-BL mode was used as a prediction mode when encoding macroblocks constituting the enhancement layer frame in S205. If the intra-BL prediction mode was not used, inverse transform is performed based on temporal inter- prediction mode or spatial intra-prediction mode in S208. If the intra-BL prediction mode was used, the transform coefficients for the sub-blocks of each macroblock are extracted in S210. Thereafter, it is determined whether the similarity between the transform coefficients has been eliminated in S215.
  • the transform coefficients may be calculated using an inverse transform in S220.
  • the inverse Hadamard transform which corresponds to the Hadamard transform performed during encoding, is an example of an inverse transform that may be used. If it is determined that the similarity has not been eliminated in S215, the process proceeds to S230. When the transform coefficients are obtained, the residual signals of each macroblock are restored based on the transform coefficients obtained in S230. The restored residual signals are added to the macroblock of the base layer frame and, thereby, the macroblock of a video image is restored in S231.
  • FlG. 10 is an exemplary diagram illustrating a bit set unit for indicating that the method of the present invention is used when intra-BL prediction is performed according to an exemplary embodiment of the present invention.
  • Video is composed of video sequences.
  • the video sequence is composed of Groups
  • One frame or picture is composed of a plurality of slices, and each of the slices includes a plurality of macroblocks.
  • one prediction mode may be selected from three prediction modes, such as directional intra-prediction, temporal inter-prediction and intra-BL prediction. Accordingly, when intra-BL prediction, proposed by an exemplary embodiment of the present invention, is performed, intra- coding may be performed on a macroblock basis. However, if one bit is additionally used to determine whether, on a macroblock basis, intra-coding or inter-coding is performed, many bits may be necessary for the overall frames or the overall slices.
  • the number of bits may be set on a macroblock basis, and the number of bits may also be set on a slice basis or on a frame basis. As shown in FlG. 10, the number of bits may be set on a macroblock basis. Furthermore, one bit may be set for all the macroblocks constituting a corresponding slice. In this case, information requirements can be reduced because one bit is assigned to each slice.
  • a compression rate may be increased by eliminating the similarity that exists between the pieces of information of the sub- blocks of each macroblock to be encoded using intra-BL prediction.
  • the compression rate may be increased by applying an intra-prediction method when video information is compressed using an intra-BL mode and, therefore, the amount of data transmitted over a network may be reduced.

Abstract

A method and apparatus for encoding and decoding macroblocks in an intra-base layer prediction mode by selectively applying intra-coding are provided. The method includes the steps of calculating a difference between an input frame and a base layer frame calculated from the input frame and obtaining residual signals, converting the residual signals using an intra-coding method, and generating an enhancement layer frame including the converted residual signals.

Description

Description
METHOD AND APPARATUS FOR ENCODING AND
DECODING VIDEO SIGNALS IN INTRA-BASE-LAYER
PREDICTION MODE BY SELECTIVELY APPLYING INTRA-
CODING
Technical Field
[1] The present invention relates generally to a method and apparatus for encoding and decoding macroblocks in an intra-base-layer prediction mode by selectively applying intra-coding.
Background Art
[2] As information and communication technology, including the Internet, develops, image-based communication as well as text-based communication and voice-based communication is increasing. The existing text-based communication is insufficient to satisfy consumers' various demands. Therefore, the provision of multimedia services capable of accommodating various types of information, such as text, images and music, is increasing. Since multimedia data files are large, they require high-capacity storage media and a broad bandwidth at the time of transmission. Therefore, to transmit multimedia data, including text, images and audio, it is essential to compress the data.
[3] The fundamental principle of data compression is to eliminate data redundancy.
Data can be compressed by eliminating spatial redundancy, such as the case where the same color or object is repeated in an image, temporal redundancy, such as the case where there is little change between neighboring frames or the same sound is repeated, or perceptual/visual redundancy, which takes into account human insensitivity to high frequencies. In a general coding method, temporal redundancy is eliminated by temporal filtering based on motion compensation, and spatial redundancy is eliminated by a spatial transform.
[4] In order to transmit multimedia data after the redundancy has been removed, transmission media are necessary. Performance differs according to the transmission medium. Currently used transmission media have various transmission speeds ranging from the speed of an ultra high-speed communication network, which can transmit data at a transmission rate of several tens of megabits per second, to the speed of a mobile communication network, which can transmit data at a transmission rate of 384 Kbits per second. In these environments, a scalable video encoding method, which can support transmission media having a variety of speeds or can transmit multimedia at a transmission speed suitable for each transmission environment, is required. Also, the size of a screen, such as the aspect ratio (e.g., 4:3 or 16:9) may vary according to the size or characteristics of a reproduction apparatus at the time of reproduction of the multimedia data.
[5] Such a scalable video coding method refers to a coding method that allows a video resolution, frame rate, signal-to-noise ratio (SNR), and other parameters to be adjusted by truncating part of an already compressed bitstream in conformity with surrounding conditions, such as the transmission bit rate, transmission error rate, and system source. With regard to the scalable video encoding method, standardization is in progress in Moving Picture Experts Group-21 (MPEG-21) Part 10. In particular, much research into multi-layer based scalability has been carried out. For example, scalability can be implemented in such a way that multiple layers, including a base layer, a first enhancement layer and a second enhancement layer, are provided, and respective layers are constructed to have different resolutions, such as a Quarter Common Intermediate Format (QCIF), a Common Intermediate Format (CIF) and a 2CIF, or different frame rates.
[6] In the case of coding for a multiple layer, as in the case of coding for a single layer, it is necessary to obtain motion vectors (MVs) for eliminating temporal redundancy from each layer. The MVs are obtained separately for each layer and are then used, or they are obtained from a single layer and are then used for other layers (without change or after up/down-sampling). When comparing the two methods, the former case has the advantage of finding exact MVs and the disadvantage that the MVs generated for each layer act as overhead. In the former case, a goal is to more efficiently eliminate redundancy between the MVs for each layer.
[7] FIG. 1 is a diagram showing an example of a conventional scalable video codec using a multi-layer structure. First, a base layer is defined as a layer having a QCIF and a frame rate of 15 Hz, a first enhancement layer is defined as a layer having a CIF and a frame rate of 30 Hz, and a second enhancement layer is defined as a layer having Standard Definition (SD) format and a frame rate of 60 Hz. If a 0.5 Mbps CIF stream is desired, a bitstream may be truncated and transmitted to reach a bit rate of 0.5 Mbps based on a CIF_30Hz_0.7Mbps first enhancement layer. In this manner, spatial scalability, temporal scalability and SNR scalability can be implemented.
[8] As shown in FIG. 1, with regard to, for example, frames 10, 20 and 30, which have an identical temporal location and correspond to different layers, it can be assumed that the images thereof will be similar. Accordingly, a method of predicting texture of a current layer based on the texture of a lower layer (directly or after up-sampling), and encoding the difference between the predicted value and the actual value of the texture of a current layer is well known. In 'Scalable Video Model 3.0 of ISO/IEC 21000-13 Scalable Video Coding' (hereinafter referred to as 'SVM 3.0'), the method is defined as intra-Base-Layer (BL) prediction.
[9] In the SVM 3.0 described above, a method of predicting a current block using correlation between a current block and a lower layer block is adopted in addition to inter-prediction and directional intra-prediction used in the existing H.264 digital video codec standard protocol to perform prediction on a block and macroblocks that constitute the current frame. Such a prediction method is called 'intra-BL prediction,' and a mode of performing encoding using the prediction is called 'intra-BL mode.'
[10] FlG. 2 is a schematic diagram illustrating the three prediction methods; it shows case (1) where intra-prediction is performed on an arbitrary macroblock 14 of a current frame 11, case (2) where inter-prediction is performed using the current frame 11 and a frame 12 existing at a different temporal location different than that of the current frame 11, and case (3) where intra-BL prediction is performed using texture data for region 16 of a base layer frame 13 corresponding to a macroblock 14.
[11] In the above-described scalable video coding standard, an advantageous method is selected from the three prediction methods and is used on a macroblock basis.
[12] FlG. 3 is a diagram illustrating an intra-BL prediction method, which is one of the three prediction methods. Since coding is performed with reference to the macroblock 22 of a base layer frame, a macroblock 24, which is constructed from residual signals obtained by calculating the difference between an original macroblock 21 and the macroblock 22 of the base layer frame, is encoded. In this case, the respective residual signals of sub-blocks constituting each macroblock can be obtained. This is similar to an inter-coding method in that residuals between two frames are obtained. That is, in FlG. 3, the residual signals, which are obtained by calculating differences between the sub-blocks 25 of the original macroblock 21 and the sub-blocks 26 of the macroblock 22 of the base layer frame, construct the sub-blocks 28 of the macroblock 24 for which intra-BL prediction is used.
Disclosure of Invention
Technical Problem
[13] However, since sub-blocks of the macroblock 24 that uses intra-BL prediction exist in a single macroblock, a uniform similarity between the residual signals of the sub- blocks exists. Accordingly, in the case of intra-BL prediction it is necessary to calculate the differences in the same macroblock, and a method and apparatus for increasing a compression rate using the similarity between the residual signals of sub- blocks are required.
Technical Solution
[14] Accordingly, the present invention has been made keeping in mind the above problems occurring in the prior art, and an aspect of the present invention increases a compression rate using the similarity existing between pieces of information of sub- blocks within a macroblock that is encoded by intra-BL prediction.
[15] Another aspect of the present invention increases a compression rate using an intra prediction method at the time of compressing video information in an intra-BL mode.
[16] Exemplary embodiments of the present invention provide methods of encoding video signals in intra-BL prediction mode by selectively applying intra coding in a multilayer-based video encoder, the method including: calculating the difference between an input frame and a base layer frame calculated from the input frame and obtaining residual signals; converting the residual signals using an intra coding method; and generating an enhancement layer frame including the converted residual signals.
[17] In addition, exemplary embodiments of the present invention provide methods of decoding video signals in intra-BL prediction mode by selectively applying intra coding in a multilayer-based video decoder, the method including: receiving a base layer frame and an enhancement layer frame; performing an inverse transform when the residual signals of the enhancement layer frame are encoded using an intra coding method; and performing restoration by adding the inversely transformed residual signals to the image signals of the base layer frame.
[18] In addition, exemplary embodiments of the present invention provide an encoder, which may include: a base layer encoder for generating a base layer frame from an input frame; and an enhancement layer encoder for generating an enhancement layer frame from the input frame; wherein, at the time of generating the macroblock of the enhancement layer frame, the enhancement layer encoder includes a conversion unit for performing intra coding on residual signals obtained by calculating the difference between a macroblock of the base layer, which corresponds to the macroblock of the enhancement layer frame, and the macroblock of the input frame.
[19] In addition, exemplary embodiments of the present invention provide a decoder, which may include: a base layer decoder for restoring a base layer frame; and an enhancement layer decoder for restoring an enhancement layer frame; wherein the enhancement layer decoder performs an inverse transform on residual signals and performs restoration by adding inversely transformed residual signals to image signals of the restored base layer frame, thus restoring the image signals when the residual signals are encoded using an intra-coding method.
Description of Drawings
[20] The above and other aspects, features and advantages of the present invention will be more clearly understood from the following detailed description of exemplary embodiments taken in conjunction with the accompanying drawings, in which:
[21] FIG. 1 is a diagram showing a scalable video codec that uses a multi-layer structure;
[22] FlG. 2 is a schematic diagram illustrating three prediction methods;
[23] FlG. 3 is a diagram illustrating the intra-BL prediction method;
[24] FlG. 4 is a conceptual diagram illustrating the encoding of macroblocks by intra-
BL prediction according to an exemplary embodiment of the present invention;
[25] FlG. 5 is a conceptual diagram illustrating the decoding of macroblocks by intra-
BL prediction according to an exemplary embodiment of the present invention;
[26] FlG. 6 is a block diagram showing the construction of an encoder according to an exemplary embodiment of the present invention;
[27] FlG. 7 is a block diagram showing the construction of a decoder according to an exemplary embodiment of the present invention;
[28] FlG. 8 is a flowchart illustrating a process of encoding a video signal according to an exemplary embodiment of the present invention;
[29] FlG. 9 is a flowchart illustrating a process of decoding a video signal according to an exemplary embodiment of the present invention; and
[30] FlG. 10 is an exemplary diagram illustrating a bit set unit for indicating that the method of the present invention is used when intra-BL prediction is performed according to an exemplary embodiment of the present invention.
Mode for Invention
[31] The present invention is described below with reference to drawings of block diagrams and flowcharts illustrating methods and apparatuses for encoding and decoding video signals using an intra-BL prediction mode which selectively applies intra-coding in accordance with exemplary embodiments of the present invention. It should be noted that each block of the flowchart illustrations, and combinations of blocks in the flowchart illustrations, can be implemented using computer program instructions. These computer program instructions can be provided to a processor of a general-purpose computer, a special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute on the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart block or blocks.
[32] These computer program instructions may also be stored in computer-usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions, which are stored in the computer-usable or computer-readable memory, enables the production of a product that includes an instruction means for implementing the functions specified in the flowchart block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operation steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process so that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks. Furthermore, each block in the flowchart illustrations may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur in a different order. For example, two blocks shown in succession may in fact be executed concurrently or may sometimes be executed in reverse order, depending upon the desired functionality.
[33] FIG. 4 is a conceptual diagram illustrating the case where a method of encoding macroblocks using intra-BL prediction according to an exemplary embodiment of the present invention is employed. The encoding of macroblocks using intra-BL prediction, as described in conjunction with FIG. 4, generates the macroblock 105 of an enhancement layer frame based on the difference between the macroblock 101 of an original video frame and the macroblock 102 of a base layer frame. In this case, respective sub-blocks are converted in order to compress information. Image signals or residual signals constituting sub-blocks can be compressed and converted using methods, such as the Discrete Cosine Transform (DCT), wavelet transform, Hadamard transform, and Fourier transform. FIG. 4 shows an example of performing the DCT transform on respective sub-blocks. In order to perform the DCT, Direct Current (DC) components are obtained from the upper-left sides of respective sub-blocks and, subsequently, Alternating Current (AC) components are obtained. The DC component of each sub-block may be regarded as a characteristic of the corresponding sub-block. However, a macroblock 105 based on intra-BL prediction is generated from the difference between the macroblock 101 of the original video frame and the macroblock 102 of the base layer frame and, as a result, the sub-blocks of the macroblock 105 have similar information values. Thus, a similarity also exists between the DC components of sub-blocks 51, 52, 53, — . Accordingly, compression can be performed in such a manner that the DC components are combined as indicated by reference numeral 151, and the similarity therebetween is eliminated, like the intra-coding applied in an intra- mode method. As shown in FIG. 4, results obtained by compressing the DC components using the Hadamard transform are indicated by reference numeral 152.
[34] In contrast to the transfer of the macroblock 105, which is constructed using reference numeral 151 having DC components and AC components corresponding to the DC components as encoding results, the transfer of data 152, which is compressed more than data 105, generates a relatively high compression rate.
[35] FIG. 5 is a conceptual diagram illustrating the case where a method of decoding macroblocks using intra-BL prediction according to an exemplary embodiment of the present invention is employed. Data 152, which are obtained by compressing the DC components generated in FlG. 4 using the Hadamard transform, are decompressed using an inverse Hadamard transform, thereby restoring the DC components. A macroblock 205 is generated by combining the restored DC components 155 and AC components 157. Since the macroblock 205 is a macroblock of an intra-BL mode, a macroblock 201 to be output as an image can be restored by adding the macroblock 205 to the macroblock 202 of the base layer.
[36] The term 'module' as used herein means, but is not limited to, a software or ha rdware component, such as a Field Programmable Gate Array (FPGA) or an Application Specific Integrated Circuit (ASIC), which performs certain tasks. A module may advantageously be configured to reside on the addressable storage medium and may be configured to execute on one or more processors. Thus, a module may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables. The components and modules may be combined into fewer components and modules or further separated into additional components and modules. Furthermore, the components and modules may be implemented to operate one or more central processing units (CPUs) residing in a device or a secure multimedia card.
[37] FlG. 6 is a block diagram showing the construction of an encoder according to an exemplary embodiment of the present invention. Although, in the description of FlG. 6 and in the description of FlG. 7, which will be given later, the case of using a single BL and a single enhancement layer is described, it should be apparent to those skilled in the art that the present invention can be applied between a lower layer and a current layer even if more layers are used.
[38] The video encoder 500 may be classified into an enhancement layer encoder 400 and a BL encoder 300. First, the construction of the base layer encoder 300 is described below.
[39] A down-sampler 310 may down-sample the input video to a resolution and frame rate suitable for the base layer, or it performs the down-sampling in accordance with a desired size of a video image. From the point of view of resolution, the down-sampling may be realized using an MPEG down-sampler or a wavelet down-sampler. From the point of view of frame rate, the down-sampling may be performed using a frame skip method, a frame interpolation method or the like. Down-sampling in accordance with a desired size of a video image refers to a process of adjusting the size thereof so that an original input video image having an aspect ratio of 16:9 can be viewed at an aspect ratio of 4:3. For this purpose, a method of eliminating information corresponding to a boundary region from video information, or a method of reducing the video information to conform to the size of a corresponding screen may be used.
[40] A motion estimation unit 350 may perform motion estimation on the base layer frame, thus obtaining MVs for partitions constituting the base layer frame. Motion estimation is a process of searching for a region that is most similar to the respective partitions of a current frame Fc; that is, a region of a previous reference frame Fr' stored in a frame buffer 380 where the error is small. Motion estimation may be performed using various methods, such as a fixed size block matching method and a hierarchical variable size block matching method. The previous reference frame Fr' may be provided from the frame buffer 380. Although the base layer encoder 300 of FIG. 6 may adopt a scheme using the restored frame as a reference frame, that is, a closed-loop encoding scheme, it may additionally or alternatively adopt an open-loop encoding scheme using the original base layer frame, which may be provided by the down-sampler 310, as a reference frame.
[41] Meanwhile, the MVs obtained by the motion estimation unit 350 may be transferred to a virtual region frame generation unit 390. The reason for this is to generate virtual region frames to which virtual regions may be added in the case where the MVs of the boundary region blocks of the current frame are headed for the center of the frame.
[42] A motion compensation unit 360 may perform motion compensation on the reference frame using the obtained MVs. A subtractor 315 may calculate the difference between the current frame Fc of the base layer and the motion-compensated reference frame, thus generating a residual frame.
[43] A conversion unit 320 may perform a spatial transform on the generated residual frame, thus generating transform coefficients. The Discrete Cosine Transform (DCT) or the wavelet transform may be used as the spatial transform method. The transform coefficients are DCT coefficients in the case where the DCT method is employed, and wavelet coefficients in the case where the wavelet transform is employed.
[44] A quantization unit 330 may quantize the transform coefficients generated by the conversion unit 320. Quantization refers to a process of representing the conversion coefficients as discrete values by dividing the conversion coefficients, which are expressed as real numbers, at predetermined intervals, and matching the discrete values to predetermined indices. As described above, the quantized result values are called quantized coefficients.
[45] The entropy encoding unit 340 may encode the transform coefficients, which have been quantized by the quantization unit 330, and MVs, which may be generated by the motion estimation unit 350, without loss, thus generating a base layer bitstream. Various lossless encoding methods, such as an arithmetic encoding method and a variable length encoding method may be used as such a lossless encoding method.
[46] Meanwhile, an inverse quantization unit 371 may dequantize the quantized coefficients output from the quantization unit 330. Such a dequantization process is the inverse of the quantization process and is a process of restoring matched quantization coefficients based on the indices, which have been generated for the quantization process, using a quantization table used in the quantization process.
[47] An inverse conversion unit 372 may perform an inverse spatial transform on the inversely quantized results. The inverse spatial transform is performed in a reverse order relative to the transform process of the conversion unit 320. The Inverse Discrete Cosine Transform (IDCT) or the inverse wavelet transform may be used as such an inverse spatial transform method.
[48] An adder 325 may add the output values of the motion compensation unit 360 and the output values of the inverse conversion unit 372 to restore the current frame (Fc'), and provide the restored frame Fc' to the frame buffer 380. The frame buffer 380 may temporarily store the restored frame and provide it as a reference frame for the inter- prediction of other base layer frames.
[49] The restored frame Fc' may be provided to the enhancement layer encoder 400 via an up-sampler 395. The up-sampling process of the up-sampler 395 may be omitted if the resolution of the base layer is identical to that of the enhancement layer.
[50] The construction of the enhancement layer encoder 400 is described below. A frame, which may be provided by the base layer encoder 300, and an input frame may be input to a subtractor 410. The subtracter 210 may calculate the difference between the input frame and the input base layer frame, which may include a virtual region, thus generating a residual frame. The residual frame may be converted into a bitstream via a conversion unit 420, a quantization unit 430, and an entropy encoding unit 440, and may then be output.
[51] The conversion unit 420 of the enhancement layer encoder 400 may perform a spatial transform on the residual signals between the macroblocks of the input frame and the macroblocks of the base layer frame. Here, the DCT or the wavelet transform may be used as the spatial transform method. Due to the characteristics of the macroblocks of the enhancement layer, a similarity exists between the DCT coefficients obtained when DCT is used; the same is true of the wavelet coefficients Accordingly, a process of eliminating the similarity existing between these coefficients and, thereby, increasing the compression rate may be performed by the conversion unit 420 of the enhancement layer encoder 400. In order to increase the compression rate, the Hadamard transform, which has been described in conjunction with FIG. 4, may be employed. [52] However, a case exists where the similarity of the coefficients of the sub-blocks of each macroblock is low. In this case, it is not necessary to perform a transform process on the transform coefficients. Macroblocks may be constructed using the difference signals between the macroblocks of the base layer frame and macroblocks of the input frame in a manner similar to the temporal inter-prediction.
[53] Since the functions and operations of the quantization unit 430 and the entropy encoding unit 440 may be identical to those of the quantization unit 330 and the entropy encoding unit 340, respectively, the description thereof is omitted.
[54] The enhancement layer encoder 400 shown in FlG. 6 has been described with emphasis on the encoding of the results of intra-BL prediction of the base layer frame. In addition, as described in conjunction with FlG. 2, it should be appreciated by those skilled in the art that selective encoding may be performed using a temporal inter- prediction method or a directional intra-prediction method.
[55] FlG. 7 is a block diagram showing the construction of a decoder according to an exemplary embodiment of the present invention. The video decoder 550 may be divided into an enhancement layer decoder 700 and a base layer decoder 600. First, the construction of the base layer decoder 600 is described below.
[56] An entropy decoding unit 610 may decode a base layer bitstream without loss, and extract texture data of a base layer frame and motion data (MVs, partition information, and a reference frame number).
[57] A inverse quantization unit 620 may dequantize the texture data. Such a de- quantization process may be the inverse of the quantization process performed in the video encoder 500 Dequantization is a process of restoring quantization coefficients based on the indices, which were generated in the quantization process, using a quant ization table used in the quantization process.
[58] An inverse conversion unit 630 may perform an inverse spatial transform on the resulting inversely quantized results, thus restoring a residual frame. The inverse spatial transform may be performed in reverse order to the transform process of the conversion unit 320 of the video encoder 500. As such, the inverse spatial transform method (IDCT) or the inverse wavelet transform may be used.
[59] An entropy decoding unit 610 may provide motion data, including MVs, to a motion compensation unit 660.
[60] The motion compensation unit 660 may perform motion compensation on a previously restored video frame, that is, a reference frame, which may be provided by a frame buffer 650, using the motion data which may be provided by the entropy decoding unit 610, thus generating a motion compensation frame.
[61] An adder 615 may add the residual frame, which may be restored by the inverse conversion unit 630, to the motion compensation frame which may be generated by the motion compensation unit 660, thus restoring the base layer video frame. The restored video frame may be temporarily stored in the frame buffer 650, and may be provided to the motion compensation unit 660 as a reference frame to restore subsequent frames.
[62] A restored frame Fc', which is restored from a current frame, may be provided to an enhancement layer decoder 700 via an up-sampler 680. Accordingly, the up- sampling process may be omitted if the resolution of the base layer is identical to that of the enhancement layer. Furthermore, the up-sampling process may be omitted if part of the region information is eliminated by the comparison of the video information of the base layer with the video information of the enhancement layer.
[63] The construction of the enhancement layer decoder 700 is described below. When an enhancement layer bitstream is input to an entropy decoding unit 710, the entropy decoding unit 710 may decode the input bitstream without loss, thus extracting the texture data of an asynchronous frame.
[64] Thereafter, the extracted texture data may be restored to the residual frame via a quantization unit 720 and an inverse conversion unit 730. The function and operation of the inverse quantization unit 720 may be identical to those of the inverse quantization unit 620 of the base layer decoder 550.
[65] An adder 715 may add the base layer frame, which is provided by the base layer decoder 600, to the restored residual frame, thus restoring the original frame.
[66] The inverse conversion unit 730 of the enhancement layer decoder 700 may perform an inverse transform based on the method by which the enhanced bitstream of a received macroblock was encoded. The encoding method, as described in conjunction with FIG. 6, may determine whether the step of eliminating the similarity between transform coefficients, such as DCT coefficients or wavelet coefficients, which exist in the sub-blocks of each macroblock, was performed in the process of obtaining the difference using the macroblocks of the base layer frame.
[67] If the step of eliminating the similarity between the coefficients has been included in the encoding process, the inverse process thereof may be performed. As described in conjunction with FIG. 5, the transform coefficients, such as DCT coefficients or wavelet coefficients, may be restored by performing an inverse Hadamard transform, and a macroblock constituted by residual signals may be restored based on the restored coefficients. This process has been described in conjunction with FIG. 5.
[68] The enhancement layer decoder 700 shown in FIG. 7 has been described based on the operation of performing decoding on the base layer frame using intra-BL prediction. In addition, as described in conjunction with FIG. 2, it should be appreciated by those skilled in the art that selective decoding may be performed using an inter-prediction method or an intra-prediction method.
[69] FIG. 8 is a flowchart illustrating a process of encoding a video signal according to an exemplary embodiment of the present invention.
[70] An input frame is received and a base layer frame is generated in S 101. When the prediction mode varies on a macroblock basis, it is determined which prediction mode (temporal inter-prediction mode, directional intra-prediction mode, and intra-BL prediction mode) provides the highest compression rate for respective macroblocks. If, as a result, the intra-BL prediction mode is selected in S 105, residuals between the corresponding macroblock of the base layer frame and the macroblock of the input frame is obtained in Sl 10. Thereafter, conversion is performed on residual signals in Sl 11. In this case, DCT transform or wavelet transform may be performed. The extent of similarity between transform coefficients obtained by the conversion is determined in S 120. If the resolution of the base layer frame is not different from that of the enhancement layer frame, the similarity between the transform coefficients is determined to be high. If the resolution of the base layer frame is different from that of the enhancement layer frame, the similarity therebetween is determined to be low. This is only one embodiment. In S 130, the actual correlation between the transform coefficients is obtained, and it is determined that the similarity between the transform coefficient is high when the obtained correlation exceeds a predetermined level. When a similarity exists between the transform coefficients, the similarity is eliminated in S 130. In this case, the above-described Hadamard transform may be employed, and the DCT, wavelet transform and Fourier transform are also employed. With respect to operational speed, the Hadamard transform may be faster than the other methods due to the use of addition and subtraction. In the case where the similarity is not high or does not exceed a predetermined level in S 120, S 131 is directly performed without performing S 130. In order to notify a decoding stage of whether the similarity has been eliminated, one bit may be set.
[71] In S131, quantization and entropy processes are performed using the similarity- eliminated transform coefficients and the conversion results obtained in Sl 11. Thereafter, the enhancement layer frame, including macroblocks based on BL prediction, is transferred in S 132.
[72] If the intra-BL prediction mode is not used in S 105, the temporal inter-prediction mode or spatial intra-prediction mode is used in S 108.
[73] FIG. 9 is a flowchart illustrating a process of decoding a video signal according to an exemplary embodiment of the present invention. First, a base layer frame and an enhancement layer frame are extracted from a received bitstream in S201. It is determined whether intra-BL mode was used as a prediction mode when encoding macroblocks constituting the enhancement layer frame in S205. If the intra-BL prediction mode was not used, inverse transform is performed based on temporal inter- prediction mode or spatial intra-prediction mode in S208. If the intra-BL prediction mode was used, the transform coefficients for the sub-blocks of each macroblock are extracted in S210. Thereafter, it is determined whether the similarity between the transform coefficients has been eliminated in S215. This may be determined using a specific bit as described in conjunction with FlG. 8. Furthermore, the determination may be performed without the specific bit in the case where the similarity between the transform coefficients has been eliminated only when the resolution of the base layer frame is identical to that of the enhancement layer frame. If, as a result, the similarity existing between the transform coefficients has been eliminated, the transform coefficients may be calculated using an inverse transform in S220. In this case, the inverse Hadamard transform, which corresponds to the Hadamard transform performed during encoding, is an example of an inverse transform that may be used. If it is determined that the similarity has not been eliminated in S215, the process proceeds to S230. When the transform coefficients are obtained, the residual signals of each macroblock are restored based on the transform coefficients obtained in S230. The restored residual signals are added to the macroblock of the base layer frame and, thereby, the macroblock of a video image is restored in S231.
[74] FlG. 10 is an exemplary diagram illustrating a bit set unit for indicating that the method of the present invention is used when intra-BL prediction is performed according to an exemplary embodiment of the present invention.
[75] Video is composed of video sequences. The video sequence is composed of Groups
Of Pictures (GOPs), each of which is composed of a plurality of frames (pictures). One frame or picture is composed of a plurality of slices, and each of the slices includes a plurality of macroblocks. For each of the macroblocks, one prediction mode may be selected from three prediction modes, such as directional intra-prediction, temporal inter-prediction and intra-BL prediction. Accordingly, when intra-BL prediction, proposed by an exemplary embodiment of the present invention, is performed, intra- coding may be performed on a macroblock basis. However, if one bit is additionally used to determine whether, on a macroblock basis, intra-coding or inter-coding is performed, many bits may be necessary for the overall frames or the overall slices. Accordingly, the number of bits may be set on a macroblock basis, and the number of bits may also be set on a slice basis or on a frame basis. As shown in FlG. 10, the number of bits may be set on a macroblock basis. Furthermore, one bit may be set for all the macroblocks constituting a corresponding slice. In this case, information requirements can be reduced because one bit is assigned to each slice.
Industrial Applicability
[76] In accordance with the present invention, a compression rate may be increased by eliminating the similarity that exists between the pieces of information of the sub- blocks of each macroblock to be encoded using intra-BL prediction. [77] Furthermore, by implementing the present invention, the compression rate may be increased by applying an intra-prediction method when video information is compressed using an intra-BL mode and, therefore, the amount of data transmitted over a network may be reduced.
[78] The exemplary embodiments of the present invention have been disclosed for illustrative purposes, and those skilled in the art will appreciate that various modifications, additions and substitutions are possible, without departing from the scope and spirit of the invention as disclosed in the accompanying claims.

Claims

Claims
[I] A method of encoding video signals in intra-Base-Layer (BL) prediction mode by selectively applying intra-coding in a multilayer-based video encoder, the method comprising:
(a) calculating a difference between an input frame and a base layer frame calculated from the input frame and obtaining residual signals;
(b) converting the residual signals using an intra-coding method; and
(c) generating an enhancement layer frame including the converted residual signals.
[2] The method of claim 1, wherein (a) comprises calculating a difference between a first macroblock constituting part of the input frame, and a second macroblock constituting part of the base layer frame and corresponding to the first macroblock, and obtaining the residual signals. [3] The method of claim 1, wherein (b) comprises converting second sub-blocks of a macroblock by referring to first sub-blocks constituting a macroblock formed of the residual signals.
[4] The method of claim 1, wherein (b) comprises (d) converting transform coefficients of a plurality of sub-blocks constituting a macroblock constructed by the residual signals. [5] The method of claim 4, wherein (d) converts the transform coefficients using a
Hadamard transform. [6] The method of claim 4, further comprising, after (b), (e) setting information indicating that the residual signals have been converted using an intra-coding method.
[7] The method of claim 6, wherein (e) sets the information on a macroblock basis.
[8] The method of claim 6, wherein (e) sets information about all blocks included in each slice. [9] The method of claim 6, wherein (e) sets information about all macroblocks included in each frame. [10] The method of claim 1, further comprising comparing results converted using an intra-coding method with results converted using an inter-coding method.
[II] A method of encoding video signals in intra-Base-Layer (BL) prediction mode by selectively applying intra-coding in a multilayer-based video encoder, the method comprising:
(a) calculating a difference between an input frame and a base layer frame calculated from the input frame and obtaining residual signals;
(b) determining if resolution of the base layer frame is identical to that of the en- hancement layer frame and converting the residual signals using an intra-coding method if the resolution is identical;
(c) generating an enhancement layer frame including the converted residual signals; and
(d) comparing results converted using an intra-coding method with results converted using an inter-coding method.
[12] A method of decoding video signals in intra-BL prediction mode by selectively applying intra-coding in a multilayer-based video decoder, the method comprising:
(a) receiving a base layer frame and an enhancement layer frame;
(b) performing an inverse transform when residual signals of the enhancement layer frame are encoded using an intra-coding method; and
(c) performing restoration by adding the inversely transformed residual signals to image signals of the base layer frame.
[13] The method of claim 12, wherein (b) comprises: restoring transform coefficients existing in the residual signals; and restoring the residual signals using restored transform coefficients.
[14] The method of claim 12, wherein (b) comprises: restoring transform coefficients of a plurality of sub-blocks constituting a macroblock formed of the residual signals; and restoring the sub-blocks using the restored transform coefficients.
[15] The method of claim 14, further comprising (d) restoring the transform coefficients using an inverse Hadamard transform.
[16] The method of claim 12, further comprising, before (b), extracting information indicating that residual signals have been converted using an intra-coding method.
[17] The method of claim 16, wherein the information is information set on a macroblock basis.
[18] The method of claim 16, wherein the information is information set for all macroblocks included in each slice.
[19] The method of claim 16, wherein the information is information set for all macroblocks included in each frame.
[20] A method of decoding video signals in intra-BL prediction mode by selectively applying intra-coding in a multilayer-based video decoder, the method comprising:
(a) receiving a base layer frame and an enhancement layer frame;
(b) determining if resolution of the base layer frame is identical to that of the enhancement layer frame and performing an inverse transform when residual signals of the enhancement layer frame are encoded using an intra-coding method if the resolution is identical; and
(c) performing restoration by adding the inversely transformed residual signals to image signals of the base layer frame.
[21] An encoder comprising : a base layer encoder generating a base layer frame from an input frame; and an enhancement layer encoder generating an enhancement layer frame from the input frame; wherein, at a time of generating a macroblock of the enhancement layer frame, the enhancement layer encoder comprises a conversion unit performing intra- coding on residual signals obtained by calculating a difference between a macroblock of the base layer frame, which corresponds to the macroblock of the enhancement layer frame, and a macroblock of the input frame.
[22] The encoder of claim 21, wherein the conversion unit converts a second sub- block, which is part of a macroblock, by referring to a first sub-block constituting part of a macroblock that is formed of the residual signals.
[23] The encoder of claim 21, wherein the conversion unit converts transform coefficients of sub-blocks constituting the macroblock that is formed of the residual signals.
[24] The encoder of claim 23, wherein the conversion unit converts the transform coefficients using a Hadamard transform.
[25] The encoder of claim 21, wherein the conversion unit sets information indicating that the residual signals have been converted using an intra-coding method.
[26] The encoder of claim 25, wherein the information is information set on a macroblock basis.
[27] The encoder of claim 25, wherein the information is information set for all macroblocks included in each slice.
[28] The encoder of claim 25, wherein the information is information set for all macroblocks included in each frame.
[29] The encoder of claim 21, wherein the conversion unit compares results encoded using an intra-coding method with results encoded using an inter-coding method.
[30] The encoder of claim 29, wherein the conversion unit determines whether resolution of the base layer frame is identical to that of the enhancement layer frame, and performs intra-coding on the residual signals if the resolution is identical.
[31] A decoder comprising: a base layer decoder for restoring a base layer frame; and an enhancement layer decoder for restoring an enhancement layer frame; wherein the enhancement layer decoder performs an inverse transform on residual signals and performs restoration by adding inversely transformed residual signals to image signals of the restored base layer frame, thus restoring the image signals when the residual signals are encoded using an intra-coding method.
[32] The decoder of claim 31 , wherein an inverse conversion unit restores transform coefficients existing in the residual signals, and restores the residual signals using the restored transform coefficients.
[33] The decoder of claim 31, wherein an inverse conversion unit restores transform coefficients of a plurality of sub-blocks constituting a macroblock formed of the residual signals, and restores the sub-blocks using the restored transform coefficients.
[34] The decoder of claim 33, wherein the inverse conversion unit converts the transform coefficients using an inverse Hadamard transform.
[35] The decoder of claim 31, wherein the enhancement layer decoder extracts information indicating that the residual signals have been converted using an intra- coding method.
[36] The decoder of claim 35, wherein the information is information set on a macroblock basis.
[37] The decoder of claim 35, wherein the information is information set for all macroblocks included in each slice.
[38] The decoder of claim 35, wherein the information is information set for all macroblocks included in each frame.
[39] The decoder of claim 31, wherein an inverse conversion unit determines whether resolution of the base layer frame is identical to that of the enhancement layer frame, and restores the residual signals by performing an inverse transform if the resolution is identical.
[40] A computer readable medium having stored therein a program for encoding video signals in intra-Base-Layer (BL) prediction mode by selectively applying intra-coding in a multilayer-based video encoder, said program including computer executable instructions for performing steps comprising:
(a) calculating a difference between an input frame and a base layer frame calculated from the input frame and obtaining residual signals;
(b) converting the residual signals using an intra-coding method; and
(c) generating an enhancement layer frame including the converted residual signals.
[41] A computer readable medium having stored therein a program for encoding video signals in intra-Base-Layer (BL) prediction mode by selectively applying intra-coding in a multilayer-based video encoder, said program including computer executable instructions for performing steps comprising:
(a) calculating a difference between an input frame and a base layer frame calculated from the input frame and obtaining residual signals;
(b) determining if resolution of the base layer frame is identical to that of the enhancement layer frame and converting the residual signals using an intra-coding method if the resolution is identical;
(c) generating an enhancement layer frame including the converted residual signals; and
(d) comparing results converted using an intra-coding method with results converted using an inter-coding method.
[42] A computer readable medium having stored therein a program for decoding video signals in intra-BL prediction mode by selectively applying intra-coding in a multilayer-based video decoder, said program including computer executable instructions for performing steps comprising:
(a) receiving a base layer frame and an enhancement layer frame;
(b) performing an inverse transform when residual signals of the enhancement layer frame are encoded using an intra-coding method; and
(c) performing restoration by adding the inversely transformed residual signals to image signals of the base layer frame.
[43] A computer readable medium having stored therein a program for decoding video signals in intra-BL prediction mode by selectively applying intra-coding in a multilayer-based video decoder, said program including computer executable instructions for performing steps comprising:
(a) receiving a base layer frame and an enhancement layer frame;
(b) determining if resolution of the base layer frame is identical to that of the enhancement layer frame and performing an inverse transform when residual signals of the enhancement layer frame are encoded using an intra-coding method if the resolution is identical; and
(c) performing restoration by adding the inversely transformed residual signals to image signals of the base layer frame.
PCT/KR2006/001338 2005-04-13 2006-04-12 Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding WO2006109985A1 (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US67070005P 2005-04-13 2005-04-13
US60/670,700 2005-04-13
US67254705P 2005-04-19 2005-04-19
US60/672,547 2005-04-19
KR1020050053661A KR100703774B1 (en) 2005-04-13 2005-06-21 Method and apparatus for encoding and decoding video signal using intra baselayer prediction mode applying selectively intra coding
KR10-2005-0053661 2005-06-21

Publications (1)

Publication Number Publication Date
WO2006109985A1 true WO2006109985A1 (en) 2006-10-19

Family

ID=37087226

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2006/001338 WO2006109985A1 (en) 2005-04-13 2006-04-12 Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding

Country Status (1)

Country Link
WO (1) WO2006109985A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011126348A2 (en) * 2010-04-09 2011-10-13 Lg Electronics Inc. Method and apparatus for processing video data
US8908763B2 (en) 2008-06-25 2014-12-09 Qualcomm Incorporated Fragmented reference in temporal compression for video coding
US8948270B2 (en) 2008-08-19 2015-02-03 Qualcomm Incorporated Power and computational load management techniques in video processing
US8948822B2 (en) 2008-04-23 2015-02-03 Qualcomm Incorporated Coordinating power management functions in a multi-media device
US8964828B2 (en) 2008-08-19 2015-02-24 Qualcomm Incorporated Power and computational load management techniques in video processing
FR3012935A1 (en) * 2014-05-20 2015-05-08 Thomson Licensing METHOD AND DEVICE FOR ENCODING VIDEO IMAGES, METHOD AND DEVICE FOR DECODING DATA STREAM, COMPUTER PROGRAM AND CORRESPONDING STORAGE MEDIUM
US10728575B2 (en) 2011-06-07 2020-07-28 Interdigital Vc Holdings, Inc. Method for encoding and/or decoding images on macroblock level using intra-prediction

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5510840A (en) * 1991-12-27 1996-04-23 Sony Corporation Methods and devices for encoding and decoding frame signals and recording medium therefor
US5805293A (en) * 1995-01-30 1998-09-08 Nec Corporation Hadamard transform coding/decoding method and apparatus for image signals
US20020071485A1 (en) * 2000-08-21 2002-06-13 Kerem Caglar Video coding
US20040264791A1 (en) * 1999-12-07 2004-12-30 Intel Corporation Video processing

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5510840A (en) * 1991-12-27 1996-04-23 Sony Corporation Methods and devices for encoding and decoding frame signals and recording medium therefor
US5805293A (en) * 1995-01-30 1998-09-08 Nec Corporation Hadamard transform coding/decoding method and apparatus for image signals
US20040264791A1 (en) * 1999-12-07 2004-12-30 Intel Corporation Video processing
US20020071485A1 (en) * 2000-08-21 2002-06-13 Kerem Caglar Video coding

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8948822B2 (en) 2008-04-23 2015-02-03 Qualcomm Incorporated Coordinating power management functions in a multi-media device
US8908763B2 (en) 2008-06-25 2014-12-09 Qualcomm Incorporated Fragmented reference in temporal compression for video coding
US8964828B2 (en) 2008-08-19 2015-02-24 Qualcomm Incorporated Power and computational load management techniques in video processing
US9565467B2 (en) 2008-08-19 2017-02-07 Qualcomm Incorporated Power and computational load management techniques in video processing
US9462326B2 (en) 2008-08-19 2016-10-04 Qualcomm Incorporated Power and computational load management techniques in video processing
US8948270B2 (en) 2008-08-19 2015-02-03 Qualcomm Incorporated Power and computational load management techniques in video processing
US9426472B2 (en) 2010-04-09 2016-08-23 Lg Electronics Inc. Method and apparatus for processing video data
WO2011126348A2 (en) * 2010-04-09 2011-10-13 Lg Electronics Inc. Method and apparatus for processing video data
US8861594B2 (en) 2010-04-09 2014-10-14 Lg Electronics Inc. Method and apparatus for processing video data
WO2011126348A3 (en) * 2010-04-09 2012-01-26 Lg Electronics Inc. Method and apparatus for processing video data
US9918106B2 (en) 2010-04-09 2018-03-13 Lg Electronics Inc. Method and apparatus for processing video data
US10321156B2 (en) 2010-04-09 2019-06-11 Lg Electronics Inc. Method and apparatus for processing video data
US10841612B2 (en) 2010-04-09 2020-11-17 Lg Electronics Inc. Method and apparatus for processing video data
US11197026B2 (en) 2010-04-09 2021-12-07 Lg Electronics Inc. Method and apparatus for processing video data
US10728575B2 (en) 2011-06-07 2020-07-28 Interdigital Vc Holdings, Inc. Method for encoding and/or decoding images on macroblock level using intra-prediction
US11197022B2 (en) 2011-06-07 2021-12-07 Interdigital Vc Holdings, Inc. Method for encoding and/or decoding images on macroblock level using intra-prediction
FR3012935A1 (en) * 2014-05-20 2015-05-08 Thomson Licensing METHOD AND DEVICE FOR ENCODING VIDEO IMAGES, METHOD AND DEVICE FOR DECODING DATA STREAM, COMPUTER PROGRAM AND CORRESPONDING STORAGE MEDIUM

Similar Documents

Publication Publication Date Title
US20060233250A1 (en) Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding
KR100763181B1 (en) Method and apparatus for improving coding rate by coding prediction information from base layer and enhancement layer
KR100791299B1 (en) Multi-layer based video encoding method and apparatus thereof
US7889793B2 (en) Method and apparatus for effectively compressing motion vectors in video coder based on multi-layer
KR100781525B1 (en) Method and apparatus for encoding and decoding FGS layers using weighting factor
JP5026965B2 (en) Method and apparatus for predecoding and decoding a bitstream including a base layer
US20060120448A1 (en) Method and apparatus for encoding/decoding multi-layer video using DCT upsampling
US20060165302A1 (en) Method of multi-layer based scalable video encoding and decoding and apparatus for the same
KR100763179B1 (en) Method for compressing/Reconstructing motion vector of unsynchronized picture and apparatus thereof
US20060280372A1 (en) Multilayer-based video encoding method, decoding method, video encoder, and video decoder using smoothing prediction
CA2543947A1 (en) Method and apparatus for adaptively selecting context model for entropy coding
WO2006137709A1 (en) Video coding method and apparatus using multi-layer based weighted prediction
JP2006304307A5 (en)
US20060165301A1 (en) Video coding method and apparatus for efficiently predicting unsynchronized frame
US20060165303A1 (en) Video coding method and apparatus for efficiently predicting unsynchronized frame
EP1659797A2 (en) Method and apparatus for compressing motion vectors in video coder based on multi-layer
WO2006109985A1 (en) Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding
US20080013624A1 (en) Method and apparatus for encoding and decoding video signal of fgs layer by reordering transform coefficients
KR100703751B1 (en) Method and apparatus for encoding and decoding referencing virtual area image
EP1889487A1 (en) Multilayer-based video encoding method, decoding method, video encoder, and video decoder using smoothing prediction
WO2006078109A1 (en) Method of multi-layer based scalable video encoding and decoding and apparatus for the same
WO2006078125A1 (en) Video coding method and apparatus for efficiently predicting unsynchronized frame
WO2006104357A1 (en) Method for compressing/decompressing motion vectors of unsynchronized picture and apparatus using the same
WO2006083107A1 (en) Method and apparatus for compressing multi-layered motion vector
KR20130080213A (en) Video encoder, video encoding method and video deconing method

Legal Events

Date Code Title Description
NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: RU

122 Ep: pct application non-entry in european phase

Ref document number: 06757431

Country of ref document: EP

Kind code of ref document: A1