WO2006078125A1 - Procede et appareil de codage video permettant une prediction efficace de trame non synchronisee - Google Patents

Procede et appareil de codage video permettant une prediction efficace de trame non synchronisee Download PDF

Info

Publication number
WO2006078125A1
WO2006078125A1 PCT/KR2006/000207 KR2006000207W WO2006078125A1 WO 2006078125 A1 WO2006078125 A1 WO 2006078125A1 KR 2006000207 W KR2006000207 W KR 2006000207W WO 2006078125 A1 WO2006078125 A1 WO 2006078125A1
Authority
WO
WIPO (PCT)
Prior art keywords
frame
lower layer
motion vector
residual
base layer
Prior art date
Application number
PCT/KR2006/000207
Other languages
English (en)
Inventor
Sang-Chang Cha
Woo-Jin Han
Original Assignee
Samsung Electronics Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020050020812A external-priority patent/KR100703746B1/ko
Application filed by Samsung Electronics Co., Ltd. filed Critical Samsung Electronics Co., Ltd.
Publication of WO2006078125A1 publication Critical patent/WO2006078125A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • H04N19/615Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding using motion compensated temporal filtering [MCTF]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/187Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/63Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]

Definitions

  • the present invention relates, in general, to a video compression method and, more particularly, to a method of efficiently predicting a frame having no corresponding lower layer frame in video frames having a multi-layered structure, and a video coding apparatus using the prediction method.
  • the basic principle of compressing data involves a process of removing redundancy. Spatial redundancy, in which the same color or object is repeated in an image, temporal redundancy, in which an adjacent frame varies little in moving image frames or in which the same sound is repeated in audio data, and psycho- visual redundancy, which takes into consideration the fact that human vision and perceptivity are insensitive to high frequencies, are removed so that data can be compressed.
  • temporal redundancy is removed using temporal filtering based on motion compensation
  • spatial redundancy is removed using a spatial transform.
  • transmission media In order to transmit generated multimedia data after the redundancy has been removed, transmission media are required. The performances of the transmission media differ. Currently used transmission media have various data rates ranging from a data rate like that of an ultra high speed communication network, capable of transmitting data at a data rate of several tens of Mbit/s, to a data rate like that of a mobile communication network, having a data rate of 384 Kbit/s. In this environment, a method of transmitting multimedia data at a data rate suitable for supporting transmission media having various data rates or depending on various transmission environments, that is, a scalable video coding method, may be more suitable for a multimedia environment.
  • Such scalable video coding denotes an encoding method of cutting part of a previously compressed bit stream depending on surrounding conditions, such as a bit rate, an error rate or system resources, thus controlling the resolution, the frame rate and the bit rate of the video.
  • MPEG-4 Moving Picture Experts Group-21
  • part 10 has already achieved the standardization thereof.
  • MPEG-4 Moving Picture Experts Group-21
  • many efforts have been made to realize multi-layered scalability. For example, multiple layers, including a base layer, a first enhancement layer, and a second enhancement layer, are provided, so that respective layers can be constructed to have different frame rates or different resolutions, such as the Quarter Common Intermediate Format (QCIF), CIF and 2CIF.
  • QCIF Quarter Common Intermediate Format
  • FIG. 1 is a diagram showing an example of a scalable video codec using a multi- layered structure.
  • a first layer is defined as a Quarter Common Intermediate Format (QCIF) having a frame rate of 15 Hz
  • a first enhancement layer is defined as a Common Intermediate Format (CIF) having a frame rate of 30 Hz
  • a second enhancement layer is defined as a Standard Definition (SD) having a frame rate of 60 Hz.
  • QCIF Quarter Common Intermediate Format
  • CIF Common Intermediate Format
  • SD Standard Definition
  • 'Scalable Video Model 3.0 of ISO/IEC 21000-13 Scalable Video Coding' (hereinafter referred to as 'SVM 3.0') calls the above method as 'Intra-BL prediction'.
  • SVM 3.0 additionally adopts a method of predicting a current block using a correlation between a current block and a corresponding lower layer block, in addition to inter-prediction and directional intra-prediction, which are used to predict blocks or macroblocks constituting a current frame, in the existing H.264 method.
  • Such a prediction method is designated 'Intra-BL prediction'
  • a mode of performing encoding using Intra-BL prediction is designated an 'intra BL mode'.
  • FIG. 2 is a schematic diagram showing the three prediction methods; case 1 when intra prediction is performed with respect to a certain macroblock 14 of a current frame 11 ; case 2 when inter prediction is performed using a frame 12 placed at a temporal location differing from that of the current frame 11 ; and case 3 when Intra-BL prediction is performed using the texture data of an area 16 of a base layer frame 13 corresponding to the macroblock 14. Disclosure of Invention
  • the frame 40 having no corresponding lower layer frame may exist.
  • Intra-BL prediction cannot be used.
  • the frame 40 is encoded using only information about a corresponding layer (that is, using inter-prediction and intra-prediction) without using information about a lower layer, so that the prediction methods may be somewhat inefficient from the standpoint of encoding performance.
  • an aspect of the present invention provides a video coding method, which can perform Intra-BL prediction with respect to an unsyn- chronized frame.
  • Another aspect of the present invention provides a scheme which can improve the performance of a multi-layered video codec using the video coding method.
  • a multi- layered video encoding method comprising (a) performing motion estimation by using a first frame of two frames of a lower layer temporally closest to an unsynchronized frame of a current layer as a reference frame; (b) obtaining a residual frame between the reference frame and a second frame of the lower layer frames; (c) generating a virtual base layer frame at the same temporal location as that of the unsynchronized frame using a motion vector obtained as a result of the motion estimation, the reference frame, and the residual frame; (d) subtracting the generated virtual base layer frame from the unsynchronized frame to generate a difference; and (e) encoding the difference.
  • a multi-layered video decoding method comprising (a) reconstructing a reference frame from a lower layer bit stream about two frames of a lower layer temporally closest to an unsynchronized frame of a current layer; (b) reconstructing a first residual frame between the two lower layer frames from the lower layer bit stream; (c) generating a virtual base layer frame at the same temporal location as the unsynchronized frame using a motion vector included in the lower layer bit stream, the reconstructed reference frame and the first residual frame; (d) extracting texture data of the unsynchronized frame from a current layer bit stream, and reconstructing a second residual frame for the unsynchronized frame from the texture data; and (e) adding the second residual frame to the virtual base layer frame.
  • a multi-layered video encoder comprising means for performing motion estimation by using a first frame of two frames of a lower layer temporally closest to an unsyn- chronized frame of a current layer as a reference frame; means for obtaining a residual frame between the reference frame and a second frame of the lower layer frames; means for generating a virtual base layer frame at the same temporal location as that of the unsynchronized frame using a motion vector obtained as a result of the motion estimation, the reference frame, and the residual frame; means for subtracting the generated virtual base layer frame from the unsynchronized frame to generate a difference; and means for encoding the difference.
  • a multi-layered video decoder comprising means for reconstructing a reference frame from a lower layer bit stream about two frames of a lower layer temporally closest to an unsynchronized frame of a current layer; means for reconstructing a first residual frame between the two lower layer frames from the lower layer bit stream; means for generating a virtual base layer frame at the same temporal location as the unsynchronized frame using a motion vector included in the lower layer bit stream, the reconstructed reference frame and the first residual frame; means for extracting texture data of the unsynchronized frame from a current layer bit stream, and reconstructing a second residual frame for the unsynchronized frame from the texture data; and means for adding the second residual frame to the virtual base layer frame.
  • FlG. 1 is a diagram showing an example of a scalable video codec using a multi- layered structure
  • FlG. 2 is a schematic diagram showing three conventional prediction methods
  • FlG. 3 is a schematic diagram showing the basic concept of Virtual Base-layer
  • VBP Prediction
  • FlG. 4 is a diagram showing an example of the implementation of VBP using the forward inter-prediction of a base layer
  • FlG. 5 is a diagram showing an example of the implementation of VBP using the backward inter-prediction of a base layer
  • FlG. 6 is a diagram showing the basic concept of a process of considering variation in motion and generating a temporary frame according to aspects of the present invention
  • FIGS. 7 to 11 are diagrams showing a process of generating a temporary frame according to a first exemplary embodiment of the present invention.
  • FIGS. 12 and 13 are diagrams showing a process of generating a virtual base layer frame from the temporary frame according to a second exemplary embodiment of the present invention.
  • FlG. 14 is a diagram showing variation in texture between the corresponding areas of frames
  • FlG. 15 is a diagram showing a concept of the application of texture variation to a temporary frame according to a first exemplary embodiment of the present invention
  • FlG. 16 is a diagram showing a concept of the application of texture variation to a temporary frame according to a second exemplary embodiment of the present invention.
  • FlG. 17 is a block diagram showing the construction of a video encoder according to an exemplary embodiment of the present invention.
  • FlG. 18 is a block diagram showing the construction of a video decoder according to an exemplary embodiment of the present invention.
  • FlG. 19 is a diagram showing the construction of a system environment in which the video encoder of FlG. 17 and the video decoder of FlG. 18 operate;
  • FlG. 20 is a flowchart showing a video encoding process according to an exemplary embodiment of the present invention.
  • FlG. 21 is a flowchart showing a video decoding process according to an exemplary embodiment of the present invention.
  • FlG. 3 is a schematic diagram showing the basic concept of Virtual Base-layer
  • VBP Prediction Prediction
  • a current layer L n has the CIF resolution and a frame rate of 30 Hz
  • a lower layer L n-l has the QCIF resolution and a frame rate of 15 Hz.
  • a current layer frame having no corresponding base layer frame is defined as an 'unsynchronized frame'
  • a current layer frame having a corresponding base layer frame is defined as a 'synchronized frame'. Since an unsynchronized frame does not have a base layer frame, the present invention proposes a method of generating a virtual base layer frame and utilizing the virtual base layer frame for Intra-BL prediction.
  • VBP Concept of VBP according to the present invention can be applied to two layers having different frame rates. Therefore, VBP can also be applied to the case in which a current layer and a lower layer use a hierarchical inter-prediction method, such as Motion Compensated Temporal Filtering (MCTF), as well as the case in which a current layer and a lower layer use a non-hierarchical inter-prediction method (I-B-P coding of an MPEG system codec). Therefore, when a current layer uses MCTF, the concept of the VBP can be applied to the temporal level of the MCTF having a frame rate higher than that of a lower layer.
  • MCTF Motion Compensated Temporal Filtering
  • FIGS. 4 and 5 are diagrams showing examples of a method of implementing VBP according to aspects of the present invention.
  • a base layer frame B may be generated using a motion vector, which may be generated by using one of two frames B and B closest to an unsynchronized frame A in a lower layer as a reference frame, a residual image and the reference frame.
  • FlG. 4 illustrates an example of implementing VBP using the forward inter- prediction of a lower layer.
  • the frame B of a base layer may be predicted through forward inter-prediction by using its previous frame B as a reference frame. That is, after a forward motion vector mv is obtained by using the previous frame B as a reference frame F , the reference frame may be motion-
  • a virtual base layer frame B may be generated using the forward motion vector mv , which may be used for inter-prediction in the base layer, the frame B , which may be used as the reference frame F , and a
  • FlG. 5 illustrates an example of implementing VBP using the backward inter- prediction of a base layer.
  • the frame B of a base layer may be predicted through backward inter-prediction by using the subsequent frame B as a reference frame. That is, after a backward motion vector mv b is obtained by using the subsequent frame B as a reference frame F , the reference frame may be motion- compensated using the obtained motion vector, and the frame B may be inter- predicted using the motion-compensated reference frame.
  • a virtual base layer frame B may be generated using the backward motion vector mv , which may be used for inter-predicti b on in the base layer, the frame B , which may be used as the reference frame F , and a residual image R generated by subtracting B from B .
  • an inter-prediction method referring to a temporally previous frame is designated forward prediction
  • an inter-prediction method referring to a temporally subsequent frame is designated backward prediction.
  • a method of generating a virtual base layer frame may include two processes.
  • the method may include a first process of considering only variation in motion and generating a temporary frame, and a second process of applying variation in texture to the temporary frame and generating a virtual base layer frame.
  • FIG. 6 a concept of a process of considering variation in motion and generating a temporary frame, as shown in FIG. 6, will be described. If there are temporally adjacent frames F and F , and a certain object A moves upwards from a lower position as time elapses in the frames F and F , the location of the object A in a virtual frame F that is disposed between the frames F and F can be predicted to be at the center 0.5u of a path u along which the object A moves between the frames F and F . As described above, a temporary (virtual) frame, generated in consideration of variation in motion between the two frames of the base layer, is based on the above-described concept.
  • FIG. 7 to FIG. 13 are diagrams showing methods of obtaining the temporary frame.
  • FIGS. 7 to 11 are diagrams showing the concept of the generation of a temporary frame according to a first exemplary embodiment of the present invention.
  • one frame 50 to be inter-predicted hereinafter referred to as an 'inter-frame'
  • the frame 50 may be B of FIG. 4, while, in the case of backward prediction, the frame 50 may be B of FIG. 5.
  • each 'partition' means a unit area used for motion estimation, that is, for searching for a motion vector.
  • the partition may have a fixed size (for example, 4x4, 8x8, or 16x16), as shown in FIG. 7, or it may have a variable size as in the case of the H.264 codec.
  • the existing H.264 utilizes Hierarchical Variable Size Block Matching (HVSBM) technology to perform inter-prediction on each macroblock (16x16 size) constituting a single frame.
  • the macroblock can be divided into sub-blocks in a 16x16 mode, 8x16 mode, 16x8 mode, or 8x8 mode.
  • Each of the sub-blocks having an 8x8 size can be further divided into sub-blocks in a 4x8 mode, 8x4 mode or 4x4 mode (if it is not divided, the 8x8 mode is used without change).
  • HVSBM technology a single frame is implemented with a set of macroblocks each having the above-described various combinations of partitions, each partition having a single motion vector.
  • a 'partition' in the present invention means a unit of area to which a motion vector is assigned. It should be apparent that the size and shape of a partition can vary according to the type of codec. However, for explanatory convenience, the inter-frame 50 is assumed to have fixed-size partitions, as shown in FIG. 7. Further, in the present specification, reference numeral 50 denotes the inter- frame of a lower layer (for example, B of FIG. 4, and B of FIG. 5), and reference numeral 60 denotes a reference frame (for example, B of FIG. 4 and B of FIG. 5) used for inter-prediction of the inter-frame.
  • an area in the reference frame 60 corresponding to the partition 1 is an area 1' at a location that is moved away from the location of the partition 1 by the motion vector.
  • a motion compensated frame 70 for the reference frame may be generated by duplicating texture data of the area 1' in the reference frame 60 to the location of the partition 1, as shown in FIG. 9.
  • a temporary frame 80 may be generated in consideration of the principles of generating the motion compensated frame, as shown in FIG. 10. That is, since a motion vector represents a direction in which a certain object moves in a frame, motion compensation may be performed to an extent corresponding to a value obtained by multiplying the motion vector by the ratio of the distance between the reference frame 60 and the location at which the temporary frame 80 is to be generated, to the distance between the reference frame 60 and the inter frame 50 (hereinafter referred to as a 'distance ratio', which is 0.5 in FIGS. 4 and 5).
  • the temporary frame 80 may be filled with texture data in such a way that the area 1' may be duplicated to a location moving away from the area 1' by -r x mv, where r is the distance ratio and mv is the motion vector.
  • r is the distance ratio
  • mv is the motion vector.
  • the temporary frame 80 generated according to the method of the first embodiment may include, for example, an unconnected pixel area and a multi-connected pixel area, as shown in FlG. 11.
  • FlG. 11 since a single-connected pixel area includes only one piece of texture data, there is no problem.
  • a method of processing pixel areas other than the single-connected pixel area may be an issue.
  • a multi-connected pixel may be replaced with a value obtained by averaging a plurality of pieces of texture data at corresponding locations connected thereto.
  • an unconnected pixel may be replaced with a corresponding pixel value in the inter-frame 50, with a corresponding pixel value in the reference frame 60, or with a value obtained by averaging corresponding pixel values in the frames 50 and 60.
  • Intra-BL prediction It is difficult to expect high performance when an unconnected pixel area or a multi-connected pixel area is used for Intra-BL prediction for an unsynchronized frame, compared to the single-connected pixel area.
  • inter-prediction or directional intra-prediction for an unsynchronized frame rather than Intra-BL prediction, will be selected as a prediction method for the above areas from the standpoint of costs, and the performance deterioration is unlikely to occur.
  • Intra-BL prediction will exhibit sufficiently high performance. Accordingly, if the pixel areas are determined to be a single frame unit, the enhancement of performance can be expected when the first exemplary embodiment is applied.
  • FIGS. 12 and 13 are diagrams showing the concept of the generation of a virtual base layer frame according to a second exemplary embodiment of the present invention.
  • the second exemplary embodiment is proposed to improve the problem whereby an unconnected pixel area and a multi-connected pixel area exist in the temporary frame 80 generated in the first exemplary embodiment.
  • the pattern of partitions of a temporary frame 90 in the second exemplary embodiment uses the pattern of partitions of the inter-frame 50 without change.
  • the temporary frame 90 may be generated in such a way that texture data of the area 1" in the reference frame 60 may be copied to the same location as the partition 1, as shown in FlG. 13.
  • the temporary frame 90 is completed. Since the temporary frame 90 generated in this way has the same partition pattern as the inter-frame 50, the temporary frame 90 includes only single-connected pixel areas without including unconnected pixel areas or multi-connected pixel areas.
  • the first and second exemplary embodiments can be independently implemented, but one embodiment, into which the embodiments are combined, can also be considered. That is, the unconnected pixel area of the temporary frame 80 in the first exemplary embodiment may be replaced with the corresponding area of the temporary frame 90 obtained in the second exemplary embodiment. Further, the unconnected pixel area and the multi-connected pixel area of the temporary frame 80 in the first exemplary embodiment may be replaced with the corresponding areas of the temporary frame 90 obtained in the second exemplary embodiment.
  • FIGS. 15 and 16 are diagrams showing a process of applying variation in texture to a temporary frame, generated in consideration of variation in motion, and generating a virtual base layer frame.
  • FlG. 15 is a diagram showing a process of considering variation in texture in the first exemplary embodiment.
  • the partition 1' in the reference frame 60 corresponding to the certain partition 1 in the inter frame 50 may be copied to a location away from the partition 1' by -r x mv in the temporary frame 80.
  • a value obtained by applying a distance ratio to the difference between the texture T of the partition 1 and the texture T of the corresponding partition 1', that is, r x (T -T ) is a texture variation in the virtual base layer frame on the basis of the reference frame.
  • the texture T of a final partition If constituting the virtual base layer frame may be obtained by adding r x (T -T ) to the texture T of the partition 1' copied to the temporary frame 80.
  • the partition If since the location of the partition If is the same as that of the partition 1' copied to the temporary frame 80, the partition If may be generated by replacing the partition 1', copied to the temporary frame 80, with T + r x
  • the texture of a reconstructed image may be used. Accordingly, the texture T of the final partition If may be expressed by Equation [1], where Rec((7) denotes a reconstructed texture image obtained by decoding an encoded texture after a certain texture has been encoded.
  • T -Rec(T ) of Equation [1] is the result obtained by subtracting a reconstructed texture in a reference frame, corresponding to a certain partition in the inter-frame 50, from the certain partition in the inter frame 50, and it denotes a residual image generated by performing inter-prediction on the frame 60.
  • Rec(T - Rec(T )) denotes a resultant image obtained by reconstructing the residual image. Therefore, in the first exemplary embodiment there is no need to execute a separate process for calculating Rec(T -Rec(T )), so reconstruction results for the inter- prediction may be used without change.
  • the temporary frame 80 may be replaced with a virtual base layer frame 85 according to the first exemplary embodiment. That is, the virtual base layer frame 85 may be generated.
  • FIG. 16 is a diagram showing a process of considering variation in texture in the second exemplary embodiment.
  • the partition 1" in the reference frame 60 corresponding to a certain partition 1 may be copied to the same location in the temporary frame 90 as that of the partition 1.
  • a value obtained by applying a distance ratio to the difference between the texture T of the partition 1 and the texture T of the corresponding partition 1", that is, r x (T -T ) is a texture variation in the virtual base layer frame on the basis of the reference frame.
  • the texture T If of a final partition If constituting the virtual base layer frame may be obtained by adding r x (T 1 -T 1" ) to the texture T of the partition 1" copied to the temporary frame 90.
  • the partition If since the location of the partition If is the same as that of the partition 1" copied to the temporary frame 90, the partition If may be generated by replacing the partition 1" with T + r x (T -T ).
  • the texture of a reconstructed image may be used. Accordingly, the texture T of the final partition If may be expressed by the following Equation [2].
  • T - Rec(T ) differs from a residual image used for inter- prediction, unlike the first exemplary embodiment. That is, in inter-prediction, the texture T at a location spaced apart from the location of the partition 1' by a motion vector may be used, but in Equation [2], the texture T at a location spaced apart from the location of the partition 1' by r x mv, may be used. Therefore, a separate process for calculating Rec(T -Rec(T )) is required.
  • the temporary frame 90 may be replaced with a virtual base layer frame 95 according to the second exemplary embodiment. That is, the virtual base layer frame 95 may be generated.
  • a method of generating a temporary frame during a process of considering motion information, and finally generating a virtual base layer frame during a process of considering texture information has been described.
  • a virtual base layer may be generated by copying the texture T of a partition, calculated using Equation [1] or [2], to the corresponding location of the virtual base layer frame.
  • FIG. 17 is a block diagram showing the construction of a video encoder 300 according to an exemplary embodiment of the present invention.
  • FIG. 17 and FIG. 18 an example in which a single base layer and a single enhancement layer are used is described, but those skilled in the art will appreciate that the present invention can be applied to a lower layer and a current layer even though the number of layers used increases.
  • the video encoder 300 may be divided into an enhancement layer encoder 200 and a base layer encoder 100. First, the construction of the base layer encoder 100 is described.
  • a downsampler 110 may downsample input video to a resolution and a frame rate appropriate for a base layer. From the standpoint of resolution, downsampling may be performed using an MPEG downsampler or wavelet downsampler. Further, from the standpoint of frame rate, downsampling may be easily performed using a frame skip method, a frame interpolation method, and others.
  • a motion estimation unit 150 may perform motion estimation on a base layer frame, and obtain a motion vector mv with respect to each partition constituting the base layer frame.
  • Such motion estimation denotes a procedure of finding an area most similar to each partition of a current frame F in a reference frame F , that is, an area c r having a minimum error, and may be performed using various methods, such as a fixed size block matching method or a hierarchical variable size block matching method.
  • the reference frame F may be provided by a frame buffer 180.
  • the base layer encoder 100 of FIG. 17 may adopt a scheme in which a reconstructed frame is used as a reference frame, that is, a closed loop encoding scheme.
  • the encoding scheme is not limited to the closed loop encoding method; the base layer encoder 100 may adopt an open loop encoding scheme in which an original base layer frame, provided by the downsampler 110, may be used as a reference frame.
  • a motion compensation unit 160 may perform motion compensation on the reference frame using the obtained motion vector.
  • a subtractor 115 may obtain the difference between the current frame F of the base layer and the motion compensated reference frame, thus generating a residual frame.
  • a transform unit 120 may perform a spatial transform on the generated residual frame and generate a transform coefficient.
  • a spatial transform method a Discrete Cosine Transform (DCT), or a wavelet transform may be used.
  • DCT Discrete Cosine Transform
  • the transform coefficient denotes a DCT coefficient
  • a wavelet transform the transform coefficient denotes a wavelet coefficient.
  • a quantization unit 130 may quantize the transform coefficient generated by the transform unit 120. Quantization refers to an operation of dividing the DCT coefficient, expressed as an arbitrary real number, into predetermined intervals based on a quantization table, representing the intervals as discrete values, and matching the discrete values to corresponding indices. A quantization result value obtained in this way is designated a quantized coefficient.
  • An entropy encoding unit 140 may perform non-lossy encoding on the quantized coefficient generated by the quantization unit 130, and the motion vector generated by the motion estimation unit 150, thus generating a base layer bit stream.
  • various non-lossy encoding methods such as Huffman coding, arithmetic coding or variable length coding may be used.
  • an inverse quantization unit 171 may perform inverse quantization on the quantized coefficient output from the quantization unit 130. Such an inverse quantization process is the inverse of the quantization process, and is a process of reconstructing indices, which are generated during the quantization process, through the use of the quantization table used in the quantization process.
  • An inverse transform unit 172 may perform an inverse spatial transform on an inverse quantization result value. This inverse spatial transform is the inverse of the transform process executed by the transform unit 120. In detail, an inverse DCT, an inverse wavelet transform, and others can be used.
  • An adder 125 may add the output value of the motion compensation unit 160 to the output value of the inverse transform unit 172, reconstruct the current frame, and provide the reconstructed current frame to the frame buffer 180.
  • the frame buffer 180 may temporarily store the reconstructed frame and provide the reconstructed frame as a reference frame to perform the inter-prediction on another base layer frame.
  • a virtual frame generation unit 190 may generate a virtual base layer frame to perform Intra-BL prediction on an unsynchronized frame of an enhancement layer. That is, the virtual frame generation unit 190 may generate the virtual base layer frame using a motion vector mv obtained between the two base layer frames temporally closest to the unsynchronized frame, a reference frame F of the two frames, and a residual frame R between the two frames.
  • the virtual frame generation unit 190 may receive the motion vector mv from the motion estimation unit 150, the reference frame F r from the frame buffer 180, and the reconstructed residual frame R from the inverse transform unit 172. [89]
  • the virtual frame generation unit 190 is described in relation to FIG. 15 according to a first exemplary embodiment.
  • the virtual frame generation unit 190 may read the texture T of the partition 1' corresponding to the partition 1 of the current frame 60 from the reference frame F r .
  • the virtual frame generation unit 190 may add a texture T - T to the texture T at the location corresponding to the partition 1 in the residual frame R, and copy the addition results to a location in the virtual base layer frame away from the location of the partition 1' by -r x mv.
  • the virtual frame generation unit 190 is described in relation to FIG. 16 according to a second exemplary embodiment.
  • the virtual frame generation unit 190 may read the texture T of the partition 1" from the reference frame F at a location away from the partition 1 by r x mv.
  • the residual frame according to the second exemplary embodiment differs from an existing residual frame R, as shown in FIG. 17, so that the residual frame may be generated through a separate process.
  • the process itself may use the construction presented in FIG. 17 without change. That is, the motion compensation unit 160 may perform motion compensation based on a new motion vector (r x mv) obtained by multiplying r by the motion vector mv, and not based on the motion vector mv.
  • a subsequent process may be the same as that of FIG.
  • the second exemplary embodiment may be realized by utilizing a residual frame R', which has been reconstructed by and output from the inverse transform unit 172.
  • the texture T - T is added to the texture T at the location corresponding to the partition 1 in the residual frame R', the addition results may be duplicated to the location of the partition 1' in the virtual base layer frame.
  • the virtual base layer frame generated by the virtual frame generation unit 190 may be selectively provided to the enhancement layer encoder 200 through an upsampler 195.
  • the upsampler 195 may upsample the virtual base layer frame at the resolution of the enhancement layer when the resolutions of the enhancement layer and the base layer are different. When the resolutions of the base layer and the enhancement layer are the same, the upsampling process may be omitted.
  • an input frame is an unsynchronized frame
  • the input frame and a virtual base layer frame, provided by the base layer encoder 100 may be input to a subtracter 210.
  • the subtractor 210 may subtract the virtual base layer frame from the input frame and generate a residual frame.
  • the residual frame may be converted into an enhancement layer bit stream through a transform unit 220, a quantization unit 230, and an entropy encoding unit 240, and the enhancement layer bit stream may be output.
  • the function and operation of the transform unit 220, the quantization unit 230 and the entropy encoding unit 240 are similar to those of the transform unit 120, the quantization unit 130 and the entropy encoding unit 140, respectively, and therefore detailed descriptions thereof are omitted.
  • the enhancement layer encoder 200 of FlG. 17 is described with respect to the encoding of an unsynchronized frame among input frames. Those skilled in the art will appreciate that if the input frame is a synchronized frame, three conventional prediction methods may be selectively used to perform encoding, as described above with reference to FlG. 2.
  • FlG. 18 is a block diagram showing the construction of a video decoder 600 according to an exemplary embodiment of the present invention.
  • the video decoder 600 may be divided into an enhancement layer decoder 500 and a base layer decoder 400. First, the construction of the base layer decoder 400 is described.
  • An entropy decoding unit 410 may perform non-lossy decoding on a base layer bit stream, thus extracting texture data of a base layer frame and motion data (a motion vector, partition information, a reference frame number, and others).
  • An inverse quantization unit 420 may perform inverse quantization on the texture data.
  • This inverse quantization process corresponds to the inverse of the quantization process executed by the video encoder 300, and is a process of reconstructing indices, which are generated during the quantization process, through the use of the quantization table used in the quantization process.
  • An inverse transform unit 430 may perform an inverse spatial transform on the inverse quantization result value, thus reconstructing a residual frame.
  • This inverse spatial transform is the inverse of the transform process executed by the transform unit 120 of the video encoder 300.
  • an inverse DCT, an inverse wavelet transform, and others may be used as the inverse transform.
  • an entropy decoding unit 410 may provide motion data, which may include a motion vector mv, both to a motion compensation unit 460 and to a virtual frame generation unit 470.
  • the motion compensation unit 460 may perform motion compensation on a previously reconstructed video frame provided by a frame buffer 450, that is, a reference frame, using the motion data provided by the entropy decoding unit 410, thus generating a motion compensated frame.
  • An adder 515 may add a residual frame reconstructed by the inverse transform unit
  • the reconstructed video frame may be temporarily stored in the frame buffer 450, and may be provided to the motion compensation unit 460 or the virtual frame generation unit 470 as a reference frame in order to reconstruct other subsequent frames.
  • the virtual frame generation unit 470 may generate a virtual base layer frame to perform Intra-BL prediction on an unsynchronized frame of an enhancement layer. That is, the virtual frame generation unit 470 may generate the virtual base layer frame using a motion vector mv obtained between the two base layer frames temporally closest to the unsynchronized frame, a reference frame F r of the two frames, and a residual frame R between the two frames. For this operation, the virtual frame generation unit 470 may receive the motion vector mv from the entropy decoding unit 410, the reference frame F r from the frame buffer 450, and the reconstructed residual frame R from the inverse transform unit 430.
  • a process of generating the virtual base layer frame using the motion vector, the reference frame and the residual frame is similar to that of the virtual frame generation unit 190 of the video encoder 300, and therefore detailed descriptions thereof are omitted.
  • a residual frame R' may be obtained by performing motion compensation on the reference frame of two reconstructed base layer frames using r x mv and subtracting the motion compensated reference frame from a current frame.
  • the virtual base layer frame generated by the virtual frame generation unit 470 may be selectively provided to the enhancement layer decoder 500 through an upsampler 480.
  • the upsampler 480 may upsample the virtual base layer frame at the resolution of the enhancement layer when the resolutions of the enhancement layer and the base layer are different. When the resolutions of the base layer and the en- hancement layer are the same, the upsampling process may be omitted.
  • the entropy decoding unit 510 may perform non-lossy decoding on the input bit stream and extract the texture data of the unsynchronized frame.
  • the extracted texture data may be reconstructed as a residual frame through an inverse quantization unit 520 and an inverse transform unit 530.
  • the function and operation of the inverse quantization unit 520 and the inverse transform unit 530 are similar to those of the inverse quantization unit 420 and the inverse transform unit 430.
  • An adder 515 may add the reconstructed residual frame to the virtual base layer frame provided by the base layer decoder 400, thus reconstructing the unsynchronized frame.
  • the enhancement layer decoder 500 of FlG. 18 has been described based on the decoding of an unsynchronized frame among input frames.
  • Those skilled in the art will appreciate that if an enhancement layer bit stream is related to a synchronized frame, reconstruction methods according to three conventional prediction methods may be selectively used, as described above with reference to FlG. 2.
  • FlG. 19 is a diagram showing the construction of a system environment, in which the video encoder 300 or video decoder 600 is operated, according to an exemplary embodiment of the present invention.
  • a system may be, for example, but not limited to, a TV, set-top box, desk-top computer, lap-top computer, palmtop computer, Personal Digital Assistant (PDA), video or image storage device (for example, a Video Cassette Recorder (VCR) or Digital Video Recorder (DVR)
  • the system may be a combination of the devices, or a specific device including another device as a part of the specific device.
  • the system may include at least one video source 910, at least one input/output device 920, a processor 940, memory 950, and a display device 930.
  • the video source 910 may be, for example, but not limited to, a TV receiver, a
  • the video source 910 may include a connection to one or more networks for receiving video from a server using the Internet, a Wide Area Network (WAN), a Local Area Network (LAN), a terrestrial broadcast system, a cable network, a satellite communication network, a wireless network, or a telephone network.
  • the video source may be a combination of the networks, or a specific network including another network as a part of the specific network.
  • the input/output device 920, the processor 940, and the memory 950 may communicate with each other through a communication medium 960.
  • the com- munication medium 960 may be a communication bus, a communication network, or one or more internal connection circuits.
  • the input video data received from the source 910 may be processed by the processor 940 using one or more software programs stored in the memory 950, or it may be executed by the processor 940 to generate output video to be output to the display device 930.
  • the software program stored in the memory 950 may include a multi- layered video codec for performing the method of the present invention.
  • the codec may be stored in the memory 950, be read from a storage medium, such as Compact Disc-Read Only Memory (CD-ROM) or a floppy disc, or be downloaded from a server through various networks.
  • the codec may be replaced with a hardware circuit implementing the software, or with a combination of software and hardware circuits.
  • FlG. 20 is a flowchart showing a video encoding process according to an exemplary embodiment of the present invention.
  • the motion estimation unit 150 may perform motion estimation by using a first frame (of two lower layer frames) that is temporally closest to the unsynchronized frame of the current layer as a reference frame in S30.
  • the motion estimation may be performed in fixed size blocks or hierarchical variable size blocks.
  • the reference frame may be a temporally previous frame of the two lower layer frames as shown in FlG. 4, or a temporally subsequent frame as shown in FlG. 5.
  • S35 may include encoding the first frame using the subtractor 115, the transform unit 120 and the quantization unit 130, and then decoding the encoded first frame using the inverse quantization unit 171, the inverse transform unit 172 and the adder 125; the motion compensation unit 160 performing motion compensation on the decoded first frame using a motion vector; the subtractor 115 subtracting the motion compensated first frame from the second frame to generate a difference; and encoding the difference using the transform unit 120 and the quantization unit 130 and then decoding the encoding results using the inverse quantization unit 171 and the inverse transform unit 172.
  • a residual frame R according to the first exemplary embodiment may be obtained.
  • S35 may include encoding the first frame using the subtractor 115, the transform unit 120 and the quantization unit 130, and then decoding the encoded first frame using the inverse quantization unit 171, the inverse transform unit 172 and the adder 125; the motion compensation unit 160 performing motion compensation on the decoded first frame using a result vector obtained by multiplying the motion vector by a distance ratio r; the subtractor 115 subtracting the motion compensated first frame from the second frame to generate a difference; and encoding the difference using the transform unit 120 and the quantization unit 130 and then decoding the encoding results using the inverse quantization unit 171 and the inverse transform unit 172.
  • a residual frame R' according to the second exemplary embodiment may be obtained.
  • the virtual frame generation unit 190 may generate a virtual base layer frame at the same temporal location as the unsynchronized frame using the motion vector, which has been obtained as a result of motion estimation, the reference frame and the residual frame in S40.
  • S40 may include reading texture data of an area 1' spaced apart from the location of a partition 1, to which the motion vector is assigned, from the reference frame; adding results, obtained by multiplying texture data T -T corresponding to the location of the partition 1 in the residual frame R by a distance ratio r, to the read texture data T ; and copying the addition results T to a location away from the area by a value, obtained by multiplying the motion vector by the distance ratio, in a direction opposite the motion vector.
  • S40 may include reading texture data T of an area 1" spaced apart from the location of the partition 1, to which the motion vector is assigned, by a value obtained by multiplying the motion vector mv by the distance ratio r, from the reference frame; adding results, obtained by multiplying the texture data T - T corresponding to the location of the partition 1' in the residual frame R' by the distance ratio r, to the read texture data T ; and copying the addition results to the location of the partition.
  • the upsampler 195 may upsample the generated virtual base layer frame at the resolution of the current layer in S50.
  • the subtractor 210 of the enhancement layer encoder 200 may subtract the upsampled virtual base layer frame from the unsynchronized frame to generate a difference in S60. Further, the transform unit 220, the quantization unit 230 and the entropy encoding unit 240 may encode the difference in S70.
  • the upsampler 190 may upsample a base layer frame at a location corresponding to the current synchronized frame at the resolution of the current layer in S80.
  • the subtractor 210 may subtract the upsampled base layer frame from the synchronized frame to generate a difference in S90.
  • the difference may also be encoded through the transform unit 220, the quantization unit 230 and the entropy encoding unit 240 in S70.
  • FlG. 21 is a flowchart showing a video decoding process according to an exemplary embodiment of the present invention.
  • the base layer decoder 400 may reconstruct a reference frame from a lower layer bit stream corresponding to the two lower layer frames that are temporally closest to the unsynchronized frame of the current layer in S 130. Further, a first residual frame between the two lower layer frames may be reconstructed from the lower layer bit stream in S 135.
  • S 135 may include the entropy decoding unit 410 extracting the texture data of an inter-frame of the two lower layer frames from the lower layer bit stream, the inverse quantization unit 420 performing inverse quantization on the extracted texture data, and the inverse transform unit 430 performing an inverse spatial transform on the inverse quantization results.
  • a first residual frame R according to the first exemplary embodiment may be reconstructed.
  • S 135 may include the entropy decoding unit 410 extracting the texture data of an inter-frame of the two lower layer frames from the lower layer bit stream, the inverse quantization unit 420 performing inverse quantization on the extracted texture data, and the inverse transform unit 430 performing an inverse spatial transform on the inverse quantization results, the motion compensation unit 460 performing motion compensation on the reconstructed reference frame using the motion vector, the adder 415 adding the inverse spatial transform results to the motion compensated reference frame, thus reconstructing an inter frame, the motion compensation unit 460 performing motion compensation on the reconstructed reference frame using a result vector obtained by multiplying the motion vector by a distance ratio, and the subtractor (not shown in FlG. 18) subtracting the motion compensated reference frame from the reconstructed inter-frame.
  • a first residual frame R' may be reconstructed.
  • the virtual frame generation unit 470 may generate a virtual base layer frame at the same temporal location as the unsynchronized frame using the motion vector included in the lower layer bit stream, the reconstructed reference frame and the first residual frame in S 140.
  • the first and second embodiments may be applied to S 140, similar to the video encoding process. This operation is described above with reference to S40 of FlG. 20, and therefore a detailed description thereof is omitted.
  • the upsampler 480 may upsample the generated virtual base layer frame at the resolution of the current layer in S 145.
  • the entropy decoding unit 510 of the enhancement layer decoder 500 may extract the texture data of the unsynchronized frame from a current layer bit stream in S150.
  • the inverse quantization unit 520 and the inverse transform unit 530 may reconstruct a second residual frame from the texture data in S 160.
  • the adder 515 may add the second residual frame to the virtual base layer frame in S 170. As a result, the unsynchronized frame may be reconstructed.
  • the base layer decoder 400 may reconstruct a base layer frame at a location corresponding to the synchronized frame in S 180.
  • the upsampler 480 may upsample the reconstructed base layer frame in S 190.
  • the entropy decoding unit 510 may extract the texture data of the synchronized frame from the current layer bit stream in S200.
  • the inverse quantization unit 520 and the inverse transform unit 530 may reconstruct a third residual frame from the texture data in S210.
  • the adder 515 may add the third residual frame to the upsampled base layer frame in S220. As a result, the synchronized frame may be reconstructed.
  • Intra-BL prediction can be performed with respect to an unsynchronized frame using a virtual base layer frame. Further, according to exemplary embodiments of the present invention, there is an advantage in that video compression efficiency can be improved by using a more efficient prediction method.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention porte sur un procédé qui permet la prédiction efficace d'une trame ne comportant pas de trame de couche inférieure correspondante dans des trames vidéo possédant une structure multicouche, et sur un appareil de codage vidéo faisant appel au procédé de prédiction précité. Selon le procédé de codage vidéo de l'invention, on procède à une estimation de mouvement en utilisant comme trame de référence une première trame parmi deux trames d'une couche inférieure temporairement la plus proche d'une trame non synchronisée d'une couche actuelle. On obtient une trame résiduelle entre la trame de référence et une seconde trame parmi les trames de couche inférieure. On produit une trame de couche de base virtuelle au même emplacement temporel que celui de la trame non synchronisée, au moyen d'un vecteur de mouvement obtenu à l'aide de l'estimation de mouvement, de la trame de référence et de la trame résiduelle. On soustrait la trame de couche de base virtuelle produite de la trame non synchronisée afin d'obtenir une différence et l'on code la différence.
PCT/KR2006/000207 2005-01-21 2006-01-19 Procede et appareil de codage video permettant une prediction efficace de trame non synchronisee WO2006078125A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US64501005P 2005-01-21 2005-01-21
US60/645,010 2005-01-21
KR10-2005-0020812 2005-03-12
KR1020050020812A KR100703746B1 (ko) 2005-01-21 2005-03-12 비동기 프레임을 효율적으로 예측하는 비디오 코딩 방법 및장치

Publications (1)

Publication Number Publication Date
WO2006078125A1 true WO2006078125A1 (fr) 2006-07-27

Family

ID=36692468

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2006/000207 WO2006078125A1 (fr) 2005-01-21 2006-01-19 Procede et appareil de codage video permettant une prediction efficace de trame non synchronisee

Country Status (1)

Country Link
WO (1) WO2006078125A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106878746A (zh) * 2010-08-17 2017-06-20 M&K控股株式会社 用于解码运动图片的设备

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001128043A (ja) * 1999-10-25 2001-05-11 Mitsubishi Electric Corp 画像撮影・蓄積装置及び画像表示・印字装置
WO2001047277A1 (fr) * 1999-12-20 2001-06-28 Sarnoff Corporation Codage vidéo échelonnable
JP2001298738A (ja) * 2000-04-14 2001-10-26 Nippon Telegr & Teleph Corp <Ntt> 画像符号化方法、画像符号化装置及び画像符号化プログラムを記憶した媒体
US6349284B1 (en) * 1997-11-20 2002-02-19 Samsung Sdi Co., Ltd. Scalable audio encoding/decoding method and apparatus

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6349284B1 (en) * 1997-11-20 2002-02-19 Samsung Sdi Co., Ltd. Scalable audio encoding/decoding method and apparatus
JP2001128043A (ja) * 1999-10-25 2001-05-11 Mitsubishi Electric Corp 画像撮影・蓄積装置及び画像表示・印字装置
WO2001047277A1 (fr) * 1999-12-20 2001-06-28 Sarnoff Corporation Codage vidéo échelonnable
JP2001298738A (ja) * 2000-04-14 2001-10-26 Nippon Telegr & Teleph Corp <Ntt> 画像符号化方法、画像符号化装置及び画像符号化プログラムを記憶した媒体

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106878746A (zh) * 2010-08-17 2017-06-20 M&K控股株式会社 用于解码运动图片的设备
CN106878747A (zh) * 2010-08-17 2017-06-20 M&K控股株式会社 用于编码运动图片的设备
CN107087198A (zh) * 2010-08-17 2017-08-22 M&K控股株式会社 用于编码运动图片的设备
CN107087197A (zh) * 2010-08-17 2017-08-22 M&K控股株式会社 用于编码运动图片的设备
CN106878747B (zh) * 2010-08-17 2019-08-30 M&K控股株式会社 用于编码运动图片的设备
CN107087197B (zh) * 2010-08-17 2020-01-17 M&K控股株式会社 用于编码运动图片的设备
CN107087198B (zh) * 2010-08-17 2020-01-17 M&K控股株式会社 用于编码运动图片的设备
CN106878746B (zh) * 2010-08-17 2020-03-27 M&K控股株式会社 用于解码运动图片的设备

Similar Documents

Publication Publication Date Title
KR100714696B1 (ko) 다계층 기반의 가중 예측을 이용한 비디오 코딩 방법 및장치
KR100703745B1 (ko) 비동기 프레임을 효율적으로 예측하는 비디오 코딩 방법 및장치
US20060165301A1 (en) Video coding method and apparatus for efficiently predicting unsynchronized frame
WO2006078115A1 (fr) Procede et appareil de codage video pour la prediction efficace de trames non synchronisees
US8249159B2 (en) Scalable video coding with grid motion estimation and compensation
KR100679031B1 (ko) 다 계층 기반의 비디오 인코딩 방법, 디코딩 방법 및 상기방법을 이용한 장치
KR101033548B1 (ko) 스무딩 예측을 이용한 다계층 기반의 비디오 인코딩 방법,디코딩 방법, 비디오 인코더 및 비디오 디코더
KR100703760B1 (ko) 시간적 레벨간 모션 벡터 예측을 이용한 비디오인코딩/디코딩 방법 및 장치
KR100888963B1 (ko) 영상 신호의 스케일러블 인코딩 및 디코딩 방법
KR100703788B1 (ko) 스무딩 예측을 이용한 다계층 기반의 비디오 인코딩 방법,디코딩 방법, 비디오 인코더 및 비디오 디코더
KR100763182B1 (ko) 다계층 기반의 가중 예측을 이용한 비디오 코딩 방법 및장치
KR100763179B1 (ko) 비동기 픽쳐의 모션 벡터를 압축/복원하는 방법 및 그방법을 이용한 장치
KR100704626B1 (ko) 다 계층 기반의 모션 벡터를 압축하는 방법 및 장치
US20060120448A1 (en) Method and apparatus for encoding/decoding multi-layer video using DCT upsampling
KR20060135992A (ko) 다계층 기반의 가중 예측을 이용한 비디오 코딩 방법 및장치
US20060233250A1 (en) Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding
KR20070075232A (ko) 모션 역변환을 사용하여 모션 예측을 수행하는 방법 및장치
WO2006109985A1 (fr) Procede et appareil destines a coder / decoder des signaux video dans un mode de prediction interne de couche de base par l&#39;application selective d&#39;intracodage
KR100703751B1 (ko) 가상 영역의 영상을 참조하여 인코딩 및 디코딩 하는 방법및 장치
EP1817918A1 (fr) Procede et appareil de codage/decodage de video multicouche par sur-echantillonnage dct
WO2006078125A1 (fr) Procede et appareil de codage video permettant une prediction efficace de trame non synchronisee
WO2006078109A1 (fr) Procede et dispositif d&#39;encodage et decodage video echelonnable multicouche
EP1847129A1 (fr) Procede et dispositif pour comprimer un vecteur de mouvement multicouche
WO2006104357A1 (fr) Procede pour la compression/decompression des vecteurs de mouvement d&#39;une image non synchronisee et appareil utilisant ce procede

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 06702731

Country of ref document: EP

Kind code of ref document: A1