CN1461563A - Method of encoding sequence of frames - Google Patents

Method of encoding sequence of frames Download PDF

Info

Publication number
CN1461563A
CN1461563A CN02801121A CN02801121A CN1461563A CN 1461563 A CN1461563 A CN 1461563A CN 02801121 A CN02801121 A CN 02801121A CN 02801121 A CN02801121 A CN 02801121A CN 1461563 A CN1461563 A CN 1461563A
Authority
CN
China
Prior art keywords
mrow
msub
temporal
mover
frames
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN02801121A
Other languages
Chinese (zh)
Other versions
CN1251509C (en
Inventor
B·佩斯奎特-波佩斯库
V·博特雷奥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Koninklijke Philips NV
Original Assignee
Koninklijke Philips Electronics NV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics NV filed Critical Koninklijke Philips Electronics NV
Publication of CN1461563A publication Critical patent/CN1461563A/en
Application granted granted Critical
Publication of CN1251509C publication Critical patent/CN1251509C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • H04N19/615Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding using motion compensated temporal filtering [MCTF]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/553Motion estimation dealing with occlusions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/583Motion compensation with overlapping blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/63Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention relates to a method for encoding a sequence of frames by means of a three-dimensional (3D) subband decomposition involving a filtering step applied to the frames considered as a 3D volume. The decomposition is applied to successive groups of frames after motion estimation and compensation which lead to double-connected pixels corresponding to uncovered regions. A minimization criterion is provided for choosing in the current frame, for the temporal filtering of a double-connected pixel of the reference frame, the pixel that leads to a minimum energy of the detail subband. The temporal part of the 3D decomposition may be a temporal wavelet transform that includes a temporal filtering based on the lifting scheme, including an overlapped motion compensation using, in the prediction step of said scheme, an average of pixels from adjacent windows in the reference frame.

Description

Method for encoding a sequence of frames
The present invention relates generally to video compression and, more particularly, to a method for encoding a sequence of frames using a three-dimensional (3D) subband decomposition, said encoding involving spatial and temporal filtering steps applied to 3D spatio-temporal data corresponding to said sequence of frames considered as a 3D quantity, said decomposition being applied to successive groups of frames only after block-based motion estimation and compensation operations have been carried out on the frames of each group in the low-frequency temporal subbands at each level of temporal decomposition; and the compensation operation results in double-connected pixels being generated in the filtered pixels, which correspond to regions of the current frame that become uncovered reference frames during motion.
Recent developments in multimedia applications and the need to deliver compressed bit streams over heterogeneous networks have led to the view of scalability as a very important feature of video encoders. Three-dimensional (3D) wavelet decomposition provides natural spatial resolution and frame rate scalability. Embedded coding algorithms, such as the 3D SPIHT algorithm described in "embedded wavelet video coder using set partitioning in three-dimensional hierarchical trees" (Proceedings of datacompresssin Conference, Snowbird, USA, March 25-27, 1997, pp.251-260), for example, b. The global coding/decoding scheme applied to a sequence of frames may comprise a temporal multi-resolution analysis along the direction of motion between successive frames in order to take into account the large displacements (more precisely, a Haar filter is applied to the motion compensated frames at each resolution level), followed by a spatial multi-resolution analysis of the resulting temporal subbands. The encoding method is an improved 3D SPIHT followed by a context arithmetic encoder. By this approach scalability is achieved in both temporal and spatial resolution as well as in quality.
However, despite some work on scalable coding algorithms and possible improvements in spatial wavelet transforms, little success has been achieved in exploiting the particularities of the temporal wavelet transforms used in 3D decomposition. The temporal wavelet transform itself is nonlinear due to the motion estimation step. The so-called lifting scheme described in "decompose wavelet transform into lifting (lifting) steps" of, for example, i.daubechies and w.sweldens (Journal of Fourier Analysis applications, vol.4, N ° 3, 1998, pp.247-269) provides a flexible framework to construct wavelet transforms (the importance of which in processing one-dimensional signals and providing appropriate decomposition for image coding or feature extraction has been demonstrated). The advantage of this approach lies in both complexity ("in-place" computation) and additional functionality: not only can each linear wavelet decomposition be expressed in this form, it also allows the introduction of non-linear operators in the prediction update step.
An important issue related to time domain multi-resolution analysis is the selection of the temporal filtering length: long filtering more advantageously exploits the temporal correlation that exists between successive frames. However, in "three-dimensional subband coding with motion compensation" (IEEE Transactions on image Processing, vol.3, No.5, 1994, pp.559-571) in j.r.ohm it is shown that the long filtering does blur the motion and increases the buffer memory requirements (corresponding to the filter length) and the reconstruction delay on the decoder side, which may be undesirable in real-time applications. Furthermore, performing temporal analysis with longer filtering does not significantly improve coding efficiency. Therefore, haar filtering is more suitable for temporal filtering. In addition, when haar filtering is used for temporal decomposition, motion estimation and motion compensation (ME/MC) are performed every two frames in the input sequence since temporal down-sampling is performed with 2. By repeating this process for the time-approximated subbands at several decomposition levels, the total number of ME/MC operations is approximately the same as for the prediction scheme.
However, motion compensated temporal filtering causes the problem of double-connected or non-connected pixels, i.e. pixels that are filtered twice or not filtered at all. A solution is proposed in "motion compensated 3D sub-band coding of video" (IEEE trans on imaging processing, vol.8, No.2, pp.155-164, feb.1999) by s.j.choi and j.w.woods, which can calculate high-pass coefficients at the same positions as the pixels in the current frame and interpret the co-existing pixels in the motion compensated reference frame as low-pass coefficients. In the case of a multi-connected pixel, the current frame may be scanned from top to bottom and from left to right and the first pixel in the current frame to which it points may be considered for the calculation of the low-pass coefficient. Nevertheless, it does not appear to be the optimal strategy.
At a given level of resolution, one can use H [ m, n ]]Representing pixels in temporal high-frequency subbands, by L m, n]Representing pixels in a low frequency subband, and using (u)m,n,vm,n) Representing the two components of the motion vector associated with pixel (m, n). If fractional pixel motion estimation is allowed, the integer part of the motion vector will be composed of (u)m,n, vm,n) And (4) showing. Thus, in the Choi method, the equations used to compute the high-pass and low-pass subbands connecting the pixels are as follows: ( 1 ) - - - H [ m , n ] = ( B [ m , n ] - A ~ [ m - u m , n , n - v m , n ] ) / 2 <math> <mrow> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>L</mi> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>=</mo> <mrow> <mo>(</mo> <mover> <mi>B</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>+</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>+</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>+</mo> <mi>A</mi> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>)</mo> </mrow> <mo>/</mo> <msqrt> <mn>2</mn> </msqrt> </mrow> </math> wherein,
Figure A0280112100053
representing the interpolated value of field X. For unconnected pixels, the high frequency component is obtained as before, while the low frequency value is simply the scaled value of the reference pixel: <math> <mrow> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>L</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>=</mo> <mi>A</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>&CenterDot;</mo> <msqrt> <mn>2</mn> </msqrt> </mrow> </math>
as a special case of the haar transform lifting scheme, it is possible to write the temporal low-pass filtering in the motion direction as: <math> <mrow> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>L</mi> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>=</mo> <mover> <mi>H</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>+</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>+</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>+</mo> <msqrt> <mn>2</mn> </msqrt> <mi>A</mi> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> </mrow> </math>
taken together (1), this equation can derive the form of the nonlinear operators P (prediction) and U (update) used in temporal haar lifting. It can be seen that P is the motion compensation operator (C), which in the case of fractional-pixel motion estimation is followed by interpolation (I). At the same time, U can be determined as the motion compensation operator, which uses the same motion vector as in P, but of opposite sign, followed by an interpolation. As a result, these operations will be expressed as:
P{.}=I{C{.}}
and U { } ═ I { C { } { (C { } { } { (C { }
And position (m-u) of (p, q)m,n,n- vm,n)。
With these symbologies, the temporal analysis of connected pixels can be written as: H [ m , n ] 1 2 ( B [ m , n ] - I { C { A [ m , n ] } } ) <math> <mrow> <mi>L</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>=</mo> <mi>I</mi> <mo>{</mo> <mover> <mi>C</mi> <mo>&OverBar;</mo> </mover> <mo>{</mo> <mi>H</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>}</mo> <mo>}</mo> <mo>+</mo> <mi>A</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>&CenterDot;</mo> <msqrt> <mn>2</mn> </msqrt> </mrow> </math>
for the synthesis part, there are: <math> <mrow> <mrow> <mo>(</mo> <mn>7</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>A</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>=</mo> <mfrac> <mn>1</mn> <msqrt> <mn>2</mn> </msqrt> </mfrac> <mrow> <mo>(</mo> <mi>L</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>-</mo> <mi>I</mi> <mo>{</mo> <mover> <mi>C</mi> <mo>&OverBar;</mo> </mover> <mo>{</mo> <mi>H</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>}</mo> <mo>}</mo> <mo>)</mo> </mrow> </mrow> </math>
for connected pixels (unconnected pixels in the reference frame are derived directly from equation (3) above), and: ( 8 ) - - - B [ m , n ] = 2 H [ m , n ] + I { C { A [ m , n ] } } }
it is therefore a first object of the present invention to adapt the update operator using such lifting formulas in order to improve the processing of uncovered areas.
To this end, the invention relates to a coding method as defined in the introductory part of the description, which is further characterized in that: -a decision step following said identification step of double-connected pixels, which provides, on the basis of a minimization criterion, said minimization criterion for choosing, in the current frame, the pixel with the lowest energy of the fine sub-band (detail sub-band) as the double-connected pixel of the reference frame for temporal filtering, the energy considered being the average energy of said double-connected pixel in the high-frequency temporal sub-band, and calculating said average energy by means of a low-pass filtering operation.
The proposed technical solution is used to correlate a pixel at a given position in the reference frame a, obtained by minimizing the energy of the fine subband, with its closest value in the current frame B.
In an advantageous specific embodiment said minimization criterion is provided for combining a second criterion for performing a minimization of the motion vector criterion related to candidate connected pixels with said energy criterion.
In this case, the minimization criterion according to the motion vector criterion may employ a threshold value that is appropriate for the value of the criterion, so as to not consider optimization of a motion vector that is too large.
It is another object of the invention to employ lifting formulas to modify the update operator in order to reduce blocking artifacts related to block-based motion compensation.
To this end, the invention relates to a coding method as defined above, in which the temporal part of the 3D decomposition is a temporal wavelet transform comprising a temporal filtering based on a so-called lifting scheme comprising, in succession, a separation step, a prediction step and an update step at each decomposition level; the method is further characterized by: the temporal filtering includes an overlapping motion compensation in the prediction step using an average of pixels of neighboring windows in the reference frame to reduce blocking artifacts caused by block-based motion estimation.
The problem of unconnected and doubly connected pixels is strongly related to the area of coverage removed by the moving object. In practice, two objects are considered that correspond to a common portion in a frame at time T and become separated at time T + Δ T. In this case, two regions in the current frame will correspond to the same region in the reference frame (or previous frame) through motion compensation. For one of these objects, the area will be an uncovered area. This region will act as a dual connection in the reference frame. For the method described in the cited document "motion compensated 3D sub-band coding of video", which relates the first block encountered during motion compensation to these double-connected pixels, the method proposes to optimize this choice by applying certain criteria based on the lifting scheme. The main structural attributes used here are: it is possible to use all the information derived from the prediction step (high frequency subbands) and the reason information in the low frequency subbands for the update step (calculation of the temporal low frequency subbands).
The basic idea corresponding to the proposed first criterion relates to the energy of the minutiae sub-bands of two moving objects. If at time T the first object is in the foreground, the uncovered area in the second object will yield a higher fine factor energy. The second criterion is the condition for motion field regularization: if several pixels are connected to the same pixel in the reference frame, the pixel with the smallest displacement will be selected for filtering.
These ideas can be formalized in the non-linear lifting framework described above. In the case of multiple connected pixels in the reference frame, if one of these pixels at position (p, q) and position (m) in the current frame are considered1,n1) And (m)2,n2) Two pixels found by the motion estimation algorithm, the two pixels having two corresponding motion vectors ( U m 1 , n 1 , v m 1 , n 1 ) And ( U m 2 , n 2 , v m 2 , n 2 ) , then there are: <math> <mrow> <mrow> <mo>(</mo> <mn>9</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <msub> <mi>m</mi> <mn>1</mn> </msub> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <msub> <mi>m</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>n</mi> <mn>1</mn> </msub> </mrow> </msub> <mo>=</mo> <msub> <mi>m</mi> <mn>2</mn> </msub> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <msub> <mi>m</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi>n</mi> <mn>2</mn> </msub> </mrow> </msub> <mo>=</mo> <mi>p</mi> <mo>,</mo> <msub> <mi>n</mi> <mn>1</mn> </msub> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <msub> <mi>m</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>n</mi> <mn>1</mn> </msub> </mrow> </msub> <mo>=</mo> <msub> <mi>n</mi> <mn>2</mn> </msub> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <msub> <mi>m</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi>n</mi> <mn>2</mn> </msub> </mrow> </msub> <mo>=</mo> <mi>q</mi> </mrow> </math>
with this result, two pixels (m) can be addressed1,n1) And (m)2,n2) Are written as equations (5) and (6), two different values H m are obtained in the fine sub-band1,n1]And H [ m ]2,n2]. Thus, the values in the approximation subband can be computed with either of these two values (it can be noted that both values allow a full reconstruction). In fact, P(p,q)Representing the set of all pixels (m, n) in the current frame connected to the pixel (p, q) in the reference frame, it can be noted that the operator f is guaranteed to be completeFull reconstruction property, such that: <math> <mrow> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>L</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>=</mo> <mi>f</mi> <mrow> <mo>(</mo> <mover> <mi>H</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>+</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>+</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>,</mo> <mrow> <mo>(</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <msub> <mi>P</mi> <mrow> <mo>(</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>)</mo> </mrow> </msub> <mo>)</mo> </mrow> <mo>+</mo> <mi>A</mi> <mo>[</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>]</mo> <mo>&CenterDot;</mo> <msqrt> <mn>2</mn> </msqrt> </mrow> </math>
one criterion for selecting operator f is to minimize the energy of the fine sub-bands, thereby minimizing A [ p, q ]]Associated with its "closest" value in frame B. This means that the pixel (m) is a pixel0,n0) Low-pass filtering is performed so that: <math> <mrow> <mrow> <mo>(</mo> <mn>11</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mo>|</mo> <mi>I</mi> <mo>{</mo> <mover> <mi>C</mi> <mo>&OverBar;</mo> </mover> <mo>{</mo> <mi>H</mi> <mo>[</mo> <msub> <mi>m</mi> <mn>0</mn> </msub> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mn>0</mn> <mo>,</mo> <mi>n</mi> <mn>0</mn> </mrow> </msub> <mo>,</mo> <msub> <mi>n</mi> <mn>0</mn> </msub> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mn>0</mn> <mo>,</mo> <mi>n</mi> <mn>0</mn> </mrow> </msub> <mo>]</mo> <mo>}</mo> <mo>}</mo> <mo>|</mo> <mo>=</mo> <mi>mi</mi> <msub> <mi>n</mi> <mrow> <mrow> <mo>(</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <msub> <mi>P</mi> <mrow> <mo>(</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>)</mo> </mrow> </msub> </mrow> </msub> <mo>|</mo> <mi>I</mi> <mo>{</mo> <mover> <mi>C</mi> <mo>&OverBar;</mo> </mover> <mo>{</mo> <mi>H</mi> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mover> <mi>u</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mi>n</mi> <mo>-</mo> <msub> <mover> <mi>v</mi> <mo>&OverBar;</mo> </mover> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>}</mo> <mo>}</mo> <mo>|</mo> </mrow> </math>
for Choi's algorithm, in the proposed algorithm, it is not necessary to send the classification map to the decoder (e.g., which pixels are connected and which pixels are not). Since the decoder follows a symmetrical process with the encoder, the same classification map will be used theoretically to derive the decision on the energy of the high frequency coefficients.
However, the decision made based on the value of a single pixel is not robust enough. In particular, in the previous example, two values in the high frequency subband may not be quantized in the same quantization step due to the fractional quantization strategy used in the SPIHT algorithm. Therefore, an erroneous decision will be made. The above decision can then be made more robust by comparing the average energy of the Displacement Frame Difference (DFD) around the considered pixel. <math> <mrow> <mrow> <mo>(</mo> <mn>12</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>e</mi> <msup> <mrow> <mo>(</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>=</mo> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>k</mi> <mo>,</mo> <mi>l</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <msub> <mi>S</mi> <mrow> <mo>(</mo> <mi>p</mi> <mo>,</mo> <mi>q</mi> <mo>)</mo> </mrow> </msub> </mrow> </munder> <msup> <mrow> <mo>(</mo> <mi>H</mi> <mrow> <mo>(</mo> <mi>p</mi> <mo>-</mo> <mi>k</mi> <mo>,</mo> <mi>q</mi> <mo>-</mo> <mi>l</mi> <mo>)</mo> </mrow> <mi>u</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>,</mo> <mi>l</mi> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> </math>
Where S (p, q) is a neighborhood around the pixel (p, q), and u (k, l) corresponds to the weighting factor of each pixel in the neighborhood S (p, q) according to its distance to the center point. For example, u (k, l) ═ α may be selected(|k|+[l])Where α > 0 is a forgetting factor.
The second term in the minimization criterion is the motion vector specification, | dm,n‖=(u2 m,n+v2 m,n)1/2. The regularization criterion may be expressed as J (p, q) ═ e (p, q)2+λ‖ dm,n|, λ is the regularization parameter. If the motion vector is too large, its value is not very reliable and therefore columns can be chosen that do not take this into account for optimization. This gives the following criteria: (13) j (p, q) ═ e (p, q)2+λ‖ dm,nII, if dm,n‖≤s(14)J(p,q)=e(p,q)2+ λ s if | dm,n‖>s
Where s is an empirically determined threshold. Therefore, some loss occurs on motion vectors larger than the threshold.
However, block-based motion estimation algorithms suffer from blocking artifacts. If the spatial transform applied to the generated motion-compensated frames is wavelet analysis, these artifacts will result in undesirably large wavelet coefficients, which in turn results in a reduction in coding efficiency. Hence, the improvement derivable from the above non-linear lifting formula relates to the possibility of introducing overlapping motion compensation in the temporal filtering algorithm, thereby reducing blocking artifacts. This operation involves using the pixel averaging of adjacent windows in the reference frame in the prediction step.
For example, if one pixel overlap is consideredThen the high-pass filtering of the pixels belonging to the first line (respectively last line) of a block is in this case: <math> <mrow> <mrow> <mo>(</mo> <mn>9</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>H</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>=</mo> <mfrac> <mn>1</mn> <msqrt> <mn>2</mn> </msqrt> </mfrac> <mo>[</mo> <mi>B</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>-</mo> <mrow> <mo>(</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&beta;</mi> <mo>)</mo> </mrow> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <msub> <mrow> <mi>n</mi> <mo>-</mo> <mi>v</mi> </mrow> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>+</mo> <mi>&beta;</mi> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <mi>l</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>-</mo> <mi>l</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>-</mo> <mi>l</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>)</mo> </mrow> <mo>]</mo> </mrow> </math> corresponding to <math> <mrow> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>H</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>=</mo> <mfrac> <mn>1</mn> <msqrt> <mn>2</mn> </msqrt> </mfrac> <mo>[</mo> <mi>B</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>-</mo> <mrow> <mo>(</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&beta;</mi> <mo>)</mo> </mrow> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>+</mo> <mi>&beta;</mi> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>+</mo> <mi>l</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>+</mo> <mi>l</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>+</mo> <mi>l</mi> <mo>-</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>+</mo> <mi>l</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>)</mo> </mrow> <mo>]</mo> </mrow> </math>
Wherein beta is a constant, beta is more than 0 and less than 1.
A similar process is applied to the first (respectively last) column of each block according to: <math> <mrow> <mrow> <mo>(</mo> <mn>11</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>H</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>=</mo> <mfrac> <mn>1</mn> <msqrt> <mn>2</mn> </msqrt> </mfrac> <mo>[</mo> <mi>B</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>-</mo> <mrow> <mo>(</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&beta;</mi> <mo>)</mo> </mrow> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <msub> <mrow> <mi>n</mi> <mo>-</mo> <mi>v</mi> </mrow> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>+</mo> <mi>&beta;</mi> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>-</mo> <mi>l</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <mi>l</mi> <mo>-</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>-</mo> <mi>l</mi> </mrow> </msub> <mo>)</mo> </mrow> <mo>]</mo> </mrow> </math> accordingly, the <math> <mrow> <mrow> <mo>(</mo> <mn>12</mn> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mi>H</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>=</mo> <mfrac> <mn>1</mn> <msqrt> <mn>2</mn> </msqrt> </mfrac> <mo>[</mo> <mi>B</mi> <mo>[</mo> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>]</mo> <mo>-</mo> <mrow> <mo>(</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&beta;</mi> <mo>)</mo> </mrow> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>-</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> </mrow> </msub> <mo>]</mo> <mo>+</mo> <mi>&beta;</mi> <mover> <mi>A</mi> <mo>~</mo> </mover> <mo>[</mo> <mi>m</mi> <mo>-</mo> <msub> <mi>u</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>+</mo> <mi>l</mi> </mrow> </msub> <mo>,</mo> <mi>n</mi> <mo>+</mo> <mi>l</mi> <mo>-</mo> <msub> <mi>v</mi> <mrow> <mi>m</mi> <mo>,</mo> <mi>n</mi> <mo>+</mo> <mi>l</mi> </mrow> </msub> <mo>]</mo> <mo>)</mo> </mrow> <mo>]</mo> </mrow> </math>
The constant β is determined experimentally. In the case of the present invention, the constant β of 0.2 yields the best results.

Claims (4)

1. A method for encoding a sequence of frames with a three-dimensional (3D) subband decomposition, said encoding involving spatial and temporal filtering steps applied to 3D spatio-temporal data corresponding to said sequence of frames considered as a 3D quantity, said decomposition being applied to successive groups of frames only after block-based motion estimation and compensation operations have been carried out on the frames of each group in the low-frequency temporal subbands at each level of temporal decomposition; and such compensation operation results in double-connected pixels in the filtered pixels, said double-connected pixels corresponding to regions of the current frame that become uncovered reference frames during motion, said method being further characterized by: said double-connected pixel identification step is followed by a decision step based on a minimization criterion provided for picking in the current frame said pixel producing the lowest energy of said fine sub-band as the double-connected pixel of said reference frame to be temporally filtered, the energy considered being the average energy of said double-connected pixels in said high-frequency temporal sub-band, and calculating said average energy by a low-pass filtering operation.
2. A coding method according to claim 1, characterized in that said minimization criterion is provided for combining a second criterion for performing minimization of a motion vector specification relating to candidate connected pixels with said energy criterion.
3. A coding method according to claim 2, characterized in that the minimization criterion according to the motion vector specification uses a threshold value that is adapted to the value of the specification so as not to take into account optimization too large motion vectors.
4. The coding method according to claim 1, characterized in that said temporal part of said 3D decomposition is a temporal wavelet transform comprising a temporal filtering based on a so-called lifting scheme comprising a separation step, a prediction step and an update step carried out in sequence at each decomposition level; the method is further characterized by: the temporal filtering includes overlapping motion compensation using an average of neighboring windowed pixels in the reference frame in the predicting step to reduce blocking artifacts caused by the block-based motion estimation.
CNB02801121XA 2001-04-10 2002-04-02 Method of encoding sequence of frames Expired - Fee Related CN1251509C (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
EP01400919.5 2001-04-10
EP01400919 2001-04-10
EP01400941.9 2001-04-12
EP01400941 2001-04-12

Publications (2)

Publication Number Publication Date
CN1461563A true CN1461563A (en) 2003-12-10
CN1251509C CN1251509C (en) 2006-04-12

Family

ID=26077231

Family Applications (1)

Application Number Title Priority Date Filing Date
CNB02801121XA Expired - Fee Related CN1251509C (en) 2001-04-10 2002-04-02 Method of encoding sequence of frames

Country Status (6)

Country Link
US (1) US6983021B2 (en)
EP (1) EP1380174A1 (en)
JP (1) JP2004523994A (en)
KR (1) KR20030014705A (en)
CN (1) CN1251509C (en)
WO (1) WO2002085026A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008089705A1 (en) * 2007-01-23 2008-07-31 Huawei Technologies Co., Ltd. Encoding and decoding method andapparatus
CN101552919B (en) * 2008-04-02 2011-03-02 联咏科技股份有限公司 Operational method with functions of executing overlapped filtration and center conversion and device thereof

Families Citing this family (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7133846B1 (en) * 1995-02-13 2006-11-07 Intertrust Technologies Corp. Digital certificate support system, methods and techniques for secure electronic commerce transaction and rights management
AUPR222500A0 (en) 2000-12-21 2001-01-25 Unisearch Limited Method for efficient scalable compression of video
JP3863775B2 (en) * 2001-12-25 2006-12-27 株式会社九州エレクトロニクスシステム Image information compression method, image information compression apparatus, and image information compression program
AUPS291002A0 (en) * 2002-06-12 2002-07-04 Unisearch Limited Method and apparatus for scalable compression of video
DE60317670T2 (en) * 2003-09-09 2008-10-30 Mitsubishi Denki K.K. Method and apparatus for 3D subband video coding
KR20060076309A (en) * 2003-09-23 2006-07-04 코닌클리케 필립스 일렉트로닉스 엔.브이. Video de-noising algorithm using inband motion-compensated temporal filtering
JP2007509516A (en) * 2003-10-10 2007-04-12 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ 3D video scalable video encoding method
WO2005081531A1 (en) * 2004-01-20 2005-09-01 Koninklijke Philips Electronics N.V. Three-dimensional video scalable video encoding method
US8442108B2 (en) * 2004-07-12 2013-05-14 Microsoft Corporation Adaptive updates in motion-compensated temporal filtering
US8340177B2 (en) * 2004-07-12 2012-12-25 Microsoft Corporation Embedded base layer codec for 3D sub-band coding
CN101019436B (en) * 2004-07-13 2013-05-08 法国电讯公司 Method and device for encoding a video a video image sequence
US8374238B2 (en) 2004-07-13 2013-02-12 Microsoft Corporation Spatial scalability in 3D sub-band decoding of SDMCTF-encoded video
US8279918B2 (en) 2005-07-15 2012-10-02 Utc Fire & Security Americas Corporation, Inc. Method and apparatus for motion compensated temporal filtering using residual signal clipping
US8483277B2 (en) 2005-07-15 2013-07-09 Utc Fire & Security Americas Corporation, Inc. Method and apparatus for motion compensated temporal filtering using split update process
US7956930B2 (en) 2006-01-06 2011-06-07 Microsoft Corporation Resampling and picture resizing operations for multi-resolution video coding and decoding
WO2008091206A1 (en) 2007-01-26 2008-07-31 Telefonaktiebolaget Lm Ericsson (Publ) Motion estimation for uncovered frame regions
US8038442B2 (en) 2007-04-23 2011-10-18 Nobel Biocare Services Ag Dental implant and dental component connection
US8953673B2 (en) 2008-02-29 2015-02-10 Microsoft Corporation Scalable video coding and decoding with sample bit depth and chroma high-pass residual layers
US8711948B2 (en) 2008-03-21 2014-04-29 Microsoft Corporation Motion-compensated prediction of inter-layer residuals
WO2010017166A2 (en) 2008-08-04 2010-02-11 Dolby Laboratories Licensing Corporation Overlapped block disparity estimation and compensation architecture
US9571856B2 (en) 2008-08-25 2017-02-14 Microsoft Technology Licensing, Llc Conversion operations in scalable video encoding and decoding
US8213503B2 (en) 2008-09-05 2012-07-03 Microsoft Corporation Skip modes for inter-layer residual video coding and decoding
CN102946539B (en) * 2012-11-21 2015-07-15 西安电子科技大学 Method for estimating motion among video image frames based on compressive sensing

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5301020A (en) * 1991-11-01 1994-04-05 Aware, Inc. Method and apparatus for coding motion pictures utilizing motion compensation
WO2001078402A1 (en) * 2000-04-11 2001-10-18 Koninklijke Philips Electronics N.V. Video encoding and decoding method

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008089705A1 (en) * 2007-01-23 2008-07-31 Huawei Technologies Co., Ltd. Encoding and decoding method andapparatus
CN101552919B (en) * 2008-04-02 2011-03-02 联咏科技股份有限公司 Operational method with functions of executing overlapped filtration and center conversion and device thereof

Also Published As

Publication number Publication date
WO2002085026A1 (en) 2002-10-24
US6983021B2 (en) 2006-01-03
US20030026339A1 (en) 2003-02-06
JP2004523994A (en) 2004-08-05
CN1251509C (en) 2006-04-12
EP1380174A1 (en) 2004-01-14
KR20030014705A (en) 2003-02-19

Similar Documents

Publication Publication Date Title
CN1251509C (en) Method of encoding sequence of frames
Pesquet-Popescu et al. Three-dimensional lifting schemes for motion compensated video compression
Andreopoulos et al. In-band motion compensated temporal filtering
KR101203338B1 (en) Adaptive updates in motion-compensated temporal filtering
KR100308627B1 (en) Low bit rate encoder using overlapping block motion compensation and zerotree wavelet coding
CN1205818C (en) Video encoding and decoding method
US6519284B1 (en) Encoding method for the compression of a video sequence
US6954500B2 (en) Video coding method using a block matching process
Heising et al. Wavelet-based very low bit-rate video coding using image warping and overlapped block motion compensation
US20080095235A1 (en) Method and apparatus for intra-frame spatial scalable video coding
US20070217513A1 (en) Method for coding video data of a sequence of pictures
JPH09121345A (en) Image encoding device and image decoding device
Andreopoulos et al. Complete-to-overcomplete discrete wavelet transforms for scalable video coding with MCTF
KR20000059799A (en) Device and method for motion compensation coding using wavelet coding
US7242717B2 (en) Wavelet domain motion compensation system
US6944225B2 (en) Resolution-scalable video compression
KR100901178B1 (en) Moving picture encoding method and device, and moving picture decoding method and device
JP4438797B2 (en) Method for video image coding taking into account the parts related to the components of the motion vector
KR20050085385A (en) Video coding method and device
JP4690371B2 (en) Scalable image encoding method, decoding method, encoding device, decoding device, program thereof and recording medium thereof
US20060012680A1 (en) Drift-free video encoding and decoding method, and corresponding devices
Mehrseresht et al. Spatially continuous orientation adaptive discrete packet wavelet decomposition for image compression
Marpe et al. Video coding using a bilinear image warping motion model and wavelet-based residual coding
WO2008051755A2 (en) Method and apparatus for intra-frame spatial scalable video coding
Padmanabhan et al. An Efficient Video Compression Encoder Based on Wavelet Lifting Scheme in LSK

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
C19 Lapse of patent right due to non-payment of the annual fee
CF01 Termination of patent right due to non-payment of annual fee