WO1995021503A1 - Video motion estimation - Google Patents

Video motion estimation Download PDF

Info

Publication number
WO1995021503A1
WO1995021503A1 PCT/GB1995/000207 GB9500207W WO9521503A1 WO 1995021503 A1 WO1995021503 A1 WO 1995021503A1 GB 9500207 W GB9500207 W GB 9500207W WO 9521503 A1 WO9521503 A1 WO 9521503A1
Authority
WO
WIPO (PCT)
Prior art keywords
motion
motion vectors
field
vectors
sequence
Prior art date
Application number
PCT/GB1995/000207
Other languages
French (fr)
Inventor
Peter Sarginson
Original Assignee
British Broadcasting Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by British Broadcasting Corporation filed Critical British Broadcasting Corporation
Priority to JP7520453A priority Critical patent/JPH09509799A/en
Priority to GB9616294A priority patent/GB2301255B/en
Priority to EP95907091A priority patent/EP0742983A1/en
Publication of WO1995021503A1 publication Critical patent/WO1995021503A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/144Movement detection
    • H04N5/145Movement estimation

Definitions

  • This invention relates to video motion estimation of the type which can be used in a video coder for reducing the bit rate of a digital video signal such as may be required for storage on digital storage media or for broadcast.
  • MPEG-l is now an international standard and is specified in ISO/IEC document IS 11172 parts 1,2 and 3.
  • ISO/IEC document IS 11172 parts 1,2 and 3 A description of the MPEG-l video compression algorithm can be found in Communications of the ACM April 1991, Vol.34, No. 4 MPEG-2 is a substantial extension of MPEG-l and is due for publication as an international standard in 1995.
  • Intra pictures are coded without reference to other pictures. They serve as access points to the coded video sequence where decoding can begin.
  • Predicted pictures are coded with reference to a motion compensated prediction derived from a previous I or P picture. Coding of P pictures is more efficient than for I pictures.
  • Bi-directionally predicted pictures are coded with reference to a forward motion compensated prediction from a previous I or P picture and a backward motion compensated prediction from a future I or P picture and provide the greatest degree of compression.
  • FIG. 1 A typical sequence of I, P and B pictures is illustrated in Figure 1.
  • frame II is an I picture
  • frames P4 and P7 are P pictures
  • frames B2, B3, B5, B6, B8 and B9 are B pictures.
  • Frame P4 is coded with reference to a forward prediction derived from frame II.
  • Frame P7 is coded with reference to a forward prediction derived from frame P4.
  • Frames B2 and B3 are coded with reference to a forward prediction derived from frame 11 and a backward prediction derived from frame P4.
  • Frames B5 and B6 are coded with reference to a forward prediction derived from frame P4 and a backward prediction derived from frame P7.
  • Frames B8 and B9 are coded with reference to a forward prediction derived from frame P7 and a backward prediction derived from frame 110.
  • a video coder When a video coder is implemented in hardware, a large part of its circuitry is devoted to the measurement of motion vectors in order to generate the required motion compensated predictions.
  • B pictures two simultaneous motion measurements are required in order to generate both the forward and backward predictions.
  • MPEG-2 algorithm twice as many motion measurements are required as both 'frame' and 'field' versions of each motion vector are needed.
  • FIG. 1 shows the sequence of I, B, and P frames described above,-
  • Figure 2 shows a conventional motion estimator arrangement required to produce vectors defining the sequence of frames shown in Figure 1; and Figure 3 shows a motion estimator embodying the inventio
  • the sequence may be split into "triples" of ⁇ BBP ⁇ or (less often) ⁇ BBl ⁇ , i.e. two interpolated and one reference picture.
  • a coder's motion estimator For each ⁇ BBP ⁇ triple, a coder's motion estimator must generate five motion vectors. For example, in the case of the triple ⁇ B12, B13, P14 ⁇ the five required motion vectors are:
  • a coder's motion estimator For each ⁇ BBl ⁇ triple a coder's motion estimator must generate four motion vectors. For example, in the case of the triple ⁇ B15, B16, 117 ⁇ the four required motion vectors are:
  • the architecture which is generally proposed to generate these motion vectors is based around two motion estimators as shown in Figure 2. This comprises a forward motion estimator 2 receiving a previous reference picture and the current picture and producing forward motion vectors and a backward motion vector estimator 4 receiving a future reference picture and the current picture and producing backward motion vectors . These are the vectors referred to above.
  • Table 1 shown below describes the passage of the sequence, B12, B13 , P14, B15, B16, 117, through the motion estimators. It will be seen that the sequence of the pictures is re-ordered to P14, B12, B13 , 117, B15, B16, prior to input to the motion estimators in accordance with the MPEG algorithm.
  • Each column of the table represents an instant in time. For example, the first column means, "when P14 is on the current picture input, Pll is applied to the previous reference picture input and the forward motion vector output generates the motion vector Pll to P14".
  • the table demonstrates how the two motion estimators of Figure 1 can together generate the required five motion vectors for each ⁇ BBP ⁇ triple and the required four motion vectors for each ⁇ BBl ⁇ triple.
  • the required motion vectors for the ⁇ BBP ⁇ triple are (Pll to P14) , (Pll to B12) , (Pll to B13), (P14 to B12) and (P14 to B13) .
  • the motion estimator has generated the first three of these.
  • the missing two motion vectors may be calculated using additional vector processing as follows :
  • the required motion vectors for the (BBl) triple are (P14 to B15), (P14 to B16) , (117 to B15) and (117 to B16) .
  • the motion estimator has generated the first two of these as well as the vector (P17 to 117) .
  • the missing two motions vectors may be calculated as follows :
  • the forward motion estimator may identify a number of motion vectors that will yield equally good predictions even though they need not represent the actual motion of the object. If the forward motion vector is incorrect, the derived backward motion vector may yield a poor prediction. (It should be noted that if a prediction resulting from a backward motion vector is poor, the coder will simply choose some other prediction mode for that particular macroblock and the erroneous backward motion vector will go unnoticed) .
  • the additional vector processing required to generate all forward and backward vectors using a single vector estimator could be implemented in hardware using vector adders and/or subtractors and some suitable storage and switching arrangement to supply the vectors to the adders at the correct times. Alternatively the processing could be carried out in software.
  • the current MPEG-2 test model details a further improved bit-rate reduction standard comprising field and frame motion vectors.
  • the idea is as follows.
  • MPEG-2 defines a macroblock as covering an area of picture 16 pixels by 16 frame lines, 8 lines originating from the odd field and the other 8 lines originating from the even field.
  • the job of a coder's prediction generator stage is to generate "predictions" of the macroblocks comprising the current frame. If the image is moving then these predictions will generally be realised by motion compensating (i.e. assigning one or more motion vectors to each macroblock of a reference frame.
  • the motion estimators (of the type referred to above) generate both 'field' and 'frame' motion vectors in order to derive two predictions of each macroblock in the current frame:- one based on the field motion vectors, the other based on the frame motion vector.
  • the coder chooses the prediction which most closely matches the actual macroblock of the current frame.
  • a macroblock is treated as a single 16 x 16 block of a frame.
  • a single "frame" motion vector is assigned to describe how a prediction of the macroblock may be generated by motion compensating a 16 x 16 pixel area of a reference frame.
  • two "field" motion vectors are used to independently generate predictions of the odd field and even field components of a macroblock. i.e. One field motion vector is used to predict the 8 odd lines of the macroblock from the odd field of a reference picture and the other field motion vector is used to predict the 8 even field lines from the even field of a reference picture.
  • the motion vectors are used to generate two predictions of a macroblock from the current picture, one generated using the frame motion vector, the other generated using the two field motion vectors. It is then the job of the coder's "decision module" to decide which is the better prediction.
  • a frame motion vector may be derived instead by averaging the two field motion vectors. If the two field motion vectors match, then the frame motion vector will have the same value resulting in similar frame and field based predictions, however, if the two field motion vectors are different, the frame motion vector is likely to be incorrect and the field based prediction is likely to be best.
  • the vector generation is explained in more detail below.
  • Motion occurring in the odd and even fields of a previous frame is independently measured and used to generate predictions of the add and even fields of a future frame, i.e. the odd field of the future frame is predicted from the odd field of the previous frame and correspondingly with the even field. This one prediction of a single future frame comprising odd and even fields has been generated.
  • the vectors measured in the odd field are averaged with those in the even field to give a set of frame motion vectors.
  • a further prediction of the future frame may now be generated by motion compensating the previous frame with the set of frame motion vectors.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Analysis (AREA)

Abstract

A method is provided for estimating motion vectors providing image displacements between reference images (IN) in a sequence of digital video images and intermediate image (BN) falling between reference images in sequence. Motion vectors are derived to estimate a prediction (PN) of one reference image (IN) from another in the sequence. Motion vectors are then derived for generating predictions of each intermediate image (BN) between two reference images (IN) in the sequence from one of the two reference images (IN) or their predictions (PN). From these estimated vectors further vectors are derived for deriving a prediction of each intermediate image (BN) from the other one of the two reference images or their predictions which it falls between.

Description

VIDEO MOTION ESTIMATION
This invention relates to video motion estimation of the type which can be used in a video coder for reducing the bit rate of a digital video signal such as may be required for storage on digital storage media or for broadcast.
Various proposals for reducing the bit rate of a digital video signal have been made such as the MPEG-l and MPEG-2 algorithms from the Motion Pictures Expert Group (MPEG) . MPEG-l is now an international standard and is specified in ISO/IEC document IS 11172 parts 1,2 and 3. A description of the MPEG-l video compression algorithm can be found in Communications of the ACM April 1991, Vol.34, No. 4 MPEG-2 is a substantial extension of MPEG-l and is due for publication as an international standard in 1995.
Both MPEG algorithms define the pictures comprising a video sequence as being one of three types. These are as follows:
A. Intra pictures (I) are coded without reference to other pictures. They serve as access points to the coded video sequence where decoding can begin.
B. Predicted pictures (P) are coded with reference to a motion compensated prediction derived from a previous I or P picture. Coding of P pictures is more efficient than for I pictures.
C. Bi-directionally predicted pictures (B) are coded with reference to a forward motion compensated prediction from a previous I or P picture and a backward motion compensated prediction from a future I or P picture and provide the greatest degree of compression.
The order of the three picture types within a video sequence is not constrained and will generally depend on the requirements of the application. A typical sequence of I, P and B pictures is illustrated in Figure 1. Here, frame II is an I picture, frames P4 and P7 are P pictures and frames B2, B3, B5, B6, B8 and B9 are B pictures. Frame P4 is coded with reference to a forward prediction derived from frame II. Frame P7 is coded with reference to a forward prediction derived from frame P4. Frames B2 and B3 are coded with reference to a forward prediction derived from frame 11 and a backward prediction derived from frame P4. Frames B5 and B6 are coded with reference to a forward prediction derived from frame P4 and a backward prediction derived from frame P7. Frames B8 and B9 are coded with reference to a forward prediction derived from frame P7 and a backward prediction derived from frame 110. When a video coder is implemented in hardware, a large part of its circuitry is devoted to the measurement of motion vectors in order to generate the required motion compensated predictions. In the case of B pictures, two simultaneous motion measurements are required in order to generate both the forward and backward predictions. In the case of the MPEG-2 algorithm, twice as many motion measurements are required as both 'frame' and 'field' versions of each motion vector are needed.
We have appreciated that the number of motion measurements, and hence the amount of motion measurement circuitry, may be considerably reduced if some additional processing of the forward motion vectors is performed. Such processing may be implemented in either hardware or software. The invention is defined in the appended claims to which reference should now be made.
Specific embodiments of the invention will now be described in detail by way of example with reference to the accompanying drawings in which: Figure 1 shows the sequence of I, B, and P frames described above,-
Figure 2 shows a conventional motion estimator arrangement required to produce vectors defining the sequence of frames shown in Figure 1; and Figure 3 shows a motion estimator embodying the inventio An MPEG picture sequence of the type shown in Figure 1, which is in display order and where I, P and B refer to the MPEG picture type is as follows.
B6, B7, P8, B9, BIO, Pll, B12, B13, P14, B15, B16, 117, B18, B19, P20 etc
The sequence may be split into "triples" of {BBP} or (less often) {BBl}, i.e. two interpolated and one reference picture.
For each {BBP} triple, a coder's motion estimator must generate five motion vectors. For example, in the case of the triple {B12, B13, P14} the five required motion vectors are:
Pll to P14 (Forward prediction of P14 from Pll)
Pll to B12 (Forward prediction of B12 from Pll) Pll to B13 (Forward prediction of B13 from Pll)
P14 to B12 (Backward prediction of B12 from P14)
P14 to B13 (Backward prediction of B13 from P14)
For each {BBl} triple a coder's motion estimator must generate four motion vectors. For example, in the case of the triple {B15, B16, 117} the four required motion vectors are:
P14 to B15 (Forward prediction of B15 from P14) P14 to B16 (Forward prediction of B16 from P14)
117 to B15 (Backward prediction of B15 from 117)
117 to B16 (Backward prediction of B16 from 117)
To date, the architecture which is generally proposed to generate these motion vectors is based around two motion estimators as shown in Figure 2. This comprises a forward motion estimator 2 receiving a previous reference picture and the current picture and producing forward motion vectors and a backward motion vector estimator 4 receiving a future reference picture and the current picture and producing backward motion vectors . These are the vectors referred to above.
Table 1 shown below describes the passage of the sequence, B12, B13 , P14, B15, B16, 117, through the motion estimators. It will be seen that the sequence of the pictures is re-ordered to P14, B12, B13 , 117, B15, B16, prior to input to the motion estimators in accordance with the MPEG algorithm. Each column of the table represents an instant in time. For example, the first column means, "when P14 is on the current picture input, Pll is applied to the previous reference picture input and the forward motion vector output generates the motion vector Pll to P14". The table demonstrates how the two motion estimators of Figure 1 can together generate the required five motion vectors for each {BBP} triple and the required four motion vectors for each {BBl} triple.
Table 1
Current Picture Input P14 B12 B13 117 B15 B16
Previous Reference Pll Pll Pll P14 P14 P14 Picture Input
Future Reference P14 P14 117 117 Picture Input
Forward Motion Vector Output Pll Pll Pll P14 P14 to to to to to P14 B12 B13 B15 B16
Backward Motion Vector Output P14 P14 117 117 to to to to B12 B13 B15 B16
The amount of hardware required for the video coder can be reduced if only a single one of the motion estimators of Figure 2 is used. This becomes feasible if further vector processing is used to generate the additional motion vectors. An arrangement for a single motion vector estimator is shown in Figure 3.Table 2 illustrates the passage of the re-ordered sequence P14, B12, B13, 117, B15, B16 through the motion estimator of Figure 3.
Table 2
Current Picture Input P14 B12 B13 117 B15 B16
Reference Picture Input Pll Pll Pll P14 P14
P14
Motion Vector Generated Pll Pll Pll P14 P14
P14 to to to to toto
P14 B12 B13 117 B15
B16
The required motion vectors for the {BBP} triple are (Pll to P14) , (Pll to B12) , (Pll to B13), (P14 to B12) and (P14 to B13) . The motion estimator has generated the first three of these. The missing two motion vectors may be calculated using additional vector processing as follows :
(P14 to B12) = (Pll to B12) - (Pll to P14) (P14 to B13) = (Pll to B13) - (Pll to P14)
Similarly, the required motion vectors for the (BBl) triple are (P14 to B15), (P14 to B16) , (117 to B15) and (117 to B16) . The motion estimator has generated the first two of these as well as the vector (P17 to 117) . The missing two motions vectors may be calculated as follows :
(117 to B15) = (P14 to B15) - (P14 to 117) (117 to B16) = (P14 to B16) - (P14 to 117) As long as the forward motion estimator generates the correct forward motion vectors, then the derived backward motion vectors are correct. This holds for objects moving with linear and non-linear motion which represents the majority of picture material. the method can break down if the image contains repetitive patterning (e.g. a chess board) . In such a case, the forward motion estimator may identify a number of motion vectors that will yield equally good predictions even though they need not represent the actual motion of the object. If the forward motion vector is incorrect, the derived backward motion vector may yield a poor prediction. (It should be noted that if a prediction resulting from a backward motion vector is poor, the coder will simply choose some other prediction mode for that particular macroblock and the erroneous backward motion vector will go unnoticed) .
The additional vector processing required to generate all forward and backward vectors using a single vector estimator could be implemented in hardware using vector adders and/or subtractors and some suitable storage and switching arrangement to supply the vectors to the adders at the correct times. Alternatively the processing could be carried out in software.
The current MPEG-2 test model details a further improved bit-rate reduction standard comprising field and frame motion vectors. The idea is as follows.
MPEG-2 defines a macroblock as covering an area of picture 16 pixels by 16 frame lines, 8 lines originating from the odd field and the other 8 lines originating from the even field. The job of a coder's prediction generator stage is to generate "predictions" of the macroblocks comprising the current frame. If the image is moving then these predictions will generally be realised by motion compensating (i.e. assigning one or more motion vectors to each macroblock of a reference frame.
At present the motion estimators (of the type referred to above) generate both 'field' and 'frame' motion vectors in order to derive two predictions of each macroblock in the current frame:- one based on the field motion vectors, the other based on the frame motion vector. The coder chooses the prediction which most closely matches the actual macroblock of the current frame.
In the case of frame motion vectors, a macroblock is treated as a single 16 x 16 block of a frame. A single "frame" motion vector is assigned to describe how a prediction of the macroblock may be generated by motion compensating a 16 x 16 pixel area of a reference frame. In the case of field motion vectors two "field" motion vectors are used to independently generate predictions of the odd field and even field components of a macroblock. i.e. One field motion vector is used to predict the 8 odd lines of the macroblock from the odd field of a reference picture and the other field motion vector is used to predict the 8 even field lines from the even field of a reference picture.
To date, the proposals for implementing field and frame motion vectors in hardware make use of two independent motion estimator blocks, one performing frame motion vector generation, the other performing field motion vector generation. Both motion estimators represent the same amount of hardware. This is because, although the motion estimator concerned with field motion vector generation produces twice as many motion vectors as the other, it needs only to compare [8X16] blocks of image data as opposed to the [16X16] blocks processed by the frame motion estimator.
The motion vectors are used to generate two predictions of a macroblock from the current picture, one generated using the frame motion vector, the other generated using the two field motion vectors. It is then the job of the coder's "decision module" to decide which is the better prediction.
If the motion seen by the motion estimator in the odd field is different from that seen in the even field then the two field motion vectors will be different and the field motion vector based prediction is highly likely to be the best.
If the motion seen by the motion estimator is the same in both fields then the two field motion vectors will be identical and further, they will probably match the frame motion vector. In this case there is unlikely to be any significant difference between the two predictions. We propose that the frame motion estimator can be dispensed with. A frame motion vector may be derived instead by averaging the two field motion vectors. If the two field motion vectors match, then the frame motion vector will have the same value resulting in similar frame and field based predictions, however, if the two field motion vectors are different, the frame motion vector is likely to be incorrect and the field based prediction is likely to be best. The vector generation is explained in more detail below.
Motion occurring in the odd and even fields of a previous frame is independently measured and used to generate predictions of the add and even fields of a future frame, i.e. the odd field of the future frame is predicted from the odd field of the previous frame and correspondingly with the even field. This one prediction of a single future frame comprising odd and even fields has been generated.
Next the vectors measured in the odd field are averaged with those in the even field to give a set of frame motion vectors. A further prediction of the future frame may now be generated by motion compensating the previous frame with the set of frame motion vectors.
Thus there are two predictions of a single specific future frame. One is derived by independently measuring the motion occurring in each field of a previous frame and motion compensating each field accordingly. The other is derived by averaging the motion vectors from the two fields of a previous frame to yield a set of frame motion vectors and using these to compensate the whole of the previous frame. This combination of field vectors can be implemented in dedicated hardware or in software in the motion estimator.

Claims

1. A method for estimating motion vectors defining image displacements between reference images in a sequence of digital video images and intermediate images falling between reference images in the sequence the method comprising the steps of estimating motion vectors for deriving a prediction of one reference image from another in the sequence, estimating motion vectors for deriving a prediction of each intermediate image between two reference images in the sequence from one of the two reference images or their predictions, and from the estimated vectors deriving further vectors for deriving a prediction of each intermediate image from the other one of the said two reference images or their predictions which it falls between.
2. A method according to claim 1 in which the thus derived vectors conform with a predetermined bit-rate reduced video signal standard.
3. A method according to claim 1 in which each image consists of two or more fields and where motion occurring in each field is measured independently in order to generate a set of field motion vectors for each field and the sets of field motion vectors are combined to produce a single set of frame motion vectors, thereby enabling two predictions of a future image to be made.
4. A method for video motion estimation between images in a sequence in which each image consists of two or more fields and where motion occurring in each field is measured independently in order to generate a set of field motion vectors for each field and the sets of field motion vectors are combined to produce a set of frame motion vectors thereby enabling two predictions of a future image to be made.
5. A method for estimating motion vectors defining image displacements between reference frames in a sequence of digital video images and intermediate frames falling between frames in the sequence substantially as herein described.
6. A method for video motion estimation substantially as herein described.
PCT/GB1995/000207 1994-02-02 1995-02-02 Video motion estimation WO1995021503A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP7520453A JPH09509799A (en) 1994-02-02 1995-02-02 Video motion estimation
GB9616294A GB2301255B (en) 1994-02-02 1995-02-02 Video motion estimation
EP95907091A EP0742983A1 (en) 1994-02-02 1995-02-02 Video motion estimation

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB9401989A GB9401989D0 (en) 1994-02-02 1994-02-02 Video motion estimation
GB9401989.0 1994-02-02

Publications (1)

Publication Number Publication Date
WO1995021503A1 true WO1995021503A1 (en) 1995-08-10

Family

ID=10749737

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB1995/000207 WO1995021503A1 (en) 1994-02-02 1995-02-02 Video motion estimation

Country Status (4)

Country Link
EP (1) EP0742983A1 (en)
JP (1) JPH09509799A (en)
GB (2) GB9401989D0 (en)
WO (1) WO1995021503A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2809572A1 (en) * 2000-05-29 2001-11-30 Cit Alcatel VIDEO SOURCE CODING WITH MOTION PREDICTION AND BLOCK EFFECT FILTERING
DE10225434A1 (en) * 2002-06-07 2004-01-08 Siemens Ag Video coding method and apparatus
EP1490982A2 (en) * 2002-03-29 2004-12-29 SONY ELECTRONICS INC. (a Delaware corporation) Method of estimating backward motion vectors within a video sequence
WO2005027520A1 (en) * 2003-09-12 2005-03-24 Institute Of Computing Technology Chinese Academy Of Sciences Bi-directional predicting method for video coding/decoding

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0424026A2 (en) * 1989-10-14 1991-04-24 Sony Corporation Video signal transmitting system
US5049991A (en) * 1989-02-20 1991-09-17 Victor Company Of Japan, Ltd. Movement compensation predictive coding/decoding method
EP0534350A2 (en) * 1991-09-23 1993-03-31 Matsushita Electric Industrial Co., Ltd. Derivation and use of motion vectors in a differential pulse code modulation system
EP0541389A2 (en) * 1991-11-08 1993-05-12 Matsushita Electric Industrial Co., Ltd. Method for predicting move compensation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5049991A (en) * 1989-02-20 1991-09-17 Victor Company Of Japan, Ltd. Movement compensation predictive coding/decoding method
EP0424026A2 (en) * 1989-10-14 1991-04-24 Sony Corporation Video signal transmitting system
EP0534350A2 (en) * 1991-09-23 1993-03-31 Matsushita Electric Industrial Co., Ltd. Derivation and use of motion vectors in a differential pulse code modulation system
EP0541389A2 (en) * 1991-11-08 1993-05-12 Matsushita Electric Industrial Co., Ltd. Method for predicting move compensation

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2809572A1 (en) * 2000-05-29 2001-11-30 Cit Alcatel VIDEO SOURCE CODING WITH MOTION PREDICTION AND BLOCK EFFECT FILTERING
EP1161101A1 (en) * 2000-05-29 2001-12-05 Alcatel Videocoding using motion prediction and filtering for block effect reduction
EP1490982A2 (en) * 2002-03-29 2004-12-29 SONY ELECTRONICS INC. (a Delaware corporation) Method of estimating backward motion vectors within a video sequence
EP1490982A4 (en) * 2002-03-29 2010-12-22 Sony Electronics Inc Method of estimating backward motion vectors within a video sequence
DE10225434A1 (en) * 2002-06-07 2004-01-08 Siemens Ag Video coding method and apparatus
DE10225434B4 (en) * 2002-06-07 2004-12-30 Siemens Ag Video coding method and apparatus
US7580458B2 (en) 2002-06-07 2009-08-25 Siemens Aktiengesellschaft Method and apparatus for video coding
US8111748B2 (en) 2002-06-07 2012-02-07 Siemens Aktiengesellschaft Method and apparatus for video coding
WO2005027520A1 (en) * 2003-09-12 2005-03-24 Institute Of Computing Technology Chinese Academy Of Sciences Bi-directional predicting method for video coding/decoding

Also Published As

Publication number Publication date
GB2301255B (en) 1998-08-26
GB9401989D0 (en) 1994-03-30
JPH09509799A (en) 1997-09-30
GB2301255A (en) 1996-11-27
GB9616294D0 (en) 1996-09-11
EP0742983A1 (en) 1996-11-20

Similar Documents

Publication Publication Date Title
US6519005B2 (en) Method of concurrent multiple-mode motion estimation for digital video
US6591015B1 (en) Video coding method and apparatus with motion compensation and motion vector estimator
US7400681B2 (en) Low-complexity motion vector prediction for video codec with two lists of reference pictures
US5424779A (en) Video coding apparatus
EP2164266B1 (en) Moving picture scalable encoding and decoding method using weighted prediction, their devices, their programs, and recording media storing the programs
US5438374A (en) System and method for filtering video signals
US5592302A (en) Coding method for coding pixel blocks and apparatus therefor
US7321626B2 (en) System and method for predictive motion estimation using a global motion predictor
US5905542A (en) Simplified dual prime video motion estimation
US7852937B2 (en) Motion vector estimation employing line and column vectors
US6415055B1 (en) Moving image encoding method and apparatus, and moving image decoding method and apparatus
US20060291563A1 (en) Interpolation apparatus and method for motion vector compensation
US20070297511A1 (en) Chroma motion vector processing apparatus, system, and method
JP2005318620A (en) Motion vector detection employing adaptive temporal prediction
KR100534192B1 (en) Motion-compensated predictive image encoding and decoding
US6556718B1 (en) Video pictures compression and coding with decision step for field/frame and motion vectors DCT
JPH11112989A (en) Motion vector encoding method and encoder therefor
JPH07193822A (en) Motion prediction processor and device therefor
US6020934A (en) Motion estimation architecture for area and power reduction
US20110051803A1 (en) Latency Rate Distortion Optimisation
EP2076047B1 (en) Video motion estimation
US20100290534A1 (en) Video Encoding Using Pixel Decimation
EP0742983A1 (en) Video motion estimation
EP0921688B1 (en) Moving vector predictive coding method and moving vector decoding method, and storage medium stored with moving vector predictive coding program and moving vector decoding program
EP0578290B1 (en) Motion estimation

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): GB JP US

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
ENP Entry into the national phase

Ref country code: US

Ref document number: 1996 687576

Date of ref document: 19960801

Kind code of ref document: A

Format of ref document f/p: F

WWE Wipo information: entry into national phase

Ref document number: 1995907091

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1995907091

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: CA

WWR Wipo information: refused in national office

Ref document number: 1995907091

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1995907091

Country of ref document: EP