WO2001096982A2 - System for the estimation of optical flow - Google Patents

System for the estimation of optical flow Download PDF

Info

Publication number
WO2001096982A2
WO2001096982A2 PCT/US2001/019012 US0119012W WO0196982A2 WO 2001096982 A2 WO2001096982 A2 WO 2001096982A2 US 0119012 W US0119012 W US 0119012W WO 0196982 A2 WO0196982 A2 WO 0196982A2
Authority
WO
WIPO (PCT)
Prior art keywords
brightness
motion field
dense
dense motion
motion
Prior art date
Application number
PCT/US2001/019012
Other languages
French (fr)
Other versions
WO2001096982A3 (en
Inventor
Siegfried Wonneberger
Max Griessl
Markus Wittkop
Original Assignee
Dynapel Systems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dynapel Systems, Inc. filed Critical Dynapel Systems, Inc.
Publication of WO2001096982A2 publication Critical patent/WO2001096982A2/en
Publication of WO2001096982A3 publication Critical patent/WO2001096982A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/207Analysis of motion for motion estimation over a hierarchy of resolutions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/269Analysis of motion using gradient-based methods

Definitions

  • This invention relates to the estimation of dense motion fields in sequences of images, e.g., video images, by a gradient based optical flow computation.
  • a dense motion field also called a dense motion vector field
  • a dense motion vector field is a set of vectors, one for each pixel of a frame from a set of motion picture frames, wherein the vectors represent the frame to frame movement of pixel-sized image components of the objects depicted in the set of sequential motion picture frames.
  • the vectors 13 representing the change of position of the image components A, B, C and D are vectors of a dense motion vector field.
  • the example of Fig. 1 is a special simplified case in which the square 11 moves without changing its size or orientation.
  • the dense motion field vectors representing the motion of the image components of the square 11 are parallel and are of equal length.
  • the motion of objects in a motion picture is more complex than that represented in the example shown in Fig. 1 and the dense motion field vectors will often not be parallel and will not be of equal length.
  • the pixel-sized image components are technically not pixels since pixels by definition do not move.
  • the image components are components of the objects depicted in the
  • optical flow computation is called optical flow computation.
  • the skeleton of this pyramid consists of a number of image pairs of
  • each image component of a depicted object is assumed to stay at the same brightness from frame to frame. This assumption which is called optical flow constraint, can formally be
  • the constant 2 should be roughly equivalent to the expected noise in the estimate of E x 2 + E y 2 .
  • the minimization is performed using the calculus of
  • correction field (u, v) to the preliminary motion field is calculated by estimating the
  • E t , E x and E y are calculated as:
  • time t lies in the middle between the two images located at
  • initial estimates of the motion field are used to predict a target image and then the motion field between the actual target image and the predicted target image is calculated as a correction field.
  • the correction field will be a dense motion field between the predicted target image and the actual target image. This correction field will then be added by vector addition to the initial estimated dense motion field to obtain a new estimate of the dense motion field.
  • the system of the invention instead of calculating a predicted target image and then determining the correction field from the predicted target image and the actual target image, calculates the correction motion field directly. In this calculation, the partial derivatives E t , E x and E y are determined from the brightness
  • the brightness partial derivatives are determined by
  • the resulting brightness partial derivatives are then used to calculate the correction to
  • Fig. 1 is a diagram used to explain a dense motion field which is calculated by
  • Fig. 2 is a block diagram illustrating the system of the present invention.
  • Fig. 3 is a flow chart of the method of the present invention.
  • Figs. 4 A, 4B and 4C illustrate graphically an example of the coordinate points
  • Fig. 5 graphically illustrates how the mean brightness value is calculated at a given coordinate point.
  • Fig. 6 graphically illustrates how an estimate of a Laplacian is calculated.
  • Fig. 7 is a schematic diagram illustrating the pyramid grid calculating used in
  • pixel frame buffer memories 21 and 22 are fed to pixel frame buffer memories 21 and 22
  • the data processor 24 computes the dense motion field from the brightness values of
  • the data processing unit computes the dense motion field between successive frames of a motion picture
  • the CPU 24 calculates the partial brightness
  • the values E are mean brightness values and calculated at coordinate points
  • the space point (x, y) is a point in the
  • each of the eight mean brightness values E are
  • brightness value E means that the brightness value is determined for a point in the
  • E( ⁇ , Xu + 1, Yr + 1) means the brightness value
  • E(T, Xu > Y ' r + 1) means the brightness value is determined for a coordinate point in
  • Fig. 4B illustrates the
  • E t , E x and E y are determined by the differences in the mean brightnesses at locations in the sequential motion picture frames determined in accordance with the initial estimate vector.
  • E t is determined by the difference between the mean
  • E x is determined by adding the brightness values at points 31
  • E x is determined by differences in mean
  • E y is determined by adding the mean brightness values at the points 33 and 34 in the second frame and at the points 37 and 38 is the first frame and by
  • E y is determined by the differences in the mean brightness values at incrementally spaced points in the Y direction at the
  • the mean brightness E are arbitrary convex combinations of the brightness values of
  • i.e., the largest integer not exceeding ⁇ .
  • Equation (10) computes the E brightness
  • FIG. 5 graphically
  • Equation (10) illustrates an example of the computation of Equation (10). As shown in Fig. 5, unit
  • square 41 surrounds the coordinate point at (X,Y) for which E is being computed.
  • the unit square 41 overlaps the boundaries of four neighboring pixels 43, 45, 47 and
  • E is the weighted average of the brightness of the pixels 43, 45, 47 and 49 with
  • the cited article are used to calculate a dense motion field. Because the brightness partial derivatives are determined as an appropriate function of the initial estimate of the dense motion field, the Horn and Schunck equations will yield a dense motion field which is a correction to the initial estimate and which, when added to the initial
  • the partial brightness derivatives can be related to the dense motion field
  • Vu and V 2 v are the Laplacians of u and v.
  • the finest grid 51 corresponds to the pixel display wherein each square of the grid represents one pixel.
  • the other grids 53 and 55 of the pyramid represent progressively coarser grids
  • the number of grids in the pyramid is an example and a

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Color Television Systems (AREA)
  • Television Systems (AREA)

Abstract

In a system for generating a dense motion field (13) representing the motion of image components in a motion picture, a correction to the initial estimate of the dense motion field is directly calculated by determining partial derivatives with respect to time and space from brightness values in two successive frames in the motion picture. The brightness partial derivatives are determined by calculating temporal and spatial differences in brightness values at positions in the successive frames determined by the vectors of the initial estimate of the motion field (13). The resulting brightness partial derivatives are used to calculate the dense motion field using a motion flow algorithm. The calculated correction to the initial estimate of the dense motion field is then added to the initial estimate to provide a new estimate of the dense motion field. The calculation of the estimated dense motion field is used in the hierarchical pyramid (51, 53, 55) wherein the calculations are carried out on successively finer grids.

Description

SYSTEM FOR THE ESTIMATION OF OPTICAL FLOW
Field of the Invention
This invention relates to the estimation of dense motion fields in sequences of images, e.g., video images, by a gradient based optical flow computation.
Background of the Invention
A dense motion field, also called a dense motion vector field, is a set of vectors, one for each pixel of a frame from a set of motion picture frames, wherein the vectors represent the frame to frame movement of pixel-sized image components of the objects depicted in the set of sequential motion picture frames. For example, as shown in Fig. 1 , if pixels A, B, C and D represent image components of a depicted square object 11 in a first motion picture frame and the square 11' represents where the square 11 has moved in a second motion picture frame, the vectors 13, representing the change of position of the image components A, B, C and D are vectors of a dense motion vector field. The example of Fig. 1 is a special simplified case in which the square 11 moves without changing its size or orientation. As a result, the dense motion field vectors representing the motion of the image components of the square 11 are parallel and are of equal length. Typically, the motion of objects in a motion picture is more complex than that represented in the example shown in Fig. 1 and the dense motion field vectors will often not be parallel and will not be of equal length. It should be noted that the pixel-sized image components are technically not pixels since pixels by definition do not move. The image components, on the other hand, are components of the objects depicted in the
motion picture and these image components change position from frame to frame
when the corresponding objects change position from frame to frame to represent
motion of these objects in the motion picture. The computation of a dense motion
field is called optical flow computation.
Motion estimation by gradient based optical flow computation between two
consecutive images of a sequence has the lack, that the amplitude of motion, which could be determined, is very limited. To overcome this deficiency, the gradient-based
method is used as a motion estimation kernel within a hierarchical pyramid as
described in PCT application Serial No. WO/9907156, which is hereby incorporated
by reference. The skeleton of this pyramid consists of a number of image pairs of
decreasing spatial resolution, derived by reducing (down sampling) the original
images. At each resolution, a correction to the motion field, determined before at the
coarser resolution, is calculated in the motion estimation kernel. After the correction
has been added, the refined motion field is filtered and expanded for use in the next
finer resolution.
The classical, gradient-based motion estimation kernel algorithm introduced
by Horn and Schunck in an article entitled "Determining Optical Flow", published in
Artificial Intelligence, 1981, Vol. 17, pp. 185-203, which is hereby incorporated by
reference. In the Horn and Schunck approach as described in this article, the image
sequence is interpreted as a discrete part of the brightness field E(t, x, y), dense in time
t and space (x, y), with the pixels being located at integer positions. The basic assumption of the Horn and Schunck article is that all brightness changes from frame
to frame of a motion picture are caused by motion. In other words, each image component of a depicted object is assumed to stay at the same brightness from frame to frame. This assumption which is called optical flow constraint, can formally be
expressed by the vanishing total temporal derivative of the brightness field E(t, x, y),
that is:
^-E = 0. (1) at
Applying the chain rule of differentiation yields the equation
E, + Exu + Eyv = 0, ' (2)
dx , dy where u := — and v .- - - dt dt
u and v representing the local components of the motion vector (flow velocity), and
the indices Et , Ex and^ expressing the partial derivatives with respect to time t and
space (x, y), that is, E, = ——EE,, EEx* ==——EE,, aanndd EEyy ==—— . E . As it is impossible to get dt ox dy
local solutions for both components of the motion vector from a single algebraic
equation, Horn and Schunck proposed an additional smoothness constraint for the
motion vector field, hi order to let the motion vector field approximately fulfill both
constraints almost everywhere, they minimize the functional
F(u,v) := ld x d y ((Et + Exu + EyV)2 + a2(ux + uy + vx 2 +vy)), (3)
wherein the integration extends over the whole space (frame) and the positive constant
a2 controls the relative contributions of the optical flow deviation term and the non-
smoothness term. The constant 2 should be roughly equivalent to the expected noise in the estimate of Ex 2 + Ey 2. The minimization is performed using the calculus of
variation, such as disclosed in Methods of Mathematical Physics, R. Courant and R.
Hubert, published by Intersciences, New York, New York 1937, 1953. The pertaining Euler-Lagrange differential equations
(Et + Exu +Eyv)Ex -a2(uxx + uyy) = 0 (4)
(Et + Exu + E,v)Ey - a2(vxx + vy,) = 0 (5)
are re-discretised by replacing the derivatives by discrete difference masks, as
explained in the Horn and Schunck article. The resulting system of linear algebraic
equations can be solved by standard numerical methods, such as disclosed in Matrix
Computation, 2d Ed., by G. H. Golub and C. F. Nan Loan, published by Johns
Hopkins University Press, Baltimore, MD 1989.
The motion refinement method, used in the above-cited PCT patent
application, at each resolution of a hierarchical pyramid, first uses an estimated
preliminary motion field (U, V) to warp one of the images called the source image, in
order to make a prediction for a second image called the target image. Then, the
correction field (u, v) to the preliminary motion field is calculated by estimating the
motion (displacement) field between this prediction and the target image. Within the motion estimation kernel of the Horn & Schunck article, partial brightness derivatives
Et , Ex and Ey are calculated as:
Et, x t, x, y)~ {
Figure imgf000005_0001
(-, +, -) E(T, X+l. Y ) }/4.
wherein
• the bracketed alternative signs correspond with the equally ordered alternative variables in the index of the partial derivative (this means that the first symbol, "+" or "-", inside parenthesis is used when determining Et, the middle symbol inside the parenthesis is used when determining Ex and the last symbol inside the parenthesis is used when determining Ey),
• the brightness values E are calculated at points, expressed with the abbreviations
-*.-=t±-, X:=x- Y -y- (7) 2
It is emphasized, that the time t lies in the middle between the two images located at
consecutive integer times andE+, that is, t e IN + 0.5. The space points (x, y) lie
on the lattice (IN + 0.5)2. The space points (x, y) lie on the lattice (IN + 0.5)2, therefore, and 7 are pixel positions.
Summary of the Invention
As explained above, in the system described in the above-cited PCT application, initial estimates of the motion field are used to predict a target image and then the motion field between the actual target image and the predicted target image is calculated as a correction field. The correction field will be a dense motion field between the predicted target image and the actual target image. This correction field will then be added by vector addition to the initial estimated dense motion field to obtain a new estimate of the dense motion field. In accordance with the present invention, instead of calculating a predicted target image and then determining the correction field from the predicted target image and the actual target image, the system of the invention calculates the correction motion field directly. In this calculation, the partial derivatives Et , Ex and Ey are determined from the brightness
values in the two successive frames for which the dense motion field is being
calculated. More specifically, the brightness partial derivatives are determined by
calculating temporal and spatial differences in brightness values at positions in the
successive frames determined by the vectors of the initial estimate of the motion field.
The resulting brightness partial derivatives are then used to calculate the correction to
the dense motion field by means of the algorithm set forth in the Horn and Schunck
article.
Brief Description of the Drawings
Fig. 1 is a diagram used to explain a dense motion field which is calculated by
the present invention.
Fig. 2 is a block diagram illustrating the system of the present invention.
Fig. 3 is a flow chart of the method of the present invention.
Figs. 4 A, 4B and 4C illustrate graphically an example of the coordinate points
at which mean brightness values are calculated in accordance with the invention.
Fig. 5 graphically illustrates how the mean brightness value is calculated at a given coordinate point.
Fig. 6 graphically illustrates how an estimate of a Laplacian is calculated.
Fig. 7 is a schematic diagram illustrating the pyramid grid calculating used in
the system of the present invention. Description of a Preferred Embodiment hi the system of the invention, as shown in Fig. 2, a source of successive pixel
based motion picture frames are fed to pixel frame buffer memories 21 and 22
wherein the first frame of the sequence is received in the pixel frame buffer memory
21 and the second frame of the sequence is received in pixel frame buffer memory 22.
The data processor 24 computes the dense motion field from the brightness values of
the pixels in the buffer memories 20 and 22.
As shown in Fig. 3, in the system of the present invention, the data processing unit computes the dense motion field between successive frames of a motion picture
by first computing a set of brightness derivatives determined as a function of changes
in brightness with space and time and also in accordance with an initial estimate of the
dense motion field U and N. Following the computation of these brightness
derivatives, the correction to the initial estimate of the dense motion field is calculated
using the Horn and Schunck equations. Following this calculation, the correction to
the dense motion field is added to the initial estimate to provide a new estimate of the
dense motion field.
In accordance with the invention, the CPU 24 calculates the partial brightness
derivatives Et , Ex and Ey talcing the preliminary motion field (U, V) into account.
They are calculated as follows:
E (t,x,y,U,V)~ { E (T+X+v+l,Y; +1)
Figure imgf000009_0001
(+,-,-) E (T+,Xv ,Y; ) (8)
Figure imgf000009_0002
(-,+,+) E (T-,X-U+1.Y +1) (-,-,+) E (T,Xυ ,7-+l)
(-,-,-) E (r-, - 7" )
(-,+,-) E (E-,N-+1, Y- ) }/4.
The bracketed alternative signs correspond with the equally ordered alternative
variables in the index of the partial derivative in the same manner as in Equation (6).
The values E are mean brightness values and calculated at coordinate points,
expressed with the abbreviations
X± u:=x--2 + U(x,y).(l2-(l±l)-λ), γ÷ := y- +V(x.y)*( (l±l)-λ). (9)
where the parameter λ s [0,1] fixes the time t = T " + λ (T+ - T~), whereby the
derivatives are calculated, at an arbitrary point between the two times T~ <T+
belonging to consecutive original images. The space point (x, y) is a point in the
frame corresponding to an image component or vector in the initial estimated field.
Normally (x, y) will lie either on the lattice (IN + 0.5)2 or on the lattice IN2.
Thus, in Equation (8), each of the eight mean brightness values E are
determined for specifically identified points in the first motion picture frame or the
second motion picture frame. For example, T+ in the parenthetical portion of a
brightness value E means that the brightness value is determined for a point in the
second frame and T" means that the brightness value of E is determined for a
coordinate point in the first frame. The coordinates of the point at which the brightness value are determined is indicated by the second and third terms in the
parenthetical expression. Thus, E(τ^ , Xu + 1, Yr + 1) means the brightness value is
determined for a point in the second frame at the
Figure imgf000010_0001
+ l, γv +l and
E(T, Xu> Y'r +1) means the brightness value is determined for a coordinate point in
the first frame at the coordinates χ' υ , Y' r +l. The coordinates
Figure imgf000010_0002
are
determined from equations (9) by using the plus (+) sign for the plus or minus
indicator (±) to compute x and γr and using the minus (-) sign for the plus or
minus indicator (±) to compute χυ ' and γ' r.
The points at which the mean brightness values E are calculated in Equation
(8) are graphically illustrated in Figs. 4A, 4B an 4C for the vector (UN) which is
positioned to pass through the point (x,y), which divides (u,v) in two parts of relative
lengths λ and 1-λ. hi these figures, λ is about 0.37. In these figures, the eight points at
which the mean brightness values are determined are designated 31 through 38. Fig.
4A represents the calculation of Et. In this figure, the plus (+) signs are on the points
31 -34 to indicate that mean brightness values at these points are added in Equation (8)
and the minus (-) signs are on the points 35-38 to indicate that mean brightness values
at these points are subtracted, h a manner similar to Fig. 4A, Fig. 4B illustrates the
calculation of the partial derivative Ex and Fig. 4C illustrates the calculation of the
partial derivative Ey. In Figs. 4A-4C, the coordinates at the initial point of the vector
and the points 35-38 are in the first frame of the two sequential frames and the
coordinates at the terminal point of the vector and the points 31-34 are in the second
of the two motion picture frames. As shown in Figs. 4A-4C, the brightness
differentials Et , Ex and Ey are determined by the differences in the mean brightnesses at locations in the sequential motion picture frames determined in accordance with the initial estimate vector. Et is determined by the difference between the mean
brightness values between the two frames at the imtial and terminal points of the
corresponding vector. Ex is determined by adding the brightness values at points 31
and 34 in the second frame and at points 35 and 38 in the first frame and subtracting
the mean brightness values at the points 32 and 33 in the second frame and at the
points 36 and 37 in the first frame. Thus, Ex is determined by differences in mean
brightness values at points incrementally spaced in the X direction at the initial point
of the vector in the first frame and at the terminal point of the vector in the second
frame. Similarly, Ey is determined by adding the mean brightness values at the points 33 and 34 in the second frame and at the points 37 and 38 is the first frame and by
subtracting the mean brightness values at the points 31 and 32 in the second frame and
at the points 35 and 36 in the first frame. Thus, Ey is determined by the differences in the mean brightness values at incrementally spaced points in the Y direction at the
initial and terminal points of the vector in the first and second frames, respectively.
The mean brightness E are arbitrary convex combinations of the brightness values of
the neighboring pixels and each mean brightness value E is an approximation of the
brightness at the corresponding coordinate point. An approximation is needed
because the initial point and termination point of a vector will not be expected to fall
at the centers of pixels. One reasonable definition for the mean brightness E is given
by E (TXY):- { (l- X-m)) (l-(7-[7j)) E(T,[X\ ,[Y] )
+ (X-[X) (l-(7-[7j)) E(T,[X]+1,[Y] )
+ (X-[X\) (7-[7j)) E(T,[X]+1,[Y]+1) (10)
+(1-(N- Y]) (7-[7j)) E(T,[X ,[7J+1) },
wherein the coefficients of the convex-combination can be interpreted as the
intersection areas of a pixel sized unit square, centered at (X, Y), with the four unit
squares, representing its neighboring pixels. The integer positions of the pixels are
expressed with the help of clipping brackets, indicating that [α] is the integral part of
α, i.e., the largest integer not exceeding α.
The above calculation of Equation (10) computes the E brightness
approximation as the weighted average of four pixels neighboring the coordinate point
for which the E brightness approximation is being computed. Fig. 5 graphically
illustrates an example of the computation of Equation (10). As shown in Fig. 5, unit
square 41 surrounds the coordinate point at (X,Y) for which E is being computed.
The unit square 41 overlaps the boundaries of four neighboring pixels 43, 45, 47 and
49. E is the weighted average of the brightness of the pixels 43, 45, 47 and 49 with
each brightness being weighted in accordance with how much it is overlapped by the
unit square 41. h this manner, an approximation of the brightness at the coordinate
point (X,Y) is determined.
Following the computation of the brightness partial derivatives Et , Ex and
Ey as described above, the equations of Horn and Schunck as set forth in the above-
cited article are used to calculate a dense motion field. Because the brightness partial derivatives are determined as an appropriate function of the initial estimate of the dense motion field, the Horn and Schunck equations will yield a dense motion field which is a correction to the initial estimate and which, when added to the initial
estimate, will provide a new estimate of the dense motion field. As described in Horn
and Schunck, the partial brightness derivatives can be related to the dense motion field
u and v as follows:
E2 xu + ExEyv = 2V2 -ExEt, (11)
ExEy + E2 yv = 2V v-EyEf (12)
In these equations, Vu and V2v are the Laplacians of u and v. As explained by Horn
and Schunck, the Laplacians of u and v can be approximated by subtracting the
magnitudes of u and v from weighted averages of the surrounding magnitudes of u
and v as follows:
V2u K(u-u) and \72v∞ K(v-v), (13)
in which u and are the weighted averages of the values of u and v surrounding the
pixels at which the Laplacians of u and v are being calculated. The weighted
averages and v at the coordinates x,y can be calculated as follows (time dependence
suppressed):
> y)=~ iu(x - V> y) + u(x> y + V + (x + 1, y) + u(x, y-1)} 6 (14)
+ — {u(x-l,y-l) + u(x-l,y + l) + u(χA-l,y + l) + u(x + l,y-l)} 12
> y) = ~ (x - 1). y) + v(x, y + l) + v(x + 1, y) + v(x, y - 1)} 6 (15)
+ — {v(x-l,y-l) + v(x-l,y + l) + v(x + l,y + l) + v(x + l,y-l)} Fig. 6 illustrates the weighting carried out by the above equations for the values at the coordinates x,y. With the approximations of the Laplacians substituted in Equations
(11) and (12) and solving for u and v the following equations result:
(a2 + E2 x + Ey)u = + (a2 + E2 y)u - ExEyv- ExEl, (16)
( +EX 2 + Ey)v = - ExEy +( 2 + E -EyEt. (17)
The above equations provide an expression for u and v at each point in the image.
These equations can be solved iteratively as follows:
u"+1 = ~ u n - Ex [Ex un + Ey 7 + Etl /(a2 + EX 2 + Ey 2), (18)
vn+1 = v" - Ey[Exu" + Eyv" + Etl /( 2 + EX 2 + Ey 2). (19)
The calculations represented by the above iterative equations are repeated until they
converge to provide a dense motion field u and v for each image component. The
calculated motion field (u, v) will be a correction to the initial estimate (U, N) and
when added to the initial estimate will provide a new estimate of the dense motion
field.
The calculation of the estimated dense motion field is then used in a
hierarchical pyramid as shown in Fig. 7. hi this pyramid, the finest grid 51 corresponds to the pixel display wherein each square of the grid represents one pixel.
The other grids 53 and 55 of the pyramid represent progressively coarser grids
representing the same image with larger pixels, hi accordance with the invention, an
initial estimate of the dense motion field is determined for the coarsest grid 55 in the pyramid and the above-described method is then used to calculate a new estimate of the dense motion field for this coarsest grid 55. This new estimate of the coarsest grid
then becomes the initial estimate for the middle grid 53 and a new motion field is
calculated by the method described above for the middle grid 53. This new motion
field estimate then becomes the initial estimate for the finest grid 51 and the
calculation is repeated for the finest grid to produce an estimate of the dense motion
field for the finest grid 51. The number of grids in the pyramid is an example and a
greater number of grids can be used if desired.
hi the above described systems, the equations for solving for the dense motion
field have been expressed in the rectangular coordinate system. It will be apparent
that the system is not limited to calculations employing rectangular coordinates and
other coordinate systems could be used in the solution, such as, for example, polar
coordinates.
The above description is of a preferred embodiment of the invention and
modification may be made thereto without departing from the spirit and scope of the
invention, as defined in the appended claims.

Claims

Claims
1. hi a method for generating a dense motion field to represent the motion of
image components from frame to frame in a motion picture wherein an initial estimate
of said dense motion field is made, a correction dense motion field is calculated and
said correction dense motion field is added to said initial estimate of said dense
motion field to provide a new estimate of said dense motion field, the improvement
wherein said correction dense motion field is determined by estimating partial
derivatives of the brightness of the images in said motion picture with respect to time
and space by the difference of pixel brightness at positions in said motion picture
frames determined in accordance with said initial estimate of said dense motion field,
and using said partial derivatives of brightness to calculate said correction dense
motion field.
2. A method as recited in claim 1, wherein said dense motion field represents the change in position of image components between sequential frames of said motion
picture and wherein said differences in pixel brightness are determined for each vector
of said initial estimate in the proximity of the corresponding image component at
locations in said sequential frames displaced from each other by the corresponding
vector.
3. A method as recited in claim 2, wherein the partial derivatives of brightness
with respect to time are determined for each vector of said initial estimate by the
differences in brightness of pixels between said sequential frames and the partial
derivatives of brightness with respect to space being determined by incremental differences in the brightnesses between pixels in both of said sequential frames.
4. A method as recited in claim 3, wherein the brightness partial derivatives
with respect to time Et and with respect to space Ex and Ey are calculated as:
Figure imgf000017_0001
(+, -, +) E (T+, Xv ,Yr +l)
(+, -, ") E (T Xv ,Yγ )
(+, +, -) E (T Xv +l,Yr )
(-, +, +) E (7-, - +l,7- +l)
(-, -, +) E T-, χ-υ J- +\)
(-, -, ") E CT, N- Yy )
(-, +, ") E (E-, X- +1, Yy ) }/4.
in wlrich the values Ε bar [do elsewhere] are mean brightness values estimated at
coordinate points and successive frames of said motion picture and wherein
Figure imgf000017_0002
in which λ is a value between 0 and 1 and x and y are the coordinate of a coordinate
point in the motion picture frames.
5. a system for generating a dense motion field comprising storage means
connected to receive and store successive frames of a motion picture and a data
processor connected to receive the data representing said motion picture frames and
programmed to generate a dense motion field representing the motion of image
components from frame to frame, wherein an initial estimate of said dense motion field is made and wherein said data processor is programmed to calculate a correction
dense motion field which is added to the initial estimate of said dense motion field to provide a new estimate of said dense motion field, the improvement wherein said data processor is programmed to determine said correction dense motion field by
estimating partial derivatives of brightness of the images in said motion picture with
respect to time and space by the difference of the pixel brightness at positions in said
motion picture frames determined in accordance with said initial estimate of said
dense motion field and using said partial derivatives of brightness to calculate said
correction dense motion field.
6. A system as recited in claim 5, wherein said dense motion field represents a
change in position of image components between sequential frames of said dense
motion field stored in said storage means and wherein said data processor determines
said differences and pixel brightness for each vector of said initial estimate and the proximity of the corresponding image component at locations in said sequential
frames displaced from each other by the corresponding vector.
7. A system as recited in claim 6, wherein the partial derivatives of brightness with respect to time are determined for each vector of said initial estimate by the
differences in brightness of pixels between said sequential frames and the partial
derivatives of brightness with respect to space are determined by incremental
differences in the brightnesses between pixels in both of said sequential frames.
8. A system as recited in claim 7,wherein said data processor is programmed to determine the brightness partial derivative with respect to time Et and the brightness
derivatives with respect to space Ex and Ey in accordance with E χy(t,x,y,U,V)- { E (T+Xv+ 1,7+ +1)
Figure imgf000019_0001
(-, +, +) E (T-,χ-υ +1,7^+1)
(-, -, +) E CT,X- ,7^+1)
(-, -, ") E (E-, " 7" )
(-, +, ") E (E-,X-+1, Yy ) }/4.
in which the values Ε bar [fix on another machine] are mean brightness values calculated at coordinate points in successive frames of said motion picture expressed
with the abbreviations
Figure imgf000019_0002
in which λ is a value between 0 and 1 and x and y are the coordinates of coordinate points in the motion picture frames.
PCT/US2001/019012 2000-06-14 2001-06-14 System for the estimation of optical flow WO2001096982A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US59352100A 2000-06-14 2000-06-14
US09/593,521 2000-06-14

Publications (2)

Publication Number Publication Date
WO2001096982A2 true WO2001096982A2 (en) 2001-12-20
WO2001096982A3 WO2001096982A3 (en) 2002-03-21

Family

ID=24375046

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2001/019012 WO2001096982A2 (en) 2000-06-14 2001-06-14 System for the estimation of optical flow

Country Status (1)

Country Link
WO (1) WO2001096982A2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002077920A2 (en) * 2001-03-26 2002-10-03 Dynapel Systems, Inc. Method and system for the estimation and compensation of brightness changes for optical flow calculations
WO2005006762A3 (en) * 2003-07-02 2005-02-10 Queen Mary & Westfield College Optical flow estimation method
US8170109B2 (en) 2006-03-23 2012-05-01 Nds Limited System for analysis of motion

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5265172A (en) * 1989-10-13 1993-11-23 Texas Instruments Incorporated Method and apparatus for producing optical flow using multi-spectral images
US5500904A (en) * 1992-04-22 1996-03-19 Texas Instruments Incorporated System and method for indicating a change between images
US5881170A (en) * 1995-03-24 1999-03-09 Matsushita Electric Industrial Co., Ltd. Contour extraction apparatus

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5265172A (en) * 1989-10-13 1993-11-23 Texas Instruments Incorporated Method and apparatus for producing optical flow using multi-spectral images
US5500904A (en) * 1992-04-22 1996-03-19 Texas Instruments Incorporated System and method for indicating a change between images
US5881170A (en) * 1995-03-24 1999-03-09 Matsushita Electric Industrial Co., Ltd. Contour extraction apparatus

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HORN BERTHOLD K.P. ET AL.: 'Determining optical flow' ARTIFICIAL INTELLIGENCE vol. 17, pages 185 - 203, XP000195787 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002077920A2 (en) * 2001-03-26 2002-10-03 Dynapel Systems, Inc. Method and system for the estimation and compensation of brightness changes for optical flow calculations
WO2002077920A3 (en) * 2001-03-26 2003-09-18 Dynapel Systems Inc Method and system for the estimation and compensation of brightness changes for optical flow calculations
US6959118B2 (en) 2001-03-26 2005-10-25 Dynapel Systems, Inc. Method and system for the estimation and compensation of brightness changes for optical flow calculations
WO2005006762A3 (en) * 2003-07-02 2005-02-10 Queen Mary & Westfield College Optical flow estimation method
US7822231B2 (en) 2003-07-02 2010-10-26 Queen Mary And Westfield College, University Of London Optical flow estimation method
US8170109B2 (en) 2006-03-23 2012-05-01 Nds Limited System for analysis of motion

Also Published As

Publication number Publication date
WO2001096982A3 (en) 2002-03-21

Similar Documents

Publication Publication Date Title
CN109859126B (en) Video noise reduction method and device, electronic equipment and storage medium
US8768069B2 (en) Image enhancement apparatus and method
KR101498532B1 (en) Digital processing method and system for determination of optical flow
Wu et al. Image registration using wavelet-based motion model
JP5385969B2 (en) Method and apparatus for super-resolution of images
Ghosal et al. A fast scalable algorithm for discontinuous optical flow estimation
KR20190065287A (en) Prediction of depth from image data using statistical model
Wu et al. Optical flow estimation using wavelet motion model
EP0652536B1 (en) Method and apparatus for estimating motion vector fields by rejecting local outliers
JP5107409B2 (en) Motion detection method and filtering method using nonlinear smoothing of motion region
KR20080063770A (en) Extracting a moving object boundary
US20110091074A1 (en) Moving object detection method and moving object detection apparatus
CN108986150B (en) Image optical flow estimation method and system based on non-rigid dense matching
US20120062548A1 (en) Reducing viewing discomfort
US20090180032A1 (en) Method and system for hierarchical motion estimation with multi-layer sub-pixel accuracy and motion vector smoothing
Adarve et al. A filter formulation for computing real time optical flow
Whitaker A level-set approach to image blending
Vazquez et al. Joint multiregion segmentation and parametric estimation of image motion by basis function representation and level set evolution
CN113259605A (en) Video matting method, system and storage medium based on prediction foreground mask prediction
Al Ismaeil et al. Real-time enhancement of dynamic depth videos with non-rigid deformations
WO2001096982A2 (en) System for the estimation of optical flow
Al-Regib et al. Hierarchical motion estimation with content-based meshes
CN103618904B (en) Motion estimation method and device based on pixels
CN113240705A (en) 3D attitude estimation method and device, electronic equipment and storage medium
Nomura Spatio-temporal optimization method for determining motion vector fields under non-stationary illumination

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): JP NO

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR

121 Ep: the epo has been informed by wipo that ep was designated in this application
AK Designated states

Kind code of ref document: A3

Designated state(s): JP NO

AL Designated countries for regional patents

Kind code of ref document: A3

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase in:

Ref country code: JP