WO2001096982A2 - System for the estimation of optical flow - Google Patents
System for the estimation of optical flow Download PDFInfo
- Publication number
- WO2001096982A2 WO2001096982A2 PCT/US2001/019012 US0119012W WO0196982A2 WO 2001096982 A2 WO2001096982 A2 WO 2001096982A2 US 0119012 W US0119012 W US 0119012W WO 0196982 A2 WO0196982 A2 WO 0196982A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- brightness
- motion field
- dense
- dense motion
- motion
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/207—Analysis of motion for motion estimation over a hierarchy of resolutions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/269—Analysis of motion using gradient-based methods
Definitions
- This invention relates to the estimation of dense motion fields in sequences of images, e.g., video images, by a gradient based optical flow computation.
- a dense motion field also called a dense motion vector field
- a dense motion vector field is a set of vectors, one for each pixel of a frame from a set of motion picture frames, wherein the vectors represent the frame to frame movement of pixel-sized image components of the objects depicted in the set of sequential motion picture frames.
- the vectors 13 representing the change of position of the image components A, B, C and D are vectors of a dense motion vector field.
- the example of Fig. 1 is a special simplified case in which the square 11 moves without changing its size or orientation.
- the dense motion field vectors representing the motion of the image components of the square 11 are parallel and are of equal length.
- the motion of objects in a motion picture is more complex than that represented in the example shown in Fig. 1 and the dense motion field vectors will often not be parallel and will not be of equal length.
- the pixel-sized image components are technically not pixels since pixels by definition do not move.
- the image components are components of the objects depicted in the
- optical flow computation is called optical flow computation.
- the skeleton of this pyramid consists of a number of image pairs of
- each image component of a depicted object is assumed to stay at the same brightness from frame to frame. This assumption which is called optical flow constraint, can formally be
- the constant 2 should be roughly equivalent to the expected noise in the estimate of E x 2 + E y 2 .
- the minimization is performed using the calculus of
- correction field (u, v) to the preliminary motion field is calculated by estimating the
- E t , E x and E y are calculated as:
- time t lies in the middle between the two images located at
- initial estimates of the motion field are used to predict a target image and then the motion field between the actual target image and the predicted target image is calculated as a correction field.
- the correction field will be a dense motion field between the predicted target image and the actual target image. This correction field will then be added by vector addition to the initial estimated dense motion field to obtain a new estimate of the dense motion field.
- the system of the invention instead of calculating a predicted target image and then determining the correction field from the predicted target image and the actual target image, calculates the correction motion field directly. In this calculation, the partial derivatives E t , E x and E y are determined from the brightness
- the brightness partial derivatives are determined by
- the resulting brightness partial derivatives are then used to calculate the correction to
- Fig. 1 is a diagram used to explain a dense motion field which is calculated by
- Fig. 2 is a block diagram illustrating the system of the present invention.
- Fig. 3 is a flow chart of the method of the present invention.
- Figs. 4 A, 4B and 4C illustrate graphically an example of the coordinate points
- Fig. 5 graphically illustrates how the mean brightness value is calculated at a given coordinate point.
- Fig. 6 graphically illustrates how an estimate of a Laplacian is calculated.
- Fig. 7 is a schematic diagram illustrating the pyramid grid calculating used in
- pixel frame buffer memories 21 and 22 are fed to pixel frame buffer memories 21 and 22
- the data processor 24 computes the dense motion field from the brightness values of
- the data processing unit computes the dense motion field between successive frames of a motion picture
- the CPU 24 calculates the partial brightness
- the values E are mean brightness values and calculated at coordinate points
- the space point (x, y) is a point in the
- each of the eight mean brightness values E are
- brightness value E means that the brightness value is determined for a point in the
- E( ⁇ , Xu + 1, Yr + 1) means the brightness value
- E(T, Xu > Y ' r + 1) means the brightness value is determined for a coordinate point in
- Fig. 4B illustrates the
- E t , E x and E y are determined by the differences in the mean brightnesses at locations in the sequential motion picture frames determined in accordance with the initial estimate vector.
- E t is determined by the difference between the mean
- E x is determined by adding the brightness values at points 31
- E x is determined by differences in mean
- E y is determined by adding the mean brightness values at the points 33 and 34 in the second frame and at the points 37 and 38 is the first frame and by
- E y is determined by the differences in the mean brightness values at incrementally spaced points in the Y direction at the
- the mean brightness E are arbitrary convex combinations of the brightness values of
- ⁇ i.e., the largest integer not exceeding ⁇ .
- Equation (10) computes the E brightness
- FIG. 5 graphically
- Equation (10) illustrates an example of the computation of Equation (10). As shown in Fig. 5, unit
- square 41 surrounds the coordinate point at (X,Y) for which E is being computed.
- the unit square 41 overlaps the boundaries of four neighboring pixels 43, 45, 47 and
- E is the weighted average of the brightness of the pixels 43, 45, 47 and 49 with
- the cited article are used to calculate a dense motion field. Because the brightness partial derivatives are determined as an appropriate function of the initial estimate of the dense motion field, the Horn and Schunck equations will yield a dense motion field which is a correction to the initial estimate and which, when added to the initial
- the partial brightness derivatives can be related to the dense motion field
- Vu and V 2 v are the Laplacians of u and v.
- the finest grid 51 corresponds to the pixel display wherein each square of the grid represents one pixel.
- the other grids 53 and 55 of the pyramid represent progressively coarser grids
- the number of grids in the pyramid is an example and a
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Color Television Systems (AREA)
- Television Systems (AREA)
Abstract
In a system for generating a dense motion field (13) representing the motion of image components in a motion picture, a correction to the initial estimate of the dense motion field is directly calculated by determining partial derivatives with respect to time and space from brightness values in two successive frames in the motion picture. The brightness partial derivatives are determined by calculating temporal and spatial differences in brightness values at positions in the successive frames determined by the vectors of the initial estimate of the motion field (13). The resulting brightness partial derivatives are used to calculate the dense motion field using a motion flow algorithm. The calculated correction to the initial estimate of the dense motion field is then added to the initial estimate to provide a new estimate of the dense motion field. The calculation of the estimated dense motion field is used in the hierarchical pyramid (51, 53, 55) wherein the calculations are carried out on successively finer grids.
Description
SYSTEM FOR THE ESTIMATION OF OPTICAL FLOW
Field of the Invention
This invention relates to the estimation of dense motion fields in sequences of images, e.g., video images, by a gradient based optical flow computation.
Background of the Invention
A dense motion field, also called a dense motion vector field, is a set of vectors, one for each pixel of a frame from a set of motion picture frames, wherein the vectors represent the frame to frame movement of pixel-sized image components of the objects depicted in the set of sequential motion picture frames. For example, as shown in Fig. 1 , if pixels A, B, C and D represent image components of a depicted square object 11 in a first motion picture frame and the square 11' represents where the square 11 has moved in a second motion picture frame, the vectors 13, representing the change of position of the image components A, B, C and D are vectors of a dense motion vector field. The example of Fig. 1 is a special simplified case in which the square 11 moves without changing its size or orientation. As a result, the dense motion field vectors representing the motion of the image components of the square 11 are parallel and are of equal length. Typically, the motion of objects in a motion picture is more complex than that represented in the example shown in Fig. 1 and the dense motion field vectors will often not be parallel and will not be of equal length. It should be noted that the pixel-sized image components are technically not pixels since pixels by definition do not move. The
image components, on the other hand, are components of the objects depicted in the
motion picture and these image components change position from frame to frame
when the corresponding objects change position from frame to frame to represent
motion of these objects in the motion picture. The computation of a dense motion
field is called optical flow computation.
Motion estimation by gradient based optical flow computation between two
consecutive images of a sequence has the lack, that the amplitude of motion, which could be determined, is very limited. To overcome this deficiency, the gradient-based
method is used as a motion estimation kernel within a hierarchical pyramid as
described in PCT application Serial No. WO/9907156, which is hereby incorporated
by reference. The skeleton of this pyramid consists of a number of image pairs of
decreasing spatial resolution, derived by reducing (down sampling) the original
images. At each resolution, a correction to the motion field, determined before at the
coarser resolution, is calculated in the motion estimation kernel. After the correction
has been added, the refined motion field is filtered and expanded for use in the next
finer resolution.
The classical, gradient-based motion estimation kernel algorithm introduced
by Horn and Schunck in an article entitled "Determining Optical Flow", published in
Artificial Intelligence, 1981, Vol. 17, pp. 185-203, which is hereby incorporated by
reference. In the Horn and Schunck approach as described in this article, the image
sequence is interpreted as a discrete part of the brightness field E(t, x, y), dense in time
t and space (x, y), with the pixels being located at integer positions. The basic assumption of the Horn and Schunck article is that all brightness changes from frame
to frame of a motion picture are caused by motion. In other words, each image
component of a depicted object is assumed to stay at the same brightness from frame to frame. This assumption which is called optical flow constraint, can formally be
expressed by the vanishing total temporal derivative of the brightness field E(t, x, y),
that is:
^-E = 0. (1) at
Applying the chain rule of differentiation yields the equation
E, + Exu + Eyv = 0, ' (2)
dx , dy where u := — and v .- - - dt dt
u and v representing the local components of the motion vector (flow velocity), and
the indices Et , Ex and^ expressing the partial derivatives with respect to time t and
space (x, y), that is, E, = ——EE,, EEx* ==——EE,, aanndd EEyy ==—— . E . As it is impossible to get dt ox dy
local solutions for both components of the motion vector from a single algebraic
equation, Horn and Schunck proposed an additional smoothness constraint for the
motion vector field, hi order to let the motion vector field approximately fulfill both
constraints almost everywhere, they minimize the functional
F(u,v) := ld x d y ((Et + Exu + EyV)2 + a2(ux + uy + vx 2 +vy)), (3)
wherein the integration extends over the whole space (frame) and the positive constant
a2 controls the relative contributions of the optical flow deviation term and the non-
smoothness term. The constant 2 should be roughly equivalent to the expected noise in the estimate of Ex 2 + Ey 2. The minimization is performed using the calculus of
variation, such as disclosed in Methods of Mathematical Physics, R. Courant and R.
Hubert, published by Intersciences, New York, New York 1937, 1953. The pertaining
Euler-Lagrange differential equations
(Et + Exu +Eyv)Ex -a2(uxx + uyy) = 0 (4)
(Et + Exu + E,v)Ey - a2(vxx + vy,) = 0 (5)
are re-discretised by replacing the derivatives by discrete difference masks, as
explained in the Horn and Schunck article. The resulting system of linear algebraic
equations can be solved by standard numerical methods, such as disclosed in Matrix
Computation, 2d Ed., by G. H. Golub and C. F. Nan Loan, published by Johns
Hopkins University Press, Baltimore, MD 1989.
The motion refinement method, used in the above-cited PCT patent
application, at each resolution of a hierarchical pyramid, first uses an estimated
preliminary motion field (U, V) to warp one of the images called the source image, in
order to make a prediction for a second image called the target image. Then, the
correction field (u, v) to the preliminary motion field is calculated by estimating the
motion (displacement) field between this prediction and the target image. Within the motion estimation kernel of the Horn & Schunck article, partial brightness derivatives
Et , Ex and Ey are calculated as:
Et, x t, x, y)~ {
wherein
• the bracketed alternative signs correspond with the equally ordered alternative
variables in the index of the partial derivative (this means that the first symbol, "+" or "-", inside parenthesis is used when determining Et, the middle symbol inside the parenthesis is used when determining Ex and the last symbol inside the parenthesis is used when determining Ey),
• the brightness values E are calculated at points, expressed with the abbreviations
-*.-=t±-, X:=x- Y -y- (7) 2
It is emphasized, that the time t lies in the middle between the two images located at
consecutive integer times andE+, that is, t e IN + 0.5. The space points (x, y) lie
on the lattice (IN + 0.5)2. The space points (x, y) lie on the lattice (IN + 0.5)2, therefore, and 7 are pixel positions.
Summary of the Invention
As explained above, in the system described in the above-cited PCT application, initial estimates of the motion field are used to predict a target image and then the motion field between the actual target image and the predicted target image is calculated as a correction field. The correction field will be a dense motion field between the predicted target image and the actual target image. This correction field will then be added by vector addition to the initial estimated dense motion field to obtain a new estimate of the dense motion field. In accordance with the present invention, instead of calculating a predicted target image and then determining the correction field from the predicted target image and the actual target image, the
system of the invention calculates the correction motion field directly. In this calculation, the partial derivatives Et , Ex and Ey are determined from the brightness
values in the two successive frames for which the dense motion field is being
calculated. More specifically, the brightness partial derivatives are determined by
calculating temporal and spatial differences in brightness values at positions in the
successive frames determined by the vectors of the initial estimate of the motion field.
The resulting brightness partial derivatives are then used to calculate the correction to
the dense motion field by means of the algorithm set forth in the Horn and Schunck
article.
Brief Description of the Drawings
Fig. 1 is a diagram used to explain a dense motion field which is calculated by
the present invention.
Fig. 2 is a block diagram illustrating the system of the present invention.
Fig. 3 is a flow chart of the method of the present invention.
Figs. 4 A, 4B and 4C illustrate graphically an example of the coordinate points
at which mean brightness values are calculated in accordance with the invention.
Fig. 5 graphically illustrates how the mean brightness value is calculated at a given coordinate point.
Fig. 6 graphically illustrates how an estimate of a Laplacian is calculated.
Fig. 7 is a schematic diagram illustrating the pyramid grid calculating used in
the system of the present invention.
Description of a Preferred Embodiment hi the system of the invention, as shown in Fig. 2, a source of successive pixel
based motion picture frames are fed to pixel frame buffer memories 21 and 22
wherein the first frame of the sequence is received in the pixel frame buffer memory
21 and the second frame of the sequence is received in pixel frame buffer memory 22.
The data processor 24 computes the dense motion field from the brightness values of
the pixels in the buffer memories 20 and 22.
As shown in Fig. 3, in the system of the present invention, the data processing unit computes the dense motion field between successive frames of a motion picture
by first computing a set of brightness derivatives determined as a function of changes
in brightness with space and time and also in accordance with an initial estimate of the
dense motion field U and N. Following the computation of these brightness
derivatives, the correction to the initial estimate of the dense motion field is calculated
using the Horn and Schunck equations. Following this calculation, the correction to
the dense motion field is added to the initial estimate to provide a new estimate of the
dense motion field.
In accordance with the invention, the CPU 24 calculates the partial brightness
derivatives Et , Ex and Ey talcing the preliminary motion field (U, V) into account.
They are calculated as follows:
(-,+,+) E (T-,X-U+1.Y +1) (-,-,+) E (T,Xυ ,7-+l)
(-,-,-) E (r-, - 7" )
(-,+,-) E (E-,N-+1, Y- ) }/4.
The bracketed alternative signs correspond with the equally ordered alternative
variables in the index of the partial derivative in the same manner as in Equation (6).
The values E are mean brightness values and calculated at coordinate points,
expressed with the abbreviations
X± u:=x--2 + U(x,y).(l2-(l±l)-λ), γ÷ := y- +V(x.y)*( (l±l)-λ). (9)
where the parameter λ s [0,1] fixes the time t = T " + λ (T+ - T~), whereby the
derivatives are calculated, at an arbitrary point between the two times T~ <T+
belonging to consecutive original images. The space point (x, y) is a point in the
frame corresponding to an image component or vector in the initial estimated field.
Normally (x, y) will lie either on the lattice (IN + 0.5)2 or on the lattice IN2.
Thus, in Equation (8), each of the eight mean brightness values E are
determined for specifically identified points in the first motion picture frame or the
second motion picture frame. For example, T+ in the parenthetical portion of a
brightness value E means that the brightness value is determined for a point in the
second frame and T" means that the brightness value of E is determined for a
coordinate point in the first frame. The coordinates of the point at which the
brightness value are determined is indicated by the second and third terms in the
parenthetical expression. Thus, E(τ^ , Xu + 1, Yr + 1) means the brightness value is
E(T, Xu> Y'r +1) means the brightness value is determined for a coordinate point in
determined from equations (9) by using the plus (+) sign for the plus or minus
indicator (±) to compute x and γr and using the minus (-) sign for the plus or
minus indicator (±) to compute χυ ' and γ' r.
The points at which the mean brightness values E are calculated in Equation
(8) are graphically illustrated in Figs. 4A, 4B an 4C for the vector (UN) which is
positioned to pass through the point (x,y), which divides (u,v) in two parts of relative
lengths λ and 1-λ. hi these figures, λ is about 0.37. In these figures, the eight points at
which the mean brightness values are determined are designated 31 through 38. Fig.
4A represents the calculation of Et. In this figure, the plus (+) signs are on the points
31 -34 to indicate that mean brightness values at these points are added in Equation (8)
and the minus (-) signs are on the points 35-38 to indicate that mean brightness values
at these points are subtracted, h a manner similar to Fig. 4A, Fig. 4B illustrates the
calculation of the partial derivative Ex and Fig. 4C illustrates the calculation of the
partial derivative Ey. In Figs. 4A-4C, the coordinates at the initial point of the vector
and the points 35-38 are in the first frame of the two sequential frames and the
coordinates at the terminal point of the vector and the points 31-34 are in the second
of the two motion picture frames. As shown in Figs. 4A-4C, the brightness
differentials Et , Ex and Ey are determined by the differences in the mean brightnesses
at locations in the sequential motion picture frames determined in accordance with the initial estimate vector. Et is determined by the difference between the mean
brightness values between the two frames at the imtial and terminal points of the
corresponding vector. Ex is determined by adding the brightness values at points 31
and 34 in the second frame and at points 35 and 38 in the first frame and subtracting
the mean brightness values at the points 32 and 33 in the second frame and at the
points 36 and 37 in the first frame. Thus, Ex is determined by differences in mean
brightness values at points incrementally spaced in the X direction at the initial point
of the vector in the first frame and at the terminal point of the vector in the second
frame. Similarly, Ey is determined by adding the mean brightness values at the points 33 and 34 in the second frame and at the points 37 and 38 is the first frame and by
subtracting the mean brightness values at the points 31 and 32 in the second frame and
at the points 35 and 36 in the first frame. Thus, Ey is determined by the differences in the mean brightness values at incrementally spaced points in the Y direction at the
initial and terminal points of the vector in the first and second frames, respectively.
The mean brightness E are arbitrary convex combinations of the brightness values of
the neighboring pixels and each mean brightness value E is an approximation of the
brightness at the corresponding coordinate point. An approximation is needed
because the initial point and termination point of a vector will not be expected to fall
at the centers of pixels. One reasonable definition for the mean brightness E is given
by
E (TXY):- { (l- X-m)) (l-(7-[7j)) E(T,[X\ ,[Y] )
+ (X-[X) (l-(7-[7j)) E(T,[X]+1,[Y] )
+ (X-[X\) (7-[7j)) E(T,[X]+1,[Y]+1) (10)
+(1-(N- Y]) (7-[7j)) E(T,[X ,[7J+1) },
wherein the coefficients of the convex-combination can be interpreted as the
intersection areas of a pixel sized unit square, centered at (X, Y), with the four unit
squares, representing its neighboring pixels. The integer positions of the pixels are
expressed with the help of clipping brackets, indicating that [α] is the integral part of
α, i.e., the largest integer not exceeding α.
The above calculation of Equation (10) computes the E brightness
approximation as the weighted average of four pixels neighboring the coordinate point
for which the E brightness approximation is being computed. Fig. 5 graphically
illustrates an example of the computation of Equation (10). As shown in Fig. 5, unit
square 41 surrounds the coordinate point at (X,Y) for which E is being computed.
The unit square 41 overlaps the boundaries of four neighboring pixels 43, 45, 47 and
49. E is the weighted average of the brightness of the pixels 43, 45, 47 and 49 with
each brightness being weighted in accordance with how much it is overlapped by the
unit square 41. h this manner, an approximation of the brightness at the coordinate
point (X,Y) is determined.
Following the computation of the brightness partial derivatives Et , Ex and
Ey as described above, the equations of Horn and Schunck as set forth in the above-
cited article are used to calculate a dense motion field. Because the brightness partial derivatives are determined as an appropriate function of the initial estimate of the
dense motion field, the Horn and Schunck equations will yield a dense motion field which is a correction to the initial estimate and which, when added to the initial
estimate, will provide a new estimate of the dense motion field. As described in Horn
and Schunck, the partial brightness derivatives can be related to the dense motion field
u and v as follows:
E2 xu + ExEyv = 2V2 -ExEt, (11)
ExEy + E2 yv = 2V v-EyEf (12)
In these equations, Vu and V2v are the Laplacians of u and v. As explained by Horn
and Schunck, the Laplacians of u and v can be approximated by subtracting the
magnitudes of u and v from weighted averages of the surrounding magnitudes of u
and v as follows:
V2u K(u-u) and \72v∞ K(v-v), (13)
in which u and are the weighted averages of the values of u and v surrounding the
pixels at which the Laplacians of u and v are being calculated. The weighted
averages and v at the coordinates x,y can be calculated as follows (time dependence
suppressed):
> y)=~ iu(x - V> y) + u(x> y + V + (x + 1, y) + u(x, y-1)} 6 (14)
+ — {u(x-l,y-l) + u(x-l,y + l) + u(χA-l,y + l) + u(x + l,y-l)} 12
> y) = ~ (x - 1). y) + v(x, y + l) + v(x + 1, y) + v(x, y - 1)} 6 (15)
+ — {v(x-l,y-l) + v(x-l,y + l) + v(x + l,y + l) + v(x + l,y-l)}
Fig. 6 illustrates the weighting carried out by the above equations for the values at the coordinates x,y. With the approximations of the Laplacians substituted in Equations
(11) and (12) and solving for u and v the following equations result:
(a2 + E2 x + Ey)u = + (a2 + E2 y)u - ExEyv- ExEl, (16)
( +EX 2 + Ey)v = - ExEy +( 2 + E -EyEt. (17)
The above equations provide an expression for u and v at each point in the image.
These equations can be solved iteratively as follows:
u"+1 = ~ u n - Ex [Ex un + Ey 7 + Etl /(a2 + EX 2 + Ey 2), (18)
vn+1 = v" - Ey[Exu" + Eyv" + Etl /( 2 + EX 2 + Ey 2). (19)
The calculations represented by the above iterative equations are repeated until they
converge to provide a dense motion field u and v for each image component. The
calculated motion field (u, v) will be a correction to the initial estimate (U, N) and
when added to the initial estimate will provide a new estimate of the dense motion
field.
The calculation of the estimated dense motion field is then used in a
hierarchical pyramid as shown in Fig. 7. hi this pyramid, the finest grid 51 corresponds to the pixel display wherein each square of the grid represents one pixel.
The other grids 53 and 55 of the pyramid represent progressively coarser grids
representing the same image with larger pixels, hi accordance with the invention, an
initial estimate of the dense motion field is determined for the coarsest grid 55 in the
pyramid and the above-described method is then used to calculate a new estimate of the dense motion field for this coarsest grid 55. This new estimate of the coarsest grid
then becomes the initial estimate for the middle grid 53 and a new motion field is
calculated by the method described above for the middle grid 53. This new motion
field estimate then becomes the initial estimate for the finest grid 51 and the
calculation is repeated for the finest grid to produce an estimate of the dense motion
field for the finest grid 51. The number of grids in the pyramid is an example and a
greater number of grids can be used if desired.
hi the above described systems, the equations for solving for the dense motion
field have been expressed in the rectangular coordinate system. It will be apparent
that the system is not limited to calculations employing rectangular coordinates and
other coordinate systems could be used in the solution, such as, for example, polar
coordinates.
The above description is of a preferred embodiment of the invention and
modification may be made thereto without departing from the spirit and scope of the
invention, as defined in the appended claims.
Claims
1. hi a method for generating a dense motion field to represent the motion of
image components from frame to frame in a motion picture wherein an initial estimate
of said dense motion field is made, a correction dense motion field is calculated and
said correction dense motion field is added to said initial estimate of said dense
motion field to provide a new estimate of said dense motion field, the improvement
wherein said correction dense motion field is determined by estimating partial
derivatives of the brightness of the images in said motion picture with respect to time
and space by the difference of pixel brightness at positions in said motion picture
frames determined in accordance with said initial estimate of said dense motion field,
and using said partial derivatives of brightness to calculate said correction dense
motion field.
2. A method as recited in claim 1, wherein said dense motion field represents the change in position of image components between sequential frames of said motion
picture and wherein said differences in pixel brightness are determined for each vector
of said initial estimate in the proximity of the corresponding image component at
locations in said sequential frames displaced from each other by the corresponding
vector.
3. A method as recited in claim 2, wherein the partial derivatives of brightness
with respect to time are determined for each vector of said initial estimate by the
differences in brightness of pixels between said sequential frames and the partial
derivatives of brightness with respect to space being determined by incremental differences in the brightnesses between pixels in both of said sequential frames.
4. A method as recited in claim 3, wherein the brightness partial derivatives
with respect to time Et and with respect to space Ex and Ey are calculated as:
(+, -, +) E (T+, Xv ,Yr +l)
(+, -, ") E (T Xv ,Yγ )
(+, +, -) E (T Xv +l,Yr )
(-, +, +) E (7-, - +l,7- +l)
(-, -, +) E T-, χ-υ J- +\)
(-, -, ") E CT, N- Yy )
(-, +, ") E (E-, X- +1, Yy ) }/4.
in wlrich the values Ε bar [do elsewhere] are mean brightness values estimated at
coordinate points and successive frames of said motion picture and wherein
point in the motion picture frames.
5. a system for generating a dense motion field comprising storage means
connected to receive and store successive frames of a motion picture and a data
processor connected to receive the data representing said motion picture frames and
programmed to generate a dense motion field representing the motion of image
components from frame to frame, wherein an initial estimate of said dense motion field is made and wherein said data processor is programmed to calculate a correction
dense motion field which is added to the initial estimate of said dense motion field to provide a new estimate of said dense motion field, the improvement wherein said data processor is programmed to determine said correction dense motion field by
estimating partial derivatives of brightness of the images in said motion picture with
respect to time and space by the difference of the pixel brightness at positions in said
motion picture frames determined in accordance with said initial estimate of said
dense motion field and using said partial derivatives of brightness to calculate said
correction dense motion field.
6. A system as recited in claim 5, wherein said dense motion field represents a
change in position of image components between sequential frames of said dense
motion field stored in said storage means and wherein said data processor determines
said differences and pixel brightness for each vector of said initial estimate and the proximity of the corresponding image component at locations in said sequential
frames displaced from each other by the corresponding vector.
7. A system as recited in claim 6, wherein the partial derivatives of brightness with respect to time are determined for each vector of said initial estimate by the
differences in brightness of pixels between said sequential frames and the partial
derivatives of brightness with respect to space are determined by incremental
differences in the brightnesses between pixels in both of said sequential frames.
8. A system as recited in claim 7,wherein said data processor is programmed to determine the brightness partial derivative with respect to time Et and the brightness
derivatives with respect to space Ex and Ey in accordance with E χy(t,x,y,U,V)- { E (T+Xv+ 1,7+ +1)
(-, +, +) E (T-,χ-υ +1,7^+1)
(-, -, +) E CT,X- ,7^+1)
(-, -, ") E (E-, " 7" )
(-, +, ") E (E-,X-+1, Yy ) }/4.
in which the values Ε bar [fix on another machine] are mean brightness values calculated at coordinate points in successive frames of said motion picture expressed
with the abbreviations
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US59352100A | 2000-06-14 | 2000-06-14 | |
US09/593,521 | 2000-06-14 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2001096982A2 true WO2001096982A2 (en) | 2001-12-20 |
WO2001096982A3 WO2001096982A3 (en) | 2002-03-21 |
Family
ID=24375046
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2001/019012 WO2001096982A2 (en) | 2000-06-14 | 2001-06-14 | System for the estimation of optical flow |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2001096982A2 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002077920A2 (en) * | 2001-03-26 | 2002-10-03 | Dynapel Systems, Inc. | Method and system for the estimation and compensation of brightness changes for optical flow calculations |
WO2005006762A3 (en) * | 2003-07-02 | 2005-02-10 | Queen Mary & Westfield College | Optical flow estimation method |
US8170109B2 (en) | 2006-03-23 | 2012-05-01 | Nds Limited | System for analysis of motion |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5265172A (en) * | 1989-10-13 | 1993-11-23 | Texas Instruments Incorporated | Method and apparatus for producing optical flow using multi-spectral images |
US5500904A (en) * | 1992-04-22 | 1996-03-19 | Texas Instruments Incorporated | System and method for indicating a change between images |
US5881170A (en) * | 1995-03-24 | 1999-03-09 | Matsushita Electric Industrial Co., Ltd. | Contour extraction apparatus |
-
2001
- 2001-06-14 WO PCT/US2001/019012 patent/WO2001096982A2/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5265172A (en) * | 1989-10-13 | 1993-11-23 | Texas Instruments Incorporated | Method and apparatus for producing optical flow using multi-spectral images |
US5500904A (en) * | 1992-04-22 | 1996-03-19 | Texas Instruments Incorporated | System and method for indicating a change between images |
US5881170A (en) * | 1995-03-24 | 1999-03-09 | Matsushita Electric Industrial Co., Ltd. | Contour extraction apparatus |
Non-Patent Citations (1)
Title |
---|
HORN BERTHOLD K.P. ET AL.: 'Determining optical flow' ARTIFICIAL INTELLIGENCE vol. 17, pages 185 - 203, XP000195787 * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002077920A2 (en) * | 2001-03-26 | 2002-10-03 | Dynapel Systems, Inc. | Method and system for the estimation and compensation of brightness changes for optical flow calculations |
WO2002077920A3 (en) * | 2001-03-26 | 2003-09-18 | Dynapel Systems Inc | Method and system for the estimation and compensation of brightness changes for optical flow calculations |
US6959118B2 (en) | 2001-03-26 | 2005-10-25 | Dynapel Systems, Inc. | Method and system for the estimation and compensation of brightness changes for optical flow calculations |
WO2005006762A3 (en) * | 2003-07-02 | 2005-02-10 | Queen Mary & Westfield College | Optical flow estimation method |
US7822231B2 (en) | 2003-07-02 | 2010-10-26 | Queen Mary And Westfield College, University Of London | Optical flow estimation method |
US8170109B2 (en) | 2006-03-23 | 2012-05-01 | Nds Limited | System for analysis of motion |
Also Published As
Publication number | Publication date |
---|---|
WO2001096982A3 (en) | 2002-03-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109859126B (en) | Video noise reduction method and device, electronic equipment and storage medium | |
US8768069B2 (en) | Image enhancement apparatus and method | |
KR101498532B1 (en) | Digital processing method and system for determination of optical flow | |
Wu et al. | Image registration using wavelet-based motion model | |
JP5385969B2 (en) | Method and apparatus for super-resolution of images | |
Ghosal et al. | A fast scalable algorithm for discontinuous optical flow estimation | |
KR20190065287A (en) | Prediction of depth from image data using statistical model | |
Wu et al. | Optical flow estimation using wavelet motion model | |
EP0652536B1 (en) | Method and apparatus for estimating motion vector fields by rejecting local outliers | |
JP5107409B2 (en) | Motion detection method and filtering method using nonlinear smoothing of motion region | |
KR20080063770A (en) | Extracting a moving object boundary | |
US20110091074A1 (en) | Moving object detection method and moving object detection apparatus | |
CN108986150B (en) | Image optical flow estimation method and system based on non-rigid dense matching | |
US20120062548A1 (en) | Reducing viewing discomfort | |
US20090180032A1 (en) | Method and system for hierarchical motion estimation with multi-layer sub-pixel accuracy and motion vector smoothing | |
Adarve et al. | A filter formulation for computing real time optical flow | |
Whitaker | A level-set approach to image blending | |
Vazquez et al. | Joint multiregion segmentation and parametric estimation of image motion by basis function representation and level set evolution | |
CN113259605A (en) | Video matting method, system and storage medium based on prediction foreground mask prediction | |
Al Ismaeil et al. | Real-time enhancement of dynamic depth videos with non-rigid deformations | |
WO2001096982A2 (en) | System for the estimation of optical flow | |
Al-Regib et al. | Hierarchical motion estimation with content-based meshes | |
CN103618904B (en) | Motion estimation method and device based on pixels | |
CN113240705A (en) | 3D attitude estimation method and device, electronic equipment and storage medium | |
Nomura | Spatio-temporal optimization method for determining motion vector fields under non-stationary illumination |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): JP NO |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
AK | Designated states |
Kind code of ref document: A3 Designated state(s): JP NO |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR |
|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
122 | Ep: pct application non-entry in european phase | ||
NENP | Non-entry into the national phase in: |
Ref country code: JP |