WO2009024349A1 - Estimating objects proper motion using optical flow, kinematics and depth information - Google Patents

Estimating objects proper motion using optical flow, kinematics and depth information Download PDF

Info

Publication number
WO2009024349A1
WO2009024349A1 PCT/EP2008/006896 EP2008006896W WO2009024349A1 WO 2009024349 A1 WO2009024349 A1 WO 2009024349A1 EP 2008006896 W EP2008006896 W EP 2008006896W WO 2009024349 A1 WO2009024349 A1 WO 2009024349A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
motion
optical flow
flow
emf
Prior art date
Application number
PCT/EP2008/006896
Other languages
French (fr)
Inventor
Julian Eggert
Sven Rebhan
Volker Willert
Jens Schmüdderich
Original Assignee
Honda Research Institute Europe Gmbh
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honda Research Institute Europe Gmbh filed Critical Honda Research Institute Europe Gmbh
Priority to AT08785662T priority Critical patent/ATE500570T1/en
Priority to DE602008005318T priority patent/DE602008005318D1/en
Priority to EP08785662A priority patent/EP2179398B1/en
Priority to JP2010519390A priority patent/JP4967062B2/en
Priority to US12/599,201 priority patent/US8422741B2/en
Publication of WO2009024349A1 publication Critical patent/WO2009024349A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/269Analysis of motion using gradient-based methods

Definitions

  • the present invention relates to a method and a corresponding device for estimating the proper motion of an object (Object Proper Motion or OPM), using optical flow and kinematics and depth information, which is applicable in controlling autonomous mobile vehicles, such as robots or cars.
  • OPM Object Proper Motion
  • a robot's autonomous movement causes the camera to undergo translational and rotational movements in 3D with sudden velocity-changes.
  • OF optical flow
  • EM ego-motion
  • EMF ego-motion-flow
  • OPM object proper motion
  • a common procedure to decompose the optical flow (OF) and extract the desired object proper motion (OPM) tries to estimate the movement of the camera from the visual flow fields, where the underlying models for camera movement may highly differ in complexity.
  • optical flow is primarily caused by ego-motion (EM) -effects and not by independently moving objects - which can not be assumed for a robot facing moving people.
  • EM ego-motion
  • the effect of translational ego-motion (EM) on points in the image is highly influenced by their dis- tance to the camera in the way that distant points induce smaller flow vectors than closer ones.
  • One idea of using existing knowledge about the movement of the camera is as follows. Letting a robot walk in a circle, one may compute the optical flow (OF) resulting from the robot's movement over a textured ground. In conjunction with the robot's gait phase and joint angles, a neural network is trained to learn the ego-motion-flow (EMF). Afterwards differences between this ego-motion-flow (EMF) and the measured optical flow (OF) can be used to detect obstacles in the path of the robot. While the movement of the robot was very constricted, Fidelman et al. ("Detecting motion in the environment with a moving quadruped robot," in RoboCup-2006: Robot Soccer World Cup X, G. Lakemeyer, E. Sklar, D.
  • the region of interest is regarded as a plane with a uniform depth z (distance to the camera) that is calculated from a distance measurement of a laser scanner and is used for all the pixels in the region of interest.
  • this assumption about the depth is not applicable in all situations in which an accurate estimate of the proper motion of an object is desired. This effect becomes more significant for objects close to the camera and in front of a distant background.
  • Fig. 1 shows an overview of a system according to an embodiment of the pre- sent invention
  • Fig. 2a shows a point /5TM, in camera- and world coordinates
  • Fig. 2b shows the description of the static point at time n using the camera- transformations
  • Fig. 3 a shows the spectrum of optical flow (OF) -motion, if ego-motion (EM)- effects are not cancelled out in advance.
  • the solid line visualizes an assumed object proper motion (OPM) of two pixels, the dashed line an assumed ego-motion (EM)-effect of three pixels.
  • OPM assumed object proper motion
  • EM assumed ego-motion
  • the interval which has to be acquired by the optical flow (OF) is a superposition of the two.
  • Fig. 3b shows the reduced spectrum for searching relative to ego-motion-flow (EMF);
  • Fig. 4a shows the disrupted results of pixel wise warping under extreme body movements.
  • Fig. 4b shows the effect of filling-in holes in the ego-motion-flow (EMF) with the averaged flow;
  • Fig. 5 shows a one-dimensional plot of the resulting distribution N-(c, C) from
  • Figure 1 shows an overview of a system according to the embodiments of the present invention:
  • OEF optical flow
  • This step-by-step movement estimation allows the reduction of the optical flow's (OFs) search-range which does not only reduce computational time but also decreases possible ambiguities in the optical flow (OF)-measurement and thereby leads to qualitatively better results.
  • the optical flow (OF) measured in this way should be zero for non-moving objects and otherwise describe their proper motion.
  • each point (xTM ,yTM) may be combined in the image with its binocular disparity dTM , to define a point in homogenous image-coordinates as
  • the way the camera moved has to be known.
  • This knowledge may be obtained from the robot kinematics.
  • the robot's state vector s m which contains the position of the upper body as well as the angles of all joints at time m is used.
  • this vector one may compute the forward kinematics, returning the transformations from the initial world-coordinate system to each robot segment, and including the transformation from world- to camera-coordinates T ⁇ w ⁇
  • T ⁇ c which describes the transformation from camera- to world-coordinates.
  • T ⁇ w can be computed by applying s" . Since this calculation includes the position of the upper body, the resulting transformations are not only rotation-matrices but also encode translational changes of the camera-position. Assuming a static point pTM, , the new position p C n l can be calculated from transferring the point to world-coordinates by multiplying the last transformation T ⁇ ⁇ _ c and passing it back to camera-coordinates with the actual transformation T ⁇ w :
  • each point may be projected back into the image- plane. This can be done by computing a homogenous projection-matrix T l ⁇ _ c and multiplying the new point p c " , with it:
  • b denotes the baseline, /the focal-length and q the pixel-size of the camera.
  • w e may be used to express a weighting of the whole variance which is beneficial for the comparison of this variance with the one from the optical flow (OF) (see below). Additionally w d e [ ⁇ ,l] may be utilized to stress the influence of either confidence or distance.
  • the ego-motion (EM)-effects may be compensated by warping the images. In the present case, warping using a forward mapping can be written as
  • the average ego -motion-flow (EMF) v" gained from valid depth-estimations may be used to fill the gaps (see Figure 4(b)).
  • a velocity estimate F 1 1 " may be obtained for each point, relative to the estimated ego-motion-flow (EMF).
  • the algorithm may also compute a covariance ⁇ TM which gives a confidence measure for the optical flow vectors, assuming pixelwise independent Gaussian noise.
  • the proposed invention also allows prediction over time, giving it the ability to iteratively make reliable calculations of motion in unstructured image regions by drawing previous estimations into account.
  • the ideal outcome of the flow-algorithm would be a vector-field, which is zero for non-moving objects and denotes their proper motion otherwise.
  • OPF optical flow
  • depth-measurements from disparity are very noisy signals, some more filtering is needed for the detection of object proper motion (OPM).
  • L 1 N 1 (P 1 +e,, ⁇ n + ⁇ e ⁇ )
  • the decision whether some optically measured velocity is classified as object proper motion is based on the absolute value of that velocity scaled by the variances of the ego-motion-flow (EMF) and the optical flow (OF).
  • EMF ego-motion-flow
  • OF optical flow
  • o may be used to extract regions of the visual input where reliably measurable object proper motion (OPM) occurs.
  • OPM object proper motion
  • the present invention provides a method and a system being capable of perceiving object proper motion (OPM) from a moving platform.
  • OPM object proper motion
  • the step-by-step movement estimation including the compensation of ego-motion (EM) prior to the optical flow (OF)- computation seems to be a central aspect for the robustness of the system against firm ego-motion (EM) of the robot.
  • Robustness against noise in the depth and flow- estimation results from the probabilistic rejection mechanism, which neglects velocities based on their size and reliability.
  • the proper object motion computed by this method and system may be used for the attraction of visual attention as well as real-time object interaction. More particularly, an autonomous robot or a car may be controlled using the proper motion of objects, as determined by the invention, e.g. for controlling their own movement or for tracking a moving object by keeping it centred in the image detector's view.

Abstract

The invention relates to a method for detecting the proper motion of an real-world object, comprising the steps of acquiring, by an image capturing device (ICD), a first image (I1) of the object at a first point in time (t1) and a second image (I2) at a second point in time (t2); obtaining a third (hypothetical) image (I3), based on an estimated effect of the motion of the image capturing device (ICD) itself (EMF) between the first and the second point in time (t1, t2), wherein the effect of the motion of the image capturing device (ICD) itself is estimated based on the forward kinematics of the image capturing device; determining an optical flow (OF) between the second image (I2) and the third image (I3); and evaluating the optical flow (OF) by incorporating uncertainties of the optical flow (OF) and the ego-motion-flow (EMF) in order to determine the proper motion of the object.

Description

ESTIMATING OBJECTS PROPER MOTION USING OPTICAL FLOW, KINEMATICS AND
DEPTH INFORMATION
The present invention relates to a method and a corresponding device for estimating the proper motion of an object (Object Proper Motion or OPM), using optical flow and kinematics and depth information, which is applicable in controlling autonomous mobile vehicles, such as robots or cars.
TECHNICAL BACKGROUND
The ability to visually perceive motion is believed to be highly beneficial for surviving in a dynamic environment. So it is not surprising to see that movement is one of the most important cues to attract visual attention. Interestingly, most mammals are able to perceive visual motion even if they are moving themselves - either by rotating the head, the eyes, moving the whole body or even while they are running. The gathered information is then used for controlling the own movement or for tracking moving objects by keeping them centred in the fovea.
In general, a robot's autonomous movement causes the camera to undergo translational and rotational movements in 3D with sudden velocity-changes.
The effects of Ego-Motion (EM) on the Optical Flow (OF), the so called ego-motion- flow (EMF), produced by a robot make this task quite challenging. Primarily, this is caused by the large number of degrees of freedom and the complex influence of each robot segment on the position of the camera: For example, bending the knee may in some situations cause the robot and hence the camera to shift and rotate violently to the side, whereas in other situations, the robot might be standing on the other leg and so knee bending does not affect the camera position at all. The domain where the estimation of object proper motion (OPM) during ego-motion (EM) is commonly addressed is the car domain. Here, this estimation is important to measure the speed of other cars relative to the observing car which allows identifying them as eventual obstacles. Most approaches rely on the calculation of optical flow (OF). This flow is a superposition of the effects caused by ego-motion (EM), the so called ego-motion-flow (EMF), and object proper motion (OPM). A common procedure to decompose the optical flow (OF) and extract the desired object proper motion (OPM) tries to estimate the movement of the camera from the visual flow fields, where the underlying models for camera movement may highly differ in complexity.
For proper extraction of the camera movement it is hence crucial that the optical flow (OF) is primarily caused by ego-motion (EM) -effects and not by independently moving objects - which can not be assumed for a robot facing moving people. The effect of translational ego-motion (EM) on points in the image is highly influenced by their dis- tance to the camera in the way that distant points induce smaller flow vectors than closer ones.
Some of the prior art approaches actually measure the distance of these points. However the majority of methods assume that all points lie on a plane running through the position of the camera and the focus of expansion at the horizon. While this procedure may be suitable for the car domain, no such simplifying assumption about the environment is appropriate for autonomously acting robots in dynamic scenes. Thus, different ways to handle ego-motion (EM) have to be found.
One idea of using existing knowledge about the movement of the camera is as follows. Letting a robot walk in a circle, one may compute the optical flow (OF) resulting from the robot's movement over a textured ground. In conjunction with the robot's gait phase and joint angles, a neural network is trained to learn the ego-motion-flow (EMF). Afterwards differences between this ego-motion-flow (EMF) and the measured optical flow (OF) can be used to detect obstacles in the path of the robot. While the movement of the robot was very constricted, Fidelman et al. ("Detecting motion in the environment with a moving quadruped robot," in RoboCup-2006: Robot Soccer World Cup X, G. Lakemeyer, E. Sklar, D. Sorenti, and T. Takahashi, Eds. Berlin: Springer Verlag, 2007, vol. 4434/2007, pp. 219-231) proved that ego-motion-flow (EMF) can be learned even for more complex movements. In their approach the neural network is provided with the recent optical flow (OF)-calculation in addition to walk phase and joint angles of a robot. The neural net predicts the flow of the next timestep, allowing to compare it with the actually measured optical flow (OF) in order to classify differences as object proper motion (OPM).
The idea of calculating the movement of the camera and using it to compute the ego- motion-flow (EMF) has also been described for the car. In US patent application 2004/0183905 (Comaniciu, D. et al.), readings from gyroscopic sensors or GPS-signals are applied to estimate the camera-movement. Similar to other approaches in cars, the depth is not measured but estimated from the plane assumption as described earlier. Another approach actually computes the ego-motion-flow (EMF) depending on the measured robot joints (del Solar J.R. et al.)( "Motion detection and tracking for an Aibo robot using camera motion compensation and Kalman filtering." in RoboCup 2004: Robot Soccer World Cup VIII, ser. Lecture Notes in Computer Science, D. Nardi, M. Riedmiller, C. Sammut, and J. Santos- Victor, Eds. Berlin: Springer, March 2005, vol. 3276, pp. 619-627) instead of using neural networks.
However, all the results of the provided approaches are not very promising, possibly due to a missing integration of depth information. Fidelman et al. ("Detecting motion in the world with a moving quadruped robot." P. Fidelman, T. Coffman, R. Miiku- lainen, and P. Stone. Technical Report TR-05-37, Department of Computer Science, University of Texas, 2005 ) implicitly include depth information by providing the cal- culated optical flow (OF)-field, which depends on the distance of points to the camera. Nevertheless, this approach can not cope properly with objects having a distance different from objects in the training-phase: The difference in distance results in an optical flow (OF) deviating from the learned one. Hence even non-moving points can be classified as object proper motion (OPM). The combination of knowledge about the robot movement with distance information and optical flow (OF)-measurements is described in a paper by Overett, G. et al. (G. Overett and D. Austin, "Stereo vision motion detection from a moving platform," in Proc. Australasian Conf. on Robotics and Automation, December 2004). Having cameras fixed to the robot, Overett et al. try to measure the odometry data and compute the resulting ego-motion-flow (EMF) considering the depth. This flow is afterwards subtracted from the estimated optical flow (OF) and residual vectors are used to indicate object proper motion (OPM). Unfortunately, the noisy data from odometry forces the authors to manually measure the dis- tance passed by the robot, preventing the system from running in real time.
None of the presented methods precomputes a movement of the camera and searches for movement relative to it. When acquiring the effects of camera movement that is superimposed with the effects of independently moving objects, the optical flow (OF) needs a large set of displacements. The higher range of acquired motion does not only increase the computational time but also corrupts optical flow (OF)-estimations by producing increased ambiguities. Very recently Rabe et al. ("Fast detection of moving objects in complex scenarios," in Proc. IEEE Symp. Intelligent Vehicles, June 2007, pp. 398-403) proposed a solution to this problem by computing flow and disparity from visual features tracked in the image instead of estimating dense depth- and flow- fields. Fusing the results in a Kalman filter provides 6D information about location and motion for a set of points. Together with readings from inertial sensors this information is used to compensate ego-motion (EM)-effects and obtain the actual 3D object proper motion (OPM). The results are very convincing, at least for the moderate movement of a car. However, apparently no procedure was presented which is able to handle the noisy character of depth, optical flow (OF) and the fierce effects of camera- movement induced by a legged robot, combining them to make reliable estimations of object proper motion (OPM).
Fardi et al. ("Motion-based pedestrian recognition from a moving vehicle," in Proc. IEEE Symp. Intelligent Vehicles, June 2006, pp. 219-224) present a method for recognizing pedestrians from a moving vehicle comprising the computation of an optical flow, including the compensation of the ego motion of the camera.
In Fardi et al, the region of interest is regarded as a plane with a uniform depth z (distance to the camera) that is calculated from a distance measurement of a laser scanner and is used for all the pixels in the region of interest. However, this assumption about the depth is not applicable in all situations in which an accurate estimate of the proper motion of an object is desired. This effect becomes more significant for objects close to the camera and in front of a distant background.
It is therefore an object of the present invention, to provide a method and device for estimating the proper motion of objects by handling depth information more precisely, being capable to provide and take advantage of the confidence of the OPM estimation for every pixel in an image sequence.
SHORT SUMMARY OF THE INVENTION This object is achieved according to the invention by a method and a system according to the independent claims. Advantageous embodiments are defined in the dependent claims.
SHORT DESCRIPTION OF THE FIGURES More aspects and advantages of the present invention will become more apparent when the following detailed description of the invention is studied, in connection with the annexed drawing, in which
Fig. 1 shows an overview of a system according to an embodiment of the pre- sent invention;
Fig. 2a shows a point /5™, in camera- and world coordinates;
Fig. 2b shows the description of the static point at time n using the camera- transformations;
Fig. 3 a shows the spectrum of optical flow (OF) -motion, if ego-motion (EM)- effects are not cancelled out in advance. The solid line visualizes an assumed object proper motion (OPM) of two pixels, the dashed line an assumed ego-motion (EM)-effect of three pixels. The interval which has to be acquired by the optical flow (OF) is a superposition of the two.
Fig. 3b shows the reduced spectrum for searching relative to ego-motion-flow (EMF);
Fig. 4a shows the disrupted results of pixel wise warping under extreme body movements. Fig. 4b shows the effect of filling-in holes in the ego-motion-flow (EMF) with the averaged flow; Fig. 5 shows a one-dimensional plot of the resulting distribution N-(c, C) from
Nv {rι + el,∑r) and N (e,. , ∑e ) as well as the likelihood L;
DETAILED DESCRIPTION OF THE INVENTION
Figure 1 shows an overview of a system according to the embodiments of the present invention:
In a first step the forward kinematics are used to compute the movement of the camera occurring in the time-interval from m to n = m +1. Combining this with the depth information from binocular disparity, one may estimate where a static point in the image at time m moved due to ego-motion (EM) and obtain the ego-motion-flow (EMF). Hence, by knowing where a point in the image at time n originated from, the ego- motion (EM)-effect may be compensated by moving the point back to its original position in a second step, resulting in an image freed from EM-effects. Afterwards this image may be used for a calculation of the optical flow (OF) measured relative to the ego-motion-flow (EMF). This step-by-step movement estimation allows the reduction of the optical flow's (OFs) search-range which does not only reduce computational time but also decreases possible ambiguities in the optical flow (OF)-measurement and thereby leads to qualitatively better results. Ideally, the optical flow (OF) measured in this way should be zero for non-moving objects and otherwise describe their proper motion.
However, since binocular disparity as well as optical flow (OF) are particularly noisy signals, suitable methods may be applied to filter out wrong velocity-estimations based on the reliability of flow-and ego-motion (EM)-vector.
COMPUTATION OF EGO-MOTION-FLOW (EMF)
The computation of ego-motion-flow (EMF) results in a flow field Em = {e™ } , which describes where the points in image Im have shifted to in I", caused by movement of the camera. For this computation it is assumed that the environment is static and the points did not move themselves.
To account for the fact that the absolute values of the vectors e™ highly depend on the distance of the corresponding points to the camera, the effect of camera movement may be calculated in a three dimensional camera-related space. Therefore, first each point (x™ ,y™) may be combined in the image with its binocular disparity d™ , to define a point in homogenous image-coordinates as
This point can then be reconstructed in camera-coordinates by computing the homogenous transformation matrix TI<_C from camera-to image-coordinates and by multiply
Figure imgf000008_0001
Pa = 1C-I - PiA (2)
To compute the position of each such point relative to the camera at the next time step, the way the camera moved has to be known. This knowledge may be obtained from the robot kinematics. For this purpose, the robot's state vector sm , which contains the position of the upper body as well as the angles of all joints at time m is used. With this vector, one may compute the forward kinematics, returning the transformations from the initial world-coordinate system to each robot segment, and including the transformation from world- to camera-coordinates T^w
Inverting this transformation results in T^c , which describes the transformation from camera- to world-coordinates. In addition, the current transformation T^w can be computed by applying s" . Since this calculation includes the position of the upper body, the resulting transformations are not only rotation-matrices but also encode translational changes of the camera-position. Assuming a static point p™, , the new position pC n l can be calculated from transferring the point to world-coordinates by multiplying the last transformation T^<_c and passing it back to camera-coordinates with the actual transformation T^w :
Figure imgf000009_0001
This process is illustrated in figure 2.
Since one is not interested in the new three dimensional position of the points but rather in the 2D-flow in the image, each point may be projected back into the image- plane. This can be done by computing a homogenous projection-matrix Tl<_c and multiplying the new point pc" , with it:
Figure imgf000009_0002
Due to the constancy of rcw , T^c , and Tl4r_c within a timestep, the concatenation of these transformations may be pre-computed, resulting in one single transformation matrix which may be multiplied with each point /J™, .
Knowing p" t and Pf1 the shift of a point in the image-plane due to camera-motion is given by
Tl _ TO
→m _ I Jxv i I „.« ,,m
To express the reliability of this shift-vector, only the quality of the measured depth z,m of ft™, has to be considered, since the computation of the kinematics is deterministic. The reliability of e™ relates to the amount of correlation found by the disparity- algorithm between the left and the right camera-image. This is expressed in terms of some confidence-value c e [θ,l] . In addition the falling accuracy with increasing distance may be taken into account.
The sensitivity for depth estimation from disparity decreases with the squared distance:
Figure imgf000010_0001
where b denotes the baseline, /the focal-length and q the pixel-size of the camera.
Since one wants to describe the likelihood for e™ in terms of some probability distribution, one may express a decay of reliability in terms of an increasing variance Σ™ by declaring it as
Figure imgf000010_0002
with 1 being the unity-matrix. The scalar we may be used to express a weighting of the whole variance which is beneficial for the comparison of this variance with the one from the optical flow (OF) (see below). Additionally wd e [θ,l] may be utilized to stress the influence of either confidence or distance.
COMPUTATION OF FLOW RELATIVE TO EGO-MOTION-FLOW (EMF)
Since the effects of ego-motion (EM) have been computed on the image, one is now able to compute the optical flow (OF) Rm = {ηm} between Im and I" relative to this ego- motion-flow (EMF). In comparison to other approaches which cancel out ego-motion (EM)-effects after the computation of the optical flow (OF), this reduces the spectrum of the motion to be acquired. This effect which is illustrated in Figure 3 does not only reduce computational time but also improves the optical flow (OF) -estimations by reducing ambiguities. The ego-motion (EM)-effects may be compensated by warping the images. In the present case, warping using a forward mapping can be written as
1 KPl. i + ei I ~ 1 KP I, i) (8)
where I m equals the old image Im but is freed from the measured ego-motion (EM)- effects. This kind of forward mapping has two major drawbacks: Since e™ usually encodes real-valued shifts, the data-points pjt + ej" may not lie inside the grid and require complicated interpolations. Additionally, it is not guaranteed that each point in the warped image is targeted by the sum of original position and shift, leading to holes in the image.
The more flexible solution lies in the usage of backward-mapping to warp the actual image /"back to 7" :
KPr.i) = 1 (Vi.i + ei ) (9)
Since data points are now passed as arguments for the resulting image / " , holes in the image can not occur. The problem of real-valued shifts may be tackled by using bilin- ear interpolation in the source-image. Apart from deviations caused by wrong depth measurements, / " and Im should only differ in points with individual object motion. Nevertheless, in some situations with extreme body-movement, holes in the depth image Dm can cause artefacts in the pixel wise warped image, as shown in Figure 4(a). These artefacts lead to errors in the computed optical flow (OF) and can hence affect surrounding regions, even if those regions have valid depth estimations. In a first step, these artefacts may be reduced by filling invalid regions with suitable assumptions obtained from valid ego-motion-flow (EMF)-estimations.
In one embodiment of the present invention, the average ego -motion-flow (EMF) v" gained from valid depth-estimations may be used to fill the gaps (see Figure 4(b)).
To account for the decreased reliability of the warped image and hence of the optical flow (OF), one may create a penalty-map P which is high for points near and inside invalid depth-regions and 0 otherwise.
By passing /m and / " to an optical flow (OF)-algorithm, a velocity estimate F1 1" may be obtained for each point, relative to the estimated ego-motion-flow (EMF). The algorithm may also compute a covariance Σ™ which gives a confidence measure for the optical flow vectors, assuming pixelwise independent Gaussian noise.
For the inclusion of the warping-based penalty, the compound variance Σ™ at point ^7» rnay be calculated as
m η = i - P(p?,) + ∑Z, (10)
For invalid points the choice of 1- p™, » Σ™ ensures a negligible influence of the con- fidence-based variance. This is useful since artefacts in the warped image usually increase the structure at that specific region and thereby lead to high confidence- measures of the optical flow (OF)-algorithm.
Besides the spatial integration, the proposed invention also allows prediction over time, giving it the ability to iteratively make reliable calculations of motion in unstructured image regions by drawing previous estimations into account. The ideal outcome of the flow-algorithm would be a vector-field, which is zero for non-moving objects and denotes their proper motion otherwise. However, since the optical flow (OF) as well as depth-measurements from disparity are very noisy signals, some more filtering is needed for the detection of object proper motion (OPM).
DETECTING OBJECT PROPER MOTION (OPM)
In the previous sections, two approaches for the calculation of image flow were introduced, which use very different methods and hence show different characteristics. Since the computation of the ego-motion-flow (EMF) is based on disparity and kinematics, it can acquire the effects of ego-motion (EM) on image-points, as long as the points are not moving. In contrast, the optical flow (OF) also works for moving points. By adding the ego-motion-flow (EMF) to the relative optical flow (OF), an overall flow may be obtained, which designates the compound retinal movement of object proper motion (OPM) and ego-motion (EM)-effects. For non-moving points, this flow should equal the ego-motion-flow (EMF), while it should be different for moving points. A measurement for the significance of this distance may be introduced and used to extract object proper motion (OPM) -vectors from the optical flow (OF).
This measurement is derived from a stochastic assumption about the estimated flows, the depth and the images. Therefore, some stochastic variables have to be defined for each point p} , in the image a. Each vector of the overall flow mentioned above is represented by the variable S1 while εt describes the ego-motion (EM)-vector for each point pl t . The random variables I = {Im,In} and D = Dm specify the observed source- and depth-images.
The principal idea for the approach is to estimate the probability of measuring the same velocity v, from the ego-motion-flow (EMF) and from the compound optical flow (OF), assuming the corresponding point is static. That is, if some point did not move, the velocity described by εt should not differ too much from S1 and the likeli- hood to measure some identical velocity Vj from both methods should be high. In contrast, a moving point results in different outcomes for εt and S1 and the likelihood to measure the same velocity Vj from the two methods is very low. Concluding, a high joint probability p(S, = Vj, ε, = vi} I, D) indicates a static point whereas a low probability indicates a moving point.
Reflecting the assumptions about the dependencies of ε, , S1 , O and I, this joint distribution can be decomposed into
p(>% = ft, e.
Figure imgf000013_0001
p(I) p{D) ,χ „ Since no prior assumptions are made about the source images and the depth, the corresponding variables are uniformly distributed and hence have no influence on the distribution. Using the precomputed results for the ego-motion-flow (EMF) and optical flow (OF), the conditional distributions may be approximated with Gaussians N(P, Σ) with mean and variance parameters, so that
p(3l =vl\ε,,I)∞NVi(rl+elr) (12) p(ε, = v, I D) ∞N(e,,Σe) (13)
Adding et in (12) accounts for the warping of the image - while P1 encodes a flow relative to the ego-motion-flow (EMF), the sum of e, and P1 makes it an absolute flow and allows the comparison with the ego-motion-flow (EMF). With this approximation the joint distribution is proportional to the product of two Gaussians, being defined as
N(Pl+el,∑r)N(e,Xe) = Lι-N(cl,Cι) with: c, = Z15(Z11eχιel +(∑eχ\Pι +e,) (14)
C, =∑(∑ney∑ (15)
L1 = N1(P1 +e,,∑n+∑)
This procedure is identical for each time step, so the time indices have been dropped for convenience.
A visualization of this product is shown in figure 5. The mean value c of the resulting distribution can be interpreted as that identical velocity which is most likely to be measured by both algorithms - the ego-motion-flow (EMF) and the optical flow (OF).
While it is always possible to find such a value which fits best the hypothesis of a common velocity the factor L is a measure to describe how well c actually fits in the light of the calculated displacements and variances. For the evaluation of object proper motion (OPM), one is not interested in the value of the vector c , but rather whether such a vector is likely to occur. So the rejection of optical flow (OF)-estimations may be based on this value L, and can be further simplified by applying the logarithm to L:
L, = Ne-ι {rl + et ,∑rι + ) (16)
= z -e 42(f '-(? ' +? ' ))r(Σ " " rl(^-(^+5')) ( „1„7)
∞ -? (Ln +Σ J'1?, (18)
That is, the decision whether some optically measured velocity is classified as object proper motion (OPM) is based on the absolute value of that velocity scaled by the variances of the ego-motion-flow (EMF) and the optical flow (OF). Defining rt τ(∑r + ∑e)'1/; as Mahalanobis norm ψtjM , a big Mahalanobis norm indicates a moving point, so one may rely on the computation of the optical flow (OF). Formally, this can be expressed as
Figure imgf000015_0001
In subsequent processing stages, o, may be used to extract regions of the visual input where reliably measurable object proper motion (OPM) occurs.
CONCLUSION
The present invention provides a method and a system being capable of perceiving object proper motion (OPM) from a moving platform. The step-by-step movement estimation including the compensation of ego-motion (EM) prior to the optical flow (OF)- computation seems to be a central aspect for the robustness of the system against firm ego-motion (EM) of the robot. Robustness against noise in the depth and flow- estimation results from the probabilistic rejection mechanism, which neglects velocities based on their size and reliability. The proper object motion computed by this method and system may be used for the attraction of visual attention as well as real-time object interaction. More particularly, an autonomous robot or a car may be controlled using the proper motion of objects, as determined by the invention, e.g. for controlling their own movement or for tracking a moving object by keeping it centred in the image detector's view.

Claims

1. Method for detecting the proper motion of an real- world object, comprising the steps of:
- acquiring, by an image capturing device (ICD), a first image (Ij) of the object at a first point in time (ti) and a second image (I2) at a second point in time (t2);
- obtaining a third (hypothetical) image (I3), based on an estimated ef- feet of the motion of the image capturing device (ICD) itself (EMF) between the first and the second point in time (tj, t2), wherein the effect of the motion of the image capturing device (ICD) itself is estimated based on the forward kinematics of the image capturing device; - determining an optical flow (OF) between the second image (I2) and the third image (I3);
- evaluating the optical flow (OF) by incorporating uncertainties of the optical flow (OF) and the ego-motion-flow (EMF) in order to determine the proper motion of the object.
2. The method according to claim 1, wherein the estimate of the effect of the motion (EMF) of the image capturing device (ICD) itself is based on estimated depth information for each pixel of the first image (I1).
3. The method according to claim 2, wherein the depth information for the first image (I1) is estimated based on binocular disparity.
4. The method according to claim 3, wherein an invalid estimated depth is replaced by a depth model gained from valid estimated depths.
5. Method according to claim 1-4, wherein the step of obtaining the third image comprises warping the first image (Ii).
6. Method according to claim 1-4, wherein the step of obtaining the third image comprises warping the second image (I2).
7. Method according to claim 1-6, wherein the step of evaluating the opti- cal flow (OF) incorporating uncertainties of the optical flow (OF) and the ego-motion-flow (EMF) comprises determining the likelihood that the optical flow (OF) and the estimated effect (EMF) are consistent with the same physical velocity.
8. Use of a method according to one of the preceding claims for controlling a mobile autonomous vehicle.
9. Use according to claim 8, wherein the mobile autonomous vehicle is an autonomous robot.
10. System for detecting the proper motion of an object, comprising:
- means for acquiring, by an image capturing device (ICD), a first image (I1) of the object at a first point in time (tj) and a second image (I2) at a second point in time (t2); - means for obtaining a third (hypothetical) image (I3), based on an estimated effect of the motion of the image capturing device (ICD) itself (EMF) between the first and the second point in time (tj, t2), wherein the effect of the motion of the image capturing device (ICD) itself is estimated based on the forward kinematics of the im- age capturing device;
- means for determining an optical flow (OF) between the second image (I2) and the third image (I3);
- means for evaluating the optical flow (OF) by incorporating uncertainties of the optical flow (OF) and the ego-motion-flow (EMF) in order to determine the proper motion of the object,
PCT/EP2008/006896 2007-08-22 2008-08-21 Estimating objects proper motion using optical flow, kinematics and depth information WO2009024349A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
AT08785662T ATE500570T1 (en) 2007-08-22 2008-08-21 ESTIMATING THE PROPER MOTION OF OBJECTS USING OPTICAL FLOW, KINEMATICS AND DEPTH INFORMATION
DE602008005318T DE602008005318D1 (en) 2007-08-22 2008-08-21 ESTIMATE THE PROPER MOVEMENT OF OBJECTS USING OPTICAL RIVER, KINEMATICS AND DEEP INFORMATION
EP08785662A EP2179398B1 (en) 2007-08-22 2008-08-21 Estimating objects proper motion using optical flow, kinematics and depth information
JP2010519390A JP4967062B2 (en) 2007-08-22 2008-08-21 A method to estimate the appropriate motion of an object using optical flow, kinematics and depth information
US12/599,201 US8422741B2 (en) 2007-08-22 2008-08-21 Estimating objects proper motion using optical flow, kinematics and depth information

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP07114793.8 2007-08-22
EP07114793 2007-08-22

Publications (1)

Publication Number Publication Date
WO2009024349A1 true WO2009024349A1 (en) 2009-02-26

Family

ID=40254410

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2008/006896 WO2009024349A1 (en) 2007-08-22 2008-08-21 Estimating objects proper motion using optical flow, kinematics and depth information

Country Status (6)

Country Link
US (1) US8422741B2 (en)
EP (1) EP2179398B1 (en)
JP (1) JP4967062B2 (en)
AT (1) ATE500570T1 (en)
DE (1) DE602008005318D1 (en)
WO (1) WO2009024349A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100183192A1 (en) * 2009-01-16 2010-07-22 Honda Research Institute Europe Gmbh System and method for object motion detection based on multiple 3d warping and vehicle equipped with such system
EP2256690A1 (en) * 2009-05-29 2010-12-01 Honda Research Institute Europe GmbH Object motion detection system based on combining 3D warping techniques and a proper object motion detection
US20100315505A1 (en) * 2009-05-29 2010-12-16 Honda Research Institute Europe Gmbh Object motion detection system based on combining 3d warping techniques and a proper object motion detection
CN102413756A (en) * 2009-04-29 2012-04-11 皇家飞利浦电子股份有限公司 Real-time depth estimation from monocular endoscope images
JP2012084140A (en) * 2010-09-30 2012-04-26 Toshiba Corp Video analysis method and system
WO2013006067A1 (en) * 2011-07-04 2013-01-10 Streeter Lee Vincent Motion compensation in range imaging
TWI425445B (en) * 2009-09-29 2014-02-01 Nat Univ Tsing Hua Method and detecting system for determining quantity of self-motion of a moving platform
DE102014202639A1 (en) 2013-02-15 2014-08-21 Robert Bosch Gmbh Method for determining movement of vehicle, involves outputting movement of vehicle when quality value formed for evaluating detected vehicle movement estimate is available in predetermined relationship to reference value
CN109903315A (en) * 2019-03-08 2019-06-18 腾讯科技(深圳)有限公司 Method, apparatus, equipment and readable storage medium storing program for executing for light stream prediction
EP3809687A1 (en) * 2019-10-15 2021-04-21 Rohde & Schwarz GmbH & Co. KG Method and system for real time video stabilization

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100246899A1 (en) * 2009-03-26 2010-09-30 Rifai Khalid El Method and Apparatus for Dynamic Estimation of Feature Depth Using Calibrated Moving Camera
JP5400718B2 (en) * 2010-07-12 2014-01-29 株式会社日立国際電気 Monitoring system and monitoring method
US10776635B2 (en) 2010-09-21 2020-09-15 Mobileye Vision Technologies Ltd. Monocular cued detection of three-dimensional structures from depth images
TR201901893T4 (en) * 2011-10-14 2019-03-21 Ecole Polytechnique Fed Lausanne Epfl Nano scale motion detector.
DE102012025463A1 (en) * 2012-12-28 2014-07-03 Connaught Electronics Ltd. Method for determining e.g. pitch angle change of passenger car to determine proper motion of car in six different degrees of freedom, involves determining parameter based on pair of image features and associated image features of images
EP2757527B1 (en) * 2013-01-16 2018-12-12 Honda Research Institute Europe GmbH System and method for distorted camera image correction
CN103745458B (en) * 2013-12-26 2015-07-29 华中科技大学 A kind of space target rotating axle based on binocular light flow of robust and mass center estimation method
TWI572846B (en) * 2015-09-18 2017-03-01 國立交通大學 3d depth estimation system and 3d depth estimation method with omni-directional images
US9794588B2 (en) 2015-09-30 2017-10-17 Sony Corporation Image processing system with optical flow recovery mechanism and method of operation thereof
KR102507248B1 (en) 2015-12-02 2023-03-06 에스케이하이닉스 주식회사 Egomotion estimation system and method thereof
US9902071B2 (en) 2015-12-18 2018-02-27 General Electric Company Control system and method for brake bleeding
CN107305635A (en) * 2016-04-15 2017-10-31 株式会社理光 Object identifying method, object recognition equipment and classifier training method
US10419669B2 (en) * 2017-01-17 2019-09-17 Disney Enterprises, Inc. Omnistereoscopic panoramic video
US20180293735A1 (en) * 2017-04-11 2018-10-11 Sony Corporation Optical flow and sensor input based background subtraction in video content
IT201700058961A1 (en) * 2017-05-30 2018-11-30 Artglass S R L METHOD AND SYSTEM OF FRUITION OF AN EDITORIAL CONTENT IN A PREFERABLY CULTURAL, ARTISTIC OR LANDSCAPE OR NATURALISTIC OR EXHIBITION OR EXHIBITION SITE
US10262224B1 (en) 2017-07-19 2019-04-16 The United States Of America As Represented By Secretary Of The Navy Optical flow estimation using a neural network and egomotion optimization
CN111386550A (en) * 2017-11-15 2020-07-07 谷歌有限责任公司 Unsupervised learning of image depth and ego-motion predictive neural networks
KR102559203B1 (en) * 2018-10-01 2023-07-25 삼성전자주식회사 Method and apparatus of outputting pose information
US11694432B2 (en) * 2019-07-23 2023-07-04 Toyota Research Institute, Inc. System and method for augmenting a visual output from a robotic device

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6269175B1 (en) * 1998-08-28 2001-07-31 Sarnoff Corporation Method and apparatus for enhancing regions of aligned images using flow estimation
JP2001084383A (en) * 1999-09-09 2001-03-30 Univ Tokyo Movement detection method
WO2001039120A2 (en) * 1999-11-26 2001-05-31 Mobileye, Inc. System and method for estimating ego-motion of a moving vehicle using successive images recorded along the vehicle's path of motion
US20060020562A1 (en) * 2004-07-21 2006-01-26 University Of Southern Mississippi Apparatus and method for estimating optical flow
JP4539291B2 (en) * 2004-10-29 2010-09-08 ソニー株式会社 Robot apparatus and operation control method thereof
US7894528B2 (en) * 2005-05-25 2011-02-22 Yissum Research Development Company Of The Hebrew University Of Jerusalem Fast and robust motion computations using direct methods
JP2007066094A (en) * 2005-08-31 2007-03-15 Matsushita Electric Ind Co Ltd Posture estimation device and posture estimation method
JP4676373B2 (en) * 2006-04-27 2011-04-27 株式会社デンソー Peripheral recognition device, peripheral recognition method, and program
US8855848B2 (en) * 2007-06-05 2014-10-07 GM Global Technology Operations LLC Radar, lidar and camera enhanced methods for vehicle dynamics estimation
US20080319664A1 (en) * 2007-06-25 2008-12-25 Tidex Systems Ltd. Navigation aid

Non-Patent Citations (9)

* Cited by examiner, † Cited by third party
Title
ANDREA GIACHETTI ET AL: "The Use of Optical Flow for Road Navigation", IEEE TRANSACTIONS ON ROBOTICS AND AUTOMATION, IEEE INC, NEW YORK, US, vol. 14, no. 1, 1 February 1998 (1998-02-01), XP011053264, ISSN: 1042-296X *
DEMIRDJIAN D ET AL: "Motion-Egomotion Discrimination and Motion Segmentation from Image-Pair Streams", COMPUTER VISION AND IMAGE UNDERSTANDING, ACADEMIC PRESS, US, vol. 78, no. 1, 1 April 2000 (2000-04-01), pages 53 - 68, XP004439286, ISSN: 1077-3142 *
FARDI B ET AL: "Motion-based pedestrian recognition from a moving vehicle", INTELLIGENT VEHICLES SYMPOSIUM, 2006 IEEE MEGURO-KU, JAPAN 13-15 JUNE 2006, PISCATAWAY, NJ, USA,IEEE, 13 June 2006 (2006-06-13), pages 219 - 224, XP010937016, ISBN: 978-4-901122-86-3 *
J. HORN, STEREO VISION BASED EGO-MOTION ESTIMATION WITH SENSOR SUPPORTED SUBSET VALIDATION
JAN HORN ET AL: "Stereo Vision Based Ego-Motion Estimation with Sensor Supported Subset Validation", INTELLIGENT VEHICLES SYMPOSIUM, 2007 IEEE, IEEE, PI, 1 June 2007 (2007-06-01), pages 741 - 748, XP031127035, ISBN: 978-1-4244-1067-5 *
KOLODKO J ET AL: "Motion estimation hardware for autonomous vehicle guidance", INDUSTRIAL ELECTRONICS SOCIETY, 2000. IECON 2000. 26TH ANNUAL CONFJERE NCE OF THE IEEE 22-28 OCT. 2000, PISCATAWAY, NJ, USA,IEEE, vol. 2, 22 October 2000 (2000-10-22), pages 930 - 935, XP010569305, ISBN: 978-0-7803-6456-1 *
MARTNEZ E ET AL: "Qualitative vision for the guidance of legged robots in unstructured environments", PATTERN RECOGNITION, ELSEVIER, GB, vol. 34, no. 8, 1 August 2001 (2001-08-01), pages 1585 - 1599, XP004362569, ISSN: 0031-3203 *
MATTHIES L ET AL: "KALMAN FILTER-BASED ALGORITHMS FOR ESTIMATING DEPTH FROM IMAGE SEQUENCES", INTERNATIONAL JOURNAL OF COMPUTER VISION, DORDRECHT, NL, vol. 3, no. 3, 1 January 1989 (1989-01-01), pages 209 - 238, XP008010388 *
TARAK GANDHI ET AL: "Parametric ego-motion estimation for vehicle surround analysis using an omnidirectional camera", MACHINE VISION AND APPLICATIONS, SPRINGER, BERLIN, DE, vol. 16, no. 2, 1 February 2005 (2005-02-01), pages 85 - 95, XP019323891, ISSN: 1432-1769 *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8233660B2 (en) * 2009-01-16 2012-07-31 Honda Research Institute Europe Gmbh System and method for object motion detection based on multiple 3D warping and vehicle equipped with such system
US20100183192A1 (en) * 2009-01-16 2010-07-22 Honda Research Institute Europe Gmbh System and method for object motion detection based on multiple 3d warping and vehicle equipped with such system
CN102413756B (en) * 2009-04-29 2015-04-01 皇家飞利浦电子股份有限公司 Real-time depth estimation from monocular endoscope images
CN102413756A (en) * 2009-04-29 2012-04-11 皇家飞利浦电子股份有限公司 Real-time depth estimation from monocular endoscope images
EP2256690A1 (en) * 2009-05-29 2010-12-01 Honda Research Institute Europe GmbH Object motion detection system based on combining 3D warping techniques and a proper object motion detection
JP2010282615A (en) * 2009-05-29 2010-12-16 Honda Research Inst Europe Gmbh Object motion detection system based on combining 3d warping technique and proper object motion (pom) detection
US8564657B2 (en) 2009-05-29 2013-10-22 Honda Research Institute Europe Gmbh Object motion detection system based on combining 3D warping techniques and a proper object motion detection
US20100315505A1 (en) * 2009-05-29 2010-12-16 Honda Research Institute Europe Gmbh Object motion detection system based on combining 3d warping techniques and a proper object motion detection
TWI425445B (en) * 2009-09-29 2014-02-01 Nat Univ Tsing Hua Method and detecting system for determining quantity of self-motion of a moving platform
JP2012084140A (en) * 2010-09-30 2012-04-26 Toshiba Corp Video analysis method and system
US8750614B2 (en) 2010-09-30 2014-06-10 Kabushiki Kaisha Toshiba Method and system for classifying features in a video sequence
WO2013006067A1 (en) * 2011-07-04 2013-01-10 Streeter Lee Vincent Motion compensation in range imaging
DE102014202639A1 (en) 2013-02-15 2014-08-21 Robert Bosch Gmbh Method for determining movement of vehicle, involves outputting movement of vehicle when quality value formed for evaluating detected vehicle movement estimate is available in predetermined relationship to reference value
CN109903315A (en) * 2019-03-08 2019-06-18 腾讯科技(深圳)有限公司 Method, apparatus, equipment and readable storage medium storing program for executing for light stream prediction
CN109903315B (en) * 2019-03-08 2023-08-25 腾讯科技(深圳)有限公司 Method, apparatus, device and readable storage medium for optical flow prediction
EP3809687A1 (en) * 2019-10-15 2021-04-21 Rohde & Schwarz GmbH & Co. KG Method and system for real time video stabilization

Also Published As

Publication number Publication date
JP4967062B2 (en) 2012-07-04
EP2179398B1 (en) 2011-03-02
EP2179398A1 (en) 2010-04-28
US8422741B2 (en) 2013-04-16
US20110243390A1 (en) 2011-10-06
DE602008005318D1 (en) 2011-04-14
JP2010535385A (en) 2010-11-18
ATE500570T1 (en) 2011-03-15

Similar Documents

Publication Publication Date Title
US8422741B2 (en) Estimating objects proper motion using optical flow, kinematics and depth information
US10275649B2 (en) Apparatus of recognizing position of mobile robot using direct tracking and method thereof
US10553026B2 (en) Dense visual SLAM with probabilistic surfel map
US10399228B2 (en) Apparatus for recognizing position of mobile robot using edge based refinement and method thereof
Audras et al. Real-time dense appearance-based SLAM for RGB-D sensors
JP5647155B2 (en) Body feature detection and human pose estimation using inner distance shape relation
KR101776620B1 (en) Apparatus for recognizing location mobile robot using search based correlative matching and method thereof
Lee et al. Particle filter with analytical inference for human body tracking
JP5012615B2 (en) Information processing apparatus, image processing method, and computer program
JP2009146406A (en) Visually tracking an object in real world using 2d appearance and multicue depth estimations
CN111354043A (en) Three-dimensional attitude estimation method and device based on multi-sensor fusion
CN111322993B (en) Visual positioning method and device
Shamwell et al. Vision-aided absolute trajectory estimation using an unsupervised deep network with online error correction
US11145072B2 (en) Methods, devices and computer program products for 3D mapping and pose estimation of 3D images
Lee et al. Human body tracking with auxiliary measurements
Kottath et al. Inertia constrained visual odometry for navigational applications
Kawanishi et al. Construction of 3D environment model from an omni-directional image sequence
Tse et al. Unified terrain mapping model with Markov random fields
Matthies et al. Probabilistic estimation mechanisms and tesselated representations for sensor fusion
KR101034675B1 (en) Method of estimating human pose using unscented kalman filter based on numerical inverse kinematics, computer readable medium and server system
Einhorn et al. Monocular detection and estimation of moving obstacles for robot navigation
JP5048381B2 (en) 3D shape restoration device
WO2022210005A1 (en) Attitude estimation system
Thompson et al. Stereo vision and sonar sensor based view registration for 2.5 dimensional map generation
Thompson et al. Revising stereo vision maps in particle filter based slam using localisation confidence and sample history

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08785662

Country of ref document: EP

Kind code of ref document: A1

DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2008785662

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 12599201

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2010519390

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE