WO2017025885A1 - Doppler time-of-flight imaging - Google Patents

Doppler time-of-flight imaging Download PDF

Info

Publication number
WO2017025885A1
WO2017025885A1 PCT/IB2016/054761 IB2016054761W WO2017025885A1 WO 2017025885 A1 WO2017025885 A1 WO 2017025885A1 IB 2016054761 W IB2016054761 W IB 2016054761W WO 2017025885 A1 WO2017025885 A1 WO 2017025885A1
Authority
WO
WIPO (PCT)
Prior art keywords
velocity
frequency
time
illumination
motion
Prior art date
Application number
PCT/IB2016/054761
Other languages
French (fr)
Inventor
Wolfgang Heidrich
Felix Heide
Gordon Wetzstein
Matthias Hullin
Original Assignee
King Abdullah University Of Science And Technology
University Of British Columbia
Stanford University
Bonn University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by King Abdullah University Of Science And Technology, University Of British Columbia, Stanford University, Bonn University filed Critical King Abdullah University Of Science And Technology
Priority to US15/739,854 priority Critical patent/US11002856B2/en
Publication of WO2017025885A1 publication Critical patent/WO2017025885A1/en

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • G01S17/08Systems determining position data of a target for measuring distance only
    • G01S17/32Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated
    • G01S17/34Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated using transmission of continuous, frequency-modulated waves while heterodyning the received signal, or a signal derived therefrom, with a locally-generated signal related to the contemporaneously transmitted signal
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/50Systems of measurement based on relative movement of target
    • G01S17/58Velocity or trajectory determination systems; Sense-of-movement determination systems
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • G01S17/8943D imaging with simultaneous measurement of time-of-flight at a 2D array of receiver pixels, e.g. time-of-flight cameras or flash lidar
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/491Details of non-pulse systems
    • G01S7/4911Transmitters

Definitions

  • the present disclosure generally relates to depth cameras and motion of objects captured by such cameras.
  • optical flow Usually, object motion or motion parallax is estimated via optical flow [Horn and Schunck 1981 ]: recognizable features are tracked across multiple video frames.
  • the computed flow field provides the basis for many computer vision algorithms, including depth estimation.
  • optical flow is computationally expensive, fails for untextured scenes that do not contain good features to track, and only measures 2D lateral motion perpendicular to the camera's line of sight.
  • the unit of optical flow is pixels; metric velocities cannot be estimated unless depth information of the scene is also available.
  • depth cameras have become increasingly popular for a range of applications, including human-computer interaction and gaming, augmented reality, machine vision, and medical imaging.
  • many limitations of optical flow estimation can be overcome using active illumination, as done by most structured illumination and time-of -flight (ToF) cameras where active illumination is temporally coded and analyzed on the camera to estimate a per-pixel depth map of the scene.
  • active illumination is temporally coded and analyzed on the camera to estimate a per-pixel depth map of the scene.
  • RGB-D imaging for example facilitated by Microsoft's Kinect One1
  • complex and untextured 3D scenes can be tracked by analyzing both color and depth information, resulting in richer visual data that has proven useful for many applications.
  • our imaging system allows for color, depth, and velocity information to be captured simultaneously.
  • Combining the optical flow computed on the RGB frames with the measured metric radial velocity allows estimation of the full 3D metric velocity field of the scene.
  • the present technique has applications in many computer graphics and vision problems, for example motion tracking, segmentation, recognition, and motion de-blurring.
  • a method for imaging object velocity can comprise the steps of: providing a Time-of-Flight camera and using the Time-of-Flight camera to capture a signal representative of an object in motion over an exposure time; coding illumination and modulation frequency of the captured motion within the exposure time; mapping a change of illumination frequency to measured pixel intensities of the captured motion within the exposure time; and extracting information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion.
  • radial velocity information of the object in motion can be simultaneously captured for each pixel captured within the exposure time.
  • the illumination frequency can be coded orthogonal to the modulation frequency of the captured motion.
  • the change of illumination frequency can correspond to radial object velocity.
  • the Time-of-Flight camera can have a receiver and a transmitter, and the frequency of the receiver can be configured to be orthogonal to the frequency of the transmitter.
  • the exposure time can be longer than the wavelength of a modulated captured signal.
  • a ratio of a heterodyne measurement and a homodyne measurement can be determined to extract the information about the Doppler shift.
  • the method can further include the step of: simultaneously capturing color, depth and velocity information concerning the object in motion during the exposure time.
  • the change of illumination frequency can correspond to radial object velocity, and optical flow of the object in motion can be computed on red, green and blue (RGB) frames within a measured change in illumination frequency.
  • the method can further include estimating a 3D velocity field for the object in motion.
  • the depth and velocity imaging can be combined using either the Time-of-Flight camera by alternating modulation frequencies between successive video frames over the exposure time or using at least two Time-of-Flight cameras.
  • the system can comprise: at least one device for capturing a signal representative of an object in motion over an exposure time; at least one computing device comprising a processor and a memory; and an application executable in the at least one computing device, the application comprising machine readable instructions stored in the memory that, when executed by the processor, cause the computing device to at least: (a) code illumination and modulation frequency of the captured motion within the exposure time; (b) map a change of illumination frequency to measured pixel intensities of the captured motion within the exposure time; and (c) extract information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion.
  • the device can be at least one Time-of-Flight camera.
  • a non-transitory computer readable medium employing an executable application in at least one computing device, the executable application comprising machine readable instructions stored in the medium that: (a) receives signals representative of an object in motion over an exposure time; (b) codes illumination and modulation frequency of the captured motion within the exposure time; (c) maps a change of illumination frequency to measured pixel intensities of the captured motion within the exposure time; and (d) extracts information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion.
  • the signals can be captured using at least one Time-of-Flight camera.
  • radial velocity information of the object in motion can be simultaneously captured for each pixel captured within the exposure time.
  • the illumination frequency can be coded orthogonal to the modulation frequency of the captured motion.
  • the change of illumination frequency can correspond to radial object velocity.
  • the Time-of-Flight camera can include a receiver and a transmitter, and the frequency of the receiver can be configured to be orthogonal to the frequency of the transmitter.
  • the logic can capture color, depth and velocity information concerning the object in motion during the exposure time.
  • FIG. 1A depicts an embodiment of an imaging system of the present disclosure that allows for metric radial velocity information to be captured. Depicted in the embodiment are a high-speed illumination source, RGB camera, and Time-of-Flight (ToF) camera.
  • RGB camera a high-speed illumination source
  • ToF Time-of-Flight
  • Fig. 1 B depicts images that can captured by an imaging system of the present disclosure of an example of an object in motion.
  • Fig. 1 C depicts velocity data that can captured by an imaging system of the present disclosure of an example of an object in motion.
  • Fig. 1 D depicts depth data that can captured by an imaging system of the present disclosure of an example of an object in motion.
  • Fig. 1 E depicts images that can captured by an imaging system of the present disclosure of an example of an object in motion.
  • Fig. 1 F depicts velocity data that can captured by an imaging system of the present disclosure of an example of an object in motion.
  • Fig. 1 G depicts depth data that can captured by an imaging system of the present disclosure of an example of an object in motion.
  • Fig. 2A depicts an embodiment of depth imaging of the present disclosure for a static scene.
  • Fig. 2B depicts an embodiment of depth imaging of the present disclosure for a scene in motion.
  • Fig. 3A depicts an embodiment of velocity imaging of the present disclosure for a static scene.
  • Fig. 3B depicts an embodiment of velocity imaging of the present disclosure for a scene in motion.
  • Fig. 4A depicts examples of simulated intensities for a large range of different velocities.
  • Fig. 4B depicts examples of simulated intensities for a small range of different velocities.
  • Fig. 4C depicts examples of measured intensities for a small range of different velocities.
  • Fig. 4D depicts examples of measured intensities for a smaller range of different velocities than Fig. 4C.
  • Fig. 5A depicts depth-dependent offset introduce by higher-order frequency components for a modulation frequency of 30 MHz.
  • Fig. 5B depicts depth-dependent offset introduce by higher-order frequency components for a modulation frequency of 80 MHz.
  • Fig. 5C depicts depth-dependent offset introduce by higher-order frequency components for a modulation frequency of 150 MHz.
  • FIG. 6A depicts an experimental verification of the imaging system for varying object velocities and depths (left)
  • Fig. 6B depicts measured intensities for a range of different pixel locations and velocity-dependent behavior for a range of different pixel locations on the sensor (right).
  • Fig. 7A depicts an embodiment of experimental setup used for an experimental validation of velocity estimation using a fan with adjustable rotation speed (three settings).
  • FIG. 7B audio recordings analyzed to generate ground truth velocity data of the rotating blades of the setup in Fig. 7 A
  • Fig. 7C shows the velocity measured by D-ToF compared to the ground truth for a varying rotation speed.
  • Fig. 7D shows the unprocessed full-field measurements of the homodyne frequency setting with the pixel indicated for which velocities were plotted in Fig. 7C.
  • Fig. 7E shows the unprocessed full-field measurements of the heterodyne frequency setting with the pixel indicated for which velocities were plotted in Fig. 7C.
  • Fig. 8A depicts images of motion within a complex scene with ambient illumination and a large depth range.
  • Fig. 8B depicts velocity data of motion within a complex scene with ambient illumination and a large depth range The velocity can be robustly estimated within the range of the illumination (approx. 5m inside), even in outdoor settings.
  • Fig. 9A depicts computed velocity maps encoded in grayscale from raw measurements.
  • Fig. 9B depicts reconstructed de-noised images based on the velocity maps encoded in grayscale from raw measurements of Fig. 9A.
  • Fig. 10A depicts an example of periodic motion of a hand along the optical axis.
  • the static scene on the left results in no response of the sensor, whereas forward (center) and backward (right) movement result in positive and negative responses respectively.
  • Fig. 10B depicts velocity data for the example in Fig. 10A.
  • Fig. 1 1 A depicts an example of periodic motion along the Z-axis for a textured object. Although the estimated velocity is mostly correct, shadows and dark scene parts can be challenging for robust velocity estimation.
  • Fig. 1 1 B depicts velocity data for the example in Fig. 1 1 A.
  • Fig. 1 1 C depicts depth data for the example in Fig. 1 1 A.
  • Fig. 12A shows an example of extremely fast motion that can be accurately captured with the present system.
  • Fig. 12B shows velocity data for the example in Fig. 12A.
  • the Airsoft gun in the example is advertised as shooting bullets with 99 m/s; a radial velocity of 98.2 m/s (average of the peak pixels) can be calculated in the present example using the system and methods of the present disclosure.
  • Fig. 13A depicts an example of a potential applications of the present disclosure, including gaming and human - computer interaction.
  • An example of a person in motion in a scene is depicted from left to right.
  • Fig. 13B shows velocity data for the example of Fig. 13A.
  • Fig. 13C shows depth data for the example of Fig. 13A.
  • Fig. 14A depicts an example of physical props for gaming, such as ping pong balls fired with a toy gun, which can be tracked with the present system and enable HCI techniques. Images of props in motion with a person in a scene is show across time from left to right.
  • Fig. 14B shows velocity data for the example of Fig. 14A.
  • FIG. 14C shows depth data for the example of Fig. 14A.
  • Fig. 15A depicts a failure case of optical flow for a moving, but un- textured, scene (left).
  • Fig. 15B shows Optical flow [Liu 2009] for two succeeding frames of the scene from Fig. 15A.
  • the 2D flow vectors can be color-coded with a color wheel (insets).
  • Fig. 15C shows SIFT flow [Liu et al. 2008] for two succeeding frames of the scene from Fig. 15A.
  • the 2D flow vectors can be color-coded with a color wheel (insets).
  • Fig. 15D shows velocity data from the scene of Fig. 15A according to the system and methods described herein.
  • Fig. 16A depicts an example of a frame where optical flow computed reasonable estimates.
  • Fig. 16B show the full 3D velocity estimate for different views of the example in Fig. 16A.
  • Optical flow can aid in 3D velocity estimates and image reconstruction.
  • Fig. 17 is a flowchart depicting an embodiment of a method of the present disclosure.
  • Fig. 18 depicts an embodiment of an apparatus that can be used in the systems and methods of the present disclosure.
  • Fig. 19 shows an embodiment of a camera system according to the present disclosure.
  • Fig. 20 shows an embodiment of a camera system according to the present disclosure.
  • Fig. 21 shows an embodiment of a camera system according to the present disclosure.
  • the motion can be velocity.
  • a Doppler effect can be analyzed in one or more Time-of-Flight cameras: object motion towards or away from the cameras can shift the temporal illumination frequency before it is recorded by the camera.
  • Conventional Time-of-Flight cameras encode phase information (and therefore scene depth) into intensity measurements.
  • Doppler Time-of-Flight (D-ToF) is used to provide a new imaging mode, whereby the change of illumination frequency (corresponding to radial object velocity) can be directly encoded into the measured intensity.
  • the camera hardware utilized can be the same as for conventional Time- of-Flight imaging, but illumination and modulation frequencies can be carefully designed. Depth and velocity imaging can be combined using either two Time-of- Flight cameras or using the same device by alternating the modulation frequencies between successive video frames; color images can be obtained with a conventional camera.
  • a fundamentally new imaging modality is provided that is ideally suited for fast motion.
  • Optical flow applied to conventional RGB video is a complimentary technique: together, optical flow and D-ToF allow for the metric 3D velocity field to be estimated, which is otherwise not easily possible.
  • the present D-ToF can be independent of the RGB flow and can work robustly for cases where optical flow often fails, including untextured scenes and extremely high object velocities.
  • Doppler radar is widely used in police speed guns, although gradually being replaced by lidar-based systems.
  • Doppler lidar is also commonly used in many meteorological applications, such as wind velocity estimation.
  • One common limitation of all Doppler measurements is that only movement along one particular direction, usually the line-of-sight, can be detected. All of these applications rely on the wave nature of light or sound, and therefore require coherent illumination or precise spectroscopic measurement apparatuses.
  • incoherent, amplitude-modulated illumination and inexpensive time-offlight (ToF) cameras can be used for instantaneous imaging of both velocity and range.
  • a full-field imaging method is provided, meaning that it does not require the scene to be sequentially scanned unlike most existing Doppler radar or lidar systems that only capture a single scene point at a time.
  • a framework and a camera system are provided implementing the described techniques; together, they can optically encode object velocity into per-pixel measurements of modified time-of-flight cameras. By combining multiple cameras, color, range, and velocity images can be captured simultaneously.
  • Pandharkar et al. [201 1 ] recently proposed a pulsed femtosecond illumination source to estimate motion of non-line-of-sight objects from differences in multiple captured images.
  • the present systems and methods can use the Doppler effect observed with conventional time-of-flight cameras within a single captured frame, as opposed to optical flow methods that track features between successive video frames.
  • Optical flow is a fundamental technique in computer vision that is vital for a wide range of applications, including tracking, segmentation, recognition, localization and mapping, video interpolation and manipulation, as well as defense.
  • Optical flow from a single camera is restricted to estimating lateral motion whereas the Doppler is observed only for radial motion towards or away from the camera.
  • the present method can involve the capture of a few sub-frames with different modulation signals.
  • appropriate hardware for example, multi-sensor cameras or custom sensors with different patterns multiplexed into pixels of a single sensor
  • the method can be implemented as a true snapshot imaging approach.
  • rapid time-sequential for example, 30-60 frames per second, and even higher with specialized equipment
  • - D-ToF is presented herein as a new modality of computational photography that allows for direct estimation on instantaneous radial velocity. In an aspect using multiple captures or implemented with multi-sensor setups, it can record velocity, range, and color information.
  • a framework for velocity estimation with Time-of-Flight cameras is provided, along with a Time-of-Flight imaging system, and the framework and system validated in simulation and with the system.
  • Time-of-flight cameras operate in continuous wave mode. That is, a light source illuminates the scene with an amplitude-modulated signal that changes periodically over time. Sinusoidal waves are often used in the ToF literature to approximate the true shape of the signals. We restrict the derivation herein to the sine wave model for simplicity of notation. Hence, the light source emits a temporal signal of the form
  • ⁇ 3 ⁇ 4 is the illumination frequency.
  • Time-of-Flight camera pixels can provide a feature that makes them distinct from conventional camera pixels: before being digitally sampled, the incident signal can be modulated by a high-frequency, periodic function fip(t) within each pixel.
  • the modulation frequency can be 10 MHz - 1 GHz, 10 MHz - 800 MHz, 10 MHz - 600 MHz, 10 MHz - 500 MHz, 10 MHz - 400 MHz, 10 MHz - 300 MHz, 10 MHz - 200 MHz, or 10 MHz - 100 MHz.
  • This on-sensor modulation can be physically performed by an electric field that rapidly redirects incident photons-converted-to-electrons into one of two buckets within each pixel.
  • the phase and frequency ⁇ ⁇ of the modulation function are programmable.
  • the general equation for the modulated signal is thus
  • the exposure time T of all cameras can act as a low-pass filter on the modulated signal before it is discretized by the sampling process of the sensor. Since the exposure time is usually significantly longer than the wavelength of the modulated signal
  • Equation 4 vanish:
  • the temporal low-pass filter rsc 3 ⁇ 41 ' is convolved with the incident signal— an operation that is analogous to the finite integration area of each sensor pixel in the spatial domain.
  • the low-pass filter resulting from spatial sensor integration is known as the detector footprint modulation transfer function [Boreman 2001 ].
  • the modulated and low- pass filtered signal can be discretely sampled. Since Equation 5 is independent of the time of measurement t', depth and albedo can be robustly estimated.
  • the conventional Time-of-Flight image formation model breaks down when objects of interest move with a non-negligible radial velocity.
  • the illumination frequency undergoes a Doppler shift [Doppler 1842] when reflected from an object in motion.
  • Doppler 1842 When reflected from an object in motion.
  • the illumination arriving at the sensor is now frequency-shifted to ⁇ 1:" where the change in temporal frequency ⁇ depends on the radial object velocity as well as the illumination frequency: (8)
  • Equation 3 Equation 3 can be used to derive a new version of the low-pass filtered sensor image (Eq. 5) for moving scenes: j S3 ⁇ 4S — ⁇ ⁇ — 9
  • heterodyne imaging mode can be beneficial in certain situations, but a major limitation of heterodyne ToF is that multiple (>2) measurements have to be captured to reliably estimate phase and depth. Since the beating pattern is usually of very low frequency (for example, in the order of a few Hz at most velocities typical to indoor environments), a significant amount of time needs to pass between the two measurements for accurate phase estimation. For moving objects, the necessity to capture multiple images may place constraints on the velocity.
  • a new computational Time-of-Flight imaging methodology is derived in the following section. Similar to orthogonal frequency-division multiplexing (OFDM), D-ToF uses illumination and on-sensor modulation frequencies that are orthogonal within the exposure time of the camera. Using these frequencies, a method is provided that allows per-pixel radial object velocity estimation.
  • OFDM orthogonal frequency-division multiplexing
  • the low-frequency beating pattern created by the Doppler effect makes it difficult or impossible to capture reliable Doppler frequency and phase information.
  • a road cyclist travels at a speed of " $ towards the camera.
  • a frequency shift of only 1 :67 Hz may seem small enough to be safely ignored. However, we show in the following that even such a minute change contains valuable information that can be used for velocity estimation.
  • the camera system can be interpreted as a communication channel, and the illumination considered as a carrier signal.
  • the carrier can be optically modified by moving objects— a change can be observed in carrier amplitude, phase, and frequency.
  • the secondary modulation in the sensor followed by a low-pass filter of the exposure time can correspond to the demodulation process in communication.
  • Conventional communication channels use orthogonal frequencies; any inter- carrier interference (which could be caused by a frequency drift) is a polluting signal.
  • the frequencies in the receiver and transmitter can be designed to be orthogonal, such that the (usually polluting) inter-carrier interference carries the desired velocity information. An example is shown in Figs. 3A and 3B.
  • the measured signal for a stationary object can be zero (or a constant intensity offset). This can be achieved by operating the ToF camera in heterodyne mode with two orthogonal frequencies ⁇ ⁇ and ⁇ ⁇ . While any two sine waves with frequencies ⁇ 3 ⁇ 4 ⁇ ⁇ ⁇ will be orthogonal for sufficiently long integration times, this is not the case for finite integrals (exposures) in the presence of low frequency beating patterns. Designing both frequencies to be orthogonal is done by setting
  • the inter-carrier interference can be used to extract valuable information about the Doppler shift. This can be achieved by computing the ratio of a heterodyne measurement and a homodyne measurement. Using only the low frequency terms from Equation 3, this ratio can be expressed, without loss of generality and assuming an exposure interval of [0... 7], as:
  • Figs. 4A-D shows the model derived here. On the left side, the full model is seen without any approximations (i.e. without neglecting high frequency components in Eq. 14). Although the image formation is nonlinear, for a relatively large range of metric velocities (Fig. 4A) it is very well approximated (Fig. 4B, center left) by our linear model (Eq. 14). The model is verified experimentally by using the camera prototype (Figs. 4C and 4D, right).
  • a third possibility is to rapidly alternate between two modulation frequencies using a single ToF camera.
  • the measurements are not truly instantaneous, and alignment problems can occur for very fast motions.
  • the two measurements can be taken immediately after each other, as fast as the camera hardware allows, e.g. at 30 or 60 Hz.
  • We follow this approach as it only requires a single ToF camera.
  • Equation 14 may only hold for sinusoidal modulation functions. If other periodic signals are being used, additional harmonic frequency components are introduced, which can distort the measurements for both stationary and moving targets. However, these offsets are systematic and can be calibrated for a specific ToF camera/lights source combination (see Implementation Section herein).
  • simultaneous velocity and range imaging may involve three distinct measurements.
  • the illumination signal may be the same for all three measurements. Only the reference signal for the camera may change.
  • velocity-only imaging this means that all three measurements can potentially be acquired at the same time using either multiple sensors with a shared optical axis, or a sensor design with interleaved pixels. If neither option is available, rapid frame-sequential imaging is also possible.
  • a Time-of-Flight camera is provided 103.
  • the camera can be used to capture a signal representative of an object in motion over an exposure time.
  • illumination and modulation frequency within the exposure time for the captured motion are coded 106.
  • illumination frequency changes are mapped 109 to measured pixel intensities of the captured motion with the exposure time.
  • Doppler shift information in the illumination frequency is extracted 1 12 to obtain a measurement of instantaneous per pixel velocity of the object in motion.
  • Hardware characteristics of the imaging system or Time- of-Flight camera as described herein can include an illumination unit, optics, an image sensor, driver electronics, an interface, and computational ability.
  • the hardware of embodiments of imaging systems as described herein can be seen in Fig. 1A, Fig. 19, Fig. 20, and Fig. 21.
  • An embodiment of a generic camera system is shown in Fig. 19.
  • the embodiment shown in Fig. 19 can be tailored to different applications by changing the characteristics of the imaging sensor.
  • the imaging sensor of Fig. 19 can be a conventional RGB imaging sensor and therefore Fig. 19 can be an RGB camera.
  • the imaging sensor of Fig. 19 can be a sensor suitable for a Time- of-Flight camera, such as the PMD Technologies PhotonlCs 19k-S3 imaging sensor, and Fig. 19 can be a Time-of-Flight camera.
  • an experimental Time-of-Flight camera system comprises a custom RF modulated light source and a demodulation camera based on the PMD Technologies PhotonlCs 19k-S3 imaging sensor (see Fig. 1A).
  • the system allows for metric radial velocity information to be captured instantaneously for each pixel (center row).
  • the illumination and modulation frequencies of a Time-of-Flight camera (left) to be orthogonal within its exposure time.
  • the Doppler effect of objects in motion is then detected as a frequency shift of the illumination, which results in a direct mapping from object velocity to recorded pixel intensity.
  • An illumination unit can be a light source which can be an array of 650 nm laser diodes driven by iC-Haus constant current driver chips, type ic-HG.
  • a PMD CamBoard nano development kit was used with a clear glass sensor that has the near IR bandpass filter removed, in combination with an external 2- channel signal generator to modulate the sensor and synchronize the light source.
  • the setup is similar to commercially-available Time-of-Flight cameras and the proposed algorithms can be easily implemented on those.
  • developers usually do not have access to illumination and modulation frequencies of these devices, requiring the construction of custom research prototype cameras.
  • the maximum illumination and demodulation frequency of our prototype is 150 MHz, but we run all of the presented results with 30 MHz.
  • the modulation signals are nearly sinusoidal, but contain multiple low-amplitude harmonic components. To avoid systematic errors in depth and velocity estimation, these components can be calibrated as described in the following.
  • Fig. 18, depicts an apparatus 1010 in which the Doppler Time-of- Flight imaging described herein may be implemented.
  • the apparatus 1010 can contain the driver electronics and computational ability for the imaging system or Time-of-Flight camera as described herein.
  • the apparatus 1010 may be embodied in any one of a wide variety of wired and/or wireless computing devices, multiprocessor computing device, and so forth.
  • the apparatus 1010 comprises memory 214, a processing device 202, a number of input/output interfaces 204, a network interface 206, a display 205, a peripheral interface 21 1 , and mass storage 226, wherein each of these devices are connected across a local data bus 210.
  • the apparatus 1010 may be coupled to one or more peripheral measurement devices (not shown) connected to the apparatus 1010 via the peripheral interface 21 1 .
  • the processing device 202 may include any custom made or commercially available processor, a central processing unit (CPU) or an auxiliary processor among several processors associated with the apparatus 1010, a semiconductor based microprocessor (in the form of a microchip), a macroprocessor, one or more application specific integrated circuits (ASICs), a plurality of suitably configured digital logic gates, and other well-known electrical configurations comprising discrete elements both individually and in various combinations to coordinate the overall operation of the computing system.
  • CPU central processing unit
  • ASICs application specific integrated circuits
  • the memory 214 can include any one of a combination of volatile memory elements (e.g., random-access memory (RAM, such as DRAM, and SRAM, etc.)) and nonvolatile memory elements (e.g., ROM, hard drive, tape, CDROM, etc.).
  • the memory 214 typically comprises a native operating system 216, one or more native applications, emulation systems, or emulated applications for any of a variety of operating systems and/or emulated hardware platforms, emulated operating systems, etc.
  • the applications may include application specific software which may be configured to perform some or all of the Doppler Time-of-Flight imaging techniques described herein.
  • the application specific software is stored in memory 214 and executed by the processing device 202.
  • the memory 214 can, and typically will, comprise other components which have been omitted for purposes of brevity.
  • Input/output interfaces 204 provide any number of interfaces for the input and output of data.
  • the apparatus 1010 comprises a personal computer
  • these components may interface with one or more user input devices 204.
  • the display 205 may comprise a computer monitor, a plasma screen for a PC, a liquid crystal display (LCD) on a hand held device, or other display device.
  • LCD liquid crystal display
  • a non-transitory computer- readable medium stores programs for use by or in connection with an instruction execution system, apparatus, or device. More specific examples of a computer- readable medium may include by way of example and without limitation: a portable computer diskette, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory), and a portable compact disc read-only memory (CDROM) (optical).
  • RAM random access memory
  • ROM read-only memory
  • EPROM erasable programmable read-only memory
  • CDROM portable compact disc read-only memory
  • network interface device 206 comprises various components used to transmit and/or receive data over a network environment.
  • the network interface 206 may include a device that can communicate with both inputs and outputs, for instance, a modulator/demodulator (e.g., a modem), wireless (e.g., radio frequency (RF)) transceiver, a telephonic interface, a bridge, a router, network card, etc.).
  • the apparatus 1010 may communicate with one or more computing devices via the network interface 206 over a network.
  • the apparatus 1010 may further comprise mass storage 226.
  • the peripheral 21 1 interface supports various interfaces including, but not limited to IEEE-1394 High Performance Serial Bus (Firewire), USB, a serial connection, and a parallel connection.
  • the apparatus 1010 shown in Fig. 18 can be electronically coupled to and in communication with a Time-of-Flight camera as shown in Fig. 19, 20, and 21. Data can be passed back and forth between the apparatus 1010 and the Time-of-Flight camera, wired (USB, Firewire, thunderbolt, SDI, Ethernet, for example) or wirelessly (Bluetooth or WiFi, for example). Alternatively, the apparatus 1010 can be a part of the Time-of-Flight camera.
  • An imaging system as described herein can be comprised of a Time-of-Flight camera or a Time-of- Flight camera in communication with an apparatus such as the apparatus 1010.
  • An imaging system as described herein can also include any conventional RGB camera and/or an illumination source.
  • An RGB camera and/or illumination source can also electronically coupled to and in communication with an apparatus 1010 along with a Time-of-Flight camera in an embodiment of an imaging system.
  • An imaging system as described herein can be configured to record successive frames of a scene.
  • the scene can contain one or more objects in motion.
  • Successive frames of a scene can be still images or from a video constructed of continuous successive frames.
  • Scenes can be captured by the Time-of-Flight camera or Time-of-Flight camera in conjunction with an RGB camera.
  • Data from the camera[s] can be sent and processed by an apparatus such as the apparatus 1010, and the apparatus 1010 can compute, process, and/or reconstruct data captured by the camera[s].
  • Data captured by the camera[s] can be one or more signals representative of one or more objects in motion.
  • the one or more signals can contain information relating to RGB images, velocity, and/or depth that are representative of a scene.
  • Embodiments of the present imaging systems are shown in Fig. 1A, Fig. 19, and Fig. 21.
  • This offset can be calibrated in an offline process and raw phase measurements can be corrected digitally using a lookup table. Note that for relatively low modulation frequencies, such as 30 MHz, we find a fairly large depth range (around 1 m) to be almost independent of this offset. In practice, it is therefore relatively easy to remove the higher-order frequency components.
  • Figs. 7A-E show another experiment that was used to verify the accuracy of our D-ToF camera system.
  • the experiment setup is shown in Fig. 7A.
  • the speed of a rotating fan was adjusted and its blades imaged such that, throughout the time it takes for a single blade to move across a pixel, forward motion is observed by that pixel.
  • the exposure time of the ToF camera was set to 1 .5 ms and the fan was captured from a frontal perspective (raw homodyne and heterodyne measurements shown in Fig. 7 bottom).
  • the slope of the fan blades was manually measured, which is constant over the entire blades.
  • the radius of the plotted position was measured, allowing calculation of the "ground truth" velocity when the rotation speed of the fan is known. Since the exact rotation speed is not actually known, it was measured by mounting a small pin on one of the blades and mounting a piece of flexible plastic in front of the fan, such that the rotating pin strikes the plastic exactly once per revolution, creating a distinct sound.
  • the sound (sampled at 44 KHz, Fig. 7B) of this setup was measured (to estimate the ground truth velocity of the fan blades, observed by one pixel, which is compared with the corresponding D- ToF estimate (Fig. 7C). For this experiment, the estimation error is always below 0.2 m/s. Errors are mainly due to the low SNR of the measured Doppler-shifted signal.
  • Figs. 1 D, 1G, 11C, 13C, and 14C also show the corresponding depth maps that can be estimated from an additional capture as well as the velocity maps (see Simultaneous Range and Velocity Section herein).
  • Optical flow computed from conventional video sequences estimates the 2D projection of the 3D flow field onto the image plane.
  • the radial component is usually lost.
  • optical flow is an ill-posed problem and may fail in many scenarios.
  • Our Doppler ToF addresses two problems of optical flow: first, it can help in cases where optical flow fails either due to large displacements or missing scene structures. Second, the present method can also help in cases where the optical flow estimation is successful; in this case, the 3D metric flow can be recovered by combining metric radial velocity and the 2D optical pixel flow.
  • Fig. 15A shows a scene where regular optical flow [Liu 2009], as well as SIFT-flow [Liu et al. 2008], fail due to limited structure in the scene (Fig. 15B and 15C respectively). Both methods cannot recover the true 2D motion of the fan and wrongly segment the scene.
  • the present orthogonal velocity estimation method successfully captures the velocity of the objects and also leads to a proper segmentation of the scene (Fig. 15D). Note that having additional depth estimates for conventional flow may only be of limited help since flat surfaces also do not deliver enough features for correspondence matching.
  • Fig. 16A shows a scene where the optical flow estimate is reasonable.
  • the orthogonal component that our method captures completes the 2D spatial flow estimates and uniquely determines the full metric 3D flow.
  • F the focal length of the lens
  • Z the corresponding depth estimate from our method.
  • the full 3D metric flow is ⁇ t ⁇ n ⁇ '
  • Fig. 16B An example is shown in Fig. 16B. Note that the optical flow helps determine that the fan's velocity is slightly rotated to the upper right, where the center of rotation is located (bottom left). Also note that 3D flow field is only as reliable as the estimated radial velocity and the RGB 2D flow.
  • a new computational imaging modality that directly captures radial object velocity via Doppler Time-of-Flight Imaging.
  • a variety of experimental results captured with a prototype camera system are demonstrated for different types of motions and outdoor settings. The methods are extensively validated in simulation and experiment.
  • the optional combination of footage captured using an RGB camera with the depth and velocity output of the present coded Time-of-Flight camera system is shown. Together, this data can represent simultaneous per-pixel RGB, depth, and velocity estimates of a scene and allow for the 3D velocity field to be estimated.
  • Applications in a wide range of computer vision problems including segmentation, recognition, tracking, super-resolution, spatially-varying motion de-blurring, and navigation of autonomous vehicles are provided.
  • the present method is complimentary to optical flow. It allows for the depth bias of xz-flow to be removed and enables recording of the metric 3D velocity field of the scene. However, if only radial velocity is required, the present method can also be used stand-alone, independent of optical flow.
  • ToF sensors today are low-resolution and their quantum efficiency and noise characteristics are not comparable with modern CMOS sensors. Future generations of ToF sensors are expected to deliver significantly higher image quality, which would directly benefit the present method as well. Higher modulation frequencies would directly improve the signal-to-noise ratio in our setup, because the Doppler effect is proportional to these frequencies.
  • laser diodes can be used that operate in the visible spectrum in combination with a ToF sensor that has its visible spectrum cutoff filter removed. The laser illumination is therefore visible in all of the RGB images as a red tint.
  • the present system can also operate the Time-of- Flight camera in the near infrared spectrum, as is common practice in commercial ToF cameras. Finally, all presented techniques can be easily be implemented on consumer Time-of-Flight cameras with the appropriate level of access to the system firmware or driver software.
  • Ratios, concentrations, amounts, and other numerical data may be expressed in a range format. It is to be understood that such a range format is used for convenience and brevity, and should be interpreted in a flexible manner to include not only the numerical values explicitly recited as the limits of the range, but also to include all the individual numerical values or sub-ranges encompassed within that range as if each numerical value and sub-range is explicitly recited.
  • a concentration range of "about 0.1 % to about 5%” should be interpreted to include not only the explicitly recited concentration of about 0.1 % to about 5 %, but also include individual concentrations (e.g., 1 %, 2%, 3%, and 4%) and the sub-ranges (e.g., 0.5%, 1 .1 %, 2.2%, 3.3%, and 4.4%) within the indicated range.
  • the term "about” can include traditional rounding according to significant figure of the numerical value.
  • the phrase "about 'x' to 'y'" includes “about 'x' to about 'y'".
  • HEIDE F., HULLIN, M. B., GREGSON, J., AND HEIDRICH, W. 2013. Low-budget transient imaging using photonic mixer devices.
  • HEIDE F., XIAO, L, HEIDRICH, W., AND HULLIN, M. B. 2014.
  • Diffuse mirrors 3D reconstruction from diffuse indirect illumination using inexpensive time-of-flight sensors. In Proc. CVPR.
  • HEIDE F., XIAO, L, KOLB, A., HULLIN, M. B., AND HEIDRICH, W.
  • KADAMBI A., WHYTE, R., BHANDARI, A., STREETER, L, BARSI, C, DORRINGTON, A., AND RASKAR, R. 2013. Coded time of flight cameras: sparse deconvolution to address multipath interference and recover time profiles.
  • KIRMANI A., HUTCHISON, T., DAVIS, J., AND RASKAR, R. 2009. Looking around the corner using transient imaging.
  • Proc. ICCV 159-166.
  • TOCCI, M., KISER, C, TOCCI, N., AND SEN, P. 201 A versatile HDR video production system.
  • VELTEN A., WU, D., JARABO, A., MASIA, B., BARSI, C, JOSHI, C, LAWSON, E., BAWENDI, M., GUTIERREZ, D., AND RASKAR, R. 2013. Femto- photography: Capturing and visualizing the propagation of light.
  • a y h are 3 ⁇ 4 e com lex phasor amplitudes containing the ampiitude and phase dependency. is the Kronecker delta; for perfectly static scenes, this expression is zero. Note that phasors can usually not be multiplied. By using the complex conjugate, this is possible while implicitly assuming that high frequency can be ignored [Ceperiey 20153.
  • Figure 1 We introduce a new computational imaging system that allows for metric radial velocity information to be captured instantaneously or each pixel (center mw).
  • the Doppler effect of objects in motion is then detected as a frequency shift of the illumination, which results in a mapping from object velocity to recorded pixel intensity.
  • the results above show several frames of two video sequences. For each example, the left-most frame shows static object (velocity map is constant), which is then moved towards (positive radial velocity) or away from (negative velocity) the camera.
  • depth cameras have become increasingly Pioneers of photography, including Eadweard Muybridge and Harpopular for a range of applications, including human-computer inold "Doc" Edgerton, advanced imaging technology to reveal otherteraction and gaming, augmented reality, machine vision, and medwise invisible motions of high-speed events.
  • Many of the commercially-available devices use the motion of objects in complex scenes is at the core of computer the time-of-fiight principle, where active illumination is temporally vision, with a wide range of applications in object tracking, segcoded and analyzed in the camera to estimate a per-pixei depth map mentation, recognition, motion deblurring, navigation of autonomof the scene.
  • optical flow illumination and modulation frequencies of the ToF camera object is computationally expensive, fails for untextured scenes that do not velocities directly map to measured pixel intensities.
  • Combinof optical flow is pixels; metric velocities cannot be estimated uning the optical flow computed on the RGB frames with the measless depth information of the scene is also available.
  • metric radial velocity allows us to further estimate the full ticular application of depth estimation, many limitations of optical 3D metric velocity field of the scene.
  • the proposed technique has How estimation can be overcome using active illumination, as done applications in many computer graphics and vision problems, for by most structured illumination and time-of-fiight (Top 7 ) cameras. example motion tracking, segmentation, recognition, and motion With the emergence of RGB-D imaging, for example facilitated by- deblurring. Microsoft's Kinect One' , complex and untextured 3D scenes can be tracked by analyzing both color and depth information, resulting in
  • the required camera hardware is the same as for 1 microsoft.com/en-us/kinectforwindows/ conventional time-of-flight imaging, hut illumination and modulajects with dark albedos, shadowed regions, and global illumination tion frequencies are carefully designed. We can combine depth and effects.
  • RGB video is a complimentary technique: together, optical flow widespread use in astronomical imaging, meteorology, traffic law and D-ToF allow for the metric 3D velocity field to be estimated, enforcement, radiology, healthcare, and aviation. Doppler spectrowhich is otherwise not easily possible. In general, however, D-ToF scopy, for example, measures radial velocity of otherwise undetectis independent of the RGB flow and works robustly for cases where able planets by observing wavelength shifts of their respective stars.
  • Laser Doppler veJocimetry is a comrange and velocity imaging. As with standard ToF imaging, our mon technique in heailhcare, for example to measure blood flow. method requires a few subframes to be captured with different modUsually, this technique uses two crossed, coherent laser beams to ulation signals.
  • Doppler radar is widely used in poshot imaging approach. In our prototype system, we instead use lice speed guns, although gradually being replaced by lidar-based rapid time-sequential acquisition of the required subframes, which systems. Doppler lidar is also commonly used in many meteorois a common strategy for regular ToF imaging. In summary, tikis logical applications, such as wind velocity estimation. One compaper makes tike following contributions: mon !imitatiotk of all Doppler measurements is that only movement alotkg one particular direction, usually the line-of- sight, can be de ⁇
  • D-ToF new modality of computational phototected. All of these applications rely on the wave nature of light or graphy that enables instantaneous radial velocity estimation. sound, and therefore require coherent illumination or precise specUsing multiple captures or implemented with multi-sensor troscopic measurement apparatuses.
  • D-ToF records velocity, range, and color information.
  • Time-of-F!ig t Photography With consumer as indoors and under strong outdoor ambient illumination. time-of-flight cameras such as Microsoft's Kinect One becoming
  • tike resolution of the PMD sensor in our prototype is limficult inverse problems, such as non- line -of- sight imaging [ irmani ited to 160 X 120 pixels and the signal-to-noise ratio (SNR) of tike et al. 2009; Velten et al. 2012; Heide et al. 2014a], BRDF estimameasured, Doppler-shifted signal is low.
  • SNR signal-to-noise ratio
  • D-ToF could be any control of illumination and on-sensor modulation.
  • D-ToF shares other limitations with ToF, including effect observed with conventional time-of-flight cameras within a the need for active illumination, limited range, and problematic single captured frame.
  • their technique is related to optical processing in the presence of strong ambient illumination or obflow methods that track features between successive video frames.
  • time-of-flight camera pixels provide a crucial fea ⁇ ture that makes them distinct from conventional camera pixels: be ⁇
  • Table 5 Notation table. fore being digitally sampled, the incident signal is modulated by a high-frequency, periodic function /,., (i) within each pixel. This on- sensor modulation is physically performed by an electric field that rapidly redirects incident photons-converted-to-electrons into one
  • ToF cameras are operated in a homodyne raode where the ventional optical flow by targeting lateral, rather than radial moillumination frequency and the reference frequency are identical: tion.
  • ojf tug Under tike common assumption of a stationary of object motion to estimate per-pixel radial velocity without the scene
  • Equation 3 simplifies need for optical flow.
  • Time-of-flight cameras operate in continuous wave mode. That is,
  • a light source illuminates the scene with an amplitude-modulated
  • the temporal low-pass filter recty ( ⁇ ) is convolved with the incident signal that changes periodically over time.
  • Sinusoidal waves are signal— an operation that is analogous to the finite integration area often used in the Top 7 literature to approximate the true shape of of each sensor pixel in the spatial domain 2 .
  • the modulated the signals are
  • Equation 5 signals in the supplemental material, we restrict the derivation in is independent of the time of measurement t,'. depth and albedo can this article to the sine wave mode! for simplicity of notation. Hence, be robustly estimated.
  • the light source emits a temporal signal of the form
  • Figure 2 Depth imaging. For static scenes, measurements are unFigure 3: Velocity imaging. Illumination > fertilization if ambiguous: different phase shifts result in unique intensity measfrequencies are designed to be orthogonal within the exposure time urements (top). For dynamic scenes, the Doppler shift results in a T. For static scenes (top), this particular choice of frequencies will low-frequency beating pattern thai makes measured intensities amintegrate to zero. The Doppler shift of moving scenes destroys the biguous, and hence prevents reliable depth estimation (bottom). orthogonality and results in an approximately linear relationship between radial velocity and recorded intensity (bottom).
  • heterodyne imaging i.e. having the exposure time T be an integer multiple of the period mode (e.g., Conroy et al. 2009]) may be beneficial in certain situof both signals. It is then easy to show from Equation 3 that ations, but a major limitation of heterodyne ToF is that multiple
  • the secondary modulation in the sensor fol compute the Doppler shift ⁇ from Equation 14.
  • the ratio image lowed by a low-pass filter of tike exposure time corresponds to tike r can be interpreted as a direct measurement of the instantaneous demodulation process in communication. Conventional communicper-pixel radial velocity.
  • ation channels use orthogonal frequencies; any inter-carrier interference (which could be caused by a frequency dri t) is a polluting
  • this approach still requires two measurements: one signal (see e.g. [Li and Stuber 2006]).
  • Doppler ToF we delibheterodyne image and one homodyne image.
  • instantaneous ence carries the desired velocity information.
  • two synchronized ToF sensors can be mounted in a co-axial setup; one of the sensors is modulated with the same
  • Figure 4 Simulated and measured intensities for a range of different velocities. Although the mapping from radial velocity to measured intensity is generally nonlinear (left), throughout a large range of velocities the conversion is approximately linear (center left).
  • a third possibility is to rapidly alternate between two modulation correct the ray,' phase measurements on a pe -pixel basis.
  • the measurements are not truly instantaneous, and alignment problems may occur for very fast motions.
  • the two measurements can be
  • time-of-flight camera system thai comprises a custom RF moduand the supplemental material. lated light source and a demodulation camera based on the PMD
  • the light source is an array of 650 rim laser diodes driven by iC-Haus con ⁇
  • Figure 6 Experimental verification of the imaging system for varythe described procedure and do not require additional processing. ing object velocities and depths (left) as well as velocity-dependent
  • Figure 7 shows another experiment, thai we used to verify the accuracy of our prototype D-ToF camera.
  • the exposure time of the ToF camera was set to 1.5 ms and die fan was captured from
  • Figure 7 Experimental validation of velocity estimation using a a frontal perspective (raw homodyne and heterodyne measurements fan with adjustable rotation speed (three settings). We measure the shown in Fig. 7 bottom). We manually measured tike slope of the fat- ground truth velocity o f the rotating blades (top left) by analyzing blades, which is constant over the entire blades. The radius of the audio recordings (top, lower left). The top right plot shows the veplotted position was measured, allowing us to calculate the "ground locity measured by D-ToF compared to the ground truth for a varytruth" velocity when the rotation speed of the fan is known. Since ing rotation speed.
  • the bottom row shows mounting a small pin on one of the blades and mounting a piece of the unprocessed full-field measurements of the homodyne (left) and flexible plastic in front of the fan, such that the rotating pin strikes the heterodyne (right) frequency setting with the pixel indicated for the plastic exactly once per revolution, creating a distinct sound. which we plotted the velocities on the top right.
  • This offset is calibrated in an offline process and raw phase measing with moving objects, the individual shots cannot, be assumed urements can be corrected digitally using a lookup table.
  • Figure 8 Complex scene with ambient illumination and a large depth range. The velocity is robustly estimated within the range of the illumination (approx. 5m inside), even in outdoor settings.
  • FIG 9 Velocity maps color-coded in grayscale. The maps comquires two frames to be captured, and they must be aligned if pondered from raw measurements ( top) are corrupted by Poisson noise. corded with a single camera, in some instances, such as Figures 10 To account for this, we apply a binning-based non-local means-type and 12, the alignment is challenging and any errors will propagdenoiser to the reconstructed velocity images (bottom). ate into the velocity maps, especially around depth-discontinuities.
  • the reconstructed velocity maps are color-coded; ab7 Towards the 3D Velocity Field solute units are indicated in the color bars. As expected, static
  • Figure 11 This result shows periodic motions in zfor a textured
  • Figure 14 Physical props for gaming, such as ping pong balls fired object. Although the estimated velocity is mostly correct, shadows with this toy gun, could be tracked and enable new HCI techniques. and dark scene parts are challenging for robust velocity estimation.
  • D-ToF is complimentary to optical flow. It allows for the deptis bias of xz-flow to be removed and enables recording of the metric where the optica? flow estimation is successful; in this case, we can 3D velocity field of the scene. However, if only radial velocity is pler Time-of-Flight, we hope to contribute a fundamentally new imaging modality that will have an impact on all of these applications. The possibility of implementing the proposed techniques on existing consumer devices makes Doppler Time-of-Flight a particularly attractive computational photography technique.
  • 3D metric flow is uniquely estimated from both 2D pixel flow and
  • A, B are the complex phasor amplitudes containing the amplitude the radial velocity maps.
  • the top images show a frame where opand phase dependency. Sui is the Kronecker delta; for perfectly- tical flow computed reasonable estimates. The bottom shows full static scenes, this expression is zero.
  • phasors can usu3D velocity estimate for different views.
  • the optical flow ally not be multiplied. By using the complex conjugate, this is helps us to determine that fan's velocity is slightly rotated to the possible while implicitly assuming that high frequency can be igupper right, 'where the center of rotatio is located (bottom left). nored [Ceperley 201 5] .
  • GUPTA ⁇ ., NAYAR, S. ., HULLIN, M., AND MARTIN, J. 2014.
  • Diffuse mirrors 3D reconstruction from diffuse indirect illuminVELTEN, A., WILLWACHER, T., GUPTA, O., VEERARAGHAVAN, ation using inexpensive time-of-flight sensors.
  • OSA Opt. Exp. VELTEN A. , Wu, D. , JARABO, A., MASIA, B . , BARSI, C , 22, 21 , 26338-26350.
  • KADAMB I A., WHYTE, R. , BHANDARI, A ., STREETER, L,, 2253.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Electromagnetism (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • General Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Optical Radar Systems And Details Thereof (AREA)

Abstract

Systems and methods for imaging object velocity are provided. In an embodiment, at least one Time-of-Flight camera is used to capture a signal representative of an object in motion over an exposure time. Illumination and modulation frequency of the captured motion are coded within the exposure time. A change of illumination frequency is mapped to measured pixel intensities of the captured motion within the exposure time, and information about a Doppler shift in the illumination frequency is extracted to obtain a measurement of instantaneous per pixel velocity of the object in motion. The radial velocity information of the object in motion can be simultaneously captured for each pixel captured within the exposure time. In one or more aspects, the illumination frequency can be coded orthogonal to the modulation frequency of the captured motion. The change of illumination frequency can correspond to radial object velocity.

Description

DOPPLER TIME-OF-FLIGHT IMAGING
CROSS-REFERENCE TO RELATED DOCUMENTS
[0001] This application claims priority to, and the benefit of, U.S. provisional patent application 62/282,708, filed August 7, 2015, the contents of which are incorporated herein by reference in its entirety.
[0002] This application also makes reference to and incorporates by reference and the following paper as if it were fully set forth herein expressly in its entirety: Doppler Time-of-Flight Imaging, Appendix B.
TECHNICAL FIELD
[0003] The present disclosure generally relates to depth cameras and motion of objects captured by such cameras.
BACKGROUND
[0004] Pioneers of photography, including Eadweard Muybridge and Harold "Doc" Edgerton, advanced imaging technology to reveal otherwise invisible motions of high-speed events. Today, understanding the motion of objects in complex scenes is at the core of computer vision, with a wide range of applications in object tracking, segmentation, recognition, motion de-blurring, navigation of autonomous vehicles, and defense.
[0005] Usually, object motion or motion parallax is estimated via optical flow [Horn and Schunck 1981 ]: recognizable features are tracked across multiple video frames. The computed flow field provides the basis for many computer vision algorithms, including depth estimation. Unfortunately, optical flow is computationally expensive, fails for untextured scenes that do not contain good features to track, and only measures 2D lateral motion perpendicular to the camera's line of sight. Further, the unit of optical flow is pixels; metric velocities cannot be estimated unless depth information of the scene is also available.
[0006] Over the last few years, depth cameras have become increasingly popular for a range of applications, including human-computer interaction and gaming, augmented reality, machine vision, and medical imaging. For the particular application of depth estimation, many limitations of optical flow estimation can be overcome using active illumination, as done by most structured illumination and time-of -flight (ToF) cameras where active illumination is temporally coded and analyzed on the camera to estimate a per-pixel depth map of the scene. With the emergence of RGB-D imaging, for example facilitated by Microsoft's Kinect One1 , complex and untextured 3D scenes can be tracked by analyzing both color and depth information, resulting in richer visual data that has proven useful for many applications. These approaches, however, still have limitations in the capture of motion.
Summary
[0007] We provide a fundamentally new imaging modality for depth cameras, in particular time-of-flight (ToF) cameras, and the capture of motion of objects. In an embodiment we provide per-pixel velocity measurement. Our technique can exploit the Doppler effect of objects in motion, which shifts the temporal frequency of the illumination before it reaches the camera. Using carefully coded illumination and modulation frequencies of the ToF camera, object velocities can directly map to measured pixel intensities.
[0008] In an embodiment our imaging system allows for color, depth, and velocity information to be captured simultaneously. Combining the optical flow computed on the RGB frames with the measured metric radial velocity allows estimation of the full 3D metric velocity field of the scene. The present technique has applications in many computer graphics and vision problems, for example motion tracking, segmentation, recognition, and motion de-blurring.
[0009] In an embodiment, provided is a method for imaging object velocity. The method can comprise the steps of: providing a Time-of-Flight camera and using the Time-of-Flight camera to capture a signal representative of an object in motion over an exposure time; coding illumination and modulation frequency of the captured motion within the exposure time; mapping a change of illumination frequency to measured pixel intensities of the captured motion within the exposure time; and extracting information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion. In any one or more aspects, radial velocity information of the object in motion can be simultaneously captured for each pixel captured within the exposure time. The illumination frequency can be coded orthogonal to the modulation frequency of the captured motion. The change of illumination frequency can correspond to radial object velocity.
[0010] In any one or more aspects, the Time-of-Flight camera can have a receiver and a transmitter, and the frequency of the receiver can be configured to be orthogonal to the frequency of the transmitter. The exposure time can be longer than the wavelength of a modulated captured signal. A ratio of a heterodyne measurement and a homodyne measurement can be determined to extract the information about the Doppler shift. The method can further include the step of: simultaneously capturing color, depth and velocity information concerning the object in motion during the exposure time. The change of illumination frequency can correspond to radial object velocity, and optical flow of the object in motion can be computed on red, green and blue (RGB) frames within a measured change in illumination frequency. The method can further include estimating a 3D velocity field for the object in motion. The depth and velocity imaging can be combined using either the Time-of-Flight camera by alternating modulation frequencies between successive video frames over the exposure time or using at least two Time-of-Flight cameras.
[0011] In an embodiment, we provide a system for imaging object velocity. The system can comprise: at least one device for capturing a signal representative of an object in motion over an exposure time; at least one computing device comprising a processor and a memory; and an application executable in the at least one computing device, the application comprising machine readable instructions stored in the memory that, when executed by the processor, cause the computing device to at least: (a) code illumination and modulation frequency of the captured motion within the exposure time; (b) map a change of illumination frequency to measured pixel intensities of the captured motion within the exposure time; and (c) extract information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion. The device can be at least one Time-of-Flight camera.
[0012] In an embodiment, we provide a non-transitory computer readable medium employing an executable application in at least one computing device, the executable application comprising machine readable instructions stored in the medium that: (a) receives signals representative of an object in motion over an exposure time; (b) codes illumination and modulation frequency of the captured motion within the exposure time; (c) maps a change of illumination frequency to measured pixel intensities of the captured motion within the exposure time; and (d) extracts information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion. The signals can be captured using at least one Time-of-Flight camera.
[0013] In any one or more aspects of the system or the computer readable medium, radial velocity information of the object in motion can be simultaneously captured for each pixel captured within the exposure time. The illumination frequency can be coded orthogonal to the modulation frequency of the captured motion. The change of illumination frequency can correspond to radial object velocity. The Time-of-Flight camera can include a receiver and a transmitter, and the frequency of the receiver can be configured to be orthogonal to the frequency of the transmitter. The logic can capture color, depth and velocity information concerning the object in motion during the exposure time.
[0014] Other systems, methods, features, and advantages of the present disclosure for Doppler time-of-flight imaging, will be or become apparent to one with skill in the art upon examination of the following drawings and detailed description. It is intended that all such additional systems, methods, features, and advantages be included within this description, be within the scope of the present disclosure, and be protected by the accompanying claims. BRIEF DESCRIPTION OF THE DRAWINGS
[0015] Many aspects of the disclosure can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present disclosure. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views.
[0016] Fig. 1A depicts an embodiment of an imaging system of the present disclosure that allows for metric radial velocity information to be captured. Depicted in the embodiment are a high-speed illumination source, RGB camera, and Time-of-Flight (ToF) camera.
[0017] Fig. 1 B depicts images that can captured by an imaging system of the present disclosure of an example of an object in motion.
[0018] Fig. 1 C depicts velocity data that can captured by an imaging system of the present disclosure of an example of an object in motion.
[0019] Fig. 1 D depicts depth data that can captured by an imaging system of the present disclosure of an example of an object in motion.
[0020] Fig. 1 E depicts images that can captured by an imaging system of the present disclosure of an example of an object in motion.
[0021] Fig. 1 F depicts velocity data that can captured by an imaging system of the present disclosure of an example of an object in motion.
[0022] Fig. 1 G depicts depth data that can captured by an imaging system of the present disclosure of an example of an object in motion.
[0023] Fig. 2A depicts an embodiment of depth imaging of the present disclosure for a static scene. [0024] Fig. 2B depicts an embodiment of depth imaging of the present disclosure for a scene in motion.
[0025] Fig. 3A depicts an embodiment of velocity imaging of the present disclosure for a static scene.
[0026] Fig. 3B depicts an embodiment of velocity imaging of the present disclosure for a scene in motion.
[0027] Fig. 4A depicts examples of simulated intensities for a large range of different velocities.
[0028] Fig. 4B depicts examples of simulated intensities for a small range of different velocities.
[0029] Fig. 4C depicts examples of measured intensities for a small range of different velocities.
[0030] Fig. 4D depicts examples of measured intensities for a smaller range of different velocities than Fig. 4C.
[0031] Fig. 5A depicts depth-dependent offset introduce by higher-order frequency components for a modulation frequency of 30 MHz.
[0032] Fig. 5B depicts depth-dependent offset introduce by higher-order frequency components for a modulation frequency of 80 MHz.
[0033] Fig. 5C depicts depth-dependent offset introduce by higher-order frequency components for a modulation frequency of 150 MHz.
[0034] Fig. 6A depicts an experimental verification of the imaging system for varying object velocities and depths (left)
[0035] Fig. 6B depicts measured intensities for a range of different pixel locations and velocity-dependent behavior for a range of different pixel locations on the sensor (right). [0036] Fig. 7A depicts an embodiment of experimental setup used for an experimental validation of velocity estimation using a fan with adjustable rotation speed (three settings).
[0037] Fig. 7B audio recordings analyzed to generate ground truth velocity data of the rotating blades of the setup in Fig. 7 A
[0038] Fig. 7C shows the velocity measured by D-ToF compared to the ground truth for a varying rotation speed.
[0039] Fig. 7D shows the unprocessed full-field measurements of the homodyne frequency setting with the pixel indicated for which velocities were plotted in Fig. 7C.
[0040] Fig. 7E shows the unprocessed full-field measurements of the heterodyne frequency setting with the pixel indicated for which velocities were plotted in Fig. 7C.
[0041] Fig. 8A depicts images of motion within a complex scene with ambient illumination and a large depth range.
[0042] Fig. 8B depicts velocity data of motion within a complex scene with ambient illumination and a large depth range The velocity can be robustly estimated within the range of the illumination (approx. 5m inside), even in outdoor settings.
[0043] Fig. 9A depicts computed velocity maps encoded in grayscale from raw measurements.
[0044] Fig. 9B depicts reconstructed de-noised images based on the velocity maps encoded in grayscale from raw measurements of Fig. 9A.
[0045] Fig. 10A depicts an example of periodic motion of a hand along the optical axis. The static scene on the left results in no response of the sensor, whereas forward (center) and backward (right) movement result in positive and negative responses respectively.
[0046] Fig. 10B depicts velocity data for the example in Fig. 10A.
[0047] Fig. 1 1 A depicts an example of periodic motion along the Z-axis for a textured object. Although the estimated velocity is mostly correct, shadows and dark scene parts can be challenging for robust velocity estimation.
[0048] Fig. 1 1 B depicts velocity data for the example in Fig. 1 1 A.
[0049] Fig. 1 1 C depicts depth data for the example in Fig. 1 1 A.
[0050] Fig. 12A shows an example of extremely fast motion that can be accurately captured with the present system.
[0051] Fig. 12B shows velocity data for the example in Fig. 12A. The Airsoft gun in the example is advertised as shooting bullets with 99 m/s; a radial velocity of 98.2 m/s (average of the peak pixels) can be calculated in the present example using the system and methods of the present disclosure.
[0052] Fig. 13A depicts an example of a potential applications of the present disclosure, including gaming and human - computer interaction. An example of a person in motion in a scene is depicted from left to right.
[0053] Fig. 13B shows velocity data for the example of Fig. 13A.
[0054] Fig. 13C shows depth data for the example of Fig. 13A.
[0055] Fig. 14A depicts an example of physical props for gaming, such as ping pong balls fired with a toy gun, which can be tracked with the present system and enable HCI techniques. Images of props in motion with a person in a scene is show across time from left to right.
[0056] Fig. 14B shows velocity data for the example of Fig. 14A.
[0057] Fig. 14C shows depth data for the example of Fig. 14A. [0058] Fig. 15A depicts a failure case of optical flow for a moving, but un- textured, scene (left).
[0059] Fig. 15B shows Optical flow [Liu 2009] for two succeeding frames of the scene from Fig. 15A. The 2D flow vectors can be color-coded with a color wheel (insets).
[0060] Fig. 15C shows SIFT flow [Liu et al. 2008] for two succeeding frames of the scene from Fig. 15A. The 2D flow vectors can be color-coded with a color wheel (insets).
[0061] Fig. 15D shows velocity data from the scene of Fig. 15A according to the system and methods described herein.
[0062] Fig. 16A depicts an example of a frame where optical flow computed reasonable estimates.
[0063] Fig. 16B show the full 3D velocity estimate for different views of the example in Fig. 16A. Optical flow can aid in 3D velocity estimates and image reconstruction.
[0064] Fig. 17 is a flowchart depicting an embodiment of a method of the present disclosure.
[0065] Fig. 18 depicts an embodiment of an apparatus that can be used in the systems and methods of the present disclosure.
[0066] Fig. 19 shows an embodiment of a camera system according to the present disclosure.
[0067] Fig. 20 shows an embodiment of a camera system according to the present disclosure.
[0068] Fig. 21 shows an embodiment of a camera system according to the present disclosure. DETAILED DESCRIPTION
[0069] Described below are various embodiments of the present systems and methods for Doppler Time-of-Flight (ToF) imaging. Although particular embodiments are described, those embodiments are mere exemplary implementations of the system and method. One skilled in the art will recognize other embodiments are possible. All such embodiments are intended to fall within the scope of this disclosure. Moreover, all references cited herein are intended to be and are hereby incorporated by reference into this disclosure as if fully set forth herein. While the disclosure will now be described in reference to the above drawings, there is no intent to limit it to the embodiment or embodiments disclosed herein. On the contrary, the intent is to cover all alternatives, modifications and equivalents included within the spirit and scope of the disclosure.
Discussion
[0070] Before the present disclosure is described in greater detail, it is to be understood that this disclosure is not limited to particular embodiments described, as such may, of course, vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting, since the scope of the present disclosure will be limited only by the appended claims.
[0071] Where a range of values is provided, it is understood that each intervening value, to the tenth of the unit of the lower limit (unless the context clearly dictates otherwise), between the upper and lower limit of that range, and any other stated or intervening value in that stated range, is encompassed within the disclosure. The upper and lower limits of these smaller ranges may independently be included in the smaller ranges and are also encompassed within the disclosure, subject to any specifically excluded limit in the stated range. Where the stated range includes one or both of the limits, ranges excluding either or both of those included limits are also included in the disclosure.
[0072] Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs. Although any methods and materials similar or equivalent to those described herein can also be used in the practice or testing of the present disclosure, the preferred methods and materials are now described.
[0073] All publications and patents cited in this specification are herein incorporated by reference as if each individual publication or patent were specifically and individually indicated to be incorporated by reference and are incorporated herein by reference to disclose and describe the methods and/or materials in connection with which the publications are cited. The citation of any publication is for its disclosure prior to the filing date and should not be construed as an admission that the present disclosure is not entitled to antedate such publication by virtue of prior disclosure. Further, the dates of publication provided could be different from the actual publication dates that may need to be independently confirmed.
[0074] As will be apparent to those of skill in the art upon reading this disclosure, each of the individual embodiments described and illustrated herein has discrete components and features which may be readily separated from or combined with the features of any of the other several embodiments without departing from the scope or spirit of the present disclosure. Any recited method can be carried out in the order of events recited or in any other order that is logically possible.
[0075] It is to be understood that, unless otherwise indicated, the present disclosure is not limited to particular materials, manufacturing processes, or the like, as such can vary. It is also to be understood that the terminology used herein is for purposes of describing particular embodiments only, and is not intended to be limiting. It is also possible in the present disclosure that steps can be executed in different sequence where this is logically possible.
[0076] It must be noted that, as used in the specification and the appended claims, the singular forms "a," "an," and "the" include plural referents unless the context clearly dictates otherwise. Thus, for example, reference to "a support" includes a plurality of supports. In this specification and in the claims that follow, reference will be made to a number of terms that shall be defined to have the following meanings unless a contrary intention is apparent.
Description
[0077] Provided herein is a new approach to directly imaging radial object motion. In an aspect, the motion can be velocity. In an aspect, a Doppler effect can be analyzed in one or more Time-of-Flight cameras: object motion towards or away from the cameras can shift the temporal illumination frequency before it is recorded by the camera. Conventional Time-of-Flight cameras encode phase information (and therefore scene depth) into intensity measurements. Instead, in various aspects herein Doppler Time-of-Flight (D-ToF) is used to provide a new imaging mode, whereby the change of illumination frequency (corresponding to radial object velocity) can be directly encoded into the measured intensity. In an aspect, the camera hardware utilized can be the same as for conventional Time- of-Flight imaging, but illumination and modulation frequencies can be carefully designed. Depth and velocity imaging can be combined using either two Time-of- Flight cameras or using the same device by alternating the modulation frequencies between successive video frames; color images can be obtained with a conventional camera.
[0078] In various aspects, a fundamentally new imaging modality is provided that is ideally suited for fast motion. Optical flow applied to conventional RGB video is a complimentary technique: together, optical flow and D-ToF allow for the metric 3D velocity field to be estimated, which is otherwise not easily possible. In general, however, the present D-ToF can be independent of the RGB flow and can work robustly for cases where optical flow often fails, including untextured scenes and extremely high object velocities.
[0079] Doppler radar is widely used in police speed guns, although gradually being replaced by lidar-based systems. Doppler lidar is also commonly used in many meteorological applications, such as wind velocity estimation. One common limitation of all Doppler measurements is that only movement along one particular direction, usually the line-of-sight, can be detected. All of these applications rely on the wave nature of light or sound, and therefore require coherent illumination or precise spectroscopic measurement apparatuses. In one or more aspects herein, incoherent, amplitude-modulated illumination and inexpensive time-offlight (ToF) cameras can be used for instantaneous imaging of both velocity and range. In various aspects, a full-field imaging method is provided, meaning that it does not require the scene to be sequentially scanned unlike most existing Doppler radar or lidar systems that only capture a single scene point at a time.
[0080] In an aspect, a framework and a camera system are provided implementing the described techniques; together, they can optically encode object velocity into per-pixel measurements of modified time-of-flight cameras. By combining multiple cameras, color, range, and velocity images can be captured simultaneously.
[0081] Pandharkar et al. [201 1 ] recently proposed a pulsed femtosecond illumination source to estimate motion of non-line-of-sight objects from differences in multiple captured images. In contrast, in an aspect, the present systems and methods can use the Doppler effect observed with conventional time-of-flight cameras within a single captured frame, as opposed to optical flow methods that track features between successive video frames.
[0082] Optical flow [Horn and Schunck 1981 ; Barron et al. 1994] is a fundamental technique in computer vision that is vital for a wide range of applications, including tracking, segmentation, recognition, localization and mapping, video interpolation and manipulation, as well as defense. Optical flow from a single camera is restricted to estimating lateral motion whereas the Doppler is observed only for radial motion towards or away from the camera.
[0083] Wei et al. [2006] and Hontani et al. [2014] have demonstrated how to use correlation image sensors to estimate optical flow of fast motion. Although correlation image sensors are conceptually similar to ToF cameras, their methods are more similar in spirit to conventional optical flow by targeting lateral, rather than radial motion. [0084] In contrast to these methods, in an aspect the present systems and methods can use the Doppler effect of object motion to estimate per-pixel radial velocity without the need for optical flow. Lindner and Kolb [2009] as well as Hoegg et al. [2013] estimate lateral optical flow to compensate for object motion between the sequentially-captured ToF phase images from which depth is usually estimated. A similar strategy can be applied herein to mitigate alignment artifacts when sub-frames are captured sequentially, but the flow is not a core part of D-ToF.
[0085] In one or more aspects, also provided herein is a mode for simultaneous range and velocity imaging. As with standard ToF imaging, the present method can involve the capture of a few sub-frames with different modulation signals. Using appropriate hardware (for example, multi-sensor cameras or custom sensors with different patterns multiplexed into pixels of a single sensor), the method can be implemented as a true snapshot imaging approach. In the present systems and methods, rapid time-sequential, (for example, 30-60 frames per second, and even higher with specialized equipment) can be used to capture the required sub-frames.
[0086] In summary among other things:
- D-ToF is presented herein as a new modality of computational photography that allows for direct estimation on instantaneous radial velocity. In an aspect using multiple captures or implemented with multi-sensor setups, it can record velocity, range, and color information. - A framework for velocity estimation with Time-of-Flight cameras is provided, along with a Time-of-Flight imaging system, and the framework and system validated in simulation and with the system.
- Evaluation of the imaging system using a range of different types of motion, for textured and untextured surfaces as well as indoors and under strong outdoor ambient illumination is also provided.
- It is demonstrated that the velocities measured with our system and method can be combined with RGB flow, allowing for the metric 3D velocity field to be estimated on a per-pixel basis.
Time-of-Flight Imaging
[0087] Time-of-flight cameras operate in continuous wave mode. That is, a light source illuminates the scene with an amplitude-modulated signal that changes periodically over time. Sinusoidal waves are often used in the ToF literature to approximate the true shape of the signals. We restrict the derivation herein to the sine wave model for simplicity of notation. Hence, the light source emits a temporal signal of the form
Figure imgf000019_0001
where <¾ is the illumination frequency. Assuming that the emitted light is reflected along a single, direct path by a stationary diffuse object at distance d, and that it is observed by a camera co-located with the light source, the signal reaching the camera is
Figure imgf000019_0002
(2) with s0 = go + b , where b is the ambient illumination. In the case of a stationary scene, the frequency at the camera is the same as the illumination frequency: <¾ = dig . In Equation 2, the amplitude sx combines the illumination amplitude gt , geometric factors such as the square distance falloff, as well as the albedo of the object. Due to the propagation distance, the phase of the received signal is shifted by φ = -2d/c <¾ .
[0088] Theoretically, s(t) can be directly sampled to estimate <p. However, illumination frequencies are usually in the order of tens to hundreds of MHz. Conventional solid state image sensors only provide sampling rates that are orders of magnitudes lower, and are hence inadequate for direct sampling of the phase. To overcome this limitation, Time-of-Flight camera pixels can provide a feature that makes them distinct from conventional camera pixels: before being digitally sampled, the incident signal can be modulated by a high-frequency, periodic function fip(t) within each pixel. In various aspects, the modulation frequency can be 10 MHz - 1 GHz, 10 MHz - 800 MHz, 10 MHz - 600 MHz, 10 MHz - 500 MHz, 10 MHz - 400 MHz, 10 MHz - 300 MHz, 10 MHz - 200 MHz, or 10 MHz - 100 MHz.
[0089] This on-sensor modulation can be physically performed by an electric field that rapidly redirects incident photons-converted-to-electrons into one of two buckets within each pixel. The phase and frequency ωΐ of the modulation function are programmable. The general equation for the modulated signal is thus
Figure imgf000021_0001
[0090] Usually, ToF cameras are operated in a homodyne mode where the illumination frequency and the reference frequency are identical: ™ w® = Under the common assumption of a stationary scene, we moreover get ω* ~ ~ ω·* and Equation 3 simplifies to
Figure imgf000021_0002
[0091] To model the discretely sampled quantities measured by the sensor, we can account for a finite integration (exposure) time. The exposure time T of all cameras can act as a low-pass filter on the modulated signal before it is discretized by the sampling process of the sensor. Since the exposure time is usually significantly longer than the wavelength of the modulated signal
-> ···· · all frequency-dependent terms in Equation 4 vanish:
Figure imgf000021_0003
[0092] The temporal low-pass filter rsc¾1') is convolved with the incident signal— an operation that is analogous to the finite integration area of each sensor pixel in the spatial domain. In the optics community, the low-pass filter resulting from spatial sensor integration is known as the detector footprint modulation transfer function [Boreman 2001 ]. Finally, the modulated and low- pass filtered signal can be discretely sampled. Since Equation 5 is independent of the time of measurement t', depth and albedo can be robustly estimated.
[0093] To distinguish the continuous function K" I from its discrete counterpart, we denote the latter as ?n ^ For depth estimation, two measurements asd an(j ίέ can be made that are usuallly recorded in quick succession, such that phase and depth can be estimated as
Figure imgf000022_0001
[0094] The same measurements can also be used to estimate the albedo:
4- !*5 (7)
[0095] More detailed discussions of the basic principle of operation of Time- of-Flight cameras can be found in the literature [Lange and Seitz 2001 ; Gokturk et al. 2004; Btittgen and Seitz 2008].
Time-of-Flight for Objects in Motion
[0096] The conventional Time-of-Flight image formation model breaks down when objects of interest move with a non-negligible radial velocity. In this case, the illumination frequency undergoes a Doppler shift [Doppler 1842] when reflected from an object in motion. The illumination arriving at the sensor is now frequency-shifted to ~ 1:" where the change in temporal frequency Δω depends on the radial object velocity as well as the illumination frequency: (8)
[0097] Consider the case of an approximately constant velocity v throughout the exposure time. If one assumes a homodyne setting with — — ki> Equation 3 can be used to derive a new version of the low-pass filtered sensor image (Eq. 5) for moving scenes: j S¾S — Δ^Ι ψ— 9
[0098] Note that this equation is now dependent on the time of measurement. Unfortunately, the introduced temporal intensity variation makes it more difficult to estimate phase and therefore also depth. In audio signal processing, this time-dependent low-frequency artifact is known as a beating pattern. This is illustrated in Figs. 2A and 2B. For static scenes, measurements are unambiguous: different phase shifts result in unique intensity measurements (Fig. 2A). For dynamic scenes, the Doppler shift results in a low-frequency beating pattern that makes measured intensities ambiguous, and hence prevents accurate depth estimation (Fig. 2B).
[0099] The phase estimate from Equation 6 is then distorted as i l I as* till! i "Γ"?Γ" I " ~ J_L&?t Vi Q) where the distortion linearly depends on the (unknown) object velocity. Note that, in practice, the estimated phase for moving objects corresponds to its average throughout the exposure. [00100] To summarize, in the homodyne setup, where the frequency of the light source and the frequency of the camera reference signal are identical, the Doppler shift introduced by moving objects results in mismatched frequencies on the image sensor. This situation is closely related to hetereodyne Time-of-Flight imaging (e.g., [Dorrington et al. 2007]), which generalizes the conventional homodyne capture mode to arbitrary combinations of illumination and sensor modulation frequencies. For static scenes, the heterodyne imaging mode can be beneficial in certain situations, but a major limitation of heterodyne ToF is that multiple (>2) measurements have to be captured to reliably estimate phase and depth. Since the beating pattern is usually of very low frequency (for example, in the order of a few Hz at most velocities typical to indoor environments), a significant amount of time needs to pass between the two measurements for accurate phase estimation. For moving objects, the necessity to capture multiple images may place constraints on the velocity.
[00101] To facilitate reliable velocity estimation, in an embodiment a new computational Time-of-Flight imaging methodology is derived in the following section. Similar to orthogonal frequency-division multiplexing (OFDM), D-ToF uses illumination and on-sensor modulation frequencies that are orthogonal within the exposure time of the camera. Using these frequencies, a method is provided that allows per-pixel radial object velocity estimation.
[00102] As illustrated in Fig. 2B, the low-frequency beating pattern created by the Doppler effect makes it difficult or impossible to capture reliable Doppler frequency and phase information. Consider the following example: a road cyclist travels at a speed of " $ towards the camera. For an illumination frequency of MHz. ?s— SO - iO¾ - the observed Doppler shift is only
Figure imgf000025_0001
[00103] A frequency shift of only 1 :67 Hz may seem small enough to be safely ignored. However, we show in the following that even such a minute change contains valuable information that can be used for velocity estimation.
Velocity Imaging via Orthogonal Frequencies
[00104] Inspired by multiplexing techniques in digital communication, an unconventional way is devised to extract velocity information from the small Doppler shift observed by a ToF camera. In an embodiment, the camera system can be interpreted as a communication channel, and the illumination considered as a carrier signal. The carrier can be optically modified by moving objects— a change can be observed in carrier amplitude, phase, and frequency. The secondary modulation in the sensor followed by a low-pass filter of the exposure time can correspond to the demodulation process in communication. Conventional communication channels use orthogonal frequencies; any inter- carrier interference (which could be caused by a frequency drift) is a polluting signal. For Doppler ToF, the frequencies in the receiver and transmitter can be designed to be orthogonal, such that the (usually polluting) inter-carrier interference carries the desired velocity information. An example is shown in Figs. 3A and 3B.
[00105] For the application of direct velocity imaging, the measured signal for a stationary object can be zero (or a constant intensity offset). This can be achieved by operating the ToF camera in heterodyne mode with two orthogonal frequencies ωβ and ωί. While any two sine waves with frequencies <¾≠ ωί will be orthogonal for sufficiently long integration times, this is not the case for finite integrals (exposures) in the presence of low frequency beating patterns. Designing both frequencies to be orthogonal is done by setting
Figure imgf000026_0001
i.e. having the exposure time T be an integer multiple of the period of both signals. It is then easy to show from Equation 3 that
Figure imgf000026_0002
for stationary objects — ½)<· In practice, we set / = k + 1, and we set k depending on T and the desired frequency ωβ.
[00106] Given these two orthogonal frequencies the inter-carrier interference can be used to extract valuable information about the Doppler shift. This can be achieved by computing the ratio of a heterodyne measurement and a homodyne measurement. Using only the low frequency terms from Equation 3, this ratio can be expressed, without loss of generality and assuming an exposure interval of [0... 7], as:
Figure imgf000027_0001
since : (" " ωβ)τ = <* ~ 92n* a Δ^ < ~ «
[00107] Figs. 4A-D shows the model derived here. On the left side, the full model is seen without any approximations (i.e. without neglecting high frequency components in Eq. 14). Although the image formation is nonlinear, for a relatively large range of metric velocities (Fig. 4A) it is very well approximated (Fig. 4B, center left) by our linear model (Eq. 14). The model is verified experimentally by using the camera prototype (Figs. 4C and 4D, right). These particular measurements were captured with a static scene, and acquired with a modulation frequency of — $©M¾ and an illumination frequency of ¾ ~ ΜίΜΗζί Ηζ + ω^ -rhus the Doppler shift for an object moving at a specific velocity was programmed into the illumination frequency for this particular experiment. With known, orthogonal illumination and modulation frequencies ω9, ωί it is therefore straightforward to compute the Doppler Δω from Equation 14. The ratio image r can be interpreted as a direct measurement of the instantaneous per-pixel radial velocity. [00108] This approach can still require two measurements: one heterodyne image and one homodyne image. There are several possible solutions for either acquiring these truly simultaneously, or they can be acquired in quick succession. For instantaneous measurements, two synchronized ToF sensors can be mounted in a co-axial setup; one of the sensors is modulated with the same frequency as the light source (<¾), while the other uses a slightly different frequency ωΐ≠ ωβ . This approach is similar in spirit to multi-sensor HDR imaging [Tocci et al. 201 1 ].
[00109] Instead of using two distinct sensors, it is also possible to multiplex pixels with two different modulation frequencies onto the same image sensor, either in alternating scanlines or in a checkerboard pattern. Again, this concept is similar in spirit to techniques that have been proposed for HDR cameras [Yasuma et al. 2010; Gu et al. 2010].
[00110] A third possibility is to rapidly alternate between two modulation frequencies using a single ToF camera. In this case, the measurements are not truly instantaneous, and alignment problems can occur for very fast motions. However, the two measurements can be taken immediately after each other, as fast as the camera hardware allows, e.g. at 30 or 60 Hz. We follow this approach as it only requires a single ToF camera. However, we can also use a setup with multiple synchronized ToF cameras. Note that, similar to heterodyne depth estimation [Dorrington et al. 2007], the Doppler shift can also be estimated directly from the low-frequency beating pattern, but at the cost of requiring multiple measurements that are much more widely spaced in time (hence not suitable for velocity estimation). [00111] Finally, the model from Equation 14 may only hold for sinusoidal modulation functions. If other periodic signals are being used, additional harmonic frequency components are introduced, which can distort the measurements for both stationary and moving targets. However, these offsets are systematic and can be calibrated for a specific ToF camera/lights source combination (see Implementation Section herein).
Simultaneous Range and Velocity
[00112] In many applications it may be useful to obtain both velocity and range measurements at the same time. As in standard ToF imaging, this can be achieved by capturing a second homodyne measurement with the phase offset by π/2. Simultaneous range and velocity imaging therefore may involve a total of three measurements: a heterodyne image with ψ = 0 , a homodyne image with ψ = 0, and a homodyne image with ψ = π/2.
[00113] As discussed in the Time-of-Flight Imaging Section above, motion introduces a velocity-dependent distortion Αωί' of the depth measurement (Eq. 10). However, since the distortion linearly depends on the Doppler shift Δω, which is known from the velocity estimation step (Eq. 14), we can now correctly estimate the phase delay (and hence the depth) from Equation 10. This may only involve a single additional calibration step to obtain Αωί' for a specific velocity, which corresponds to estimating the time offset t' between the start of the exposure time and the reference time for signal generation in the camera and light source.
[00114] As mentioned, simultaneous velocity and range imaging may involve three distinct measurements. The illumination signal may be the same for all three measurements. Only the reference signal for the camera may change. As in the case of velocity-only imaging, this means that all three measurements can potentially be acquired at the same time using either multiple sensors with a shared optical axis, or a sensor design with interleaved pixels. If neither option is available, rapid frame-sequential imaging is also possible.
[00115] The following examples are put forth so as to provide those of ordinary skill in the art with a complete disclosure and description of how to perform the methods and use the compositions and compounds disclosed and claimed herein. Efforts have been made to ensure accuracy with respect to numbers (e.g., amounts, temperature, etc.), but some errors and deviations should be accounted for.
EXAMPLES
Implementation
[00116] Method. A generic embodiment of a method 100 according to the present disclosure is shown in Fig. 17. Briefly, first a Time-of-Flight camera is provided 103. The camera can be used to capture a signal representative of an object in motion over an exposure time. Second, illumination and modulation frequency within the exposure time for the captured motion are coded 106. Third, illumination frequency changes are mapped 109 to measured pixel intensities of the captured motion with the exposure time. Last, Doppler shift information in the illumination frequency is extracted 1 12 to obtain a measurement of instantaneous per pixel velocity of the object in motion.
[00117] Hardware. Hardware characteristics of the imaging system or Time- of-Flight camera as described herein can include an illumination unit, optics, an image sensor, driver electronics, an interface, and computational ability. The hardware of embodiments of imaging systems as described herein can be seen in Fig. 1A, Fig. 19, Fig. 20, and Fig. 21. An embodiment of a generic camera system is shown in Fig. 19. The embodiment shown in Fig. 19 can be tailored to different applications by changing the characteristics of the imaging sensor. In an embodiment, the imaging sensor of Fig. 19 can be a conventional RGB imaging sensor and therefore Fig. 19 can be an RGB camera. In another embodiment, the imaging sensor of Fig. 19 can be a sensor suitable for a Time- of-Flight camera, such as the PMD Technologies PhotonlCs 19k-S3 imaging sensor, and Fig. 19 can be a Time-of-Flight camera.
[00118] For all physical experiments, an experimental Time-of-Flight camera system was used that comprises a custom RF modulated light source and a demodulation camera based on the PMD Technologies PhotonlCs 19k-S3 imaging sensor (see Fig. 1A). The system allows for metric radial velocity information to be captured instantaneously for each pixel (center row). The illumination and modulation frequencies of a Time-of-Flight camera (left) to be orthogonal within its exposure time. The Doppler effect of objects in motion is then detected as a frequency shift of the illumination, which results in a direct mapping from object velocity to recorded pixel intensity. By capturing a few coded Time-of-Flight measurements and adding a conventional RGB camera to the setup, it can be demonstrated in Figs. 1 B-G that color, velocity, and depth information of a scene can be recorded simultaneously. The results of Fig. 1 B and Fig. 1G show several frames of two video sequences. For each example in Fig. 1 B and Fig. 1G, the left-most frame shows a static object (velocity map is constant), which is then moved towards (positive radial velocity) or away (negative velocity) from the camera. [00119] An illumination unit can be a light source which can be an array of 650 nm laser diodes driven by iC-Haus constant current driver chips, type ic-HG. A PMD CamBoard nano development kit was used with a clear glass sensor that has the near IR bandpass filter removed, in combination with an external 2- channel signal generator to modulate the sensor and synchronize the light source. The setup is similar to commercially-available Time-of-Flight cameras and the proposed algorithms can be easily implemented on those. Unfortunately, developers usually do not have access to illumination and modulation frequencies of these devices, requiring the construction of custom research prototype cameras. The maximum illumination and demodulation frequency of our prototype is 150 MHz, but we run all of the presented results with 30 MHz. The modulation signals are nearly sinusoidal, but contain multiple low-amplitude harmonic components. To avoid systematic errors in depth and velocity estimation, these components can be calibrated as described in the following.
[00120] Fig. 18, depicts an apparatus 1010 in which the Doppler Time-of- Flight imaging described herein may be implemented. The apparatus 1010 can contain the driver electronics and computational ability for the imaging system or Time-of-Flight camera as described herein. The apparatus 1010 may be embodied in any one of a wide variety of wired and/or wireless computing devices, multiprocessor computing device, and so forth. As shown in Fig. 18, the apparatus 1010 comprises memory 214, a processing device 202, a number of input/output interfaces 204, a network interface 206, a display 205, a peripheral interface 21 1 , and mass storage 226, wherein each of these devices are connected across a local data bus 210. The apparatus 1010 may be coupled to one or more peripheral measurement devices (not shown) connected to the apparatus 1010 via the peripheral interface 21 1 .
[00121] The processing device 202 may include any custom made or commercially available processor, a central processing unit (CPU) or an auxiliary processor among several processors associated with the apparatus 1010, a semiconductor based microprocessor (in the form of a microchip), a macroprocessor, one or more application specific integrated circuits (ASICs), a plurality of suitably configured digital logic gates, and other well-known electrical configurations comprising discrete elements both individually and in various combinations to coordinate the overall operation of the computing system.
[00122] The memory 214 can include any one of a combination of volatile memory elements (e.g., random-access memory (RAM, such as DRAM, and SRAM, etc.)) and nonvolatile memory elements (e.g., ROM, hard drive, tape, CDROM, etc.). The memory 214 typically comprises a native operating system 216, one or more native applications, emulation systems, or emulated applications for any of a variety of operating systems and/or emulated hardware platforms, emulated operating systems, etc. For example, the applications may include application specific software which may be configured to perform some or all of the Doppler Time-of-Flight imaging techniques described herein. In accordance with such embodiments, the application specific software is stored in memory 214 and executed by the processing device 202. One of ordinary skill in the art will appreciate that the memory 214 can, and typically will, comprise other components which have been omitted for purposes of brevity.
[00123] Input/output interfaces 204 provide any number of interfaces for the input and output of data. For example, where the apparatus 1010 comprises a personal computer, these components may interface with one or more user input devices 204. The display 205 may comprise a computer monitor, a plasma screen for a PC, a liquid crystal display (LCD) on a hand held device, or other display device.
[00124] In the context of this disclosure, a non-transitory computer- readable medium stores programs for use by or in connection with an instruction execution system, apparatus, or device. More specific examples of a computer- readable medium may include by way of example and without limitation: a portable computer diskette, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory), and a portable compact disc read-only memory (CDROM) (optical).
[00125] With further reference to Fig. 18, network interface device 206 comprises various components used to transmit and/or receive data over a network environment. For example, the network interface 206 may include a device that can communicate with both inputs and outputs, for instance, a modulator/demodulator (e.g., a modem), wireless (e.g., radio frequency (RF)) transceiver, a telephonic interface, a bridge, a router, network card, etc.). The apparatus 1010 may communicate with one or more computing devices via the network interface 206 over a network. The apparatus 1010 may further comprise mass storage 226. The peripheral 21 1 interface supports various interfaces including, but not limited to IEEE-1394 High Performance Serial Bus (Firewire), USB, a serial connection, and a parallel connection.
[00126] The apparatus 1010 shown in Fig. 18 can be electronically coupled to and in communication with a Time-of-Flight camera as shown in Fig. 19, 20, and 21. Data can be passed back and forth between the apparatus 1010 and the Time-of-Flight camera, wired (USB, Firewire, thunderbolt, SDI, Ethernet, for example) or wirelessly (Bluetooth or WiFi, for example). Alternatively, the apparatus 1010 can be a part of the Time-of-Flight camera. An imaging system as described herein can be comprised of a Time-of-Flight camera or a Time-of- Flight camera in communication with an apparatus such as the apparatus 1010. An imaging system as described herein can also include any conventional RGB camera and/or an illumination source. An RGB camera and/or illumination source can also electronically coupled to and in communication with an apparatus 1010 along with a Time-of-Flight camera in an embodiment of an imaging system.
[00127] An imaging system as described herein can be configured to record successive frames of a scene. The scene can contain one or more objects in motion. Successive frames of a scene can be still images or from a video constructed of continuous successive frames. Scenes can be captured by the Time-of-Flight camera or Time-of-Flight camera in conjunction with an RGB camera. Data from the camera[s] can be sent and processed by an apparatus such as the apparatus 1010, and the apparatus 1010 can compute, process, and/or reconstruct data captured by the camera[s]. Data captured by the camera[s] can be one or more signals representative of one or more objects in motion. The one or more signals can contain information relating to RGB images, velocity, and/or depth that are representative of a scene. Embodiments of the present imaging systems are shown in Fig. 1A, Fig. 19, and Fig. 21.
[00128] Correcting for Higher-order Harmonics. The present camera prototype has the drawback that the periodic modulation functions are not perfectly sinusoidal, although they are very close. In addition to the fundamental frequency, this introduces higher-order harmonic components to the modulation signal. Unfortunately, the higher-order components are generally not orthogonal, thus they can cause a phase-dependent offset. This offset can be calibrated for different modulation frequencies and phase shifts using a static target. The depth-dependent offsets can be plotted for different modulation frequencies in Figs. 5A-C. These offsets can be calibrated in a one-time offline process and then used to correct the raw phase measurement on a per-pixel on basis.
[00129] This offset can be calibrated in an offline process and raw phase measurements can be corrected digitally using a lookup table. Note that for relatively low modulation frequencies, such as 30 MHz, we find a fairly large depth range (around 1 m) to be almost independent of this offset. In practice, it is therefore relatively easy to remove the higher-order frequency components.
[00130] Calibrating Phase Response. As is standard practice in Time-of- Flight cameras, the physical intensity response can be calibrated for different phase shifts φ in an offline calibration. Following [Lindner and Kolb 2006], the physical intensity response can be measured for a phase sweep of the illumination frequency and fit a fifth-order polynomial to the measurements. This can be used as a lookup table for converting phase to depth rather than solving Equation 6 directly. With the present prototype, a notable zeroth-order component of the fitted polynomial can be measured, corresponding to fixed pattern phase noise. This is easily corrected.
[00131] Verification of Calibration Procedure. The two calibration procedures described above are performed for all spatial locations on the sensor independently. To verify the calibration routines, a static target was imaged and a frequency and phase sweep applied to the modulation function, simulating objects at different velocities and depths. The results shown in Figs. 4C-D demonstrate that the measured intensities for a constant phase but varying Doppler shift follow the model derived in the Doppler-based Velocity Imaging Section herein. Other than a small amount of noise, which is mostly due to a relatively low signal-to-noise ratio, the curve is linear and behaves as predicted. In Fig. 6A, measurements for a range of different phase offsets in the modulation frequency was verified experimentally. This simulates objects at various depths, as indicated in the legend. Finally, the velocity-dependent behavior was tested for a range of different pixels over the sensor location and show results in Fig. 6B. All of this data is captured using a large planar target perpendicular to the camera and sweeping the illumination frequency (to simulate different Doppler shifts) and phase (to simulate different object distances). The remaining variance over pixel locations and phases is minimal.
[00132] Figs. 7A-E show another experiment that was used to verify the accuracy of our D-ToF camera system. The experiment setup is shown in Fig. 7A. In this example, the speed of a rotating fan was adjusted and its blades imaged such that, throughout the time it takes for a single blade to move across a pixel, forward motion is observed by that pixel. The exposure time of the ToF camera was set to 1 .5 ms and the fan was captured from a frontal perspective (raw homodyne and heterodyne measurements shown in Fig. 7 bottom). The slope of the fan blades was manually measured, which is constant over the entire blades. The radius of the plotted position was measured, allowing calculation of the "ground truth" velocity when the rotation speed of the fan is known. Since the exact rotation speed is not actually known, it was measured by mounting a small pin on one of the blades and mounting a piece of flexible plastic in front of the fan, such that the rotating pin strikes the plastic exactly once per revolution, creating a distinct sound. The sound (sampled at 44 KHz, Fig. 7B) of this setup was measured (to estimate the ground truth velocity of the fan blades, observed by one pixel, which is compared with the corresponding D- ToF estimate (Fig. 7C). For this experiment, the estimation error is always below 0.2 m/s. Errors are mainly due to the low SNR of the measured Doppler-shifted signal.
[00133] Subframe Alignment. Although the required heterodyne and homodyne shots could be captured simultaneously using multi-sensor configurations, they are captured in an alternating fashion using the single- sensor solution used herein. Examples are shown in Figs. 7C-7D. Since moving objects are involved, the individual shots cannot be assumed to be perfectly aligned, which results in velocity artifacts around edges in the scene. The artifacts can be mitigated, although not completely removed, by computing a SIFT flow on the raw data and warping them to a reference frame. While not perfect, the SIFT flow delivered sufficiently good warps for most captures.
[00134] Denoising. With the present system, an extremely small frequency shift (in the Hz range; for example a few Hz; for example 20 Hz or less, 15 Hz or less, 10Hz or less, 7 Hz or less, 5 Hz or less) can be captured relative to the modulation frequency (the MHz range). Additionally, the quantum efficiency of emerging time-of-flight sensors is still far from that of modern solid state sensors [Erz and Jahne 2009]. Therefore, the slight Doppler shift in the present prototype can be affected by Poisson noise. Standard denoising methods fail in strong Poisson noise scenarios. In Figs. 9A-B, velocity maps are coded in grayscale. The maps computed from raw measurements (Fig. 9A) are corrupted by Poisson noise. To account for this, a binning-based non-local means-type denoiser (denoising strategy) was applied to all captured or reconstructed velocity images or maps (Fig. 9B).
Experimental Results
[00135] The results captured with our prototype imaging system are shown in Figs. 1A-G, 8A-B, 10A-B, 11A-C, 12A-B, 13A-C, 14A-C. The results validate the proposed imaging system for a variety of challenging indoor and outdoor scenes. Color images can be recorded with the same exposure time as the Time-of-Flight camera. Most of the scenes have a slight red tint. This is due to use of eye-safe red illumination in the visible spectrum. Like current commercial ToF cameras, future implementations of this system would most likely use invisible, near infrared wavelengths to encode velocity and depth information. The reconstructed velocity maps can be color-coded; absolute units can be indicated in the color bars. As expected, static scenes result in a constant velocity map whereas velocity is directly encoded in the measurements and subsequently reconstructed for each sensor pixel independently. In addition to the velocity maps, Figs. 1 D, 1G, 11C, 13C, and 14C also show the corresponding depth maps that can be estimated from an additional capture as well as the velocity maps (see Simultaneous Range and Velocity Section herein).
[00136] The selection of scenes shows a wide range of motion types that can be reconstructed with the proposed method, but it also highlights several challenges of D-ToF and ToF in general. D-ToF requires two frames were captured, and aligned, recorded with a single camera. In some instances, such as Figs. 10A-B and 12A-B, the alignment is challenging and any errors will propagate into the velocity maps, especially around depth-discontinuities. These artifacts can be mitigated by optimizing the camera firmware to minimizing switching time between the sub-frames or by using two co-axial ToF cameras. Objects with dark albedos, as for example observed in Fig. 11 A, are challenging for any ToF method because only a small amount of the coded illumination is reflected back to the camera. Similarly, shadows are challenging and can result in either no depth/velocity estimation or errors (sweater in Fig. 8A and regions between fingers in Fig. 13A). Whereas some of these limitations can be overcome with better hardware, others are inherent to the time-of-flight approach.
Towards the 3D Velocity Field
[00137] Optical flow computed from conventional video sequences estimates the 2D projection of the 3D flow field onto the image plane. The radial component is usually lost. Furthermore, optical flow is an ill-posed problem and may fail in many scenarios. Our Doppler ToF addresses two problems of optical flow: first, it can help in cases where optical flow fails either due to large displacements or missing scene structures. Second, the present method can also help in cases where the optical flow estimation is successful; in this case, the 3D metric flow can be recovered by combining metric radial velocity and the 2D optical pixel flow.
[00138] Fig. 15A shows a scene where regular optical flow [Liu 2009], as well as SIFT-flow [Liu et al. 2008], fail due to limited structure in the scene (Fig. 15B and 15C respectively). Both methods cannot recover the true 2D motion of the fan and wrongly segment the scene. The present orthogonal velocity estimation method successfully captures the velocity of the objects and also leads to a proper segmentation of the scene (Fig. 15D). Note that having additional depth estimates for conventional flow may only be of limited help since flat surfaces also do not deliver enough features for correspondence matching.
[00139] Fig. 16A shows a scene where the optical flow estimate is reasonable. In this case, the orthogonal component that our method captures completes the 2D spatial flow estimates and uniquely determines the full metric 3D flow. Given the optical flow estimates fx,fy for the horizontal and vertical image coordinates, one can compute the metric velocity vectors
Figure imgf000041_0001
F ' where F is the focal length of the lens and Z the corresponding depth estimate from our method (see [Honegger et al. 2013]). In conjunction with the velocity estimate vz in the orthogonal direction along the optical axis, the full 3D metric flow is ^ t ^n ^ ' An example is shown in Fig. 16B. Note that the optical flow helps determine that the fan's velocity is slightly rotated to the upper right, where the center of rotation is located (bottom left). Also note that 3D flow field is only as reliable as the estimated radial velocity and the RGB 2D flow.
[00140] In summary, provided herein is a new computational imaging modality that directly captures radial object velocity via Doppler Time-of-Flight Imaging. A variety of experimental results captured with a prototype camera system are demonstrated for different types of motions and outdoor settings. The methods are extensively validated in simulation and experiment. In an aspect, the optional combination of footage captured using an RGB camera with the depth and velocity output of the present coded Time-of-Flight camera system is shown. Together, this data can represent simultaneous per-pixel RGB, depth, and velocity estimates of a scene and allow for the 3D velocity field to be estimated. Applications in a wide range of computer vision problems, including segmentation, recognition, tracking, super-resolution, spatially-varying motion de-blurring, and navigation of autonomous vehicles are provided.
[00141] The present method is complimentary to optical flow. It allows for the depth bias of xz-flow to be removed and enables recording of the metric 3D velocity field of the scene. However, if only radial velocity is required, the present method can also be used stand-alone, independent of optical flow.
[00142] Commercially available ToF sensors today are low-resolution and their quantum efficiency and noise characteristics are not comparable with modern CMOS sensors. Future generations of ToF sensors are expected to deliver significantly higher image quality, which would directly benefit the present method as well. Higher modulation frequencies would directly improve the signal-to-noise ratio in our setup, because the Doppler effect is proportional to these frequencies. For eye-safe operation, laser diodes can be used that operate in the visible spectrum in combination with a ToF sensor that has its visible spectrum cutoff filter removed. The laser illumination is therefore visible in all of the RGB images as a red tint. The present system can also operate the Time-of- Flight camera in the near infrared spectrum, as is common practice in commercial ToF cameras. Finally, all presented techniques can be easily be implemented on consumer Time-of-Flight cameras with the appropriate level of access to the system firmware or driver software.
[00143] Conclusion. Time-of -flight cameras have entered the consumer market only a few years ago, but transformed the way machines perceive the world. Human-computer interaction, medical imaging, robotics and machine vision, navigation for self-driving cars and quadcopters, and many other fundamental computer vision tasks have seen dramatic improvements using these devices. With Doppler Time-of-Flight, we provide a fundamentally new imaging modality that can impact all of these applications. Implementation of our method on existing consumer devices makes Doppler Time-of-Flight an attractive computational photography technique.
[00144] As will be apparent to those of skill in the art upon reading this disclosure, each of the individual embodiments described and illustrated has discrete components and features which may be readily separated from or combined with the features of any of the other several embodiments without departing from the scope or spirit of the present disclosure. Any recited method can be carried out in the order of events recited or in any other order logically possible.
[00145] Ratios, concentrations, amounts, and other numerical data may be expressed in a range format. It is to be understood that such a range format is used for convenience and brevity, and should be interpreted in a flexible manner to include not only the numerical values explicitly recited as the limits of the range, but also to include all the individual numerical values or sub-ranges encompassed within that range as if each numerical value and sub-range is explicitly recited. To illustrate, a concentration range of "about 0.1 % to about 5%" should be interpreted to include not only the explicitly recited concentration of about 0.1 % to about 5 %, but also include individual concentrations (e.g., 1 %, 2%, 3%, and 4%) and the sub-ranges (e.g., 0.5%, 1 .1 %, 2.2%, 3.3%, and 4.4%) within the indicated range. In an embodiment, the term "about" can include traditional rounding according to significant figure of the numerical value. In addition, the phrase "about 'x' to 'y'" includes "about 'x' to about 'y'".
[00146] It should be emphasized that the above-described embodiments are merely examples of possible implementations. Many variations and modifications may be made to the above-described embodiments without departing from the principles of the present disclosure. All such modifications and variations are intended to be included herein within the scope of this disclosure and protected by the following claims.
Table 1 - Notation Table l)es€ripi.»ii
ffit) Olsiroiealfe m rnl at te light soaree
i) illumination signal indent' at the ¾F sensor hit} sensor refef^nc signal
<*¾ illiimi atiois freqesncj
seos r modulates fe|ae £¥
programmabl ase offset, lor seos r sgnal
«fepth-dspeiKfe«t phase shft in ifuHitealioR
Doppte fe isesc shift.
costtiMo s, lo -pas filte red sensor image dfecrglely-sais led, Imv-pass Altered, sensor Image
References
[00147] BARRON, J., FLEET, D., AND BEAUCHEMIN, S. 1994. Performance of optical flow techniques. IJCV 12, 1 , 43-77.
[00148] BOREMAN, G. D. 2001 . Modulation Transfer Function in Optical and ElectroOptical Systems. SPIE Publications.
[00149] BiJTTGEN, B., AND SEITZ, P. 2008. Robust optical time-of-flight range imaging based on smart pixel structures. IEEE Trans. Circuits and Systems 55, 6, 1512-1525.
[00150] CEPERLEY, P., 2015. Resonances, waves and fields.
; ^^ 201 1/04/28-valid-method-of- multiplying-two. html. [Online; accessed 20-January-2015].
[00151] DOPPLER, C. J. 1842. LJber das farbige Licht der Doppelsterne und einiger anderer Gestirne des Himmels. Abhandlungen der Konigl. Bohm. Gesellschaft der Wissenschaften 12, 2, 465-482.
[00152] DORRINGTON, A. A., CREE, M. J., PAYNE, A. D., CONROY, R. M., AND CARNEGIE, D. A. 2007. Achieving sub-millimetre precision with a solid- state full-field heterodyning range imaging camera. In Proc. Meas. Sci. Technol., vol. 18.
[00153] ERZ, M., AND JAHNE, B. 2009. Radiometric and spectrometric calibrations, and distance noise measurement of ToF cameras. In Dynamic 3D Imaging. Springer, 28-41 .
[00154] GOKTURK, S., YALCIN, H., AND BAMJI, C. 2004. A time-offlight depth sensor - system description, issues and solutions. In Proc. CVPR, 35-35.
[00155] GU, J., HITOMI, Y., MITSUNAGA, T., AND NAYAR, S. 2010. Coded Rolling Shutter Photography: Flexible Space-Time Sampling. In Proc. ICCP. [00156] GUPTA, M., NAYAR, S. K., HULLIN, M., AND MARTIN, J. 2014. Phasor Imaging: A Generalization Of Correlation-Based Time-of-Flight Imaging. Tech. rep., Jun.
[00157] HEIDE, F., HULLIN, M. B., GREGSON, J., AND HEIDRICH, W. 2013. Low-budget transient imaging using photonic mixer devices. ACM Trans. Graph. (SIGGRAPH) 32, 4, 45: 1-45: 10.
[00158] HEIDE, F., XIAO, L, HEIDRICH, W., AND HULLIN, M. B. 2014. Diffuse mirrors: 3D reconstruction from diffuse indirect illumination using inexpensive time-of-flight sensors. In Proc. CVPR.
[00159] HEIDE, F., XIAO, L, KOLB, A., HULLIN, M. B., AND HEIDRICH, W.
2014. Imaging in scattering media using correlation image sensors and sparse convolutional coding. OSA Opt. Exp. 22, 21 , 26338-26350.
[00160] HONEGGER, D., MEIER, L, TANSKANEN, P., AND POLLEFEYS,
M. 2013. An open source and open hardware embedded metric optical flow
CMOS camera for indoor and outdoor applications. In Proc. ICRA, IEEE, 1736-
1741 .
[00161] HONTANI, H., OISHI, G., AND KITAGAWA, T. 2014. Local estimation of high velocity optical flow with correlation image sensor. In Proc. ECCV, 235- 249.
[00162] HORN, B., AND SCHUNCK, B. 1981 . Determining optical flow.Artificial Intelligence 17, 185-203.
[00163] KADAMBI, A., WHYTE, R., BHANDARI, A., STREETER, L, BARSI, C, DORRINGTON, A., AND RASKAR, R. 2013. Coded time of flight cameras: sparse deconvolution to address multipath interference and recover time profiles. ACM Trans. Graph. (SIGGRAPH Asia) 32, 6. [00164] KIRMANI, A., HUTCHISON, T., DAVIS, J., AND RASKAR, R. 2009. Looking around the corner using transient imaging. In Proc. ICCV, 159-166.
[00165] LANGE, R., AND SEITZ, P. 2001 . Solid-state time-of-flight range camera. IEEE J. Quantum Electronics 37, 3, 390-397.
[00166] LINDNER, M., AND KOLB, A. 2006. Lateral and depth calibration of PMD-distance sensors. In Advances in Visual Computing. Springer, 524-533.
[00167] LIU, C, YUEN, J., TORRALBA, A., SMC, J., AND FREEMAN, W. T. 2008. SIFT flow: Dense correspondence across different scenes. In Computer Vision-ECCV 2008. Springer, 28-42.
[00168] LIU, C. 2009. Beyond pixels: exploring new representations and applications for motion analysis. PhD thesis, MIT.
[00169] NAIK, N., ZHAO, S., VELTEN, A., RASKAR, R., AND BALA, K. 201 1 . Single view reflectance capture using multiplexed scattering and time-of-flight imaging. ACM Trans. Graph. (SIGGRAPH Asia) 30, 6, 171 :1-171 : 10.
[00170] OTOOLE, M., HEIDE, F., XIAO, L, HULLIN, M. B., HEIDRICH, W., AND KUTULAKOS, K. N. 2014. Temporal frequency probing for 5d transient analysis of global light transport. ACM Trans. Graph. (SIGGRAPH) 33, 4, 87: 1- 87: 1 1 .
[00171] PANDHARKAR, R., VELTEN, A., BARDAGJY, A., LAWSON, E.,BAWENDI, M., AND RASKAR, R. 201 1 . Estimating motion and size of moving non-line-of-sight objects in cluttered environments. In Proc. CVPR, 265-272.
[00172] TOCCI, M., KISER, C, TOCCI, N., AND SEN, P. 201 1. A versatile HDR video production system. ACM Trans. Graph. (SIGGRAPH) 30, 4, 41 . [00173] VELTEN, A., WILLWACHER, T., GUPTA, 0., VEERARAGHAVAN, A., BAWENDI, M., AND RASKAR, R. 2012. Recovering three dimensional shape around a corner using ultrafast time-of-flight imaging. Nat Commun 745, 3.
[00174] VELTEN, A., WU, D., JARABO, A., MASIA, B., BARSI, C, JOSHI, C, LAWSON, E., BAWENDI, M., GUTIERREZ, D., AND RASKAR, R. 2013. Femto- photography: Capturing and visualizing the propagation of light. ACM Trans. Graph. (SIGGRAPH) 32, 4, 44: 1-44:8.
[00175] WEI, D., MASUREL, P., KURIHARA, T., AND ANDO, S. 2006. Optical flow determination with complex-sinusoidally modulated imaging. In Proc. ICSP, vol. 2.
[00176] WU, D., WETZSTEIN, G., BARSI, C, WILLWACHER, T., OTOOLE, M., NAIK, N., DAI, Q., KUTULAKOS, K., AND RASKAR, R. 2012. Frequency analysis of transient light transport with applications in bare sensor imaging. In Proc. ECCV, 542-555.
[00177] YASUMA, F., MITSUNAGA, T., ISO, D., AND NAYAR, S. K. 2010. Generalized assorted pixel camera: postcapture control of resolution, dynamic range, and spectrum. IEEE TIP 19, 9, 2241-2253.
APPENDIX A
In this Appendix, we derive expressions for the proposed imaging systems using phasor notation. These may be more fntuitive for some readers, especially those familiar with the communication literature. In particular, we assume orthogonal illumination and modulation frequencies and derive the measured intensity of a dynamic scene (see Eqs. 13, 14) as
Figure imgf000050_0001
Ay h are ¾e com lex phasor amplitudes containing the ampiitude and phase dependency. is the Kronecker delta; for perfectly static scenes, this expression is zero. Note that phasors can usually not be multiplied. By using the complex conjugate, this is possible while implicitly assuming that high frequency can be ignored [Ceperiey 20153.
Assuming that objects are moving, a Doppier shift ^ is introduced. Frequency shifts like this have been analyzed for OFDM as inter carrier interference. We get the following complex interference pattern
Figure imgf000050_0002
49 APPENDIX B
Figure imgf000051_0001
Felix Heide1'2,3 Wolfgang Heidrich2-1 Matthias Hullin4 Gordon Wetzstein3 1 University of British Columbia 2KAUST "Stanford University 4University of Bonn
Figure imgf000051_0002
Figure 1: We introduce a new computational imaging system that allows for metric radial velocity information to be captured instantaneously or each pixel (center mw). For this purpose, we design the temporal illumination and modulation frequencies of a time-of-fiight camera (left) to be orthogonal within its exposure time. The Doppler effect of objects in motion is then detected as a frequency shift of the illumination, which results in a mapping from object velocity to recorded pixel intensity. By capturing a few coded time-of-fiight measurements and adding a conventional RGB camera to the setup, we demonstrate that color, velocity, and depth information of a scene can be recorded simultaneously. The results above show several frames of two video sequences. For each example, the left-most frame shows static object (velocity map is constant), which is then moved towards (positive radial velocity) or away from (negative velocity) the camera.
Abstract 1 Introduction
Over the last few years, depth cameras have become increasingly Pioneers of photography, including Eadweard Muybridge and Harpopular for a range of applications, including human-computer inold "Doc" Edgerton, advanced imaging technology to reveal otherteraction and gaming, augmented reality, machine vision, and medwise invisible motions of high-speed events. Today, understanding ical imaging. Many of the commercially-available devices use the motion of objects in complex scenes is at the core of computer the time-of-fiight principle, where active illumination is temporally vision, with a wide range of applications in object tracking, segcoded and analyzed in the camera to estimate a per-pixei depth map mentation, recognition, motion deblurring, navigation of autonomof the scene. In this paper, we propose a fundamentally new imaous vehicles, and defense. Usually, object motion or motion parging modality for ail time-of-fiight (ToF) cameras: per-pixei radial allax are estimated via optical flow [Horn and Schunck 1981]: revelocity measure ment. The proposed technique exploits the Dopcognizable features are tracked across multiple video frames. The pler effect of objects in motion, which shifts the temporal illuminacomputed flow field provides the basis for many computer vision tion frequency before it reaches the camera. Using carefully coded algorithms, including depth estimation. Unfortunately, optical flow illumination and modulation frequencies of the ToF camera, object is computationally expensive, fails for untextured scenes that do not velocities directly map to measured pixel intensities. We show that contain good features to track, and it only measures 2D lateral moa slight modification of our imaging system allows for color, depth, tion perpendicular to the camera's line of sight. Further, the unit and velocity information to be captured simultaneously. Combinof optical flow is pixels; metric velocities cannot be estimated uning the optical flow computed on the RGB frames with the measless depth information of the scene is also available. For the parured metric radial velocity allows us to further estimate the full ticular application of depth estimation, many limitations of optical 3D metric velocity field of the scene. The proposed technique has How estimation can be overcome using active illumination, as done applications in many computer graphics and vision problems, for by most structured illumination and time-of-fiight (Top7) cameras. example motion tracking, segmentation, recognition, and motion With the emergence of RGB-D imaging, for example facilitated by- deblurring. Microsoft's Kinect One' , complex and untextured 3D scenes can be tracked by analyzing both color and depth information, resulting in
CR Categories: 1.3.3 [Computer Graphics]: Picture/Image richer visual data that has proven useful for many applications. Generation— Digitizing and scanning
In this paper, we introduce a new approach to directly imaging ra¬
Keywords: computational photography, time-of-fiight dial object velocity. Our approach analyzes the Doppler effect in time-of-fiight cameras: object motion towards or away from the camera shifts the temporal illumination frequency before it is recorded. Conventional time-of-fiight cameras e code phase information (and therefore scene depth) into intensity measurements, instead, we propose Doppler Time -of- Flight (D-ToF) as a new imaging mode, whereby the change of illumination frequency (corresponding to radial object velocity) is directly encoded into the measured intensity. The required camera hardware is the same as for
Figure imgf000051_0003
1 microsoft.com/en-us/kinectforwindows/ conventional time-of-flight imaging, hut illumination and modulajects with dark albedos, shadowed regions, and global illumination tion frequencies are carefully designed. We can combine depth and effects.
velocity imaging using either two time- of - flight cameras or using
the same device by alternating the modulation frequencies between
successive video frames; color images can be obtained with a con2 Related Work
ventional camera.
Doppier-effect Measurements Since Christian Doppler dis¬
Our technique offers a fundamentally new imaging modality that is covered that the spectrum of astronomical objects shifts dependideally suited for fast motion. Optical flow applied to conventional ing on their velocity [Doppler 1842], the Doppler effect has found RGB video is a complimentary technique: together, optical flow widespread use in astronomical imaging, meteorology, traffic law and D-ToF allow for the metric 3D velocity field to be estimated, enforcement, radiology, healthcare, and aviation. Doppler spectrowhich is otherwise not easily possible. In general, however, D-ToF scopy, for example, measures radial velocity of otherwise undetectis independent of the RGB flow and works robustly for cases where able planets by observing wavelength shifts of their respective stars. optical flow often fails, including untextured scenes and extremely The rate of expansion of the universe can be estimated by Dophigh object velocities. We also discuss a mode for simultaneous pler spectroscopy as well. Laser Doppler veJocimetry is a comrange and velocity imaging. As with standard ToF imaging, our mon technique in heailhcare, for example to measure blood flow. method requires a few subframes to be captured with different modUsually, this technique uses two crossed, coherent laser beams to ulation signals. Using appropriate hardware (multi-sensor cameras create a small volume of bright and dark fringe patterns; the rate or custom sensors with different patterns multiplexed into pixels of of intensity fluctuation of particles moving through this volume a single sensor), the method could be implemented as a true snapgives rise to their velocity. Doppler radar is widely used in poshot imaging approach. In our prototype system, we instead use lice speed guns, although gradually being replaced by lidar-based rapid time-sequential acquisition of the required subframes, which systems. Doppler lidar is also commonly used in many meteorois a common strategy for regular ToF imaging. In summary, tikis logical applications, such as wind velocity estimation. One compaper makes tike following contributions: mon !imitatiotk of all Doppler measurements is that only movement alotkg one particular direction, usually the line-of- sight, can be de¬
• We introduce D-ToF as new modality of computational phototected. All of these applications rely on the wave nature of light or graphy that enables instantaneous radial velocity estimation. sound, and therefore require coherent illumination or precise specUsing multiple captures or implemented with multi-sensor troscopic measurement apparatuses. We are the first to exploit incosetups, D-ToF records velocity, range, and color information. herent, amplitude-modulated illumination and inexpensive time-of- flight (ToF) cameras for instantaneous imaging of both velocity and
9 We derive a mathematical framework for velocity estimation
range. Our approach is a full-field imaging method, meaning that it with time-of-flight cameras, implement a prototype time-of- does not require the scene to be sequential!); scanned unlike most flight imaging system, and validate the proposed model exexisting Doppler radar or lidar systems that only capture a single tensively in simulation and with the prototype,
scene point at a time.
• We evaluate the imaging system using a range of different
types of motion, for textured and untextured surfaces as well Computaiioiiai Time-of-F!ig t Photography With consumer as indoors and under strong outdoor ambient illumination. time-of-flight cameras such as Microsoft's Kinect One becoming
• We demonstrate that the velocities measured with our system widely available, research on computational time-of-flight imaging can be combined with RGB flow, allowing for the metric 3D has become an emerging area throughout the last few years. New- velocity field to be estimated on a per-pixel basis. approaches to capture and visualize light transport have allowed physical lighting effects to be recorded and replayed [Velten et al. 2013; Heide et al. 2013] that the field of computer graphics has thus
1.1 Limitations far only been able to simulate. Detailed analyses of temporal light transport in the frequency domain [Wu et al. 2012] or in the pres¬
As a fundamentally new imaging modality, D-ToF implemented ence of global illumination [O' Toole et al. 2014; Gupta et al. 2014] with our experimental hardware has several limitations. Forehave facilitated entirely new imaging modalities. For example, difmost, tike resolution of the PMD sensor in our prototype is limficult inverse problems, such as non- line -of- sight imaging [ irmani ited to 160 X 120 pixels and the signal-to-noise ratio (SNR) of tike et al. 2009; Velten et al. 2012; Heide et al. 2014a], BRDF estimameasured, Doppler-shifted signal is low. Together, these limitation [Naik et al. 2011], descattering [Heide et al. 2014b], and multi- tions result in low-resolution and noisy footage. We apply state- path separation [Kadambi et al. 2013], have become tractable. of-the-art denoising strategies which filter out most of the noise
but sometimes result in "blobby" images. Further, D-ToF requires Itk this manuscript, we analyze an effect not studied in prior work on two frames to be acquired with different modulation frequencies. computational time-of-flight imaging: the Doppler shift of objects Currently, we capture these frames in sequence, whic results in in motion. We derive a mathematical framework and build a camera slight misalignment between the frames observed as velocity artiprototype implementing the described techniques; together, they alfacts around depth discontinuities. However, there is a clear path to low us to optically encode object velocity into per-pixel measureaddressing all of these challenges: low-cost time-of-flight sensors ments of modified time-of-fiight cameras. By combining multiple providing QVGA or higher resolutions and significantly improved cameras, we also demonstrate how to capture color, range, and venoise characteristics are already on the market. With access to siglocity images simultaneously.
nal control of illumination and on-sensor modulation, D-ToF could
be readily implemented on high-quality consumer ToF cameras. A technique loosely related to ours was recently proposed by Pand- Combining two synchronized ToF cameras with different frequenharkar et al. [2011]. Whereas they use a pulsed femtosecond ilcies would allow for misalignment artifacts to be mitigated. lumination source to estimate motion of non-line-of-sight objects from differences in multiple captured images, we use the Doppler
Nevertheless, D-ToF shares other limitations with ToF, including effect observed with conventional time-of-flight cameras within a the need for active illumination, limited range, and problematic single captured frame. In effect, their technique is related to optical processing in the presence of strong ambient illumination or obflow methods that track features between successive video frames. 31 combines the illumination amplitude g , geometric factors such as the square distance fa!loff, as well as the albedo of the object. Due to the propagation distance, the phase of the received signal is shifted by φ = -2d/c ug.
Theoretically, sit) could be directly sampled to estimate φ. However, illumination frequencies are usually in the order of tens to hundreds of MHz. Conventional solid state image sensors only provide sampling rates thai are orders of magnitudes lower, and are hence inadequate for direct sampling of the phase. To overcome this limitation, time-of-flight camera pixels provide a crucial fea¬
Figure imgf000053_0001
ture that makes them distinct from conventional camera pixels: be¬
Table 5 : Notation table. fore being digitally sampled, the incident signal is modulated by a high-frequency, periodic function /,., (i) within each pixel. This on- sensor modulation is physically performed by an electric field that rapidly redirects incident photons-converted-to-electrons into one
Optica! F!o in Computer Vision Optical How [Horn and
of two buckets within each pixel. The phase ψ and frequency u f of Schunck 1981 ; Barron ei al, 1994] is a fundamental technique in
the modulation function are programmable. The general equation computer vision that is vital for a wide range of applications, infor the modulated signal is thus
cluding tracking, segmentation, recognition, localization and mapping, video interpolation and manipulation, as well as defense. Optical flow from a single camera is restricted to estimating lateral
motion whereas tike Doppler is observed only for radial motion to'/',;, (ί)
Figure imgf000053_0002
· (,sj cos .»si + (i) - so) wards or away from the camera.
)ί + ψ - φ) + (3)
Wei et al. [2006] and Hontani et al. [2014] have demonstrated how- to use correlation image sensors to estimate optical flow of fast mo cos( ( >f -÷- w,)t + 'φ ÷ φ) + so <x (u>f t + ¾'') ·
2
tion. Although correlation image sensors are conceptually similar
to ToF cameras, their methods are more similar in spirit to conUsually, ToF cameras are operated in a homodyne raode where the ventional optical flow by targeting lateral, rather than radial moillumination frequency and the reference frequency are identical: tion. In contrast to these methods, we analyze the Doppler effect ojf = tug Under tike common assumption of a stationary of object motion to estimate per-pixel radial velocity without the scene, we moreover get w, = w„ = i, and Equation 3 simplifies need for optical flow. Lindner and Kolb [2009] as well as Hoegg to
et al. [2013] estimate lateral optical flow to compensate for object
motion between the sequentially-captured ToF phase images from s{2 rt + (, - 1p) + S{) COS 'i -÷- ? >) . which depth is usually estimated. We can apply a similar strategy to 2 2~
(4) mitigate alignment artifacts when subframes are captured sequenTo model the discretely sampled quantities measured by the sensor, tially, but the flow is not a core part of D-ToF. we must account for a finite integration (exposure) time. The exposure time T of all cameras acts as a low-pass filter on the mod¬
3 Review of Time-of-Flight Imaging ulated signal before it is discreiized by the sampling process of the sensor. Since the exposure time is usually significantly longer than
In this section, we first review the conventional ToF image formthe wavelength of the modulated signal T all frequency- ation model for static scenes and then analyze how it behaves for dependent terms in Equation 4 vanish:
objects in motion.
ίφ {ΐ') = f¾. * rectr ) ( « ~ cc (5)
Time-of-flight cameras operate in continuous wave mode. That is,
a light source illuminates the scene with an amplitude-modulated The temporal low-pass filter recty ( ·) is convolved with the incident signal that changes periodically over time. Sinusoidal waves are signal— an operation that is analogous to the finite integration area often used in the Top7 literature to approximate the true shape of of each sensor pixel in the spatial domain2. Finally, the modulated the signals. Although we derive a full mode! for arbitrary periodic
and low-pass-filtered signal is discretely sampled. Since Equation 5 signals in the supplemental material, we restrict the derivation in is independent of the time of measurement t,'. depth and albedo can this article to the sine wave mode! for simplicity of notation. Hence, be robustly estimated.
the light source emits a temporal signal of the form
To distinguish the continuous function ίψ (ί') from its discretizag(i)— gi cos( )gt) + g0. (1) tion, we denote the latter as i,p \t'} . Fo depth estimation, at least two measurements io \V] and i.„-/2 \i' \ are necessary that are usuwhere w„ is the illumination frequency. Assuming that the emitted ally recorded in quick succession, such that phase and depth can be light is reflected along a single, direct path by a stationary diffuse estimated as
object at distance d, and
with the light source, the s(i) = si
— si
Figure imgf000053_0003
with so = go + b, where 6 is the ambient illumination. In the case 'In the optics community, the low-pass filter resulting from spatial of a stationary scene, the frequency at the camera is the same as sor integration is known as the detector footprint modulation transfer the illumination frequency: tos = ug. In Equation 2, the amplitude ct-on [Boreman 2001].
Figure imgf000054_0001
Figure 2: Depth imaging. For static scenes, measurements are unFigure 3: Velocity imaging. Illumination >„ and modulation if ambiguous: different phase shifts result in unique intensity measfrequencies are designed to be orthogonal within the exposure time urements (top). For dynamic scenes, the Doppler shift results in a T. For static scenes (top), this particular choice of frequencies will low-frequency beating pattern thai makes measured intensities amintegrate to zero. The Doppler shift of moving scenes destroys the biguous, and hence prevents reliable depth estimation (bottom). orthogonality and results in an approximately linear relationship between radial velocity and recorded intensity (bottom).
More detailed discussions of the basic principle of operation of For intuition, we consider the case of an approximately constant time- of- flight cameras can be found in the literature [Lange and velocity v throughout the exposure time. If we continue to assume Seitz 2001; Gokturk et al. 2004; Bfittgen and Seitz 2008], a homodyne setting with ω ι = ω3 = ω, Equation 3 can be used to derive a new version of the low-pass- filtered sensor image (Eq. 5) for moving scenes:
Time-of-Fiight for Objects in Motion
ί,φ (ί') ss cos( -Awi' + Ψ - Φ). (9)
The conventional time-of-ffight image formation mode! breaks
down when objects of interest move with a non-negligible radial Note thai this equation is now dependent, on the time of measurevelocity. In this case, the illumination frequency undergoes a Dopment. Unfortunately, the introduced temporal intensity variation pler shift [Doppler 1842] when reflected from an object in motion. makes it more difficult to estimate phase and therefore also depth. The illumination arriving at the sensor is now frequency- shifted to In audio signal processing, this time-dependent low-frequency artiw, = u¾ 4- Αω, where the change in temporal frequency ω defact is known as a beating pattern. We illustrate it in Figure 2. pends on the radial object velocity as well as tike iUumination fre¬
The phase estimate from Equation 6 is then distorted as quency:
(8)
Figure imgf000054_0002
where the distortion AuA' linearly depends on the (unknown) obconstant intensity offset). We can achieve this by operating the ToF ject velocity. Note that, in practice, the estimated phase for moving camera in heterodyne mode with two orthogonal frequencies ug objects corresponds to its average throughout the exposure. atkd ω; . While any two sine waves with frequencies ω9 i; will be orthogonal for sufficiently long integration times, this is not the
To summarize, in the homodyne setup, where the frequency of the case for finite integrals (exposures) in the presence of low frequency light source and the frequency of the camera reference signal are beating patterns. Designing both frequencies to be orthogonal is identical, the Doppler shift introduced by moving objects results done by setting
in mismatched frequencies on the image sensor. This situation is
closely related to hetereodyne time-of-flight imaging (e.g., [Dar2 2π
lington et a!. 2007]), which generalizes the conventional homodyne with A:, 2 £ N, k =/ /, (12) capture mode to arbitrary combinations of illumination and sensor
modulation frequencies. For static scenes, the heterodyne imaging i.e. having the exposure time T be an integer multiple of the period mode (e.g., Conroy et al. 2009]) may be beneficial in certain situof both signals. It is then easy to show from Equation 3 that ations, but a major limitation of heterodyne ToF is that multiple
(>2) measurements have to be captured to reliably estimate phase
and depth. Since the beating pattern is usually of very low frei,p (t) di = 0 (13) quency, a significant amount of time needs to pass between the two
measurements for reliable phase estimation. For moving objects, for stationary objects (u s = ωβ). In practice, we set I = k + 1 the necessity to capture multiple images would place severe conatkd we set k = ω9Τ/2π, which depends on T and the desired straints on the velocity. To facilitate reliable velocity estimation, frequency ω,Ί.
we derive a new computational time-of-flight imaging methodology in the following section. Inspired by the general concept of Given these two orthogonal frequencies we now use the inter- orthogonal frequency-division multiplexing (OFDM, e.g. [Li and carrier interference to extract valuable information about the DopStuber 2006]), D-ToF uses illumination and on-sensor modulation pler shift. We achieve this by computing the ratio of a heterodyne frequencies that are orthogonal within the exposure time of the cammeasurement and a homodyne measurement. Using only the low- era. Using this choice of frequencies along with a newly-devised refrequency terms from Equation 3, this ratio can be expressed as3 : construction method, we demonstrate the first approach to per-pixe!
radial velocity estimation.
jg cos(w/i + ) (si cos((w? -f Δω)ί + φ)÷ so ) dt
4 Dopp!er-based Velocity Imaging fp ois(tjjgi + ψ)■ (si cos((u.>ff 4- Αω)ί + ) + So ) dt
Jo ~2 cos( (w/ - )g - A )t + ψ— φ) dt
As illustrated in Figure 2 (bottom), the low- frequency beating pattern created by the Doppler effect makes it difficult or impossible Jn* ¾" COs(—Awt -r ψ— Φ) dt
to capture reliable Doppler frequency and phase information. Consider the following example: a road cyclist travels at a speed of 2(ί).- ≠α ((ωί - ¾ - Δ-'ί + ' ~
Figure imgf000055_0001
v = 10— towards the camera. For an illumination frequency of
50 MHz (i.e. ω3 = 50 · iO6 · 2ir/s), the observed Doppler shift is —Αω
only - ω„ - Δ(
10- 2 sin ( (uj ~ !g)T— ΑωΤ 4- ψ — φ)— sin('¾/> - Φ)
Αω = 50 · 10' ¾ 1.67- (Π)
300 · sin(- -ΑωΤ + φ - φ) - siii(¾ - φ)
··.
A frequency shift of only 1.67 Hz may seem small enough to be
safely ignored. However, we show in the following that even such
a minute change contains valuable information that can be used for
velocity estimation. (14) since (ί. (k— I) 2 τ, and Αω -
4.1 Velocity imaging via Orthogonal Frequencies Figure 4 shows the model derived here. On the left side, we see the full model without any approximations (i.e. without neglecting
Inspired by multiplexing techniques in digital communication, we high frequency components in Eq. 14). Although the image formdevise an unconventional way to extract velocity information from ation is nonlinear, for a relative large range of metric velocities it the small Doppler shift observed by a ToF camera. We can interis very well approximated (Fig. 4, center left) by our linear model pret the camera system as a communication channel and consider (Eq. 14). We experimentally verify the model using our camera the illumination a carrier signal. The carrier is optically modified prototype (Fig. 4, right). With known, orthogonal illumination and by moving objects— we observe a change in carrier amplitude, modulation frequencies ujg, u<t, it is therefore straightforward to phase, and frequency. The secondary modulation in the sensor folcompute the Doppler shift Αω from Equation 14. The ratio image lowed by a low-pass filter of tike exposure time corresponds to tike r can be interpreted as a direct measurement of the instantaneous demodulation process in communication. Conventional communicper-pixel radial velocity.
ation channels use orthogonal frequencies; any inter-carrier interference (which could be caused by a frequency dri t) is a polluting We note that this approach still requires two measurements: one signal (see e.g. [Li and Stuber 2006]). For Doppler ToF, we delibheterodyne image and one homodyne image. There are several erately design the frequencies in the receiver and transmitter to be possible solutions for either acquiring these truly simultaneously, orthogonal, suc thai the (usually polluting) inter-carrier interferor they can be acquired in quick succession. For instantaneous ence carries the desired velocity information. measurements, two synchronized ToF sensors can be mounted in a co-axial setup; one of the sensors is modulated with the same
For the application of direct velocity imaging, we would like to
ensure that the measured signal for a stationary object is zero (or a Without loss of generality, we assume an exposure interval of |G .
Figure imgf000056_0001
Figure 4: Simulated and measured intensities for a range of different velocities. Although the mapping from radial velocity to measured intensity is generally nonlinear (left), throughout a large range of velocities the conversion is approximately linear (center left). We verify the predicted mapping using our prototype camera (right). These particular measurements were captured with, a static scene, and acquired with a modulation frequency ofuj ~- 60 MHz and an illumination frequency ofu., ----- 60 MHz + \ KHz -\- Δω, Thus, the Doppler shift for an object moving at a specific velocity was programmed into the illumination frequency for this particular experiment. frequency as the light source (u¾), while the other uses a slightly
different frequency if u¾. This approach is similar in spirit to
multi-sensor HDR imaging [Tocci et al. 2011].
Instead of using two distinct sensors, it would also be possible to
multiplex pixels with two different modulation frequencies onto the
same image sensor, either in alternating scanlines or in a checker
Figure imgf000056_0002
board pattern. Again, this concept is similar in spirit to techniques
that have been proposed for HDR cameras [Yasuma et al. 2010; Gu 5: Depth-dependent offset introduced by higher-order freet al. 2010]. quency components for a range of modulation frequencies. These offsets are calibrated in a one-time offline process and then used to
A third possibility is to rapidly alternate between two modulation correct the ray,' phase measurements on a pe -pixel basis.
frequencies using a single ToF camera. In this case, the measurements are not truly instantaneous, and alignment problems may occur for very fast motions. However, the two measurements can be
taken immediately after each other, as fast as the camera hardware As mentioned, simultaneous velocity and range imaging requires allows, e.g. at 30 or 60 Hz. We follow this approach as it only rethree distinct measurements. We note that the illumination signal quires a single ToF camera. Note that, similar to heterodyne depth is the same for all three measurements, only the reference signal estimation [Dorrington et al. 2007], the Doppler shift can also be for the camera changes. As in the case of velocity-only imaging, estimated directly from the low-frequency beating pattern, but at this means that all three measurements can potentially be acquired the cost of requiring multiple measurements that are ranch more at the same time using either multiple sensors with a shared optical widely spaced in time (hence not suitable for velocity estimation). axis, or a special sensor design with interleaved pixels, if neither option is available, rapid frame- sequential imaging is also possible.
Finally, we note that the model from Equation 14 only holds for
sinusoidal modulation functions. If other periodic signals are being
used, additional harmonic frequency components are introduced, 5 Imp!lementat on
which distort the measurements for both stationary and moving targets. However, these offsets are systematic and can be calibrated Hardware For all physical experiments, we use an experimental for a specific ToF camera/lights source combination (see Section 5, time-of-flight camera system thai comprises a custom RF moduand the supplemental material). lated light source and a demodulation camera based on the PMD
Technologies PhoionlCs 19k-S3 sensor (see Fig, J .). The light source is an array of 650 rim laser diodes driven by iC-Haus con¬
4.2 Simultaneous Range and Velocity stant current driver chips, type ic-HG. We use a PMD CamBoard nano development kit with a clear glass sensor that has the near in many applications it may be useful to obtain both velocity and IR bandpass filter removed, in combination with an external 2- range measurements at the same time. As in standard ToF imaging, channel signal generator to modulate the sensor and synchronize this can be achieved by capturing a second homodyne measurement the light source. Our setup is similar to commercially- available with the phase offset by /2. Simultaneous range and velocity imatime-of-flight cameras and the proposed algorithms could be easging therefore requires a total of three measurements: a heterodyne ily implemented on those. Unfortunately, developers usually do not image with φ = 0, a homodyne image with ψ = 0, and a homo- have access to illumination and modulation frequencies of these dyne image with = it/ 2. devices, requiring the construction of custom research prototype cameras. The maximum illumination and demodulation frequency
As discussed in Section 3, motion introduces a velocity-dependent of our prototype is 150 MHz, but we run all of the presented results distortion Δωί' of the depth measurement (Eq. 10). However, since with 30 MHz. The modulation signals are nearly sinusoidal, but the distortion linearly depends on the Doppler shift Δω, which is contain multiple low-amplitude harmonic components. To avoid known from the velocity estimation step (Eq, 14), we can now corsystematic errors in depth and velocity estimation, these componrectly estimate the phase delay (and hence the depth) from Equaents must be calibrated as described in the following.
tion 10. This only requires an additional calibration step to obtain
Δωί' for a specific velocity, which corresponds to estimating the
time offset t' between the start of the exposure time and the referCorrecting tor Higher-order Harmonics Our camera prototype ence time for signal generation in the camera and light source. has tike drawback that the periodic modulation functions are not tensities for
Range of Different Pixel Locations As is standard practice in time- of-flight cameras, we calibrate the physical intensity response for different phase shifts φ in an offline calibration. Following [Lindner and Kolb 2006], we measure tike physical intensity response for a phase sweep of the illumination frequency and fit a fifth-order polynomial to the measurements. This is used as a lookup table for converting phase to depth rather than solving Equation 6 directly. With our prototype, we measure a notable zeroth-order component of the fitted polynomial, corresponding to fixed pattern phase noise. This is easily corrected with the lookup table. Any other illumination-specific terms, for example introduced by the baseline
Figure imgf000057_0001
between camera and light source, are automatically calibrated with
Figure 6: Experimental verification of the imaging system for varythe described procedure and do not require additional processing. ing object velocities and depths (left) as well as velocity-dependent
behavior for a range of different pixel locations on the sensor
( right). All of this data is captured using a large planar target perVerif icatioit of Ca!ibration Procedure The two calibration propendicular to the camera and sweeping the illumination frequency cedures described above are performed for all spatial locations on (to simulate different Doppler shifts) and phase (to simulate differthe sensor independently. To verify our calibration routines, we ent object distances i. image a static target and apply a frequency and phase sweep to the modulation function, simulating objects at different velocities and depths. The results shown in Figure 4 (left) demonstrate that the measured intensities for a constant phase but varying Doppler shift follow the model derived in the Section 4. Other than a small amount of noise, which is mostly due to a relatively low signal-to- noise ratio, the curve is linear and behaves as predicted. In Figure 6 (left), we verify experimental measurements for a range of different phase offsets in the modulation frequency. This simulates objects at various depths, as indicated in the legend. Finally, we also test the velocity-dependent behavior for a range of different pixels over the sensor location and show results in Figure 6 (right). The remaining variance over pixel locations and phases is minimal.
Figure 7 shows another experiment, thai we used to verify the accuracy of our prototype D-ToF camera. In this example, we adjusted the speed of a rotating fan and imaged its blades such that, throughout the time it takes for a single blade to move across a
Figure imgf000057_0002
pixel, forward motion is observed by that pixel. The exposure time of the ToF camera was set to 1.5 ms and die fan was captured from
Figure 7: Experimental validation of velocity estimation using a a frontal perspective (raw homodyne and heterodyne measurements fan with adjustable rotation speed (three settings). We measure the shown in Fig. 7 bottom). We manually measured tike slope of the fat- ground truth velocity o f the rotating blades (top left) by analyzing blades, which is constant over the entire blades. The radius of the audio recordings (top, lower left). The top right plot shows the veplotted position was measured, allowing us to calculate the "ground locity measured by D-ToF compared to the ground truth for a varytruth" velocity when the rotation speed of the fan is known. Since ing rotation speed. As the speed becomes larger, estimation errors the exact rotation speed is not actually known, we measure it by increase to maximum of about 0.2 m/s. The bottom row shows mounting a small pin on one of the blades and mounting a piece of the unprocessed full-field measurements of the homodyne (left) and flexible plastic in front of the fan, such that the rotating pin strikes the heterodyne (right) frequency setting with the pixel indicated for the plastic exactly once per revolution, creating a distinct sound. which we plotted the velocities on the top right. We record the sound (sampled at 44 KHz) of this setup to estimate the ground truth velocity of the fan blades, observed by one pixel, which is compared with the corresponding D-ToF estimate (Fig. 7, perfectly sinusoidal, although they are very close. In addition to top right). For this experiment, die estimation error is always below the fundamental frequency, this introduces higher-order harmonic 0.2 m/s. Errors are mainly due to the low SNR of the measured components to the modulation signal. Please refer to the suppleDoppler- shifted signal.
mental material for a detailed derivation of the image formation in
these conditions. Unfortunately, the higher-order components are
generally not orthogonal, thus they can cause a phase-dependent
offset. We calibrate this offset for different modulation frequencies Subframe Alignment Although the required heterodyne and hoand phase shifts tp using a static target. The depth-dependent offsets modyne shots could be captured simultaneously using multi-sensor are plotted for different modulation frequencies in Figure 5. configurations, they have to be captured in an alternating fashion using the single-sensor solution used in this paper. Since we are deal¬
This offset is calibrated in an offline process and raw phase measing with moving objects, the individual shots cannot, be assumed urements can be corrected digitally using a lookup table. Note that to be perfectly aligned, which results in velocity artifacts around for relatively low modulation frequencies, such as 30 MHz, we find edges in the scene. We can mitigate, although not completely rea fairly large depth range (around 1 m) to be almost independent of move, these artifacts by computing a SIFT flow on the raw data and this offset. In practice, it is therefore relatively easy to remove the warping them to a reference frame. While not perfect, the SIFT higher-order frequency components. flow delivered sufficiently good warps for most captures.
Figure imgf000058_0001
Figure 8: Complex scene with ambient illumination and a large depth range. The velocity is robustly estimated within the range of the illumination (approx. 5m inside), even in outdoor settings.
Figure imgf000058_0002
lights several challenges of D-ToF and ToF in general. D-ToF re¬
Figure 9: Velocity maps color-coded in grayscale. The maps comquires two frames to be captured, and they must be aligned if reputed from raw measurements ( top) are corrupted by Poisson noise. corded with a single camera, in some instances, such as Figures 10 To account for this, we apply a binning-based non-local means-type and 12, the alignment is challenging and any errors will propagdenoiser to the reconstructed velocity images (bottom). ate into the velocity maps, especially around depth-discontinuities.
These artifacts could be mitigated by optimizing the camera firmware to minimizing switching time between the subframes or by
8 Experimental Results using two co-axial ToF cameras. Objects with dark albedos, as for example observed in Figure 11, are challenging for any ToF method
We show results captured with our prototype imaging system in because only a small amount of the coded illumination is reflected Figures 1, 8, 10, 11, 12, 13, 14, and in the supplement. The resback to the camera. Similarly, shadows are very challenging and ults validate the proposed imaging system for a variety of challenoften result in either no depth velocity estimation or errors (sweater ging indoor and outdoor scenes. Color images are recorded with the in Fig. 8 and regions between fingers in Fig. 13). Whereas some same exposure time as the iime-of-flight camera. Most of the scenes of these limitations can be overcome with better hardware, others have a slight red tint, because we work with eye-safe red illuminare inherent to the time- of- flight approach. Please see Section 8 ation in the visible spectrum. Like current commercial ToF camfor a more detailed discussion and the supplemental video for more eras, future implementations of this system would most likely use results.
invisible, near infrared wavelengths to encode velocity and depth
information. The reconstructed velocity maps are color-coded; ab7 Towards the 3D Velocity Field solute units are indicated in the color bars. As expected, static
scenes result in a constant velocity map whereas velocity is directly
encoded in the measurements and subsequently reconstructed for Optical flow computed from conventional video sequences estimeach sensor pixel independently. In addition to the velocity maps, ates the 2D projection of the 3D flow field onto the image plane. Figures 1, 11, 13, 14 also show the corresponding depth maps that The radial component is usually lost Furthermore, optical flow can be estimated from an additional capture as well as the velocity is an ill-posed problem and may fail in many scenarios. Doppler maps (see Sec. 4.2). ToF addresses two problems of optical flow: first, it can help in cases where optical flow fails either due to large displacements or
The selection of scenes shows a wide range of motion types that tsiissing scene structures. Second, our technique also helps in cases
Figure imgf000059_0001
Figure imgf000059_0002
Figure 11: This result shows periodic motions in zfor a textured Figure 14: Physical props for gaming, such as ping pong balls fired object. Although the estimated velocity is mostly correct, shadows with this toy gun, could be tracked and enable new HCI techniques. and dark scene parts are challenging for robust velocity estimation.
Figure imgf000059_0003
D-ToF is complimentary to optical flow. It allows for the deptis bias of xz-flow to be removed and enables recording of the metric where the optica? flow estimation is successful; in this case, we can 3D velocity field of the scene. However, if only radial velocity is pler Time-of-Flight, we hope to contribute a fundamentally new imaging modality that will have an impact on all of these applications. The possibility of implementing the proposed techniques on
Figure imgf000060_0001
existing consumer devices makes Doppler Time-of-Flight a particularly attractive computational photography technique.
Figure 15: Failure case of optical flow for a moving, but untextured
scene (left). Optical flow [ Liu 2009] and SIFT flow [Liu et al. 2008]
or two succeeding color frames are shown in the second and third
column; the 2D flow vectors are color-coded with the shown color
wheel (insets). Both methods cannot recover the true 2D motion In this appendix, we derive expressions for the proposed imaging of the fan and wrongly segment the scene. Our orthogonal velocity systems using phasor notation. These may be more intuitive for estimate can resolve this problem and properly segment the scene. some readers, especially those familiar with the communication literature. In particular, we assume orthogonal illumination and modulation frequencies and derive the measured intensity of a dynamic
Figure imgf000060_0002
Figure 16: Towards 3D flow: when optical flow succeeds, the full
3D metric flow is uniquely estimated from both 2D pixel flow and A, B are the complex phasor amplitudes containing the amplitude the radial velocity maps. The top images show a frame where opand phase dependency. Sui is the Kronecker delta; for perfectly- tical flow computed reasonable estimates. The bottom shows full static scenes, this expression is zero. Note that phasors can usu3D velocity estimate for different views. Note that the optical flow ally not be multiplied. By using the complex conjugate, this is helps us to determine that fan's velocity is slightly rotated to the possible while implicitly assuming that high frequency can be igupper right, 'where the center of rotatio is located (bottom left). nored [Ceperley 201 5] .
required, our technique can also be used stand-alone, independent Assuming that objects are moving, a Doppler shift δ/Τ is introof optical flow. duced. Frequency shifts like this have been analyzed for OFDM as inter carrier interference. We get the following complex interference pattern I :
Limitations and Future Work Commercially available ToF
sensors today are low-resolution and their quantum efficiency
and noise characteristics are not comparable with modern CMOS
sensors. We expect future generations of ToF sensors to deliver - significantly higher image quality, which would directly benefit D- ToF as well. Higher modulation frequencies would directly improve the signal-to-noise ratio in our setup, because the Doppler
Figure imgf000060_0003
effect is proportional to these frequencies. For eye-safe operation,
we use diffused laser diodes that operate in the visible spectrum in
combination with a ToF sensor that has its visible spectrum cutoff
filter removed. The laser illumination is therefore visible in ail of The authors would like to thank Lei Xiao, Refaei Whyte, Adrian the RGB images as a red tint. Future implementations of our sysDorrington, Tom Malzbender, Charles Fraccia, Achuta Kadambi, tem would operate the time-of-flight camera in the near infrared and the anonymous reviewers for valuable feedback and help with spectrum, as is common practice in commercial ToF cameras. Fithe experiments, Felix Heide was supported by a Four-year Felnally, all presented techniques could easily be implemented on conlowship from the University of British Columbia. Matthias Hullin sumer time-of-flight cameras with the appropriate level of access to was supported by the X-Rite Chair for Digital Material Appearance, the system firmware or driver software. We hope that vendors will Gordon Wetzstein was supported by a Terman Faculty Fellowship. give researchers and developers the opportunity to modify modulaWolfgang Heidrich was supported by Baseline Funding of the King tion and illumination frequencies of their devices in the near future. Abdullah University of Science and Technology and a NSERC DisWith access to these frequencies, we could port die proposed techcovery Grant.
niques to available consumer devices.
References
Coric!l!Sior! Time-of-flight cameras have entered the consumer
market only a few years ago, but transformed the way machines BARRON , J. , FLEET, D,, AND BEAUCHEMEN, S, 1 994. Performperceive the world. Human-computer interaction, medical imaging, ance of optical flow techniques. IJCV 12, 1, 43-77.
robotics and machine vision, navigation for self-driving cars and
quadcopters, and many other fundamental computer vision tasks B OREMAN , G. D. 2001. Modulation Transfer Function in Optical have seen dramatic improvements using these devices. With Dopand ElectroOptical Systems. SPIE Publications. BOTTGE , B., AND SEITZ, P. 2008. Robust optical time-of-flight KIRMANI, A., HUTCHISON, Τ., DAVIS, J., AND RASKAR, R. range imaging based on smart pixel structures. IEEE Trans. Cir2009. Looking around the corner using transient imaging. In cuits and Systems 55, 6, 1512-1525. Proc. 1CCV, 159-166.
CEPERLEY, P., 2015. Resonances, waves and fields, http: LANGE, R., AND SEITZ, P. 2001. Solid-state time-of-flight range / /resonan ceswavesan cifie1ds . b1 ogspot . coin/ camera. IEEE J. Quantum Electronics 37, 3, 390-397.
2011/04/28- alid-method-cf-multiplying-two .
Li, Y., AND STUBER, G. 2006. Orthogonal Frequency Division html. [Online; accessed 20-January -2015].
Multiplexing for Wireless Communications. Springer.
CONROY, R., DORRINGTON, A., KUNNEMEYER, R., AND CREE,
LINDNER, M,, AND KOLB, A. 2006. Lateral and depth calibration M . 2009. Range imager Performance Comparison in Homodyne
of PMD-distance sensors. In Advances in Visual Computing. and Heterodyne Operating Modes. In Proc. SPIE 7239.
Spnnger, 524-533.
DOPPLER, C. J. 1842. fiber das farbige Licht der Doppelsterne LINDNER, M., AND KOLB, A. 2009. Compensation of Motion Arund einiger anderer Gestirne des Himmels. Abhandlungen der tifacts for Time-of-Flight Cameras. In Proc. Dynamic 3D imaKonigl. Bohm. Ge.sellsch.afi der Wissenschaften 12, 2, 465-482. ging. 16-27.
DORRINGTON, A. A., CREE, M. J., PAYNE, A. D., CONROY, Liu, C, YUEN, TORRALBA, A., SIVIC, J., AND FREEMAN,
R. M., AND CARNEGIE, D. A. 2007. Achieving sub- millimetre W. T. 2008. SIFT flow: Dense correspondence across different precision with a solid-state full-field heterodyning range imaging scenes. In Computer Vision-ECCV 2008. Springer, 28-42. camera. In Proc. Meas. Sci. Technol., vol. 18.
Liu, C. 2009. Beyond pixels: exploring new representations and
ERZ, M., AND JAHNE, B . 2009. Radiometric and spectrometric applications for motion analysis. PhD thesis, MIT.
calibrations, and distance noise measurement of ToF cameras. In
Dynamic 3D Imaging. Springer, 28-41. AIK, N., ZHAO, S ., VELTEN, A., RASKAR, R., AND BALA, K.
2011. Single view reflectance capture using multiplexed scatter¬
GOKTURK, S., YALCIN, H., AND BAMJI, C. 2004. A time-of- ing and time-of-flight imaging. ACM Trans. Graph. (SIGGRAPH flight depth sensor - system description, issues and solutions. In Asia) 30, 6, 171 :1-171: 10.'
Proc. CVPR, 35-35. '
O 'TOOLE, M., HEIDE, P., XIAO, L., HULLIN, M. Β., HEID¬
GU, J. , HITOMI, Y., M ITS UN AG A, T., AND NAYAR, S . 2010. RICH, W., AND K UTULAKOS, K. N. 2014. Temporal frequency Coded Rolling Shutter Photography: Flexible Space-Time probing for 5c! transient analysis of global light transport. ACM Sampling. In Proc. ICCP. Trans. Graph. (SIGGRAPH) 33, 4, 87: 1 -87:11.
GUPTA, Μ., NAYAR, S. ., HULLIN, M., AND MARTIN, J. 2014. PANDHARKAR, R ,, VELTEN, A., BARDAGJY, A. , LAWSON, E., Phasor Imaging: A Generalization Of Correlation-Based Time- BAWENDI, M. , AND RASKAR, R. 2011. Estimating motion of-Flight Imaging. Tech. rep., Jiui. and size of moving non-line-of-sight objects in cluttered environments. In Proc. CVPR, 265-272.
HEIDE, F., HULLIN, M . B ., GREGSON , J . , AND HEIDRICH,
W. 2013. Low-budget transient imaging using photonic mixer TOCCI, M., KISER, C, TOCCI, N., AND SEN, P. 2011. A verdevices. ACM Trans. Graph. (SIGGRAPH) 32. , 45: 1-45: 10. satile HDR video production system. ACM Trans. Graph. (SIGGRAPH) 30, 4, 4 i.
HEIDE, P., XIAO, L., HEIDRICH, VV„ AND HULLIN, M . B . 2014.
Diffuse mirrors: 3D reconstruction from diffuse indirect illuminVELTEN, A., WILLWACHER, T., GUPTA, O., VEERARAGHAVAN, ation using inexpensive time-of-flight sensors. In Proc. CVPR. A., BAWENDI, M., AND RASKAR , R . 2012. Recovering tfwee- dimensiona) shape around a corner using ultrafast time-of-flight
HEIDE, F., XIAO, L., KOLB, A., HULLIN, M. B ., AND HEIDimaging. Nat Commun 745, 3,
RICH, W. 2014. Imaging in scattering media using correlation
image sensors and sparse convolutional coding. OSA Opt. Exp. VELTEN, A. , Wu, D. , JARABO, A., MASIA, B . , BARSI, C , 22, 21 , 26338-26350. JOSHI, C, LAWSON, E., BAWENDI, Μ., GUTIERREZ, D., AND
RASKAR, R. 2013. Femto-photography: Capturing and visual¬
HOEGG, T., LEFLOCH, D., AND KOLB, A. 201 3. Real-time Moizing the propagation of light. ACM Trans. Graph. (SIGGRAPH) tion Compensation for PMD-ToF Images. In Lecture Notes in 52, 4, 44:1-44-8.
Computer Science, vol. 8200.
WEI, D., MAS UREL, P., URIHARA, T., AND ANDO, S . 2006.
HONEGGER , D., MEIER, I.,, TANS ANEN, P., AND POLLEFEYS, Optical flow determination with complex-sinusoidally moduM. 2013. An open source and open hardware embedded metric lated imaging. In Proc. ICSP, vol. 2.
optical flow CMOS camera for indoor and outdoor applications.
In Proc. ICRA, IEEE, Wu, D. , WETZSTEIN, G., BARSI, C, WILLWACHER, T.,
1736-1741 .
O' TOOLE, Μ. , NAIK, N., DAI, Q., KUTULAKOS, K. , AND
HONTAN i, H., OlSHi, G., AND KriAGAWA, Τ, 2014. Local estimRASKAR, R , 2012. Frequency analysis of transient light transation of high velocity optical flow with correlation image sensor. port with applications in bare sensor imaging. In Proc. ECCV, In Proc. ECCV, 235-249. 542-555.
HORN, B . , AND SCHUNCK, B. 1981. Determining optical flow. YAS UMA, F., MITS UNAGA, T., Iso, D., AND NAYAR, S. K. 2010. Artificial Intelligence 17, 185-203. Generalized assorted pixel camera: postcapture control of resolution, dynamic range, and spectrum. IEEE TIP 19, 9, 2241-
KADAMB I, A., WHYTE, R. , BHANDARI, A ., STREETER, L,, 2253.
BARSI, C, DORRINGTON, A., AND RASKAR, R. 2013. Coded
time of flight cameras: sparse deconvolution to address multipatli
interference and recover time profiles. ACM Trans. Graph. (SIGGRAPH Asia) 32, 6.

Claims

CLAIMS Therefore, the following is claimed:
1. A method for imaging object velocity, comprising the steps of:
(a) providing a Time-of-Flight camera and using the Time-of-Flight camera to capture a signal representative of an object in motion over an exposure time;
(b) coding illumination and modulation frequency of the captured signal within the exposure time;
(c) mapping a change of illumination frequency to measured pixel
intensities of the captured signal within the exposure time; and
(d) extracting information about a Doppler shift in the illumination
frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion.
2. The method of claim 1 , wherein radial velocity information of the object in motion is simultaneously captured for each pixel captured within the exposure time.
3. The method of claim 1 or 2, wherein the illumination frequency is coded orthogonal to the modulation frequency of the captured signal.
4. The method of any of claims 1 - 3, wherein the change of illumination frequency corresponds to radial object velocity.
5. The method of any of claims 1 - 4, wherein the Time-of-Flight camera has a receiver and a transmitter, and the frequency of the receiver is configured to be orthogonal to the frequency of the transmitter.
6. The method of any of claims 1 - 5, wherein the exposure time is longer than the wavelength of a modulated captured signal.
7. The method of any of claims 1 - 6, wherein a ratio of a heterodyne measurement and a homodyne measurement is determined to extract the information about the Doppler shift.
8. The method of any of claims 1 - 7, further including the step of:
simultaneously capturing color, depth and velocity information concerning the object in motion during the exposure time.
9. The method of claim 8, wherein the change of illumination frequency corresponds to radial object velocity and optical flow of the object in motion is computed on red, green and blue (RGB) frames within a measured change in illumination frequency.
10. The method of claim 9, including estimating a 3D velocity field for the object in motion.
1 1. The method of any of claims 1 -10, wherein depth and velocity imaging are combined either using the Time-of-Flight camera by alternating modulation frequencies between successive video frames over the exposure time or using at least two Time-of-Flight cameras.
12. A system for imaging object velocity, comprising:
at least one device for capturing a signal representative of an object in motion over an exposure time;
at least one computing device comprising a processor and a memory; and
an application executable on the at least one computing device, the application comprising machine readable instructions stored in the memory that, when executed by the processor, cause the computing device to at least:
(a) code illumination and modulation frequency of the captured signal within the exposure time;
(b) map a change of illumination frequency to measured pixel intensities of the captured signal within the exposure time; and
(c) extract information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion.
13. The system of claim 12, wherein the device is at least one Time-of-Flight camera.
14. The system of claim 12 or 13, wherein radial velocity information of the object in motion is simultaneously captured for each pixel captured within the exposure time.
15. The system of any of claims 12-14, wherein the illumination frequency is coded orthogonal to the modulation frequency of the captured signal.
16. The system of any of claims 12-15, wherein the change of illumination frequency corresponds to radial object velocity.
17. The system of any of claims 13-16, wherein the Time-of-Flight camera includes a receiver and a transmitter, and the frequency of the receiver is configured to be orthogonal to the frequency of the transmitter.
18. The system of any of claims 12-17, wherein the logic captures color, depth and velocity information concerning the object in motion during the exposure time.
19. A non-transitory computer readable medium employing an executable application in at least one computing device, the executable application comprising machine readable instructions stored in the medium that:
(a) receives one or more signals representative of an object in motion over an exposure time;
(b) codes illumination and modulation frequency of the one or more
signals within the exposure time;
(c) maps a change of illumination frequency to measured pixel intensities of the one or more signals within the exposure time; and
(d) extracts information about a Doppler shift in the illumination frequency to obtain a measurement of instantaneous per pixel velocity of the object in motion.
20. The non-transitory computer readable medium of claim 19, wherein the one or more signals are captured using at least one Time-of-Flight camera.
PCT/IB2016/054761 2015-08-07 2016-08-05 Doppler time-of-flight imaging WO2017025885A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/739,854 US11002856B2 (en) 2015-08-07 2016-08-05 Doppler time-of-flight imaging

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201562282708P 2015-08-07 2015-08-07
US62/282,708 2015-08-07

Publications (1)

Publication Number Publication Date
WO2017025885A1 true WO2017025885A1 (en) 2017-02-16

Family

ID=56801663

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2016/054761 WO2017025885A1 (en) 2015-08-07 2016-08-05 Doppler time-of-flight imaging

Country Status (2)

Country Link
US (1) US11002856B2 (en)
WO (1) WO2017025885A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108254757A (en) * 2018-04-11 2018-07-06 大连理工大学 A kind of speed measuring device based on double TOF sensors
EP3620821A1 (en) * 2018-09-05 2020-03-11 Infineon Technologies AG Time of flight camera and method for calibrating a time of flight camera
EP3644094A1 (en) 2018-10-26 2020-04-29 Sick Ag 3d time of flight camera and method for recording three-dimensional image data
CN111837159A (en) * 2018-03-09 2020-10-27 三星电子株式会社 Method and apparatus for performing depth estimation of an object
US11163064B1 (en) 2020-07-13 2021-11-02 Palo Alto Research Center Incorporated Doppler imager

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018102190A1 (en) 2016-11-29 2018-06-07 Blackmore Sensors and Analytics Inc. Method and system for classification of an object in a point cloud data set
WO2018102188A1 (en) 2016-11-30 2018-06-07 Blackmore Sensors and Analytics Inc. Method and system for automatic real-time adaptive scanning with optical ranging systems
CN110140063B (en) 2016-11-30 2023-10-13 布莱克莫尔传感器和分析有限责任公司 Method and system for adaptive scanning using optical ranging system
KR102254468B1 (en) * 2016-11-30 2021-05-20 블랙모어 센서스 앤드 애널리틱스, 엘엘씨 Method and apparatus for Doppler detection and Doppler correction of optical chirp distance detection
US10422880B2 (en) 2017-02-03 2019-09-24 Blackmore Sensors and Analytics Inc. Method and system for doppler detection and doppler correction of optical phase-encoded range detection
US10401495B2 (en) 2017-07-10 2019-09-03 Blackmore Sensors and Analytics Inc. Method and system for time separated quadrature detection of doppler effects in optical range measurements
JP2021515241A (en) 2018-04-23 2021-06-17 ブラックモア センサーズ アンド アナリティクス エルエルシー Control methods and systems for autonomous vehicles using coherent distance Doppler optical sensors
US10984539B2 (en) * 2018-07-03 2021-04-20 Eys3D Microelectronics, Co. Image device for generating velocity maps
CN113167899B (en) * 2018-12-04 2022-08-16 Iee国际电子工程股份公司 Method for depth measurement with time-of-flight camera using amplitude modulated continuous light for correction
US11822010B2 (en) 2019-01-04 2023-11-21 Blackmore Sensors & Analytics, Llc LIDAR system
US11828848B2 (en) * 2019-03-14 2023-11-28 Aeva, Inc. Velocity estimation using doppler per point LiDAR systems
WO2021087800A1 (en) * 2019-11-06 2021-05-14 Oppo广东移动通信有限公司 Information processing method, encoding apparatus, decoding apparatus, system, and storage medium
JP2021081290A (en) * 2019-11-19 2021-05-27 ソニーグループ株式会社 Ranging device and ranging method
US11702102B2 (en) * 2020-11-19 2023-07-18 Waymo Llc Filtering return points in a point cloud based on radial velocity measurement
CN114095713A (en) * 2021-11-23 2022-02-25 京东方科技集团股份有限公司 Imaging module, processing method, system, device and medium thereof
US20230320551A1 (en) * 2022-04-11 2023-10-12 Vorwerk & Co. Interholding Gmb Obstacle avoidance using fused depth and intensity from nnt training

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013012335A1 (en) * 2011-07-21 2013-01-24 Ziv Attar Imaging device for motion detection of objects in a scene, and method for motion detection of objects in a scene
US8446575B1 (en) * 2005-11-08 2013-05-21 Sandia Corporation Imaging doppler velocimeter with downward heterodyning in the optical domain
EP2890125A1 (en) * 2013-12-24 2015-07-01 Softkinetic Sensors N.V. A time-of-flight camera system

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU6633798A (en) * 1998-03-09 1999-09-27 Gou Lite Ltd. Optical translation measurement
US6860350B2 (en) * 2002-12-20 2005-03-01 Motorola, Inc. CMOS camera with integral laser ranging and velocity measurement
US7440113B2 (en) * 2005-12-23 2008-10-21 Agilent Technologies, Inc. Littrow interferometer
US7746450B2 (en) * 2007-08-28 2010-06-29 Science Applications International Corporation Full-field light detection and ranging imaging system
US8279420B2 (en) * 2009-08-06 2012-10-02 ISC8 Inc. Phase sensing and scanning time of flight LADAR using atmospheric absorption bands
GB201208088D0 (en) * 2012-05-09 2012-06-20 Ncam Sollutions Ltd Ncam
ES2581554T3 (en) * 2012-09-10 2016-09-06 Philips Lighting Holding B.V. Light detection system and method
AT513589B1 (en) * 2012-11-08 2015-11-15 Bluetechnix Gmbh Recording method for at least two ToF cameras
US9374512B2 (en) * 2013-02-24 2016-06-21 Pelican Imaging Corporation Thin form factor computational array cameras and modular array cameras

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8446575B1 (en) * 2005-11-08 2013-05-21 Sandia Corporation Imaging doppler velocimeter with downward heterodyning in the optical domain
WO2013012335A1 (en) * 2011-07-21 2013-01-24 Ziv Attar Imaging device for motion detection of objects in a scene, and method for motion detection of objects in a scene
EP2890125A1 (en) * 2013-12-24 2015-07-01 Softkinetic Sensors N.V. A time-of-flight camera system

Non-Patent Citations (65)

* Cited by examiner, † Cited by third party
Title
BA O , J.; FLEET, D.; BEAUCHEMIN, S: "Performance of optical flow techniques", LJCV 12, vol. 1, 1994, pages 43 - 77, XP001109171, DOI: doi:10.1007/BF01420984
BARRON, J.; FLEET, D.; BEAUCHEMIN, S.: "Performance of optical flow techniques", IJCV, vol. 12, no. 1, 1994, pages 43 - 77, XP001109171, DOI: doi:10.1007/BF01420984
BO A , G. D.: "Modulation Transfer Function in Optical and ElectroOptical Systems", SPIE PUBLICATIONS, 2001
BOREMAN, G. D: "Modulation Transfer Function in Optical and ElectroOptical Systems", SPIE PUBLICATIONS, 2001
BUTTGEN, B.; SEITZ, P.: "Robust optical time-of-flight range imaging based on smart pixel structures", IEEE TRANS. CIRCUITS AND SYSTEMS, vol. 55, no. 6, 2008, pages 1512 - 1525, XP011224743, DOI: doi:10.1109/TCSI.2008.916679
BÜTTGEN, B.; SEITZ, P.: "Robust optical time-of-flight range imaging based on smart pixel structures", IEEE TRANS. CIRCUITS AND SYSTEMS, vol. 55, no. 6, 2008, pages 1512 - 1525, XP011224743, DOI: doi:10.1109/TCSI.2008.916679
CEPERLEY, P., RESONANCES, 20 January 2015 (2015-01-20), Retrieved from the Internet <URL:http: //resonanceswavesandfields.blogspot.com/ 2011/04/28-valid-metho-d-of-multiplying-two. html.>
CEPERLEY, P., RESONANCES, WAVES AND FIELDS, 20 January 2015 (2015-01-20), Retrieved from the Internet <URL:http://resonanceswavesandfields.blogspot.com/ 2011/04/28-valid-method-of-multiplying-two. html.>
CONROY, R.; DORRINGTON, A.; KUNNEMEYER, R.; CREE, M.: "Range Imager Performance Comparison in Homodyne and Heterodyne Operating Modes", IN PROC. SPIE 7239, 2009
DOPPLER, C. J.: "Uber das farbige Licht der Doppelsteme und einiger anderer Gestime des Himmels", ABHANDLUNGEN DER KÖNIGL. BÖHM. GESELLSCHAFT DER WISSENSCHAFTEN, vol. 12, no. 2, pages 465,482
DOPPLER, C. J.: "Uber das farbige Licht der Doppelsterne und einiger anderer Gestirne des Himmels. Abhandlungen der Konigl", BOHM. GESELLSCHAFT DER WISSENSCHAFTEN, vol. 12, no. 2, pages 465 - 482
DORRINGTON, A. A.; CREE, M. J.; PAYNE, A. D.; CONROY, R. M.; CARNEGIE, D. A.: "Achieving sub-millimetre precision with a solid-state full-field heterodyning range imaging camera", IN PROC. MEAS. SCI. TECHNOL., vol. 18, 2007, XP020118817, DOI: doi:10.1088/0957-0233/18/9/010
ERZ, M.; JÄHNE, B.: "In Dynamic 3D Imaging", 2009, SPRINGER, article "Radiometric and spectrometric calibrations, and distance noise measurement of ToF cameras", pages: 28 - 41
ERZ, M.; JAHNE, B.: "Radiometric and spectrometric calibrations, and distance noise measurement of ToF cameras", IN DYNAMIC 3D IMAGING. SPRINGER, 2009, pages 28 - 41, XP047390409, DOI: doi:10.1007/978-3-642-03778-8_3
FELIX HEIDE ET AL: "Doppler time-of-flight imaging", ACM TRANSACTIONS ON GRAPHICS (TOG), vol. 34, no. 4, 27 July 2015 (2015-07-27), US, pages 36:1 - 36:11, XP055318130, ISSN: 0730-0301, DOI: 10.1145/2766953 *
GOKTURK, S.; YALCIN, H.; BAMJI, C.: "A time-offlight depth sensor - system description, issues and solutions", IN PROC. CVPR, 2004, pages 35 - 35, XP010761980
GOKTURK, S.; YALCIN, H.; BAMJT, C.: "A time-of-flight depth sensor - system description, issues and solutions", IN PROC. CVPR, 2004, pages 35 - 35, XP010761980
GU, J.; HITOMI, Y.; MITSUNAGA, T.; NAYAR, S.: "Coded Rolling Shutter Photography: Flexible Space-Time Sampling", IN PROC. ICCP, 2010
GU, J.; HITOMI, Y.; MITSUNAGA, T.; NAYAR, S.: "Coded Rolling Shutter Photography: Flexible Space-Time Sampling", IN PROC. ICCP., 2010
GUPTA, M.; NAYAR, S. K.; HULLIN, M.; MARTIN, J.: "Phasor Imaging: A Generalization Of Correlation-Based Time-of-Flight Imaging", TECH. REP., June 2014 (2014-06-01)
GUPTA, M.; NAYAR. S. K.; HULLIN, M.; MARTIN, J.: "Phasor Imaging: A Generalization Of Correlation-Based Time-of-Flight Imaging", TECH. REP., 2014
HEIDE, F.; HULLIN, M. B.; GREGSON, J.; HEIDRICH, W.: "Low-budget transient imaging using photonic mixer devices", ACM TRANS. GRAPH. (SIGGRAPH, vol. 32, no. 4, 2013, pages 45.1 - 45.10
HEIDE, F.; HULLIN, M. B.; GREGSON, J.; HEIDRICH. W.: "Low-budget transient imaging using photonic mixer devices", ACM TRANS. GRAPH. (SIGGRAPH), vol. 32, no. 4, 2013, pages 45.1 - 45.10
HEIDE, F.; XIAO, L.; HEIDRICH, W.; HULLIN, M. B.: "Diffuse mirrors: 3D reconstruction from diffuse indirect illumination using inexpensive time-of-flight sensors", IN PROC. CVPR, 2014
HEIDE, F.; XIAO, L.; HEIDRICH, W.; HULLIN, M. B: "Diffuse mirrors: 3D reconstruction from diffuse indirect illumination using inexpensive time-of-flight sensors", IN PROC. CVPR., 2014
HEIDE, F.; XIAO, L.; KOLB, A.; HULLIN, M. B.; HEID-RICH, W.: "Imaging in scattering media using correlation image sensors and sparse convolutional coding", OSA OPT. EXP., vol. 22, no. 21, 2014, pages 26338 - 26350
HEIDE, F.; XIAO, L.; KOLB, A.; HULLIN, M. B.; HEIDRICH, W.: "Imaging in scattering media using correlation image sensors and sparse convolutional coding", OSA OPT. EXP., vol. 22, no. 21, 2014, pages 26338 - 26350
HEIDE. F.; HEIDRICH W.; HULLIN. M; WETZSTEIN, G., DOPPLER TIME-OF-RIGHT IMAGING ACM TRANS. GRAPH., vol. 34, 2015, pages 4
HOEGG, T.; LEFLOCH, D.; KOLB, A.: "Real-time Mo-tion Compensation for PMD-ToF Images", IN LECTURE NOTES IN COMPUTER SCIENCE, vol. 8200, 2013
HONEGGER, D.; MEIER, L.; TANSKANEN, P.; POLLEFEYS, M.: "An open source and open hardware embedded metric optical flow CMOS camera for indoor and outdoor applications", IN PROC. ICRA, 2013, pages 1736 - 1741, XP032505966, DOI: doi:10.1109/ICRA.2013.6630805
HONEGGER, D.; METER, L.; TANSKANEN, P.; POLLEFEYS, M.: "An open source and open hardware embedded metric optical flow CMOS camera for indoor and outdoor applications", IN PROC. ICRA. IEEE, 2013, pages 1736 - 1741, XP032505966, DOI: doi:10.1109/ICRA.2013.6630805
HONTANI, H.; OISHI, G.; KITAGAWA, T.: "Local estimation of high velocity optical flow with correlation image sensor", IN PROC. ECCV, 2014, pages 235 - 249
HONTANI, H.; OISHI, G.; KITAGAWA, T: "Local estimation of high velocity optical flow with correlation image sensor", IN PROC. ECCV., 2014, pages 235 - 249
HORN, B.; SCHUNCK, B., DETERMINING OPTICAL FLOW ARTIFICIAL INTELLIGENCE, vol. 17, 1981, pages 185 - 203
HORN, B.; SCHUNCK, B.: "Determining optical flow", ARTIFICIAL INTELLIGENCE, vol. 17, 1981, pages 185 - 203, XP000195787, DOI: doi:10.1016/0004-3702(81)90024-2
KADAMBI, A.; WHYTE, R.; BHANDARI, A.; STREETER, L.; BARSI, C.; DORRINGTON, A.; R SK R, R.: "Coded time of flight cameras: sparse deconvolution to address multipath interference and recover time profiles", ACM TRANS. GRAPH. (SIGGRAPH ASIA), vol. 32, 2013, pages 6
KADAMBI, A.; WHYTE, R.; BHANDARI, A.; STREETER, L.; BARSI, C.; DORRINGTON, A.; RASKAR, R.: "Coded time of flight cameras: sparse deconvolution to address multipath interference and recover time profiles", ACM TRANS. GRAPH. (SIGGRAPH ASIA, vol. 32, 2013, pages 6
KIRMANI, A.; HUTCHISON, T.; DAVIS, J.; RASKAR, R.: "Looking around the corner using transient imaging", IN PROC. ICCV, 2009, pages 159 - 166, XP031672511
KIRMANI, A.; HUTCHISON. T.; DAVIS. J.; RASKAR. R.: "Looking around the corner using transient imaging", IN PROC. ICCV, 2009, pages 159 - 166, XP031672511
LANGE, R.; SEITZ, P.: "Solid-state time-of-flight range camera", IEE J. QUANTUM ELECTRONICS, vol. 37, no. 3, 2001, pages 390 - 397, XP001086832, DOI: doi:10.1109/3.910448
LANGE, R.; SEITZ, P.: "Solid-state time-of-flight range camera", IEEE J. QUANTUM ELECTRONICS, vol. 37, no. 3, 2001, pages 390 - 397, XP001086832, DOI: doi:10.1109/3.910448
LI, Y.; STUBER, G.: "Orthogonal Frequency Division Multiplexing for Wireless Communications", 2006, SPRINGER
LINDNER, M..; KOLB, A.: "Compensation of Motion Artifacts for Time-of-Flight Cameras", IN PROC. DYNAMIC 3D IMAGING., 2009, pages 16 - 27, XP047390408, DOI: doi:10.1007/978-3-642-03778-8_2
LINDNER, M.; KOLB, A.: "In Advances in Visual Computing", 2006, SPRINGER, article "Lateral and depth calibration of PMD-distance sensors", pages: 524 - 533
LINDNER, M.; KOLB, A.: "In Advances in Visual Computing", 2006, SPRINGER, article "Lateral and depth calibration of PMD-distance sensors", pages: 524,533
LIU, C.: "Beyond pixels: exploring new representations and applications for motion analysis", PHD THESIS, MIT, 2009
LIU, C.; YUEN, J.; TORRALBA, A.; SIVIC, J.; FREEMAN, W. T.: "In Computer Vision-ECCV", 2008, SPRINGER, article "SIFT flow: Dense correspondence across different scenes", pages: 28 - 42
LIU, C: "PhD thesis", 2009, MIT, article "Beyond pixels: exploring new representations and applications for motion analysis"
NAIK, N.; ZHAO, S.; VELTEN, A.; RASKAR, R.; BALA, K.: "Single view reflectance capture using multiplexed scattering and time-of-flight imaging", ACM TRANS. GRAPH. (SIGGRAPH ASIA, vol. 30, no. 6, 2011, pages 171.1 - 171.10, XP058035096, DOI: doi:10.1145/2070781.2024205
NAIK. N.; ZHAO, S.; VELTEN, A.; RASKAR. R.; BALA, K.: "Single view reflectance capture using multiplexed scattering and time-of-flight imaging", ACM TRANS. GRAPH. (SIGGRAPH ASIA, vol. 30, no. 6, 2011, pages 171.1 - 171.10, XP058035096, DOI: doi:10.1145/2070781.2024205
O'TOOLE, M.; HEIDE, F.; XIAO, L.; HULLIN, M. B.; HEIDRICH, W.; KUTULAKOS, K. N.: "Temporal frequency probing for 5d transient analysis of global light transport", ACM TRANS. GRAPH. (SIGGRAPH, vol. 33, no. 4, 2014, pages 87.1 - 87.11, XP058051918, DOI: doi:10.1145/2601097.2601103
O'TOOLE, M.; HEIDE, F.; XIAO, L.; HULLIN, M. B.; HEIDRICH, W.; KUTULAKOS. K. N.: "Temporal frequency probing for 5d transient analysis of global light transport", ACM TRANS. GRAPH. (SIGGRAPH, vol. 33, no. 4, 2014, pages 87.1 - 87.11, XP058051918, DOI: doi:10.1145/2601097.2601103
PANDHARKAR, R.; VELTEN, A.; BARDAGJY, A.; LAWSON, E.; BAWENDI, M.; RASKAR, R.: "Estimating motion and size of moving non-line-of-sight objects in cluttered environments", IN PROC. CVPR, 2011, pages 265 - 272, XP032037967, DOI: doi:10.1109/CVPR.2011.5995465
SIGGRAPH '15 TECHNICAL PAPER, 9 August 2015 (2015-08-09)
TOCCI, M.; KISER, C.; TOCCI, N.; SEN, P., A VERSATILE HDR VIDEO PRODUCTION SYSTEM. ACM TRANS. GRAPH. (SIGGRAPH, vol. 30, no. 4, 2011, pages 41
TOCCI, M.; KISER, C.; TOCCI, N.; SEN, P.: "A versatile HDR video production system", ACM TRANS. GRAPH. (SIGGRAPH, vol. 30, no. 4, 2011, pages 41
VELTEN, A.; WILLWACHER, T.; GUPTA, 0.; VEERARAGHAVAN, A.; BAWENDI, M.; RASKAR, R.: "Recovering three dimensional shape around a corner using ultrafast time-of-flight imaging", NAT COMMUN, vol. 745, 2012, pages 3
VELTEN, A.; WILLWACHER, T.; GUPTA, O.; VEERARAGHAVAN. A.; BAWENDI, M.; RASKAR, R.: "Recovering three-dimensional shape around a corner using ultrafast time-of-flight imaging", NAT COMMUN, vol. 745, 2012, pages 3
VELTEN, A.; WU, D.; JARABO, A.; MASIA, B.; BARSI, C.; JOSHI, C.; LAWSON, E.; BAWENDI, M.; GUTIERREZ, D.; RASKAR, R.: "Femto-photography: Capturing and visualizing the propagation of light", ACM TRANS. GRAPH. (SIGGRAPH, vol. 32, no. 4, 2013, pages 44.1 - 44.8
VELTEN, A.; WU, D.; JARABO, A.; MASIA, B.; BARSI, C.; JOSHI, C.; LAWSON, E; BAWENDI. M.; GUTIERREZ. D.; RASKAR, R.: "Femto-photography: Capturing and visualizing the propagation of light", ACM TRANS. GRAPH. (SIGGRAPH, vol. 3-1, no. 4, 2013, pages 44.1 - 44.8
WEI, D.; MASUREL, P.; KURIHARA, T.; ANDO, S.: "Optical flow determination with complex-sinusoidally modulated imaging", IN PROC. ICSP, vol. 2, 2006
WU, D.; WETZSTEIN, G.; BARSI, C.; WILLWACHER, T.; O'TOOLE, M.; NAIK, N.; DAI, Q.; KUTULAKOS, K.; RAS A , R.: "Frequency analysis of transient light transport with applications in bare sensor imaging", IN PROC. ECCV, 2012, pages 542 - 555, XP047018580
WU, D.; WETZSTEIN, G.; BARSI, C.; WILLWACHER, T.; O'TOOLE, M.; NAIK, N.; DAI, Q.; KUTULAKOS, K.; RASKAR, R.: "Frequency analysis of transient light transport with applications in bare sensor imaging", IN PROC. ECCV, 2012, pages 542 - 555, XP047018580
YASUMA, F.; MITSUNAGA, T.; ISO, D.; NAYAR, S. K.: "Generalized assorted pixel camera: postcapture control of resolution, dynamic range, and spectrum", IEEE TIP, vol. 19, no. 9, 2010, pages 2241 - 2253, XP011306450
YASUMA. F.; MITSUNAGA. T.; ISO, D.; NAYAR. S. K.: "Generalized assorted pixel camera: postcapture control of resolution, dynamic range, and spectrum", IEEE TIP, vol. 19, no. 9, 2010, pages 2241 - 2253, XP011306450

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111837159A (en) * 2018-03-09 2020-10-27 三星电子株式会社 Method and apparatus for performing depth estimation of an object
CN108254757A (en) * 2018-04-11 2018-07-06 大连理工大学 A kind of speed measuring device based on double TOF sensors
EP3620821A1 (en) * 2018-09-05 2020-03-11 Infineon Technologies AG Time of flight camera and method for calibrating a time of flight camera
EP3644094A1 (en) 2018-10-26 2020-04-29 Sick Ag 3d time of flight camera and method for recording three-dimensional image data
DE102018126841A1 (en) 2018-10-26 2020-04-30 Sick Ag 3D time-of-flight camera and method for capturing three-dimensional image data
US11163064B1 (en) 2020-07-13 2021-11-02 Palo Alto Research Center Incorporated Doppler imager

Also Published As

Publication number Publication date
US11002856B2 (en) 2021-05-11
US20190154834A1 (en) 2019-05-23

Similar Documents

Publication Publication Date Title
US11002856B2 (en) Doppler time-of-flight imaging
Jarabo et al. Recent advances in transient imaging: A computer graphics and vision perspective
Kadambi et al. 3d depth cameras in vision: Benefits and limitations of the hardware: With an emphasis on the first-and second-generation kinect models
US10302424B2 (en) Motion contrast depth scanning
Kim et al. Design and calibration of a multi-view TOF sensor fusion system
Shrestha et al. Computational imaging with multi-camera time-of-flight systems
US10557921B2 (en) Active brightness-based strategy for invalidating pixels in time-of-flight depth-sensing
Lefloch et al. Technical foundation and calibration methods for time-of-flight cameras
US8619122B2 (en) Depth camera compatibility
CN104104941B (en) Acquiring three-dimensional images device and the method for generating depth image
US8687044B2 (en) Depth camera compatibility
US11563911B2 (en) Method and system for time-of-flight imaging with high lateral resolution
EP3504682A1 (en) Simultaneous localization and mapping with an event camera
WO2013052781A1 (en) Method and apparatus to determine depth information for a scene of interest
Falquez et al. Inertial aided dense & semi-dense methods for robust direct visual odometry
WO2012107574A1 (en) Method of enhanced depth image acquisition
CN111047650B (en) Parameter calibration method for time-of-flight camera
Castaneda et al. Stereo time-of-flight
Mufti et al. Statistical analysis of signal measurement in time-of-flight cameras
Caspi et al. Vertical parallax from moving shadows
KR20160026189A (en) Method and apparatus for generating depth image
Ivanov et al. An iterative technique for calculating aliasing probability of linear feedback signature registers
Islam et al. Interference mitigation technique for Time-of-Flight (ToF) camera
IL269260A (en) Target-less calibration of turbid media
EP3552044B1 (en) Time-of-flight imaging apparatus and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16757361

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16757361

Country of ref document: EP

Kind code of ref document: A1