CA3051102A1 - Method for epipolar time of flight imaging - Google Patents

Method for epipolar time of flight imaging Download PDF

Info

Publication number
CA3051102A1
CA3051102A1 CA3051102A CA3051102A CA3051102A1 CA 3051102 A1 CA3051102 A1 CA 3051102A1 CA 3051102 A CA3051102 A CA 3051102A CA 3051102 A CA3051102 A CA 3051102A CA 3051102 A1 CA3051102 A1 CA 3051102A1
Authority
CA
Canada
Prior art keywords
sensor
epipolar
illuminated
depth
pixel
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CA3051102A
Other languages
French (fr)
Inventor
Srinivasa Narasimhan
Supreeth ACHAR
Kiriakos KUTULAKOS
Joseph BARTELS
William Whittaker
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Toronto
Carnegie Mellon University
Original Assignee
University of Toronto
Carnegie Mellon University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Toronto, Carnegie Mellon University filed Critical University of Toronto
Publication of CA3051102A1 publication Critical patent/CA3051102A1/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C11/00Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • G01S17/08Systems determining position data of a target for measuring distance only
    • G01S17/10Systems determining position data of a target for measuring distance only using transmission of interrupted, pulse-modulated waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/491Details of non-pulse systems
    • G01S7/4912Receivers
    • G01S7/4913Circuits for detection, sampling, integration or read-out
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/491Details of non-pulse systems
    • G01S7/4912Receivers
    • G01S7/4915Time delay measurement, e.g. operational details for pixel components; Phase measurement

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Electromagnetism (AREA)
  • Multimedia (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Measurement Of Optical Distance (AREA)
  • Optical Radar Systems And Details Thereof (AREA)

Abstract

Energy-efficient epipolar imaging is applied to the ToF domain to significantly expand the versatility of ToF sensors. The described system exhibits 15+ m range outdoors in bright sunlight; robustness to global transport effects such as specular and diffuse inter-reflections; interference-free 3D imaging in the presence of many ToF sensors, even when they are all operating at the same optical wavelength and modulation frequency; and blur- and distortion-free 3D video in the presence of severe camera shake. The described embodiments are broadly applicable in consumer and robotics domains.

Description

METHOD FOR EPIPOLAR TIME OF FLIGHT IMAGING
Related Applications [0001] This application claims the benefit of U.S. Provisional Patent Application Serial No. 62/499,193, filed January 20, 2017.
Government Rights
[0002] This invention was made with government support under N000141512358 awarded by the ONR, 11S1317749 awarded by the NSF, HR00111620021 awarded by DARPA, and grants NNX16AD98G and NNX14AM53H awarded by NASA. The government has certain rights in the invention.
Background of the Invention
[0003] Time-of-flight (ToF) depth sensors have become the technology of choice in diverse applications, from automotive and aviation to robotics, gaming and consumer electronics. These sensors come in two general flavors: LIDAR-based systems that rely on extremely brief pulses of light to sense depth, and continuous-wave (CW) systems that emit a modulated light signal over much longer duration.
The LIDAR-based systems can acquire centimeter-accurate depth maps up to a kilometer away in broad daylight, but have low measurement rates.
Additionally, the cost per pixel is orders of magnitude higher than CW systems, whose range, outdoor operation and robustness are extremely limited. Because low cost, large-scale production and high measurement rate often trump other considerations, continuous-wave time-of-flight (CW-ToF) sensors continue to dominate the consumer electronics and low-end robotics space despite their shortcomings. Further, consumer grade time-of-flight depth cameras like Kinect and PMD are cheap, compact and produce video-rate depth maps in short-range applications.
Summary of the Invention
[0004] The present invention significantly reduces the shortcomings of CW-ToF
through the use of energy-efficient epipolar imaging. In certain embodiments, a continuously-modulated sheet of laser light is projected along a sequence of carefully chosen epipolar planes that collectively span the field of view. For each projected sheet, only a strip of CW-ToF pixels corresponding to each epipolar plane is exposed.
As shown in FIG. 2, a prototype implementation of the invention couples a specially built projection system to a CW-ToF sensor that has a controllable region of interest.
In some embodiments, an off-the-shelf CW-ToF sensor may be used. The off-the-shelf sensor may output live 320 x 240 3D video at 7.5 frames per second, with the frame rate only limited by the sensor's API.
[0005] Epipolar imaging was first proposed for acquiring live direct-only or global-only video with a conventional (non-ToF) video sensor. The approach has been extended to the ToF domain, but its energy efficiency is very low and it involves capturing more than 500 images to calculate a single "direct-only" ToF image.
In the context of triangulation-based 3D imaging, significant improvements in energy efficiency and robustness can be achieved with a 2D scanning-laser projector and a rolling shutter camera. The present invention extends this idea to the ToF
domain. As such, it inherits all the advantages of non-ToF energy-efficient epipolar imaging while also addressing challenges that are specific to CW-ToF.
[0006] The primary difficulty is that the range of CW-ToF sensors is severely limited by power consumption and eye safety considerations. Although most CW-ToF
sensors electronically subtract the DC component of incident light, photon noise from strong ambient sources such as sunlight can easily overwhelm the CW-ToF signal at distances of more than a few meters outdoors at typical frame rates. By concentrating the energy of the light source into a single sheet, epipolar ToF boosts this range to 10m and acquires a useful, albeit noisier, depth signal at over 15m outdoors.
[0007] A secondary difficulty is that the depth accuracy of CW-ToF sensors is strongly affected by global illumination effects, such as inter-reflections and global illumination transport. These effects produce longer light paths and show up as a source of structured additive noise. These effects cannot be cancelled a posteriori without imposing strong assumptions on the scene's geometry and reflectance properties, yet are extremely common indoors (e,g. , corners between walls, shiny surfaces of tables and floors, mirrors, etc,). The present invention demonstrates significant robustness to all forms of global transport, and to specular inter-reflections in particular, a form of global illumination transport that has never been possible to handle in live CW-ToF.
[0008] As devices equipped with CW-ToF depth sensors become increasingly common indoors and outdoors, they must be able to operate without interfering with each other. While non-interference between devices of a given make and model can be achieved by varying modulation frequency across them, robustness against the broader ecosystem of CW-ToF sensors is desirable. The present invention demonstrates that epipolar ToF enables interference-free live 3D imaging, even for devices that have the exact same modulation frequency and light source wavelength.
[0009] Lastly, CW-ToF sensors must acquire two or more frames with a different phase of emitted light to compute a single depth map. This makes them highly sensitive to camera shake, unlike conventional cameras where shaking merely blurs the image, camera shake in CW-ToF causes the static-scene assumption to be violated, leading to depth maps that are both blurry and corrupted by motion artifacts.
Epipolar ToF makes it possible to address both problems: motion blur is minimized by relying on very short exposures for each epipolar plane, motion artifacts and depth errors are minimized by acquiring multiple phase measurements per epipolar plane, rather than per frame and rolling-shutter-like distortions due to the sequential nature of epipolar-plane ToF are reduced by scheduling the sequence of epipolar planes so that post-acquisition distortion correction becomes easier.
Brief Description of the Drawings
[0010] FIGS. 1(a) ¨ 1(d) show a comparison of various scenes scanned using a regular (ToF) system versus the epipolar ToF system of the present invention.
[0011] FIG. 2 is a schematic view of a system for performing epipolar ToF
imaging.
[0012] FIGS. 3(a) ¨ 3(f) shows several possible epipolar plane sampling schemes and row exposures in ToF imaging.
[0013] FIG. 4(a) is a depiction of the prototype of the present application.
[0014] FIG. 4(b) depicts the laser light source components.
[0015] FIG. 5 shows timing diagrams for camera exposure, readout and mirror position for a particular sequence of the rows.
[0016] FIG. 6(a) ¨ 6(d) shows the results of imaging a white planar target at a range of distances from the sensor in cloudy weather and bright sunshine.
[0017] FIG. 7(a) ¨ 7(b) are graphs showing (a) the standard deviation in depth measurements obtained using regular and epipolar ToF imaging; and (b) the working range of the same simulated camera at different levels of acceptable range accuracy.
Note that the simulated camera's parameters differ from the prototype.
[0018] FIG. 8 shows that epipolar ToF imaging provides accurate depth returns from the surface of the light bulbs even when they are turned on.
[0019] FIG. 9 compares depth maps with epipolar and regular ToF imaging in the presence of global light transport.
Detailed Description
[0020] The term microcontroller, as used herein, may mean a dedicated hardware device, circuitry, an ASIC, an FPGA, a microprocessor running software, or any other means known in the art. It is further understood that the microcontroller will include connections to both the sensor and the laser light projector for sending control signals, and for receiving data. The invention is not intended to be limited to one method of implementing the functions of the controller.
[0021] As used herein, the terms camera and sensor are used interchangeably.
Continuous Wave Time of Flight
[0022] CW-ToF cameras use a temporally modulated light source and a sensor where the exposure is also modulated during integration. If the illumination modulation function is ft = cos(cot) and the sensor modulation function is = cos(cot + 4)) where co is the modulation frequency in rad/s and 4) is the phase offset between the source and sensor modulation functions, then the measurement at a pixel x is:
16),0-(X) = f 0 (t) * [Mt) + Ax],g(t)dt (1) T r CO
= -2 j 0 COS(COT h, (r)cly (2)
[0023] where h(t) represents a pixel 's transient response to the active light source and A, is the light received due to ambient light and the DC component of the active light source. Although, A, drops out of the integral, in practice,'&ig(x) is measured by integrating the incoming light to two different storage sites (called taps) depending on whether (t) is positive or negative and then taking the difference between the stored values so the ambient light still adds to the measurement shot noise.
[0024] If there are no indirect light paths between the light source and sensor pixel x, then h(t) a (t ¨ 1(x) II c) where c is the speed of light and /(x) is the length of the path from the light source to the scene point corresponding to x and back to the sensor.
[0025] Assuming the scene is static, the path length /(x) can be recovered by capturing a pair of images at the same frequency but two different modulation phases = 0 and 4) = 772:
1(x) = ¨ atan2 (3)
26.) The pixel depth z(x) can be computed from /(x) using the geometric calibration parameters of the light source and sensor.
Epipolar Time of Flight [0026] FIG. 2 is a schematic view of a system for performing epipolar ToF
imaging. A projector that generates a steerable sheet of modulated laser light is combined with a ToF sensor whose rows can be exposed one at a time. The projector and sensor are placed in a rectified stereo configuration so that the light sheet always lies in an epipolar plane between the projector and the camera. At any given instant, only the row of camera pixels in the epipolar plane are exposed to light.
[0027] To realize the geometry of FIG. 2, a line laser source with a 1D
scanning mirror that projects a steerable light sheet out onto the scene is used, as shown in FIG.
4(b). No current CW-ToF sensor provides controllable exposure coding across the 2D
pixel array. Taking into account available off-the-shelf hardware, there are three ways to restrict exposure to pixels on an epipolar plane:
1. use a digital micro-mirror device (DMD) to mask all other pixels;
2. use a 1D sensor and a controllable mirror to select the epipolar plane it should image; or 3. use a 2D sensor with a controllable region of interest (ROI).
[0028] In a preferred embodiment, the third option is chosen because it is more light-efficient than using a DMD mask, and it leads to a simpler design. The ROI is set to one row tall to match the requirements of epipolar ToF.
Epipolar Plane Sampling
[0029] CW-ToF requires at least two images to recover depth. To cover an entire scene using epipolar ToF, the active epipolar plane must be swept across the field-of-view. This offers flexibility to choose the order by which epipolar planes are sampled.
[0030] FIG. 3 illustrates several such ordering schemes. FIG. 3(a) shows conventional prior art ToF, wherein all epipolar planes are illuminated simultaneously and all camera rows are exposed at the same time. This requires long exposures and leads to severe artifacts due to motion, ambient light, global light transport and interference between devices. FIG. 3(b) shows that sending a very brief, high-intensity pulse of light for CW-ToF confers resistance to ambient light, but it is still prone to artifacts due to global light transport and motion.
[0031] FIG. 3(c) shows an ordering the epipolar ToF planes which produces an effect similar to a rolling-shutter camera, where one complete image is acquired for each modulation phase. This results in robustness to ambient light, global illumination and motion blur. Sensitivity to motion remains, however, because of the significant delay between the multiple phase measurements acquired for each row. This scheme is undesirable because if the scene or camera move while acquiring these images, the recovered depth map will contain hard to correct errors.
[0032] Another embodiment in FIG. 3(d) shows that interleaving measurements plane by plane minimizes such artifacts. The ordering strategy shown in FIG
3(d) loops through the set of modulation phases one epipolar plane at a time.
Because the exposure time of each row is very short, all phases required for a single row can be acquired quickly enough to minimize depth and motion blur artifacts due to camera/scene motion.
[0033] Using this strategy, each row is captured at a slightly different time.
Although this induces a rolling shutter-like effect in the acquired depth map, the individual depth values will be blur- and artifact-free and can be combined into a consistent model by post- processing.
[0034] To make such post-processing even easier while obeying the kinematic constraints of the mirror's actuator, epipolar planes are ordered in a sawtooth pattern, as shown in FIG. 3(e). In this scheme, the entire field of view is scanned twice within the same total exposure time, yielding a higher temporal sampling of the scene and making consistent merging of individual depth map rows easier. This essentially provides full field-of-view depth maps at twice the frame rate but half the vertical resolution, making depth correction easier for fast camera shake and/or scene motions.
[0035] More generally, FIG. 3(f) shows that, for certain applications, scanning different portions of the field of view with different temporal sampling rates can be beneficial. An example of a non-uniform sampling scheme in which epipolar planes corresponding to lower image rows are sampled more frequently is shown. This type of sampling could be useful, for example, on a vehicle where lower portions of the field of view are usually closer and move faster, requiring acquisition at a faster sampling rate.
[0036] In operation, the projector generates a sheet of modulated laser light and sequentially illuminates epipolar planes defined between the laser projector and the sensor. The planes may be illuminated in any order, but, in a preferred embodiment, are illuminated from top-to-bottom and then bottom-to-top. The actual order in which the planes are illuminated may be dependent upon the particular environment in which the platform is being used or the application for which the depth map is being created. Also, any number of planes may be defined within the field-of-view, limited only by the capabilities of the laser and the sensor, and the desired frame rate. In a preferred embodiment, there are 240 planes defined in the field-of-view, with each plane being 320 x 240 pixels.
[0037] The region of interest of the sensor can be set to any portion of the field-of-view and, in operation, a microcontroller synchronizes the laser projector and the sensor such that the ROT of the sensor is set to sense a row of pixels within the currently illuminated epipolar plane. Phase is estimated using two images. In general, the sensor uses 4, measurements for correlating the incoming signal with shifted input signals (angles 0, 90, 180, 270). Either 2 or 4 of these images can be used for phase estimation, however, using 4 images gives more accuracy but takes longer to capture and reduces frame rate. If phase unwrapping is necessary, the phase estimation process will need to be performed at different modulation frequencies, and, as such, 4 images instead of 2 images will be required for phase unwrapping. In certain embodiments of the invention, an inertial measurement unit (IMU) may be attached to the sensor and is used to compensate for motion of the platform.
Epipolar ToF Prototype
[0038] A prototype device for epipolar ToF imaging, shown in FIG. 4(a) was constructed using a galvomirror-based light sheet projector for illumination and a ToF
sensor with adjustable region of interest for imaging. FIG. 4 is a depiction of the prototype. A DME660 camera with fast ROT control to capture arbitrary rows of pixels, and a custom-built, steerable light sheet projector as the light source were used in the prototype. It should be realized by one of skill in the art that the prototype described herein is only an exemplar of one particular embodiment of the invention, and that other embodiments utilizing different equipment and operational parameters fall within the scope of the invention.
[0039] The ToF sensor used is the EPC660 (from Espros Photonics) which has a resolution of 320x240 and the pixels implement ambient saturation prevention.
The sensor is fitted with an 8mm F1.6 low distortion lens and an optical bandpass filter (650nm center frequency, 20nm bandwidth). The sensor allows the ROT to be changed with every sensor readout and this feature is used to select different rows to image. To read data out of the sensor, the sensor development kit (DME660) from the manufacturer is utilized. It should be realized that the invention is not limited to the use of the described ToF sensor, but that any ToF sensor might be used.
[0040] The line projector utilized for the prototype uses a 638nm laser diode with a peak power of 700mW as its light source. Light from the diode is collimated and passed through a Powell lens that stretches the beam cross-section into a diverging, almost uniformly illuminated straight line with a 45 degree fanout angle. The laser light is directed at a 1D scanning galvomirror that can be rotated to deflect the sheet.
The rotational range of the mirror gives the projector a 40 degree vertical field of view. The projector's effective center of projection moves as the mirror rotates, but because the distance between the fanout point and the galvomirror is very small compared to depths in the scene, this effect can be ignored.
[0041] A microcontroller is used to synchronize the sensor and light source. The microcontroller may communicate with the sensor over an I2C bus to set the exposure time, modulation frequency/phase, region of interest, row and to trigger each capture.
The microcontroller may also actuate the projector's galvomirror. In addition, the microcontroller can read the camera's rotational velocity using a MEMs inertial magnetic unit (IMU) that is attached to the sensor. A frequency generator circuit allows the selection of a modulation frequency (between 11 MHz and 24 MHz in steps of 1 MHz).
[0042] The projector and camera are aligned side-by-side in a rectified stereo configuration, as required for epipolar imaging. When correctly aligned, the projected light sheet illuminates a single row of pixels in the camera, and this row is independent of depth. A mirror calibration is performed to determine the mapping between the galvomirror angle and the illuminated camera row.
Sensor Calibration
[0043] In practice, the measurements read out from the sensor, as observed, do not match their expected values. There are a number of reasons for this discrepancy, including fixed pattern noise, unequal sensitivity and crosstalk between taps and variations in the phase of the actual exposure modulation function at each pixel. The relation between the expected sensor measurements /(x) and the observed measurements !(x) is modelled using a projective correction H (,)(x) at each pixel.
(,),o I 7,1= H,D(x)[f 6) (4) 7 '7
[0044] To find H,)(x), the sensor is placed fronto-parallel to a planar surface at a set of known distances zk,k = 1, , K. For each position of the plane, sensor measurements are collected at different aperture settings (s = 1, S) to simulate the effect of varying scene albedos. For each plane position k, the path length can be (x) computed at a pixel /k (x) and from it the expected phase 26)/k. The H,)(x) that best explains the sensor measurements 16),k,s(X) can be computed by finding the correction 1(x) that minimizes the least square error between the corrected measurements and the expected phase.
[0045] These calibration parameters are dependent on both modulation frequency and exposure time so the process is repeated for all the frequencies and exposure times. Although the modulation signals passed to the sensor and light source driver are square waves, at modulation frequencies of 20 MHz and above, the harmonics were largely suppressed and so the modulation functions were well approximated by sinusoids.
Timing
[0046] The time needed to image a row (and by extension the frame rate) with the prototype is a function of n, the number of readouts per row, exposure time texp, the readout time for a row tread and tmirror, the time taken by the galvomirror to move to the next row position in the sampling sequence.
trOW ntexp (n ¨ 1)tread MaX(tread, tniirror) (5)
[0047] With a two-tap sensor like the one used in our prototype, at least n = 2 readouts are needed to measure depth using a single modulation frequency. FIG.

shows timing diagrams for camera exposure, readout and mirror position for a particular sequence of the rows. First, the scanning mirror is moved to the new active row and takes t -mirror time to settle in the position. When the previous row readout is complete (which takes tread time) and the mirror is in position, the camera is triggered. In this example, tmirror > td so the speed of the mirror is a bottleneck for capture rate. Each exposure lasts for time texp and at the end of each exposure the row is read. FIG. 5 shows a timing example. troy, is 175 ps and texp is set to 100 ps.
In a row sampling sequence, the mirror rotates through two rows (approximately 0.33 ) per step and the settling time tniir, for this step size is roughly 100 ps. In total, troy, works out to 550 ps when n= 2, which yields a framerate of 7.5 fps (3.8 fps when n= 4).
Limitations
[0048] Currently, the main bottleneck for the frame rate is the readout time.
Embodiments of the present invention need data from only one row of the sensor per readout, but the smallest region of interest the EPC660 sensor supports is 4 rows tall, the reading of 4 rows is forced when in actuality, only one row is used. In addition, the development kit limits the sensor data bus to 20 MHz, but the sensor itself supports bus rates up to 80 MHz. The minimum value of texp depends on the peak power of the light source and desired range. The described prototype of the present invention has a source with a peak power of 700 mW, while most other experimental time-of-flight systems have a peak light source power in the 3 W to 10 W
range. With a brighter light source, a shorter exposure time could be used without loss of range.
Lastly, the low cost galvomirror could be replaced with a faster 1D MEMs mirror.
With these improvements, a system based on the described prototype would operate at video frame rates.
[0049] The sensor used in the described prototype supports a maximum modulation frequency of only 24 MHz, whereas most other time-of-flight sensors can run in the 50 MHz to 100 MHz range. This limits the ability of the prototype to accurately scan smaller objects or to be used for transient imaging. The datasheet specifies that the sensor ADC returns 12 bit values, but the version of the sensor which was used returns only 10 bits, which effects range and makes the output depth maps noisier.
Results
[0050] To run the sensor in regular imaging mode for comparing performance under ambient light and global illumination, the entire sensor is exposed at once instead of using a small ROT and the sensor is left exposed until the sheet projector has finished a sweep across the field of view. For regular ToF imaging in the multi-device interference and camera motion experiments, the sheet projector can be replaced with a diffused source.
Ambient Light
[0051] The benefits of applying epipolar imaging to time-of-flight in brightly lit environments were simulated, and the results shown in FIG. 7. For a given light source power, with regular imaging, depth accuracy degrades rapidly as ambient light levels increase from 0 lx (complete darkness) to 100 klx (direct sunlight).
With epipolar imaging, the degradation is much more gradual.
[0052] FIG. 7(a) shows the results of a simulation of the standard deviation in depth measurements obtained using regular and epipolar ToF imaging (15 MHz modulation frequency) for a target 10m from the camera as a function of ambient light level. For both cases, the peak light source power is 2W and the total exposure time is the same (7.2 ms per image), but epipolar ToF is more robust to ambient light because it concentrates light source power and uses a short exposure for each row (30 [is).
[0053] FIG. 7(b) shows the working range of the same simulated camera at different levels of acceptable range accuracy. Note that the simulated camera's parameters differ from the prototype.
[0054] FIG. 6 quantitatively compares the described prototype in regular ToF and epipolar ToF modes in cloudy and sunny conditions. Regular ToF mode fails in bright sunlight, while epipolar ToF is considerably more robust. FIG. 1(a) shows a live 3D
CW-ToF imaging in sunlight (70 klx) with 15m range (people walking on stairs, phase wraps around distant building), and FIG. 8 shows an example scene with both strong ambient light and global illumination effects. Reflections from the table service cause errors with regular ToF, but these are suppressed with epipolar imaging.
[0055] FIG. 6(d) shows standard deviation in depth measurements versus distance to target (slower rising curves are better). The prototype of the present invention has depth error of around 3% at 10m in bright sunlight.
Global Illumination
[0056] FIG. 9 demonstrates the ability of epipolar imaging to suppress the effects of global illumination in a few common indoor environments. These results are generated using a single modulation frequency (24 MHz). At the corner of the room, diffuse inter-reflections between the walls and ceiling cause depths to be overestimated and a rounding out of the corner with regular imaging.
[0057] The conference table in the second row of FIG. 9, also shown in FIG.
1(b), appears specular at grazing angles. In the bathroom scene, the ghosting on the wall due to reflections from the mirror is suppressed by epipolar imaging. The water fountain is particularly challenging because the direct return from its metallic surface is very weak, but the surface reflects a lot of indirect light back to the sensor. For epipolar imaging, 3 exposures are combined to try to recover a useable direct signal.
Longer exposures do not help regular imaging because the inter-reflections cause the sensor to saturate.
[0058] With epipolar imaging, the walls appear straight and meet at a sharp right angle. diffuse inter-reflections at the corner, glossy inter-projections from the projection screen onto a shiny conference table, reflections from the mirrors in the restroom and in between the wall and the shiny water fountain. Epipolar ToF
eliminates most of the global light transport, resulting in depth maps that are significantly more accurate than regular ToF.
Multi-Camera Interference
[0059] With epipolar CW-ToF imaging, two cameras running at the same modulation frequency can usually only interfere with each other at a sparse set of pixels in each image. Each camera illuminates and images a single line in the scene at a time, so at any point of time the second camera can only interfere with the first camera at the points where its illuminated line intersects with the first camera's exposed row of pixels. A degenerate case occurs when the light source of one camera forms a rectified stereo pair with the sensor of the second camera and both cameras happen to be synchronized, but this can be considered a rare occurrence.
[0060] If more than two cameras are present, each pair of cameras has a sparse set of points where they interfere with each other. When a set of epipolar ToF
cameras are running at different modulation frequencies, the contribution of each camera to shot noise in the other cameras is greatly reduced. FIG. 1(c) shows the result of operating two CW-ToF cameras simultaneously at the same frequency with regular and epipolar imaging. Epipolar imaging shoes the lack of interference between ToF
devices operating at the same frequency. There are observable errors (i.e., the wall and chair) with regular ToF.
Camera Motion
[0061] With a rotating camera having a known rotational trajectory (obtained from a MEMS gyroscope), with regular imaging, each captured ToF measurement has motion blur and strong artefacts at depth discontinuities because the measurements are not aligned to each other. In theory, these could be collected using a spatially varying deconvolution but this is computationally expensive and does a poor job of recovering high frequency components. With epipolar ToF imaging, motion blur has basically no effect and a depth map with a rolling shutter like effect is acquired. This can be corrected with a simple image warp computed from the rotation. FIG.
1(d) shows an example from a rapidly panning camera, showing that non-distorted depth maps can be obtained even in the presence of severe camera shake during scene exposure (hard-to-remove ghosting errors in regular ToF can be observed).
Further, as previously mentioned, the sensor may be equipped with and I am you, which is used to compensate promotion of the platform.
[0062] Epipolar imaging for time-of-flight depth cameras mitigates many of the problems commonly encountered with depth cameras, such as poor performance in brightly lit conditions, systemic errors due to global illumination, inter-device interference and errors due to camera motion. Compared to depth cameras, systems like scanning LIDAR that illuminate and image a single point at a time are very robust to all these effects but have a low measurement rate. Epipolar imaging can be thought of a compromise between these two extremes of full-field capture and point-by-point capture. Because epipolar imaging illuminates and captures a single line at a time, it allows a depth camera to have most of the robustness of point scanning while still having a high measurement rate.
[0063] Cycling through patterns row-by-row, as is done here for ToF, is directly applicable to structured light as well. It would make it possible to apply multi-image structured light methods that generate high quality depth maps to dynamic scenes where currently only single-shot methods can be used.
[0064] In the described prototype, the scanning mirror follows a sawtooth pattern and captures rows in an orderly sequence. However, with a faster scanning minor, pseudo random row sampling strategies could be implemented that might allow epipolar imaging to be used in conjunction with compressed sensing or similar techniques to recover temporally super-resolved depth maps of fast moving scenes.
Embodiments of the invention have been described herein using specific identified components, however, the invention is not meant to be limited thereby. The scope of the claimed invention is defined by the claim set presented below.

Claims (22)

We Claim:
1. A system comprising:
a. a modulated light source for projecting a steerable sheet of modulated light and;
b. a sensor, the modulated light source and sensor in a rectified stereo configuration such that the light sheets projected by the modulated light source lies in an epipolar plane between the projector and the sensor; and c. a microcontroller for synchronizing the modulated light source and the sensor.
2. The system of claim 1, wherein the modulated light source comprises:
a. a laser source;
b. an optical element, configured to generate a light sheet from the collimated output of the laser source; and c. a means for steering the sheet of laser light along a series of epipolar planes between the projector and the sensor.
3. The system of claim 2 wherein the means for steering the sheet of laser light is selected from a group comprising a rotatable galvomirror and a MEMS mirror.
4. The system of claim 2, wherein the sensor is a continuous wave time-of-flight camera having a controllable region of interest.
5. The system of claim 4, where in the projected light sheet illuminates a single row of pixels in the sensor, and further wherein the controllable region of interest of the sensor is all set to sense the single row of illuminated pixels.
6. The system of claim 5, wherein the laser light projector and the sensor are synchronized such that the region of interest of the sensor is set to sense a row of pixels within the currently illuminated epipolar plane.
7. The system of claim 6, wherein the sensor captures at least two images from each illuminated epipolar plane.
8. The system of claim 7 wherein the projected light sheet is modulated as a repeating wave and further wherein the sensed depth may be calculated for each sensed pixel by the phase of the returned reflection.
9. The system of claim 8 wherein a depth map of the entire field of view is created based upon the depth calculated for each sensed pixel from each illuminated epipolar plane within the field of view.
10. The system of claim 9, wherein the microcontroller reads data from the sensor regarding the previously illuminated epipolar plane while the galvomirror is rotated to a position to illuminate the next epipolar in the series of epipolar planes.
11. A method comprising:
a. projecting a sheet of modulated light along a series of epipolar planes defined by a modulated light source and a sensor placed in a rectified stereo configuration, the series of epipolar planes defining a field-of-view;
b. imaging a single row of illuminated pixels in the illuminated epipolar plane; and c. determining the depth of each pixel in the single row of illuminated pixels
12. The method of claim 11, wherein determining the depth of each pixel comprises calculating the depth of each pixel based on the phase of the reflected light from each pixel.
13. The method of claim 12 wherein calculating the depth of each pixel in an illuminated epipolar plane further comprises determining the difference in phase of reflected light contained in two or more separate images of the illuminated epipolar plane.
14. The method of claim 12, further comprising creating a depth map based on the depth of each pixel in each illuminated epipolar plane within the defined field-of-view.
15. The method of claim 11, further comprising synchronizing the modulated light source and the sensor, such that the region of interest of the sensor corresponds to the currently illuminated epipolar plane.
16. The method of claim 11 wherein the epipolar planes in the field-of-view are illuminated in a varying order.
17. The method of claim 11 wherein the modulated light source comprises:
a. a laser source;
b. an optical element, configured to generate a light sheet from the collimated output of the laser source; and c. a means for steering the sheet of laser light along a series of epipolar planes between the projector and the sensor.
18. A non-transitory computer-readable media containing software that when executed, performs the functions of:
a. projecting a sheet of modulated light along a series of epipolar planes defined by a modulated light source and a sensor placed in a rectified stereo configuration, the series of epipolar planes defining a field-of-view;
b. imaging a single row of illuminated pixels in the illuminated epipolar plane;

c. determining the depth of each pixel in the single row of illuminated pixels; and d. creating a depth map based on the depth of each pixel in each illuminated epipolar plane within the defined field-of-view.
19. The non-transitory computer-readable media of claim 18 wherein determining the depth of each pixel comprises calculating the depth of each pixel based on the phase of the reflected light from each pixel and wherein calculating the depth of each pixel in an illuminated epipolar plane further comprises determining the difference in phase of reflected light contained in two or more separate images of the illuminated epipolar plane.
20. The non-transitory computer-readable media of claim 18, wherein the software performs the further function of synchronizing the modulated light source and the sensor, such that the region of interest of the sensor corresponds to the currently illuminated epipolar plane.
21. The non-transitory computer-readable media of claim 20 wherein the epipolar planes in the field-of-view are illuminated in a varying order.
22. The method of claim 11 wherein the modulated light source comprises:
a. a laser source;
b. an optical element, configured to generate a light sheet from the collimated output of the laser source; and c. a means for steering the sheet of laser light along a series of epipolar planes between the projector and the sensor.
CA3051102A 2017-01-20 2018-01-19 Method for epipolar time of flight imaging Pending CA3051102A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201762499193P 2017-01-20 2017-01-20
US62/499,193 2017-01-20
PCT/US2018/014369 WO2018136709A1 (en) 2017-01-20 2018-01-19 Method for epipolar time of flight imaging

Publications (1)

Publication Number Publication Date
CA3051102A1 true CA3051102A1 (en) 2018-07-26

Family

ID=62909043

Family Applications (1)

Application Number Title Priority Date Filing Date
CA3051102A Pending CA3051102A1 (en) 2017-01-20 2018-01-19 Method for epipolar time of flight imaging

Country Status (4)

Country Link
EP (1) EP3571467A4 (en)
JP (1) JP7244013B2 (en)
CA (1) CA3051102A1 (en)
WO (1) WO2018136709A1 (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020214201A1 (en) * 2019-04-17 2020-10-22 Carnegie Mellon University Agile depth sensing using triangulation light curtains
US11972586B2 (en) 2015-02-13 2024-04-30 Carnegie Mellon University Agile depth sensing using triangulation light curtains
US11747135B2 (en) 2015-02-13 2023-09-05 Carnegie Mellon University Energy optimized imaging system with synchronized dynamic control of directable beam light source and reconfigurably masked photo-sensor
EP3620821A1 (en) * 2018-09-05 2020-03-11 Infineon Technologies AG Time of flight camera and method for calibrating a time of flight camera
US11467270B2 (en) * 2019-03-27 2022-10-11 Asmpt Singapore Pte. Ltd. Apparatus and method for calibrating or testing an imaging device
US11619723B2 (en) 2019-09-23 2023-04-04 Microsoft Technology Licensing, Llc Multiple-mode frequency sharing for time-of-flight camera
CN111077538A (en) * 2019-12-29 2020-04-28 中国科学院西安光学精密机械研究所 Dynamic high-precision optical combined imaging method and system for marine complex environment
WO2021177045A1 (en) * 2020-03-04 2021-09-10 ソニーグループ株式会社 Signal processing device, signal processing method, and range-finding module
EP4138385A4 (en) * 2020-07-20 2024-05-01 Wuxi Idata Technology Company Ltd. Scanning system and method for controlling aiming light source

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002039716A (en) * 2000-07-25 2002-02-06 Olympus Optical Co Ltd Depth map input device
EP1971820B1 (en) * 2006-01-08 2012-10-17 Hermann Tropf Creation of a distance image
JP2012168049A (en) * 2011-02-15 2012-09-06 Stanley Electric Co Ltd Distance image generation device and method
JP5858688B2 (en) * 2011-08-30 2016-02-10 スタンレー電気株式会社 Distance image generator
JP6309459B2 (en) * 2012-02-15 2018-04-11 ヘプタゴン・マイクロ・オプティクス・ピーティーイー・エルティーディーHeptagon Micro Optics Pte.Ltd. Time-of-flight camera with stripe lighting
US9476695B2 (en) * 2013-07-03 2016-10-25 Faro Technologies, Inc. Laser tracker that cooperates with a remote camera bar and coordinate measurement device
WO2015115797A1 (en) * 2014-01-29 2015-08-06 엘지이노텍 주식회사 Device for extracting depth information and method thereof
US9389069B2 (en) * 2014-03-26 2016-07-12 Alces Technology, Inc. Compact 3D depth capture systems
WO2016131036A1 (en) * 2015-02-13 2016-08-18 Carnegie Mellon University Imaging system with synchronized dynamic control of directable beam light source and reconfigurably masked photo-sensor

Also Published As

Publication number Publication date
WO2018136709A1 (en) 2018-07-26
EP3571467A1 (en) 2019-11-27
JP2020504310A (en) 2020-02-06
JP7244013B2 (en) 2023-03-22
EP3571467A4 (en) 2020-08-12

Similar Documents

Publication Publication Date Title
CA3051102A1 (en) Method for epipolar time of flight imaging
US11425357B2 (en) Method for epipolar time of flight imaging
US10764517B2 (en) Stereo assist with rolling shutters
US11002856B2 (en) Doppler time-of-flight imaging
US10935371B2 (en) Three-dimensional triangulational scanner with background light cancellation
US11375165B2 (en) Image calibration for projected images
US20130148102A1 (en) Method to Compensate for Errors in Time-of-Flight Range Cameras Caused by Multiple Reflections
JP7371443B2 (en) 3D measuring device
Wang et al. Programmable triangulation light curtains
JP7201592B2 (en) System for characterizing vehicle surroundings
CN106896370B (en) Structured light ranging device and method
US20200018592A1 (en) Energy optimized imaging system with synchronized dynamic control of directable beam light source and reconfigurably masked photo-sensor
CN112513670B (en) Distance meter, distance measuring system, distance measuring method and program
CN206740977U (en) Structure light measurement device
JP6369897B2 (en) Self-position calculation device and self-position calculation method
CN109100740B (en) Three-dimensional image imaging device, imaging method and system
WO2021084891A1 (en) Movement amount estimation device, movement amount estimation method, movement amount estimation program, and movement amount estimation system
US11610339B2 (en) Imaging processing apparatus and method extracting a second RGB ToF feature points having a correlation between the first RGB and TOF feature points
US20170287140A1 (en) High quality Lightning resilient segmentation system using active background
Langmann et al. Real-time image stabilization for ToF cameras on mobile platforms
Maas Close range photogrammetry sensors
WO2023176127A1 (en) Three-dimensional measurement device
Langmann et al. PMD imaging
US20180286062A1 (en) Information processing device, information processing method, program, and image capturing device
JP2010281733A (en) Three-dimensional optical image-forming system

Legal Events

Date Code Title Description
EEER Examination request

Effective date: 20220727

EEER Examination request

Effective date: 20220727

EEER Examination request

Effective date: 20220727

EEER Examination request

Effective date: 20220727

EEER Examination request

Effective date: 20220727