WO2016149570A1 - Capteurs et procédés d'imagerie spectrale avec détection de temps de vol - Google Patents

Capteurs et procédés d'imagerie spectrale avec détection de temps de vol Download PDF

Info

Publication number
WO2016149570A1
WO2016149570A1 PCT/US2016/023012 US2016023012W WO2016149570A1 WO 2016149570 A1 WO2016149570 A1 WO 2016149570A1 US 2016023012 W US2016023012 W US 2016023012W WO 2016149570 A1 WO2016149570 A1 WO 2016149570A1
Authority
WO
WIPO (PCT)
Prior art keywords
light
array
spectral imaging
apertures
imaging sensor
Prior art date
Application number
PCT/US2016/023012
Other languages
English (en)
Other versions
WO2016149570A8 (fr
Inventor
Daniel L. Lau
Gonzalo R. ARCE
Original Assignee
University Of Delaware
The University Of Kentucky
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University Of Delaware, The University Of Kentucky filed Critical University Of Delaware
Priority to US15/559,228 priority Critical patent/US10151629B2/en
Publication of WO2016149570A1 publication Critical patent/WO2016149570A1/fr
Publication of WO2016149570A8 publication Critical patent/WO2016149570A8/fr

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/02Details
    • G01J3/0205Optical elements not provided otherwise, e.g. optical manifolds, diffusers, windows
    • G01J3/0229Optical elements not provided otherwise, e.g. optical manifolds, diffusers, windows using masks, aperture plates, spatial light modulators or spatial filters, e.g. reflective filters
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/02Details
    • G01J3/10Arrangements of light sources specially adapted for spectrometry or colorimetry
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/02Details
    • G01J3/10Arrangements of light sources specially adapted for spectrometry or colorimetry
    • G01J3/108Arrangements of light sources specially adapted for spectrometry or colorimetry for measurement in the infrared range
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/28Investigating the spectrum
    • G01J3/2823Imaging spectrometer
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/28Investigating the spectrum
    • G01J3/2889Rapid scan spectrometers; Time resolved spectrometry
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • G01S17/08Systems determining position data of a target for measuring distance only
    • G01S17/32Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated
    • G01S17/36Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated with phase comparison between the received signal and the contemporaneously transmitted signal
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • G01S17/8943D imaging with simultaneous measurement of time-of-flight at a 2D array of receiver pixels, e.g. time-of-flight cameras or flash lidar
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/481Constructional features, e.g. arrangements of optical elements
    • G01S7/4816Constructional features, e.g. arrangements of optical elements of receivers alone
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/483Details of pulse systems
    • G01S7/486Receivers
    • G01S7/4861Circuits for detection, sampling, integration or read-out
    • G01S7/4863Detector arrays, e.g. charge-transfer gates

Definitions

  • the present invention relates generally to spectral imaging, and more particularly, to spectral imaging sensors and methods employing time of flight sensing.
  • spectral information In recent years, the ability to acquire spectral information for a field of view has become desirable in numerous applications. Such information can include a complete spectrum or just some spectral information at every location in an image plane. As such, spectral image sensors must be capable of obtaining a large amount of spatial information across a multitude of wavelengths in a field of view ("spatio-spectral information").
  • aspects of the present invention are directed to spectral imaging sensors and methods.
  • a spectral imaging sensor includes a light source, an array of coded apertures, one or more optical elements, and a photodetector.
  • the light source is configured to emit a plurality of pulses of light toward an object to be imaged.
  • the array of coded apertures is positioned to spatially modulate light received from the object to be imaged.
  • the optical elements are configured to redirect light from the array of coded apertures.
  • the photodetector is positioned to receive light from the one or more optical elements.
  • the photodetector comprise a plurality of light sensing elements. The plurality of light sensing elements are operable to sense the light from the one or more optical elements in a plurality of time periods.
  • the plurality of time periods have a same frequency as the plurality of pulses of light.
  • a spectral imaging method includes emitting a plurality of pulses of light toward an objected to be imaged, spatially modulating light from the object to be imaged with an array of coded apertures, redirecting the spatially modulated light with one or more optical elements, and receiving the redirected light with a photodetector comprising a plurality of light sensing elements, the plurality of light sensing elements operated to sense the redirected light in a plurality of time periods, the plurality of time periods having a same frequency as the plurality of pulses of light.
  • FIG. 1 is a diagram illustrating an exemplary spectral imaging sensor in accordance with aspects of the present invention
  • FIG. 2 is a diagram illustrating an optical path of the exemplary sensor of
  • FIG. 1 A first figure.
  • FIG. 3 is a timing diagram illustrating the operation of a photodetector of the exemplary sensor of FIG. 1 ;
  • FIG. 4 is a flowchart illustrating an exemplary spectral imaging method in accordance with aspects of the present invention.
  • FIGS. 1 and 2 illustrate an exemplary spectral imaging sensor 100 in accordance with aspects of the present invention.
  • Sensor 100 may be usable in photographic or spectroscopic systems.
  • sensor 100 includes a light source, an array 120 of coded apertures, optical elements 130, and a photodetector 140. Additional details of sensor 100 are described below.
  • Light source 110 is configured to emit light toward the object 50 to be imaged.
  • Light source 110 operates by emitting a plurality of pulses of light.
  • light source 110 may be operated with a 50% duty cycle (e.g., light source is on and off for the same period of time).
  • Suitable frequencies for pulsing light source 110 include from 10 to 80 MHz (i .e. with a pulse length from 0.1 to 0.0125 ps).
  • light source 110 is a laser diode.
  • Suitable laser diodes for use as light source 110 include, for example, near-infrared laser diodes. 5
  • Other suitable light sources will be known to one of ordinary skill in the art from the description herein.
  • An array 120 of coded apertures is positioned to receive light from object 50.
  • Array 120 receives both ambient light reflected by object 50 and light emitted by light source 110 and reflected by object 50.
  • Array 120 is formed from a two-dimensional l o array of coded apertures 122, and is configured to spatially modulate light from object 50.
  • Array 120 thus represents a 3D array, where the two axes orthogonal to the light from object 50 represent a spatial distribution, and the third axis parallel to the light from object 50 represents a distribution of spatial modulation results from apertures 122.
  • spatial modulation is intended to encompass a
  • apertures 122 may allow all of the light from object 50 to pass therethrough. Apertures 122 may allow a portion of the light from object 50 to pass therethrough. Apertures 122 may
  • apertures 122 may be controlled to absorb or pass all or a portion of the light from object 50. Apertures 122 may also be controlled to redirect the light from object 50 in one or more different directions. Other examples of spatial modulation performed by apertures 122 will be known to one of ordinary skill in the art from the description
  • array 120 of coded apertures produces (either by passing or reflecting/redirecting) a beam of spatially modulated light from object 50.
  • array 120 is an array of static coded apertures.
  • the spatial modulation by each aperture 122 does not change.
  • the array of static coded apertures may be, for example, a color-coded array of
  • Such an array may include a first plurality of apertures configured to pass light in a first predetermined wavelength range and a second plurality of apertures configured to pass light in a second predetermined wavelength range different from the first predetermined wavelength range. It may further be preferable that all wavelength ranges incorporated by apertures in the color-coded array be designed to pass a
  • array 120 is an array of dynamic coded apertures.
  • the spatial modulation by each aperture 122 may be changed (e.g., through electronic or mechanical adjustment).
  • the array of dynamic coded apertures may be produced, for example, by a spatial light modulator.
  • a spatial light modulator (such as a liquid crystal on-silicon modulator or a liquid crystal phase- only modulator) is a modulator that imposes a spatially varying modulation on a beam of light.
  • the modulation of light through each aperture may binary (e.g. on/off) or may be continuously varying (e.g. from 0-100% intensity) at each location (or aperture).
  • the modulation may be controlled by a computer or other digital processing element.
  • Suitable spatial light modulators for use as array 120 will be known to one of ordinary skill in the art from the description herein.
  • the array of dynamic coded apertures may also be, for example, a digital mirror device.
  • a digital micromirror device includes an array of individually controllable microscopic mirrors that impose a spatially varying reflection on a beam of light. The reflection may be toward a desired optical element (e.g . lens) or may be out of the intended optical path (e.g . to a heat sink) at each location (or aperture). The direction of reflection may be controlled by a computer or other digital processing element. Suitable digital micromirror devices for use as array 120 will be known to one of ordinary skill in the art from the description herein.
  • Optical elements 130 are positioned to receive light spatially modulated by array 120. Optical elements 130 are configured to redirect the light from array 120 onto photodetector 140.
  • optical elements 130 include an imaging lens 132 and one or more prisms 134, as shown in FIG. 2.
  • Imaging lens 132 focuses the light from array 120, and prisms 134 refract the light passing through imaging lens 132.
  • prisms 134 are an Amici prism, as shown in FIG. 2.
  • Suitable lenses and prisms for use as optical elements 130 will be known to one of ordinary skill in the art from the description herein.
  • Refraction of the light with prisms 134 separates the different wavelength bands of light modulated by array 120. This is because the amount of refraction caused by prisms 134 is dependent on the wavelength of the light passing therethrough. In other words, prisms 134 will refract longer wavelength (e.g. red) light to a greater extent than they will shorter wavelength (e.g. blue) light. Due to this refraction, prisms 134 will cause different wavelengths of light from the same region of object 50 to strike photodetector 140 at different locations (due to the different amounts of refraction). Where photodetector 140 is a photodetector array of light sensing elements (e.g.
  • each element will receive spectral information about object 50 from differential spatial areas of the light from object 50.
  • Optical elements 130 are not limited to the elements described herein. Other suitable optical elements for use in sensor 100 will be known to one of ordinary skill in the art from the description herein. For example, suitable optic elements include diffractive elements such as gratings. Other suitable elements for use as optical elements 130 will be known to one of ordinary skill in the art from the description herein.
  • optical elements 130 include a beam splitter.
  • a beam splitter would allow light from object 50 to pass therethrough to array 120, and then reflect light from array toward other optical elements 130 and/or photodetector 140. In other embodiments, the beam splitter would reflect light from object 50 onto array 120, and allow light reflected by array 120 to pass therethrough on toward other optical elements 130 and/or photodetector 140. Suitable structures for use as a beam splitter will be known to one of ordinary skill in the art from the description herein.
  • Photodetector 140 is positioned to receive light from optical elements 130. Photodetector 140 collects the light passing from array 120 and converts it into spatio- spectral image information and surface depth information about object 50.
  • Photodetector 140 may be configured to detect light in any region of the optical electromagnetic spectrum. In particular, photodetector 140 collects both ambient light reflected by object 50 and light emitted by light source 110 and reflected by object 50. Data relating to these separate types of light may be processed and stored separately by photodetector 140 (and related processing elements).
  • photodetector 140 comprises a plurality of light sensing elements 142, as shown in FIG. 2.
  • the light sensing elements 142 may be an array of pixels, such as a focal plane array. Suitable pixel arrays will be known to one of ordinary skill in the art from the description herein.
  • Light sensing elements 142 sense the light modulated by array 120 and redirected by optical elements 130. Light sensing elements 142 are operated to sense this light in a plurality of separate time periods. Conventional pixel arrays obtain an image during a single time period. In such an operation (known as shuttering), each pixel is turned on for a predetermined time period, and then turned off. Light sensing elements 142, by contrast, are turned on and off a plurality of times to obtain information used to create an image.
  • the plurality of time periods in which light sensing elements 142 sense light are each shorter than the single time period for conventional pixel arrays, and may collectively cover approximately the same length of time as the single time period for conventional pixel arrays (e.g., 15.0 ms).
  • Suitable image sensors for use as photodetector include, for example, the image sensor of the epc660 3D Time-of-Flight QVGA Imager, provided by ESPROS Photonics Corporation.
  • Other suitable photodetectors will be known to one of ordinary skill in the art from the description herein.
  • each light sensing element 142 is connected to a processing element (not shown).
  • the processing element(s) may actuate all of the light sensing elements 142 together (in phase), or may subdivide the light sensing elements 142 into sections, columns, or rows, and actuate each subdivision of light sensing elements 142 according to its own timing.
  • a single processing element may individually actuate all of the light sensing elements 142, or multiple processing elements may be used (e.g. one for each section, column, or row of light sensing elements).
  • Suitable processing elements may be found, for example, in the DME660 3D Time-of-Flight camera, also provided by ESPROS Photonics Corporation. Other suitable processing elements will be known to one of ordinary skill in the art from the description herein.
  • the plurality of time periods in which light sensing elements 142 sense light have the same frequency as the pulses of light emitted by light source 110.
  • light sensing elements 142 sense light in a pulsed fashion, similar to the pulsing of light source 110.
  • Light sensing elements 142 may sense light with a 50% duty cycle, and at the same frequency or pulse length recited above for the light pulses from light source 110.
  • the pulses of light from light source 110 are timed to be emitted in phase with the time periods of light sensing elements 142.
  • FIG. 3 is a timing diagram 160 illustrating the operation of photodetector 140.
  • the x-axis of FIG. 3 represents time; the y-axis of FIG. 3 represents an estimated range relative to object 50 from a predetermined point.
  • pulses of light from light source 110 are received by photodetector 140, they will overlap with the time periods of sensing by light sensing elements 142.
  • the columns labeled A and B in FIG. 3 represent two groups of light sensing elements 142, or pixels.
  • a pixels are timed to collect light when light source 110 is turned on, while B pixels are timed to collect light when the light source 110 is turned off.
  • a pixels and B pixels may be groups of d ifferent pixels.
  • a pixels and B pixels may include the same pixels of photodetector 140.
  • each pixel is switched between a first electron collection region (e.g ., an A capacitor) and a second electron collection region (e.g . , a B capacitor) .
  • each pixel may produce two separate values of light received for A and B.
  • the top row of pulses overlap perfectly with the on period for the A pixels; in other words, the entire pulse is received by photodetector 140 when the A pixels are on. Because the light pulses and time periods share the same frequency, this relationship is maintained over a series of pulses. In the next row down, there is a short lag between the start of the on time period for the A pixels and receipt of the light pulse (resulting from the light pulse travelling a longer distance to object 50) . As a result, less than the entire pulse is received by photodetector 140 within the on time period of the A pixels. This relationship is continued down the diagram of FIG.
  • sensor 100 is not limited to the above-described components, but may include alternative or additional components as would be understood to one of ordinary skill in the art.
  • sensor 100 may include an objective lens 150, as shown in FIG. 2.
  • Objective lens 150 is positioned to receive the light from object 50, and is positioned between object 50 and array 120 of coded apertures. Objective lens 150 thereby functions to focus light from object 50 onto array 120 of coded apertures. Objective lens 150 focuses both ambient light reflected by object 50 and light emitted by light source 110 and reflected by object 50.
  • the distance f illustrated in FIG. 2 directly depends on the used objective lens. This relates to the focal length from the lens to the plane where the optical images are formed (i. e. , the image plane of the objective lens) .
  • Sensor 100 may also include an adjustment mechanism.
  • the adjustment mechanism is configured to adjust a position of array 120 relative to object 50.
  • the adjustment mechanism may be configured to move array 120 orthogonally to the light from object 50 by a predetermined number of apertures. Such movement may be particularly useful for arrays of static coded apertures, in order to adjust the modulation by the array.
  • the adjustment mechanism may be configured to adjust a position of the one or more optical elements 130 relative to object 50. Suitable adjustment mechanisms will be known to one of ordinary skill in the art from the description herein.
  • FIG. 4 illustrates an exemplary method 200 for spectral imaging in accordance with aspects of the present invention .
  • Method 200 may be usable for photographic or spectroscopic imaging .
  • method 200 includes emitting light, spatially modulating light with an array of coded apertures, redirecting the spatially modulated light, and receiving the redirected light. Additional details of method 200 are described below with respect to the components of sensor 100.
  • step 210 light is emitted from a light source.
  • a plu rality of light pulses are emitted by light source 110.
  • the light pulses may be emitted at a 50% duty cycle.
  • the light pulses may be emitted at a frequency from 10 to 80 MHz (i .e. with a pulse length from 0.1 to 0.0125 [is) .
  • step 220 light from an object to be imaged is spatially modulated .
  • light from object 50 is spatially modulated by array 120 of coded apertures.
  • Array 120 is positioned to receive light from object 50 and spatially modulate that light.
  • the light from object 50 may be statically spatially modulated with an array of static coded apertures.
  • the light from object 50 may be dynamically spatially modulated with an array of dynamic coded apertures.
  • the spatially modu lated light is redirected .
  • optical elements 130 redirect the spatially modulated light from array 120.
  • one or more prisms 134 may refract the modulated light in order to separate the light modulated by array 120 into different wavelength bands.
  • the redirected light is received.
  • photodetector 140 receives the light redirected by optical elements 130. Where photodetector 140 includes a plurality of light sensing elements 142, the light sensing elements are operated to sense the redirected light in a plurality of separate time periods. The plurality of time periods in which light sensing elements 142 sense light have the same frequency as the pulses of light emitted by light source 110. In a further embodiment, the pulses of light are emitted in phase with the time periods of light sensing elements 142.
  • method 200 is not limited to the above-described steps, but may include alternative or additional components as would be understood to one of ordinary skill in the art.
  • a spectral three-dimensional image may be created.
  • photodetector 140 converts the light modulated by array 120 and redirected by optical elements 130 into spatio-spectral image information and surface depth information about object 50. This information may then utilized by photodetector 140, along with several processing elements (not shown) to create a spectral three-dimensional image of object 50.
  • Suitable algorithms for creating a spectral three-dimensional image from the spatially modulated light received by photodetector 140 will be known to one of ordinary skill in the art from the description herein.
  • sensor 100 may include an objective lens 150, as shown in FIG. 2.
  • method 200 may include the step of focusing the light from object 50 onto array 120 of apertures with objective lens 150.
  • the light received with photodetector 140 may be processed.
  • processing may include a step of separating ambient light from object 50 and modulated light reflected from object 50.
  • the ambient light may be processed in accordance with compressive sensing in order to procure spatio-spectral image information regarding object 50.
  • the modulated light may be processed as set forth above to obtain surface depth information regarding object 50.

Landscapes

  • Physics & Mathematics (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Electromagnetism (AREA)
  • Measurement Of Optical Distance (AREA)
  • Spectrometry And Color Measurement (AREA)
  • Color Television Image Signal Generators (AREA)
  • Optical Radar Systems And Details Thereof (AREA)

Abstract

L'invention concerne des capteurs et des procédés d'imagerie spectrale. Un capteur d'imagerie spectrale comprend une source de lumière, un réseau d'ouvertures codées, un ou plusieurs éléments optiques, et un photodétecteur. La source de lumière est configurée pour émettre une pluralité d'impulsions de lumière vers un objet devant subir une imagerie. Le réseau d'ouvertures codées est positionné de manière à moduler spatialement la lumière reçue en provenance de l'objet devant subir une imagerie. Les éléments optiques sont configurés pour rediriger la lumière provenant du réseau d'ouvertures codées. Le photodétecteur est positionné pour recevoir de la lumière à partir du ou des éléments optiques. Le photodétecteur comprend une pluralité d'éléments de détection de lumière. La pluralité d'éléments de détection de lumière peut être utilisée pour détecter la lumière provenant d'un ou de plusieurs éléments optiques pendant une pluralité de périodes de temps. Les périodes de la pluralité de périodes de temps ont la même fréquence que les impulsions de la pluralité d'impulsions de lumière.
PCT/US2016/023012 2015-03-19 2016-03-18 Capteurs et procédés d'imagerie spectrale avec détection de temps de vol WO2016149570A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/559,228 US10151629B2 (en) 2015-03-19 2016-03-18 Spectral imaging sensors and methods with time of flight sensing

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201562135233P 2015-03-19 2015-03-19
US62/135,233 2015-03-19

Publications (2)

Publication Number Publication Date
WO2016149570A1 true WO2016149570A1 (fr) 2016-09-22
WO2016149570A8 WO2016149570A8 (fr) 2017-10-05

Family

ID=56920165

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2016/023012 WO2016149570A1 (fr) 2015-03-19 2016-03-18 Capteurs et procédés d'imagerie spectrale avec détection de temps de vol

Country Status (2)

Country Link
US (1) US10151629B2 (fr)
WO (1) WO2016149570A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022018527A1 (fr) 2020-07-20 2022-01-27 CSEM Centre Suisse d'Electronique et de Microtechnique SA - Recherche et Développement Dispositif multi-spectral

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10987071B2 (en) * 2017-06-29 2021-04-27 University Of Delaware Pixelated K-edge coded aperture system for compressive spectral X-ray imaging
US10739196B2 (en) * 2018-07-03 2020-08-11 Electronics And Telecommunications Research Institute Spectroscopic apparatus and spectroscopic method using orthogonal code
US11137287B2 (en) * 2019-01-11 2021-10-05 The United States Of America As Represented By The Secretary Of The Army Compressive spectral imaging via polar coded aperture
US11592542B2 (en) * 2019-09-24 2023-02-28 Technion Research & Development Foundation Limited Coded aperture ultrasound detector
CN111929862B (zh) * 2020-07-09 2021-07-27 中国科学院西安光学精密机械研究所 一种共光路编码型快照式短波成像光学系统
JP2024519198A (ja) * 2021-05-13 2024-05-09 ヴァイアヴィ・ソリューションズ・インコーポレイテッド 光センサデバイス

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7888626B2 (en) * 2005-05-23 2011-02-15 Qinetiq Limited Coded aperture imaging system having adjustable imaging performance with a reconfigurable coded aperture mask
US20130234029A1 (en) * 2012-03-06 2013-09-12 Omnivision Technologies, Inc. Image sensor for two-dimensional and three-dimensional image capture
CN103472455A (zh) * 2013-09-13 2013-12-25 中国科学院空间科学与应用研究中心 稀疏孔径压缩计算关联飞行时间四维光谱成像系统及方法
CN103743482A (zh) * 2013-11-22 2014-04-23 中国科学院光电研究院 一种光谱成像装置及光谱成像反演方法
US20150028763A1 (en) * 2012-01-17 2015-01-29 Koninklijke Philips N.V. Modulation of light emitted by a lighting device, using plurality of different modulation periods

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090040516A1 (en) * 2007-08-10 2009-02-12 Honeywell International Inc. Spectroscopic system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7888626B2 (en) * 2005-05-23 2011-02-15 Qinetiq Limited Coded aperture imaging system having adjustable imaging performance with a reconfigurable coded aperture mask
US20150028763A1 (en) * 2012-01-17 2015-01-29 Koninklijke Philips N.V. Modulation of light emitted by a lighting device, using plurality of different modulation periods
US20130234029A1 (en) * 2012-03-06 2013-09-12 Omnivision Technologies, Inc. Image sensor for two-dimensional and three-dimensional image capture
CN103472455A (zh) * 2013-09-13 2013-12-25 中国科学院空间科学与应用研究中心 稀疏孔径压缩计算关联飞行时间四维光谱成像系统及方法
CN103743482A (zh) * 2013-11-22 2014-04-23 中国科学院光电研究院 一种光谱成像装置及光谱成像反演方法

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022018527A1 (fr) 2020-07-20 2022-01-27 CSEM Centre Suisse d'Electronique et de Microtechnique SA - Recherche et Développement Dispositif multi-spectral

Also Published As

Publication number Publication date
US10151629B2 (en) 2018-12-11
US20180080822A1 (en) 2018-03-22
WO2016149570A8 (fr) 2017-10-05

Similar Documents

Publication Publication Date Title
US10151629B2 (en) Spectral imaging sensors and methods with time of flight sensing
CN110325879B (zh) 用于压缩三维深度感测的系统和方法
CA2901100C (fr) Systeme et procede pour scanner une surface et programme d'ordinateur qui met en oeuvre le procede
JP7569334B2 (ja) 電子走査lidarシステム用同期画像捕捉
US9402067B2 (en) Imaging optical system for 3D image acquisition apparatus, and 3D image acquisition apparatus including the imaging optical system
KR20200024914A (ko) 전자적으로 스캔되는 방출기 어레이 및 동기화된 센서 어레이를 갖는 광 레인징 장치
CN113272624A (zh) 包括具有多个通带的带通滤波器的三维传感器
US20160265906A1 (en) Object detection device
US20200300977A1 (en) Time of flight-based three-dimensional sensing system
KR102184042B1 (ko) 카메라 장치
US20190041519A1 (en) Device and method of optical range imaging
US10982836B2 (en) Tunable spectral illuminator for camera
US20190058871A1 (en) Nanostructured optical element, depth sensor, and electronic device
KR20220012319A (ko) 필드 각도에 따라 조명 파장을 변화시키기 위한 능동 조명 시스템들
US20180313759A1 (en) Image acquisition device, image acquisition method, and spatial light modulation unit
EP4107546A1 (fr) Imagerie tridimensionnelle efficace de puissance sélective
US20200184662A1 (en) Structured light projector
JP2024019680A (ja) 走査装置及び光検出装置
US10931894B2 (en) Tunable spectral illuminator for camera
WO2022194006A1 (fr) Appareil de détection
US20240056669A1 (en) Camera module
EP4058762A1 (fr) Illuminateur spectral accordable pour caméra
RU2543688C2 (ru) Камера и оптическая система для получения 3d изображений (варианты)
US20240045068A1 (en) LiDAR SYSTEM AND RESOLUSION IMPROVEMENT METHOD THEREOF
WO2020022150A1 (fr) Dispositif de détection d'ondes électromagnétiques et système d'acquisition d'informations

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16765804

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 15559228

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16765804

Country of ref document: EP

Kind code of ref document: A1