US11190750B2 - Optical imaging system with a plurality of sense channels - Google Patents

Optical imaging system with a plurality of sense channels Download PDF

Info

Publication number
US11190750B2
US11190750B2 US16/046,643 US201816046643A US11190750B2 US 11190750 B2 US11190750 B2 US 11190750B2 US 201816046643 A US201816046643 A US 201816046643A US 11190750 B2 US11190750 B2 US 11190750B2
Authority
US
United States
Prior art keywords
aperture
pixel
pixels
imaging system
bulk
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/046,643
Other versions
US20180359460A1 (en
Inventor
Angus Pacala
Mark Frichtl
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ouster Inc
Sense Photonics Inc
Original Assignee
Ouster Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
US case filed in International Trade Commission litigation Critical https://portal.unifiedpatents.com/litigation/International%20Trade%20Commission/case/337-TA-1363 Source: International Trade Commission Jurisdiction: International Trade Commission "Unified Patents Litigation Data" by Unified Patents is licensed under a Creative Commons Attribution 4.0 International License.
US case filed in Delaware District Court litigation https://portal.unifiedpatents.com/litigation/Delaware%20District%20Court/case/1%3A23-cv-00406 Source: District Court Jurisdiction: Delaware District Court "Unified Patents Litigation Data" by Unified Patents is licensed under a Creative Commons Attribution 4.0 International License.
First worldwide family litigation filed litigation https://patents.darts-ip.com/?family=59960008&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US11190750(B2) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
PTAB case IPR2023-01375 filed (Procedural Termination) litigation https://portal.unifiedpatents.com/ptab/case/IPR2023-01375 Petitioner: "Unified Patents PTAB Data" by Unified Patents is licensed under a Creative Commons Attribution 4.0 International License.
PTAB case IPR2023-01457 filed (Pending) litigation https://portal.unifiedpatents.com/ptab/case/IPR2023-01457 Petitioner: "Unified Patents PTAB Data" by Unified Patents is licensed under a Creative Commons Attribution 4.0 International License.
US case filed in International Trade Commission litigation https://portal.unifiedpatents.com/litigation/International%20Trade%20Commission/case/337-TA-3675 Source: International Trade Commission Jurisdiction: International Trade Commission "Unified Patents Litigation Data" by Unified Patents is licensed under a Creative Commons Attribution 4.0 International License.
Priority to US16/046,643 priority Critical patent/US11190750B2/en
Application filed by Ouster Inc filed Critical Ouster Inc
Assigned to Ouster, Inc. reassignment Ouster, Inc. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FRICHTL, Mark, PACALA, ANGUS
Publication of US20180359460A1 publication Critical patent/US20180359460A1/en
Priority to US16/584,515 priority patent/US11202056B2/en
Assigned to SILICON VALLEY BANK reassignment SILICON VALLEY BANK SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Ouster, Inc.
Publication of US11190750B2 publication Critical patent/US11190750B2/en
Application granted granted Critical
Priority to US17/549,675 priority patent/US11627298B2/en
Assigned to Ouster, Inc. reassignment Ouster, Inc. RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: SILICON VALLEY BANK
Assigned to HERCULES CAPITAL, INC., AS AGENT reassignment HERCULES CAPITAL, INC., AS AGENT SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Ouster, Inc., Sense Photonics, Inc.
Priority to US18/179,559 priority patent/US20230319253A1/en
Assigned to Ouster, Inc. reassignment Ouster, Inc. RELEASE OF INTELLECTUAL PROPERTY SECURITY INTEREST AT REEL/FRAME NO. 059859/0035 Assignors: HERCULES CAPITAL, INC.
Assigned to Sense Photonics, Inc., Ouster, Inc. reassignment Sense Photonics, Inc. CORRECTIVE ASSIGNMENT TO CORRECT THE ADD THE SECOND ASSIGNEE PREVIOUSLY RECORDED AT REEL: 65350 FRAME: 826. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT. Assignors: HERCULES CAPITAL, INC.
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J1/00Photometry, e.g. photographic exposure meter
    • G01J1/02Details
    • G01J1/029Multi-channel photometry
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J1/00Photometry, e.g. photographic exposure meter
    • G01J1/02Details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/218Image signal generators using stereoscopic image cameras using a single 2D image sensor using spatial multiplexing
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J1/00Photometry, e.g. photographic exposure meter
    • G01J1/02Details
    • G01J1/04Optical or mechanical part supplementary adjustable parts
    • G01J1/0407Optical elements not provided otherwise, e.g. manifolds, windows, holograms, gratings
    • G01J1/0411Optical elements not provided otherwise, e.g. manifolds, windows, holograms, gratings using focussing or collimating elements, i.e. lenses or mirrors; Aberration correction
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J1/00Photometry, e.g. photographic exposure meter
    • G01J1/02Details
    • G01J1/04Optical or mechanical part supplementary adjustable parts
    • G01J1/0407Optical elements not provided otherwise, e.g. manifolds, windows, holograms, gratings
    • G01J1/0437Optical elements not provided otherwise, e.g. manifolds, windows, holograms, gratings using masks, aperture plates, spatial light modulators, spatial filters, e.g. reflective filters
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J1/00Photometry, e.g. photographic exposure meter
    • G01J1/02Details
    • G01J1/08Arrangements of light sources specially adapted for photometry standard sources, also using luminescent or radioactive material
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/481Constructional features, e.g. arrangements of optical elements
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/481Constructional features, e.g. arrangements of optical elements
    • G01S7/4814Constructional features, e.g. arrangements of optical elements of transmitters alone
    • G01S7/4815Constructional features, e.g. arrangements of optical elements of transmitters alone using multiple transmitters
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/481Constructional features, e.g. arrangements of optical elements
    • G01S7/4816Constructional features, e.g. arrangements of optical elements of receivers alone
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/42Diffraction optics, i.e. systems including a diffractive element being designed for providing a diffractive effect
    • G02B27/46Systems using spatial filters
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B3/00Simple or compound lenses
    • G02B3/0006Arrays
    • G02B3/0037Arrays characterized by the distribution or form of lenses
    • G02B3/0056Arrays characterized by the distribution or form of lenses arranged along two different directions in a plane, e.g. honeycomb arrangement of lenses
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B30/00Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
    • G02B30/20Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
    • G02B30/26Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type
    • G02B30/27Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type involving lenticular arrays
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B5/00Optical elements other than lenses
    • G02B5/005Diaphragms
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B5/00Optical elements other than lenses
    • G02B5/20Filters
    • G02B5/208Filters for use with infrared or ultraviolet radiation, e.g. for separating visible light from infrared and/or ultraviolet radiation
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B6/00Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B6/00Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
    • G02B6/0001Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings specially adapted for lighting devices or systems
    • G02B6/0011Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings specially adapted for lighting devices or systems the light guides being planar or of plate-like form
    • G02B6/0013Means for improving the coupling-in of light from the light source into the light guide
    • G02B6/0015Means for improving the coupling-in of light from the light source into the light guide provided on the surface of the light guide or in the bulk of it
    • G02B6/002Means for improving the coupling-in of light from the light source into the light guide provided on the surface of the light guide or in the bulk of it by shaping at least a portion of the light guide, e.g. with collimating, focussing or diverging surfaces
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/14625Optical elements or arrangements associated with the device

Definitions

  • This invention relates generally to the field of optical sensors and more specifically to a new and useful optical system for collecting distance information in the field of optical sensors.
  • FIG. 1 is a schematic representation of a system.
  • FIG. 2 is a schematic representation in accordance with one variation of the system.
  • FIG. 3 is a schematic representation in accordance with one variation of the system.
  • FIG. 4 is a schematic representation in accordance with one variation of the system.
  • FIG. 5 is a schematic representation in accordance with one variation of the system.
  • FIG. 6 is a schematic representation in accordance with one variation of the system.
  • FIG. 7 is a schematic representation in accordance with one variation of the system.
  • FIG. 8 is a schematic representation in accordance with one variation of the system.
  • FIG. 9 is a flowchart representation in accordance with one variation of the system.
  • FIG. 10 is a schematic representation in accordance with one variation of the system.
  • FIG. 11 is a schematic representation in accordance with one variation of the system.
  • a one-dimensional optical system 100 for collecting distance information within a field includes: a set of illumination sources 110 arranged along a first axis, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in the field ahead of the illumination source; a bulk imaging optic 130 characterized by a focal plane opposite the field; an aperture layer 140 coincident the focal plane, defining a set of apertures 144 in a line array parallel to the first axis, and defining a stop region 146 around the set of apertures 144 , each aperture in the set of apertures 144 defining a field of view in the field coincident a discrete spot output by a corresponding illumination source in the set of illumination sources 110 , the stop region 146 absorbing light rays reflected from surfaces in the field outside of fields of view defined by the set of apertures 144 and passing through the bulk imaging optic 130 ; a set of lenses 150 , each lens in the set of lenses 150 , each lens in the set of
  • the one-dimensional optical system 100 functions as an image sensor that, when rotated about an axis parallel to a column of apertures, collects three-dimensional distance data of a volume occupied by the system.
  • the one-dimensional optical system 100 can scan a volume to collect three-dimensional distance data that can then be reconstructed into a virtual three-dimensional representation of the volume, such as based on recorded times between transmission of illuminating beams from the illumination sources and detection of photons—likely originating from the illumination sources—incident on the set of pixels 170 , based on phase-based measurements techniques, or based on any other suitable distance measurement technique.
  • the system 100 includes: a column of offset apertures arranged behind a bulk imaging optic 130 and defining discrete fields of view in a field ahead of the bulk imaging optic 130 (that is non-overlapping fields of view beyond a threshold distance from the system); a set of illumination sources 110 that project discrete illuminating beams at an operating wavelength into (and substantially only into) the fields of view defined by the apertures; a column of lenses that collimate light rays passed by corresponding apertures; and an optical filter 160 that selectively passes a narrow band of wavelengths of light (i.e., electromagnetic radiation) including the operating wavelength; and a set of pixels 170 that detect incident photons (e.g., count incident photons, tracks times between consecutive incident photons).
  • incident photons e.g., count incident photons, tracks times between consecutive incident photons.
  • the system can therefore selectively project illuminating beams into a field ahead of the system according to an illumination pattern that substantially matches—in size and geometry across a range of distances from the system—the fields of view of the apertures.
  • the illumination sources are configured to illuminate substantially only surfaces in the field ahead of the system that can be detected by pixels in the system such that minimal power output by the system (via the illumination sources) is wasted by illuminating surfaces in the field for which the pixels are blind.
  • the system can therefore achieve a relatively high ratio of output signal (i.e., illuminating beam power) to input signal (i.e., photons passed to an incident on the pixel array).
  • the set of lenses 150 can collimate light rays passed by adjacent apertures such that light rays incident on the optical filter 160 meet the optical filter 160 at an angle of incidence of approximately 0°, thereby maintaining a relatively narrow band of wavelengths of light passed by the optical filter 160 and achieving a relatively high signal-to-noise ratio (“SNR”) for light rays reaching the set of pixels 170 .
  • SNR signal-to-noise ratio
  • the system includes pixels arranged in a column and aligned with the apertures, and each pixel can be non-square in geometry (e.g., short and wide) to extend the sensing area of the system for a fixed aperture pitch and pixel column height.
  • the system also includes a diffuser 180 that spreads light rays passed from an aperture through the optical filter 160 across the area of a corresponding pixel such that the pixel can detect incident photons across its full width and height thereby increasing the dynamic range of the system.
  • the system is described herein as projecting electromagnetic radiation into a field and detecting electromagnetic radiation reflected from a surface in the field back to bulk receiver optic.
  • Terms “illumination beam,” “light,” “light rays,” and “photons” recited herein refer to such electromagnetic radiation.
  • the term “channel” recited herein refers to one aperture in the aperture layer 140 , a corresponding lens in the set of lenses 150 , and a corresponding pixel in the set of pixels 170 .
  • the system includes a bulk imaging optic 130 characterized by a focal plane opposite the field.
  • the bulk imaging optic 130 functions to project incident light rays from outside the system toward the focal plane where light rays incident on a stop region 146 of the aperture layer 140 are rejected (e.g., mirrored or absorbed) and where light rays incident on apertures in the aperture layer 140 are passed into a lens characterized by a focal length and offset from the focal plane by the focal length.
  • the bulk imaging optic 130 includes a converging lens, such as a bi-convex lens (shown in FIG. 2 ) or a plano-convex lens, characterized by a particular focal length at the operating wavelength of the system.
  • the bulk imaging optic 130 can also include multiple discrete lens that cooperate to project light rays toward the aperture layer 140 and that are characterized by a composite focal plane opposite the field, as shown in FIG. 11 .
  • the bulk imaging optic 130 can be any other suitable type of lens or combination of lenses of any other type or geometry.
  • the system includes an aperture layer 140 coincident the focal plane, defining a set of apertures 144 in a line array parallel to the axes of the illumination sources, and defining a stop region 146 around the set of apertures 144 , wherein each aperture in the set of apertures 144 defines a field of view in the field coincident a discrete spot output by a corresponding illumination source in the set of illumination sources 110 , and wherein the stop region 146 absorbs and/or reflects light rays reflected from surfaces in the field outside of fields of view defined by the set of apertures 144 and passing through the bulk imaging optic 130 .
  • the aperture layer 140 defines an array of open regions (i.e., apertures, including one aperture per lens) and closed regions (“stop regions”) between adjacent opens.
  • Each aperture in the aperture layer 140 defines a “pinhole” that defines a field of view for its corresponding sense channel and passes light rights reflected from an external surface within its field of the view into its corresponding lens, and each stop region 146 can block light rays incident on select regions of the focal plane from passing into the lens array, as shown in FIG. 6 .
  • the aperture layer 140 includes a relatively thin opaque structure coinciding with (e.g., arranged along) the focal plane of the bulk imaging optic 130 , as shown in FIGS. 1 and 2 .
  • the aperture layer 140 can include a 10 micrometer-thick copper, silver, or nickel film deposited (e.g., plated) over a photocurable transparent polymer and then selectively etched to form the array of apertures.
  • a reflective metalized layer or a light-absorbing photopolymer e.g., a photopolymer mixed with a light absorbing dye
  • a photomask can be deposited onto a glass wafer and selectively cured with a photomask to form the aperture layer 140 and the set of apertures 144 .
  • the aperture layer 140 can include a discrete metallic film that is mechanically or chemically perforated to form the array of apertures, bonded to the lens array, and then installed over the bulk imaging optic 130 along the focal plane.
  • the aperture layer 140 can include any other reflective (e.g., mirrored) or light-absorbing material formed in any other way to define the array of apertures along the focal plane of the bulk imaging optic 130 .
  • the aperture layer 140 can define a single column of multiple discrete circular apertures of substantially uniform diameter, wherein each aperture defines an axis substantially parallel to and aligned with one lens in the lens array, as shown in FIG. 3 . Adjacent apertures are offset by an aperture pitch distance greater than the aperture diameter and substantially similar to the lens pitch distance, and the aperture layer 140 defines a stop region 146 (i.e., an opaque or reflecting region) between adjacent apertures such that the apertures define discrete, non-overlapping fields of view for their corresponding sense channels.
  • a stop region 146 i.e., an opaque or reflecting region
  • an aperture defines a narrower field of view (i.e., a field of view of smaller diameter) and passes a sharper but lower-intensity (attenuated) signal from the bulk imaging optic 130 into its corresponding lens.
  • the aperture layer 140 can therefore define apertures of diameter: greater than the diffraction-limited diameter for the wavelength of light output by the illumination sources (e.g., 900 nm); substantially greater than the thickness of the aperture layer 140 ; and less than the aperture pitch distance, which is substantially equivalent to the lens pitch distance and the pixel pitch distance.
  • aperture layer 140 can define apertures of diameters approaching the diffraction-limited diameter to maximize geometrical selectivity of the field of view of each sense channel.
  • the apertures can be of diameter less that the diffraction-limited diameter for the wavelength of light output by the illumination sources.
  • the aperture layer 140 can define apertures of diameters matched to a power output of illumination sources in the system and to a number and photon detection capacity of subpixel photodetectors in each pixel in the set of pixels 170 to achieve a target number of photons incident on each pixel within each sampling period.
  • each aperture can define a particular diameter that achieves target attenuation range for pixels originating from a corresponding illumination source and incident on the bulk imaging optic 130 during a sampling period.
  • the diameter of the aperture can be matched to the dynamic range of its corresponding pixel.
  • a first aperture 141 in the aperture layer 140 passes light rays—reflected from a discrete region of a surface in the field (the field of view of the sense channel) ahead of the bulk imaging optic 130 —into its corresponding lens; a stop region 146 interposed between the first aperture 141 and adjacent apertures in the aperture layer 140 blocks light rays—reflected from a region of the surface outside of the field of view of the first aperture 141 —from passing into the lens corresponding to the first aperture 141 .
  • the aperture layer 140 therefore defines a column of apertures that define multiple discrete, non-overlapping fields of view of substantially infinite depth of field, as shown in FIG. 2 .
  • a first aperture 141 in the aperture layer 140 defines a field of view that is distinct and that does not intersect a field of view defined by another aperture in the aperture layer 140 , as shown in FIG. 2 .
  • the set of illumination sources 110 includes a first illumination source 111 paired with the first aperture 141 and configured to project an illuminating beam substantially aligned with (i.e., overlapping) the field of view of the first aperture 141 in the field ahead of the bulk imaging optic 130 .
  • the first illumination source 111 and a bulk transmitting optic 120 can cooperate to project an illuminating beam of a cross-section substantially similar to (and slightly larger than) the cross section of the field of view of the first aperture 141 as various distances from the bulk imaging optic 130 . Therefore light output by the first illumination source 111 —paired with the first aperture 141 —and projected into the field of view of the first aperture 141 can remain substantially outside the fields of view of other apertures in the aperture layer 140 .
  • photons projected into the field by the first illumination source 111 illuminate a particular region of a surface (or multiple surfaces) in the field within the field of view of the first sense channel and are reflected (e.g., scattered) by the surface(s); at least some of these photons reflected by the particular region of a surface may reach the bulk imaging optic 130 , which directs these photons toward the focal plane.
  • the bulk imaging optic 130 may project these photons into the first aperture 141 , and the first aperture 141 may pass these photons into the first lens 151 (or a subset of these photons incident at an angle relative to the axis of the first aperture 141 below a threshold angle).
  • a second aperture 142 in the aperture layer 140 is offset from the first aperture 141 and because the particular region of the surface in the field illuminated via the first illumination source 111 does not (substantially) coincide with the field of view of the second aperture 142 , photons reflected by the particular region of the surface and reaching the bulk imaging optic 130 are projected into the second aperture 142 and passed to a second lens 152 behind the second aperture 142 , and vice versa, as shown in FIG. 2 .
  • a stop region 146 between the first and second apertures 142 can block photons directed toward the focal plane between the first and second apertures 142 reflected by the bulk imaging optic 130 , thereby reducing crosstalk between the first and second sense channels.
  • the first aperture 141 in the aperture layer 140 For a first aperture 141 in the aperture layer 140 paired with a first illumination source 111 in the set of illumination sources 110 , the first aperture 141 in the aperture layer 140 defines a first field of view and passes—into the first lens 151 —incident light rays originating at or reflected from a surface in the field coinciding with the first field of view. Because the first illumination source 111 projects an illuminating beam that is substantially coincident (and substantially the same size as or minimally larger than) the field of view defined by the first aperture 141 (as shown in FIG.
  • a signal passed into the first lens 151 by the first aperture 141 in the aperture layer 140 can exhibit a relatively high ratio of light rays originating from the first illumination source 111 to light rays originating from other illumination sources in the system.
  • various illumination sources in the system may output illuminating beams at different frequencies, duty cycles, and/or power levels, etc. at a particular time during operation, light rays passed from the bulk imaging optic 130 into a first pixel 171 in the set of pixels 170 but originating from an illumination source other than the first illumination source 111 paired with the first pixel 171 constitute noise at the first pixel 171 .
  • each aperture in the aperture layer 140 may pass a relatively high proportion of photons originating from its corresponding illumination source than from other illumination sources in the system; that is, due to the geometry of a particular aperture and its corresponding illumination source, a particular aperture may pass a signal exhibiting a relatively high SNR to its corresponding lens and thus into its corresponding pixel.
  • the system can pass less noise from solar radiation or other ambient light sources to the set of pixels 170 .
  • the system includes a second aperture layer interposed between the lens array and the optical filter 160 , wherein the second aperture layer defines a second set of apertures 144 , each aligned with a corresponding lens in the set of lenses 150 , as described above.
  • an aperture in the second aperture layer 140 can absorb or reflect errant light rays passed by a corresponding lens, as described above, to further reduce crosstalk between channels, thereby improving SNR within the system.
  • the system can additionally or alternatively include a third aperture layer interposed between the optical filter 160 and the diffuser(s) 180 , wherein the third aperture layer defines a third set of apertures 144 , each aligned with a corresponding lens in the set of lenses 150 , as described above.
  • an aperture in the third aperture layer can absorb or reflect errant light rays passed by the light filter, as described above, to again reduce crosstalk between channels, thereby improving SNR within the system.
  • the system includes a set of lenses 150 , wherein each lens in the set of lenses 150 is characterized by a second focal length, is offset from the focal plane opposite the bulk imaging optic 130 by the second focal length, is aligned with a corresponding aperture in the set of apertures 144 , and is configured to collimate light rays passed by the corresponding aperture.
  • a lens in the set of lenses 150 functions to collimate lights rays passed by its corresponding aperture and to pass these collimated light rays into the optical filter 160 .
  • each lens can include a converging lens characterized by a second focal length and can be offset from the focal plane of the bulk imaging optic 130 —opposite the bulk imaging optic 130 —by the second focal length to preserve the aperture of the bulk imaging optic 130 and to collimate light incident on the bulk imaging optic 130 and passed by a corresponding aperture.
  • Each lens in the set of lens can be characterized by a relatively short focal length (i.e., less than a focal length of the bulk imaging optic 130 ) and a relatively large marginal ray angle (e.g., a relatively high numeric aperture lens) such that the lens can capture highly-angled light rays projected toward the lens by the extent of the bulk imaging optic 130 . That is, each lens in the set of lens can be characterized by a ray cone substantially matched to a ray cone of the bulk imaging optic 130 .
  • Lenses in the set of lenses 150 can be substantially similar.
  • a lens in the set of lenses 150 is configured to collimate light rays focused into its corresponding aperture by the bulk imaging optic 130 .
  • a lens in the set of lenses 150 can include a bi-convex or plano-convex lens characterized by a focal length selected based on the size (e.g., diameter) of its corresponding aperture and the operating wavelength of the system.
  • the focal length (f) of a lens in the set of lenses 150 can be calculated according to the formula:
  • d the diameter of the corresponding aperture in the aperture layer and ⁇ is the operating wavelength of light output by the illumination source (e.g., 900 nm).
  • the geometry of a lens in the set of lenses 150 can therefore be matched to the geometry of a corresponding aperture in the aperture layer such that the lens passes a substantially sharp image of light rays—at or near the operating wavelength—into the optical filter 160 and thus on to the pixel array.
  • the set of lenses 150 can include lenses of any other geometry and arranged in any other way adjacent the aperture layer.
  • the system includes an optical filter 160 adjacent the set of lenses 150 opposite the aperture layer and configured to pass light rays at the operating wavelength.
  • the optical filter 160 receives electromagnetic radiation across a spectrum from the set of lenses 150 , passes a relatively narrow band of electromagnetic radiation—including radiation at the operating wavelength—to the pixel array, and blocks electromagnetic radiation outside of the band.
  • electromagnetic radiation other than electromagnetic radiation output by the illumination source—such as ambient light—incident on a pixel in the set of pixels 170 constitutes noise in the system.
  • the optical filter 160 therefore functions to reject electromagnetic radiation outside of the operating wavelength or, more pragmatically, outside of a narrow wavelength band, thereby reducing noise in the system and increasing SNR.
  • the optical filter 160 includes an optical bandpass filter that passes a narrow band of electromagnetic radiation substantially centered at the operating wavelength of the system.
  • the illumination sources output light (predominantly) at an operating wavelength of 900 nm, and the optical filter 160 is configured to pass light between 899.95 nm and 900.05 nm and to block light outside of this band.
  • the optical filter 160 may selectively pass and reject wavelengths of light as a function of angle of incidence on the optical filter 160 .
  • optical bandpass filters may pass wavelengths of light inversely proportional to their angle of incidence on the light optical bandpass filter.
  • the optical filter 160 may pass over 95% of electromagnetic radiation over a sharp band from 899.75 nm to 900.25 nm and reject approximately 100% of electromagnetic radiation below 899.70 nm and above 900.30 nm for light rays incident on the optical filter 160 at an angle of incidence of approximately 0°.
  • the optical filter 160 may pass over 95% of electromagnetic radiation over a narrow band from 899.5 nm to 900.00 nm and reject approximately 100% of electromagnetic radiation over a much wider band below 899.50 nm and above 900.30 nm for light rays incident on the optical filter 160 at an angle of incidence of approximately 15°. Therefore, the incidence plane of the optical filter 160 can be substantially normal to the axes of the lenses, and the set of lenses 150 can collimate light rays received through a corresponding aperture and output these light rays substantially normal to the incidence plane of the optical filter 160 (i.e., at an angle of incidence of approximately 0° on the optical filter). Specifically, the set of lenses 150 can output light rays toward the optical filter 160 at angles of incidence approximating 0° such that substantially all electromagnetic radiation passed by the optical filter 160 is at or very near the operating wavelength of the system.
  • the system can include a single optical filter 160 that spans the column of lens in the set of lenses 150 .
  • the system can include multiple optical filters 160 , each adjacent a single lens or a subset of lenses in the set of lenses 150 .
  • the optical filter 160 can define any other geometry and can function in any other way to pass only a limited band of wavelengths of light.
  • the system includes a set of pixels 170 adjacent the optical filter 160 opposite the set of lenses 150 , each pixel in the set of pixels 170 corresponding to a lens in the set of lenses 150 and including a set of subpixels arranged along a second axis non-parallel to the first axis.
  • the set of pixels 170 are offset from the optical filter 160 opposite the set of lenses 150 , and each pixel in the set of pixels 170 functions to output a single signal or stream of signals corresponding to the count of photons incident on the pixel within one or more sampling periods, wherein each sampling period may be picoseconds, nanoseconds, microseconds, or milliseconds in duration.
  • the system also includes a diffuser 180 interposed between the optical filter 160 and the set of pixels 170 and configured to spread collimated light output from each lens in the set of lenses 150 across a set of subpixels of a single corresponding pixel in the set of pixels 170 .
  • the diffuser 180 functions to spread light rays—previously collimated by the lens and passed by the optical filter 160 —across the width and height of a sensing area within a corresponding pixel.
  • the diffuser 180 can define a single optic element spanning the set of lenses 150 , or the diffuser 180 can include multiple discrete optical elements, such as including one optical diffuser element aligned with each channel in the system.
  • a first pixel 171 in the set of pixels 170 includes an array of single-photon avalanche diode detectors (hereinafter “SPADs”), and the diffuser 180 spreads lights rays—previously passed by a corresponding first aperture 141 , collimated by a corresponding first lens 151 , and passed by the optical filter 160 —across the area of the first pixel 171 , as shown in FIGS. 3, 5, and 6 .
  • SPADs single-photon avalanche diode detectors
  • adjacent apertures can be aligned and offset vertically by an aperture pitch distance
  • adjacent lenses can be aligned and offset vertically by a lens pitch distance substantially identical to the aperture pitch distance
  • adjacent pixels can be aligned and offset vertically by a pixel pitch distance substantially identical to the lens and aperture pitch distances.
  • the pixel pitch distance may accommodate only a relatively small number of (e.g., two) vertically-stacked SPADs.
  • Each pixel in the set of pixels 170 can therefore define an aspect ratio greater than 1:1, and the diffuser 180 can spread light rays passed by the optical filter 160 according to the geometry of a corresponding pixel in order to accommodate a larger sensing area per pixel.
  • each pixel in the set of pixels 170 is arranged on an image sensor, and a first pixel 171 in the set of pixels 170 includes a single row of 16 SPADs spaced along a lateral axis perpendicular to a vertical axis bisecting the column of apertures and lenses.
  • the height of a single SPAD in the first pixel 171 can be less than the height (e.g., diameter) of the first lens 151 , but the total length of the 16 SPADs can be greater than the width (e.g., diameter) of the first lens 151 ; the diffuser 180 can therefore converge light rays output from the first lens 151 to a height corresponding to the height of a SPAD at the plane of the first pixel 171 and can diverge light rays output from the first lens 151 to a width corresponding to the width of the 16 SPADs at the plane of the first pixel 171 .
  • the remaining pixels in the set of pixels 170 can include similar rows of SPADs, and the diffuser 180 can similarly converge and diverge light rays passed by corresponding apertures onto corresponding pixels.
  • the aperture layer can include a column of 16 like apertures
  • the set of lenses 150 can include a column of 16 like lenses arranged behind the aperture layer
  • the set of pixels 170 can include a set of 16 like pixels—each including a similar array of SPADs—arranged behind the set of lenses 150 .
  • each pixel can include a single row of 16 SPADs, wherein each SPAD is electrically coupled to a remote analog front-end processing electronics/digital processing electronics circuit 240 .
  • Each SPAD can be arranged in a 400 ⁇ m-wide, 400 ⁇ m-tall SPAD area and can define an active sensing area approaching 400 ⁇ m in diameter.
  • Adjacent SPADs can be offset by a SPAD pitch distance of 400 ⁇ m.
  • the aperture pitch distance along the vertical column of apertures, the lens pitch distance along the vertical column of lenses, and the pixel pitch distance along the vertical column of pixels can each be approximately 400 ⁇ m accordingly.
  • a first diffuser 180 can diverge a cylindrical column of light rays passed from the first lens 151 through the optical filter 160 —such as a column of light approximately 100 ⁇ m in diameter for an aperture layer aspect ratio of 1:4—to a height of approximately 400 ⁇ m aligned vertically with the row of SPADs in the first pixel 171 .
  • the first diffuser can similarly diverge the cylindrical column of light rays passed from the first lens 151 through the optical filter 160 to a width of approximately 6.4 ⁇ m centered horizontally across the row of SPADs in the first pixel 171 .
  • diffusers 180 in the system can similarly diverge (or converge) collimated light passed by corresponding lenses across corresponding pixels in the set of pixels 170 . Therefore, in this example, by connecting each SPAD (or each pixel) to a remote analog front-end processing electronics/digital processing electronics circuit 240 and by incorporating diffusers 180 that spread light passed by the optical filter 160 across the breadths and heights of corresponding pixels, the system can achieve a relatively high sensing area fill factor across the imaging sensor.
  • pixels in the set of pixels 170 can include an array of multiple SPADS arranged in aspect ratio exceeding 1:1, and the diffuser 180 can spread light rays across corresponding non-square pixels that enables a relatively large numbers of SPADs to be tiled across a single pixel to achieve a greater dynamic range across the image sensor than an image sensor with a single SPAD per pixel, as shown in FIG. 3 .
  • a first sense channel in the system can detect multiple incident photons—originating from a surface in the field bound by a field of view defined by the first aperture 141 —within the span of the dead time characteristic of the SPADs.
  • the first sense channel can therefore detect a “brighter” surface in its field of view. Additionally or alternatively, the first pixel 171 in the first sense channel can be sampled faster than the dead time characteristic of SPADs in the first pixel 171 because, though a first subset of SPADs in the first pixel 171 may be down (or “dead”) during a first sampling period due to collection of incident photons during the first sampling period, other SPADs in the first pixel 171 remain on (or “alive”) and can therefore collect incident photons during a subsequent sampling period. Furthermore, by incorporating pixels characterized by relatively high aspect ratios of photodetectors, the image sensor can include pixels offset by a relatively small pixel pitch, but the system 100 can still achieve a relatively high dynamic range pixel.
  • pixels in the set of pixels 170 can include any other number of SPADs arranged in any other arrays, such as in a 64-by-1 grid array (as described above), in a 32-by-2 grid array, or in a 16-by-4 grid array, and the diffuser 180 can converge and/or diverge collimated light rays onto corresponding pixels accordingly in any other suitable way.
  • each pixel in the set of pixels 170 can include one or more linear avalanche photodiodes, Geiger mode avalanche photodiodes, photomultipliers, resonant cavity photodiodes, QUANTUM DOT detectors, or other types of photodetectors arranged as described above, and the diffuser(s) 180 can similarly converge and diverge signals passed by the optical filter(s) 160 across corresponding pixels, as described herein.
  • the system includes a set of illumination sources 110 arranged along a first axis, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in a field ahead of the illumination source.
  • each illumination source functions to output an illuminating beam coincident a field of view defined by a corresponding aperture in the set of apertures 144 , as shown in FIGS. 1 and 2 .
  • the set of illumination sources 110 includes a bulk transmitter optic and one discrete emitter per sense channel.
  • the set of illumination sources 110 can include a monolithic VCSEL arrays including a set of discrete emitters.
  • the bulk transmitter optic can be substantially identical to the bulk imaging optic 130 in material, geometry (e.g., focal length), thermal isolation, etc., and the bulk transmitter optic is adjacent and offset laterally and/or vertically from the bulk imaging optic 130 .
  • set of illumination sources 110 includes a laser array including discrete emitters arranged in a column with adjacent emitters offset by an emitter pitch distance substantially identical to the aperture pitch distance.
  • each emitter outputs an illuminating beam of diameter substantially identical to or slightly greater than the diameter of a corresponding aperture in the apertures layer
  • the column of emitters is arranged along the focal plane of the bulk transmitter optic such that each illuminating beam projected from the bulk transmitter optic into the field intersects and is of substantially the same size and geometry as the field of view of the corresponding sense channel, as shown in FIG. 4 . Therefore, substantially all power output by each emitter in the set of illumination sources 110 can be projected into the field of view of its corresponding sense channel with relatively minimal power wasted illuminating surfaces in the field outside of the fields of view of the sense channels.
  • the discrete emitters are similarly arranged in a column with adjacent emitters offset by an emitter pitch distance twice the aperture pitch distance, as shown in FIG. 2 .
  • each emitter is characterized by an illuminating active area (or aperture) of diameter approximately (or slightly greater than) twice the diameter of a corresponding aperture in the apertures layer, and the column of emitters is offset behind the bulk transmitter optic by twice the focal length of the bulk transmitter optic such that each illuminating beam projected from the bulk transmitter optic into the field intersects and is of substantially the same size and geometry as the field of view of the corresponding sense channel, as described above.
  • an illuminating beam output by an emitter in this second example may contain four times the power of an illuminating beam output by an emitter in the first example described above.
  • the system can therefore include a set of emitter arranged according to an emitter pitch distance, configured to output illuminating beams of diameter, and offset behind the bulk transmitter optic by an offset distance as a function of a scale factor (e.g., 2.0 or 3.0) and 1) the aperture pitch distance in the aperture layer, 2) the diameter of apertures in the aperture layer, and 3) the focal length of bulk transmitter optic, respectively.
  • the system can therefore include an illuminating subsystem that is proportionally larger than a corresponding receiver subsystem to achieve greater total output illumination power within the same beam angles and fields of view of corresponding channels in the receiver subsystem.
  • the system can also include multiple discrete sets of illumination sources, each set of illumination sources 110 paired with a discrete bulk transmitter optic adjacent the bulk imaging optic 130 .
  • the system can include a first bulk transmitter optic, a second bulk transmitter optic, and a third bulk transmitter optic patterned radially about the bulk imaging optic 130 at a uniform radial distance from the center of the bulk imaging optic 130 and spaced apart by an angular distance of 120°.
  • the system can include a laser array with one emitter—as described above—behind each of the first, second, and third bulk transmitter optics. Each discrete laser array and its corresponding bulk transmitter optic can thus project a set of illuminating beams into the fields of view of defined by corresponding in the apertures in the aperture layer.
  • the three discrete laser arrays and the three corresponding bulk transmitter optics can cooperate to project three times the power onto the fields of view of the sense channels in the system, as compared to a single laser array and one bulk transmitter optic.
  • the system can include multiple discrete layer arrays and bulk transmitter optics to both: 1) achieve a target illumination power output into the field of view of each sensing channel in the receiver subsystem with multiple lower-power emitters per sensing channel; and 2) distribute optical energy over a larger area in the near-field to achieve an optical energy density less than a threshold allowable optical energy density for the human eye.
  • the system can include any other number and configuration of illumination source sets and bulk transmitter optics configured to illuminate fields of view defined by the sense channels.
  • the set of illumination sources 110 can also include any other suitable type of optical transmitter, such as a 1 ⁇ 16 optical splitter powered by a single laser diode, a side-emitting laser diode array, an LED array, or a quantum dot LED array, etc.
  • the bulk receiver lens, the aperture layer, the set of lenses 150 , the optical filter 160 , and the diffuser 180 are fabricated and then aligned with and mounted onto an image sensor.
  • the optical filter 160 can be fabricated by coating a fused silica substrate. Photoactive optical polymer can then be deposited over the optical filter 160 , and a lens mold can be placed over the photoactive optical polymer and a UV light source activated to cure the photoactive optical polymer in the form of lenses patterned across the optical filter 160 .
  • Standoffs can be similarly molded or formed across the optical filter 160 via photolithography techniques, and an aperture layer defined by a selectively-cured, metallized glass wafer can then be bonded or otherwise mounted to the standoffs to form the aperture layer.
  • the assembly can then be inverted, and a set of discrete diffusers and standoffs can be similarly fabricated across the opposite side of the optical filter 160 .
  • a discrete image sensor can then be aligned with and bonded to the standoffs, and a bulk imaging optic 130 can be similarly mounted over the aperture layer.
  • photolithography and wafer level bonding techniques can be implemented to fabricate the bulk imaging optics, the aperture layer, the set of lenses 150 , the optical filter 160 , and the diffuser 180 directly on to the un-diced semiconductor wafer containing the detector chips in order to simplify manufacturing, reduce cost, and reduce optical stack height for decreased pixel crosstalk.
  • One variation of the system includes: a set of illumination sources 110 arranged along a first axis, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in a field ahead of the illumination source; a bulk imaging optic 130 characterized by a focal plane opposite the field; a set of lens tubes 210 arranged in a line array parallel to the first axis, each lens tube in the set of lens tubes 210 including: a lens characterized by a focal length, offset from the focal plane by the focal length, and configured to collimate light rays reflected into the bulk imaging optic 130 from a discrete spot in the field illuminated by a corresponding illumination source in the set of optics into the bulk imaging optic 130 ; and a cylindrical wall 218 extending from the lens opposite the focal plane, defining a long axis substantially perpendicular to the first axis, and configured to absorb incident light rays reflected into the bulk imaging optic 130 from a region in the field outside the discrete spot illuminate
  • the system also includes: an optical filter 160 adjacent the set of lens tubes 210 opposite the focal plane and configured to pass light rays at the operating wavelength; a set of pixels 170 adjacent the optical filter 160 opposite the set of lenses 150 , each pixel in the set of pixels 170 corresponding to a lens in the set of lenses 150 and including a set of subpixels aligned along a third axis perpendicular to the first axis; and a diffuser 180 interposed between the optical filter 160 and the set of pixels 170 and configured to spread collimated light output from each lens in the set of lenses 150 across a set of subpixels of a corresponding pixel in the set of pixels 170 .
  • the system includes a lens tube in replacement of (or in addition to) each aperture and lens pair described above.
  • each lens tube can be characterized by a second (short) focal length and can be offset from the focal plane of the bulk imaging optic 130 by the second focal length to preserve the aperture of the bulk imaging optic 130 and to collimate incident light received from the bulk imaging optic 130 , as described above and as shown in FIGS. 5 and 7 .
  • Each lens tube also defines an opaque cylindrical wall 218 defining an axis normal to the incidence plane of the adjacent optical filter 160 and configured to absorb incident light rays, as shown in FIG. 5 .
  • the cylindrical wall 218 of a lens tube may absorb light rays passing through the lens tube at shallower angles to the axis of the lens tube, thereby reducing the field of view of the lens tube (which may be similar to decreasing the diameter of an aperture in the aperture layer up to the diffraction-limited diameter, as described above) and yielding an output signal of collimated light rays nearer to perpendicular to the incidence plane of the optical filter 160 .
  • Each lens tube can therefore define an elongated cylindrical wall 218 of length sufficient to achieve a target field of view and to pass collimated light rays at maximum angles to the axis of the lens tube less than a threshold angle.
  • a lens tube can thus function as an aperture-sense pair described above to define a narrow field of view and to output substantially collimated light to the adjacent optical filter 160 .
  • the cylindrical wall 218 of a lens tube can define a coarse or patterned opaque interface about a transparent (or translucent) lens material, as shown in FIG. 5 , to increase absorption and decrease reflection of light rays incident on the cylindrical wall 218 .
  • Each lens tube (and each lens described above) can also be coated with an anti-reflective coating.
  • the set of lens tubes 210 can be fabricated by implementing photolithography techniques to pattern a photoactive optical polymer (e.g., SU8) onto the optical filter 160 (e.g., on a silicon wafer defining the optical filter).
  • a light-absorbing polymer can then be poured between the lens tubes and cured.
  • a set of lenses 150 can then be fabricated (e.g., molded) separately and then bonded over the lens tubes.
  • lenses can be fabricated directly onto the lens tubes by photolithography techniques.
  • a mold for lenses can be cast directly onto the lens tubes by injecting polymer into a mold arranged over the lens tubes.
  • a singular diffuser 180 or multiple discrete diffusers 180 can be similarly fabricated and/or assembled on the optical filter 160 opposite the lens tubes. Standoffs extending from the optical filter 160 can be similarly fabricated or installed around the diffuser(s) 180 , and the image sensor can be aligned with and bonded to the standoffs opposite the optical filter 160 .
  • Other optical elements within the system e.g., the bulk imaging lens, the bulk transmitting lens, etc.
  • Another variation of the system includes: a set of illumination sources 110 arranged in a first rectilinear grid array, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in a field ahead of the illumination source; a bulk imaging optic 130 characterized by a focal plane opposite the field; an aperture layer coincident the focal plane, defining a set of apertures 144 in a second rectilinear grid array proportional to the first rectilinear grid array, and defining a stop region 146 around the set of apertures 144 , each aperture in the set of apertures 144 defining a field of view in the field coincident a discrete spot output by a corresponding illumination source in the set of illumination sources 110 , the stop region 146 absorbing light rays reflected from surfaces in the field outside of fields of view defined by the set of apertures 144 and passing through the bulk imaging optic 130 ; a set of lenses 150 , each lens in the set of lenses 150 characterized by
  • the system includes a two-dimensional grid array of channels (i.e., aperture, lens, and pixel sets or lens tube and pixel sets) and is configured to image a volume occupied by the system in two dimensions.
  • the system can collect one-dimensional distance data—such as counts of incident photons within a sampling period and/or times between consecutive photons incident on pixels of known position corresponding to known fields of view in the field—across a two-dimensional field.
  • the one-dimensional distance data can then be merged with known positions of the fields of view for each channel in the system to reconstruct a virtual three-dimensional representation of the field ahead of the system.
  • the aperture layer can define a grid array of apertures
  • the set of lenses 150 can be arranged in a similar grid array with one lens aligned with one aperture in the aperture layer
  • the set of pixels 170 can include one pixel per aperture and lens pair, as described above.
  • the aperture layer can define a 24-by-24 grid array of 200- ⁇ m-diameter apertures offset vertically and laterally by an aperture pitch distance of 300 ⁇ m
  • the set of lenses 150 can similarly define a 24-by-24 grid array of lenses offset vertically and laterally by a lens pitch distance of 300 ⁇ m
  • the set of pixels 170 can include a 24-by-24 grid array of 300- ⁇ m-square pixels, wherein each pixel includes a 3 ⁇ 3 square array of nine 100- ⁇ m-square SPADs.
  • the set of pixels 170 can include one pixel per group of multiple aperture and lens pairs.
  • the set of pixels 170 can alternatively include a 12-by-12 grid array of 600- ⁇ m-square pixels, wherein each pixel includes a 6 ⁇ 6 square array of 36 100- ⁇ m-square SPADs and wherein each pixel is aligned with a group of four adjacent lenses in a square grid.
  • the diffuser 180 can bias collimated light rays output from a lens in the (1,1) position in the square grid upward and to the right to spread light rays passing through the (1,1) lens across the full breadth and width of the corresponding pixel; can bias collimated light rays output from a lens in the (2,1) position in the square grid upward and to the left to spread light rays passing through the (2,1) lens across the full breadth and width of the corresponding pixel; can bias collimated light rays output from a lens in the (1,2) position in the square grid downward and to the right to spread light rays passing through the (1,2) lens across the full breadth and width of the corresponding pixel; and can bias collimated light rays output from a lens in the (2,2) position in the square grid downward and to the left to spread light rays passing through the (2,2) lens across the full breadth and width of the corresponding pixel, as shown in FIG. 8 .
  • the system can actuate one illumination source in the group of four illumination sources at any given instance in time.
  • the system can actuate a first illumination source 111 in a (1,1) position during a first sampling period to illuminate a field of view defined by a first aperture 141 corresponding to a lens in the (1,1) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the first sampling period.
  • the system can then shut down the first illumination source 111 and actuate a second illumination source 112 in a (1,2) position during a subsequent second sampling period to illuminate a field of view defined by a second aperture 142 corresponding to a lens in the (1,2) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the second sampling period.
  • the system can then shut down the first and second illumination sources 112 and actuate a third illumination source in a (2,1) position during a subsequent third sampling period to illuminate a field of view defined by a third aperture corresponding to a lens in the (2,1) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the third sampling period.
  • the system can shut down the first, second, and third illumination sources and actuate a fourth illumination source in a (2,2) position during a fourth sampling period to illuminate a field of view defined by a fourth aperture corresponding to a lens in the (2,2) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the fourth sampling period.
  • the system can repeat this process throughout its operation.
  • the system can include a set of pixels 170 arranged across an image sensor 7.2 mm in width and 7.2 mm in length and can implement a scanning schema such that each channel in the system can access (can project light rays onto) a number of SPADs otherwise necessitating a substantially larger image sensor (e.g., a 14.4 mm by 14.4 mm image sensor).
  • the system can implement a serial scanning schema per group of illumination sources to achieve an exponential increase in the dynamic range of each channel in the system.
  • the system can implement the foregoing imaging techniques to increase imaging resolution of the system.
  • the system can also include a shutter 182 between each channel and the image sensor, and the system can selectively open and close each shutter 182 when the illumination source for the corresponding channel is actuated and deactivated, respectively.
  • the system can include one independently-operable electrochromic shutter 182 interposed between each lens, and the system can open the electrochromic shutter 182 over the (1,1) lens in the square-gridded group of four lenses and close electrochromic shutters 182 over the (1,2), (2,1), and (2,2) lens when the (1,1) illumination source is activated, thereby rejecting noise passing through the (1,2), (2,1), and (2,2) lens from reaching the corresponding pixel on the image sensor.
  • the system can therefore selectively open and close shutters 182 between each channel and the image sensor to increase SNR per channel during operation.
  • the system can include one independently-operable electrochromic shutter 182 arranged over select regions of each pixel, as shown in FIG. 8 , wherein each electrochromic shutter 182 is aligned with a single channel (i.e., with a single lens in the set of lenses).
  • the system can alternatively include MEMS mechanical shutters or any other suitable type of shutter interposed between the set of lenses 150 and the image sensor.
  • the system can define two-dimension grid arrays of apertures, lenses, diffusers, and/or pixels characterized by a first pitch distance along a first (e.g., X) axis and a second pitch distance—different from the first pitch distance—along a second (e.g., Y) axis.
  • the image sensor can include pixels offset by a 25 ⁇ m horizontal pitch and a 300 ⁇ m vertical pitch, wherein each pixel includes a single row of twelve subpixels.
  • the two-dimensional optical system can include an array of any other number and pattern of channels (e.g., apertures, lenses (or lens tubes), and diffusers) and pixels and can execute any other suitable scanning schema to achieve higher spatial resolutions per channel than the raw pixel resolution of the image sensor.
  • the system can additionally or alternatively include a converging optic, a diverging optic, and/or any other suitable type of optical element to spread light rights passed from a channel across the breadth of a corresponding pixel.

Abstract

An optical system for collecting distance information within a field is provided. The optical system may include lenses for collecting photons from a field and may include lenses for distributing photons to a field. The optical system may include lenses that collimate photons passed by an aperture, optical filters that reject normally incident light outside of the operating wavelength, and pixels that detect incident photons. The optical system may further include illumination sources that output photons at an operating wavelength.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application is a continuation of U.S. patent application Ser. No. 15/880,491, filed Jan. 25, 2018, which is a continuation of U.S. patent application Ser. No. 15/276,532 (“the '532 application”), filed Sep. 26, 2016; which claims the benefit of U.S. Provisional Patent Application No. 62/232,222 (“the '222 application”) filed Sep. 24, 2015. The present application is also related to U.S. patent application Ser. No. 15/861,330, filed on Jan. 3, 2018, which is a continuation-in-part of the '532 application. The disclosures of the '532 and '222 applications are each incorporated by reference herein in their entirety.
BACKGROUND OF THE INVENTION Field of the Invention
This invention relates generally to the field of optical sensors and more specifically to a new and useful optical system for collecting distance information in the field of optical sensors.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a schematic representation of a system.
FIG. 2 is a schematic representation in accordance with one variation of the system.
FIG. 3 is a schematic representation in accordance with one variation of the system.
FIG. 4 is a schematic representation in accordance with one variation of the system.
FIG. 5 is a schematic representation in accordance with one variation of the system.
FIG. 6 is a schematic representation in accordance with one variation of the system.
FIG. 7 is a schematic representation in accordance with one variation of the system.
FIG. 8 is a schematic representation in accordance with one variation of the system.
FIG. 9 is a flowchart representation in accordance with one variation of the system.
FIG. 10 is a schematic representation in accordance with one variation of the system.
FIG. 11 is a schematic representation in accordance with one variation of the system.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
The following description of embodiments of the invention is not intended to limit the invention to these embodiments but rather to enable a person skilled in the art to make and use this invention. Variations, configurations, implementations, example implementations, and examples described herein are optional and are not exclusive to the variations, configurations, implementations, example implementations, and examples they describe. The invention described herein can include any and all permutations of these variations, configurations, implementations, example implementations, and examples.
1. One-Dimensional Optical System: Aperture Array
As shown in FIG. 1, a one-dimensional optical system 100 for collecting distance information within a field includes: a set of illumination sources 110 arranged along a first axis, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in the field ahead of the illumination source; a bulk imaging optic 130 characterized by a focal plane opposite the field; an aperture layer 140 coincident the focal plane, defining a set of apertures 144 in a line array parallel to the first axis, and defining a stop region 146 around the set of apertures 144, each aperture in the set of apertures 144 defining a field of view in the field coincident a discrete spot output by a corresponding illumination source in the set of illumination sources 110, the stop region 146 absorbing light rays reflected from surfaces in the field outside of fields of view defined by the set of apertures 144 and passing through the bulk imaging optic 130; a set of lenses 150, each lens in the set of lenses 150 characterized by a second focal length, offset from the focal plane opposite the bulk imaging optic 130 by the second focal length, aligned with an aperture in the set of apertures 144, and configured to collimate light rays passed by the aperture; an optical filter 160 adjacent the set of lenses 150 opposite the aperture layer 140 and configured to pass light rays at the operating wavelength; a set of pixels 170 adjacent the optical filter 160 opposite the set of lenses 150, each pixel in the set of pixels 170 corresponding to a lens in the set of lenses 150 and including a set of subpixels arranged along a second axis non-parallel to the first axis; and a diffuser 180 interposed between the optical filter 160 and the set of pixels 170 and configured to spread collimated light output from each lens in the set of lenses 150 across a set of subpixels of a corresponding pixel in the set of pixels 170.
1.1 Applications
Generally, the one-dimensional optical system 100 (the “system”) functions as an image sensor that, when rotated about an axis parallel to a column of apertures, collects three-dimensional distance data of a volume occupied by the system. Specifically, the one-dimensional optical system 100 can scan a volume to collect three-dimensional distance data that can then be reconstructed into a virtual three-dimensional representation of the volume, such as based on recorded times between transmission of illuminating beams from the illumination sources and detection of photons—likely originating from the illumination sources—incident on the set of pixels 170, based on phase-based measurements techniques, or based on any other suitable distance measurement technique. The system 100 includes: a column of offset apertures arranged behind a bulk imaging optic 130 and defining discrete fields of view in a field ahead of the bulk imaging optic 130 (that is non-overlapping fields of view beyond a threshold distance from the system); a set of illumination sources 110 that project discrete illuminating beams at an operating wavelength into (and substantially only into) the fields of view defined by the apertures; a column of lenses that collimate light rays passed by corresponding apertures; and an optical filter 160 that selectively passes a narrow band of wavelengths of light (i.e., electromagnetic radiation) including the operating wavelength; and a set of pixels 170 that detect incident photons (e.g., count incident photons, tracks times between consecutive incident photons). The system can therefore selectively project illuminating beams into a field ahead of the system according to an illumination pattern that substantially matches—in size and geometry across a range of distances from the system—the fields of view of the apertures. In particular, the illumination sources are configured to illuminate substantially only surfaces in the field ahead of the system that can be detected by pixels in the system such that minimal power output by the system (via the illumination sources) is wasted by illuminating surfaces in the field for which the pixels are blind. The system can therefore achieve a relatively high ratio of output signal (i.e., illuminating beam power) to input signal (i.e., photons passed to an incident on the pixel array). Furthermore, the set of lenses 150 can collimate light rays passed by adjacent apertures such that light rays incident on the optical filter 160 meet the optical filter 160 at an angle of incidence of approximately 0°, thereby maintaining a relatively narrow band of wavelengths of light passed by the optical filter 160 and achieving a relatively high signal-to-noise ratio (“SNR”) for light rays reaching the set of pixels 170.
The system includes pixels arranged in a column and aligned with the apertures, and each pixel can be non-square in geometry (e.g., short and wide) to extend the sensing area of the system for a fixed aperture pitch and pixel column height. The system also includes a diffuser 180 that spreads light rays passed from an aperture through the optical filter 160 across the area of a corresponding pixel such that the pixel can detect incident photons across its full width and height thereby increasing the dynamic range of the system.
The system is described herein as projecting electromagnetic radiation into a field and detecting electromagnetic radiation reflected from a surface in the field back to bulk receiver optic. Terms “illumination beam,” “light,” “light rays,” and “photons” recited herein refer to such electromagnetic radiation. The term “channel” recited herein refers to one aperture in the aperture layer 140, a corresponding lens in the set of lenses 150, and a corresponding pixel in the set of pixels 170.
1.2 Bulk Imaging Optic
The system includes a bulk imaging optic 130 characterized by a focal plane opposite the field. Generally, the bulk imaging optic 130 functions to project incident light rays from outside the system toward the focal plane where light rays incident on a stop region 146 of the aperture layer 140 are rejected (e.g., mirrored or absorbed) and where light rays incident on apertures in the aperture layer 140 are passed into a lens characterized by a focal length and offset from the focal plane by the focal length.
In one implementation, the bulk imaging optic 130 includes a converging lens, such as a bi-convex lens (shown in FIG. 2) or a plano-convex lens, characterized by a particular focal length at the operating wavelength of the system. The bulk imaging optic 130 can also include multiple discrete lens that cooperate to project light rays toward the aperture layer 140 and that are characterized by a composite focal plane opposite the field, as shown in FIG. 11. However, the bulk imaging optic 130 can be any other suitable type of lens or combination of lenses of any other type or geometry.
1.3 Aperture Layer
As shown in FIGS. 1 and 2, the system includes an aperture layer 140 coincident the focal plane, defining a set of apertures 144 in a line array parallel to the axes of the illumination sources, and defining a stop region 146 around the set of apertures 144, wherein each aperture in the set of apertures 144 defines a field of view in the field coincident a discrete spot output by a corresponding illumination source in the set of illumination sources 110, and wherein the stop region 146 absorbs and/or reflects light rays reflected from surfaces in the field outside of fields of view defined by the set of apertures 144 and passing through the bulk imaging optic 130. Generally, the aperture layer 140 defines an array of open regions (i.e., apertures, including one aperture per lens) and closed regions (“stop regions”) between adjacent opens. Each aperture in the aperture layer 140 defines a “pinhole” that defines a field of view for its corresponding sense channel and passes light rights reflected from an external surface within its field of the view into its corresponding lens, and each stop region 146 can block light rays incident on select regions of the focal plane from passing into the lens array, as shown in FIG. 6.
The aperture layer 140 includes a relatively thin opaque structure coinciding with (e.g., arranged along) the focal plane of the bulk imaging optic 130, as shown in FIGS. 1 and 2. For example, the aperture layer 140 can include a 10 micrometer-thick copper, silver, or nickel film deposited (e.g., plated) over a photocurable transparent polymer and then selectively etched to form the array of apertures. In a similar example, a reflective metalized layer or a light-absorbing photopolymer (e.g., a photopolymer mixed with a light absorbing dye) can be deposited onto a glass wafer and selectively cured with a photomask to form the aperture layer 140 and the set of apertures 144. Alternatively, the aperture layer 140 can include a discrete metallic film that is mechanically or chemically perforated to form the array of apertures, bonded to the lens array, and then installed over the bulk imaging optic 130 along the focal plane. However, the aperture layer 140 can include any other reflective (e.g., mirrored) or light-absorbing material formed in any other way to define the array of apertures along the focal plane of the bulk imaging optic 130.
In the one-dimensional optical system 100, the aperture layer 140 can define a single column of multiple discrete circular apertures of substantially uniform diameter, wherein each aperture defines an axis substantially parallel to and aligned with one lens in the lens array, as shown in FIG. 3. Adjacent apertures are offset by an aperture pitch distance greater than the aperture diameter and substantially similar to the lens pitch distance, and the aperture layer 140 defines a stop region 146 (i.e., an opaque or reflecting region) between adjacent apertures such that the apertures define discrete, non-overlapping fields of view for their corresponding sense channels. At increasingly smaller diameters up to a diffraction-limited diameter—which is a function of wavelength of incident light and numeric aperture of the bulk imaging lens—an aperture defines a narrower field of view (i.e., a field of view of smaller diameter) and passes a sharper but lower-intensity (attenuated) signal from the bulk imaging optic 130 into its corresponding lens. The aperture layer 140 can therefore define apertures of diameter: greater than the diffraction-limited diameter for the wavelength of light output by the illumination sources (e.g., 900 nm); substantially greater than the thickness of the aperture layer 140; and less than the aperture pitch distance, which is substantially equivalent to the lens pitch distance and the pixel pitch distance. In one example, aperture layer 140 can define apertures of diameters approaching the diffraction-limited diameter to maximize geometrical selectivity of the field of view of each sense channel. Alternatively, the apertures can be of diameter less that the diffraction-limited diameter for the wavelength of light output by the illumination sources. In one example, the aperture layer 140 can define apertures of diameters matched to a power output of illumination sources in the system and to a number and photon detection capacity of subpixel photodetectors in each pixel in the set of pixels 170 to achieve a target number of photons incident on each pixel within each sampling period. In this example, each aperture can define a particular diameter that achieves target attenuation range for pixels originating from a corresponding illumination source and incident on the bulk imaging optic 130 during a sampling period. In particular, because an aperture in the aperture layer 140 attenuates a signal passed to its corresponding lens and on to its corresponding pixel, the diameter of the aperture can be matched to the dynamic range of its corresponding pixel.
In one implementation, a first aperture 141 in the aperture layer 140 passes light rays—reflected from a discrete region of a surface in the field (the field of view of the sense channel) ahead of the bulk imaging optic 130—into its corresponding lens; a stop region 146 interposed between the first aperture 141 and adjacent apertures in the aperture layer 140 blocks light rays—reflected from a region of the surface outside of the field of view of the first aperture 141—from passing into the lens corresponding to the first aperture 141. In the one-dimensional optical system 100, the aperture layer 140 therefore defines a column of apertures that define multiple discrete, non-overlapping fields of view of substantially infinite depth of field, as shown in FIG. 2.
In this implementation, a first aperture 141 in the aperture layer 140 defines a field of view that is distinct and that does not intersect a field of view defined by another aperture in the aperture layer 140, as shown in FIG. 2. The set of illumination sources 110 includes a first illumination source 111 paired with the first aperture 141 and configured to project an illuminating beam substantially aligned with (i.e., overlapping) the field of view of the first aperture 141 in the field ahead of the bulk imaging optic 130. Furthermore, the first illumination source 111 and a bulk transmitting optic 120 can cooperate to project an illuminating beam of a cross-section substantially similar to (and slightly larger than) the cross section of the field of view of the first aperture 141 as various distances from the bulk imaging optic 130. Therefore light output by the first illumination source 111—paired with the first aperture 141—and projected into the field of view of the first aperture 141 can remain substantially outside the fields of view of other apertures in the aperture layer 140.
Generally, photons projected into the field by the first illumination source 111 illuminate a particular region of a surface (or multiple surfaces) in the field within the field of view of the first sense channel and are reflected (e.g., scattered) by the surface(s); at least some of these photons reflected by the particular region of a surface may reach the bulk imaging optic 130, which directs these photons toward the focal plane. Because these photons were reflected by a region of a surface within the field of view of the first aperture 141, the bulk imaging optic 130 may project these photons into the first aperture 141, and the first aperture 141 may pass these photons into the first lens 151 (or a subset of these photons incident at an angle relative to the axis of the first aperture 141 below a threshold angle). However, because a second aperture 142 in the aperture layer 140 is offset from the first aperture 141 and because the particular region of the surface in the field illuminated via the first illumination source 111 does not (substantially) coincide with the field of view of the second aperture 142, photons reflected by the particular region of the surface and reaching the bulk imaging optic 130 are projected into the second aperture 142 and passed to a second lens 152 behind the second aperture 142, and vice versa, as shown in FIG. 2. Furthermore, a stop region 146 between the first and second apertures 142 can block photons directed toward the focal plane between the first and second apertures 142 reflected by the bulk imaging optic 130, thereby reducing crosstalk between the first and second sense channels.
For a first aperture 141 in the aperture layer 140 paired with a first illumination source 111 in the set of illumination sources 110, the first aperture 141 in the aperture layer 140 defines a first field of view and passes—into the first lens 151—incident light rays originating at or reflected from a surface in the field coinciding with the first field of view. Because the first illumination source 111 projects an illuminating beam that is substantially coincident (and substantially the same size as or minimally larger than) the field of view defined by the first aperture 141 (as shown in FIG. 4), a signal passed into the first lens 151 by the first aperture 141 in the aperture layer 140 can exhibit a relatively high ratio of light rays originating from the first illumination source 111 to light rays originating from other illumination sources in the system. Generally, because various illumination sources in the system may output illuminating beams at different frequencies, duty cycles, and/or power levels, etc. at a particular time during operation, light rays passed from the bulk imaging optic 130 into a first pixel 171 in the set of pixels 170 but originating from an illumination source other than the first illumination source 111 paired with the first pixel 171 constitute noise at the first pixel 171. Though the relatively small diameters of apertures in the aperture layer 140 may attenuate a total light signal passed from the bulk imaging optic 130 into the set of lenses 150, each aperture in the aperture layer 140 may pass a relatively high proportion of photons originating from its corresponding illumination source than from other illumination sources in the system; that is, due to the geometry of a particular aperture and its corresponding illumination source, a particular aperture may pass a signal exhibiting a relatively high SNR to its corresponding lens and thus into its corresponding pixel. Furthermore, at smaller aperture diameters in the aperture layer 140—and therefore smaller fields of view of corresponding channels—the system can pass less noise from solar radiation or other ambient light sources to the set of pixels 170.
In one variation, the system includes a second aperture layer interposed between the lens array and the optical filter 160, wherein the second aperture layer defines a second set of apertures 144, each aligned with a corresponding lens in the set of lenses 150, as described above. In this variation, an aperture in the second aperture layer 140 can absorb or reflect errant light rays passed by a corresponding lens, as described above, to further reduce crosstalk between channels, thereby improving SNR within the system. Similarly, the system can additionally or alternatively include a third aperture layer interposed between the optical filter 160 and the diffuser(s) 180, wherein the third aperture layer defines a third set of apertures 144, each aligned with a corresponding lens in the set of lenses 150, as described above. In this variation, an aperture in the third aperture layer can absorb or reflect errant light rays passed by the light filter, as described above, to again reduce crosstalk between channels, thereby improving SNR within the system.
1.4 Lens Array
The system includes a set of lenses 150, wherein each lens in the set of lenses 150 is characterized by a second focal length, is offset from the focal plane opposite the bulk imaging optic 130 by the second focal length, is aligned with a corresponding aperture in the set of apertures 144, and is configured to collimate light rays passed by the corresponding aperture. Generally, a lens in the set of lenses 150 functions to collimate lights rays passed by its corresponding aperture and to pass these collimated light rays into the optical filter 160.
In the one-dimensional optical system 100, the lenses are arranged in a single column, and adjacent lenses are offset by a uniform lens pitch distance (i.e., a center-to-center-distance between adjacent pixels), as shown in FIG. 3. The set of lenses 150 is interposed between the aperture layer and the optical filter 160. In particular, each lens can include a converging lens characterized by a second focal length and can be offset from the focal plane of the bulk imaging optic 130—opposite the bulk imaging optic 130—by the second focal length to preserve the aperture of the bulk imaging optic 130 and to collimate light incident on the bulk imaging optic 130 and passed by a corresponding aperture. Each lens in the set of lens can be characterized by a relatively short focal length (i.e., less than a focal length of the bulk imaging optic 130) and a relatively large marginal ray angle (e.g., a relatively high numeric aperture lens) such that the lens can capture highly-angled light rays projected toward the lens by the extent of the bulk imaging optic 130. That is, each lens in the set of lens can be characterized by a ray cone substantially matched to a ray cone of the bulk imaging optic 130.
Lenses in the set of lenses 150 can be substantially similar. A lens in the set of lenses 150 is configured to collimate light rays focused into its corresponding aperture by the bulk imaging optic 130. For example, a lens in the set of lenses 150 can include a bi-convex or plano-convex lens characterized by a focal length selected based on the size (e.g., diameter) of its corresponding aperture and the operating wavelength of the system. In this example, the focal length (f) of a lens in the set of lenses 150 can be calculated according to the formula:
f = d 2 2 λ
where d is the diameter of the corresponding aperture in the aperture layer and λ is the operating wavelength of light output by the illumination source (e.g., 900 nm). The geometry of a lens in the set of lenses 150 can therefore be matched to the geometry of a corresponding aperture in the aperture layer such that the lens passes a substantially sharp image of light rays—at or near the operating wavelength—into the optical filter 160 and thus on to the pixel array.
However, the set of lenses 150 can include lenses of any other geometry and arranged in any other way adjacent the aperture layer.
1.5 Optical Filter
As shown in FIG. 3, the system includes an optical filter 160 adjacent the set of lenses 150 opposite the aperture layer and configured to pass light rays at the operating wavelength. Generally, the optical filter 160 receives electromagnetic radiation across a spectrum from the set of lenses 150, passes a relatively narrow band of electromagnetic radiation—including radiation at the operating wavelength—to the pixel array, and blocks electromagnetic radiation outside of the band. In particular, electromagnetic radiation other than electromagnetic radiation output by the illumination source—such as ambient light—incident on a pixel in the set of pixels 170 constitutes noise in the system. The optical filter 160 therefore functions to reject electromagnetic radiation outside of the operating wavelength or, more pragmatically, outside of a narrow wavelength band, thereby reducing noise in the system and increasing SNR.
In one implementation, the optical filter 160 includes an optical bandpass filter that passes a narrow band of electromagnetic radiation substantially centered at the operating wavelength of the system. In one example, the illumination sources output light (predominantly) at an operating wavelength of 900 nm, and the optical filter 160 is configured to pass light between 899.95 nm and 900.05 nm and to block light outside of this band.
The optical filter 160 may selectively pass and reject wavelengths of light as a function of angle of incidence on the optical filter 160. Generally, optical bandpass filters may pass wavelengths of light inversely proportional to their angle of incidence on the light optical bandpass filter. For example, for an optical filter 160 including a 0.5 nm-wide optical bandpass filter, the optical filter 160 may pass over 95% of electromagnetic radiation over a sharp band from 899.75 nm to 900.25 nm and reject approximately 100% of electromagnetic radiation below 899.70 nm and above 900.30 nm for light rays incident on the optical filter 160 at an angle of incidence of approximately 0°. However, in this example, the optical filter 160 may pass over 95% of electromagnetic radiation over a narrow band from 899.5 nm to 900.00 nm and reject approximately 100% of electromagnetic radiation over a much wider band below 899.50 nm and above 900.30 nm for light rays incident on the optical filter 160 at an angle of incidence of approximately 15°. Therefore, the incidence plane of the optical filter 160 can be substantially normal to the axes of the lenses, and the set of lenses 150 can collimate light rays received through a corresponding aperture and output these light rays substantially normal to the incidence plane of the optical filter 160 (i.e., at an angle of incidence of approximately 0° on the optical filter). Specifically, the set of lenses 150 can output light rays toward the optical filter 160 at angles of incidence approximating 0° such that substantially all electromagnetic radiation passed by the optical filter 160 is at or very near the operating wavelength of the system.
In the one-dimensional optical system 100, the system can include a single optical filter 160 that spans the column of lens in the set of lenses 150. Alternatively, the system can include multiple optical filters 160, each adjacent a single lens or a subset of lenses in the set of lenses 150. However, the optical filter 160 can define any other geometry and can function in any other way to pass only a limited band of wavelengths of light.
1.6 Pixel Array and Diffuser
The system includes a set of pixels 170 adjacent the optical filter 160 opposite the set of lenses 150, each pixel in the set of pixels 170 corresponding to a lens in the set of lenses 150 and including a set of subpixels arranged along a second axis non-parallel to the first axis. Generally, the set of pixels 170 are offset from the optical filter 160 opposite the set of lenses 150, and each pixel in the set of pixels 170 functions to output a single signal or stream of signals corresponding to the count of photons incident on the pixel within one or more sampling periods, wherein each sampling period may be picoseconds, nanoseconds, microseconds, or milliseconds in duration.
The system also includes a diffuser 180 interposed between the optical filter 160 and the set of pixels 170 and configured to spread collimated light output from each lens in the set of lenses 150 across a set of subpixels of a single corresponding pixel in the set of pixels 170. Generally, for each lens in the set of lenses 150, the diffuser 180 functions to spread light rays—previously collimated by the lens and passed by the optical filter 160—across the width and height of a sensing area within a corresponding pixel. The diffuser 180 can define a single optic element spanning the set of lenses 150, or the diffuser 180 can include multiple discrete optical elements, such as including one optical diffuser element aligned with each channel in the system.
In one implementation, a first pixel 171 in the set of pixels 170 includes an array of single-photon avalanche diode detectors (hereinafter “SPADs”), and the diffuser 180 spreads lights rays—previously passed by a corresponding first aperture 141, collimated by a corresponding first lens 151, and passed by the optical filter 160—across the area of the first pixel 171, as shown in FIGS. 3, 5, and 6. Generally, adjacent apertures can be aligned and offset vertically by an aperture pitch distance, adjacent lenses can be aligned and offset vertically by a lens pitch distance substantially identical to the aperture pitch distance, and adjacent pixels can be aligned and offset vertically by a pixel pitch distance substantially identical to the lens and aperture pitch distances. However, the pixel pitch distance may accommodate only a relatively small number of (e.g., two) vertically-stacked SPADs. Each pixel in the set of pixels 170 can therefore define an aspect ratio greater than 1:1, and the diffuser 180 can spread light rays passed by the optical filter 160 according to the geometry of a corresponding pixel in order to accommodate a larger sensing area per pixel.
In one example, each pixel in the set of pixels 170 is arranged on an image sensor, and a first pixel 171 in the set of pixels 170 includes a single row of 16 SPADs spaced along a lateral axis perpendicular to a vertical axis bisecting the column of apertures and lenses. In this example, the height of a single SPAD in the first pixel 171 can be less than the height (e.g., diameter) of the first lens 151, but the total length of the 16 SPADs can be greater than the width (e.g., diameter) of the first lens 151; the diffuser 180 can therefore converge light rays output from the first lens 151 to a height corresponding to the height of a SPAD at the plane of the first pixel 171 and can diverge light rays output from the first lens 151 to a width corresponding to the width of the 16 SPADs at the plane of the first pixel 171. In this example, the remaining pixels in the set of pixels 170 can include similar rows of SPADs, and the diffuser 180 can similarly converge and diverge light rays passed by corresponding apertures onto corresponding pixels.
In the foregoing example, the aperture layer can include a column of 16 like apertures, the set of lenses 150 can include a column of 16 like lenses arranged behind the aperture layer, and the set of pixels 170 can include a set of 16 like pixels—each including a similar array of SPADs—arranged behind the set of lenses 150. For a 6.4 mm-wide, 6.4 mm-tall image sensor, each pixel can include a single row of 16 SPADs, wherein each SPAD is electrically coupled to a remote analog front-end processing electronics/digital processing electronics circuit 240. Each SPAD can be arranged in a 400 μm-wide, 400 μm-tall SPAD area and can define an active sensing area approaching 400 μm in diameter. Adjacent SPADs can be offset by a SPAD pitch distance of 400 μm. In this example, the aperture pitch distance along the vertical column of apertures, the lens pitch distance along the vertical column of lenses, and the pixel pitch distance along the vertical column of pixels can each be approximately 400 μm accordingly. For the first sense channel in the system (i.e., the first aperture 141, the first lens 151, and the first pixel 171, etc.), a first diffuser 180 can diverge a cylindrical column of light rays passed from the first lens 151 through the optical filter 160—such as a column of light approximately 100 μm in diameter for an aperture layer aspect ratio of 1:4—to a height of approximately 400 μm aligned vertically with the row of SPADs in the first pixel 171. The first diffuser can similarly diverge the cylindrical column of light rays passed from the first lens 151 through the optical filter 160 to a width of approximately 6.4 μm centered horizontally across the row of SPADs in the first pixel 171. Other diffusers 180 in the system can similarly diverge (or converge) collimated light passed by corresponding lenses across corresponding pixels in the set of pixels 170. Therefore, in this example, by connecting each SPAD (or each pixel) to a remote analog front-end processing electronics/digital processing electronics circuit 240 and by incorporating diffusers 180 that spread light passed by the optical filter 160 across the breadths and heights of corresponding pixels, the system can achieve a relatively high sensing area fill factor across the imaging sensor.
Therefore, in the one-dimensional optical system 100, pixels in the set of pixels 170 can include an array of multiple SPADS arranged in aspect ratio exceeding 1:1, and the diffuser 180 can spread light rays across corresponding non-square pixels that enables a relatively large numbers of SPADs to be tiled across a single pixel to achieve a greater dynamic range across the image sensor than an image sensor with a single SPAD per pixel, as shown in FIG. 3. In particular, by incorporating multiple SPADs per pixel (i.e., per sense channel), a first sense channel in the system can detect multiple incident photons—originating from a surface in the field bound by a field of view defined by the first aperture 141—within the span of the dead time characteristic of the SPADs. The first sense channel can therefore detect a “brighter” surface in its field of view. Additionally or alternatively, the first pixel 171 in the first sense channel can be sampled faster than the dead time characteristic of SPADs in the first pixel 171 because, though a first subset of SPADs in the first pixel 171 may be down (or “dead”) during a first sampling period due to collection of incident photons during the first sampling period, other SPADs in the first pixel 171 remain on (or “alive”) and can therefore collect incident photons during a subsequent sampling period. Furthermore, by incorporating pixels characterized by relatively high aspect ratios of photodetectors, the image sensor can include pixels offset by a relatively small pixel pitch, but the system 100 can still achieve a relatively high dynamic range pixel.
However, pixels in the set of pixels 170 can include any other number of SPADs arranged in any other arrays, such as in a 64-by-1 grid array (as described above), in a 32-by-2 grid array, or in a 16-by-4 grid array, and the diffuser 180 can converge and/or diverge collimated light rays onto corresponding pixels accordingly in any other suitable way. Furthermore, rather than (or in addition to) SPADs, each pixel in the set of pixels 170 can include one or more linear avalanche photodiodes, Geiger mode avalanche photodiodes, photomultipliers, resonant cavity photodiodes, QUANTUM DOT detectors, or other types of photodetectors arranged as described above, and the diffuser(s) 180 can similarly converge and diverge signals passed by the optical filter(s) 160 across corresponding pixels, as described herein.
1.7 Illumination Sources
The system includes a set of illumination sources 110 arranged along a first axis, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in a field ahead of the illumination source. Generally, each illumination source functions to output an illuminating beam coincident a field of view defined by a corresponding aperture in the set of apertures 144, as shown in FIGS. 1 and 2.
In one implementation, the set of illumination sources 110 includes a bulk transmitter optic and one discrete emitter per sense channel. For example, the set of illumination sources 110 can include a monolithic VCSEL arrays including a set of discrete emitters. In this implementation, the bulk transmitter optic can be substantially identical to the bulk imaging optic 130 in material, geometry (e.g., focal length), thermal isolation, etc., and the bulk transmitter optic is adjacent and offset laterally and/or vertically from the bulk imaging optic 130. In a first example, set of illumination sources 110 includes a laser array including discrete emitters arranged in a column with adjacent emitters offset by an emitter pitch distance substantially identical to the aperture pitch distance. In this first example, each emitter outputs an illuminating beam of diameter substantially identical to or slightly greater than the diameter of a corresponding aperture in the apertures layer, and the column of emitters is arranged along the focal plane of the bulk transmitter optic such that each illuminating beam projected from the bulk transmitter optic into the field intersects and is of substantially the same size and geometry as the field of view of the corresponding sense channel, as shown in FIG. 4. Therefore, substantially all power output by each emitter in the set of illumination sources 110 can be projected into the field of view of its corresponding sense channel with relatively minimal power wasted illuminating surfaces in the field outside of the fields of view of the sense channels.
In a second example, the discrete emitters are similarly arranged in a column with adjacent emitters offset by an emitter pitch distance twice the aperture pitch distance, as shown in FIG. 2. In this second example, each emitter is characterized by an illuminating active area (or aperture) of diameter approximately (or slightly greater than) twice the diameter of a corresponding aperture in the apertures layer, and the column of emitters is offset behind the bulk transmitter optic by twice the focal length of the bulk transmitter optic such that each illuminating beam projected from the bulk transmitter optic into the field intersects and is of substantially the same size and geometry as the field of view of the corresponding sense channel, as described above. Furthermore, for the same illumination beam power density, an illuminating beam output by an emitter in this second example may contain four times the power of an illuminating beam output by an emitter in the first example described above. The system can therefore include a set of emitter arranged according to an emitter pitch distance, configured to output illuminating beams of diameter, and offset behind the bulk transmitter optic by an offset distance as a function of a scale factor (e.g., 2.0 or 3.0) and 1) the aperture pitch distance in the aperture layer, 2) the diameter of apertures in the aperture layer, and 3) the focal length of bulk transmitter optic, respectively. The system can therefore include an illuminating subsystem that is proportionally larger than a corresponding receiver subsystem to achieve greater total output illumination power within the same beam angles and fields of view of corresponding channels in the receiver subsystem.
The system can also include multiple discrete sets of illumination sources, each set of illumination sources 110 paired with a discrete bulk transmitter optic adjacent the bulk imaging optic 130. For example, the system can include a first bulk transmitter optic, a second bulk transmitter optic, and a third bulk transmitter optic patterned radially about the bulk imaging optic 130 at a uniform radial distance from the center of the bulk imaging optic 130 and spaced apart by an angular distance of 120°. In this example, the system can include a laser array with one emitter—as described above—behind each of the first, second, and third bulk transmitter optics. Each discrete laser array and its corresponding bulk transmitter optic can thus project a set of illuminating beams into the fields of view of defined by corresponding in the apertures in the aperture layer. Therefore, in this example, the three discrete laser arrays and the three corresponding bulk transmitter optics can cooperate to project three times the power onto the fields of view of the sense channels in the system, as compared to a single laser array and one bulk transmitter optic. Additionally or alternatively, the system can include multiple discrete layer arrays and bulk transmitter optics to both: 1) achieve a target illumination power output into the field of view of each sensing channel in the receiver subsystem with multiple lower-power emitters per sensing channel; and 2) distribute optical energy over a larger area in the near-field to achieve an optical energy density less than a threshold allowable optical energy density for the human eye.
However, the system can include any other number and configuration of illumination source sets and bulk transmitter optics configured to illuminate fields of view defined by the sense channels. The set of illumination sources 110 can also include any other suitable type of optical transmitter, such as a 1×16 optical splitter powered by a single laser diode, a side-emitting laser diode array, an LED array, or a quantum dot LED array, etc.
1.8 Fabrication
In one implementation, the bulk receiver lens, the aperture layer, the set of lenses 150, the optical filter 160, and the diffuser 180 are fabricated and then aligned with and mounted onto an image sensor. For example, the optical filter 160 can be fabricated by coating a fused silica substrate. Photoactive optical polymer can then be deposited over the optical filter 160, and a lens mold can be placed over the photoactive optical polymer and a UV light source activated to cure the photoactive optical polymer in the form of lenses patterned across the optical filter 160. Standoffs can be similarly molded or formed across the optical filter 160 via photolithography techniques, and an aperture layer defined by a selectively-cured, metallized glass wafer can then be bonded or otherwise mounted to the standoffs to form the aperture layer. The assembly can then be inverted, and a set of discrete diffusers and standoffs can be similarly fabricated across the opposite side of the optical filter 160. A discrete image sensor can then be aligned with and bonded to the standoffs, and a bulk imaging optic 130 can be similarly mounted over the aperture layer.
Alternatively, photolithography and wafer level bonding techniques can be implemented to fabricate the bulk imaging optics, the aperture layer, the set of lenses 150, the optical filter 160, and the diffuser 180 directly on to the un-diced semiconductor wafer containing the detector chips in order to simplify manufacturing, reduce cost, and reduce optical stack height for decreased pixel crosstalk.
2. One-Dimensional Optical System: Lens Tube
One variation of the system includes: a set of illumination sources 110 arranged along a first axis, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in a field ahead of the illumination source; a bulk imaging optic 130 characterized by a focal plane opposite the field; a set of lens tubes 210 arranged in a line array parallel to the first axis, each lens tube in the set of lens tubes 210 including: a lens characterized by a focal length, offset from the focal plane by the focal length, and configured to collimate light rays reflected into the bulk imaging optic 130 from a discrete spot in the field illuminated by a corresponding illumination source in the set of optics into the bulk imaging optic 130; and a cylindrical wall 218 extending from the lens opposite the focal plane, defining a long axis substantially perpendicular to the first axis, and configured to absorb incident light rays reflected into the bulk imaging optic 130 from a region in the field outside the discrete spot illuminated by the corresponding illumination source. In this variation, the system also includes: an optical filter 160 adjacent the set of lens tubes 210 opposite the focal plane and configured to pass light rays at the operating wavelength; a set of pixels 170 adjacent the optical filter 160 opposite the set of lenses 150, each pixel in the set of pixels 170 corresponding to a lens in the set of lenses 150 and including a set of subpixels aligned along a third axis perpendicular to the first axis; and a diffuser 180 interposed between the optical filter 160 and the set of pixels 170 and configured to spread collimated light output from each lens in the set of lenses 150 across a set of subpixels of a corresponding pixel in the set of pixels 170.
Generally, in this variation, the system includes a lens tube in replacement of (or in addition to) each aperture and lens pair described above. In this variation, each lens tube can be characterized by a second (short) focal length and can be offset from the focal plane of the bulk imaging optic 130 by the second focal length to preserve the aperture of the bulk imaging optic 130 and to collimate incident light received from the bulk imaging optic 130, as described above and as shown in FIGS. 5 and 7.
Each lens tube also defines an opaque cylindrical wall 218 defining an axis normal to the incidence plane of the adjacent optical filter 160 and configured to absorb incident light rays, as shown in FIG. 5. Generally, at greater axial lengths, the cylindrical wall 218 of a lens tube may absorb light rays passing through the lens tube at shallower angles to the axis of the lens tube, thereby reducing the field of view of the lens tube (which may be similar to decreasing the diameter of an aperture in the aperture layer up to the diffraction-limited diameter, as described above) and yielding an output signal of collimated light rays nearer to perpendicular to the incidence plane of the optical filter 160. Each lens tube can therefore define an elongated cylindrical wall 218 of length sufficient to achieve a target field of view and to pass collimated light rays at maximum angles to the axis of the lens tube less than a threshold angle. In this variation, a lens tube can thus function as an aperture-sense pair described above to define a narrow field of view and to output substantially collimated light to the adjacent optical filter 160.
The cylindrical wall 218 of a lens tube can define a coarse or patterned opaque interface about a transparent (or translucent) lens material, as shown in FIG. 5, to increase absorption and decrease reflection of light rays incident on the cylindrical wall 218. Each lens tube (and each lens described above) can also be coated with an anti-reflective coating.
As shown in FIG. 9, in this variation, the set of lens tubes 210 can be fabricated by implementing photolithography techniques to pattern a photoactive optical polymer (e.g., SU8) onto the optical filter 160 (e.g., on a silicon wafer defining the optical filter). A light-absorbing polymer can then be poured between the lens tubes and cured. A set of lenses 150 can then be fabricated (e.g., molded) separately and then bonded over the lens tubes. Alternatively, lenses can be fabricated directly onto the lens tubes by photolithography techniques. Yet alternatively, a mold for lenses can be cast directly onto the lens tubes by injecting polymer into a mold arranged over the lens tubes. A singular diffuser 180 or multiple discrete diffusers 180 can be similarly fabricated and/or assembled on the optical filter 160 opposite the lens tubes. Standoffs extending from the optical filter 160 can be similarly fabricated or installed around the diffuser(s) 180, and the image sensor can be aligned with and bonded to the standoffs opposite the optical filter 160. Other optical elements within the system (e.g., the bulk imaging lens, the bulk transmitting lens, etc.) can be fabricated according to similar techniques and with similar materials.
3. Two-Dimensional Optical System
Another variation of the system includes: a set of illumination sources 110 arranged in a first rectilinear grid array, each illumination source in the set of illumination sources 110 configured to output an illuminating beam of an operating wavelength toward a discrete spot in a field ahead of the illumination source; a bulk imaging optic 130 characterized by a focal plane opposite the field; an aperture layer coincident the focal plane, defining a set of apertures 144 in a second rectilinear grid array proportional to the first rectilinear grid array, and defining a stop region 146 around the set of apertures 144, each aperture in the set of apertures 144 defining a field of view in the field coincident a discrete spot output by a corresponding illumination source in the set of illumination sources 110, the stop region 146 absorbing light rays reflected from surfaces in the field outside of fields of view defined by the set of apertures 144 and passing through the bulk imaging optic 130; a set of lenses 150, each lens in the set of lenses 150 characterized by a second focal length, offset from the focal plane opposite the bulk imaging optic 130 by the second focal length, aligned with an aperture in the set of apertures 144, and configured to collimate light rays passed by the aperture; an optical filter 160 adjacent the set of lenses 150 opposite the aperture layer and configured to pass light rays at the operating wavelength; a set of pixels 170 adjacent the optical filter 160 opposite the set of lenses 150, each pixel in the set of pixels 170 aligned with a subset of lenses in the set of lenses 150; and a diffuser 180 interposed between the optical filter 160 and the set of pixels 170 and configured to spread collimated light output from each lens in the set of lenses 150 across a corresponding pixel in the set of pixels 170.
Generally, in this variation, the system includes a two-dimensional grid array of channels (i.e., aperture, lens, and pixel sets or lens tube and pixel sets) and is configured to image a volume occupied by the system in two dimensions. The system can collect one-dimensional distance data—such as counts of incident photons within a sampling period and/or times between consecutive photons incident on pixels of known position corresponding to known fields of view in the field—across a two-dimensional field. The one-dimensional distance data can then be merged with known positions of the fields of view for each channel in the system to reconstruct a virtual three-dimensional representation of the field ahead of the system.
In this variation, the aperture layer can define a grid array of apertures, the set of lenses 150 can be arranged in a similar grid array with one lens aligned with one aperture in the aperture layer, and the set of pixels 170 can include one pixel per aperture and lens pair, as described above. For example, the aperture layer can define a 24-by-24 grid array of 200-μm-diameter apertures offset vertically and laterally by an aperture pitch distance of 300 μm, and the set of lenses 150 can similarly define a 24-by-24 grid array of lenses offset vertically and laterally by a lens pitch distance of 300 μm. In this example, the set of pixels 170 can include a 24-by-24 grid array of 300-μm-square pixels, wherein each pixel includes a 3×3 square array of nine 100-μm-square SPADs.
Alternatively, in this variation, the set of pixels 170 can include one pixel per group of multiple aperture and lens pairs. In the foregoing example, the set of pixels 170 can alternatively include a 12-by-12 grid array of 600-μm-square pixels, wherein each pixel includes a 6×6 square array of 36 100-μm-square SPADs and wherein each pixel is aligned with a group of four adjacent lenses in a square grid. In this example, for each group of four adjacent lenses, the diffuser 180: can bias collimated light rays output from a lens in the (1,1) position in the square grid upward and to the right to spread light rays passing through the (1,1) lens across the full breadth and width of the corresponding pixel; can bias collimated light rays output from a lens in the (2,1) position in the square grid upward and to the left to spread light rays passing through the (2,1) lens across the full breadth and width of the corresponding pixel; can bias collimated light rays output from a lens in the (1,2) position in the square grid downward and to the right to spread light rays passing through the (1,2) lens across the full breadth and width of the corresponding pixel; and can bias collimated light rays output from a lens in the (2,2) position in the square grid downward and to the left to spread light rays passing through the (2,2) lens across the full breadth and width of the corresponding pixel, as shown in FIG. 8.
In the foregoing example, for each group of four illumination sources in a square grid and corresponding to one group of four lenses in a square grid, the system can actuate one illumination source in the group of four illumination sources at any given instance in time. In particular, for each group of four illumination sources in a square grid corresponding to one pixel in the set of pixels 170, the system can actuate a first illumination source 111 in a (1,1) position during a first sampling period to illuminate a field of view defined by a first aperture 141 corresponding to a lens in the (1,1) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the first sampling period. The system can then shut down the first illumination source 111 and actuate a second illumination source 112 in a (1,2) position during a subsequent second sampling period to illuminate a field of view defined by a second aperture 142 corresponding to a lens in the (1,2) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the second sampling period. Subsequently, the system can then shut down the first and second illumination sources 112 and actuate a third illumination source in a (2,1) position during a subsequent third sampling period to illuminate a field of view defined by a third aperture corresponding to a lens in the (2,1) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the third sampling period. Finally, the system can shut down the first, second, and third illumination sources and actuate a fourth illumination source in a (2,2) position during a fourth sampling period to illuminate a field of view defined by a fourth aperture corresponding to a lens in the (2,2) position in the corresponding group of four lenses, and the system can sample all 36 SPADs in the corresponding pixel during the fourth sampling period. The system can repeat this process throughout its operation.
Therefore, in the foregoing example, the system can include a set of pixels 170 arranged across an image sensor 7.2 mm in width and 7.2 mm in length and can implement a scanning schema such that each channel in the system can access (can project light rays onto) a number of SPADs otherwise necessitating a substantially larger image sensor (e.g., a 14.4 mm by 14.4 mm image sensor). In particular, the system can implement a serial scanning schema per group of illumination sources to achieve an exponential increase in the dynamic range of each channel in the system. In particular, in this variation, the system can implement the foregoing imaging techniques to increase imaging resolution of the system.
In the foregoing implementation, the system can also include a shutter 182 between each channel and the image sensor, and the system can selectively open and close each shutter 182 when the illumination source for the corresponding channel is actuated and deactivated, respectively. For example, the system can include one independently-operable electrochromic shutter 182 interposed between each lens, and the system can open the electrochromic shutter 182 over the (1,1) lens in the square-gridded group of four lenses and close electrochromic shutters 182 over the (1,2), (2,1), and (2,2) lens when the (1,1) illumination source is activated, thereby rejecting noise passing through the (1,2), (2,1), and (2,2) lens from reaching the corresponding pixel on the image sensor. The system can therefore selectively open and close shutters 182 between each channel and the image sensor to increase SNR per channel during operation. Alternatively, the system can include one independently-operable electrochromic shutter 182 arranged over select regions of each pixel, as shown in FIG. 8, wherein each electrochromic shutter 182 is aligned with a single channel (i.e., with a single lens in the set of lenses). The system can alternatively include MEMS mechanical shutters or any other suitable type of shutter interposed between the set of lenses 150 and the image sensor.
In this variation, the system can define two-dimension grid arrays of apertures, lenses, diffusers, and/or pixels characterized by a first pitch distance along a first (e.g., X) axis and a second pitch distance—different from the first pitch distance—along a second (e.g., Y) axis. For example, the image sensor can include pixels offset by a 25 μm horizontal pitch and a 300 μm vertical pitch, wherein each pixel includes a single row of twelve subpixels.
However, in this variation, the two-dimensional optical system can include an array of any other number and pattern of channels (e.g., apertures, lenses (or lens tubes), and diffusers) and pixels and can execute any other suitable scanning schema to achieve higher spatial resolutions per channel than the raw pixel resolution of the image sensor. The system can additionally or alternatively include a converging optic, a diverging optic, and/or any other suitable type of optical element to spread light rights passed from a channel across the breadth of a corresponding pixel.
As a person skilled in the art will recognize from the previous detailed description and from the figures and claims, modifications and changes can be made to the embodiments of the invention without departing from the scope of this invention as defined in the following claims.

Claims (50)

What is claimed is:
1. An optical imaging system comprising:
a bulk imaging optic having a focal plane;
an array of pixels; and
an aperture layer disposed between the bulk imaging optic and the array of pixels, the aperture layer including a plurality of discrete apertures coincident with the focal plane, wherein the aperture layer and pixel layer are arranged to form a plurality of sense channels defining a plurality of discrete, non-overlapping fields of view beyond a threshold distance in a field external to the optical imaging system, each sense channel in the plurality of sense channels including a pixel from the array of pixels and an aperture from the plurality of apertures that defines the field of view for its respective pixel, and wherein a pixel area of each pixel in the plurality of sense channels is larger than an area of its corresponding aperture in the plurality of apertures.
2. The optical imaging system of claim 1 wherein each pixel in the array of pixels comprises a plurality of subpixels distributed across its pixel area.
3. The optical imaging system of claim 2 wherein the plurality of subpixels in each pixel is a plurality of SPADs.
4. The optical imaging system of claim 1 further comprising, within each channel in the plurality of sense channels, an optical filter disposed between the bulk imaging optic and the array of pixels, the optical filter configured to receive light passed through the bulk imaging optic and pass a narrow band of radiation while blocking radiation outside the band.
5. The optical imaging system of claim 1 further comprising:
a bulk transmitter optic; and
an illumination source comprising a plurality of emitters corresponding in number to the plurality of sense channels, each emitter in the plurality of emitters configured to project an illuminating beam at an operating wavelength through the bulk transmitter optic into a field of view that is substantially the same size and geometry as a field of view defined by the bulk imaging optic and a corresponding aperture in the plurality of sense channels.
6. The optical imaging system of claim 5 wherein each emitter in the plurality of emitters comprises a vertical-cavity surface-emitting laser (VCSEL).
7. The optical imaging system of claim 5 wherein a diameter of each illuminating beam projected from each emitter in the illumination source is substantially similar to a cross-section to a field of view of its respective sense channel at various distances from the optical imaging system.
8. The optical imaging system of claim 6 further comprising, within each channel in the plurality of sense channels, an optical filter disposed between the bulk imaging optic and the array of pixels, the optical filter configured to receive light passed through the bulk imaging optic and pass a narrow band of radiation that includes the operating wavelength of the plurality of lasers to the plurality of pixels while blocking radiation outside the band.
9. The optical imaging system of claim 8 further comprising, within each channel in the plurality of sense channels, a diffuser disposed within the channel between the aperture and the pixel.
10. The optical imaging system of claim 5 wherein each emitter in the plurality of emitters comprises a vertical-cavity surface-emitting laser VCSEL), each pixel in the array of pixels comprises a plurality of SPADs and an optical filter is disposed between the bulk imaging optic and the array of pixels, the optical filter configured to receive light passed through the bulk imaging optic and pass a narrow band of radiation that includes the operating wavelength of the plurality of VCSELs to the array of pixels while blocking radiation outside the band.
11. The optical imaging system of claim 1 further comprising a plurality of lenses, wherein each channel in the plurality of channels includes a lens from the plurality of lenses disposed between the aperture of the channel and its corresponding pixel.
12. The optical imaging system of claim 1 wherein adjacent apertures in the aperture layer are offset by an aperture pitch distance greater than a diameter of each aperture.
13. An optical imaging system comprising:
a bulk imaging optic having a focal plane;
an array of pixels of substantially uniform pixel area, each pixel in the array of pixels including a plurality of subpixels distributed across its pixel area; and
an aperture layer disposed between the bulk imaging optic and the array of pixels, the aperture layer including a plurality of discrete apertures of substantially uniform area formed coincident with the focal plane and stop regions interposed between adjacent apertures in the aperture layer, wherein the aperture layer and pixel layer are arranged to form a plurality of sense channels defining a plurality of discrete, non-overlapping fields of view beyond a threshold distance in a field external to the optical imaging system, each sense channel in the plurality of sense channels including an aperture from the plurality of apertures and a pixel from the array of pixels with the aperture of each channel defining the field of view for its respective pixel, and wherein an area of the aperture in each channel is smaller than a pixel area of its respective pixel.
14. The optical imaging system of claim 13 wherein the plurality of subpixels in each pixel is a plurality of SPADs.
15. The optical imaging system of claim 13 further comprising:
a bulk transmitter optic; and
an illumination source comprising a plurality of lasers corresponding in number to the plurality of sense channels, each laser in the plurality of lasers configured to project a discrete illuminating beam at an operating wavelength through the bulk transmitter optic into a field of view that is substantially the same size and geometry as a field of view defined by the bulk imaging optic and a corresponding aperture in the plurality of sense channels.
16. The optical imaging system of claim 15 wherein each aperture in the plurality of discrete apertures defines a field of view in the field coincident a discrete spot output by a corresponding laser in the plurality of lasers.
17. An optical imaging system comprising:
a bulk receiver imaging optic having a focal plane;
an optical assembly having a plurality of sense channels defining a plurality of discrete, non-overlapping fields of view beyond a threshold distance in a field external to the bulk imaging optic, the optical assembly comprising:
an array of pixels of substantially uniform pixel area, each pixel in the array of pixels including a plurality of subpixels distributed across its pixel area;
an aperture layer disposed between the bulk receiver optic and the array of pixels, the aperture layer including a plurality of discrete apertures of substantially uniform area formed coincident with the focal plane and stop regions between adjacent apertures;
a lens layer including a plurality of lenses arranged to receive light passed through the aperture and pass the received light towards the array of pixels; and
an optical filter layer disposed between the bulk imaging optic and the array of pixels, the optical filter layer configured to receive light passed through the bulk imaging optic and pass a narrow band of radiation while blocking radiation outside that band;
wherein each sense channel in the optical assembly includes a pixel from the array of pixels, an aperture from the aperture layer that defines the field of view for its respective pixel and has an area that is smaller than a sensing area of its respective pixel, a lens from the lens layer and a filter from the optical filter layer.
18. The optical imaging system of claim 17 wherein adjacent apertures in the aperture layer are offset by an aperture pitch distance greater than a diameter of each aperture and substantially equal to a pitch between adjacent lenses in the lens layer.
19. The optical imaging system of claim 18 wherein the optical filter layer is disposed between the lens layer and the array of pixels and wherein each lens is configured to collimate light rays passed by its respective aperture so that light rays incident on the optical filter meet the optical filter at an angle of incidence of approximately 0 degrees.
20. The optical imaging system of claim 18 further comprising:
a bulk transmitter optic; and
an illumination source comprising a plurality of emitters corresponding in number to the plurality of sense channels, each emitter in the plurality of emitters configured to project an illuminating beam at an operating wavelength through the bulk transmitter optic into a field of view that overlaps with a field of view defined by the bulk imaging optic and a corresponding aperture in the plurality of sense channels.
21. The optical imaging system of claim 20 wherein the plurality of subpixels in each pixel is a plurality of SPADs and where the plurality of emitters in the illumination source is a plurality of vertical-cavity surface-emitting lasers (VCSELs).
22. The optical imaging system of claim 21 wherein each aperture in the plurality of discrete apertures has a diameter of 200 microns or less.
23. The optical imaging system of claim 21 wherein each aperture in the plurality of discrete apertures has a diameter approaching the diffraction-limited diameter for the plurality of VCSELs.
24. The optical imaging system of claim 21 wherein each aperture in the plurality of discrete apertures has a diameter less than the diffraction-limited diameter for the operating wavelength of the plurality of emitters.
25. The optical imaging system of claim 21 wherein each aperture in the plurality of discrete apertures has a diameter matched to a dynamic range of its corresponding pixel.
26. The optical imaging system of claim 21 wherein the stop region surrounding each aperture in the plurality of discrete apertures blocks light rays reflected from a region of the surface outside the field of view of its corresponding aperture.
27. An optical imaging system comprising:
a bulk imaging optic having a focal plane;
a plurality of pixels arranged in a pixel array in which adjacent pixels in the array are spaced apart from each other in a first dimension by a pixel pitch; and
an aperture layer disposed between the bulk imaging optic and the plurality of pixels arranged in the pixel array, the aperture layer including a plurality of discrete apertures of substantially uniform size coincident with the focal plane and stop regions between adjacent apertures, wherein the plurality of discrete apertures are arranged in an aperture array in which adjacent apertures in the aperture array are spaced apart from each other in the first dimension by an aperture pitch and wherein the aperture array is aligned with the pixel array and a maximum linear dimension of each aperture in the aperture array is less than the pixel pitch; and
wherein the aperture layer and pixel layer are arranged to form a plurality of sense channels defining a plurality of discrete, non-overlapping fields of view beyond a threshold distance in a field external to the optical imaging system, each sense channel in the plurality of sense channels including a pixel from the plurality of pixels and an aperture from the plurality of apertures that defines the field of view for its respective pixel.
28. The optical imaging system of claim 27 wherein each aperture in the plurality of discrete apertures is a circular aperture of substantially uniform diameter.
29. The optical imaging system of claim 27 wherein the aperture pitch is substantially equal to the pixel pitch.
30. The optical imaging system of claim 27 wherein each pixel in the plurality of pixels comprises a plurality of subpixels distributed across a pixel area.
31. The optical imaging system of claim 30 wherein the plurality of subpixels in each pixel is a plurality of SPADs.
32. The optical imaging system of claim 27 further comprising, within each channel in the plurality of sense channels, an optical filter disposed between the bulk imaging optic and the pixel within the channel, the optical filter configured to receive light passed through the bulk imaging optic and pass a narrow band of radiation while blocking radiation outside the band.
33. The optical imaging system of claim 32 further comprising:
a bulk transmitter optic; and
an illumination source comprising a plurality of emitters corresponding in number to the plurality of sense channels, each emitter in the plurality of emitters configured to project electromagnetic radiation at an operating wavelength through the bulk transmitter optic into a field of view that is aligned with and overlaps with a field of view defined by the bulk imaging optic and a corresponding aperture in the plurality of sense channels; and
wherein the narrow band of radiation passed by the optical filter includes the operating wavelength.
34. The optical imaging system of claim 33 wherein each emitter in the plurality of emitters comprises a laser.
35. An optical imaging system comprising:
a bulk imaging optic having a focal plane;
an array of pixels; and
an aperture layer disposed between the bulk imaging optic and the array of pixels, the aperture layer including a plurality of discrete apertures coincident with the focal plane, wherein the aperture layer and pixel layer are arranged to form a plurality of sense channels defining a plurality of discrete, non-overlapping fields of view beyond a threshold distance in a field external to the optical imaging system, each sense channel in the plurality of sense channels including a pixel from the array of pixels and an aperture from the plurality of apertures that defines the field of view for its respective pixel.
36. The optical imaging system of claim 35 further comprising a plurality of lenses disposed between the aperture layer and the array of pixels with each lens in the plurality of lenses being characterized by a second focal length, and wherein each channel in the plurality of channels includes a lens from the plurality of lenses offset from the bulk imaging optic focal plane by the second focal length and aligned with a corresponding aperture in the channel.
37. The optical system of claim 36 further comprising, within each channel in the plurality of sense channels, an optical filter disposed between the bulk imaging optic and the array of pixels, the optical filter configured to receive light passed through the bulk imaging optic and pass a narrow band of radiation to the array of pixels while blocking radiation outside the band.
38. The optical imaging system of claim 36 wherein, within each channel, the lens from the plurality of lenses is disposed between an aperture and optical filter and is configured to collimate light rays passed by the aperture and to pass the collimated light rays into the optical filter.
39. The optical imaging system of claim 35 wherein the pixel in each of the plurality of sense channels comprises a plurality of SPADs distributed across a pixel area of the pixel that is larger than an area of its corresponding aperture.
40. The optical imaging system of claim 39 further comprising, within each channel in the plurality of sense channels, a diffuser disposed within the channel between the aperture and its corresponding pixel, the diffuser configured to spread light rays passed by the aperture across the plurality of SPADs distributed across the pixel area of the pixel.
41. The optical imaging system of claim 35 wherein each pixel in the array of pixels comprises a plurality of SPADs.
42. The optical imaging system of claim 41 wherein each pixel in the array of pixels is operable to output a signal corresponding to a count of photons incident on the pixel within one sample period.
43. The optical imaging system of claim 35 further comprising:
a bulk transmitting optic;
a plurality of emitters, each emitter in the plurality of emitters configured to output electromagnetic radiation of an operating wavelength through the bulk transmitting optic and into a field of view in the field external to the optical imaging system coincident a field of view of a defined by a corresponding aperture in the plurality of discrete apertures.
44. An optical imaging system comprising:
a bulk imaging optic;
a plurality of pixels arranged behind the bulk imaging optic, wherein each pixel in the plurality of pixels comprises a plurality of single photon avalanche diodes (SPADs) and has a field of view that is non-overlapping beyond a threshold distance in a field external to the optical imaging system with fields of view of all other pixels in the plurality of pixels, and wherein the plurality of pixels includes at least a first pixel and a second pixel;
an aperture layer disposed between the bulk imaging optic and the plurality of pixels, the aperture layer comprising a plurality of apertures separated from each other by a stop region, the plurality of apertures including a first aperture disposed in a first optical path extending between the bulk imaging optic and the first pixel and a second aperture disposed in a second optical path extending between the bulk imaging optic and the second pixel;
a bulk transmitting optic disposed adjacent to the bulk imaging optic; and
a plurality of emitters arranged to project electromagnetic radiation through the bulk transmitting optic and into the field according to an illumination pattern where electromagnetic radiation output from each emitter in the illumination pattern is coincident with a field of view of a corresponding pixel in the plurality of pixels.
45. The optical imaging system of claim 44 wherein the plurality of pixels further includes third and fourth pixels, and the plurality of apertures further includes a third aperture disposed in a third optical path extending between the bulk imaging optic and the third pixel and a fourth aperture disposed in a fourth optical path extending between the bulk imaging optic and the fourth pixel.
46. The optical imaging system of claim 44 wherein the plurality of emitters comprises a plurality of vertical-cavity surface-emitting lasers (VCSELs).
47. The optical imaging system of claim 46 wherein each emitter in the plurality of emitters comprises a single VCSEL.
48. The optical imaging system of claim 44 wherein each pixel in the plurality of pixels is operable to output a signal corresponding to a count of photons incident on the pixel within one sample period.
49. The optical imaging system of claim 44 wherein the bulk transmitting optic and the bulk imaging optic are each image-space telecentric lenses.
50. The optical imaging system of claim 44 wherein the optical imaging system is configured to collect distance data based on recorded times between transmission of beams of electromagnetic radiation emitted from the plurality of emitters and detection of photons originating from the beams of electromagnetic radiation.
US16/046,643 2015-09-24 2018-07-26 Optical imaging system with a plurality of sense channels Active 2038-02-25 US11190750B2 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US16/046,643 US11190750B2 (en) 2015-09-24 2018-07-26 Optical imaging system with a plurality of sense channels
US16/584,515 US11202056B2 (en) 2015-09-24 2019-09-26 Optical system with multiple light emitters sharing a field of view of a pixel detector
US17/549,675 US11627298B2 (en) 2015-09-24 2021-12-13 Optical system for collecting distance information within a field
US18/179,559 US20230319253A1 (en) 2015-09-24 2023-03-07 Optical system for collecting distance information within a field

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201562232222P 2015-09-24 2015-09-24
US15/276,532 US9992477B2 (en) 2015-09-24 2016-09-26 Optical system for collecting distance information within a field
US15/880,491 US11025885B2 (en) 2015-09-24 2018-01-25 Optical system for collecting distance information within a field
US16/046,643 US11190750B2 (en) 2015-09-24 2018-07-26 Optical imaging system with a plurality of sense channels

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US15/880,491 Continuation US11025885B2 (en) 2015-09-24 2018-01-25 Optical system for collecting distance information within a field

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/584,515 Continuation US11202056B2 (en) 2015-09-24 2019-09-26 Optical system with multiple light emitters sharing a field of view of a pixel detector

Publications (2)

Publication Number Publication Date
US20180359460A1 US20180359460A1 (en) 2018-12-13
US11190750B2 true US11190750B2 (en) 2021-11-30

Family

ID=59960008

Family Applications (8)

Application Number Title Priority Date Filing Date
US15/276,532 Active 2036-10-15 US9992477B2 (en) 2015-09-24 2016-09-26 Optical system for collecting distance information within a field
US15/880,491 Active 2038-09-01 US11025885B2 (en) 2015-09-24 2018-01-25 Optical system for collecting distance information within a field
US16/046,643 Active 2038-02-25 US11190750B2 (en) 2015-09-24 2018-07-26 Optical imaging system with a plurality of sense channels
US16/584,515 Active US11202056B2 (en) 2015-09-24 2019-09-26 Optical system with multiple light emitters sharing a field of view of a pixel detector
US17/317,809 Active US11196979B2 (en) 2015-09-24 2021-05-11 Optical system for collecting distance information within a field
US17/323,962 Active US11178381B2 (en) 2015-09-24 2021-05-18 Optical system for collecting distance information within a field
US17/549,675 Active US11627298B2 (en) 2015-09-24 2021-12-13 Optical system for collecting distance information within a field
US18/179,559 Pending US20230319253A1 (en) 2015-09-24 2023-03-07 Optical system for collecting distance information within a field

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US15/276,532 Active 2036-10-15 US9992477B2 (en) 2015-09-24 2016-09-26 Optical system for collecting distance information within a field
US15/880,491 Active 2038-09-01 US11025885B2 (en) 2015-09-24 2018-01-25 Optical system for collecting distance information within a field

Family Applications After (5)

Application Number Title Priority Date Filing Date
US16/584,515 Active US11202056B2 (en) 2015-09-24 2019-09-26 Optical system with multiple light emitters sharing a field of view of a pixel detector
US17/317,809 Active US11196979B2 (en) 2015-09-24 2021-05-11 Optical system for collecting distance information within a field
US17/323,962 Active US11178381B2 (en) 2015-09-24 2021-05-18 Optical system for collecting distance information within a field
US17/549,675 Active US11627298B2 (en) 2015-09-24 2021-12-13 Optical system for collecting distance information within a field
US18/179,559 Pending US20230319253A1 (en) 2015-09-24 2023-03-07 Optical system for collecting distance information within a field

Country Status (11)

Country Link
US (8) US9992477B2 (en)
JP (1) JP7214629B2 (en)
KR (2) KR102425304B1 (en)
CN (2) CN109983312B (en)
CA (1) CA3038038A1 (en)
DE (2) DE202017007509U1 (en)
DK (1) DK180852B1 (en)
GB (2) GB2601435B (en)
IL (1) IL265562B2 (en)
SE (1) SE544865C2 (en)
WO (1) WO2018057084A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210041570A1 (en) * 2017-05-15 2021-02-11 Ouster, Inc. Augmenting panoramic lidar results with color
US11422236B2 (en) 2016-08-24 2022-08-23 Ouster, Inc. Optical system for collecting distance information within a field
US11627298B2 (en) 2015-09-24 2023-04-11 Ouster, Inc. Optical system for collecting distance information within a field
US11789128B2 (en) 2021-03-01 2023-10-17 Innovusion, Inc. Fiber-based transmitter and receiver channels of light detection and ranging systems

Families Citing this family (95)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11609336B1 (en) 2018-08-21 2023-03-21 Innovusion, Inc. Refraction compensation for use in LiDAR systems
US10063849B2 (en) 2015-09-24 2018-08-28 Ouster, Inc. Optical system for collecting distance information within a field
AU2017212835B2 (en) 2016-01-29 2019-03-14 Ouster, Inc. Systems and methods for calibrating an optical distance sensor
US10761195B2 (en) 2016-04-22 2020-09-01 OPSYS Tech Ltd. Multi-wavelength LIDAR system
JP2019523422A (en) * 2016-08-04 2019-08-22 オフィール オプトロニクス ソリューションズ リミテッド Photometric test system for light emitting devices
US11275155B1 (en) * 2016-11-08 2022-03-15 Lockheed Martin Coherent Technologies, Inc. Laser-array lidar devices
US10605984B2 (en) * 2016-12-01 2020-03-31 Waymo Llc Array of waveguide diffusers for light detection using an aperture
US10502618B2 (en) 2016-12-03 2019-12-10 Waymo Llc Waveguide diffuser for light detection using an aperture
JP7088937B2 (en) 2016-12-30 2022-06-21 イノビュージョン インコーポレイテッド Multi-wavelength rider design
US10942257B2 (en) 2016-12-31 2021-03-09 Innovusion Ireland Limited 2D scanning high precision LiDAR using combination of rotating concave mirror and beam steering devices
US11009605B2 (en) 2017-01-05 2021-05-18 Innovusion Ireland Limited MEMS beam steering and fisheye receiving lens for LiDAR system
DE102017101945A1 (en) * 2017-02-01 2018-08-02 Osram Opto Semiconductors Gmbh Measuring arrangement with an optical transmitter and an optical receiver
EP3596492A4 (en) 2017-03-13 2020-12-16 Opsys Tech Ltd Eye-safe scanning lidar system
CN110603461B (en) * 2017-05-11 2021-10-22 华为技术有限公司 Time-of-flight device
AU2018269000B2 (en) 2017-05-15 2021-03-11 Ouster, Inc. Optical imaging transmitter with brightness enhancement
SG11201913642VA (en) 2017-07-05 2020-01-30 Ouster Inc Light ranging device with electronically scanned emitter array and synchronized sensor array
KR102218679B1 (en) 2017-07-28 2021-02-23 옵시스 테크 엘티디 VCSEL Array LIDAR Transmitter with Small Angle Divergence
US10791283B2 (en) * 2017-09-01 2020-09-29 Facebook Technologies, Llc Imaging device based on lens assembly with embedded filter
US10785400B2 (en) * 2017-10-09 2020-09-22 Stmicroelectronics (Research & Development) Limited Multiple fields of view time of flight sensor
WO2019079642A1 (en) 2017-10-19 2019-04-25 Innovusion Ireland Limited Lidar with large dynamic range
US11802943B2 (en) 2017-11-15 2023-10-31 OPSYS Tech Ltd. Noise adaptive solid-state LIDAR system
JP2019101244A (en) * 2017-12-04 2019-06-24 富士通株式会社 Optical module
JP7120756B2 (en) * 2017-12-05 2022-08-17 シャープ株式会社 Photodetector, time-of-flight measuring device and optical radar device
US10969490B2 (en) 2017-12-07 2021-04-06 Ouster, Inc. Light ranging system with opposing circuit boards
US11493601B2 (en) 2017-12-22 2022-11-08 Innovusion, Inc. High density LIDAR scanning
US11675050B2 (en) 2018-01-09 2023-06-13 Innovusion, Inc. LiDAR detection systems and methods
WO2019165130A1 (en) 2018-02-21 2019-08-29 Innovusion Ireland Limited Lidar detection systems and methods with high repetition rate to observe far objects
WO2019164961A1 (en) 2018-02-21 2019-08-29 Innovusion Ireland Limited Lidar systems with fiber optic coupling
US11422234B2 (en) 2018-02-23 2022-08-23 Innovusion, Inc. Distributed lidar systems
WO2020013890A2 (en) 2018-02-23 2020-01-16 Innovusion Ireland Limited Multi-wavelength pulse steering in lidar systems
US11567182B2 (en) 2018-03-09 2023-01-31 Innovusion, Inc. LiDAR safety systems and methods
KR102506579B1 (en) 2018-04-01 2023-03-06 옵시스 테크 엘티디 Noise Adaptive Solid-State LIDAR System
US11789132B2 (en) 2018-04-09 2023-10-17 Innovusion, Inc. Compensation circuitry for lidar receiver systems and method of use thereof
WO2019199775A1 (en) 2018-04-09 2019-10-17 Innovusion Ireland Limited Lidar systems and methods for exercising precise control of a fiber laser
DE102018109544A1 (en) * 2018-04-20 2019-10-24 Sick Ag Optoelectronic sensor and method for distance determination
WO2019241396A1 (en) 2018-06-15 2019-12-19 Innovusion Ireland Limited Lidar systems and methods for focusing on ranges of interest
DE102018118653B4 (en) 2018-08-01 2020-07-30 Sick Ag Optoelectronic sensor and method for detecting an object
CN112543875A (en) * 2018-08-03 2021-03-23 欧普赛斯技术有限公司 Distributed modular solid-state LIDAR system
US10739189B2 (en) * 2018-08-09 2020-08-11 Ouster, Inc. Multispectral ranging/imaging sensor arrays and systems
US10760957B2 (en) 2018-08-09 2020-09-01 Ouster, Inc. Bulk optics for a scanning array
US11579300B1 (en) 2018-08-21 2023-02-14 Innovusion, Inc. Dual lens receive path for LiDAR system
US11860316B1 (en) 2018-08-21 2024-01-02 Innovusion, Inc. Systems and method for debris and water obfuscation compensation for use in LiDAR systems
US11614526B1 (en) 2018-08-24 2023-03-28 Innovusion, Inc. Virtual windows for LIDAR safety systems and methods
US11796645B1 (en) 2018-08-24 2023-10-24 Innovusion, Inc. Systems and methods for tuning filters for use in lidar systems
US11579258B1 (en) 2018-08-30 2023-02-14 Innovusion, Inc. Solid state pulse steering in lidar systems
US20200088512A1 (en) * 2018-09-18 2020-03-19 Shenzhen GOODIX Technology Co., Ltd. Depth information construction system, associated electronic device, and method for constructing depth information
CN113167866A (en) 2018-11-14 2021-07-23 图达通爱尔兰有限公司 LIDAR system and method using polygon mirror
US10852434B1 (en) * 2018-12-11 2020-12-01 Facebook Technologies, Llc Depth camera assembly using fringe interferometery via multiple wavelengths
US11585906B2 (en) 2018-12-26 2023-02-21 Ouster, Inc. Solid-state electronic scanning laser array with high-side and low-side switches for increased channels
DE112020000407B4 (en) 2019-01-10 2024-02-15 Innovusion, Inc. LIDAR SYSTEMS AND METHODS WITH RADIATION DEFLECTION AND WIDE ANGLE SIGNAL DETECTION
US11486970B1 (en) 2019-02-11 2022-11-01 Innovusion, Inc. Multiple beam generation from a single source beam for use with a LiDAR system
EP3699640B1 (en) 2019-02-19 2022-01-26 Sick Ag Optoelectronic sensor and method for detecting an object
JP2020154008A (en) * 2019-03-18 2020-09-24 キヤノン株式会社 Lens device and imaging apparatus
JP2020153796A (en) 2019-03-19 2020-09-24 株式会社リコー Distance measuring device and method for measuring distance
JP2022526998A (en) 2019-04-09 2022-05-27 オプシス テック リミテッド Solid-state LIDAR transmitter with laser control
EP3948344A4 (en) * 2019-05-01 2022-12-21 Sense Photonics, Inc. Event driven shared memory pixel
US11531111B2 (en) 2019-05-21 2022-12-20 Northrop Grumman Systems Corporation 360 degrees field of view scanning lidar with no movable parts
US11448732B2 (en) 2019-05-21 2022-09-20 Northrop Grumman Systems Corporation Frequency modulated scanning LIDAR with 360 degrees field of view
US11555891B2 (en) 2019-05-21 2023-01-17 Northrop Grumman Systems Corporation Methods for large angle field of view scanning LIDAR with no movable parts
WO2020242834A1 (en) 2019-05-30 2020-12-03 OPSYS Tech Ltd. Eye-safe long-range lidar system using actuator
KR102538137B1 (en) 2019-06-10 2023-05-31 옵시스 테크 엘티디 Eye-safe long-range solid-state LIDAR system
US10861819B1 (en) * 2019-07-05 2020-12-08 Asm Technology Singapore Pte Ltd High-precision bond head positioning method and apparatus
US10701326B1 (en) * 2019-07-10 2020-06-30 Lightspace Technologies, SIA Image display system, method of operating image display system and image projecting device
KR20210020469A (en) 2019-08-14 2021-02-24 삼성전자주식회사 Spectral camera
DE102019126982A1 (en) 2019-10-08 2021-04-08 Sick Ag Optoelectronic sensor and method for detecting objects
KR20210046971A (en) 2019-10-21 2021-04-29 삼성전자주식회사 Multi-input folded camera and mobile device including the same
DE102019129986A1 (en) 2019-11-07 2021-05-12 Sick Ag Optoelectronic sensor and method for detecting objects
CN115097463A (en) * 2019-12-20 2022-09-23 深圳市速腾聚创科技有限公司 Receiving device, transmitting/receiving device, and laser radar
CN114902168A (en) 2019-12-26 2022-08-12 ams国际有限公司 Sensing system
DE102020102247A1 (en) 2020-01-30 2021-08-05 Sick Ag Optoelectronic sensor and method for detecting objects
CN113472995B (en) * 2020-03-31 2022-12-06 深圳市安思疆科技有限公司 Dynamic automatic focusing system and method based on SPAD sensor receiving module
US11695911B2 (en) 2020-04-27 2023-07-04 Ouster, Inc. Stereoscopic image capturing systems
US11876109B2 (en) * 2020-05-06 2024-01-16 Semiconductor Components Industries, Llc Semiconductor devices with single-photon avalanche diodes and light spreading lenses
US11579265B2 (en) 2020-08-19 2023-02-14 Continental Autonomous Mobility US, LLC Lidar system with crosstalk reduction comprising a power supply circuit layer stacked between an avalanche-type diode layer and a read-out circuit layer
US20220276345A1 (en) * 2020-08-24 2022-09-01 Innoviz Technologies Ltd. Multi-beam laser emitter with common optical path
EP4001866A1 (en) * 2020-11-20 2022-05-25 STMicroelectronics S.r.l. Radiation sensor with an integrated mechanical optical modulator and related manufacturing process
US20240050608A1 (en) * 2021-01-04 2024-02-15 Signify Holding B.V. A disinfection system comprising an optical arrangement for a far uv light source to filter out undesired wavelengths
US11921234B2 (en) 2021-02-16 2024-03-05 Innovusion, Inc. Attaching a glass mirror to a rotating metal motor frame
US11422267B1 (en) 2021-02-18 2022-08-23 Innovusion, Inc. Dual shaft axial flux motor for optical scanners
CN112965243B (en) * 2021-03-10 2022-10-04 北京航空航天大学 Compact eye pattern (sclera blood vessel) imaging device
WO2022195537A1 (en) * 2021-03-17 2022-09-22 The Trustees Of Princeton University Microlens amplitude masks for flying pixel removal in time-of-flight imaging
US11555895B2 (en) 2021-04-20 2023-01-17 Innovusion, Inc. Dynamic compensation to polygon and motor tolerance using galvo control profile
US20220333911A1 (en) * 2021-04-20 2022-10-20 Egis Technology Inc. Distance sensing apparatus
US11614521B2 (en) 2021-04-21 2023-03-28 Innovusion, Inc. LiDAR scanner with pivot prism and mirror
US11662439B2 (en) 2021-04-22 2023-05-30 Innovusion, Inc. Compact LiDAR design with high resolution and ultra-wide field of view
CN113156557B (en) * 2021-04-30 2023-02-21 浙江光珀智能科技有限公司 Optical mask and optical system
EP4314885A1 (en) 2021-05-12 2024-02-07 Innovusion, Inc. Systems and apparatuses for mitigating lidar noise, vibration, and harshness
US11662440B2 (en) 2021-05-21 2023-05-30 Innovusion, Inc. Movement profiles for smart scanning using galvonometer mirror inside LiDAR scanner
US11768294B2 (en) 2021-07-09 2023-09-26 Innovusion, Inc. Compact lidar systems for vehicle contour fitting
WO2023018146A1 (en) * 2021-08-09 2023-02-16 주식회사 위멤스 Scanning mirror-based lidar device
DE102021124430B3 (en) 2021-09-21 2022-11-03 Sick Ag Visualize lidar measurement data
WO2023131509A1 (en) 2022-01-04 2023-07-13 Signify Holding B.V. A radiation generating system comprising an optical arrangement for a far uv light source minimizing the impact of unfiltered undesired wavelengths
WO2023147143A1 (en) * 2022-01-30 2023-08-03 Ouster, Inc. Overlapping sub-ranges with power stepping
US11871130B2 (en) 2022-03-25 2024-01-09 Innovusion, Inc. Compact perception device
US20230375679A1 (en) * 2022-05-17 2023-11-23 Stmicroelectronics (Research & Development) Limited Quantum film direct time of flight sensor circuit for low cost short wave infrared operation

Citations (237)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4358851A (en) 1980-02-28 1982-11-09 Xerox Corporation Fiber optic laser device and light emitter utilizing the device
EP0095725A1 (en) 1982-05-31 1983-12-07 Kabushiki Kaisha Toshiba Area sensor
US4634272A (en) 1982-06-02 1987-01-06 Nissan Motor Company, Limited Optical radar system with an array of photoelectric sensors
US4676599A (en) 1985-08-14 1987-06-30 The United States Of America As Represented By The Secretary Of The Air Force Micro-optical lens holder
US4702600A (en) 1984-07-20 1987-10-27 Litef Gmbh Method and apparatus for measuring angular rate with a passive optical resonator
US4744667A (en) 1986-02-11 1988-05-17 University Of Massachusetts Microspectrofluorimeter
US4851664A (en) 1988-06-27 1989-07-25 United States Of America As Represented By The Secretary Of The Navy Narrow band and wide angle hemispherical interference optical filter
JPH036407A (en) 1989-06-03 1991-01-11 Daido Steel Co Ltd Measuring device for shape of outer periphery
US5188286A (en) 1991-12-18 1993-02-23 International Business Machines Corporation Thermoelectric piezoelectric temperature control
US5267016A (en) 1991-11-27 1993-11-30 United Technologies Corporation Laser diode distance measurement
US5268748A (en) 1991-02-13 1993-12-07 U.S. Philips Corporation Arrangement for measuring the reflection and/or transmission of an object
US5288992A (en) 1992-12-15 1994-02-22 Gte Laboratories Incorporated Wide angle, narrow band optical filter
JPH0749417A (en) 1993-08-06 1995-02-21 Fujitsu Ltd Interference filter assembly
JPH07181023A (en) 1993-09-30 1995-07-18 Komatsu Ltd Cofocal optical system
JPH07318325A (en) 1994-05-26 1995-12-08 Matsushita Electric Works Ltd Method and apparatus for detection of shape
US5659420A (en) 1993-09-30 1997-08-19 Kabushiki Kaisha Komatsu Seisakusho Confocal optical apparatus
JPH09257440A (en) 1996-03-26 1997-10-03 Takaoka Electric Mfg Co Ltd Two-dimensional-array-type confocal optic device
JPH10311950A (en) 1997-05-14 1998-11-24 Olympus Optical Co Ltd Scanning microscope
US5953110A (en) 1998-04-23 1999-09-14 H.N. Burns Engineering Corporation Multichannel laser radar
US5982552A (en) 1994-12-29 1999-11-09 Nippon Sheet Glass Co., Ltd Planar microlens array having high converging efficiency
US6014232A (en) 1996-09-07 2000-01-11 U.S. Philips Corporation Electrical device comprising an array of pixels
US6043873A (en) 1997-01-10 2000-03-28 Advanced Optical Technologies, Llc Position tracking system
US6104516A (en) 1996-06-11 2000-08-15 Canon Kabushiki Kaisha Wavelength-changeable light source capable of changing wavelength of output light, optical communication network using the same and wavelength control method for controlling wavelength
US6133989A (en) 1993-02-09 2000-10-17 Advanced Scientific Concepts, Inc. 3D imaging laser radar
US6255133B1 (en) 1998-11-19 2001-07-03 Xerox Corporation Electro optical devices with reduced filter thinning on the edge pixel photosites and method of producing same
US20020003617A1 (en) 1999-03-18 2002-01-10 Guenter Doemens Spatially resolving range-finding system
US6374024B1 (en) 1998-10-30 2002-04-16 Sharp Kabushiki Kaisha Image sensor and method of manufacturing the same
US6414746B1 (en) 1999-11-24 2002-07-02 Advanced Scientific Concepts, Inc. 3-D imaging multiple target laser radar
US20030006676A1 (en) 2001-05-29 2003-01-09 Smith Stuart T. Closed loop control systems employing relaxor ferroelectric actuators
US20030047752A1 (en) 2000-09-29 2003-03-13 Campbell Joe C. Avalanche photodiodes with an impact-ionization-engineered multiplication region
US6690019B2 (en) 1997-09-16 2004-02-10 Advanced Scientific Concepts, Inc. High data rate smart sensor
US20040061502A1 (en) 1999-09-14 2004-04-01 Hasser Christopher J. High-resolution optical encoder with phased-array photodetectors
US6721262B1 (en) 1997-09-22 2004-04-13 Seagate Technology Llc Aperture stop for a flying optical head
JP2004247461A (en) 2003-02-13 2004-09-02 Toyota Central Res & Dev Lab Inc Monitor device and method
US20040223071A1 (en) 2003-05-08 2004-11-11 David Wells Multiple microlens system for image sensors or display units
US20050030409A1 (en) 2003-08-08 2005-02-10 Matherson Kevin J. Method and apparatus for generating data representative of an image
US20050046850A1 (en) 2003-07-11 2005-03-03 Svt Associates, Inc. Film mapping system
US7091462B2 (en) 2002-08-26 2006-08-15 Jds Uniphase Corporation Transmitter with laser monitoring and wavelength stabilization circuit
US20060244851A1 (en) 2005-05-02 2006-11-02 Andrew Cartlidge Increasing fill-factor on pixelated sensors
USD531525S1 (en) 2004-11-02 2006-11-07 Sick Ag Optoelectronic sensor
US20070007563A1 (en) 2004-06-01 2007-01-11 Chandra Mouli Silicon-based resonant cavity photodiode for image sensors
US7170542B2 (en) 2003-01-23 2007-01-30 Orbotech Ltd. System and method for providing high brightness illumination
US20070057211A1 (en) 2005-05-25 2007-03-15 Karsten Bahlman Multifocal imaging systems and method
US20070060806A1 (en) 2005-04-27 2007-03-15 Martin Hunter Raman spectroscopy for non-invasive glucose measurements
JP2007103590A (en) 2005-10-03 2007-04-19 Nikon Corp Image pick-up device, focus detector and image pick-up system
US20070228262A1 (en) 2005-12-19 2007-10-04 Daniel Cantin Object-detecting lighting system and method
US7295298B2 (en) 2001-06-05 2007-11-13 Ibeo Automobile Sensor Gmbh Detection method and a detection apparatus
KR20070112679A (en) 2006-05-22 2007-11-27 삼성에스디아이 주식회사 Laser irradiation device and fabrication method of organic light emitting display device using the same
US20080037114A1 (en) 2006-05-05 2008-02-14 Visitech International Ltd. Laser confocal scanning microscope and methods of improving image quality in such microscope
US7345271B2 (en) 2002-09-25 2008-03-18 Ibeo Automobile Sensor Gmbh Optoelectric sensing device with common deflection device
US20080153189A1 (en) 2003-09-30 2008-06-26 Osram Opto Semiconductors Gmbh Method for producing a radiation-emitting-and-receiving semiconductor chip
US7421159B2 (en) 2005-12-13 2008-09-02 Board of Supervisor of Louisiana State University and Agricultural and Mechanical College Integral pre-aligned micro-optical systems
US20080218612A1 (en) 2007-03-09 2008-09-11 Border John N Camera using multiple lenses and image sensors in a rangefinder configuration to provide a range map
US7433042B1 (en) 2003-12-05 2008-10-07 Surface Optics Corporation Spatially corrected full-cubed hyperspectral imager
US20090016642A1 (en) 2000-07-14 2009-01-15 Massachusetts Institute Of Technology Method and system for high resolution, ultra fast 3-d imaging
US20090040629A1 (en) 2006-03-24 2009-02-12 Bechtel Jon H Vision System
US20090179142A1 (en) 2006-01-23 2009-07-16 Jacques Duparre Image Detection System and Method For Production at Least One Image Detection System
US20090236505A1 (en) 2005-09-19 2009-09-24 Nereo Pallaro Multifunctional optical sensor comprising a photodetectors matrix coupled to a microlenses matrix
CN101563595A (en) 2006-12-12 2009-10-21 皇家飞利浦电子股份有限公司 Sample concentration detector with temperature compensation
EP2124069A1 (en) 2008-03-20 2009-11-25 IBEO Automobile Sensor GmbH Omnidirectional lidar system
US20090295910A1 (en) 2005-03-24 2009-12-03 Jose Mir Hyperspectral Imaging System and Methods Thereof
US20100008588A1 (en) 2008-07-08 2010-01-14 Chiaro Technologies LLC Multiple channel locating
US20100020306A1 (en) 2006-07-13 2010-01-28 Velodyne Acoustics, Inc. High definition lidar system
US20100110275A1 (en) 2007-04-06 2010-05-06 Gilles Mathieu Large depth-of-field imaging system and iris recogniton system
US20100123893A1 (en) 2008-11-17 2010-05-20 Dezhong Yang Laser distance measuring device
JP2010128122A (en) 2008-11-27 2010-06-10 Olympus Corp Imaging apparatus
US20100204964A1 (en) 2009-02-09 2010-08-12 Utah State University Lidar-assisted multi-image matching for 3-d model and sensor pose refinement
US20100250700A1 (en) 2009-03-30 2010-09-30 Sun Microsystems, Inc. Data storage system and method of processing a data access request
US7808706B2 (en) 2004-02-12 2010-10-05 Tredegar Newco, Inc. Light management films for displays
US7873601B1 (en) 2006-06-29 2011-01-18 Emc Corporation Backup of incremental metadata in block based backup systems
US7876456B2 (en) 2009-05-11 2011-01-25 Mitutoyo Corporation Intensity compensation for interchangeable chromatic point sensor components
US20110022566A1 (en) 2009-06-26 2011-01-27 Simplivt Corporation File system
US20110025843A1 (en) 2009-07-31 2011-02-03 Mesa Imaging Ag Time of Flight Camera with Rectangular Field of Illumination
US20110032398A1 (en) 2009-08-06 2011-02-10 Victor Lenchenkov Image sensor with multilayer interference filters
US20110037849A1 (en) 2008-04-11 2011-02-17 Cristiano Niclass Time-of-flight based imaging system using a display as illumination source
US20110038459A1 (en) 2009-08-14 2011-02-17 Varian Medical Systems, Inc. X-ray tube bearing shaft and hub
US20110116262A1 (en) 2009-11-13 2011-05-19 Phoseon Technology, Inc. Economical partially collimating reflective micro optical array
WO2011063347A2 (en) 2009-11-20 2011-05-26 Pelican Imaging Corporation Capturing and processing of images using monolithic camera array with heterogeneous imagers
US8013983B2 (en) 2007-12-28 2011-09-06 Industrial Technology Research Institute Three-dimensional surround scanning device and method thereof
US20110216304A1 (en) 2006-07-13 2011-09-08 Velodyne Acoustics, Inc. High definition lidar system
US20110228142A1 (en) 2009-10-14 2011-09-22 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Device, image processing device and method for optical imaging
US20110283113A1 (en) 2010-05-14 2011-11-17 Oracle International Corporation Method and system for encrypting data
US20110320865A1 (en) 2010-06-25 2011-12-29 International Business Machines Corporation Deduplication in a hybrid storage environment
US20120044476A1 (en) 2008-05-09 2012-02-23 Ball Aerospace & Technologies Corp. Systems and methods of scene and action capture using imaging system incorporating 3d lidar
US8130367B2 (en) 2005-12-08 2012-03-06 Roger Stettner Laser ranging, tracking and designation using 3-D focal planes
WO2012055966A1 (en) 2010-10-27 2012-05-03 MAX-PLANCK-Gesellschaft zur Förderung der Wissenschaften e.V. Protecting data integrity with storage leases
USD659030S1 (en) 2010-07-30 2012-05-08 Sick Ag Optoelectronic sensor
US20120141104A1 (en) 2010-12-03 2012-06-07 Research In Motion Limited Dynamic lighting control in hybrid camera - projector device
US20120140109A1 (en) 2010-12-06 2012-06-07 Primesense Ltd. Lens Arrays for Pattern Projection and Imaging
US20120154914A1 (en) 2010-12-16 2012-06-21 Denso Corporation Interference filter assembly
US20120182464A1 (en) 2011-01-18 2012-07-19 Primesense Ltd. Objective optics with interference filter
JP2012202776A (en) 2011-03-24 2012-10-22 Toyota Central R&D Labs Inc Distance measuring device
US20120287417A1 (en) 2011-05-11 2012-11-15 Yvan Mimeault Multiple-field-of-view scannerless optical rangefinder in high ambient background light
US8319949B2 (en) 2007-06-18 2012-11-27 Leddartech Inc. Method for detecting objects with visible light
JP2012530917A (en) 2009-06-22 2012-12-06 トヨタ モーター ヨーロッパ ナームロゼ フェンノートシャップ/ソシエテ アノニム Optical distance meter using pulsed light
US20120320164A1 (en) 2011-06-16 2012-12-20 Lenny Lipton Stereoscopic camera with polarizing apertures
KR20120137432A (en) 2010-04-05 2012-12-20 퀄컴 인코포레이티드 Combining data from multiple image sensors
US8374405B2 (en) 2009-08-28 2013-02-12 Hon Hai Precision Industry Co., Ltd. Camera device and identity recognition method utilizing the same
US20130038941A1 (en) 2011-08-09 2013-02-14 Primesense Ltd. Lens Array Projector
US20130044187A1 (en) 2011-08-18 2013-02-21 Sick Ag 3d camera and method of monitoring a spatial zone
US20130044310A1 (en) 2007-12-21 2013-02-21 Leddartech Inc. Distance detection method and system
US8384997B2 (en) 2008-01-21 2013-02-26 Primesense Ltd Optical pattern projection
CN103048046A (en) 2012-12-21 2013-04-17 浙江大学 Double-beam spectrometer
KR20130038388A (en) 2010-07-19 2013-04-17 프라운호퍼-게젤샤프트 츄어 푀르더룽 데어 안게반텐 포르슝에.파우. Image recording device and method for recording an image
US20130110779A1 (en) 2010-05-03 2013-05-02 Panzura, Inc. Archiving data for a distributed filesystem
US20130111262A1 (en) 2010-05-03 2013-05-02 Panzura, Inc. Providing disaster recovery for a distributed filesystem
US20130141549A1 (en) 2010-06-29 2013-06-06 Cyclomedia Technology B.V. Method for Producing a Digital Photo Wherein at Least Some of the Pixels Comprise Position Information, and Such a Digital Photo
EP2615566A2 (en) 2012-01-13 2013-07-17 Nexenta Systems, Inc. Unified local storage supporting file and cloud object access
US8494252B2 (en) 2007-06-19 2013-07-23 Primesense Ltd. Depth mapping using optical elements having non-uniform focal characteristics
US8504529B1 (en) 2009-06-19 2013-08-06 Netapp, Inc. System and method for restoring data to a storage device based on a backup image
CN103234527A (en) 2013-04-07 2013-08-07 南京理工大学 Multispectral light-field camera
US20130206967A1 (en) 2012-02-15 2013-08-15 Primesense Ltd. Integrated optoelectronic modules
US20130229646A1 (en) 2012-03-02 2013-09-05 Seiko Epson Corporation Component analyzer
US20130294089A1 (en) 2007-04-02 2013-11-07 Primesense Ltd. Pattern projection using micro-lenses
US20130300840A1 (en) 2010-12-23 2013-11-14 Fastree 3D S.A. 3d landscape real-time imager and corresponding imaging methods
US20140006354A1 (en) 2010-05-03 2014-01-02 Panzura, Inc. Executing a cloud command for a distributed filesystem
US8645810B2 (en) 2011-07-31 2014-02-04 Sandisk Technologies Inc. Fast detection of convergence or divergence in iterative decoding
US8675181B2 (en) 2009-06-02 2014-03-18 Velodyne Acoustics, Inc. Color LiDAR scanner
US20140118335A1 (en) 2012-10-30 2014-05-01 Primesense Ltd. Depth mapping with enhanced resolution
US20140118493A1 (en) 2009-04-16 2014-05-01 Primesense Ltd. Three-dimensional mapping and imaging
US8743176B2 (en) 2009-05-20 2014-06-03 Advanced Scientific Concepts, Inc. 3-dimensional hybrid camera and production system
US8742325B1 (en) 2013-07-31 2014-06-03 Google Inc. Photodetector array on curved substrate
US20140153001A1 (en) 2012-03-22 2014-06-05 Primesense Ltd. Gimbaled scanning mirror array
US20140158900A1 (en) 2012-12-11 2014-06-12 Samsung Electronics Co., Ltd. Photon-counting detector and readout circuit
US20140168631A1 (en) 2012-12-18 2014-06-19 Pouch Holdings LLC Multi-clad Fiber Based Optical Apparatus and Methods for Light Detection and Ranging Sensors
US8762798B2 (en) 2011-11-16 2014-06-24 Stec, Inc. Dynamic LDPC code rate solution
US20140211194A1 (en) 2013-01-27 2014-07-31 Quanergy Systems, Inc. Cost-effective lidar sensor for multi-signal detection, weak signal detection and signal disambiguation and method of using same
US8829406B2 (en) 2009-11-15 2014-09-09 Primesense Ltd. Optical projector with beam monitor including sensing intensity of beam pattern not projected toward an object
US8836922B1 (en) 2013-08-20 2014-09-16 Google Inc. Devices and methods for a rotating LIDAR platform with a shared transmit/receive path
US20140269796A1 (en) 2012-01-31 2014-09-18 Flir Systems, Inc. Multi-wavelength vcsel array to reduce speckle
US20140267878A1 (en) 2011-11-04 2014-09-18 Imec Spectral camera with overlapping segments of image copies interleaved onto sensor array
WO2014150856A1 (en) 2013-03-15 2014-09-25 Pelican Imaging Corporation Array camera implementing quantum dot color filters
US20140285628A1 (en) 2010-02-23 2014-09-25 Primesense Ltd. Wideband ambient light rejection
US8848039B2 (en) 2008-07-09 2014-09-30 Primesense Ltd. Integrated processor for 3D mapping
US20140291491A1 (en) 2012-03-22 2014-10-02 Primesense Ltd. Calibration of time-of-flight measurement using stray reflections
US20140313519A1 (en) 2013-03-15 2014-10-23 Primesense Ltd. Depth scanning with multiple emitters
US20140375977A1 (en) 2013-06-21 2014-12-25 David Ludwig LIDAR Device and Method for Clear and Degraded Environmental Viewing Conditions
US20140376092A1 (en) 2013-06-19 2014-12-25 Primesense Ltd. Integrated structured-light projector
US20150002636A1 (en) 2013-06-28 2015-01-01 Cable Television Laboratories, Inc. Capturing Full Motion Live Events Using Spatially Distributed Depth Sensing Cameras
WO2015037211A1 (en) 2013-09-11 2015-03-19 Sony Corporation Image processing device and method
WO2015052616A1 (en) 2013-10-09 2015-04-16 Koninklijke Philips N.V. Monolithic led arrays for uniform and high-brightness light sources
US20150115131A1 (en) 2013-10-28 2015-04-30 Omnivision Technologies, Inc. Stacked chip spad image sensor
US20150124094A1 (en) 2013-11-05 2015-05-07 Delphi Technologies, Inc. Multiple imager vehicle optical sensor system
US20150131080A1 (en) 2013-11-12 2015-05-14 Facet Technology Corp. Methods and Apparatus for Array Based Lidar Systems with Reduced Interference
US9063549B1 (en) 2013-03-06 2015-06-23 Google Inc. Light detection and ranging device with oscillating mirror driven by magnetically interactive coil
KR20150072439A (en) 2012-10-19 2015-06-29 퀄컴 인코포레이티드 Multi-camera system using folded optics
US9071763B1 (en) 2012-09-26 2015-06-30 Google Inc. Uniform illumination image capture
US20150184999A1 (en) 2005-05-10 2015-07-02 Advanced Scientific Concepts Inc. Dimensioning system
US20150186287A1 (en) 2013-12-30 2015-07-02 Michael Henry Kass Using Memory System Programming Interfacing
US20150192677A1 (en) 2014-01-03 2015-07-09 Quanergy Systems, Inc. Distributed lidar sensing system for wide field of view three dimensional mapping and method of using same
US9086273B1 (en) 2013-03-08 2015-07-21 Google Inc. Microrod compression of laser beam in combination with transmit lens
JP2015137987A (en) 2014-01-24 2015-07-30 アズビル株式会社 Distance sensor and distance measurement method
US9111444B2 (en) 2012-10-31 2015-08-18 Raytheon Company Video and lidar target detection and tracking system and method for segmenting moving targets
EP2908166A2 (en) 2014-02-17 2015-08-19 Yokogawa Electric Corporation Confocal optical scanner
WO2015136100A2 (en) 2014-03-14 2015-09-17 Mesa Imaging Ag Optoelectronic modules operable to recognize spurious reflections and to compensate for errors caused by spurious reflections
US20150260830A1 (en) 2013-07-12 2015-09-17 Princeton Optronics Inc. 2-D Planar VCSEL Source for 3-D Imaging
US20150292948A1 (en) 2013-08-02 2015-10-15 Verifood, Ltd. Spectrometry system with diffuser
US20150293224A1 (en) 2013-05-09 2015-10-15 Quanergy Systems, Inc. Solid state optical phased array lidar and method of using same
US9164511B1 (en) 2013-04-17 2015-10-20 Google Inc. Use of detected objects for image processing
WO2015163074A1 (en) 2014-04-22 2015-10-29 シャープ株式会社 Optical sensor system, optical-type gas sensor system, microparticle sensor system, light-emitting device, and image printing device
US9176051B2 (en) 2011-06-29 2015-11-03 Karlsruher Institut Fuer Technologie Micro-optical element, micro-optical array, and optical sensor system
US20150316473A1 (en) 2014-05-01 2015-11-05 Rebellion Photonics, Inc. Mobile gas and chemical imaging camera
US20150319349A1 (en) 2012-11-21 2015-11-05 Nokia Technologies Oy A Module for Plenoptic Camera System
US20150355470A1 (en) 2012-12-31 2015-12-10 Iee International Electronics & Engineering S.A. Optical system generating a structured light field from an array of light sources by means of a refracting or reflecting light structuring element
US20150358601A1 (en) 2009-12-23 2015-12-10 Mesa Imaging Ag Optical Filter on Objective Lens for 3D Cameras
US20150362585A1 (en) 2013-07-12 2015-12-17 Princeton Optronics Inc. 2-D Planar VCSEL Source for 3-D Imaging
US20150379371A1 (en) 2014-06-30 2015-12-31 Microsoft Corporation Object Detection Utilizing Geometric Information Fused With Image Data
US20150378241A1 (en) 2014-06-30 2015-12-31 Quanergy Systems, Inc. Planar beam forming and steering optical phased array chip and method of using same
US9229109B2 (en) 2011-09-15 2016-01-05 Advanced Scientific Concepts, Inc. Automatic range corrected flash ladar camera
US20160003946A1 (en) 2014-07-03 2016-01-07 Advanced Scientific Concepts, Inc. Ladar sensor for a dense environment
US20160047898A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Method and System for Ladar Transmission with Spiral Dynamic Scan Patterns
US20160049765A1 (en) 2014-08-15 2016-02-18 Quanergy Systems, Inc. Three-dimensional-mapping two-dimensional-scanning lidar based on one-dimensional-steering optical phased arrays and method of using same
US20160047901A1 (en) 2012-12-25 2016-02-18 Quanergy Systems, Inc. Robust lidar sensor for broad weather, shock and vibration conditions
US9281841B2 (en) 2012-10-31 2016-03-08 Avago Technologies General Ip (Singapore) Pte. Ltd. Load balanced decoding of low-density parity-check codes
US9285477B1 (en) 2013-01-25 2016-03-15 Apple Inc. 3D depth point cloud from timing flight of 2D scanned light beam pulses
US9299731B1 (en) 2013-09-30 2016-03-29 Google Inc. Systems and methods for selectable photodiode circuits
EP3002548A1 (en) 2014-10-02 2016-04-06 Sick Ag Illumination device and method for generating an illumination field
US20160097858A1 (en) 2014-10-06 2016-04-07 The Boeing Company Backfilling clouds of 3d coordinates
US9330464B1 (en) 2014-12-12 2016-05-03 Microsoft Technology Licensing, Llc Depth camera feedback
US20160150963A1 (en) 2011-09-26 2016-06-02 Michael Lee Roukes One-photon integrated neurophotonic systems
US20160161600A1 (en) 2013-08-19 2016-06-09 Quanergy Systems, Inc. Optical phased array lidar system and method of using same
US9368936B1 (en) 2013-09-30 2016-06-14 Google Inc. Laser diode firing system
US9369689B1 (en) 2015-02-24 2016-06-14 HypeVR Lidar stereo fusion live action 3D model video reconstruction for six degrees of freedom 360° volumetric virtual reality video
US9383753B1 (en) 2012-09-26 2016-07-05 Google Inc. Wide-view LIDAR with areas of special attention
EP3045935A1 (en) 2015-01-13 2016-07-20 XenomatiX BVBA Surround sensing system with dome-filter assembly
WO2016116733A1 (en) 2015-01-20 2016-07-28 Milan Momcilo Popovich Holographic waveguide lidar
US20160218727A1 (en) 2014-10-31 2016-07-28 Seiko Epson Corporation Quantum interference device, atomic oscillator, electronic apparatus, and moving object
WO2016125165A2 (en) 2015-02-05 2016-08-11 Verifood, Ltd. Spectrometry system with visible aiming beam
US9425654B2 (en) 2013-09-30 2016-08-23 Google Inc. Contactless electrical coupling for a rotatable LIDAR device
US9435891B2 (en) 2012-02-15 2016-09-06 Heptagon Micro Optics Pte. Ltd. Time of flight camera with stripe illumination
US20160265902A1 (en) 2015-03-12 2016-09-15 Apple Inc. Thin psd for laser-scanning systems
US20160291134A1 (en) 2015-04-06 2016-10-06 Google Inc. Long Range Steerable LIDAR System
US9470520B2 (en) 2013-03-14 2016-10-18 Apparate International C.V. LiDAR scanner
US20160306032A1 (en) 2013-11-22 2016-10-20 Brent S. SCHWARZ Lidar scanner calibration
US9489601B2 (en) 2012-09-05 2016-11-08 X Development Llc Construction zone sign detection
US20160328619A1 (en) 2013-12-24 2016-11-10 Lg Electronics Inc. Vehicle driving assistance device and vehicle having the same
US20160327779A1 (en) 2014-01-17 2016-11-10 The Trustees Of Columbia University In The City Of New York Systems And Methods for Three Dimensional Imaging
US9525863B2 (en) 2015-04-29 2016-12-20 Apple Inc. Time-of-flight depth mapping with flexible scan pattern
US9529079B1 (en) 2015-03-26 2016-12-27 Google Inc. Multiplexed multichannel photodetector
US9551791B2 (en) 2013-07-09 2017-01-24 Xenomatix Nv Surround sensing system
US20170146640A1 (en) 2015-11-25 2017-05-25 Velodyne Lidar, Inc. Three Dimensional LIDAR System With Targeted Field of View
US20170219695A1 (en) 2016-01-31 2017-08-03 Velodyne Lidar, Inc. Multiple Pulse, LIDAR Based 3-D Imaging
WO2017132704A1 (en) 2016-01-31 2017-08-03 Velodyne Lidar, Inc. Lidar based 3-d imaging with far-field illumination overlap
US20170219426A1 (en) 2016-01-29 2017-08-03 Ouster, Inc. Systems and Methods for Calibrating an Optical Distance Sensor
US20170269197A1 (en) 2016-03-21 2017-09-21 Velodyne Lidar, Inc. LIDAR Based 3-D Imaging With Varying Illumination Intensity
US20170269215A1 (en) 2016-03-19 2017-09-21 Velodyne Lidar, Inc. Integrated Illumination And Detection For LIDAR Based 3-D Imaging
US20170269209A1 (en) 2016-03-21 2017-09-21 Velodyne Lidar, Inc. LIDAR Based 3-D Imaging With Varying Pulse Repetition
US20170269198A1 (en) 2016-03-21 2017-09-21 Velodyne Lidar, Inc. LIDAR Based 3-D Imaging With Varying Illumination Field Density
US20170289524A1 (en) 2015-09-24 2017-10-05 Ouster, Inc. Optical System for Collecting Distance Information Within a Field
US20170299700A1 (en) 2014-10-20 2017-10-19 Quanergy Systems, Inc. Three-dimensional lidar sensor based on two-dimensional scanning of one-dimensional optical emitter and method of using same
US20170303829A1 (en) 2014-08-21 2017-10-26 I.r Med Ltd. System and method for noninvasive analysis of subcutaneous tissue
US20170350983A1 (en) 2016-06-01 2017-12-07 Velodyne Lidar, Inc. Multiple Pixel Scanning LIDAR
US9866241B2 (en) 2015-09-25 2018-01-09 SK Hynix Inc. Techniques for adaptive LDPC decoding
US20180032396A1 (en) 2016-07-29 2018-02-01 Sandisk Technologies Llc Generalized syndrome weights
US20180038959A1 (en) 2013-12-06 2018-02-08 Kazuhiro Akatsu Object detector and sensor
US20180059222A1 (en) 2016-08-24 2018-03-01 Ouster, Inc. Optical system for collecting distance information within a field
WO2018065427A1 (en) 2016-10-03 2018-04-12 Xenomatix Nv System for determining a distance to an object
WO2018065429A1 (en) 2016-10-03 2018-04-12 Xenomatix Nv System and method for determining a distance to an object
WO2018065426A1 (en) 2016-10-03 2018-04-12 Xenomatix Nv Method for subtracting background light from an exposure value of a pixel in an imaging array, and pixel for use in same
WO2018065428A2 (en) 2016-10-03 2018-04-12 Xenomatix Nv System for determining a distance to an object
EP3316000A1 (en) 2016-10-28 2018-05-02 Xenomatix NV Vehicular system for measuring a distance to an object and method of installing same
US20180167602A1 (en) 2015-09-24 2018-06-14 Ouster, Inc. Optical system for collecting distance information within a field
WO2018122415A1 (en) 2016-12-30 2018-07-05 Xenomatix Nv System for characterizing surroundings of a vehicle
US20180299554A1 (en) 2015-10-23 2018-10-18 Xenomatix Nv System and method for determining a distance to an object
WO2018197441A1 (en) 2017-04-23 2018-11-01 Xenomatix Nv A pixel structure
US20180329065A1 (en) 2017-05-15 2018-11-15 Ouster, Inc. Optical imaging transmitter with brightness enhancement
US10183541B2 (en) 2015-01-13 2019-01-22 Xenomatix Nv Surround sensing system with telecentric optics
US20190097653A1 (en) 2017-09-22 2019-03-28 SK Hynix Memory Solutions America Inc. Memory system with decoders and method of operating such memory system and decoders
US20190103885A1 (en) 2017-09-29 2019-04-04 Nyquist Semiconductor Limited Systems and methods for decoding error correcting codes
US10291261B2 (en) 2016-04-25 2019-05-14 SK Hynix Inc. Early selection decoding and automatic tuning
AU2017330180A1 (en) 2016-09-26 2019-05-16 Ouster, Inc. Optical system for collecting distance information within a field
US20190288713A1 (en) 2018-03-16 2019-09-19 SK Hynix Memory Solutions America Inc. Memory system with hybrid decoding scheme and method of operating such memory system
US20200041646A1 (en) 2017-07-05 2020-02-06 Ouster, Inc. Light ranging device with electronically scanned emitter array and synchronized sensor array
US20200116560A1 (en) 2018-08-09 2020-04-16 Ouster, Inc. Scanning sensor array with overlapping pass bands
US20200209355A1 (en) 2018-12-26 2020-07-02 Ouster, Inc. Solid-state electronic scanning laser array with high-side and low-side switches for increased channels
US10739189B2 (en) * 2018-08-09 2020-08-11 Ouster, Inc. Multispectral ranging/imaging sensor arrays and systems
US10809380B2 (en) 2017-05-15 2020-10-20 Ouster, Inc. Augmenting panoramic LIDAR results with color

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US672262A (en) * 1900-11-16 1901-04-16 Charles Hook Adrean Harness attachment.
US8605761B2 (en) * 2010-01-18 2013-12-10 Optical Physics Company Multi-beam laser control system and method
TWI533179B (en) * 2013-10-25 2016-05-11 緯創資通股份有限公司 Optical touch system, method of touch detection, and computer program product
US11105925B2 (en) * 2017-03-01 2021-08-31 Ouster, Inc. Accurate photo detector measurements for LIDAR
US10969490B2 (en) * 2017-12-07 2021-04-06 Ouster, Inc. Light ranging system with opposing circuit boards
WO2020180997A1 (en) * 2019-03-04 2020-09-10 Ouster, Inc. Driver visualization and semantic monitoring of a vehicle using lidar data
EP4025934A4 (en) * 2019-09-06 2023-08-16 Ouster, Inc. Processing of lidar images
US11695911B2 (en) * 2020-04-27 2023-07-04 Ouster, Inc. Stereoscopic image capturing systems

Patent Citations (278)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4358851A (en) 1980-02-28 1982-11-09 Xerox Corporation Fiber optic laser device and light emitter utilizing the device
EP0095725A1 (en) 1982-05-31 1983-12-07 Kabushiki Kaisha Toshiba Area sensor
US4634272A (en) 1982-06-02 1987-01-06 Nissan Motor Company, Limited Optical radar system with an array of photoelectric sensors
US4702600A (en) 1984-07-20 1987-10-27 Litef Gmbh Method and apparatus for measuring angular rate with a passive optical resonator
US4676599A (en) 1985-08-14 1987-06-30 The United States Of America As Represented By The Secretary Of The Air Force Micro-optical lens holder
US4744667A (en) 1986-02-11 1988-05-17 University Of Massachusetts Microspectrofluorimeter
US4851664A (en) 1988-06-27 1989-07-25 United States Of America As Represented By The Secretary Of The Navy Narrow band and wide angle hemispherical interference optical filter
JPH036407A (en) 1989-06-03 1991-01-11 Daido Steel Co Ltd Measuring device for shape of outer periphery
US5268748A (en) 1991-02-13 1993-12-07 U.S. Philips Corporation Arrangement for measuring the reflection and/or transmission of an object
US5267016A (en) 1991-11-27 1993-11-30 United Technologies Corporation Laser diode distance measurement
US5188286A (en) 1991-12-18 1993-02-23 International Business Machines Corporation Thermoelectric piezoelectric temperature control
US5288992A (en) 1992-12-15 1994-02-22 Gte Laboratories Incorporated Wide angle, narrow band optical filter
US6133989A (en) 1993-02-09 2000-10-17 Advanced Scientific Concepts, Inc. 3D imaging laser radar
JPH0749417A (en) 1993-08-06 1995-02-21 Fujitsu Ltd Interference filter assembly
JPH07181023A (en) 1993-09-30 1995-07-18 Komatsu Ltd Cofocal optical system
US5659420A (en) 1993-09-30 1997-08-19 Kabushiki Kaisha Komatsu Seisakusho Confocal optical apparatus
JPH07318325A (en) 1994-05-26 1995-12-08 Matsushita Electric Works Ltd Method and apparatus for detection of shape
US5982552A (en) 1994-12-29 1999-11-09 Nippon Sheet Glass Co., Ltd Planar microlens array having high converging efficiency
JPH09257440A (en) 1996-03-26 1997-10-03 Takaoka Electric Mfg Co Ltd Two-dimensional-array-type confocal optic device
US6104516A (en) 1996-06-11 2000-08-15 Canon Kabushiki Kaisha Wavelength-changeable light source capable of changing wavelength of output light, optical communication network using the same and wavelength control method for controlling wavelength
US6014232A (en) 1996-09-07 2000-01-11 U.S. Philips Corporation Electrical device comprising an array of pixels
US6043873A (en) 1997-01-10 2000-03-28 Advanced Optical Technologies, Llc Position tracking system
JPH10311950A (en) 1997-05-14 1998-11-24 Olympus Optical Co Ltd Scanning microscope
US6028306A (en) 1997-05-14 2000-02-22 Olympus Optical Co., Ltd. Scanning microscope
US6690019B2 (en) 1997-09-16 2004-02-10 Advanced Scientific Concepts, Inc. High data rate smart sensor
US6721262B1 (en) 1997-09-22 2004-04-13 Seagate Technology Llc Aperture stop for a flying optical head
US5953110A (en) 1998-04-23 1999-09-14 H.N. Burns Engineering Corporation Multichannel laser radar
US6374024B1 (en) 1998-10-30 2002-04-16 Sharp Kabushiki Kaisha Image sensor and method of manufacturing the same
US6255133B1 (en) 1998-11-19 2001-07-03 Xerox Corporation Electro optical devices with reduced filter thinning on the edge pixel photosites and method of producing same
US20020003617A1 (en) 1999-03-18 2002-01-10 Guenter Doemens Spatially resolving range-finding system
US20040061502A1 (en) 1999-09-14 2004-04-01 Hasser Christopher J. High-resolution optical encoder with phased-array photodetectors
US6414746B1 (en) 1999-11-24 2002-07-02 Advanced Scientific Concepts, Inc. 3-D imaging multiple target laser radar
US20090016642A1 (en) 2000-07-14 2009-01-15 Massachusetts Institute Of Technology Method and system for high resolution, ultra fast 3-d imaging
US20030047752A1 (en) 2000-09-29 2003-03-13 Campbell Joe C. Avalanche photodiodes with an impact-ionization-engineered multiplication region
US20030006676A1 (en) 2001-05-29 2003-01-09 Smith Stuart T. Closed loop control systems employing relaxor ferroelectric actuators
US7295298B2 (en) 2001-06-05 2007-11-13 Ibeo Automobile Sensor Gmbh Detection method and a detection apparatus
US7091462B2 (en) 2002-08-26 2006-08-15 Jds Uniphase Corporation Transmitter with laser monitoring and wavelength stabilization circuit
US7345271B2 (en) 2002-09-25 2008-03-18 Ibeo Automobile Sensor Gmbh Optoelectric sensing device with common deflection device
US7170542B2 (en) 2003-01-23 2007-01-30 Orbotech Ltd. System and method for providing high brightness illumination
JP2004247461A (en) 2003-02-13 2004-09-02 Toyota Central Res & Dev Lab Inc Monitor device and method
US20040223071A1 (en) 2003-05-08 2004-11-11 David Wells Multiple microlens system for image sensors or display units
US20050046850A1 (en) 2003-07-11 2005-03-03 Svt Associates, Inc. Film mapping system
US20050030409A1 (en) 2003-08-08 2005-02-10 Matherson Kevin J. Method and apparatus for generating data representative of an image
US20080153189A1 (en) 2003-09-30 2008-06-26 Osram Opto Semiconductors Gmbh Method for producing a radiation-emitting-and-receiving semiconductor chip
US7433042B1 (en) 2003-12-05 2008-10-07 Surface Optics Corporation Spatially corrected full-cubed hyperspectral imager
US7808706B2 (en) 2004-02-12 2010-10-05 Tredegar Newco, Inc. Light management films for displays
US20070007563A1 (en) 2004-06-01 2007-01-11 Chandra Mouli Silicon-based resonant cavity photodiode for image sensors
USD531525S1 (en) 2004-11-02 2006-11-07 Sick Ag Optoelectronic sensor
US20090295910A1 (en) 2005-03-24 2009-12-03 Jose Mir Hyperspectral Imaging System and Methods Thereof
US20070060806A1 (en) 2005-04-27 2007-03-15 Martin Hunter Raman spectroscopy for non-invasive glucose measurements
US20060244851A1 (en) 2005-05-02 2006-11-02 Andrew Cartlidge Increasing fill-factor on pixelated sensors
US20150184999A1 (en) 2005-05-10 2015-07-02 Advanced Scientific Concepts Inc. Dimensioning system
US20070057211A1 (en) 2005-05-25 2007-03-15 Karsten Bahlman Multifocal imaging systems and method
US20090236505A1 (en) 2005-09-19 2009-09-24 Nereo Pallaro Multifunctional optical sensor comprising a photodetectors matrix coupled to a microlenses matrix
JP2007103590A (en) 2005-10-03 2007-04-19 Nikon Corp Image pick-up device, focus detector and image pick-up system
US8130367B2 (en) 2005-12-08 2012-03-06 Roger Stettner Laser ranging, tracking and designation using 3-D focal planes
US7421159B2 (en) 2005-12-13 2008-09-02 Board of Supervisor of Louisiana State University and Agricultural and Mechanical College Integral pre-aligned micro-optical systems
US20070228262A1 (en) 2005-12-19 2007-10-04 Daniel Cantin Object-detecting lighting system and method
US20090179142A1 (en) 2006-01-23 2009-07-16 Jacques Duparre Image Detection System and Method For Production at Least One Image Detection System
US20090040629A1 (en) 2006-03-24 2009-02-12 Bechtel Jon H Vision System
US20080037114A1 (en) 2006-05-05 2008-02-14 Visitech International Ltd. Laser confocal scanning microscope and methods of improving image quality in such microscope
KR20070112679A (en) 2006-05-22 2007-11-27 삼성에스디아이 주식회사 Laser irradiation device and fabrication method of organic light emitting display device using the same
US7873601B1 (en) 2006-06-29 2011-01-18 Emc Corporation Backup of incremental metadata in block based backup systems
US20100020306A1 (en) 2006-07-13 2010-01-28 Velodyne Acoustics, Inc. High definition lidar system
US7969558B2 (en) 2006-07-13 2011-06-28 Velodyne Acoustics Inc. High definition lidar system
US20110216304A1 (en) 2006-07-13 2011-09-08 Velodyne Acoustics, Inc. High definition lidar system
US8767190B2 (en) 2006-07-13 2014-07-01 Velodyne Acoustics, Inc. High definition LiDAR system
CN101563595A (en) 2006-12-12 2009-10-21 皇家飞利浦电子股份有限公司 Sample concentration detector with temperature compensation
US7683962B2 (en) 2007-03-09 2010-03-23 Eastman Kodak Company Camera using multiple lenses and image sensors in a rangefinder configuration to provide a range map
US20080218612A1 (en) 2007-03-09 2008-09-11 Border John N Camera using multiple lenses and image sensors in a rangefinder configuration to provide a range map
US20130294089A1 (en) 2007-04-02 2013-11-07 Primesense Ltd. Pattern projection using micro-lenses
US20100110275A1 (en) 2007-04-06 2010-05-06 Gilles Mathieu Large depth-of-field imaging system and iris recogniton system
US8319949B2 (en) 2007-06-18 2012-11-27 Leddartech Inc. Method for detecting objects with visible light
US8761495B2 (en) 2007-06-19 2014-06-24 Primesense Ltd. Distance-varying illumination and imaging techniques for depth mapping
US8494252B2 (en) 2007-06-19 2013-07-23 Primesense Ltd. Depth mapping using optical elements having non-uniform focal characteristics
US20130044310A1 (en) 2007-12-21 2013-02-21 Leddartech Inc. Distance detection method and system
US8013983B2 (en) 2007-12-28 2011-09-06 Industrial Technology Research Institute Three-dimensional surround scanning device and method thereof
US8384997B2 (en) 2008-01-21 2013-02-26 Primesense Ltd Optical pattern projection
EP2124069A1 (en) 2008-03-20 2009-11-25 IBEO Automobile Sensor GmbH Omnidirectional lidar system
US20110037849A1 (en) 2008-04-11 2011-02-17 Cristiano Niclass Time-of-flight based imaging system using a display as illumination source
US9041915B2 (en) 2008-05-09 2015-05-26 Ball Aerospace & Technologies Corp. Systems and methods of scene and action capture using imaging system incorporating 3D LIDAR
US20120044476A1 (en) 2008-05-09 2012-02-23 Ball Aerospace & Technologies Corp. Systems and methods of scene and action capture using imaging system incorporating 3d lidar
US20100008588A1 (en) 2008-07-08 2010-01-14 Chiaro Technologies LLC Multiple channel locating
US8848039B2 (en) 2008-07-09 2014-09-30 Primesense Ltd. Integrated processor for 3D mapping
US8089618B2 (en) 2008-11-17 2012-01-03 Chervon Limited Laser distance measuring device
US20100123893A1 (en) 2008-11-17 2010-05-20 Dezhong Yang Laser distance measuring device
US20100149389A1 (en) 2008-11-27 2010-06-17 Hisashi Goto Image pickup device
JP2010128122A (en) 2008-11-27 2010-06-10 Olympus Corp Imaging apparatus
US20100204964A1 (en) 2009-02-09 2010-08-12 Utah State University Lidar-assisted multi-image matching for 3-d model and sensor pose refinement
US20100250700A1 (en) 2009-03-30 2010-09-30 Sun Microsystems, Inc. Data storage system and method of processing a data access request
US20140118493A1 (en) 2009-04-16 2014-05-01 Primesense Ltd. Three-dimensional mapping and imaging
US7876456B2 (en) 2009-05-11 2011-01-25 Mitutoyo Corporation Intensity compensation for interchangeable chromatic point sensor components
US8743176B2 (en) 2009-05-20 2014-06-03 Advanced Scientific Concepts, Inc. 3-dimensional hybrid camera and production system
US8675181B2 (en) 2009-06-02 2014-03-18 Velodyne Acoustics, Inc. Color LiDAR scanner
US8504529B1 (en) 2009-06-19 2013-08-06 Netapp, Inc. System and method for restoring data to a storage device based on a backup image
JP2012530917A (en) 2009-06-22 2012-12-06 トヨタ モーター ヨーロッパ ナームロゼ フェンノートシャップ/ソシエテ アノニム Optical distance meter using pulsed light
US20110022566A1 (en) 2009-06-26 2011-01-27 Simplivt Corporation File system
US20110025843A1 (en) 2009-07-31 2011-02-03 Mesa Imaging Ag Time of Flight Camera with Rectangular Field of Illumination
US8330840B2 (en) 2009-08-06 2012-12-11 Aptina Imaging Corporation Image sensor with multilayer interference filters
US20110032398A1 (en) 2009-08-06 2011-02-10 Victor Lenchenkov Image sensor with multilayer interference filters
US20110038459A1 (en) 2009-08-14 2011-02-17 Varian Medical Systems, Inc. X-ray tube bearing shaft and hub
US8374405B2 (en) 2009-08-28 2013-02-12 Hon Hai Precision Industry Co., Ltd. Camera device and identity recognition method utilizing the same
US20110228142A1 (en) 2009-10-14 2011-09-22 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Device, image processing device and method for optical imaging
US20110116262A1 (en) 2009-11-13 2011-05-19 Phoseon Technology, Inc. Economical partially collimating reflective micro optical array
US8829406B2 (en) 2009-11-15 2014-09-09 Primesense Ltd. Optical projector with beam monitor including sensing intensity of beam pattern not projected toward an object
WO2011063347A2 (en) 2009-11-20 2011-05-26 Pelican Imaging Corporation Capturing and processing of images using monolithic camera array with heterogeneous imagers
US20150358601A1 (en) 2009-12-23 2015-12-10 Mesa Imaging Ag Optical Filter on Objective Lens for 3D Cameras
US20140285628A1 (en) 2010-02-23 2014-09-25 Primesense Ltd. Wideband ambient light rejection
KR20120137432A (en) 2010-04-05 2012-12-20 퀄컴 인코포레이티드 Combining data from multiple image sensors
US20130111262A1 (en) 2010-05-03 2013-05-02 Panzura, Inc. Providing disaster recovery for a distributed filesystem
US20130110779A1 (en) 2010-05-03 2013-05-02 Panzura, Inc. Archiving data for a distributed filesystem
US20140006354A1 (en) 2010-05-03 2014-01-02 Panzura, Inc. Executing a cloud command for a distributed filesystem
US20110283113A1 (en) 2010-05-14 2011-11-17 Oracle International Corporation Method and system for encrypting data
US20110320865A1 (en) 2010-06-25 2011-12-29 International Business Machines Corporation Deduplication in a hybrid storage environment
US20130141549A1 (en) 2010-06-29 2013-06-06 Cyclomedia Technology B.V. Method for Producing a Digital Photo Wherein at Least Some of the Pixels Comprise Position Information, and Such a Digital Photo
KR20130038388A (en) 2010-07-19 2013-04-17 프라운호퍼-게젤샤프트 츄어 푀르더룽 데어 안게반텐 포르슝에.파우. Image recording device and method for recording an image
USD659030S1 (en) 2010-07-30 2012-05-08 Sick Ag Optoelectronic sensor
WO2012055966A1 (en) 2010-10-27 2012-05-03 MAX-PLANCK-Gesellschaft zur Förderung der Wissenschaften e.V. Protecting data integrity with storage leases
US20120141104A1 (en) 2010-12-03 2012-06-07 Research In Motion Limited Dynamic lighting control in hybrid camera - projector device
US20120140109A1 (en) 2010-12-06 2012-06-07 Primesense Ltd. Lens Arrays for Pattern Projection and Imaging
US20120154914A1 (en) 2010-12-16 2012-06-21 Denso Corporation Interference filter assembly
US20130300840A1 (en) 2010-12-23 2013-11-14 Fastree 3D S.A. 3d landscape real-time imager and corresponding imaging methods
US20120182464A1 (en) 2011-01-18 2012-07-19 Primesense Ltd. Objective optics with interference filter
US8717488B2 (en) 2011-01-18 2014-05-06 Primesense Ltd. Objective optics with interference filter
JP2012202776A (en) 2011-03-24 2012-10-22 Toyota Central R&D Labs Inc Distance measuring device
US20120287417A1 (en) 2011-05-11 2012-11-15 Yvan Mimeault Multiple-field-of-view scannerless optical rangefinder in high ambient background light
US20120320164A1 (en) 2011-06-16 2012-12-20 Lenny Lipton Stereoscopic camera with polarizing apertures
US9176051B2 (en) 2011-06-29 2015-11-03 Karlsruher Institut Fuer Technologie Micro-optical element, micro-optical array, and optical sensor system
US8645810B2 (en) 2011-07-31 2014-02-04 Sandisk Technologies Inc. Fast detection of convergence or divergence in iterative decoding
US20130038941A1 (en) 2011-08-09 2013-02-14 Primesense Ltd. Lens Array Projector
US20130044187A1 (en) 2011-08-18 2013-02-21 Sick Ag 3d camera and method of monitoring a spatial zone
US9229109B2 (en) 2011-09-15 2016-01-05 Advanced Scientific Concepts, Inc. Automatic range corrected flash ladar camera
US20160150963A1 (en) 2011-09-26 2016-06-02 Michael Lee Roukes One-photon integrated neurophotonic systems
US20140267878A1 (en) 2011-11-04 2014-09-18 Imec Spectral camera with overlapping segments of image copies interleaved onto sensor array
US8762798B2 (en) 2011-11-16 2014-06-24 Stec, Inc. Dynamic LDPC code rate solution
EP2615566A2 (en) 2012-01-13 2013-07-17 Nexenta Systems, Inc. Unified local storage supporting file and cloud object access
US20140269796A1 (en) 2012-01-31 2014-09-18 Flir Systems, Inc. Multi-wavelength vcsel array to reduce speckle
US20130206967A1 (en) 2012-02-15 2013-08-15 Primesense Ltd. Integrated optoelectronic modules
US9157790B2 (en) 2012-02-15 2015-10-13 Apple Inc. Integrated optoelectronic modules with transmitter, receiver and beam-combining optics for aligning a beam axis with a collection axis
US9435891B2 (en) 2012-02-15 2016-09-06 Heptagon Micro Optics Pte. Ltd. Time of flight camera with stripe illumination
JP2013181912A (en) 2012-03-02 2013-09-12 Seiko Epson Corp Component analyzer
US20130229646A1 (en) 2012-03-02 2013-09-05 Seiko Epson Corporation Component analyzer
US20140153001A1 (en) 2012-03-22 2014-06-05 Primesense Ltd. Gimbaled scanning mirror array
US20140291491A1 (en) 2012-03-22 2014-10-02 Primesense Ltd. Calibration of time-of-flight measurement using stray reflections
US9489601B2 (en) 2012-09-05 2016-11-08 X Development Llc Construction zone sign detection
US9383753B1 (en) 2012-09-26 2016-07-05 Google Inc. Wide-view LIDAR with areas of special attention
US9071763B1 (en) 2012-09-26 2015-06-30 Google Inc. Uniform illumination image capture
KR20150072439A (en) 2012-10-19 2015-06-29 퀄컴 인코포레이티드 Multi-camera system using folded optics
US20140118335A1 (en) 2012-10-30 2014-05-01 Primesense Ltd. Depth mapping with enhanced resolution
US9111444B2 (en) 2012-10-31 2015-08-18 Raytheon Company Video and lidar target detection and tracking system and method for segmenting moving targets
US9281841B2 (en) 2012-10-31 2016-03-08 Avago Technologies General Ip (Singapore) Pte. Ltd. Load balanced decoding of low-density parity-check codes
US20150319349A1 (en) 2012-11-21 2015-11-05 Nokia Technologies Oy A Module for Plenoptic Camera System
US20140158900A1 (en) 2012-12-11 2014-06-12 Samsung Electronics Co., Ltd. Photon-counting detector and readout circuit
US20140168631A1 (en) 2012-12-18 2014-06-19 Pouch Holdings LLC Multi-clad Fiber Based Optical Apparatus and Methods for Light Detection and Ranging Sensors
US20140176933A1 (en) 2012-12-18 2014-06-26 Pouch Holdings LLC Light detecting and ranging sensing apparatus and methods
CN103048046A (en) 2012-12-21 2013-04-17 浙江大学 Double-beam spectrometer
US20160047901A1 (en) 2012-12-25 2016-02-18 Quanergy Systems, Inc. Robust lidar sensor for broad weather, shock and vibration conditions
US20150355470A1 (en) 2012-12-31 2015-12-10 Iee International Electronics & Engineering S.A. Optical system generating a structured light field from an array of light sources by means of a refracting or reflecting light structuring element
US9285477B1 (en) 2013-01-25 2016-03-15 Apple Inc. 3D depth point cloud from timing flight of 2D scanned light beam pulses
US20140211194A1 (en) 2013-01-27 2014-07-31 Quanergy Systems, Inc. Cost-effective lidar sensor for multi-signal detection, weak signal detection and signal disambiguation and method of using same
US9063549B1 (en) 2013-03-06 2015-06-23 Google Inc. Light detection and ranging device with oscillating mirror driven by magnetically interactive coil
US9086273B1 (en) 2013-03-08 2015-07-21 Google Inc. Microrod compression of laser beam in combination with transmit lens
US9470520B2 (en) 2013-03-14 2016-10-18 Apparate International C.V. LiDAR scanner
US20140313519A1 (en) 2013-03-15 2014-10-23 Primesense Ltd. Depth scanning with multiple emitters
WO2014150856A1 (en) 2013-03-15 2014-09-25 Pelican Imaging Corporation Array camera implementing quantum dot color filters
CN103234527A (en) 2013-04-07 2013-08-07 南京理工大学 Multispectral light-field camera
US9164511B1 (en) 2013-04-17 2015-10-20 Google Inc. Use of detected objects for image processing
US20150293224A1 (en) 2013-05-09 2015-10-15 Quanergy Systems, Inc. Solid state optical phased array lidar and method of using same
US20140376092A1 (en) 2013-06-19 2014-12-25 Primesense Ltd. Integrated structured-light projector
US20140375977A1 (en) 2013-06-21 2014-12-25 David Ludwig LIDAR Device and Method for Clear and Degraded Environmental Viewing Conditions
US20150002636A1 (en) 2013-06-28 2015-01-01 Cable Television Laboratories, Inc. Capturing Full Motion Live Events Using Spatially Distributed Depth Sensing Cameras
US9551791B2 (en) 2013-07-09 2017-01-24 Xenomatix Nv Surround sensing system
US20150362585A1 (en) 2013-07-12 2015-12-17 Princeton Optronics Inc. 2-D Planar VCSEL Source for 3-D Imaging
US20150260830A1 (en) 2013-07-12 2015-09-17 Princeton Optronics Inc. 2-D Planar VCSEL Source for 3-D Imaging
US8742325B1 (en) 2013-07-31 2014-06-03 Google Inc. Photodetector array on curved substrate
US20150292948A1 (en) 2013-08-02 2015-10-15 Verifood, Ltd. Spectrometry system with diffuser
US20160161600A1 (en) 2013-08-19 2016-06-09 Quanergy Systems, Inc. Optical phased array lidar system and method of using same
JP2016534346A (en) 2013-08-20 2016-11-04 グーグル インコーポレイテッド Apparatus and method for rotating LIDAR platform with shared transmission / light receiving path
US9285464B2 (en) 2013-08-20 2016-03-15 Google Inc. Devices and methods for a rotating LIDAR platform with a shared transmit/receive path
US8836922B1 (en) 2013-08-20 2014-09-16 Google Inc. Devices and methods for a rotating LIDAR platform with a shared transmit/receive path
WO2015037211A1 (en) 2013-09-11 2015-03-19 Sony Corporation Image processing device and method
US9425654B2 (en) 2013-09-30 2016-08-23 Google Inc. Contactless electrical coupling for a rotatable LIDAR device
US9368936B1 (en) 2013-09-30 2016-06-14 Google Inc. Laser diode firing system
US9299731B1 (en) 2013-09-30 2016-03-29 Google Inc. Systems and methods for selectable photodiode circuits
WO2015052616A1 (en) 2013-10-09 2015-04-16 Koninklijke Philips N.V. Monolithic led arrays for uniform and high-brightness light sources
US20150115131A1 (en) 2013-10-28 2015-04-30 Omnivision Technologies, Inc. Stacked chip spad image sensor
US20150124094A1 (en) 2013-11-05 2015-05-07 Delphi Technologies, Inc. Multiple imager vehicle optical sensor system
US20150131080A1 (en) 2013-11-12 2015-05-14 Facet Technology Corp. Methods and Apparatus for Array Based Lidar Systems with Reduced Interference
US20160306032A1 (en) 2013-11-22 2016-10-20 Brent S. SCHWARZ Lidar scanner calibration
US20180038959A1 (en) 2013-12-06 2018-02-08 Kazuhiro Akatsu Object detector and sensor
US20160328619A1 (en) 2013-12-24 2016-11-10 Lg Electronics Inc. Vehicle driving assistance device and vehicle having the same
US20150186287A1 (en) 2013-12-30 2015-07-02 Michael Henry Kass Using Memory System Programming Interfacing
US20150192677A1 (en) 2014-01-03 2015-07-09 Quanergy Systems, Inc. Distributed lidar sensing system for wide field of view three dimensional mapping and method of using same
US20160327779A1 (en) 2014-01-17 2016-11-10 The Trustees Of Columbia University In The City Of New York Systems And Methods for Three Dimensional Imaging
JP2015137987A (en) 2014-01-24 2015-07-30 アズビル株式会社 Distance sensor and distance measurement method
EP2908166A2 (en) 2014-02-17 2015-08-19 Yokogawa Electric Corporation Confocal optical scanner
WO2015136100A2 (en) 2014-03-14 2015-09-17 Mesa Imaging Ag Optoelectronic modules operable to recognize spurious reflections and to compensate for errors caused by spurious reflections
US20170038459A1 (en) 2014-03-14 2017-02-09 Heptogaon Micro Optics Pte Ltd. Optoelectronic modules operable to recognize spurious reflections and to compensate for errors caused by spurious reflections
WO2015163074A1 (en) 2014-04-22 2015-10-29 シャープ株式会社 Optical sensor system, optical-type gas sensor system, microparticle sensor system, light-emitting device, and image printing device
US20150316473A1 (en) 2014-05-01 2015-11-05 Rebellion Photonics, Inc. Mobile gas and chemical imaging camera
US20150379371A1 (en) 2014-06-30 2015-12-31 Microsoft Corporation Object Detection Utilizing Geometric Information Fused With Image Data
US20150378241A1 (en) 2014-06-30 2015-12-31 Quanergy Systems, Inc. Planar beam forming and steering optical phased array chip and method of using same
US20160003946A1 (en) 2014-07-03 2016-01-07 Advanced Scientific Concepts, Inc. Ladar sensor for a dense environment
US20160049765A1 (en) 2014-08-15 2016-02-18 Quanergy Systems, Inc. Three-dimensional-mapping two-dimensional-scanning lidar based on one-dimensional-steering optical phased arrays and method of using same
US20160047898A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Method and System for Ladar Transmission with Spiral Dynamic Scan Patterns
US20160047903A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Ladar Point Cloud Compression
US20160047900A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Method and System for Scanning Ladar Transmission with Pulse Modulation
US20160047899A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Method and System for Ladar Transmission Employing Dynamic Scan Patterns with Macro Patterns and Base Patterns
US20160047897A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Method and System for Ladar Transmission with Interline Skipping for Dynamic Scan Patterns
US20160047896A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Method and System for Ladar Transmission with Spinning Polygon Mirror for Dynamic Scan Patterns
US20160047895A1 (en) 2014-08-15 2016-02-18 US LADAR, Inc. Method and System for Ladar Transmission with Closed Loop Feedback Control of Dynamic Scan Patterns
US20170303829A1 (en) 2014-08-21 2017-10-26 I.r Med Ltd. System and method for noninvasive analysis of subcutaneous tissue
EP3002548A1 (en) 2014-10-02 2016-04-06 Sick Ag Illumination device and method for generating an illumination field
US20160097858A1 (en) 2014-10-06 2016-04-07 The Boeing Company Backfilling clouds of 3d coordinates
US20170299700A1 (en) 2014-10-20 2017-10-19 Quanergy Systems, Inc. Three-dimensional lidar sensor based on two-dimensional scanning of one-dimensional optical emitter and method of using same
US20180306907A1 (en) 2014-10-20 2018-10-25 Quanergy Systems, Inc. Three-dimensional lidar sensor based on two-dimensional scanning of one-dimensional optical emitter and method of using same
US20160218727A1 (en) 2014-10-31 2016-07-28 Seiko Epson Corporation Quantum interference device, atomic oscillator, electronic apparatus, and moving object
US9330464B1 (en) 2014-12-12 2016-05-03 Microsoft Technology Licensing, Llc Depth camera feedback
US10183541B2 (en) 2015-01-13 2019-01-22 Xenomatix Nv Surround sensing system with telecentric optics
EP3045935A1 (en) 2015-01-13 2016-07-20 XenomatiX BVBA Surround sensing system with dome-filter assembly
WO2016116733A1 (en) 2015-01-20 2016-07-28 Milan Momcilo Popovich Holographic waveguide lidar
WO2016125165A2 (en) 2015-02-05 2016-08-11 Verifood, Ltd. Spectrometry system with visible aiming beam
US9369689B1 (en) 2015-02-24 2016-06-14 HypeVR Lidar stereo fusion live action 3D model video reconstruction for six degrees of freedom 360° volumetric virtual reality video
US20160265902A1 (en) 2015-03-12 2016-09-15 Apple Inc. Thin psd for laser-scanning systems
US9529079B1 (en) 2015-03-26 2016-12-27 Google Inc. Multiplexed multichannel photodetector
US20160291134A1 (en) 2015-04-06 2016-10-06 Google Inc. Long Range Steerable LIDAR System
US9525863B2 (en) 2015-04-29 2016-12-20 Apple Inc. Time-of-flight depth mapping with flexible scan pattern
US20180152691A1 (en) 2015-09-24 2018-05-31 Ouster, Inc. Optical system for collecting distance information within a field
US20180167602A1 (en) 2015-09-24 2018-06-14 Ouster, Inc. Optical system for collecting distance information within a field
US9992477B2 (en) 2015-09-24 2018-06-05 Ouster, Inc. Optical system for collecting distance information within a field
US20170289524A1 (en) 2015-09-24 2017-10-05 Ouster, Inc. Optical System for Collecting Distance Information Within a Field
WO2018057084A1 (en) 2015-09-24 2018-03-29 Ouster, Inc. Optical system for collecting distance information within a field
US10063849B2 (en) 2015-09-24 2018-08-28 Ouster, Inc. Optical system for collecting distance information within a field
US20200036959A1 (en) * 2015-09-24 2020-01-30 Ouster, Inc. Optical system with multiple light emitters sharing a field of view of a pixel detector
US11025885B2 (en) 2015-09-24 2021-06-01 Ouster, Inc. Optical system for collecting distance information within a field
US9866241B2 (en) 2015-09-25 2018-01-09 SK Hynix Inc. Techniques for adaptive LDPC decoding
US20180299554A1 (en) 2015-10-23 2018-10-18 Xenomatix Nv System and method for determining a distance to an object
US20170146640A1 (en) 2015-11-25 2017-05-25 Velodyne Lidar, Inc. Three Dimensional LIDAR System With Targeted Field of View
US20170219426A1 (en) 2016-01-29 2017-08-03 Ouster, Inc. Systems and Methods for Calibrating an Optical Distance Sensor
US9989406B2 (en) 2016-01-29 2018-06-05 Ouster, Inc. Systems and methods for calibrating an optical distance sensor
US10557750B2 (en) 2016-01-29 2020-02-11 Ouster, Inc. Systems and methods for calibrating an optical distance sensor
AU2017212835A1 (en) 2016-01-29 2018-09-06 Ouster, Inc. Systems and methods for calibrating an optical distance sensor
US20180209841A1 (en) 2016-01-29 2018-07-26 Ouster, Inc. Systems and methods for calibrating an optical distance sensor
US20170219695A1 (en) 2016-01-31 2017-08-03 Velodyne Lidar, Inc. Multiple Pulse, LIDAR Based 3-D Imaging
WO2017132704A1 (en) 2016-01-31 2017-08-03 Velodyne Lidar, Inc. Lidar based 3-d imaging with far-field illumination overlap
US20170269215A1 (en) 2016-03-19 2017-09-21 Velodyne Lidar, Inc. Integrated Illumination And Detection For LIDAR Based 3-D Imaging
US20170269198A1 (en) 2016-03-21 2017-09-21 Velodyne Lidar, Inc. LIDAR Based 3-D Imaging With Varying Illumination Field Density
US20170269209A1 (en) 2016-03-21 2017-09-21 Velodyne Lidar, Inc. LIDAR Based 3-D Imaging With Varying Pulse Repetition
US20170269197A1 (en) 2016-03-21 2017-09-21 Velodyne Lidar, Inc. LIDAR Based 3-D Imaging With Varying Illumination Intensity
US10291261B2 (en) 2016-04-25 2019-05-14 SK Hynix Inc. Early selection decoding and automatic tuning
US20170350983A1 (en) 2016-06-01 2017-12-07 Velodyne Lidar, Inc. Multiple Pixel Scanning LIDAR
US20180032396A1 (en) 2016-07-29 2018-02-01 Sandisk Technologies Llc Generalized syndrome weights
US20180059222A1 (en) 2016-08-24 2018-03-01 Ouster, Inc. Optical system for collecting distance information within a field
US10809359B2 (en) 2016-08-24 2020-10-20 Ouster, Inc. Optical system for collecting distance information within a field
US20190018111A1 (en) 2016-08-24 2019-01-17 Ouster, Inc. Optical system for collecting distance information within a field
AU2017330180A1 (en) 2016-09-26 2019-05-16 Ouster, Inc. Optical system for collecting distance information within a field
WO2018065428A2 (en) 2016-10-03 2018-04-12 Xenomatix Nv System for determining a distance to an object
WO2018065429A1 (en) 2016-10-03 2018-04-12 Xenomatix Nv System and method for determining a distance to an object
WO2018065426A1 (en) 2016-10-03 2018-04-12 Xenomatix Nv Method for subtracting background light from an exposure value of a pixel in an imaging array, and pixel for use in same
WO2018065427A1 (en) 2016-10-03 2018-04-12 Xenomatix Nv System for determining a distance to an object
EP3316000A1 (en) 2016-10-28 2018-05-02 Xenomatix NV Vehicular system for measuring a distance to an object and method of installing same
WO2018122415A1 (en) 2016-12-30 2018-07-05 Xenomatix Nv System for characterizing surroundings of a vehicle
WO2018197441A1 (en) 2017-04-23 2018-11-01 Xenomatix Nv A pixel structure
US11086013B2 (en) * 2017-05-15 2021-08-10 Ouster, Inc. Micro-optics for imaging module with multiple converging lenses per channel
US10222475B2 (en) * 2017-05-15 2019-03-05 Ouster, Inc. Optical imaging transmitter with brightness enhancement
US10663586B2 (en) * 2017-05-15 2020-05-26 Ouster, Inc. Optical imaging transmitter with brightness enhancement
US20180329065A1 (en) 2017-05-15 2018-11-15 Ouster, Inc. Optical imaging transmitter with brightness enhancement
US10809380B2 (en) 2017-05-15 2020-10-20 Ouster, Inc. Augmenting panoramic LIDAR results with color
US20200041646A1 (en) 2017-07-05 2020-02-06 Ouster, Inc. Light ranging device with electronically scanned emitter array and synchronized sensor array
US20190097653A1 (en) 2017-09-22 2019-03-28 SK Hynix Memory Solutions America Inc. Memory system with decoders and method of operating such memory system and decoders
US20190103885A1 (en) 2017-09-29 2019-04-04 Nyquist Semiconductor Limited Systems and methods for decoding error correcting codes
US20190288713A1 (en) 2018-03-16 2019-09-19 SK Hynix Memory Solutions America Inc. Memory system with hybrid decoding scheme and method of operating such memory system
US20200116559A1 (en) * 2018-08-09 2020-04-16 Ouster, Inc. Bulk optics for a scanning array
US10739189B2 (en) * 2018-08-09 2020-08-11 Ouster, Inc. Multispectral ranging/imaging sensor arrays and systems
US10760957B2 (en) * 2018-08-09 2020-09-01 Ouster, Inc. Bulk optics for a scanning array
US20200116836A1 (en) 2018-08-09 2020-04-16 Ouster, Inc. Subpixel apertures for channels in a scanning sensor array
US20200116560A1 (en) 2018-08-09 2020-04-16 Ouster, Inc. Scanning sensor array with overlapping pass bands
US20200209355A1 (en) 2018-12-26 2020-07-02 Ouster, Inc. Solid-state electronic scanning laser array with high-side and low-side switches for increased channels

Non-Patent Citations (79)

* Cited by examiner, † Cited by third party
Title
"U.S. Appl. No. 17/317,809", Optical System for Collecting Distance Information Within a Field, filed May 11, 2021, 40 pages.
"U.S. Appl. No. 17/323,962", Optical System for Collecting Distance Information Within a Field, filed May 18, 2021, 39 pages.
Application No. CN201780015714.1 , Office Action, dated Mar. 11, 2021, 18 pages.
Application No. DKPA201970244 , Office Action, dated Feb. 12, 2021, 3 pages.
Application No. EP17745102.8 , Extended European Search Report, dated Sep. 9, 2019, 6 pages.
Application No. EP17745102.8 , Office Action, dated Jul. 16, 2020, 4 pages.
Application No. EP17745102.8 , Office Action, dated Mar. 6, 2020, 5 pages.
Application No. JP2018-559175 , Notice of Decision to Grant, dated Aug. 11, 2020, 1 page.
Application No. JP2018-559175 , Office Action, dated Dec. 23, 2019, 6 pages.
Application No. JP2019-516177 , Office Action, dated Apr. 1, 2021, 4 pages.
Application No. PCT/US2017/015683 , International Preliminary Report on Patentability, dated Aug. 9, 2018, 12 pages.
Application No. PCT/US2017/015683 , International Search Report and Written Opinion, dated Apr. 17, 2017, 12 pages.
Application No. RU2018130989 , Office Action, dated Apr. 27, 2020, 21 pages.
Application No. SG11201806442Y , Written Opinion, dated Jan. 28, 2020, 9 pages.
Application No. TH1801004505 , Office Action, dated Aug. 6, 2019, 2 pages.
Application No. ZA2018/05645 , Notice of Allowance, dated Sep. 3, 2019, 1 page.
AU2017212835 , "First Examination Report", dated Sep. 24, 2018, 5 pages.
AU2017212835 , "Notice of Acceptance", dated Feb. 28, 2019, 3 pages.
AU2017212835 , "Second Examination Report", dated Oct. 24, 2018, 3 pages.
AU2017330180 , "Notice of Acceptance", dated Jun. 17, 2019, 3 pages.
Bronzi, Danilo, "100 000 Frames/s 64x32 Single Photon Detector Array for 2-D Imaging and 3-D Ranging", IEEE Journal of Selected Topic in Quantum Electronics, vol. 20, No. 6, Nov./Dec. 2014; 10 pages.
Charbon, Edoardo, et al. "SPAD-Based Sensors"; TOF Range-Imaging Cameras, Remondino, F.,; Stoppa D. (Eds.), 2013, V, 240 p. 138 Illus., 85 Illus. In color., Hardcover ISBN 978-3-642-27522-7, 2013.
China National Intellectual Property Administration, Office Action received in Chinese application No. CN201780015714.1, dated Jun. 28, 2020, 10 pages.
CN201780015714.1, "Supplemental Search Report", dated Mar. 4, 2021, 1 page.
CN201780068584.8 , "Office Action", dated Nov. 3, 2020, 20 pages.
Contini et al., "Time-Domain Diffuse Optics: Towards Next Generation Devices", Proceedings SPIE Diffuse Optical Imaging V, vol. 9538, Jul. 10, 2015, 95380A page.
Cova, Sergio D.; Single-Photon Counting Detectors, IEEE Photonics Journal; vol. 3, No. 2, Apr. 2011, 5 pages.
DKPA201970244 , "Office Action", dated Apr. 24, 2020, 8 pages.
Guerrieri, Fabrizio, Two-Dimensional SPAD Imaging Camera for Photon Counting, vol. 2, No. 5, Oct. 2010, 17 pages.
IN201847031610 , "First Examination Report", dated Jan. 28, 2021, 5 pages.
Itzler, Mark A., "Geiger-mode avalance photodiode focal plane arrays for three-dimensional imaging LADAR" Princeton Lightwave, Inc., Proc of SPIE vol. 7808 780890C-, 14 pages.
KR10-2019-7011974 , "Office Action", dated Dec. 16, 2020, 4 pages.
Le et al., "On the Use of Hard-Decision LDPC Decoders on MLC NAND Flash Memory", 15th International Multi-Conference on Systems, Signals & Devices (SSD), Mar. 2018, 6 pages.
Non-Final Office Action dated Jul. 28, 2017 in U.S. Appl. No. 15/419,053, filed Jan. 30, 2017, 26 pages.
PCT/US2017/039306 , "International Preliminary Report on Patentability", dated Apr. 4, 2019, 17 pages.
PCTUS2017/048379 "International Search Report and Written Opinion" dated Nov. 2, 2017, 15 pages.
PCTUS2017039306 "International Search Report and Written Opinion" dated Nov. 7, 2017 21 pages.
PCTUs2017039306 "Invitation to Pay Add'l Fees and Partial Search Report" dated Aug. 31, 2017 2 pages.
Quanergy Systems Ex, 1005, Review of Scientific Instruments; vol. 72, No. 4, Apr. 2001, 13 pages.
RU2018130989 , "Notice of Decision to Grant", dated Dec. 14, 2020, 21 pages.
SE1950477-8 , "Office Action", dated Jan. 27, 2020, 8 pages.
SG11201806442Y , "Notice of Decision to Grant", dated Dec. 22, 2020, 5 pages.
U.S. Appl. No. 15/276,532 , "Non-Final Office Action", dated Dec. 15, 2017, 15 pages.
U.S. Appl. No. 15/276,532 , "Notice of Allowance", dated Apr. 11, 2018, 7 pages.
U.S. Appl. No. 15/419,053 , "Corrected Notice of Allowability", dated Mar. 12, 2018, 8 pages.
U.S. Appl. No. 15/419,053 , Notice of Allowance, dated Jan. 26, 2018, 12 pages.
U.S. Appl. No. 15/861,330 , "Notice of Allowance", dated May 29, 2018, 11 pages.
U.S. Appl. No. 15/880,491 , "Corrected Notice of Allowability", dated Feb. 25, 2021, 2 pages.
U.S. Appl. No. 15/880,491 , "Non-Final Office Action", dated Nov. 12, 2020, 9 pages.
U.S. Appl. No. 15/880,491 , "Notice of Allowability", dated Mar. 16, 2021, 2 pages.
U.S. Appl. No. 15/880,491 , "Notice of Allowance", dated Feb. 2, 2021, 7 pages.
U.S. Appl. No. 15/934,338 , "Corrected Notice of Allowability", dated Jan. 13, 2020, 2 pages.
U.S. Appl. No. 15/934,338 , Non-Final Office Action, dated May 15, 2019, 16 pages.
U.S. Appl. No. 15/934,338 , Notice of Allowance, dated Sep. 17, 2019, 10 pages.
U.S. Appl. No. 16/584,515 , Notice of Allowance, dated Mar. 12, 2021, 11 pages.
U.S. Appl. No. 16/584,515, Notice of Allowance, dated Aug. 13, 2021, 10 pages.
U.S. Appl. No. 17/317,809, filed May 11, 2021, 81 pages.
U.S. Appl. No. 17/317,809, Notice of Allowance, dated Aug. 18, 2021, 33 pages.
U.S. Appl. No. 17/323,962 , Non-Final Office Action, dated Aug. 31, 2021, 9 pages.
U.S. Appl. No. 17/323,962, "Notice of Allowance", dated Sep. 29, 2021, 7 pages.
U.S. Appl. No. 17/323,962, filed May 18, 2021, 78 pages.
Velodyne Lidar Inc., Velodyne Lidar Puck; User's Manual and Programming Guide; 2014.
Velodyne Lidar, Inc., HDL-32E Data Sheet 2010, 2017.
Velodyne Lidar, Inc., HDL-32E, HDL-32E, webpage: http://www.velodynelidar.com/hdl-32e.html; retrieved Dec. 6, 2017.
Velodyne Lidar, Inc., HDL-32E, User's Manual 2010.; Aug. 2016.
Velodyne Lidar, Inc., HDL-64E Data Sheet, 2017.
Velodyne Lidar, Inc., HDL-64E S2 and S2.1; User's Manual and Programming Guide 2007; Firmware version 4; 2007, revision Nov. 2012.
Velodyne Lidar, Inc., HDL-64E, S3; User's Manual and Programming Guide revision J; 2007.; Dec. 2017.
Velodyne Lidar, Inc., HDL-64E; webpage: http://www.velodynelidar.com/hdl-64e.html; retrieved Dec. 6, 2017.
Velodyne Lidar, Inc., VLP-16 data sheet, Puck, Real Time 3D Lidar Sensor, 2014.
Velodyne Lidar, Inc., VLP-16; retrieved via website: http://www.velodynelidar.com/vlp-16.html; Dec. 6, 2017.
Velodyne Lidar, Inc.,; Ultra Puck VLP-32C; Nov. 2017; retrieved via website: http://www.velodynelidar.com/vlp-32c.html; Dec. 13, 2017.
Velodyne Lidar, Inc.; Puck Hi-Res retrieved via website: http://www.velodynelidar.com/vlp-16-hi-res.html; Dec. 13, 2017.
Velodyne Lidar, Inc.; Puck Hi-Res User Manual; Sep. 2016.
Velodyne Lidar, Inc.; Puck Lite Data Sheet; Feb. 2016.
Velodyne Lidar, Inc.; Puck Lite User Manual; Feb. 2016.
Velodyne Lidar, Inc.; Puck Lite, Our Lightest Sensor Ever, Apr. 2016; retrieved via website: http://www.velodynelidar.com/vlp-16-lite.html; Dec. 13, 2017.
Velodyne, Lidar, Inc.; Puck Hi-Res Data Sheet; Sep. 2016.
Yang et al., "Wavelength Tuning of Vertical-Cavity Surface-Emitting Lasers by an Internal Device Heater", IEEE Photonics Technology Letters, vol. 20, Issue 20, Oct. 15, 2008, pp. 1679-1681.

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11627298B2 (en) 2015-09-24 2023-04-11 Ouster, Inc. Optical system for collecting distance information within a field
US11422236B2 (en) 2016-08-24 2022-08-23 Ouster, Inc. Optical system for collecting distance information within a field
US20210041570A1 (en) * 2017-05-15 2021-02-11 Ouster, Inc. Augmenting panoramic lidar results with color
US11789128B2 (en) 2021-03-01 2023-10-17 Innovusion, Inc. Fiber-based transmitter and receiver channels of light detection and ranging systems

Also Published As

Publication number Publication date
CN109983312A (en) 2019-07-05
US11178381B2 (en) 2021-11-16
IL265562B2 (en) 2023-06-01
KR102425304B1 (en) 2022-07-25
US11627298B2 (en) 2023-04-11
CA3038038A1 (en) 2018-03-29
US20220201267A1 (en) 2022-06-23
GB2569749B (en) 2022-04-13
KR102309478B1 (en) 2021-10-07
US20230319253A1 (en) 2023-10-05
CN114624728A (en) 2022-06-14
KR20190058588A (en) 2019-05-29
KR20210122905A (en) 2021-10-12
US20200036959A1 (en) 2020-01-30
US11202056B2 (en) 2021-12-14
GB2601435A (en) 2022-06-01
US20170289524A1 (en) 2017-10-05
US11196979B2 (en) 2021-12-07
SE544865C2 (en) 2022-12-13
GB201905868D0 (en) 2019-06-12
GB202202140D0 (en) 2022-04-06
DK180852B1 (en) 2022-05-19
US9992477B2 (en) 2018-06-05
JP2020501109A (en) 2020-01-16
DE112017004806T5 (en) 2019-06-19
US20180359460A1 (en) 2018-12-13
GB2569749A (en) 2019-06-26
US20210274148A1 (en) 2021-09-02
IL265562A (en) 2019-05-30
JP7214629B2 (en) 2023-01-30
US20210306609A1 (en) 2021-09-30
GB2601435B (en) 2022-11-16
WO2018057084A1 (en) 2018-03-29
US11025885B2 (en) 2021-06-01
CN109983312B (en) 2022-03-22
DK201970244A1 (en) 2019-04-30
SE1950477A1 (en) 2019-04-17
US20180152691A1 (en) 2018-05-31
DE202017007509U1 (en) 2022-03-15

Similar Documents

Publication Publication Date Title
US11196979B2 (en) Optical system for collecting distance information within a field
US10063849B2 (en) Optical system for collecting distance information within a field
AU2018269000B2 (en) Optical imaging transmitter with brightness enhancement
AU2017330180B2 (en) Optical system for collecting distance information within a field

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: OUSTER, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PACALA, ANGUS;FRICHTL, MARK;REEL/FRAME:047697/0993

Effective date: 20161125

AS Assignment

Owner name: SILICON VALLEY BANK, CALIFORNIA

Free format text: SECURITY INTEREST;ASSIGNOR:OUSTER, INC.;REEL/FRAME:052035/0339

Effective date: 20200303

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: OUSTER, INC., CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:SILICON VALLEY BANK;REEL/FRAME:059763/0484

Effective date: 20220425

AS Assignment

Owner name: HERCULES CAPITAL, INC., AS AGENT, CALIFORNIA

Free format text: SECURITY INTEREST;ASSIGNORS:OUSTER, INC.;SENSE PHOTONICS, INC.;REEL/FRAME:059859/0035

Effective date: 20220429

IPR Aia trial proceeding filed before the patent and appeal board: inter partes review

Free format text: TRIAL NO: IPR2023-01375

Opponent name: HESAI GROUP, HESAI TECHNOLOGY CO. LTD., AND HESAI, INC.

Effective date: 20230905

AS Assignment

Owner name: OUSTER, INC., CALIFORNIA

Free format text: RELEASE OF INTELLECTUAL PROPERTY SECURITY INTEREST AT REEL/FRAME NO. 059859/0035;ASSIGNOR:HERCULES CAPITAL, INC.;REEL/FRAME:065350/0826

Effective date: 20231025

IPR Aia trial proceeding filed before the patent and appeal board: inter partes review

Free format text: TRIAL NO: IPR2023-01457

Opponent name: HESAI GROUP, HESAI TECHNOLOGY CO. LTD., AND HESAI, INC.

Effective date: 20230925

AS Assignment

Owner name: SENSE PHOTONICS, INC., NORTH CAROLINA

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ADD THE SECOND ASSIGNEE PREVIOUSLY RECORDED AT REEL: 65350 FRAME: 826. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:HERCULES CAPITAL, INC.;REEL/FRAME:066432/0458

Effective date: 20231025

Owner name: OUSTER, INC., CALIFORNIA

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ADD THE SECOND ASSIGNEE PREVIOUSLY RECORDED AT REEL: 65350 FRAME: 826. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:HERCULES CAPITAL, INC.;REEL/FRAME:066432/0458

Effective date: 20231025