US20110187878A1 - Synchronization of projected illumination with rolling shutter of image sensor - Google Patents

Synchronization of projected illumination with rolling shutter of image sensor Download PDF

Info

Publication number
US20110187878A1
US20110187878A1 US12/762,373 US76237310A US2011187878A1 US 20110187878 A1 US20110187878 A1 US 20110187878A1 US 76237310 A US76237310 A US 76237310A US 2011187878 A1 US2011187878 A1 US 2011187878A1
Authority
US
United States
Prior art keywords
scene
radiation
radiation sources
image
rolling shutter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/762,373
Inventor
Zafrir Mor
Alexander Shpunt
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Apple Inc
Original Assignee
PrimeSense Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by PrimeSense Ltd filed Critical PrimeSense Ltd
Priority to US12/762,373 priority Critical patent/US20110187878A1/en
Assigned to PRIMESENSE LTD. reassignment PRIMESENSE LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MOR, ZAFRIR, SHPUNT, ALEXANDER
Priority to EP11150668A priority patent/EP2363686A1/en
Priority to US13/008,042 priority patent/US20110188054A1/en
Priority to JP2011009310A priority patent/JP2011160420A/en
Priority to CN201110035986.2A priority patent/CN102193295B/en
Priority to CN201110035990.9A priority patent/CN102143342B/en
Priority to CN201410641036.8A priority patent/CN104360571B/en
Publication of US20110187878A1 publication Critical patent/US20110187878A1/en
Priority to US13/765,706 priority patent/US9736459B2/en
Priority to US14/231,764 priority patent/US9091413B2/en
Assigned to APPLE INC. reassignment APPLE INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PRIMESENSE LTD.
Assigned to APPLE INC. reassignment APPLE INC. CORRECTIVE ASSIGNMENT TO CORRECT THE APPLICATION # 13840451 AND REPLACE IT WITH CORRECT APPLICATION # 13810451 PREVIOUSLY RECORDED ON REEL 034293 FRAME 0092. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT. Assignors: PRIMESENSE LTD.
Priority to US14/727,884 priority patent/US20150292709A1/en
Priority to US15/068,661 priority patent/US9740019B2/en
Priority to US15/648,473 priority patent/US10571709B2/en
Priority to US15/673,433 priority patent/US10063835B2/en
Priority to US16/045,779 priority patent/US10609357B2/en
Priority to US16/798,463 priority patent/US11320666B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/02Structural details or components not essential to laser action
    • H01S5/022Mountings; Housings
    • H01S5/023Mount members, e.g. sub-mount members
    • H01S5/02315Support members, e.g. bases or carriers
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B21/00Projectors or projection-type viewers; Accessories therefor
    • G03B21/14Details
    • G03B21/20Lamp housings
    • G03B21/2006Lamp housings characterised by the light source
    • G03B21/2033LED or laser light sources
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B19/00Condensers, e.g. light collectors or similar non-imaging optics
    • G02B19/0004Condensers, e.g. light collectors or similar non-imaging optics characterised by the optical means employed
    • G02B19/0028Condensers, e.g. light collectors or similar non-imaging optics characterised by the optical means employed refractive and reflective surfaces, e.g. non-imaging catadioptric systems
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B19/00Condensers, e.g. light collectors or similar non-imaging optics
    • G02B19/0033Condensers, e.g. light collectors or similar non-imaging optics characterised by the use
    • G02B19/0047Condensers, e.g. light collectors or similar non-imaging optics characterised by the use for use with a light source
    • G02B19/0052Condensers, e.g. light collectors or similar non-imaging optics characterised by the use for use with a light source the light source comprising a laser diode
    • G02B19/0057Condensers, e.g. light collectors or similar non-imaging optics characterised by the use for use with a light source the light source comprising a laser diode in the form of a laser diode array, e.g. laser diode bar
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/09Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
    • G02B27/0938Using specific optical elements
    • G02B27/0977Reflective elements
    • G02B27/0983Reflective elements being curved
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B15/00Special procedures for taking photographs; Apparatus therefor
    • G03B15/02Illuminating scene
    • G03B15/06Special arrangements of screening, diffusing, or reflecting devices, e.g. in studio
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B21/00Projectors or projection-type viewers; Accessories therefor
    • G03B21/14Details
    • G03B21/20Lamp housings
    • G03B21/2066Reflectors in illumination beam
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B21/00Projectors or projection-type viewers; Accessories therefor
    • G03B21/14Details
    • G03B21/28Reflectors in projection beam
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/02Structural details or components not essential to laser action
    • H01S5/022Mountings; Housings
    • H01S5/0225Out-coupling of light
    • H01S5/02255Out-coupling of light using beam deflecting elements
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/09Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
    • G02B27/0938Using specific optical elements
    • G02B27/0944Diffractive optical elements, e.g. gratings, holograms
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/09Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
    • G02B27/0938Using specific optical elements
    • G02B27/095Refractive optical elements
    • G02B27/0972Prisms
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L33/00Semiconductor devices with at least one potential-jump barrier or surface barrier specially adapted for light emission; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof
    • H01L33/48Semiconductor devices with at least one potential-jump barrier or surface barrier specially adapted for light emission; Processes or apparatus specially adapted for the manufacture or treatment thereof or of parts thereof; Details thereof characterised by the semiconductor body packages
    • H01L33/58Optical field-shaping elements
    • H01L33/60Reflective elements
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/005Optical components external to the laser cavity, specially adapted therefor, e.g. for homogenisation or merging of the beams or for manipulating laser pulses, e.g. pulse shaping
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/005Optical components external to the laser cavity, specially adapted therefor, e.g. for homogenisation or merging of the beams or for manipulating laser pulses, e.g. pulse shaping
    • H01S5/0071Optical components external to the laser cavity, specially adapted therefor, e.g. for homogenisation or merging of the beams or for manipulating laser pulses, e.g. pulse shaping for beam steering, e.g. using a mirror outside the cavity to change the beam direction
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/02Structural details or components not essential to laser action
    • H01S5/022Mountings; Housings
    • H01S5/0225Out-coupling of light
    • H01S5/02253Out-coupling of light using lenses
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/02Structural details or components not essential to laser action
    • H01S5/022Mountings; Housings
    • H01S5/0225Out-coupling of light
    • H01S5/02257Out-coupling of light using windows, e.g. specially adapted for back-reflecting light to a detector inside the housing
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/02Structural details or components not essential to laser action
    • H01S5/022Mountings; Housings
    • H01S5/023Mount members, e.g. sub-mount members
    • H01S5/02325Mechanically integrated components on mount members or optical micro-benches
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/40Arrangement of two or more semiconductor lasers, not provided for in groups H01S5/02 - H01S5/30
    • H01S5/4025Array arrangements, e.g. constituted by discrete laser diodes or laser bar
    • H01S5/4031Edge-emitting structures
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S5/00Semiconductor lasers
    • H01S5/40Arrangement of two or more semiconductor lasers, not provided for in groups H01S5/02 - H01S5/30
    • H01S5/42Arrays of surface emitting lasers
    • H01S5/423Arrays of surface emitting lasers having a vertical cavity

Definitions

  • the present invention relates generally to systems and methods for electronic imaging, and specifically to methods of illumination for enhancing the quality of captured images.
  • CMOS image sensors use a rolling shutter, in which successive rows of sensor elements are triggered sequentially to capture light. This method of image acquisition thus records each individual frame not as a single snapshot at a point in time, but rather as a sequence of image stripes scanning across the frame.
  • the result of the rolling shutter is that not all parts of the optical image are recorded at exactly the same time (although the frame is stored as a single electronic image).
  • imaging apparatus including an illumination assembly, including a plurality of radiation sources and projection optics, which are configured to project radiation from the radiation sources onto different, respective regions of a scene.
  • An imaging assembly includes an image sensor and objective optics configured to form an optical image of the scene on the image sensor, which includes an array of sensor elements arranged in multiple groups, which are triggered by a rolling shutter to capture the radiation from the scene in successive, respective exposure periods from different, respective areas of the scene so as to form an electronic image of the scene.
  • a controller is coupled to actuate the radiation sources sequentially in a pulsed mode so that the illumination assembly illuminates the different, respective areas of the scene in synchronization with the rolling shutter.
  • each group includes one or more rows of the sensor elements, and the regions define stripes that extend across the scene in a direction parallel to the rows.
  • each stripe illuminates a respective region that contains the areas of the scene from which the sensor elements in a respective set of multiple rows capture the radiation, and the controller is configured to actuate the radiation sources so that the projected radiation sweeps across the scene in a direction perpendicular to the rows.
  • the rolling shutter defines a frame time for capturing the entire electronic image
  • the controller is configured to actuate each of the radiation sources for a respective actuation period that is less than half the frame time.
  • the controller may actuate each of the radiation sources so that the illumination assembly illuminates each area of the scene only during a respective exposure period of a corresponding group of the sensor elements that captures the radiation from the area.
  • the projection optics include a patterning element, which is configured so that the radiation is projected onto the scene in a predefined pattern, which is detectable in the electronic image formed by the imaging assembly.
  • the controller is configured to analyze the pattern in the electronic image so as to generate a depth map of the scene.
  • the radiation sources include a matrix of light-emitting elements, which are arranged on a substrate and are configured to emit the radiation in a direction perpendicular to the substrate.
  • a method for imaging including arranging a plurality of radiation sources to project radiation onto different, respective regions of the scene.
  • An image sensor which includes an array of sensor elements arranged in multiple groups, is configured to receive an optical image of the scene, in which the groups of the sensor elements receive the radiation from different, respective areas of the scene.
  • the groups of the sensor elements are triggered with a rolling shutter to capture the radiation from the scene in successive, respective exposure periods so as to form an electronic image of the scene.
  • the radiation sources are actuated sequentially in a pulsed mode so as to illuminate the different, respective areas of the scene in synchronization with the rolling shutter.
  • configuring the image sensor includes arranging multiple image sensors, having respective rolling shutters, together with multiple, respective pluralities of the radiation sources to form respective electronic images of different, respective, overlapping parts of a scene, and actuating the radiation sources includes synchronizing the respective pluralities of the radiation sources over the multiple image sensors so as to control an overlap of the respective areas of the scene illuminated by the radiation sources at any given time.
  • the method may include analyzing the pattern over the electronic images formed by the multiple image sensors in order to generate a depth map of the scene.
  • imaging apparatus including multiple imaging units.
  • the imaging units include respective pluralities of radiation sources and projection optics, which are configured to project radiation from the radiation sources onto different, respective regions of a scene, and respective imaging assemblies.
  • the imaging assemblies include respective image sensors and objective optics configured to form respective optical images of different, respective, overlapping parts of the scene on the respective image sensors.
  • Each image sensor includes an array of sensor elements arranged in multiple groups, which are triggered by a rolling shutter to capture the radiation from the scene in successive, respective exposure periods from different, respective areas of the scene so as to form respective electronic images of the scene.
  • the radiation sources are actuated sequentially in a pulsed mode so that the illumination assembly illuminates the different, respective areas of the scene in synchronization with the rolling shutter, while synchronizing the respective pluralities of the radiation sources over the multiple image sensors so as to control an overlap of the respective areas of the scene illuminated by the radiation sources at any given time.
  • the overlap is controlled so that the respective areas of the scene illuminated by the radiation sources at any given time are non-overlapping.
  • FIG. 1 is a schematic side view of an imaging system, in accordance with an embodiment of the present invention.
  • FIG. 2 is a schematic representation of a portion of an image frame captured using stripe illumination, in accordance with an embodiment of the present invention
  • FIG. 3 is a timing diagram showing synchronization of stripe illumination with rolling shutter operation, in accordance with an embodiment of the present invention
  • FIG. 4A is a schematic side view of a projection module, in accordance with an embodiment of the present invention.
  • FIGS. 5A and 5B are schematic side and top views, respectively, of an optoelectronic subassembly, in accordance with another embodiment of the present invention.
  • FIG. 6 is a schematic side view of an illumination assembly, in accordance with an alternative embodiment of the present invention.
  • FIG. 7 is a schematic representation of a portion of an image frame illuminated by the illumination assembly of FIG. 6 ;
  • FIG. 8 is a schematic side view of an imaging system, in accordance with another embodiment of the present invention.
  • FIG. 9 is a schematic pictorial view of an imaging system, in accordance with yet another embodiment of the present invention.
  • imaging systems include optical projectors for illuminating the scene of interest.
  • a projector may be used to cast a pattern of coded or structured light onto an object for purposes of three-dimensional (3D) depth mapping.
  • 3D three-dimensional
  • U.S. Patent Application Publication 2008/0240502 whose disclosure is incorporated herein by reference, describes an illumination assembly in which a light source, such as a laser diode or LED, transilluminates a transparency with optical radiation so as to project a pattern onto the object.
  • the rolling shutter is operated so as to cause different groups (typically successive rows) of sensor elements in the image sensor to capture radiation in different, successive exposure periods, which are much shorter than the total frame period (typically less than half, and possibly less than 10%).
  • Each such group collects radiation from a different, respective area of the scene, which is focused onto the image sensor by objective optics.
  • the illumination assembly is controlled so as to sweep the projected radiation over those areas of the scene in synchronization with the rolling shutter, so that each area of the scene is illuminated during the specific time that the corresponding group of sensor elements is active.
  • the output power of the illumination assembly is concentrated, in each area of the scene, in the specific exposure periods during which the corresponding sensor elements are able to collect radiation from that area.
  • Limitation of the exposure periods by the rolling shutter reduces the total amount of ambient radiation that is collected, without wasting any of the projected radiation. Therefore, the signal/background ratio of the system is enhanced substantially even without increasing the average power of the illumination.
  • the principles of the present invention may similarly be applied to enhance the performance of other projection-based imaging systems.
  • the rolling shutter in these embodiments is assumed to activate the sensor elements in the image sensor row by row, as in conventional CMOS image sensors that are known in the art; but the principles of the present invention may similarly be applied in conjunction with image sensors that use other sorts of sequential activation of groups of sensor elements, such as block-by-block activation.
  • Imaging assembly 28 comprises objective optics 36 , which form an optical image of the scene containing object 26 on an image sensor 38 , such as a CMOS integrated circuit image sensor.
  • the image sensor comprises an array of sensor elements 40 , arranged in multiple rows.
  • the sensor elements generate respective signals in response to the radiation focused onto them by optics 36 , wherein the pixel value of each pixel in the electronic images output by image sensor 38 corresponds to the signal from a respective sensor element 40 .
  • the sensor elements are activated and deactivated, row by row, by a rolling shutter, whose timing is set by controller 31 . This sort of rolling shutter operation is a standard feature of many CMOS image sensors.
  • Illumination assembly 22 comprises a projection module 32 , which generates a beam of patterned light, and projection optics 34 , which project the beam onto field 24 .
  • Module 32 typically comprises multiple radiation sources, along with optics for pattern generation.
  • Controller 31 actuates the radiation sources sequentially, in a pulsed mode, in synchronization with the rolling shutter of image sensor 38 .
  • the design of module 32 and the synchronization of its operation with the rolling shutter are described in detail hereinbelow.
  • FIG. 2 is a schematic representation of a portion of an image frame 42 captured by system 20 , in accordance with an embodiment of the present invention.
  • Frame 42 comprises a matrix of pixels 44 , each corresponding to the signal generated by a corresponding sensor element 40 in image sensor 38 .
  • each row of pixels 44 corresponds to the area in the scene from which radiation is captured by the corresponding row of sensor elements.
  • Illumination assembly 22 generates multiple stripes 46 , 48 , 50 , 52 , . . . of illumination.
  • Each such stripe is generated by a respective radiation source or group of radiation sources.
  • the region defined by each stripe covers the area of a number of the rows of pixels 44 .
  • each stripe illuminates a certain area of the scene from which the image sensors in the corresponding rows capture radiation.
  • stripes 46 , 48 , 50 , 52 are shown in FIG. 2 , for the sake of simplicity, as being precisely adjacent to one another and non-overlapping, in practical systems there is generally a certain amount of overlap between the stripes in order to ensure that all areas of the scene are illuminated.
  • Traces 62 , 64 , . . . correspond to actuation of the respective radiation sources that generate stripes 46 , 48 , . . . .
  • the radiation source that generates stripe 46 is actuated, and so on.
  • the actuation period of the corresponding radiation source is set so as to fall entirely within the exposure periods of all the rows in the group.
  • the illumination assembly illuminates each area of the scene only during the exposure periods of the sensor elements that capture the radiation from the area, and none of the illumination is wasted.
  • the actuation period of each illumination stripe is approximately 1/(N*FR), while the exposure period of each row of sensor elements 40 is approximately 2/(N*FR), wherein FR is the frame rate, such as 30 frames/sec.
  • timing relations may be used between the frame rate, actuation periods and exposure times. These alternative timing arrangements may be advantageous in situations in which the geometrical relationships between illumination stripes and sensor rows are not maintained as precisely as in FIG. 2 , and particularly when successive stripes partially overlap.
  • the reflector may be integrally formed in substrate 72 , as shown in FIG. 4A , or it may alternatively comprise a separate element, which is positioned on the substrate and aligned with optoelectronic elements 70 .
  • Reflector 74 may simply comprise a flat reflecting surface, or it may alternatively comprise one or more curved surfaces or multiple flat surfaces in order to spread or focus the radiation, as illustrated in FIG. 4B , as well as FIG. 5C .
  • Each of optoelectronic elements 70 emits radiation that forms a respective stripe 80 , 82 , 84 , . . . , as shown in FIG. 4B .
  • Reflector 74 may be slightly curved, as shown in the figure, so that the stripes spread over a wider area and overlap the adjacent stripes at their edges.
  • controller 31 FIG. 1
  • each region of the scene is illuminated during the exposure periods of the corresponding rows of sensor elements 40 .
  • each stripe 80 , 82 , 84 , . . . passes through a different, respective region of the transparency, and thus creates a respective part of the overall illumination pattern corresponding to the pattern embedded in the transparency.
  • Projection optics 34 projects this pattern onto the object.
  • either lens 76 or one of elements 78 is typically configured to create an appropriate “carrier” angle for the beam emitted by each of the optoelectronic elements.
  • the beams emitted by the different optoelectronic elements use different parts of lens 76 , which may therefore be designed so that the collimated beams exit at respective angles corresponding to the desired vertical fan-out.
  • the illumination module may comprise some other type of optics, such as a blazed grating with as many different zones as there are optoelectronic elements.
  • illumination module 32 Further details of the fabrication of illumination module 32 , as well as other, similar sorts of modules, are described in the above-mentioned U.S. Provisional Patent Application 61/300,465.
  • FIGS. 5A and 5B are schematic side and top views, respectively, of an optoelectronic subassembly 90
  • FIG. 5C is a schematic pictorial view of a prism 92 used in subassembly 90 , in accordance with another embodiment of the present invention.
  • Subassembly 90 may be used in place of the corresponding components in module 32 .
  • Optoelectronic subassembly 90 comprises a row of edge-emitting optoelectronic elements 70 , such as laser diodes, which may be fabricated on a suitable substrate as in the preceding embodiment.
  • the radiation emitted by elements 70 is reflected internally from an interior surface 94 (typically with a suitable reflective coating) of prism 92 .
  • the radiation from elements 70 enters prism 92 via a curved entry surface 96 .
  • Controller 31 actuates elements 70 to emit radiation sequentially during each image frame in synchronization with the rolling shutter of image sensor 38
  • elements 110 comprise surface-emitting devices, such as light-emitting diodes (LEDs) or vertical-cavity surface-emitting laser (VCSEL) diodes, which emit radiation directly into the Z-direction.
  • An array of microlenses (or other suitable micro-optics, such as total internal reflection-based micro-structures) 112 is aligned with elements 110 , so that a respective microlens collects the radiation from each element and directs it into an optical module 104 .
  • the optical module comprises, inter alia, a suitable patterning element 106 , as described above, and a projection lens 108 , which projects the resulting pattern onto the scene.
  • FIG. 7 is a schematic representation of a portion of an image frame illuminated by assembly 100 , in accordance with an embodiment of the present invention.
  • Each microlens 112 spreads the radiation from the corresponding optoelectronic element 110 over a region of the scene that corresponds to a group of pixels 44 .
  • Elements 110 are arranged in multiple rows 114 , 116 , . . . .
  • controller 31 actuates all the optoelectronic elements in each row in turn in synchronization with the rolling shutter of image sensor 38 , in accordance with the scheme shown in FIG. 3 , for example.
  • the area of each pixel 44 is illuminated during the exposure period of the corresponding sensor element 40 .
  • FIG. 8 is a schematic side view of an imaging system 120 , in accordance with another embodiment of the present invention.
  • a synchronization controller 121 synchronizes the operation of multiple sensing units 122 , 124 , 126 , 128 .
  • Each of these sensing units typically comprises an illumination assembly and an imaging assembly, which operate in concert as in system 20 .
  • Each sensing unit 122 , 124 , 126 , 128 projects a respective patterned beam 132 , 134 , 136 , 138 onto a scene 130 and forms a respective image of the part of the scene that is illuminated by the respective pattern.
  • the projected patterned beams typically overlap in overlap regions 140 .
  • the overlap of the patterns could lead to inability of sensing units 122 , 124 , 126 , 128 to detect their own patterns reliably in regions 140 and thus to loss of 3D information in these regions.
  • One way to overcome this problem could be to operate the sensing units at different wavelengths, so that each unit senses only its own pattern. This solution, however, can be cumbersome and require costly optoelectronics and optical filters.
  • controller 121 controls the timing of the illumination assemblies and the rolling shutters of the imaging assemblies in sensing units 122 , 124 , 126 , 128 so as to control the overlap between the regions that are illuminated at any given time.
  • the sensing units are controlled so that they illuminate and capture radiation from respective non-overlapping stripes 142 , 144 , 146 , 148 .
  • the illumination stripe and the sensing area that is triggered to receive radiation by the rolling shutter are internally synchronized as described above. Furthermore, the timing of all the sensing units is coordinated to avoid interference.
  • each sensing unit simultaneously activate their respective stripes 142 , followed by stripes 144 , and so on, so that no more than a single sensing unit is active within each overlap region 140 at any given time.
  • Each sensing unit provides 3D mapping data with respect to its own part of scene 130 , and a processing unit (such as controller 121 or another computer) stitches the data together into a combined depth map.
  • sensing units 122 , 124 , 126 , 128 may operate together without a centralized controller to regulate synchronization. For example, each sensing unit may adjust its own timing so as to maximize its depth readings. Thus, the entire system will converge to an optimal synchronization. Additionally or alternatively, the sensing units may communicate with one another using a token ring type protocol, without centralized control.
  • FIG. 9 is a schematic pictorial view of an imaging system 150 , in accordance with yet another embodiment of the present invention. This embodiment is similar in its principles of operation to the embodiment of FIG. 8 : Multiple sensing units 152 , 154 , . . . , project respective patterned beams 156 , 158 , . . . , onto a scene, while controlling the timing of their respective illumination assemblies and rolling shutters so as to illuminate and capture radiation from respective sequences of stripes 160 . Beams 156 and 158 overlap in an overlap region 162 . Although for the sake of simplicity, only two sensing units are shown in FIG. 9 , any suitable number of sensing units may be arranged in this matter.

Abstract

Imaging apparatus includes an illumination assembly, including a plurality of radiation sources and projection optics, which are configured to project radiation from the radiation sources onto different, respective regions of a scene. An imaging assembly includes an image sensor and objective optics configured to form an optical image of the scene on the image sensor, which includes an array of sensor elements arranged in multiple groups, which are triggered by a rolling shutter to capture the radiation from the scene in successive, respective exposure periods from different, respective areas of the scene so as to form an electronic image of the scene. A controller is coupled to actuate the radiation sources sequentially in a pulsed mode so that the illumination assembly illuminates the different, respective areas of the scene in synchronization with the rolling shutter.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims the benefit of U.S. Provisional Patent Application 61/300,465, filed Feb. 2, 2010, which is incorporated herein by reference.
  • FIELD OF THE INVENTION
  • The present invention relates generally to systems and methods for electronic imaging, and specifically to methods of illumination for enhancing the quality of captured images.
  • BACKGROUND OF THE INVENTION
  • Most low-cost CMOS image sensors use a rolling shutter, in which successive rows of sensor elements are triggered sequentially to capture light. This method of image acquisition thus records each individual frame not as a single snapshot at a point in time, but rather as a sequence of image stripes scanning across the frame. The result of the rolling shutter is that not all parts of the optical image are recorded at exactly the same time (although the frame is stored as a single electronic image).
  • The use of a rolling shutter introduces a temporal shear in the image frame, which can create artifacts in imaging of moving objects. Bradley et al. address this problem in “Synchronization and Rolling Shutter Compensation for Consumer Video Camera Arrays,” IEEE International Workshop on Projector-Camera Systems—PROCAMS 2009 (Miami Beach, Fla., 2009), which is incorporated herein by reference. The authors propose to solve the problem using synchronized stroboscopic illumination.
  • SUMMARY
  • Embodiments of the present invention that are described hereinbelow provide apparatus and methods for illuminating an object that can be advantageous when the object is imaged using a sensor with a rolling shutter.
  • There is therefore provided, in accordance with an embodiment of the present invention, imaging apparatus, including an illumination assembly, including a plurality of radiation sources and projection optics, which are configured to project radiation from the radiation sources onto different, respective regions of a scene. An imaging assembly includes an image sensor and objective optics configured to form an optical image of the scene on the image sensor, which includes an array of sensor elements arranged in multiple groups, which are triggered by a rolling shutter to capture the radiation from the scene in successive, respective exposure periods from different, respective areas of the scene so as to form an electronic image of the scene. A controller is coupled to actuate the radiation sources sequentially in a pulsed mode so that the illumination assembly illuminates the different, respective areas of the scene in synchronization with the rolling shutter.
  • In disclosed embodiments, each group includes one or more rows of the sensor elements, and the regions define stripes that extend across the scene in a direction parallel to the rows. Typically, each stripe illuminates a respective region that contains the areas of the scene from which the sensor elements in a respective set of multiple rows capture the radiation, and the controller is configured to actuate the radiation sources so that the projected radiation sweeps across the scene in a direction perpendicular to the rows.
  • In a disclosed embodiment, the rolling shutter defines a frame time for capturing the entire electronic image, and the controller is configured to actuate each of the radiation sources for a respective actuation period that is less than half the frame time. The controller may actuate each of the radiation sources so that the illumination assembly illuminates each area of the scene only during a respective exposure period of a corresponding group of the sensor elements that captures the radiation from the area.
  • In some embodiments, the projection optics include a patterning element, which is configured so that the radiation is projected onto the scene in a predefined pattern, which is detectable in the electronic image formed by the imaging assembly. Typically, the controller is configured to analyze the pattern in the electronic image so as to generate a depth map of the scene. In one embodiment, the radiation sources include a matrix of light-emitting elements, which are arranged on a substrate and are configured to emit the radiation in a direction perpendicular to the substrate. In another embodiment, the radiation sources include a row of edge-emitting elements, which are arranged on a substrate and are configured to emit the radiation in a direction parallel to the substrate, and the illumination assembly includes a reflector disposed on the substrate so as to turn the radiation emitted by the edge-emitting elements away from the substrate and toward the patterning element.
  • There is also provided, in accordance with an embodiment of the present invention, a method for imaging, including arranging a plurality of radiation sources to project radiation onto different, respective regions of the scene. An image sensor, which includes an array of sensor elements arranged in multiple groups, is configured to receive an optical image of the scene, in which the groups of the sensor elements receive the radiation from different, respective areas of the scene. The groups of the sensor elements are triggered with a rolling shutter to capture the radiation from the scene in successive, respective exposure periods so as to form an electronic image of the scene. The radiation sources are actuated sequentially in a pulsed mode so as to illuminate the different, respective areas of the scene in synchronization with the rolling shutter.
  • In one embodiment, configuring the image sensor includes arranging multiple image sensors, having respective rolling shutters, together with multiple, respective pluralities of the radiation sources to form respective electronic images of different, respective, overlapping parts of a scene, and actuating the radiation sources includes synchronizing the respective pluralities of the radiation sources over the multiple image sensors so as to control an overlap of the respective areas of the scene illuminated by the radiation sources at any given time. The method may include analyzing the pattern over the electronic images formed by the multiple image sensors in order to generate a depth map of the scene.
  • There is additionally provided, in accordance with an embodiment of the present invention, imaging apparatus, including multiple imaging units. The imaging units include respective pluralities of radiation sources and projection optics, which are configured to project radiation from the radiation sources onto different, respective regions of a scene, and respective imaging assemblies. The imaging assemblies include respective image sensors and objective optics configured to form respective optical images of different, respective, overlapping parts of the scene on the respective image sensors. Each image sensor includes an array of sensor elements arranged in multiple groups, which are triggered by a rolling shutter to capture the radiation from the scene in successive, respective exposure periods from different, respective areas of the scene so as to form respective electronic images of the scene. The radiation sources are actuated sequentially in a pulsed mode so that the illumination assembly illuminates the different, respective areas of the scene in synchronization with the rolling shutter, while synchronizing the respective pluralities of the radiation sources over the multiple image sensors so as to control an overlap of the respective areas of the scene illuminated by the radiation sources at any given time.
  • Typically, the overlap is controlled so that the respective areas of the scene illuminated by the radiation sources at any given time are non-overlapping.
  • The present invention will be more fully understood from the following detailed description of the embodiments thereof, taken together with the drawings in which:
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a schematic side view of an imaging system, in accordance with an embodiment of the present invention;
  • FIG. 2 is a schematic representation of a portion of an image frame captured using stripe illumination, in accordance with an embodiment of the present invention;
  • FIG. 3 is a timing diagram showing synchronization of stripe illumination with rolling shutter operation, in accordance with an embodiment of the present invention;
  • FIG. 4A is a schematic side view of a projection module, in accordance with an embodiment of the present invention;
  • FIG. 4B is a schematic top view of an optoelectronic subassembly used in the projection module of FIG. 4A;
  • FIGS. 5A and 5B are schematic side and top views, respectively, of an optoelectronic subassembly, in accordance with another embodiment of the present invention;
  • FIG. 5C is a schematic pictorial view of a prism used in the subassembly of FIGS. 5A and 5B;
  • FIG. 6 is a schematic side view of an illumination assembly, in accordance with an alternative embodiment of the present invention;
  • FIG. 7 is a schematic representation of a portion of an image frame illuminated by the illumination assembly of FIG. 6;
  • FIG. 8 is a schematic side view of an imaging system, in accordance with another embodiment of the present invention; and
  • FIG. 9 is a schematic pictorial view of an imaging system, in accordance with yet another embodiment of the present invention.
  • DETAILED DESCRIPTION OF EMBODIMENTS OVERVIEW
  • Various types of imaging systems include optical projectors for illuminating the scene of interest. For example, a projector may be used to cast a pattern of coded or structured light onto an object for purposes of three-dimensional (3D) depth mapping. In this regard, U.S. Patent Application Publication 2008/0240502, whose disclosure is incorporated herein by reference, describes an illumination assembly in which a light source, such as a laser diode or LED, transilluminates a transparency with optical radiation so as to project a pattern onto the object. (The terms “optical,” “light” and “illumination” as used herein refer generally to any of visible, infrared, and ultraviolet radiation.) An image sensor captures an image of the pattern that is projected onto the object, and a processor processes the image so as to reconstruct a three-dimensional (3D) map of the object.
  • Systems based on projection of patterned light may suffer from low signal/background ratio due to limitations on the power of the projector, particularly in conditions of strong ambient light. Embodiments of the present invention address this problem by projecting radiation onto the scene of interest in a synchronized spatial sweep, which is timed to take advantage of the rolling shutter of the image sensor in order to improve the signal/background ratio of the system.
  • In embodiments of the present invention, the rolling shutter is operated so as to cause different groups (typically successive rows) of sensor elements in the image sensor to capture radiation in different, successive exposure periods, which are much shorter than the total frame period (typically less than half, and possibly less than 10%). Each such group collects radiation from a different, respective area of the scene, which is focused onto the image sensor by objective optics. The illumination assembly is controlled so as to sweep the projected radiation over those areas of the scene in synchronization with the rolling shutter, so that each area of the scene is illuminated during the specific time that the corresponding group of sensor elements is active. As a result, the output power of the illumination assembly is concentrated, in each area of the scene, in the specific exposure periods during which the corresponding sensor elements are able to collect radiation from that area. Limitation of the exposure periods by the rolling shutter reduces the total amount of ambient radiation that is collected, without wasting any of the projected radiation. Therefore, the signal/background ratio of the system is enhanced substantially even without increasing the average power of the illumination.
  • In the embodiments that are disclosed hereinbelow, the illumination assembly comprises an array of radiation sources, with projection optics that project radiation from the radiation sources onto different, respective regions of the scene. The spatial sweep of the projected radiation is accomplished by pulsing the radiation sources sequentially. The respective region of the scene that is illuminated by each radiation source overlaps the areas in the scene that are sensed by one or more of the groups of the sensor elements. Each radiation source is thus pulsed on only during the time that the corresponding groups of sensor elements are active. This sequential pulsed operation of the array of radiation sources provides full flexibility in choosing the optimal timing for the spatial sweep of radiation, as well as high reliability in that no moving parts or active optical elements (other than the radiation sources themselves) are required to implement the sweep.
  • Although the embodiments that are described below relate specifically to projection of patterned light in a 3D sensing system, the principles of the present invention may similarly be applied to enhance the performance of other projection-based imaging systems. The rolling shutter in these embodiments is assumed to activate the sensor elements in the image sensor row by row, as in conventional CMOS image sensors that are known in the art; but the principles of the present invention may similarly be applied in conjunction with image sensors that use other sorts of sequential activation of groups of sensor elements, such as block-by-block activation.
  • System Description
  • FIG. 1 is a schematic side view of an imaging system 20, in accordance with an embodiment of the present invention. A set of X-Y-Z axes is used in this figure and throughout the description that follows to aid in understanding the orientation of the figures, wherein the X-Y plane is the frontal plane of system 20, and the Z-axis extends perpendicularly from this plane toward the scene. The choice of axes, however, is arbitrary and is made solely for the sake of convenience in describing embodiments of the invention.
  • An illumination assembly 22 projects a patterned radiation field 24 onto an object 26 (in this case a hand of a user of the system) in a scene. An imaging assembly 28 captures an image of the scene within a field of view 30. A controller 31 or other electronic processor processes the image in order to generate a 3D depth map of object 26. Further details of this sort of mapping process are described, for example, in the above-mentioned US 2008/0240502 and in PCT International Publication WO 2007/105205, whose disclosure is also incorporated herein by reference. The 3D map of the user's hand (and/or other parts of the user's body) may be used in a gesture-based computer interface, but this sort of functionality is beyond the scope of the present patent application.
  • Imaging assembly 28 comprises objective optics 36, which form an optical image of the scene containing object 26 on an image sensor 38, such as a CMOS integrated circuit image sensor. The image sensor comprises an array of sensor elements 40, arranged in multiple rows. The sensor elements generate respective signals in response to the radiation focused onto them by optics 36, wherein the pixel value of each pixel in the electronic images output by image sensor 38 corresponds to the signal from a respective sensor element 40. The sensor elements are activated and deactivated, row by row, by a rolling shutter, whose timing is set by controller 31. This sort of rolling shutter operation is a standard feature of many CMOS image sensors.
  • Illumination assembly 22 comprises a projection module 32, which generates a beam of patterned light, and projection optics 34, which project the beam onto field 24. Module 32 typically comprises multiple radiation sources, along with optics for pattern generation. Controller 31 actuates the radiation sources sequentially, in a pulsed mode, in synchronization with the rolling shutter of image sensor 38. The design of module 32 and the synchronization of its operation with the rolling shutter are described in detail hereinbelow.
  • FIG. 2 is a schematic representation of a portion of an image frame 42 captured by system 20, in accordance with an embodiment of the present invention. Frame 42 comprises a matrix of pixels 44, each corresponding to the signal generated by a corresponding sensor element 40 in image sensor 38. Thus, each row of pixels 44 corresponds to the area in the scene from which radiation is captured by the corresponding row of sensor elements.
  • Illumination assembly 22 generates multiple stripes 46, 48, 50, 52, . . . of illumination. Each such stripe is generated by a respective radiation source or group of radiation sources. (Example arrangements of radiation sources that can be used to generate this sort of multi-stripe illumination are shown in the figures that follow.) The region defined by each stripe covers the area of a number of the rows of pixels 44. In other words, each stripe illuminates a certain area of the scene from which the image sensors in the corresponding rows capture radiation. Although stripes 46, 48, 50, 52 are shown in FIG. 2, for the sake of simplicity, as being precisely adjacent to one another and non-overlapping, in practical systems there is generally a certain amount of overlap between the stripes in order to ensure that all areas of the scene are illuminated.
  • FIG. 3 is a timing diagram illustrating synchronization of the sort of stripe illumination shown in FIG. 2 with the operation of a rolling shutter in image sensor 38, in accordance with an embodiment of the present invention. Traces 56 correspond to the operation of the rolling shutter on successive rows of sensor elements, wherein the elements are active (i.e., convert received photons to electrons in the output signal from the image sensor) when the corresponding trace is high. The period during which a given row is active is referred to herein as the exposure period of that row. The exposure periods of successive rows are staggered, so that each row is activated shortly after the preceding row. The rows are arranged in groups 58, 60, . . . , each group corresponding to the region covered by one of stripes 46, 48, . . . .
  • Traces 62, 64, . . . correspond to actuation of the respective radiation sources that generate stripes 46, 48, . . . . In other words, when trace 62 is high, the radiation source that generates stripe 46 is actuated, and so on. For each group 58, 60, . . . , of the rows, the actuation period of the corresponding radiation source is set so as to fall entirely within the exposure periods of all the rows in the group. Thus, the illumination assembly illuminates each area of the scene only during the exposure periods of the sensor elements that capture the radiation from the area, and none of the illumination is wasted.
  • Trace 64 goes high just as trace 62 goes low, and so forth over all the radiation sources in illumination assembly 22. Thus, the stripe output of the illumination assembly sweeps across the scene in a sweep direction perpendicular to the rows of pixels 44 (and sensor elements 40), completing one such sweep in each image frame, in synchronization with the sweep of the rolling shutter of image sensor 38. The duty cycle of each radiation source is roughly 1:N, wherein N is the number of stripes (each illuminated by a respective radiation source or group of radiation sources). In the timing scheme of FIG. 3, the actuation period of each illumination stripe is approximately 1/(N*FR), while the exposure period of each row of sensor elements 40 is approximately 2/(N*FR), wherein FR is the frame rate, such as 30 frames/sec. These timing relations typically make optimal use of the available illumination power and provide the greatest possible enhancement of signal/background ratio.
  • Alternatively, other timing relations may be used between the frame rate, actuation periods and exposure times. These alternative timing arrangements may be advantageous in situations in which the geometrical relationships between illumination stripes and sensor rows are not maintained as precisely as in FIG. 2, and particularly when successive stripes partially overlap.
  • Illumination Module with Edge Emitters
  • FIG. 4A is a schematic side view of illumination module 32, while FIG. 4B is a schematic top view of an optoelectronic subassembly used in illumination module 32, in accordance with an embodiment of the present invention. Module 32 comprises a row of edge-emitting optoelectronic elements 70, such as laser diodes, which are formed on a substrate 72, such as a silicon wafer. (Only one of the elements can be seen in the side view of FIG. 4A.) Elements 70 emit radiation in a direction parallel to the substrate. A reflector 74 on the substrate turns the radiation emitted by elements 70 away from the substrate, which is oriented in the X-Y plane, toward the Z-axis. The reflector may be integrally formed in substrate 72, as shown in FIG. 4A, or it may alternatively comprise a separate element, which is positioned on the substrate and aligned with optoelectronic elements 70. Reflector 74 may simply comprise a flat reflecting surface, or it may alternatively comprise one or more curved surfaces or multiple flat surfaces in order to spread or focus the radiation, as illustrated in FIG. 4B, as well as FIG. 5C.
  • A collecting lens 76 collimates and directs the radiation from optoelectronic elements 70 through one or more patterning elements 78. The patterning elements cause the radiation from elements 70 to be projected onto the scene in a predefined pattern, which is detectable in the electronic image formed by imaging assembly 28. This pattern in the image is processed in order to compute the depth map of the scene. Patterning elements 78 may comprise a patterned transparency, which may comprise a micro-lens array (MLA), as described, for example, in the above-mentioned US 2008/0240502 or WO 2007/105205, and/or one or more diffractive optical elements (DOEs), as described in U.S. Patent Application Publication 2009/0185274, whose disclosure is also incorporated herein by reference. Additionally or alternatively, when elements 70 emit coherent radiation, patterning elements 78 may comprise a diffuser, which casts a laser speckle pattern on the scene.
  • Each of optoelectronic elements 70 emits radiation that forms a respective stripe 80, 82, 84, . . . , as shown in FIG. 4B. (Although the figure shows six such elements and respective stripes, a larger or smaller number of elements and stripes may be used, depending on application requirements.) Reflector 74 may be slightly curved, as shown in the figure, so that the stripes spread over a wider area and overlap the adjacent stripes at their edges. As explained above, controller 31 (FIG. 1) activates elements 70 to emit radiation sequentially, in synchronization with the rolling shutter of image sensor 38, during each image frame captured by imaging assembly 28. Thus, each region of the scene is illuminated during the exposure periods of the corresponding rows of sensor elements 40.
  • In embodiments in which patterning elements 78 comprise a MLA or other transparency, each stripe 80, 82, 84, . . . , passes through a different, respective region of the transparency, and thus creates a respective part of the overall illumination pattern corresponding to the pattern embedded in the transparency. Projection optics 34 projects this pattern onto the object.
  • On the other hand, in embodiments in which patterning elements 78 comprise a DOE, either lens 76 or one of elements 78 (or the geometry of optoelectronic elements 70) is typically configured to create an appropriate “carrier” angle for the beam emitted by each of the optoelectronic elements. In such embodiments, the beams emitted by the different optoelectronic elements use different parts of lens 76, which may therefore be designed so that the collimated beams exit at respective angles corresponding to the desired vertical fan-out. Alternatively, the illumination module may comprise some other type of optics, such as a blazed grating with as many different zones as there are optoelectronic elements.
  • Further details of the fabrication of illumination module 32, as well as other, similar sorts of modules, are described in the above-mentioned U.S. Provisional Patent Application 61/300,465.
  • FIGS. 5A and 5B are schematic side and top views, respectively, of an optoelectronic subassembly 90, while FIG. 5C is a schematic pictorial view of a prism 92 used in subassembly 90, in accordance with another embodiment of the present invention. Subassembly 90 may be used in place of the corresponding components in module 32.
  • Optoelectronic subassembly 90 comprises a row of edge-emitting optoelectronic elements 70, such as laser diodes, which may be fabricated on a suitable substrate as in the preceding embodiment. In subassembly 90, however, the radiation emitted by elements 70 is reflected internally from an interior surface 94 (typically with a suitable reflective coating) of prism 92. The radiation from elements 70 enters prism 92 via a curved entry surface 96. As a result, respective beams generated by elements 70 spread apart and overlap partially with the adjacent beams. Controller 31 actuates elements 70 to emit radiation sequentially during each image frame in synchronization with the rolling shutter of image sensor 38
  • Illumination Module with Surface Emitters
  • FIG. 6 is a schematic side view of an illumination assembly 100, in accordance with an alternative embodiment of the present invention. Assembly 100 may be used in system 20 in place of illumination assembly 22. Assembly 100 comprises radiation sources in the form of a two-dimensional matrix of optoelectronic elements 110, which are arranged on a substrate 102 and emit radiation in a direction perpendicular to the substrate. Although FIG. 6 shows only a single row 114 of elements arrayed along the X-axis, assembly 100 actually comprises multiple, parallel rows of this sort, forming a grid in the X-Y plane. FIG. 6 illustrates an 8×8 grid, but larger or smaller matrices, not necessarily square or rectilinear, may alternatively be used.
  • In contrast to the preceding embodiments, elements 110 comprise surface-emitting devices, such as light-emitting diodes (LEDs) or vertical-cavity surface-emitting laser (VCSEL) diodes, which emit radiation directly into the Z-direction. An array of microlenses (or other suitable micro-optics, such as total internal reflection-based micro-structures) 112 is aligned with elements 110, so that a respective microlens collects the radiation from each element and directs it into an optical module 104. The optical module comprises, inter alia, a suitable patterning element 106, as described above, and a projection lens 108, which projects the resulting pattern onto the scene.
  • FIG. 7 is a schematic representation of a portion of an image frame illuminated by assembly 100, in accordance with an embodiment of the present invention. Each microlens 112 spreads the radiation from the corresponding optoelectronic element 110 over a region of the scene that corresponds to a group of pixels 44. (Typically there is some overlap between neighboring regions, as in the preceding embodiments.) Elements 110 are arranged in multiple rows 114, 116, . . . . In typical operation, controller 31 actuates all the optoelectronic elements in each row in turn in synchronization with the rolling shutter of image sensor 38, in accordance with the scheme shown in FIG. 3, for example. Thus, as described above, the area of each pixel 44 is illuminated during the exposure period of the corresponding sensor element 40.
  • Although the above embodiments are described, for the sake of clarity, in the context of system 20 and certain specific geometrical configurations of illumination and sensing, the principles of the present invention may similarly be applied in systems and configurations of other sorts.
  • Synchronization Over Multiple Sensors
  • FIG. 8 is a schematic side view of an imaging system 120, in accordance with another embodiment of the present invention. In this system, a synchronization controller 121 synchronizes the operation of multiple sensing units 122, 124, 126, 128. Each of these sensing units typically comprises an illumination assembly and an imaging assembly, which operate in concert as in system 20. Each sensing unit 122, 124, 126, 128 projects a respective patterned beam 132, 134, 136, 138 onto a scene 130 and forms a respective image of the part of the scene that is illuminated by the respective pattern.
  • In order to cover scene 130 completely, the projected patterned beams typically overlap in overlap regions 140. In conventional operation, the overlap of the patterns could lead to inability of sensing units 122, 124, 126, 128 to detect their own patterns reliably in regions 140 and thus to loss of 3D information in these regions. One way to overcome this problem could be to operate the sensing units at different wavelengths, so that each unit senses only its own pattern. This solution, however, can be cumbersome and require costly optoelectronics and optical filters.
  • Therefore, in system 120, controller 121 controls the timing of the illumination assemblies and the rolling shutters of the imaging assemblies in sensing units 122, 124, 126, 128 so as to control the overlap between the regions that are illuminated at any given time. Typically, the sensing units are controlled so that they illuminate and capture radiation from respective non-overlapping stripes 142, 144, 146, 148. Within each sensing unit, the illumination stripe and the sensing area that is triggered to receive radiation by the rolling shutter are internally synchronized as described above. Furthermore, the timing of all the sensing units is coordinated to avoid interference. Thus, for example, all of the sensing units simultaneously activate their respective stripes 142, followed by stripes 144, and so on, so that no more than a single sensing unit is active within each overlap region 140 at any given time. Each sensing unit provides 3D mapping data with respect to its own part of scene 130, and a processing unit (such as controller 121 or another computer) stitches the data together into a combined depth map.
  • The scheme illustrated in FIG. 8 is just one example of a possible synchronization pattern, and alternative geometrical and timing patterns may also be implemented to achieve similar objectives. For example, the synchronized sensing units may be arranged in a two-dimensional array in order to cover a wider area of scene 130. Depending on the geometrical arrangement and the timing of the sensing units, systems of multiple synchronized sensing units may be used to capture depth information over greater areas of substantially any desired size and profile, or, alternatively or additionally, with greater speed.
  • Alternatively, sensing units 122, 124, 126, 128 may operate together without a centralized controller to regulate synchronization. For example, each sensing unit may adjust its own timing so as to maximize its depth readings. Thus, the entire system will converge to an optimal synchronization. Additionally or alternatively, the sensing units may communicate with one another using a token ring type protocol, without centralized control.
  • FIG. 9 is a schematic pictorial view of an imaging system 150, in accordance with yet another embodiment of the present invention. This embodiment is similar in its principles of operation to the embodiment of FIG. 8: Multiple sensing units 152, 154, . . . , project respective patterned beams 156, 158, . . . , onto a scene, while controlling the timing of their respective illumination assemblies and rolling shutters so as to illuminate and capture radiation from respective sequences of stripes 160. Beams 156 and 158 overlap in an overlap region 162. Although for the sake of simplicity, only two sensing units are shown in FIG. 9, any suitable number of sensing units may be arranged in this matter.
  • In system 150, however, sensing units 152 and 154 and their beams 156 and 158 are offset from one another in a direction perpendicular to the scan direction of the illumination and rolling shutter (horizontal offset with vertical scan in the view shown in FIG. 9), as opposed to the parallel offset shown in FIG. 8. Therefore, most or all of stripes 160 may overlap with certain stripes of the neighboring sensing unit. The scans of sensing units 152, 154, . . . , are therefore synchronized so that each stripe is illuminated in different time periods from its overlapping neighbors. As shown in FIG. 9, there is no need for precise overlap between stripes 160 of the different sensing units, nor do the stripes need to be exactly parallel. Generally speaking, the sensing units may be arranged in any desired arrangement, as long as the synchronization schedule can make overlapping stripes disjoint in time.
  • It will thus be appreciated that the embodiments described above are cited by way of example, and that the present invention is not limited to what has been particularly shown and described hereinabove. Rather, the scope of the present invention includes both combinations and subcombinations of the various features described hereinabove, as well as variations and modifications thereof which would occur to persons skilled in the art upon reading the foregoing description and which are not disclosed in the prior art.

Claims (27)

1. Imaging apparatus, comprising:
an illumination assembly, comprising a plurality of radiation sources and projection optics, which are configured to project radiation from the radiation sources onto different, respective regions of a scene;
an imaging assembly, comprising an image sensor and objective optics configured to form an optical image of the scene on the image sensor, which comprises an array of sensor elements arranged in multiple groups, which are triggered by a rolling shutter to capture the radiation from the scene in successive, respective exposure periods from different, respective areas of the scene so as to form an electronic image of the scene; and
a controller, which is coupled to actuate the radiation sources sequentially in a pulsed mode so that the illumination assembly illuminates the different, respective areas of the scene in synchronization with the rolling shutter.
2. The apparatus according to claim 1, wherein each group comprises one or more rows of the sensor elements, and wherein the regions define stripes that extend across the scene in a direction parallel to the rows.
3. The apparatus according to claim 2, wherein each stripe illuminates a respective region that contains the areas of the scene from which the sensor elements in a respective set of multiple rows capture the radiation.
4. The apparatus according to claim 2, wherein the controller is configured to actuate the radiation sources so that the projected radiation sweeps across the scene in a direction perpendicular to the rows.
5. The apparatus according to claim 1, wherein the rolling shutter defines a frame time for capturing the entire electronic image, and wherein the controller is configured to actuate each of the radiation sources for a respective actuation period that is less than half the frame time.
6. The apparatus according to claim 5, wherein the controller is configured to actuate each of the radiation sources so that the illumination assembly illuminates each area of the scene only during a respective exposure period of a corresponding group of the sensor elements that captures the radiation from the area.
7. The apparatus according to claim 1, wherein the projection optics comprise a patterning element, which is configured so that the radiation is projected onto the scene in a predefined pattern, which is detectable in the electronic image formed by the imaging assembly.
8. The apparatus according to claim 7, wherein the controller is configured to analyze the pattern in the electronic image so as to generate a depth map of the scene.
9. The apparatus according to claim 7, wherein the radiation sources comprise a matrix of light-emitting elements, which are arranged on a substrate and are configured to emit the radiation in a direction perpendicular to the substrate.
10. The apparatus according to claim 7, wherein the radiation sources comprise a row of edge-emitting elements, which are arranged on a substrate and are configured to emit the radiation in a direction parallel to the substrate.
11. The apparatus according to claim 10, wherein the illumination assembly comprises a reflector disposed on the substrate so as to turn the radiation emitted by the edge-emitting elements away from the substrate and toward the patterning element.
12. A method for imaging, comprising:
arranging a plurality of radiation sources to project radiation onto different, respective regions of the scene;
configuring an image sensor, which comprises an array of sensor elements arranged in multiple groups, to receive an optical image of the scene, in which the groups of the sensor elements receive the radiation from different, respective areas of the scene;
triggering the groups of the sensor elements with a rolling shutter to capture the radiation from the scene in successive, respective exposure periods so as to form an electronic image of the scene; and
actuating the radiation sources sequentially in a pulsed mode so as to illuminate the different, respective areas of the scene in synchronization with the rolling shutter.
13. The method according to claim 12, wherein each group comprises one or more rows of the sensor elements, and wherein the regions define stripes that extend across the scene in a direction parallel to the rows.
14. The method according to claim 13, wherein each stripe illuminates a respective region that contains the areas of the scene from which the sensor elements in a respective set of multiple rows capture the radiation.
15. The method according to claim 13, wherein actuating the radiation sources causes the projected radiation to sweep across the scene in a direction perpendicular to the rows.
16. The method according to claim 12, wherein the rolling shutter defines a frame time for capturing the entire electronic image, and wherein actuating the radiation sources comprises actuating each of the radiation sources for a respective actuation period that is less than half the frame time.
17. The method according to claim 16, wherein actuating each of the radiation sources comprises illuminating each area of the scene only during a respective exposure period of a corresponding group of the sensor elements that captures the radiation from the area.
18. The method according to claim 12, wherein arranging the plurality of the radiation source comprises projecting the radiation onto the scene in a predefined pattern, which is detectable in the electronic image.
19. The method according to claim 18, and comprising analyzing the pattern in the electronic image so as to generate a depth map of the scene.
20. The method according to claim 18, wherein the radiation sources comprise a matrix of light-emitting elements, which are arranged on a substrate and are configured to emit the radiation in a direction perpendicular to the substrate.
21. The method according to claim 18, wherein the radiation sources comprise a row of edge-emitting elements, which are arranged on a substrate and are configured to emit the radiation in a direction parallel to the substrate.
22. The method according to claim 21, wherein arranging the plurality of the radiation sources comprises providing a reflector on the substrate so as to turn the radiation emitted by the edge-emitting elements away from the substrate and toward the scene.
23. The method according to claim 12, wherein configuring the image sensor comprises arranging multiple image sensors, having respective rolling shutters, together with multiple, respective pluralities of the radiation sources to form respective electronic images of different, respective, overlapping parts of a scene, and
wherein actuating the radiation sources comprises synchronizing the respective pluralities of the radiation sources over the multiple image sensors so as to control an overlap of the respective areas of the scene illuminated by the radiation sources at any given time.
24. The method according to claim 23, and comprising analyzing the pattern over the electronic images formed by the multiple image sensors in order to generate a depth map of the scene.
25. Imaging apparatus, comprising:
multiple imaging units, comprising:
respective pluralities of radiation sources and projection optics, which are configured to project radiation from the radiation sources onto different, respective regions of a scene; and
respective imaging assemblies, comprising respective image sensors and objective optics configured to form respective optical images of different, respective, overlapping parts of the scene on the respective image sensors, each image sensor comprising an array of sensor elements arranged in multiple groups, which are triggered by a rolling shutter to capture the radiation from the scene in successive, respective exposure periods from different, respective areas of the scene so as to form respective electronic images of the scene,
wherein the radiation sources are actuated sequentially in a pulsed mode so that the illumination assembly illuminates the different, respective areas of the scene in synchronization with the rolling shutter, while synchronizing the respective pluralities of the radiation sources over the multiple image sensors so as to control an overlap of the respective areas of the scene illuminated by the radiation sources at any given time.
26. The apparatus according to claim 25, and comprising a controller, which is configured to analyze the pattern over the electronic images formed by the image sensors in the multiple imaging units so as to generate a depth map of the scene.
27. The apparatus according to claim 25, wherein the overlap is controlled so that the respective areas of the scene illuminated by the radiation sources at any given time are non-overlapping.
US12/762,373 2010-02-02 2010-04-19 Synchronization of projected illumination with rolling shutter of image sensor Abandoned US20110187878A1 (en)

Priority Applications (15)

Application Number Priority Date Filing Date Title
US12/762,373 US20110187878A1 (en) 2010-02-02 2010-04-19 Synchronization of projected illumination with rolling shutter of image sensor
EP11150668A EP2363686A1 (en) 2010-02-02 2011-01-11 Optical apparatus, an imaging system and a method for producing a photonics module
US13/008,042 US20110188054A1 (en) 2010-02-02 2011-01-18 Integrated photonics module for optical projection
JP2011009310A JP2011160420A (en) 2010-02-02 2011-01-19 Synchronization of projected illumination with rolling shutter of image sensor
CN201410641036.8A CN104360571B (en) 2010-02-02 2011-02-01 Optical device and imaging system
CN201110035986.2A CN102193295B (en) 2010-02-02 2011-02-01 Integrated photonics module for optical projection
CN201110035990.9A CN102143342B (en) 2010-02-02 2011-02-01 Synchronization of projected illumination with rolling shutter of image sensor
US13/765,706 US9736459B2 (en) 2010-02-02 2013-02-13 Generation of patterned radiation
US14/231,764 US9091413B2 (en) 2010-02-02 2014-04-01 Integrated photonics module for optical projection
US14/727,884 US20150292709A1 (en) 2010-02-02 2015-06-02 Integrated photonics module for optical projection
US15/068,661 US9740019B2 (en) 2010-02-02 2016-03-14 Integrated structured-light projector
US15/648,473 US10571709B2 (en) 2010-02-02 2017-07-13 Integrated structured-light projector
US15/673,433 US10063835B2 (en) 2010-02-02 2017-08-10 Generation of patterned radiation
US16/045,779 US10609357B2 (en) 2010-02-02 2018-07-26 Synchronization of projected illumination with activation of sensing elements
US16/798,463 US11320666B2 (en) 2010-02-02 2020-02-24 Integrated structured-light projector

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US30046510P 2010-02-02 2010-02-02
US12/762,373 US20110187878A1 (en) 2010-02-02 2010-04-19 Synchronization of projected illumination with rolling shutter of image sensor

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US13/008,042 Continuation-In-Part US20110188054A1 (en) 2010-02-02 2011-01-18 Integrated photonics module for optical projection
US13/765,706 Continuation US9736459B2 (en) 2010-02-02 2013-02-13 Generation of patterned radiation

Publications (1)

Publication Number Publication Date
US20110187878A1 true US20110187878A1 (en) 2011-08-04

Family

ID=44341315

Family Applications (4)

Application Number Title Priority Date Filing Date
US12/762,373 Abandoned US20110187878A1 (en) 2010-02-02 2010-04-19 Synchronization of projected illumination with rolling shutter of image sensor
US13/765,706 Active 2032-05-14 US9736459B2 (en) 2010-02-02 2013-02-13 Generation of patterned radiation
US15/673,433 Active US10063835B2 (en) 2010-02-02 2017-08-10 Generation of patterned radiation
US16/045,779 Active US10609357B2 (en) 2010-02-02 2018-07-26 Synchronization of projected illumination with activation of sensing elements

Family Applications After (3)

Application Number Title Priority Date Filing Date
US13/765,706 Active 2032-05-14 US9736459B2 (en) 2010-02-02 2013-02-13 Generation of patterned radiation
US15/673,433 Active US10063835B2 (en) 2010-02-02 2017-08-10 Generation of patterned radiation
US16/045,779 Active US10609357B2 (en) 2010-02-02 2018-07-26 Synchronization of projected illumination with activation of sensing elements

Country Status (4)

Country Link
US (4) US20110187878A1 (en)
EP (1) EP2363686A1 (en)
JP (1) JP2011160420A (en)
CN (3) CN104360571B (en)

Cited By (110)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120140096A1 (en) * 2010-12-01 2012-06-07 Sony Ericsson Mobile Communications Ab Timing Solution for Projector Camera Devices and Systems
US8237835B1 (en) * 2011-05-19 2012-08-07 Aeon Imaging, LLC Confocal imaging device using spatially modulated illumination with electronic rolling shutter detection
US20130021474A1 (en) * 2011-07-20 2013-01-24 Raytheon Company Rolling-shutter imaging system with synchronized scanning illumination and methods for higher-resolution imaging
US8374397B2 (en) 2005-10-11 2013-02-12 Primesense Ltd Depth-varying light fields for three dimensional sensing
WO2013094347A1 (en) * 2011-12-19 2013-06-27 Canon Kabushiki Kaisha Information processing apparatus, control method therefor, and computer-readable storage medium
US20130335546A1 (en) * 2012-06-18 2013-12-19 Randall T. Crane Selective imaging
US8615108B1 (en) 2013-01-30 2013-12-24 Imimtek, Inc. Systems and methods for initializing motion tracking of human hands
US8655021B2 (en) 2012-06-25 2014-02-18 Imimtek, Inc. Systems and methods for tracking human hands by performing parts based template matching using images from multiple viewpoints
US20140111619A1 (en) * 2012-10-19 2014-04-24 Electronics And Telecommunications Research Institute Device and method for acquiring image
US8749796B2 (en) 2011-08-09 2014-06-10 Primesense Ltd. Projectors of structured light
WO2014106843A2 (en) * 2013-01-01 2014-07-10 Inuitive Ltd. Method and system for light patterning and imaging
US20140240469A1 (en) * 2013-02-28 2014-08-28 Motorola Mobility Llc Electronic Device with Multiview Image Capture and Depth Sensing
US8830312B2 (en) 2012-06-25 2014-09-09 Aquifi, Inc. Systems and methods for tracking human hands using parts based template matching within bounded regions
US8908277B2 (en) 2011-08-09 2014-12-09 Apple Inc Lens array projector
US20150002734A1 (en) * 2013-07-01 2015-01-01 Motorola Mobility Llc Electronic Device with Modulated Light Flash Operation for Rolling Shutter Image Sensor
US8982182B2 (en) 2010-03-01 2015-03-17 Apple Inc. Non-uniform spatial resource allocation for depth mapping
US9030528B2 (en) 2011-04-04 2015-05-12 Apple Inc. Multi-zone imaging sensor and lens array
US9036158B2 (en) 2010-08-11 2015-05-19 Apple Inc. Pattern projector
US9066084B2 (en) 2005-10-11 2015-06-23 Apple Inc. Method and system for object reconstruction
US9066087B2 (en) 2010-11-19 2015-06-23 Apple Inc. Depth mapping using time-coded illumination
US9092665B2 (en) 2013-01-30 2015-07-28 Aquifi, Inc Systems and methods for initializing motion tracking of human hands
US9098931B2 (en) 2010-08-11 2015-08-04 Apple Inc. Scanning projectors and image capture modules for 3D mapping
US9131136B2 (en) 2010-12-06 2015-09-08 Apple Inc. Lens arrays for pattern projection and imaging
US20150264335A1 (en) * 2014-03-13 2015-09-17 Samsung Electronics Co., Ltd. Image pickup apparatus and method for generating image having depth information
US9157790B2 (en) 2012-02-15 2015-10-13 Apple Inc. Integrated optoelectronic modules with transmitter, receiver and beam-combining optics for aligning a beam axis with a collection axis
US20150323311A1 (en) * 2012-06-13 2015-11-12 Koninklijke Philips N.V. Determining a propagation velocity for a surface wave
US9201237B2 (en) 2012-03-22 2015-12-01 Apple Inc. Diffraction-based sensing of mirror position
WO2015127328A3 (en) * 2014-02-21 2015-12-23 The Lightco Inc. Illumination methods and apparatus
US9239467B2 (en) 2008-01-21 2016-01-19 Apple Inc. Optical pattern projection
US9270876B2 (en) 2013-01-05 2016-02-23 The Lightco Inc. Methods and apparatus for using multiple optical chains in parallel with multiple different exposure times
US9298266B2 (en) 2013-04-02 2016-03-29 Aquifi, Inc. Systems and methods for implementing three-dimensional (3D) gesture based graphical user interfaces (GUI) that incorporate gesture reactive interface objects
US9310891B2 (en) 2012-09-04 2016-04-12 Aquifi, Inc. Method and system enabling natural user interface gestures with user wearable glasses
US9325906B2 (en) 2013-10-18 2016-04-26 The Lightco Inc. Methods and apparatus relating to a thin camera device
US9330324B2 (en) 2005-10-11 2016-05-03 Apple Inc. Error compensation in three-dimensional mapping
US9374514B2 (en) 2013-10-18 2016-06-21 The Lightco Inc. Methods and apparatus relating to a camera including multiple optical chains
US9398287B2 (en) 2013-02-28 2016-07-19 Google Technology Holdings LLC Context-based depth sensor control
US9398229B2 (en) 2012-06-18 2016-07-19 Microsoft Technology Licensing, Llc Selective illumination of a region within a field of view
US9407837B2 (en) 2013-02-28 2016-08-02 Google Inc. Depth sensor using modulated light projector and image sensor with color and IR sensing
US9423588B2 (en) 2013-10-18 2016-08-23 The Lightco Inc. Methods and apparatus for supporting zoom operations
US9426365B2 (en) 2013-11-01 2016-08-23 The Lightco Inc. Image stabilization related methods and apparatus
US9462170B2 (en) 2014-02-21 2016-10-04 The Lightco Inc. Lighting methods and apparatus
US9467627B2 (en) 2013-10-26 2016-10-11 The Lightco Inc. Methods and apparatus for use with multiple optical chains
US9504920B2 (en) 2011-04-25 2016-11-29 Aquifi, Inc. Method and system to create three-dimensional mapping in a two-dimensional game
US9507417B2 (en) 2014-01-07 2016-11-29 Aquifi, Inc. Systems and methods for implementing head tracking based graphical user interfaces (GUI) that incorporate gesture reactive interface objects
US9525863B2 (en) 2015-04-29 2016-12-20 Apple Inc. Time-of-flight depth mapping with flexible scan pattern
US9528906B1 (en) 2013-12-19 2016-12-27 Apple Inc. Monitoring DOE performance using total internal reflection
US9544503B2 (en) 2014-12-30 2017-01-10 Light Labs Inc. Exposure control methods and apparatus
US9554031B2 (en) 2013-12-31 2017-01-24 Light Labs Inc. Camera focusing related methods and apparatus
EP3047641A4 (en) * 2013-09-16 2017-03-08 Intel Corporation Camera and light source synchronization for object tracking
US9600078B2 (en) 2012-02-03 2017-03-21 Aquifi, Inc. Method and system enabling natural user interface gestures with an electronic system
US9619105B1 (en) 2014-01-30 2017-04-11 Aquifi, Inc. Systems and methods for gesture based interaction with viewpoint dependent user interfaces
US9736459B2 (en) 2010-02-02 2017-08-15 Apple Inc. Generation of patterned radiation
US9736365B2 (en) 2013-10-26 2017-08-15 Light Labs Inc. Zoom related methods and apparatus
US9749549B2 (en) 2015-10-06 2017-08-29 Light Labs Inc. Methods and apparatus for facilitating selective blurring of one or more image portions
US20170288366A1 (en) * 2016-04-05 2017-10-05 Aquifi, Inc. Thin laser package for optical applications
US9798388B1 (en) 2013-07-31 2017-10-24 Aquifi, Inc. Vibrotactile system to augment 3D input systems
US9825425B2 (en) 2013-06-19 2017-11-21 Apple Inc. Integrated structured-light projector comprising light-emitting elements on a substrate
US9824427B2 (en) 2015-04-15 2017-11-21 Light Labs Inc. Methods and apparatus for generating a sharp image
US9857868B2 (en) 2011-03-19 2018-01-02 The Board Of Trustees Of The Leland Stanford Junior University Method and system for ergonomic touch-free interface
US9857584B2 (en) 2015-04-17 2018-01-02 Light Labs Inc. Camera device methods, apparatus and components
TWI616711B (en) * 2015-12-11 2018-03-01 Ningbo Sunny Opotech Co Ltd Lighting module and its application
US9912864B2 (en) 2014-10-17 2018-03-06 Light Labs Inc. Methods and apparatus for using a camera device to support multiple modes of operation
US9930233B2 (en) 2015-04-22 2018-03-27 Light Labs Inc. Filter mounting methods and apparatus and related camera apparatus
US9948832B2 (en) 2016-06-22 2018-04-17 Light Labs Inc. Methods and apparatus for synchronized image capture in a device including optical chains with different orientations
US9967535B2 (en) 2015-04-17 2018-05-08 Light Labs Inc. Methods and apparatus for reducing noise in images
US9979878B2 (en) 2014-02-21 2018-05-22 Light Labs Inc. Intuitive camera user interface methods and apparatus
US9998638B2 (en) 2014-12-17 2018-06-12 Light Labs Inc. Methods and apparatus for implementing and using camera devices
US10003738B2 (en) 2015-12-18 2018-06-19 Light Labs Inc. Methods and apparatus for detecting and/or indicating a blocked sensor or camera module
WO2018111860A1 (en) * 2016-12-12 2018-06-21 Princeton Identity, Inc. Systems and methods of biometric analysis
US10012831B2 (en) 2015-08-03 2018-07-03 Apple Inc. Optical monitoring of scan parameters
US10073004B2 (en) 2016-09-19 2018-09-11 Apple Inc. DOE defect monitoring utilizing total internal reflection
US10075651B2 (en) 2015-04-17 2018-09-11 Light Labs Inc. Methods and apparatus for capturing images using multiple camera modules in an efficient manner
US10091447B2 (en) 2015-04-17 2018-10-02 Light Labs Inc. Methods and apparatus for synchronizing readout of multiple image sensors
US10110794B2 (en) 2014-07-09 2018-10-23 Light Labs Inc. Camera device including multiple optical chains and related methods
US10129483B2 (en) 2015-06-23 2018-11-13 Light Labs Inc. Methods and apparatus for implementing zoom using one or more moveable camera modules
US10153614B1 (en) 2017-08-31 2018-12-11 Apple Inc. Creating arbitrary patterns on a 2-D uniform grid VCSEL array
US10158845B2 (en) 2017-01-18 2018-12-18 Facebook Technologies, Llc Tileable structured light projection for wide field-of-view depth sensing
US10191356B2 (en) 2014-07-04 2019-01-29 Light Labs Inc. Methods and apparatus relating to detection and/or indicating a dirty lens condition
US10225445B2 (en) 2015-12-18 2019-03-05 Light Labs Inc. Methods and apparatus for providing a camera lens or viewing point indicator
US10306218B2 (en) 2016-03-22 2019-05-28 Light Labs Inc. Camera calibration apparatus and methods
US10365480B2 (en) 2015-08-27 2019-07-30 Light Labs Inc. Methods and apparatus for implementing and/or using camera devices with one or more light redirection devices
US10366296B2 (en) 2016-03-31 2019-07-30 Princeton Identity, Inc. Biometric enrollment systems and methods
US10373008B2 (en) 2016-03-31 2019-08-06 Princeton Identity, Inc. Systems and methods of biometric analysis with adaptive trigger
US20190271828A1 (en) * 2015-09-11 2019-09-05 Ams Sensors Singapore Pte. Ltd. Imaging devices having autofocusd control
US10425814B2 (en) 2014-09-24 2019-09-24 Princeton Identity, Inc. Control of wireless communication device capability in a mobile device with a biometric key
US10447424B2 (en) 2018-01-18 2019-10-15 Apple Inc. Spatial multiplexing scheme
US10452936B2 (en) 2016-01-12 2019-10-22 Princeton Identity Systems and methods of biometric analysis with a spectral discriminator
US10484584B2 (en) 2014-12-03 2019-11-19 Princeton Identity, Inc. System and method for mobile device biometric add-on
US10491806B2 (en) 2015-08-03 2019-11-26 Light Labs Inc. Camera device control related methods and apparatus
US10607096B2 (en) 2017-04-04 2020-03-31 Princeton Identity, Inc. Z-dimension user feedback biometric system
EP3693785A4 (en) * 2017-10-25 2020-12-16 Huawei Technologies Co., Ltd. Projector, camera module and terminal device
US10877285B2 (en) 2018-03-28 2020-12-29 Apple Inc. Wavelength-based spatial multiplexing scheme
US10902104B2 (en) 2017-07-26 2021-01-26 Princeton Identity, Inc. Biometric security systems and methods
US11121288B2 (en) 2019-01-16 2021-09-14 Lextar Electronics Corporation Package structure of light-emitting element
US11137618B2 (en) * 2018-02-23 2021-10-05 Jabil Optics Germany GmbH Projector module
US20210409582A1 (en) * 2015-02-02 2021-12-30 Apple Inc. Focusing Lighting Module
US11237370B2 (en) 2018-04-30 2022-02-01 University Of Central Florida Research Foundation, Inc. Multiple inclined beam line-scanning imaging apparatus, methods, and applications
US11245891B2 (en) * 2015-01-21 2022-02-08 Nevermind Capital Llc Methods and apparatus for environmental measurements and/or stereoscopic image capture
US11303817B2 (en) * 2018-12-27 2022-04-12 Koito Manufaciuring Co., Ltd. Active sensor, object identification system, vehicle and vehicle lamp
US11322910B2 (en) 2019-02-21 2022-05-03 Apple Inc. Indium-phosphide VCSEL with dielectric DBR
US11333895B1 (en) 2019-01-11 2022-05-17 Facebook Technologies, Llc Systems and methods for structured light projector operational safety
US20220174187A1 (en) * 2020-12-01 2022-06-02 Samsung Electronics Co., Ltd. Vision sensor, image processing device including the same, and operating method of the vision sensor
CN114636546A (en) * 2022-03-10 2022-06-17 杭州海康威视数字技术股份有限公司 System for imaging synchronicity detection
US11374381B1 (en) 2019-06-10 2022-06-28 Apple Inc. Integrated laser module
US11418010B2 (en) 2019-04-01 2022-08-16 Apple Inc. VCSEL array with tight pitch and high efficiency
US11422292B1 (en) 2018-06-10 2022-08-23 Apple Inc. Super-blazed diffractive optical elements with sub-wavelength structures
US11493606B1 (en) 2018-09-12 2022-11-08 Apple Inc. Multi-beam scanning system
US11506762B1 (en) 2019-09-24 2022-11-22 Apple Inc. Optical module comprising an optical waveguide with reference light path
US11681019B2 (en) 2019-09-18 2023-06-20 Apple Inc. Optical module with stray light baffle
US11754767B1 (en) 2020-03-05 2023-09-12 Apple Inc. Display with overlaid waveguide

Families Citing this family (88)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8908995B2 (en) 2009-01-12 2014-12-09 Intermec Ip Corp. Semi-automatic dimensioning with imager on a portable device
US9329080B2 (en) 2012-02-15 2016-05-03 Aplle Inc. Modular optics for scanning engine having beam combining optics with a prism intercepted by both beam axis and collection axis
AU2015203089B2 (en) * 2012-02-15 2016-06-30 Apple Inc. Scanning depth engine
CN203385981U (en) * 2012-03-15 2014-01-08 普莱姆森斯有限公司 Projector of structured light
JP5913726B2 (en) 2012-03-22 2016-04-27 アップル インコーポレイテッド Gimbal scanning mirror array
US9715107B2 (en) 2012-03-22 2017-07-25 Apple Inc. Coupling schemes for gimbaled scanning mirror arrays
CN104135908B (en) * 2012-03-28 2016-07-06 富士胶片株式会社 Camera head and possess its endoscope apparatus
US9779546B2 (en) 2012-05-04 2017-10-03 Intermec Ip Corp. Volume dimensioning systems and methods
US10007858B2 (en) 2012-05-15 2018-06-26 Honeywell International Inc. Terminals and methods for dimensioning objects
US9188433B2 (en) * 2012-05-24 2015-11-17 Qualcomm Incorporated Code in affine-invariant spatial mask
US10321127B2 (en) 2012-08-20 2019-06-11 Intermec Ip Corp. Volume dimensioning system calibration systems and methods
US9841311B2 (en) 2012-10-16 2017-12-12 Hand Held Products, Inc. Dimensioning system
JP2014085215A (en) * 2012-10-23 2014-05-12 Ayako Sato System for unified analysis and measurement of various forms of carbon and nitrogen which employs calibration curve based on organic compounds
EP2946336B1 (en) 2013-01-15 2023-06-21 Mobileye Vision Technologies Ltd. Stereo assist with rolling shutters
CN104010150B (en) * 2013-02-27 2018-06-01 联想(北京)有限公司 The method and electronic equipment of a kind of Image Acquisition
US10228452B2 (en) 2013-06-07 2019-03-12 Hand Held Products, Inc. Method of error correction for 3D imaging device
SG10201807071TA (en) * 2014-03-13 2018-09-27 Nat Univ Singapore An optical interference device
US9558436B2 (en) * 2014-06-20 2017-01-31 Qualcomm Incorporated Coded light pattern having hermitian symmetry
US9823059B2 (en) 2014-08-06 2017-11-21 Hand Held Products, Inc. Dimensioning system with guided alignment
MX363283B (en) * 2014-08-10 2019-03-19 Facebook Inc Structured light for 3d sensing.
US10775165B2 (en) 2014-10-10 2020-09-15 Hand Held Products, Inc. Methods for improving the accuracy of dimensioning-system measurements
US9779276B2 (en) 2014-10-10 2017-10-03 Hand Held Products, Inc. Depth sensor based auto-focus system for an indicia scanner
US10810715B2 (en) 2014-10-10 2020-10-20 Hand Held Products, Inc System and method for picking validation
US9897434B2 (en) 2014-10-21 2018-02-20 Hand Held Products, Inc. Handheld dimensioning system with measurement-conformance feedback
US9752864B2 (en) 2014-10-21 2017-09-05 Hand Held Products, Inc. Handheld dimensioning system with feedback
US9784838B1 (en) 2014-11-26 2017-10-10 Apple Inc. Compact scanner with gimbaled optics
US9835853B1 (en) 2014-11-26 2017-12-05 Apple Inc. MEMS scanner with mirrors of different sizes
JP6516453B2 (en) * 2014-11-26 2019-05-22 株式会社ミツトヨ Image measuring device and measuring device
US9798135B2 (en) 2015-02-16 2017-10-24 Apple Inc. Hybrid MEMS scanning module
US10326252B2 (en) * 2015-05-06 2019-06-18 Microsoft Technology Licensing, Llc Beam projection for fast axis expansion
US9786101B2 (en) 2015-05-19 2017-10-10 Hand Held Products, Inc. Evaluating image values
US20160377414A1 (en) * 2015-06-23 2016-12-29 Hand Held Products, Inc. Optical pattern projector
US9835486B2 (en) 2015-07-07 2017-12-05 Hand Held Products, Inc. Mobile dimensioner apparatus for use in commerce
US20170017301A1 (en) 2015-07-16 2017-01-19 Hand Held Products, Inc. Adjusting dimensioning results using augmented reality
US10380728B2 (en) * 2015-08-31 2019-08-13 Kla-Tencor Corporation Model-based metrology using images
US9703096B2 (en) 2015-09-30 2017-07-11 Apple Inc. Asymmetric MEMS mirror assembly
US9897801B2 (en) 2015-09-30 2018-02-20 Apple Inc. Multi-hinge mirror assembly
US10249030B2 (en) 2015-10-30 2019-04-02 Hand Held Products, Inc. Image transformation for indicia reading
KR20230004905A (en) 2015-11-10 2023-01-06 루미리즈 홀딩 비.브이. Adaptive light source
US10225544B2 (en) 2015-11-19 2019-03-05 Hand Held Products, Inc. High resolution dot pattern
US10101154B2 (en) * 2015-12-21 2018-10-16 Intel Corporation System and method for enhanced signal to noise ratio performance of a depth camera system
US10025314B2 (en) 2016-01-27 2018-07-17 Hand Held Products, Inc. Vehicle positioning and object avoidance
US10375330B2 (en) * 2016-05-27 2019-08-06 Verily Life Sciences Llc Systems and methods for surface topography acquisition using laser speckle
US10339352B2 (en) 2016-06-03 2019-07-02 Hand Held Products, Inc. Wearable metrological apparatus
US10163216B2 (en) 2016-06-15 2018-12-25 Hand Held Products, Inc. Automatic mode switching in a volume dimensioner
US10488652B2 (en) 2016-09-21 2019-11-26 Apple Inc. Prism-based scanner
AU2017336066B2 (en) * 2016-09-30 2022-04-14 Magic Leap, Inc. Projector with spatial light modulation
US10909708B2 (en) 2016-12-09 2021-02-02 Hand Held Products, Inc. Calibrating a dimensioner using ratios of measurable parameters of optic ally-perceptible geometric elements
US10120111B2 (en) * 2016-12-14 2018-11-06 Google Llc Thin ceramic imaging screen for camera systems
US10825952B2 (en) * 2017-01-16 2020-11-03 Apple Inc. Combining light-emitting elements of differing divergence on the same substrate
US11047672B2 (en) 2017-03-28 2021-06-29 Hand Held Products, Inc. System for optically dimensioning
US10511828B2 (en) * 2017-03-29 2019-12-17 Intel Corporation Camera platforms with rolling light projection
CN106990660A (en) * 2017-05-09 2017-07-28 深圳奥比中光科技有限公司 Structured light projection module
CN107390461A (en) * 2017-06-20 2017-11-24 深圳奥比中光科技有限公司 3D is imaged submodule group and its electronic equipment
US10830879B2 (en) 2017-06-29 2020-11-10 Apple Inc. Time-of-flight depth mapping with parallax compensation
US10733748B2 (en) 2017-07-24 2020-08-04 Hand Held Products, Inc. Dual-pattern optical 3D dimensioning
WO2019054933A1 (en) * 2017-09-18 2019-03-21 Profoto Aktiebolag A flash housing for photographic purposes and a method for simulating a flash light
US10955552B2 (en) 2017-09-27 2021-03-23 Apple Inc. Waveform design for a LiDAR system with closely-spaced pulses
US11269193B2 (en) * 2017-11-27 2022-03-08 Liqxtal Technology Inc. Optical sensing device and structured light projector
CN108040212B (en) * 2017-11-29 2019-12-31 长光卫星技术有限公司 High-precision target tracking method based on staggered rolling shutter
US11852727B2 (en) 2017-12-18 2023-12-26 Apple Inc. Time-of-flight sensing using an addressable array of emitters
CN110072065B (en) * 2018-01-23 2021-04-27 舜宇光学(浙江)研究院有限公司 Projector working time control method suitable for roller shutter exposure depth camera and application thereof
CN108181777A (en) * 2018-02-27 2018-06-19 广东欧珀移动通信有限公司 Laser projection module, depth camera and electronic device
CN108344376A (en) 2018-03-12 2018-07-31 广东欧珀移动通信有限公司 Laser projection module, depth camera and electronic device
US10584962B2 (en) 2018-05-01 2020-03-10 Hand Held Products, Inc System and method for validating physical-item security
CN110471192B (en) * 2018-05-11 2021-09-21 宁波舜宇光电信息有限公司 Projection device, diffractive optical element, method for manufacturing the same, and electronic apparatus with projection device
WO2019236563A1 (en) * 2018-06-06 2019-12-12 Magik Eye Inc. Distance measurement using high density projection patterns
WO2019243046A1 (en) 2018-06-18 2019-12-26 Lumileds Holding B.V. Lighting device comprising led and grating
EP3832741A4 (en) * 2018-08-03 2022-03-16 SZ DJI Technology Co., Ltd. Laser diode package module, distance detection apparatus, and electronic device
CN109143755A (en) * 2018-09-08 2019-01-04 深圳阜时科技有限公司 A kind of optical module, optical projection mould group, sensing device and equipment
US11223816B2 (en) 2018-09-26 2022-01-11 Himax Technologies Limited Multi-image projector and electronic device having multi-image projector
EP3637044B1 (en) * 2018-10-09 2022-11-16 Himax Technologies Limited Multi-image projector and electronic device having the multi-image projector
WO2020150131A1 (en) 2019-01-20 2020-07-23 Magik Eye Inc. Three-dimensional sensor including bandpass filter having multiple passbands
EP3887852A1 (en) 2019-02-11 2021-10-06 Apple Inc. Depth sensing using a sparse array of pulsed beams
US11885613B2 (en) 2019-03-15 2024-01-30 Shanghai Percipio Technology Limited Depth data measuring head, measurement device and measuring method
WO2020197813A1 (en) 2019-03-25 2020-10-01 Magik Eye Inc. Distance measurement using high density projection patterns
US11500094B2 (en) 2019-06-10 2022-11-15 Apple Inc. Selection of pulse repetition intervals for sensing time of flight
US11404844B2 (en) * 2019-06-13 2022-08-02 Omnivision Technologies, Inc. Semiconductor structure and method of manufacturing the same
US11555900B1 (en) 2019-07-17 2023-01-17 Apple Inc. LiDAR system with enhanced area coverage
WO2021034371A1 (en) 2019-08-18 2021-02-25 Apple Inc. Force-balanced micromirror with electromagnetic actuation
US11639846B2 (en) 2019-09-27 2023-05-02 Honeywell International Inc. Dual-pattern optical 3D dimensioning
US11733359B2 (en) 2019-12-03 2023-08-22 Apple Inc. Configurable array of single-photon detectors
CN111556225B (en) * 2020-05-20 2022-11-22 杭州海康威视数字技术股份有限公司 Image acquisition device and image acquisition control method
CN112859359B (en) * 2021-02-05 2022-02-08 中国工程物理研究院激光聚变研究中心 Focal spot control method
US11681028B2 (en) 2021-07-18 2023-06-20 Apple Inc. Close-range measurement of time of flight using parallax shift
WO2023065222A1 (en) * 2021-10-21 2023-04-27 深圳先进技术研究院 Microscope synchronization control system and method
DE102021129698A1 (en) * 2021-11-15 2023-05-17 OSRAM Opto Semiconductors Gesellschaft mit beschränkter Haftung OPTOELECTRONIC DEVICE AND METHOD FOR MANUFACTURING OPTOELECTRONIC DEVICE
CN115864976B (en) * 2022-12-22 2023-08-18 广东工业大学 Control circuit of step-adjustable photonic transformer

Citations (79)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4542376A (en) * 1983-11-03 1985-09-17 Burroughs Corporation System for electronically displaying portions of several different images on a CRT screen through respective prioritized viewports
US4802759A (en) * 1986-08-11 1989-02-07 Goro Matsumoto Three-dimensional shape measuring apparatus
US4843568A (en) * 1986-04-11 1989-06-27 Krueger Myron W Real time perception of and response to the actions of an unencumbered participant/user
US4850673A (en) * 1987-11-23 1989-07-25 U. S. Philips Corporation Optical scanning apparatus which detects scanning spot focus error
US5075562A (en) * 1990-09-20 1991-12-24 Eastman Kodak Company Method and apparatus for absolute Moire distance measurements using a grating printed on or attached to a surface
US5477383A (en) * 1993-02-05 1995-12-19 Apa Optics, Inc. Optical array method and apparatus
US5483261A (en) * 1992-02-14 1996-01-09 Itu Research, Inc. Graphical input controller and method with rear screen image detection
US5606181A (en) * 1994-03-29 1997-02-25 Ricoh Company, Ltd. Edge emitting type light emitting diode array heads
US5630043A (en) * 1995-05-11 1997-05-13 Cirrus Logic, Inc. Animated texture map apparatus and method for 3-D image displays
US5636025A (en) * 1992-04-23 1997-06-03 Medar, Inc. System for optically measuring the surface contour of a part using more fringe techniques
US5648951A (en) * 1990-04-12 1997-07-15 Matsushita Electic Industrial Co. Ltd. Movable optical head integrally incorporated with objective lens and hologram element
US5691989A (en) * 1991-07-26 1997-11-25 Accuwave Corporation Wavelength stabilized laser sources using feedback from volume holograms
US5742262A (en) * 1993-06-23 1998-04-21 Olympus Optical Co., Ltd. Image display apparatus
US5781332A (en) * 1995-09-04 1998-07-14 Nec Corporation Variable wavelength optical filter
US5838428A (en) * 1997-02-28 1998-11-17 United States Of America As Represented By The Secretary Of The Navy System and method for high resolution range imaging with split light source and pattern mask
US5856871A (en) * 1993-08-18 1999-01-05 Applied Spectral Imaging Ltd. Film thickness mapping using interferometric spectral imaging
US6002520A (en) * 1997-04-25 1999-12-14 Hewlett-Packard Company Illumination system for creating a desired irradiance profile using diffractive optical elements
US6101269A (en) * 1997-12-19 2000-08-08 Lifef/X Networks, Inc. Apparatus and method for rapid 3D image parametrization
US6100517A (en) * 1995-06-22 2000-08-08 3Dv Systems Ltd. Three dimensional camera
US6167151A (en) * 1996-12-15 2000-12-26 Cognitens, Ltd. Apparatus and method for 3-dimensional surface geometry reconstruction
US6259561B1 (en) * 1999-03-26 2001-07-10 The University Of Rochester Optical system for diffusing light
US6262740B1 (en) * 1997-08-01 2001-07-17 Terarecon, Inc. Method for rendering sections of a volume data set
US20020075456A1 (en) * 2000-12-20 2002-06-20 Olympus Optical Co., Ltd. 3D image acquisition apparatus and 3D image acquisition method
US20030057972A1 (en) * 1999-07-26 2003-03-27 Paul Pfaff Voltage testing and measurement
US6560019B2 (en) * 1998-02-05 2003-05-06 Canon Kabushiki Kaisha Diffractive optical element and optical system having the same
US6583873B1 (en) * 2000-09-25 2003-06-24 The Carnegie Institution Of Washington Optical devices having a wavelength-tunable dispersion assembly that has a volume dispersive diffraction grating
US20030156756A1 (en) * 2002-02-15 2003-08-21 Gokturk Salih Burak Gesture recognition system using depth perceptive sensors
US6611000B2 (en) * 2001-03-14 2003-08-26 Matsushita Electric Industrial Co., Ltd. Lighting device
US20040012958A1 (en) * 2001-04-23 2004-01-22 Takuma Hashimoto Light emitting device comprising led chip
US6686921B1 (en) * 2000-08-01 2004-02-03 International Business Machines Corporation Method and apparatus for acquiring a set of consistent image maps to represent the color of the surface of an object
US6741251B2 (en) * 2001-08-16 2004-05-25 Hewlett-Packard Development Company, L.P. Method and apparatus for varying focus in a scene
US6751344B1 (en) * 1999-05-28 2004-06-15 Champion Orthotic Investments, Inc. Enhanced projector system for machine vision
US6754370B1 (en) * 2000-08-14 2004-06-22 The Board Of Trustees Of The Leland Stanford Junior University Real-time structured light range scanning of moving scenes
US20040130730A1 (en) * 2002-11-21 2004-07-08 Michel Cantin Fast 3D height measurement method and system
US20040174770A1 (en) * 2002-11-27 2004-09-09 Rees Frank L. Gauss-Rees parametric ultrawideband system
US6813440B1 (en) * 2000-10-10 2004-11-02 The Hong Kong Polytechnic University Body scanner
US20040218262A1 (en) * 2003-02-21 2004-11-04 Chuang Yung-Ho Inspection system using small catadioptric objective
US20040228519A1 (en) * 2003-03-10 2004-11-18 Cranial Technologies, Inc. Automatic selection of cranial remodeling device trim lines
US6825985B2 (en) * 2001-07-13 2004-11-30 Mems Optical, Inc. Autostereoscopic display with rotated microlens and method of displaying multidimensional images, especially color images
US6927852B2 (en) * 2000-05-05 2005-08-09 Applera Corporation Optical system and method for optically analyzing light from a sample
US20050178950A1 (en) * 2004-02-18 2005-08-18 Fujinon Corporation Electronic imaging system
US6940583B2 (en) * 2003-07-28 2005-09-06 International Business Machines Corporation Method and apparatus for amplitude filtering in the frequency plane of a lithographic projection system
US20050200925A1 (en) * 1999-12-10 2005-09-15 Xyz Imaging, Inc. Holographic printer
US20050271279A1 (en) * 2004-05-14 2005-12-08 Honda Motor Co., Ltd. Sign based human-machine interaction
US20060001055A1 (en) * 2004-02-23 2006-01-05 Kazuhiko Ueno Led and fabrication method of same
US20060072851A1 (en) * 2002-06-15 2006-04-06 Microsoft Corporation Deghosting mosaics using multiperspective plane sweep
US7076024B2 (en) * 2004-12-01 2006-07-11 Jordan Valley Applied Radiation, Ltd. X-ray apparatus with dual monochromators
US20060156756A1 (en) * 2005-01-20 2006-07-20 Becke Paul E Phase change and insulating properties container and method of use
US7112774B2 (en) * 2003-10-09 2006-09-26 Avago Technologies Sensor Ip (Singapore) Pte. Ltd CMOS stereo imaging system and method
US7120228B2 (en) * 2004-09-21 2006-10-10 Jordan Valley Applied Radiation Ltd. Combined X-ray reflectometer and diffractometer
US20060252167A1 (en) * 2005-05-04 2006-11-09 Wang Tak K Silicon optical package with 45 degree turning mirror
US20060252169A1 (en) * 2004-10-07 2006-11-09 Takeshi Ashida Transparent member, optical device using transparent member and method of manufacturing optical device
US20060269896A1 (en) * 2005-05-27 2006-11-30 Yongqian Liu High speed 3D scanner and uses thereof
US20070019909A1 (en) * 2005-07-22 2007-01-25 Seiko Epson Corporation Illumination device, image display device, and projector
US20070060336A1 (en) * 2003-09-15 2007-03-15 Sony Computer Entertainment Inc. Methods and systems for enabling depth and direction detection when interfacing with a computer program
US20070057946A1 (en) * 2003-07-24 2007-03-15 Dan Albeck Method and system for the three-dimensional surface reconstruction of an object
US7227618B1 (en) * 2004-03-24 2007-06-05 Baokang Bi Pattern generating systems
US7304735B2 (en) * 2004-04-02 2007-12-04 Kla-Tencor Technologies Broadband wavelength selective filter
US20080018595A1 (en) * 2000-07-24 2008-01-24 Gesturetek, Inc. Video-based image control system
US7335898B2 (en) * 2004-07-23 2008-02-26 Ge Healthcare Niagara Inc. Method and apparatus for fluorescent confocal microscopy
US7369685B2 (en) * 2002-04-05 2008-05-06 Identix Corporation Vision-based operating method and system
US20080106746A1 (en) * 2005-10-11 2008-05-08 Alexander Shpunt Depth-varying light fields for three dimensional sensing
US20080198355A1 (en) * 2006-12-27 2008-08-21 Cambridge Research & Instrumentation, Inc Surface measurement of in-vivo subjects using spot projector
US20080212835A1 (en) * 2007-03-01 2008-09-04 Amon Tavor Object Tracking by 3-Dimensional Modeling
US20080240502A1 (en) * 2007-04-02 2008-10-02 Barak Freedman Depth mapping using projected patterns
US7433024B2 (en) * 2006-02-27 2008-10-07 Prime Sense Ltd. Range mapping using speckle decorrelation
US20080278572A1 (en) * 2007-04-23 2008-11-13 Morteza Gharib Aperture system with spatially-biased aperture shapes and positions (SBPSP) for static and dynamic 3-D defocusing-based imaging
US20090096783A1 (en) * 2005-10-11 2009-04-16 Alexander Shpunt Three-dimensional sensing using speckle patterns
US20090183125A1 (en) * 2008-01-14 2009-07-16 Prime Sense Ltd. Three-dimensional user interface
US20090183152A1 (en) * 2008-01-16 2009-07-16 Dell Products, Lp Method to Dynamically Provision Additional Computer Resources to Handle Peak Database Workloads
US20090185274A1 (en) * 2008-01-21 2009-07-23 Prime Sense Ltd. Optical designs for zero order reduction
US20100007717A1 (en) * 2008-07-09 2010-01-14 Prime Sense Ltd Integrated processor for 3d mapping
US20100013860A1 (en) * 2006-03-08 2010-01-21 Electronic Scripting Products, Inc. Computer interface employing a manipulated object with absolute pose detection component and a display
US20100020078A1 (en) * 2007-01-21 2010-01-28 Prime Sense Ltd Depth mapping using multi-beam illumination
US7700904B2 (en) * 2007-07-18 2010-04-20 Funai Electric Co., Ltd. Compound-eye imaging device
US20100118123A1 (en) * 2007-04-02 2010-05-13 Prime Sense Ltd Depth mapping using projected patterns
US20100142014A1 (en) * 2006-11-17 2010-06-10 Celloptic, Inc. System, apparatus and method for extracting three-dimensional information of an object from received electromagnetic radiation
US20100284082A1 (en) * 2008-01-21 2010-11-11 Primesense Ltd. Optical pattern projection
US20110091258A1 (en) * 2009-10-20 2011-04-21 Canon Kabushiki Kaisha Image forming apparatus

Family Cites Families (72)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB1353470A (en) 1970-10-19 1974-05-15 Post D Position measuring apparatus utilizing moire fringe multiplication
JPH0710017B2 (en) * 1985-07-08 1995-02-01 松下電器産業株式会社 Method of manufacturing semiconductor laser device
US5325386A (en) * 1992-04-21 1994-06-28 Bandgap Technology Corporation Vertical-cavity surface emitting laser assay display system
US5406543A (en) 1993-04-07 1995-04-11 Olympus Optical Co., Ltd. Optical head with semiconductor laser
JP3042278B2 (en) 1993-09-17 2000-05-15 三菱電機株式会社 Distance measuring device
DE4433763A1 (en) 1994-09-22 1996-05-23 Hell Ag Linotype Beam splitting device
US5646928A (en) 1995-06-22 1997-07-08 The Regents Of The University Of California Free-space integrated micro-pickup head for optical data storage and a micro-optical bench
US6075799A (en) 1996-08-28 2000-06-13 Canon Kabushiki Kaisha Polarization selective semiconductor laser, optical transmitter using the same, optical communication system using the same and fabrication method of the same
JPH10123512A (en) * 1996-08-28 1998-05-15 Sanyo Electric Co Ltd Light source for liquid crystal display device and color liquid crystal display device using it
JP3433647B2 (en) * 1997-04-23 2003-08-04 富士ゼロックス株式会社 Projector device
US6031611A (en) 1997-06-03 2000-02-29 California Institute Of Technology Coherent gradient sensing method and system for measuring surface curvature
US6636538B1 (en) 1999-03-29 2003-10-21 Cutting Edge Optronics, Inc. Laser diode packaging
US6288815B1 (en) 1999-03-31 2001-09-11 Philips Electronics North America Corporation Light scanner with cylindrical lenses
AU3994799A (en) * 1999-05-14 2000-12-05 3Dmetrics, Incorporated Color structured light 3d-imaging system
US6888871B1 (en) * 2000-07-12 2005-05-03 Princeton Optronics, Inc. VCSEL and VCSEL array having integrated microlenses for use in a semiconductor laser pumped solid state laser system
EP1334463B1 (en) 2000-11-06 2010-03-31 Koninklijke Philips Electronics N.V. Method of measuring the movement of an input device.
WO2002082201A1 (en) * 2001-04-04 2002-10-17 Instro Precision Limited Image analysis apparatus
JP2002372701A (en) * 2001-06-13 2002-12-26 Ricoh Co Ltd Picture display device
US7133022B2 (en) 2001-11-06 2006-11-07 Keyotee, Inc. Apparatus for image projection
KR100440959B1 (en) * 2001-11-08 2004-07-21 삼성전자주식회사 Illumination system and projection system employing it
US20040184270A1 (en) * 2003-03-17 2004-09-23 Halter Michael A. LED light module with micro-reflector cavities
JP4022498B2 (en) * 2003-04-18 2007-12-19 インターナショナル・ビジネス・マシーンズ・コーポレーション OPTICAL LINK MODULE, OPTICAL CONNECTION METHOD, INFORMATION PROCESSING DEVICE INCLUDING THE OPTICAL LINK MODULE, SIGNAL TRANSFER METHOD, PRISM, AND MANUFACTURING METHOD THEREOF
US7289090B2 (en) * 2003-12-10 2007-10-30 Texas Instruments Incorporated Pulsed LED scan-ring array for boosting display system lumens
EP1569276A1 (en) * 2004-02-27 2005-08-31 Heptagon OY Micro-optics on optoelectronics
GB2417824A (en) * 2004-09-02 2006-03-08 Custom Interconnect Ltd LED light source
IL165212A (en) 2004-11-15 2012-05-31 Elbit Systems Electro Optics Elop Ltd Device for scanning light
CN1725042B (en) 2005-06-30 2010-11-24 昆明理工大学 Scanning type optical grating write in method based on palbot interferometer and scanning type Talbot interferometer
US8400494B2 (en) 2005-10-11 2013-03-19 Primesense Ltd. Method and system for object reconstruction
US8792978B2 (en) 2010-05-28 2014-07-29 Lockheed Martin Corporation Laser-based nerve stimulators for, E.G., hearing restoration in cochlear prostheses and method
US7544945B2 (en) 2006-02-06 2009-06-09 Avago Technologies General Ip (Singapore) Pte. Ltd. Vertical cavity surface emitting laser (VCSEL) array laser scanner
DE102006011284A1 (en) * 2006-02-28 2007-08-30 Osram Opto Semiconductors Gmbh Semiconductor laser device, has pumping radiation source and vertical section is configured so as to be index guiding for pumping radiation in lateral direction perpendicular to main direction of pumping radiation
WO2007105215A2 (en) * 2006-03-14 2007-09-20 Prime Sense Ltd. Depth-varying light fields for three dimensional sensing
US20070242254A1 (en) 2006-03-17 2007-10-18 Nikon Corporation Exposure apparatus and device manufacturing method
JP5239124B2 (en) 2006-03-31 2013-07-17 株式会社ニコン Camera with built-in projector
JP4823743B2 (en) * 2006-04-03 2011-11-24 三星電子株式会社 Imaging apparatus and imaging method
KR101439434B1 (en) 2007-10-05 2014-09-12 삼성전자주식회사 Image sensor and method of fabricating the same
WO2009101236A1 (en) 2008-02-13 2009-08-20 Nokia Corporation Display device and a method for illuminating a light modulator array of a display device
US9244287B2 (en) 2008-05-09 2016-01-26 Reald Inc. Optical systems with compact back focal lengths
KR20120048018A (en) 2009-08-17 2012-05-14 쓰리엠 이노베이티브 프로퍼티즈 컴파니 Dual total internal reflection polarizing beamsplitter
JP5588310B2 (en) 2009-11-15 2014-09-10 プライムセンス リミテッド Optical projector with beam monitor
JP5452197B2 (en) 2009-12-03 2014-03-26 パナソニック株式会社 MEMS optical scanner
US20110187878A1 (en) 2010-02-02 2011-08-04 Primesense Ltd. Synchronization of projected illumination with rolling shutter of image sensor
US20110188054A1 (en) 2010-02-02 2011-08-04 Primesense Ltd Integrated photonics module for optical projection
WO2012020380A1 (en) 2010-08-11 2012-02-16 Primesense Ltd. Scanning projectors and image capture modules for 3d mapping
US9036158B2 (en) 2010-08-11 2015-05-19 Apple Inc. Pattern projector
WO2012066501A1 (en) 2010-11-19 2012-05-24 Primesense Ltd. Depth mapping using time-coded illumination
US9131136B2 (en) 2010-12-06 2015-09-08 Apple Inc. Lens arrays for pattern projection and imaging
DE102011001387A1 (en) 2011-03-18 2012-09-20 First Sensor AG Sampling method for sampling field by optical sampling system or scanning system, involves transmitting test light signals by transmitters of transmitter arrangement of optical sampling system
US8908277B2 (en) 2011-08-09 2014-12-09 Apple Inc Lens array projector
US8749796B2 (en) 2011-08-09 2014-06-10 Primesense Ltd. Projectors of structured light
JP5984950B2 (en) 2011-10-24 2016-09-06 スリーエム イノベイティブ プロパティズ カンパニー Tilted dichroic polarizing beam splitter
WO2014093085A1 (en) 2012-12-10 2014-06-19 3M Innovative Properties Company Reflective fly eye array illuminator
US9690107B2 (en) 2013-03-15 2017-06-27 Trumpf Laser Gmbh Device for wavelength combining of laser beams
US20150260830A1 (en) 2013-07-12 2015-09-17 Princeton Optronics Inc. 2-D Planar VCSEL Source for 3-D Imaging
CN104460196A (en) 2013-09-18 2015-03-25 光宝科技股份有限公司 Projection module
US10203399B2 (en) 2013-11-12 2019-02-12 Big Sky Financial Corporation Methods and apparatus for array based LiDAR systems with reduced interference
DE102014104027B4 (en) 2014-03-24 2020-10-29 Sick Ag Optoelectronic device and method for capturing object information
DE112015002814B4 (en) 2014-06-13 2022-04-07 Mitsubishi Electric Corporation Virtual image display device
US20170142393A1 (en) 2014-06-27 2017-05-18 Heptagon Micro Optics Pte. Ltd. Structured Light Imaging System and Method
CN106574963B (en) 2014-08-12 2018-03-06 三菱电机株式会社 Laser radar apparatus
US10088558B2 (en) 2014-08-15 2018-10-02 Aeye, Inc. Method and system for ladar transmission with spiral dynamic scan patterns
US9553423B2 (en) 2015-02-27 2017-01-24 Princeton Optronics Inc. Miniature structured light illuminator
JP6548727B2 (en) 2015-05-29 2019-07-24 オリンパス株式会社 Lighting device and measuring device
US10057553B2 (en) 2015-06-19 2018-08-21 Seiko Epson Corporation Light source device, illumination device, and projector
US10620300B2 (en) 2015-08-20 2020-04-14 Apple Inc. SPAD array with gated histogram construction
US9997551B2 (en) 2015-12-20 2018-06-12 Apple Inc. Spad array with pixel-level bias control
US10324171B2 (en) 2015-12-20 2019-06-18 Apple Inc. Light detection and ranging sensor
US10627490B2 (en) 2016-01-31 2020-04-21 Velodyne Lidar, Inc. Multiple pulse, LIDAR based 3-D imaging
KR101641479B1 (en) 2016-03-24 2016-07-20 김상수 Apparatus for displaying a stereoscopic image
US10305247B2 (en) 2016-08-30 2019-05-28 Apple Inc. Radiation source with a small-angle scanning array
US10362295B2 (en) 2016-09-19 2019-07-23 Apple Inc. Optical apparatus with beam steering and position feedback
US11852727B2 (en) 2017-12-18 2023-12-26 Apple Inc. Time-of-flight sensing using an addressable array of emitters

Patent Citations (85)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4542376A (en) * 1983-11-03 1985-09-17 Burroughs Corporation System for electronically displaying portions of several different images on a CRT screen through respective prioritized viewports
US4843568A (en) * 1986-04-11 1989-06-27 Krueger Myron W Real time perception of and response to the actions of an unencumbered participant/user
US4802759A (en) * 1986-08-11 1989-02-07 Goro Matsumoto Three-dimensional shape measuring apparatus
US4850673A (en) * 1987-11-23 1989-07-25 U. S. Philips Corporation Optical scanning apparatus which detects scanning spot focus error
US5648951A (en) * 1990-04-12 1997-07-15 Matsushita Electic Industrial Co. Ltd. Movable optical head integrally incorporated with objective lens and hologram element
US5075562A (en) * 1990-09-20 1991-12-24 Eastman Kodak Company Method and apparatus for absolute Moire distance measurements using a grating printed on or attached to a surface
US5691989A (en) * 1991-07-26 1997-11-25 Accuwave Corporation Wavelength stabilized laser sources using feedback from volume holograms
US5483261A (en) * 1992-02-14 1996-01-09 Itu Research, Inc. Graphical input controller and method with rear screen image detection
US5636025A (en) * 1992-04-23 1997-06-03 Medar, Inc. System for optically measuring the surface contour of a part using more fringe techniques
US5477383A (en) * 1993-02-05 1995-12-19 Apa Optics, Inc. Optical array method and apparatus
US5742262A (en) * 1993-06-23 1998-04-21 Olympus Optical Co., Ltd. Image display apparatus
US5856871A (en) * 1993-08-18 1999-01-05 Applied Spectral Imaging Ltd. Film thickness mapping using interferometric spectral imaging
US5606181A (en) * 1994-03-29 1997-02-25 Ricoh Company, Ltd. Edge emitting type light emitting diode array heads
US5630043A (en) * 1995-05-11 1997-05-13 Cirrus Logic, Inc. Animated texture map apparatus and method for 3-D image displays
US6100517A (en) * 1995-06-22 2000-08-08 3Dv Systems Ltd. Three dimensional camera
US5781332A (en) * 1995-09-04 1998-07-14 Nec Corporation Variable wavelength optical filter
US6167151A (en) * 1996-12-15 2000-12-26 Cognitens, Ltd. Apparatus and method for 3-dimensional surface geometry reconstruction
US6438263B2 (en) * 1996-12-15 2002-08-20 Dan Albeck Apparatus and method for 3-dimensional surface geometry reconstruction
US5838428A (en) * 1997-02-28 1998-11-17 United States Of America As Represented By The Secretary Of The Navy System and method for high resolution range imaging with split light source and pattern mask
US6002520A (en) * 1997-04-25 1999-12-14 Hewlett-Packard Company Illumination system for creating a desired irradiance profile using diffractive optical elements
US6262740B1 (en) * 1997-08-01 2001-07-17 Terarecon, Inc. Method for rendering sections of a volume data set
US6101269A (en) * 1997-12-19 2000-08-08 Lifef/X Networks, Inc. Apparatus and method for rapid 3D image parametrization
US6560019B2 (en) * 1998-02-05 2003-05-06 Canon Kabushiki Kaisha Diffractive optical element and optical system having the same
US6259561B1 (en) * 1999-03-26 2001-07-10 The University Of Rochester Optical system for diffusing light
US6751344B1 (en) * 1999-05-28 2004-06-15 Champion Orthotic Investments, Inc. Enhanced projector system for machine vision
US20030057972A1 (en) * 1999-07-26 2003-03-27 Paul Pfaff Voltage testing and measurement
US6803777B2 (en) * 1999-07-26 2004-10-12 Paul Pfaff Voltage testing and measurement
US7009742B2 (en) * 1999-12-10 2006-03-07 Xyz Imaging, Inc. Holographic printer
US20050200925A1 (en) * 1999-12-10 2005-09-15 Xyz Imaging, Inc. Holographic printer
US6927852B2 (en) * 2000-05-05 2005-08-09 Applera Corporation Optical system and method for optically analyzing light from a sample
US20080018595A1 (en) * 2000-07-24 2008-01-24 Gesturetek, Inc. Video-based image control system
US6686921B1 (en) * 2000-08-01 2004-02-03 International Business Machines Corporation Method and apparatus for acquiring a set of consistent image maps to represent the color of the surface of an object
US6754370B1 (en) * 2000-08-14 2004-06-22 The Board Of Trustees Of The Leland Stanford Junior University Real-time structured light range scanning of moving scenes
US6583873B1 (en) * 2000-09-25 2003-06-24 The Carnegie Institution Of Washington Optical devices having a wavelength-tunable dispersion assembly that has a volume dispersive diffraction grating
US6813440B1 (en) * 2000-10-10 2004-11-02 The Hong Kong Polytechnic University Body scanner
US20020075456A1 (en) * 2000-12-20 2002-06-20 Olympus Optical Co., Ltd. 3D image acquisition apparatus and 3D image acquisition method
US7013040B2 (en) * 2000-12-20 2006-03-14 Olympus Optical Co., Ltd. 3D image acquisition apparatus and 3D image acquisition method
US6611000B2 (en) * 2001-03-14 2003-08-26 Matsushita Electric Industrial Co., Ltd. Lighting device
US20040012958A1 (en) * 2001-04-23 2004-01-22 Takuma Hashimoto Light emitting device comprising led chip
US6825985B2 (en) * 2001-07-13 2004-11-30 Mems Optical, Inc. Autostereoscopic display with rotated microlens and method of displaying multidimensional images, especially color images
US6741251B2 (en) * 2001-08-16 2004-05-25 Hewlett-Packard Development Company, L.P. Method and apparatus for varying focus in a scene
US20030156756A1 (en) * 2002-02-15 2003-08-21 Gokturk Salih Burak Gesture recognition system using depth perceptive sensors
US7369685B2 (en) * 2002-04-05 2008-05-06 Identix Corporation Vision-based operating method and system
US20060072851A1 (en) * 2002-06-15 2006-04-06 Microsoft Corporation Deghosting mosaics using multiperspective plane sweep
US20040130730A1 (en) * 2002-11-21 2004-07-08 Michel Cantin Fast 3D height measurement method and system
US20040174770A1 (en) * 2002-11-27 2004-09-09 Rees Frank L. Gauss-Rees parametric ultrawideband system
US20040218262A1 (en) * 2003-02-21 2004-11-04 Chuang Yung-Ho Inspection system using small catadioptric objective
US20040228519A1 (en) * 2003-03-10 2004-11-18 Cranial Technologies, Inc. Automatic selection of cranial remodeling device trim lines
US7127101B2 (en) * 2003-03-10 2006-10-24 Cranul Technologies, Inc. Automatic selection of cranial remodeling device trim lines
US20070057946A1 (en) * 2003-07-24 2007-03-15 Dan Albeck Method and system for the three-dimensional surface reconstruction of an object
US6940583B2 (en) * 2003-07-28 2005-09-06 International Business Machines Corporation Method and apparatus for amplitude filtering in the frequency plane of a lithographic projection system
US20070060336A1 (en) * 2003-09-15 2007-03-15 Sony Computer Entertainment Inc. Methods and systems for enabling depth and direction detection when interfacing with a computer program
US7112774B2 (en) * 2003-10-09 2006-09-26 Avago Technologies Sensor Ip (Singapore) Pte. Ltd CMOS stereo imaging system and method
US20050178950A1 (en) * 2004-02-18 2005-08-18 Fujinon Corporation Electronic imaging system
US20060001055A1 (en) * 2004-02-23 2006-01-05 Kazuhiko Ueno Led and fabrication method of same
US7227618B1 (en) * 2004-03-24 2007-06-05 Baokang Bi Pattern generating systems
US7304735B2 (en) * 2004-04-02 2007-12-04 Kla-Tencor Technologies Broadband wavelength selective filter
US20050271279A1 (en) * 2004-05-14 2005-12-08 Honda Motor Co., Ltd. Sign based human-machine interaction
US7335898B2 (en) * 2004-07-23 2008-02-26 Ge Healthcare Niagara Inc. Method and apparatus for fluorescent confocal microscopy
US7120228B2 (en) * 2004-09-21 2006-10-10 Jordan Valley Applied Radiation Ltd. Combined X-ray reflectometer and diffractometer
US7551719B2 (en) * 2004-09-21 2009-06-23 Jordan Valley Semiconductord Ltd Multifunction X-ray analysis system
US20060252169A1 (en) * 2004-10-07 2006-11-09 Takeshi Ashida Transparent member, optical device using transparent member and method of manufacturing optical device
US7076024B2 (en) * 2004-12-01 2006-07-11 Jordan Valley Applied Radiation, Ltd. X-ray apparatus with dual monochromators
US20060156756A1 (en) * 2005-01-20 2006-07-20 Becke Paul E Phase change and insulating properties container and method of use
US20060252167A1 (en) * 2005-05-04 2006-11-09 Wang Tak K Silicon optical package with 45 degree turning mirror
US20060269896A1 (en) * 2005-05-27 2006-11-30 Yongqian Liu High speed 3D scanner and uses thereof
US20070019909A1 (en) * 2005-07-22 2007-01-25 Seiko Epson Corporation Illumination device, image display device, and projector
US20090096783A1 (en) * 2005-10-11 2009-04-16 Alexander Shpunt Three-dimensional sensing using speckle patterns
US20080106746A1 (en) * 2005-10-11 2008-05-08 Alexander Shpunt Depth-varying light fields for three dimensional sensing
US7433024B2 (en) * 2006-02-27 2008-10-07 Prime Sense Ltd. Range mapping using speckle decorrelation
US20100013860A1 (en) * 2006-03-08 2010-01-21 Electronic Scripting Products, Inc. Computer interface employing a manipulated object with absolute pose detection component and a display
US20100142014A1 (en) * 2006-11-17 2010-06-10 Celloptic, Inc. System, apparatus and method for extracting three-dimensional information of an object from received electromagnetic radiation
US20080198355A1 (en) * 2006-12-27 2008-08-21 Cambridge Research & Instrumentation, Inc Surface measurement of in-vivo subjects using spot projector
US20100020078A1 (en) * 2007-01-21 2010-01-28 Prime Sense Ltd Depth mapping using multi-beam illumination
US20080212835A1 (en) * 2007-03-01 2008-09-04 Amon Tavor Object Tracking by 3-Dimensional Modeling
US20080240502A1 (en) * 2007-04-02 2008-10-02 Barak Freedman Depth mapping using projected patterns
US20100118123A1 (en) * 2007-04-02 2010-05-13 Prime Sense Ltd Depth mapping using projected patterns
US20080278572A1 (en) * 2007-04-23 2008-11-13 Morteza Gharib Aperture system with spatially-biased aperture shapes and positions (SBPSP) for static and dynamic 3-D defocusing-based imaging
US7700904B2 (en) * 2007-07-18 2010-04-20 Funai Electric Co., Ltd. Compound-eye imaging device
US20090183125A1 (en) * 2008-01-14 2009-07-16 Prime Sense Ltd. Three-dimensional user interface
US20090183152A1 (en) * 2008-01-16 2009-07-16 Dell Products, Lp Method to Dynamically Provision Additional Computer Resources to Handle Peak Database Workloads
US20090185274A1 (en) * 2008-01-21 2009-07-23 Prime Sense Ltd. Optical designs for zero order reduction
US20100284082A1 (en) * 2008-01-21 2010-11-11 Primesense Ltd. Optical pattern projection
US20100007717A1 (en) * 2008-07-09 2010-01-14 Prime Sense Ltd Integrated processor for 3d mapping
US20110091258A1 (en) * 2009-10-20 2011-04-21 Canon Kabushiki Kaisha Image forming apparatus

Cited By (161)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9330324B2 (en) 2005-10-11 2016-05-03 Apple Inc. Error compensation in three-dimensional mapping
US9066084B2 (en) 2005-10-11 2015-06-23 Apple Inc. Method and system for object reconstruction
US8374397B2 (en) 2005-10-11 2013-02-12 Primesense Ltd Depth-varying light fields for three dimensional sensing
US9239467B2 (en) 2008-01-21 2016-01-19 Apple Inc. Optical pattern projection
US9736459B2 (en) 2010-02-02 2017-08-15 Apple Inc. Generation of patterned radiation
US10609357B2 (en) 2010-02-02 2020-03-31 Apple Inc. Synchronization of projected illumination with activation of sensing elements
US8982182B2 (en) 2010-03-01 2015-03-17 Apple Inc. Non-uniform spatial resource allocation for depth mapping
US9098931B2 (en) 2010-08-11 2015-08-04 Apple Inc. Scanning projectors and image capture modules for 3D mapping
US9036158B2 (en) 2010-08-11 2015-05-19 Apple Inc. Pattern projector
US9066087B2 (en) 2010-11-19 2015-06-23 Apple Inc. Depth mapping using time-coded illumination
US20120140096A1 (en) * 2010-12-01 2012-06-07 Sony Ericsson Mobile Communications Ab Timing Solution for Projector Camera Devices and Systems
US9167138B2 (en) 2010-12-06 2015-10-20 Apple Inc. Pattern projection and imaging using lens arrays
US9131136B2 (en) 2010-12-06 2015-09-08 Apple Inc. Lens arrays for pattern projection and imaging
US9857868B2 (en) 2011-03-19 2018-01-02 The Board Of Trustees Of The Leland Stanford Junior University Method and system for ergonomic touch-free interface
US9030528B2 (en) 2011-04-04 2015-05-12 Apple Inc. Multi-zone imaging sensor and lens array
US9504920B2 (en) 2011-04-25 2016-11-29 Aquifi, Inc. Method and system to create three-dimensional mapping in a two-dimensional game
US8237835B1 (en) * 2011-05-19 2012-08-07 Aeon Imaging, LLC Confocal imaging device using spatially modulated illumination with electronic rolling shutter detection
US20130021474A1 (en) * 2011-07-20 2013-01-24 Raytheon Company Rolling-shutter imaging system with synchronized scanning illumination and methods for higher-resolution imaging
US8908277B2 (en) 2011-08-09 2014-12-09 Apple Inc Lens array projector
US8749796B2 (en) 2011-08-09 2014-06-10 Primesense Ltd. Projectors of structured light
US20140354803A1 (en) * 2011-12-19 2014-12-04 Canon Kabushiki Kaisha Information processing apparatus, control method therefor, and computer-readable storage medium
WO2013094347A1 (en) * 2011-12-19 2013-06-27 Canon Kabushiki Kaisha Information processing apparatus, control method therefor, and computer-readable storage medium
US9600078B2 (en) 2012-02-03 2017-03-21 Aquifi, Inc. Method and system enabling natural user interface gestures with an electronic system
US9651417B2 (en) 2012-02-15 2017-05-16 Apple Inc. Scanning depth engine
US9157790B2 (en) 2012-02-15 2015-10-13 Apple Inc. Integrated optoelectronic modules with transmitter, receiver and beam-combining optics for aligning a beam axis with a collection axis
US9201237B2 (en) 2012-03-22 2015-12-01 Apple Inc. Diffraction-based sensing of mirror position
US10240912B2 (en) * 2012-06-13 2019-03-26 Koninklijke Philips N.V. Determining a propagation velocity for a surface wave
US20150323311A1 (en) * 2012-06-13 2015-11-12 Koninklijke Philips N.V. Determining a propagation velocity for a surface wave
US10063846B2 (en) 2012-06-18 2018-08-28 Microsoft Technology Licensing, Llc Selective illumination of a region within a field of view
US9674436B2 (en) * 2012-06-18 2017-06-06 Microsoft Technology Licensing, Llc Selective imaging zones of an imaging sensor
US20130335546A1 (en) * 2012-06-18 2013-12-19 Randall T. Crane Selective imaging
US9398229B2 (en) 2012-06-18 2016-07-19 Microsoft Technology Licensing, Llc Selective illumination of a region within a field of view
US8934675B2 (en) 2012-06-25 2015-01-13 Aquifi, Inc. Systems and methods for tracking human hands by performing parts based template matching using images from multiple viewpoints
US9098739B2 (en) 2012-06-25 2015-08-04 Aquifi, Inc. Systems and methods for tracking human hands using parts based template matching
US9111135B2 (en) 2012-06-25 2015-08-18 Aquifi, Inc. Systems and methods for tracking human hands using parts based template matching using corresponding pixels in bounded regions of a sequence of frames that are a specified distance interval from a reference camera
US8655021B2 (en) 2012-06-25 2014-02-18 Imimtek, Inc. Systems and methods for tracking human hands by performing parts based template matching using images from multiple viewpoints
US8830312B2 (en) 2012-06-25 2014-09-09 Aquifi, Inc. Systems and methods for tracking human hands using parts based template matching within bounded regions
US9310891B2 (en) 2012-09-04 2016-04-12 Aquifi, Inc. Method and system enabling natural user interface gestures with user wearable glasses
US20140111619A1 (en) * 2012-10-19 2014-04-24 Electronics And Telecommunications Research Institute Device and method for acquiring image
US20150341619A1 (en) * 2013-01-01 2015-11-26 Inuitive Ltd. Method and system for light patterning and imaging
WO2014106843A2 (en) * 2013-01-01 2014-07-10 Inuitive Ltd. Method and system for light patterning and imaging
WO2014106843A3 (en) * 2013-01-01 2014-09-04 Inuitive Ltd. Method and system for light patterning and imaging
US10466359B2 (en) * 2013-01-01 2019-11-05 Inuitive Ltd. Method and system for light patterning and imaging
US9270876B2 (en) 2013-01-05 2016-02-23 The Lightco Inc. Methods and apparatus for using multiple optical chains in parallel with multiple different exposure times
US9690079B2 (en) 2013-01-05 2017-06-27 Light Labs Inc. Camera methods and apparatus using optical chain modules which alter the direction of received light
US9568713B2 (en) 2013-01-05 2017-02-14 Light Labs Inc. Methods and apparatus for using multiple optical chains in parallel to support separate color-capture
US9282228B2 (en) 2013-01-05 2016-03-08 The Lightco Inc. Camera methods and apparatus using optical chain modules which alter the direction of received light
US9547160B2 (en) 2013-01-05 2017-01-17 Light Labs Inc. Methods and apparatus for capturing and/or processing images
US9671595B2 (en) 2013-01-05 2017-06-06 Light Labs Inc. Methods and apparatus for using multiple optical chains in paralell
US9129155B2 (en) 2013-01-30 2015-09-08 Aquifi, Inc. Systems and methods for initializing motion tracking of human hands using template matching within bounded regions determined using a depth map
US8615108B1 (en) 2013-01-30 2013-12-24 Imimtek, Inc. Systems and methods for initializing motion tracking of human hands
US9092665B2 (en) 2013-01-30 2015-07-28 Aquifi, Inc Systems and methods for initializing motion tracking of human hands
US9407837B2 (en) 2013-02-28 2016-08-02 Google Inc. Depth sensor using modulated light projector and image sensor with color and IR sensing
US10038893B2 (en) 2013-02-28 2018-07-31 Google Llc Context-based depth sensor control
US9398287B2 (en) 2013-02-28 2016-07-19 Google Technology Holdings LLC Context-based depth sensor control
US20140240469A1 (en) * 2013-02-28 2014-08-28 Motorola Mobility Llc Electronic Device with Multiview Image Capture and Depth Sensing
US9298266B2 (en) 2013-04-02 2016-03-29 Aquifi, Inc. Systems and methods for implementing three-dimensional (3D) gesture based graphical user interfaces (GUI) that incorporate gesture reactive interface objects
US9825425B2 (en) 2013-06-19 2017-11-21 Apple Inc. Integrated structured-light projector comprising light-emitting elements on a substrate
US20150002734A1 (en) * 2013-07-01 2015-01-01 Motorola Mobility Llc Electronic Device with Modulated Light Flash Operation for Rolling Shutter Image Sensor
US10250789B2 (en) * 2013-07-01 2019-04-02 Google Llc Electronic device with modulated light flash operation for rolling shutter image sensor
US20170150021A1 (en) * 2013-07-01 2017-05-25 Google Inc. Electronic Device with Modulated Light Flash Operation for Rolling Shutter Image Sensor
US9798388B1 (en) 2013-07-31 2017-10-24 Aquifi, Inc. Vibrotactile system to augment 3D input systems
US10142553B2 (en) * 2013-09-16 2018-11-27 Intel Corporation Camera and light source synchronization for object tracking
EP3047641A4 (en) * 2013-09-16 2017-03-08 Intel Corporation Camera and light source synchronization for object tracking
US9549127B2 (en) 2013-10-18 2017-01-17 Light Labs Inc. Image capture control methods and apparatus
US9851527B2 (en) 2013-10-18 2017-12-26 Light Labs Inc. Methods and apparatus for capturing and/or combining images
US9578252B2 (en) 2013-10-18 2017-02-21 Light Labs Inc. Methods and apparatus for capturing images using optical chains and/or for using captured images
US9557520B2 (en) 2013-10-18 2017-01-31 Light Labs Inc. Synchronized image capture methods and apparatus
US9557519B2 (en) 2013-10-18 2017-01-31 Light Labs Inc. Methods and apparatus for implementing a camera device supporting a number of different focal lengths
US10509208B2 (en) * 2013-10-18 2019-12-17 Light Labs Inc. Methods and apparatus for implementing and/or using a camera device
US9551854B2 (en) 2013-10-18 2017-01-24 Light Labs Inc. Methods and apparatus for controlling sensors to capture images in a synchronized manner
US9423588B2 (en) 2013-10-18 2016-08-23 The Lightco Inc. Methods and apparatus for supporting zoom operations
US10038860B2 (en) * 2013-10-18 2018-07-31 Light Labs Inc. Methods and apparatus for controlling sensors to capture images in a synchronized manner
US10120159B2 (en) 2013-10-18 2018-11-06 Light Labs Inc. Methods and apparatus for supporting zoom operations
US10009530B2 (en) * 2013-10-18 2018-06-26 Light Labs Inc. Methods and apparatus for synchronized image capture using camera modules with different focal lengths
US9544501B2 (en) 2013-10-18 2017-01-10 Light Labs Inc. Methods and apparatus for implementing and/or using a camera device
US20170208230A1 (en) * 2013-10-18 2017-07-20 Light Labs Inc. Image capture related methods and apparatus
US9451171B2 (en) 2013-10-18 2016-09-20 The Lightco Inc. Zoom related methods and apparatus
US9374514B2 (en) 2013-10-18 2016-06-21 The Lightco Inc. Methods and apparatus relating to a camera including multiple optical chains
US9749511B2 (en) 2013-10-18 2017-08-29 Light Labs Inc. Methods and apparatus relating to a camera including multiple optical chains
US9563033B2 (en) 2013-10-18 2017-02-07 Light Labs Inc. Methods and apparatus for capturing images and/or for using captured images
US9325906B2 (en) 2013-10-18 2016-04-26 The Lightco Inc. Methods and apparatus relating to a thin camera device
US9467627B2 (en) 2013-10-26 2016-10-11 The Lightco Inc. Methods and apparatus for use with multiple optical chains
US9736365B2 (en) 2013-10-26 2017-08-15 Light Labs Inc. Zoom related methods and apparatus
US9426365B2 (en) 2013-11-01 2016-08-23 The Lightco Inc. Image stabilization related methods and apparatus
US9686471B2 (en) 2013-11-01 2017-06-20 Light Labs Inc. Methods and apparatus relating to image stabilization
US9528906B1 (en) 2013-12-19 2016-12-27 Apple Inc. Monitoring DOE performance using total internal reflection
US9554031B2 (en) 2013-12-31 2017-01-24 Light Labs Inc. Camera focusing related methods and apparatus
US9507417B2 (en) 2014-01-07 2016-11-29 Aquifi, Inc. Systems and methods for implementing head tracking based graphical user interfaces (GUI) that incorporate gesture reactive interface objects
US9619105B1 (en) 2014-01-30 2017-04-11 Aquifi, Inc. Systems and methods for gesture based interaction with viewpoint dependent user interfaces
WO2015127328A3 (en) * 2014-02-21 2015-12-23 The Lightco Inc. Illumination methods and apparatus
US9979878B2 (en) 2014-02-21 2018-05-22 Light Labs Inc. Intuitive camera user interface methods and apparatus
US9462170B2 (en) 2014-02-21 2016-10-04 The Lightco Inc. Lighting methods and apparatus
US20150264335A1 (en) * 2014-03-13 2015-09-17 Samsung Electronics Co., Ltd. Image pickup apparatus and method for generating image having depth information
US10375292B2 (en) * 2014-03-13 2019-08-06 Samsung Electronics Co., Ltd. Image pickup apparatus and method for generating image having depth information
US10191356B2 (en) 2014-07-04 2019-01-29 Light Labs Inc. Methods and apparatus relating to detection and/or indicating a dirty lens condition
US10110794B2 (en) 2014-07-09 2018-10-23 Light Labs Inc. Camera device including multiple optical chains and related methods
US10425814B2 (en) 2014-09-24 2019-09-24 Princeton Identity, Inc. Control of wireless communication device capability in a mobile device with a biometric key
US9912865B2 (en) 2014-10-17 2018-03-06 Light Labs Inc. Methods and apparatus for supporting burst modes of camera operation
US9912864B2 (en) 2014-10-17 2018-03-06 Light Labs Inc. Methods and apparatus for using a camera device to support multiple modes of operation
US10484584B2 (en) 2014-12-03 2019-11-19 Princeton Identity, Inc. System and method for mobile device biometric add-on
US9998638B2 (en) 2014-12-17 2018-06-12 Light Labs Inc. Methods and apparatus for implementing and using camera devices
US9544503B2 (en) 2014-12-30 2017-01-10 Light Labs Inc. Exposure control methods and apparatus
US11245891B2 (en) * 2015-01-21 2022-02-08 Nevermind Capital Llc Methods and apparatus for environmental measurements and/or stereoscopic image capture
US20210409582A1 (en) * 2015-02-02 2021-12-30 Apple Inc. Focusing Lighting Module
US11588961B2 (en) * 2015-02-02 2023-02-21 Apple Inc. Focusing lighting module
US9824427B2 (en) 2015-04-15 2017-11-21 Light Labs Inc. Methods and apparatus for generating a sharp image
US9967535B2 (en) 2015-04-17 2018-05-08 Light Labs Inc. Methods and apparatus for reducing noise in images
US10091447B2 (en) 2015-04-17 2018-10-02 Light Labs Inc. Methods and apparatus for synchronizing readout of multiple image sensors
US9857584B2 (en) 2015-04-17 2018-01-02 Light Labs Inc. Camera device methods, apparatus and components
US10075651B2 (en) 2015-04-17 2018-09-11 Light Labs Inc. Methods and apparatus for capturing images using multiple camera modules in an efficient manner
US9930233B2 (en) 2015-04-22 2018-03-27 Light Labs Inc. Filter mounting methods and apparatus and related camera apparatus
US9525863B2 (en) 2015-04-29 2016-12-20 Apple Inc. Time-of-flight depth mapping with flexible scan pattern
US10129483B2 (en) 2015-06-23 2018-11-13 Light Labs Inc. Methods and apparatus for implementing zoom using one or more moveable camera modules
US10491806B2 (en) 2015-08-03 2019-11-26 Light Labs Inc. Camera device control related methods and apparatus
US10012831B2 (en) 2015-08-03 2018-07-03 Apple Inc. Optical monitoring of scan parameters
US10365480B2 (en) 2015-08-27 2019-07-30 Light Labs Inc. Methods and apparatus for implementing and/or using camera devices with one or more light redirection devices
US10663691B2 (en) * 2015-09-11 2020-05-26 Ams Sensors Singapore Pte. Ltd. Imaging devices having autofocus control in response to the user touching the display screen
US20190271828A1 (en) * 2015-09-11 2019-09-05 Ams Sensors Singapore Pte. Ltd. Imaging devices having autofocusd control
US9749549B2 (en) 2015-10-06 2017-08-29 Light Labs Inc. Methods and apparatus for facilitating selective blurring of one or more image portions
TWI616711B (en) * 2015-12-11 2018-03-01 Ningbo Sunny Opotech Co Ltd Lighting module and its application
US10225445B2 (en) 2015-12-18 2019-03-05 Light Labs Inc. Methods and apparatus for providing a camera lens or viewing point indicator
US10003738B2 (en) 2015-12-18 2018-06-19 Light Labs Inc. Methods and apparatus for detecting and/or indicating a blocked sensor or camera module
US10643087B2 (en) 2016-01-12 2020-05-05 Princeton Identity, Inc. Systems and methods of biometric analysis to determine a live subject
US10452936B2 (en) 2016-01-12 2019-10-22 Princeton Identity Systems and methods of biometric analysis with a spectral discriminator
US10943138B2 (en) 2016-01-12 2021-03-09 Princeton Identity, Inc. Systems and methods of biometric analysis to determine lack of three-dimensionality
US10762367B2 (en) 2016-01-12 2020-09-01 Princeton Identity Systems and methods of biometric analysis to determine natural reflectivity
US10643088B2 (en) 2016-01-12 2020-05-05 Princeton Identity, Inc. Systems and methods of biometric analysis with a specularity characteristic
US10306218B2 (en) 2016-03-22 2019-05-28 Light Labs Inc. Camera calibration apparatus and methods
US10366296B2 (en) 2016-03-31 2019-07-30 Princeton Identity, Inc. Biometric enrollment systems and methods
US10373008B2 (en) 2016-03-31 2019-08-06 Princeton Identity, Inc. Systems and methods of biometric analysis with adaptive trigger
US20170288366A1 (en) * 2016-04-05 2017-10-05 Aquifi, Inc. Thin laser package for optical applications
US10122146B2 (en) * 2016-04-05 2018-11-06 Aquifi, Inc. Thin laser package for optical applications
US9948832B2 (en) 2016-06-22 2018-04-17 Light Labs Inc. Methods and apparatus for synchronized image capture in a device including optical chains with different orientations
US10073004B2 (en) 2016-09-19 2018-09-11 Apple Inc. DOE defect monitoring utilizing total internal reflection
WO2018111860A1 (en) * 2016-12-12 2018-06-21 Princeton Identity, Inc. Systems and methods of biometric analysis
US10469832B1 (en) 2017-01-18 2019-11-05 Facebook Technologes, LLC Tileable structured light projection for wide field-of-view depth sensing
US10158845B2 (en) 2017-01-18 2018-12-18 Facebook Technologies, Llc Tileable structured light projection for wide field-of-view depth sensing
US10607096B2 (en) 2017-04-04 2020-03-31 Princeton Identity, Inc. Z-dimension user feedback biometric system
US10902104B2 (en) 2017-07-26 2021-01-26 Princeton Identity, Inc. Biometric security systems and methods
US10153614B1 (en) 2017-08-31 2018-12-11 Apple Inc. Creating arbitrary patterns on a 2-D uniform grid VCSEL array
EP3693785A4 (en) * 2017-10-25 2020-12-16 Huawei Technologies Co., Ltd. Projector, camera module and terminal device
US10447424B2 (en) 2018-01-18 2019-10-15 Apple Inc. Spatial multiplexing scheme
US20210397013A1 (en) * 2018-02-23 2021-12-23 Jabil Optics Germany GmbH Projector Module
US11137618B2 (en) * 2018-02-23 2021-10-05 Jabil Optics Germany GmbH Projector module
US10877285B2 (en) 2018-03-28 2020-12-29 Apple Inc. Wavelength-based spatial multiplexing scheme
US11237370B2 (en) 2018-04-30 2022-02-01 University Of Central Florida Research Foundation, Inc. Multiple inclined beam line-scanning imaging apparatus, methods, and applications
US11422292B1 (en) 2018-06-10 2022-08-23 Apple Inc. Super-blazed diffractive optical elements with sub-wavelength structures
US11493606B1 (en) 2018-09-12 2022-11-08 Apple Inc. Multi-beam scanning system
US11303817B2 (en) * 2018-12-27 2022-04-12 Koito Manufaciuring Co., Ltd. Active sensor, object identification system, vehicle and vehicle lamp
US11333895B1 (en) 2019-01-11 2022-05-17 Facebook Technologies, Llc Systems and methods for structured light projector operational safety
US11121288B2 (en) 2019-01-16 2021-09-14 Lextar Electronics Corporation Package structure of light-emitting element
US11322910B2 (en) 2019-02-21 2022-05-03 Apple Inc. Indium-phosphide VCSEL with dielectric DBR
US11418010B2 (en) 2019-04-01 2022-08-16 Apple Inc. VCSEL array with tight pitch and high efficiency
US11374381B1 (en) 2019-06-10 2022-06-28 Apple Inc. Integrated laser module
US11681019B2 (en) 2019-09-18 2023-06-20 Apple Inc. Optical module with stray light baffle
US11506762B1 (en) 2019-09-24 2022-11-22 Apple Inc. Optical module comprising an optical waveguide with reference light path
US11754767B1 (en) 2020-03-05 2023-09-12 Apple Inc. Display with overlaid waveguide
US20220174187A1 (en) * 2020-12-01 2022-06-02 Samsung Electronics Co., Ltd. Vision sensor, image processing device including the same, and operating method of the vision sensor
US11695895B2 (en) * 2020-12-01 2023-07-04 Samsung Electronics Co., Ltd. Vision sensor, image processing device including the same, and operating method of the vision sensor
CN114636546A (en) * 2022-03-10 2022-06-17 杭州海康威视数字技术股份有限公司 System for imaging synchronicity detection

Also Published As

Publication number Publication date
US9736459B2 (en) 2017-08-15
CN102193295A (en) 2011-09-21
US20190068951A1 (en) 2019-02-28
US20180070073A1 (en) 2018-03-08
US20130147921A1 (en) 2013-06-13
US10063835B2 (en) 2018-08-28
EP2363686A1 (en) 2011-09-07
CN102143342A (en) 2011-08-03
US10609357B2 (en) 2020-03-31
CN104360571A (en) 2015-02-18
JP2011160420A (en) 2011-08-18
CN102193295B (en) 2014-12-10
CN102143342B (en) 2015-03-25
CN104360571B (en) 2016-05-11

Similar Documents

Publication Publication Date Title
US10609357B2 (en) Synchronization of projected illumination with activation of sensing elements
US11320666B2 (en) Integrated structured-light projector
KR102338174B1 (en) Projectors of structured light
CN107424188B (en) Structured light projection module based on VCSEL array light source
CN106997603B (en) Depth camera based on VCSEL array light source
US9825425B2 (en) Integrated structured-light projector comprising light-emitting elements on a substrate
TWI719325B (en) Light ranging device with electronically scanned emitter array and synchronized sensor array
US10466359B2 (en) Method and system for light patterning and imaging
JP6977045B2 (en) Systems and methods for determining the distance to an object
US7745771B2 (en) Synchronous imaging using segmented illumination
US20140211215A1 (en) Projectors of structured light
CN109597211B (en) Projection module, depth camera and depth image acquisition method
KR102481774B1 (en) Image apparatus and operation method thereof
CN212135135U (en) 3D imaging device
US20200292306A1 (en) 3d camera system with rolling-shutter image sensor
WO2021200016A1 (en) Distance measurement device and light emitting device
RU2543688C2 (en) Camera and optical system for obtaining 3d images (versions)

Legal Events

Date Code Title Description
AS Assignment

Owner name: PRIMESENSE LTD., ISRAEL

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MOR, ZAFRIR;SHPUNT, ALEXANDER;SIGNING DATES FROM 20100414 TO 20100418;REEL/FRAME:024249/0390

STCB Information on status: application discontinuation

Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION

AS Assignment

Owner name: APPLE INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PRIMESENSE LTD.;REEL/FRAME:034293/0092

Effective date: 20140828

AS Assignment

Owner name: APPLE INC., CALIFORNIA

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE APPLICATION # 13840451 AND REPLACE IT WITH CORRECT APPLICATION # 13810451 PREVIOUSLY RECORDED ON REEL 034293 FRAME 0092. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:PRIMESENSE LTD.;REEL/FRAME:035624/0091

Effective date: 20140828