US8908063B2 - Method and apparatus for a time-of-flight sensor with charge storage - Google Patents
Method and apparatus for a time-of-flight sensor with charge storage Download PDFInfo
- Publication number
- US8908063B2 US8908063B2 US13/792,756 US201313792756A US8908063B2 US 8908063 B2 US8908063 B2 US 8908063B2 US 201313792756 A US201313792756 A US 201313792756A US 8908063 B2 US8908063 B2 US 8908063B2
- Authority
- US
- United States
- Prior art keywords
- time
- storage circuit
- charges
- frame storage
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H04N5/232—
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/483—Details of pulse systems
- G01S7/486—Receivers
- G01S7/4861—Circuits for detection, sampling, integration or read-out
- G01S7/4863—Detector arrays, e.g. charge-transfer gates
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
- G01S17/894—3D imaging with simultaneous measurement of time-of-flight at a 2D array of receiver pixels, e.g. time-of-flight cameras or flash lidar
Definitions
- This invention relates generally to time-of-flight sensors.
- Time-of-flight sensors are used to determine a distance of an object or plurality of objects from the sensor.
- Time-of-flight sensors operate by detecting light reflected off of a surface of the object where the light is emitted from an illuminator that is generally co-located with the sensor.
- a processing device can determine the round-trip travel time of light emitted from the light source and reflected off of the object back to the sensor. With this information, and with the knowledge of the speed of light (constant c), the processing device can determine the distance of the object.
- eight data points per pixel are used to calculate distances, though more or less may be used.
- two data points (X and Y) are acquired per pixel per frame with four frames captured sequentially.
- a processor cannot perform the calculations on the acquired data until all data has been captured.
- Time-of-flight sensors will store the captured data points in binary form in a digital memory, such as SRAM, DRAM, or other common memory formats, until such time as all data points are captured and the distance calculation can be completed.
- digital storage of all the data points can increase the cost of devices utilizing time-of-flight sensors, increase processing power, increase processing or information transportation time, increase space requirements, or may otherwise be impracticable.
- the memory required might be too large to supply on chip.
- such memory requirements may increase the cost of an end device by requiring a separate memory chip and interface to store the data points until processing time. Accordingly, it is desirable to provide an efficient and low cost solution to store the data points.
- a time-of-flight sensor includes on-chip analog memory to store at least one set of charges associated with an image frame. So configured, an amount of digital memory required for storage of image data prior to processing time is greatly reduced or eliminated, thus reducing overall cost associated with devices utilizing time-of-flight sensors while providing an efficient performance.
- FIG. 2 is a chart of signal charges representing the operation of a time-of-flight sensor in accordance with various approaches
- FIG. 3 illustrates a simplified view of a depth map frame creation process in accordance with at least one approach
- FIG. 4 illustrates a time-of-flight sensor in accordance with various approaches
- FIG. 5 illustrates charge generation by the pixel array of the time-of-flight sensor in accordance with at least one approach
- FIG. 6 illustrates a transfer of a first set of charges in the time-of-flight sensor in accordance with at least one approach
- FIG. 7 illustrates a transfer of a first and a second set of charges in the time-of-flight sensor in accordance with at least one approach
- FIG. 8 illustrates a transfer of a first, second, and third set of charges in the time-of-flight sensor in accordance with at least one approach
- FIG. 9 illustrates a transfer of a first, second, third, and fourth set of charges in the time-of-flight sensor in accordance with at least one approach
- FIG. 10 illustrates an output of the charges from the time-of-flight sensor in accordance with at least one approach
- FIG. 11 illustrates an alternate output of the changes from the time-of-flight sensor in accordance with at least one approach.
- FIG. 1 illustrates a block diagram of an example contextual apparatus 100 for a time-of-flight sensor 102 in accordance with various approaches.
- the example contextual apparatus 100 is provided merely to aid the reader in understanding the use and purpose of the various embodiments of the time-of-flight sensor 102 and is not meant to be limiting in any sense.
- the illustrated contextual apparatus 100 may be a dedicated time-of-flight apparatus or other non-time-of-flight-dedicated apparatus that utilizes such features (such as measurement devices, a mobile phone, a PDA, a tablet, or the like).
- the specific details of the contextual apparatus 100 are not required to form an understanding of the specifics of the disclosed time-of-flight sensor 102 .
- the example contextual apparatus 100 includes the time-of-flight sensor 102 operatively coupled to at least one processing device 104 . At least one illuminator 106 is operatively coupled to the processing device 104 as well.
- the illuminator 106 and the time-of-flight sensor 102 may be equipped with or coupled to various optical elements 108 and 110 , such as lenses, prisms, or the like.
- the illuminator 106 may be equipped with a lens 108 to effect dispersion or focusing of emitted light 112 , or simply a clear or translucent membrane to protect the illuminator 106 .
- the time-of-flight sensor 102 may be configured to receive incoming light 114 through one or more lenses 110 to focus the image upon the sensor 102 , in much the same manner as an image sensor of a conventional digital camera.
- the example contextual time-of-flight apparatus 100 is configured to generate depth map frames 312 (see FIG. 3 ) as are commensurate with time-of-flight sensing and/or imagery or other uses.
- Depth map frames 312 can be generated by many means, but time-of-flight apparatuses 100 generally operate on the principle that a distance to a surface of a detected object 116 can be calculated if the speed of an item, substance, or phenomenon (i.e., photons or light) is known and the time required to travel that distance is known.
- the speed of light is known (the constant c, approximately 300,000,000 meters per second), and the time that light takes to travel from the illuminator 106 to the detected object 116 and back to the sensor 102 can be measured.
- the illuminator 106 and sensor 102 are not necessarily required to be co-located (or nearly co-located as compared to the distance of the object), if they are co-located, then the distance is simply half of the total distance measured for the round trip.
- IR light is often utilized as it is invisible to humans, can be relatively inexpensive to produce, and behaves similarly to visible light. Accordingly, IR light is assumed to be used in this disclosure as a non-limiting example for the sake of description and widespread applicability.
- the illuminator 106 may comprise various known or unknown light emitters, such as a light emitting diode (LED), lasers, precision incandescent lights, shuttered light sources, and other precision light sources.
- LED light emitting diode
- lasers precision incandescent lights
- shuttered light sources and other precision light sources.
- an LED is ideally suited for use as the illuminator 106 as its light emission start and stop times can be controlled with high precision to produce well defined pulses of emitted light 112 , which can aid in the precision of the measurements. Further, LEDs provide an efficient solution as the illuminator output power is high compared to the input electrical power.
- a depth z can be calculated or estimated for each pixel of a time-of-flight sensor 102 based on the following equation:
- aliasing issues can be alleviated or accommodated for by utilizing techniques involving varying the frequency or period of the emitted pulses of light 112 so that the apparatus 100 can determine which outgoing pulse it detected (i.e., the one it is currently emitting or just sent out, or the previous pulse, or a pulse even prior to that).
- the weaker the power of the reflected pulse of light 114 will be, to a point where a given apparatus 100 may not be able to detect the reflected pulse of light 114 due to being lost in the noise or being simply too weak.
- FIG. 2 illustrates a signal-based representation of an example operation of a time-of-flight apparatus 100 .
- the top signal represents an outgoing emitted pulse of light 112 .
- the outgoing emitted pulse of light 112 is emitted with a frequency of f, which, when depicted in the time-based scale of FIG. 2 , is represented by its pulse width period T, which is simply 1/f.
- the emitted pulse of light 112 may be re-transmitted after a time t r .
- the second signal represents the reflected pulse of light 114 received at a pixel of the time-of-flight sensor 102 . As is shown in FIG.
- the reflected pulse of light 114 is delayed as compared to the emitted pulse of light 112 by an angular phase ⁇ of the period T of the emitted pulse of light 112 .
- FIG. 2 shows the phase ⁇ as approximately 225° (i.e., five-eighths), meaning the beginning of the reflected pulse of light 114 first hit the sensor 102 about five-eighths through the period T during which that same pulse of light 112 was being emitted.
- One method of determining the phase ⁇ of the reflected pulse of light 114 is to utilize two photosites in each pixel, called photosite X and photosite Y.
- the photosites each measure the light received at the pixel during different time segments to generate a charge pair comprising an X charge and a Y charge.
- photosite X may be active during the same period T during which the emitted pulse of light 112 is emitted.
- the time period T is chosen to be 1 ⁇ 2 that of the entire period to maximize the time for light emission and thus lower the average power required for a given total flux.
- photosite X may discontinue measurement of received light 114 while at the same time photosite Y may begin measurement of received light 114 .
- Photosite Y may make a measurement for the same duration of time period T as did photosite X.
- photons from the reflected pulse of light 114 hits each photosite they will each build and store a charge 202 during the time which each photosite is enabled to measure the reflected pulse of light 114 .
- the charge 202 generated at photosite X is represented by the line x 0°
- the charge 202 generated at photosite Y is represented by the line y 0° .
- photosite X will build a charge X as more photons continue to be collected in photosite X. This is shown by the positive slope of the line x 0° during the time the reflected pulse of light 114 encounters photosite X while it is active.
- photosite Y becomes active, it will generate and store a charge Y as is shown by the positive slope of the line y 0° during the time when the reflected pulse of light 114 encounters photosite Y while it is active.
- the charge pair x 0° and y 0° can be eventually collected and converted into digital signals to be processed by the processing device 104 to calculate the phase ⁇ of the reflected pulse of light 114 .
- a single photosite can be utilized that employs a fast switching gate arrangement that will store the accumulated charge 202 into two separate charge wells or charge repositories during their respective times, one storing the X charge representing the time during which the representative “photosite X” is active and the other storing the Y charge representing the time during which the representative “photosite Y” is active.
- phase ⁇ of the reflected pulses of light may be beneficial in certain examples to take measurements of the phase ⁇ of the reflected pulses of light across a plurality of measurement period phases ⁇ , wherein the angular phase ⁇ of the active times of photosites X and Y are shifted by various amounts.
- the four depicted phase measurements periods represent four different times measuring four different reflected pulse of light 114 , but are depicted in relation to a single emitted and reflected pulse of light 114 for simplicity.
- the phased measurement periods will provide different values of the charges X and Y 202 for each phase ⁇ measurement period (i.e., x 0° and y 0° , x 90° and y 90° , x 180° and y 180° , x 270° and y 270° ) to provide additional information with which to make a more accurate reflected pulse of light 114 phase ⁇ estimate.
- Phase ⁇ can then be calculated according to the following formula:
- phase ⁇ estimate can then be inserted into formula [1] above to calculate distance z measured at the subject pixel.
- a depth map frame 312 can be formed.
- the depth map frame 312 represents distances z for each pixel, resulting in a three-dimensional (3D) reading or image.
- FIG. 3 illustrates a simplified view of the depth map frame 312 creation process in accordance with at least one embodiment.
- each pixel 310 of the depth map frame 312 has a corresponding calculated z value represented by “D1”, “D2”, and so forth.
- the depth map frame 312 can then be used for many purposes, one of which includes combining it with a standard two-dimensional image taken simultaneously, or near simultaneously, to provide depth to each 2D pixel resulting in a visible 3D image.
- each pixel 310 of each image frame 302 , 304 , 306 , 308 comprising two data points (the X and Y charges 202 ), though other configurations involving the use of two, three, or more images frames may exist.
- other approaches may incorporate a single data point per pixel 310 , or more than two data points per pixel 310 .
- the remainder of the description assumes an approach with four images frames 302 , 304 , 306 , 308 and two data points (X and Y) per pixel 310 . This results in 8 data points per pixel 310 to generate the depth map frame 312 .
- a processing device 104 cannot perform necessary calculations on the acquired data for a pixel 310 until all the data points have been captured for that pixel 310 (i.e., X and Y for each image frame).
- Present solutions that do not utilize the disclosed on-chip analog memory require the data from each image frame 302 , 304 , 306 , 308 to be offloaded from the time-of-flight sensor 102 , converted to digital signals, and stored into a memory until such time as all the data points have been captured, at which time the processing device 104 can perform the calculations.
- the additional memory to temporarily store the multiple sets of image data can increase the cost of devices utilizing time-of-flight sensors 102 , increase processing power, increase processing or information transportation time, or may otherwise be impracticable.
- a new time-of-flight sensor 102 is disclosed herein that omits the requirement for use of the additional memory.
- FIG. 4 illustrates a new time-of-flight sensor 102 in accordance with various embodiments.
- the time-of-flight sensor 102 includes a time-of-flight pixel array 402 including a plurality of time-of-flight sensor pixels 310 configured to generate at least one charge 202 (i.e., the X charge, the Y charge, or both).
- the pixel array 402 may be of many different known or unknown types of pixel sensor arrays, including but not limited to CMOS image sensors and Charge-Coupled Device (CCD) image sensors.
- CCD Charge-Coupled Device
- the time-of-flight sensor 102 also has a first frame storage circuit 404 and a second frame storage circuit 406 .
- the time-of-flight sensor 102 may also comprise a third frame storage circuit 408 or even a fourth frame storage circuit 410 .
- Each frame storage circuit has a set of charge storage devices 412 , 414 , 416 , 418 .
- the charge storage devices 412 , 414 , 416 , 418 may be of any known or presently unknown charge storage devices, including but not limited to capacitors, n-doped or p-doped silicon, potential wells, charge repositories, gates, and the like.
- the number of charge storage devices 412 , 414 , 416 , 418 in each frame storage circuit 404 , 406 , 408 , 410 will correspond directly to the number of pixels 310 of the pixel array 402 , or twice the number of pixels 310 when the two changes 202 (X and Y) are both captured for a single image frame 302 , 304 , 306 , 308 , so as to store the charge data 202 generated during each capture of each image frame 302 , 304 , 306 , 308 .
- Each frame storage circuit 404 , 406 , 408 , 410 also includes an output port 420 , 422 , 424 , 426 configured to output the charges 202 stored in each frame storage circuit 404 , 406 , 408 , 410 for use by a processing device 104 or other device at such time as is appropriate according to the various approaches described herein.
- the set of charge storage devices 412 of the first frame storage circuit 404 are operatively coupled to the plurality of time-of-flight pixels 310 of the pixel array 402 .
- the set of charge storage devices 414 of the second frame storage circuit 406 are operatively coupled to the first set of charge storage devices 412 of the first frame storage circuit 404 .
- the set of charge storage devices 416 of the optional third frame storage circuit 408 are operatively coupled to the set of charge storage devices 414 of the second frame storage circuit 406 .
- the set of charge storage devices 418 of the optional fourth frame storage circuit 410 are operatively coupled to the set of charge storage devices 416 of the third frame storage circuit 408 .
- the sets of charge storage devices 412 , 414 , 416 , 418 of each frame storage circuit 404 , 406 , 408 , 410 may be operatively coupled to the plurality of time-of-flight pixels 310 of the pixel array 402 , possibly through one or more gates, to direct the charges 202 from the pixel array 402 to the various frame storage circuits directly 404 , 406 , 408 , 410 .
- the time-of-flight pixel array 402 will generate a first set of charges 502 that constitute the first image frame 302 as described above (possibly including the X and Y charge pair for each pixel 310 ).
- the first set of charges 502 (or a set of charges corresponding to the first set of charges 502 ) will be shifted into the first frame storage circuit 404 for storage in the first frame storage circuit's 404 set of charge storage devices 412 .
- the pixel array 402 can capture the second image frame 304 and generate a second set of charges 702 corresponding to the second image frame.
- the first set of charges 502 is shifted out of the first frame storage circuit 404 and into the second frame storage circuit 406 for storage in the second frame storage circuit's 406 set of charge storage devices 414 .
- the second frame storage circuit 406 may receive the first set of charges 502 while the second set of charges 702 is being generated during the second image frame 304 capture.
- the second set of charges 702 is shifted into the first frame storage circuit 404 for storage in its set of charge storage devices 412 .
- the output ports 420 , 422 of the first frame storage circuit 404 and the second frame storage circuit 406 output the contents of the first and second sets of charge storage devices 412 , 414 (the first set of charges 502 being stored in the set of charge storage devices 414 of the second frame storage circuit 406 , the second set of charges 702 being stored in the set of charge storage devices 412 of the first frame storage circuit 404 ) parallel to one another for use by the processing device 104 .
- the above described process will continue at the third time to capture, transfer, and store additional charge data.
- the third frame storage circuit 408 receives the first set of charges 502 from the second frame storage circuit 406 for storage in the third frame storage circuit's 408 set of charge storage devices 416 .
- the second set of charges 702 is transferred to the second frame storage circuit 406 from the first frame storage circuit 404
- the third set of charges 802 is transferred to the first frame storage circuit 404 from the pixel array 402 for storage in the first set of storage devices 412 .
- the above described process will continue at a fourth time, which is after the pixel array 402 generates a fourth set of charges 902 corresponding to a fourth image frame 308 .
- the fourth frame storage circuit 410 receives the first set of charges 502 from the third frame storage circuit 408 for storage in the fourth frame storage circuit's 410 set of charge storage devices 418 .
- the second set of charges 702 is transferred to the third frame storage circuit 408 from the second frame storage circuit 406
- the third set of charges 802 is transferred to the second frame storage circuit 406 from the first frame storage circuit 404
- the fourth set of charges 902 is transferred to the first frame storage circuit 404 from the pixel array 402 for storage in the first set of storage devices 412 .
- the entire set of data comprising the four image frames 302 , 304 , 306 , 308 will be stored in the four frame storage circuits 404 , 406 , 408 , 410 . It is not until this point (all image frames 302 , 304 , 306 , 308 being captured) that the processing device 104 can make the depth map frame calculations 312 . As is illustrated in FIG. 10 , with all of the charge data 502 , 702 , 802 , 902 captured and stored, the data is then ready for processing by the processing device 104 to determine one or more depth map frames 312 .
- each frame storage circuit 404 , 406 , 408 , 410 can output 1002 its charge data 502 , 702 , 802 , 902 through its respective output port 420 , 422 , 424 , 426 in parallel to one another for processing by at least one processing device 104 .
- the charge packets for each pixel 310 i.e., X and Y charges 202
- ADC analog-to-digital converter
- each frame storage circuit 404 , 406 , 408 , 410 can output 1002 the charge data for the same pixel 310 (i.e., charge X and Y for a specific pixel 310 ) in parallel to one another.
- the processing device 104 can receive all charge data for that specific pixel 310 and can immediately calculate the depth data for that specific pixel 310 and store the calculated depth data into the depth map frame 312 .
- the charge data has been collected and processed, it can be discarded and the next set of charge data for the next pixel 310 can be collected and processed.
- the described approaches eliminate the need for additional memory to store the entirety of the plurality of image frames 302 , 304 , 306 , 308 prior to processing because the charge data 502 , 702 , 802 , 902 is stored on the time-of-flight sensor 102 as analog charge packets until such time as all charge data 502 , 702 , 802 , 902 for all image frames 302 , 304 , 306 , 308 has been collected and can be processed.
- a processing device 104 may utilize a buffer to store at least some of the charge data 502 , 702 , 802 , 902 for various pixels 310 during processing so that the process of converting the charge data 502 , 702 , 802 , 902 to digital signals, receiving the charge data 502 , 702 , 802 , 902 , and processing the charge data 502 , 702 , 802 , 902 can occur in parallel for different pixels, like an assembly line, rather than exclusively in sequence (i.e., rather than waiting for one pixel to be completely processed before converting the charged data for the next pixel), so as to ensure the data 502 , 702 , 802 , 902 is removed and processed as quickly as possible.
- this buffer may be sized to store far less data than makes up all the pixels 310 of each image frame 302 , 304 , 306 , 308 so that the benefits of the various disclosed approaches may still be realized.
- the pixel array 402 can begin to capture the first image frame 1004 of the next set of data for the next depth map frame (i.e., a fifth set of charges 1004 ).
- the process can repeat itself and the newly captured first image frame 1004 can be transferred to the first frame storage circuit 404 for storage therein without waiting to generate the fifth set of charges 1004 .
- the time-of-flight sensor 102 may be provided with one less frame storage circuit than the number of image frames to be captured. This alternative is shown in FIG. 11 , where the optional fourth frame storage circuit 410 has been omitted.
- the operation of outputting 1002 the charge data 502 , 702 , 802 , 902 to the processing device 104 can begin immediately.
- Each of the frame storage circuits 404 , 406 , 408 outputs 1002 the charge data 502 , 702 , 802 through their respective output port 420 , 422 , 424 parallel to one another as previously described (with the first frame storage circuit 404 outputting 1002 the third set of charges 802 , the second frame storage circuit 406 outputting 1002 the second set of charges 702 , and the third frame storage circuit 408 outputting 1002 the first set of charges 502 ).
- the pixel array 402 outputs 1002 the fourth set of charges 902 to the processing device in parallel to the frame storage circuits 404 , 406 , 408 instead of transferring the fourth set of charges 902 to the first frame storage circuit 404 .
- This alternative approach eliminates at least one extra transfer step to transfer all of all the charges 502 , 702 , 802 , 902 to the next frame storage circuit, as well as eliminates an entire frame storage circuit (i.e., the fourth frame storage circuit 410 ), thus saving transfer time and silicon real estate.
- the approaches described herein can still be utilized to at least reduce the number and/or size of memory required to store all the image frames as at least some of the image frames can be stored on the time-of-flight sensor 102 until all data is collected to be processed.
- the various approaches described herein provide a solution that circumvents the need for additional memory storage space when collecting multiple time-of-flight image frames (i.e., image frames 302 , 304 , 306 , 308 ) required for calculating a depth map frame 312 .
- image frames 302 , 304 , 306 , 308 the required memory storage in conventional systems can scale rapidly. For example, 12-16 bits of digital data per pixel 310 is typically required to precisely capture the charge data 202 .
Abstract
Description
where c is the speed of light, Φ is the angular phase of the detected pulse of
This more accurate phase Φ estimate can then be inserted into formula [1] above to calculate distance z measured at the subject pixel. When this occurs for an entire set of pixels for a time-of-
Claims (15)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/792,756 US8908063B2 (en) | 2013-03-11 | 2013-03-11 | Method and apparatus for a time-of-flight sensor with charge storage |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/792,756 US8908063B2 (en) | 2013-03-11 | 2013-03-11 | Method and apparatus for a time-of-flight sensor with charge storage |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140253758A1 US20140253758A1 (en) | 2014-09-11 |
US8908063B2 true US8908063B2 (en) | 2014-12-09 |
Family
ID=51487406
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/792,756 Active US8908063B2 (en) | 2013-03-11 | 2013-03-11 | Method and apparatus for a time-of-flight sensor with charge storage |
Country Status (1)
Country | Link |
---|---|
US (1) | US8908063B2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11543501B2 (en) * | 2016-10-03 | 2023-01-03 | Xenomatix Nv | Method for subtracting background light from an exposure value of a pixel in an imaging array, and pixel for use in same |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9581696B2 (en) * | 2014-12-22 | 2017-02-28 | Google Inc. | Image sensor and light source driver integrated in a same semiconductor package |
US9615013B2 (en) * | 2014-12-22 | 2017-04-04 | Google Inc. | Image sensor having multiple output ports |
EP3159711A1 (en) | 2015-10-23 | 2017-04-26 | Xenomatix NV | System and method for determining a distance to an object |
EP3301480A1 (en) | 2016-10-03 | 2018-04-04 | Xenomatix NV | System and method for determining a distance to an object |
EP3301477A1 (en) | 2016-10-03 | 2018-04-04 | Xenomatix NV | System for determining a distance to an object |
EP3301478A1 (en) * | 2016-10-03 | 2018-04-04 | Xenomatix NV | System for determining a distance to an object |
EP3343246A1 (en) | 2016-12-30 | 2018-07-04 | Xenomatix NV | System for characterizing surroundings of a vehicle |
EP3392674A1 (en) | 2017-04-23 | 2018-10-24 | Xenomatix NV | A pixel structure |
EP3625589B1 (en) | 2017-12-15 | 2020-11-18 | Xenomatix NV | System and method for determining a distance to an object |
WO2021142091A1 (en) * | 2020-01-09 | 2021-07-15 | Sense Photonics, Inc. | Pipelined histogram pixel |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080048100A1 (en) * | 2005-04-19 | 2008-02-28 | Canesta, Inc. | Method and system to differentially enhance sensor dynamic range using enhanced common mode reset |
US20090167923A1 (en) * | 2007-12-27 | 2009-07-02 | Ati Technologies Ulc | Method and apparatus with depth map generation |
US20100053405A1 (en) * | 2008-08-28 | 2010-03-04 | Mesa Imaging Ag | Demodulation Pixel with Daisy Chain Charge Storage Sites and Method of Operation Therefor |
US20100231774A1 (en) * | 2009-03-12 | 2010-09-16 | Canon Kabushiki Kaisha | Solid-state imaging device, driving method thereof, and imaging system |
USRE42292E1 (en) * | 1998-08-19 | 2011-04-12 | Round Rock Research, Llc | Pinned photodiode photodetector with common pixel transistors and binning capability |
US20110273561A1 (en) * | 2009-11-09 | 2011-11-10 | Mesa Imaging Ag | Multistage Demodulation Pixel and Method |
US20130070132A1 (en) * | 2011-09-20 | 2013-03-21 | Mesa Imaging Ag | Time of Flight Sensor with Subframe Compression and Method |
US20130076993A1 (en) * | 2011-09-28 | 2013-03-28 | Stephen Wyllie | Moving display system |
US8654219B2 (en) * | 2011-04-26 | 2014-02-18 | Lg Electronics Inc. | Method and apparatus for restoring dead pixel using light intensity map in a time-of-flight camera |
-
2013
- 2013-03-11 US US13/792,756 patent/US8908063B2/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USRE42292E1 (en) * | 1998-08-19 | 2011-04-12 | Round Rock Research, Llc | Pinned photodiode photodetector with common pixel transistors and binning capability |
US20080048100A1 (en) * | 2005-04-19 | 2008-02-28 | Canesta, Inc. | Method and system to differentially enhance sensor dynamic range using enhanced common mode reset |
US20090167923A1 (en) * | 2007-12-27 | 2009-07-02 | Ati Technologies Ulc | Method and apparatus with depth map generation |
US20100053405A1 (en) * | 2008-08-28 | 2010-03-04 | Mesa Imaging Ag | Demodulation Pixel with Daisy Chain Charge Storage Sites and Method of Operation Therefor |
US20100231774A1 (en) * | 2009-03-12 | 2010-09-16 | Canon Kabushiki Kaisha | Solid-state imaging device, driving method thereof, and imaging system |
US20110273561A1 (en) * | 2009-11-09 | 2011-11-10 | Mesa Imaging Ag | Multistage Demodulation Pixel and Method |
US8654219B2 (en) * | 2011-04-26 | 2014-02-18 | Lg Electronics Inc. | Method and apparatus for restoring dead pixel using light intensity map in a time-of-flight camera |
US20130070132A1 (en) * | 2011-09-20 | 2013-03-21 | Mesa Imaging Ag | Time of Flight Sensor with Subframe Compression and Method |
US20130076993A1 (en) * | 2011-09-28 | 2013-03-28 | Stephen Wyllie | Moving display system |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11543501B2 (en) * | 2016-10-03 | 2023-01-03 | Xenomatix Nv | Method for subtracting background light from an exposure value of a pixel in an imaging array, and pixel for use in same |
Also Published As
Publication number | Publication date |
---|---|
US20140253758A1 (en) | 2014-09-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8908063B2 (en) | Method and apparatus for a time-of-flight sensor with charge storage | |
US8642938B2 (en) | Shared time of flight pixel | |
TWI544232B (en) | Calibration circuitry and method for a time of flight imaging system | |
CN107300705B (en) | Laser radar ranging system and method based on carrier modulation | |
US10291895B2 (en) | Time of flight photosensor | |
US9784822B2 (en) | Time of flight sensor binning | |
US9313476B2 (en) | Precharged latched pixel cell for a time of flight 3D image sensor | |
CN107247269B (en) | Detection device, pixel unit and array for collecting and processing laser signals | |
US11221400B2 (en) | Dual mode stacked photomultipliers suitable for use in long range time of flight applications | |
US9621868B2 (en) | Depth sensor, image capture method, and image processing system using depth sensor | |
US8829408B2 (en) | Sensor pixel array and separated array of storage and accumulation with parallel acquisition and readout wherein each pixel includes storage sites and readout nodes | |
JP7100049B2 (en) | Systric processor system for optical ranging system | |
CN108282626B (en) | TOF image sensor pixel structure and TOF imaging system | |
US11181419B2 (en) | Photon sensing with threshold detection using capacitor-based comparator | |
US20130176550A1 (en) | Image sensor, image sensing method, and image photographing apparatus including the image sensor | |
WO2018180660A1 (en) | Solid-state imaging device | |
US9681083B2 (en) | Method and system to detect a light-emitting diode | |
US10645319B2 (en) | Pixel circuit and image sensing system | |
US20220011437A1 (en) | Distance measuring device, distance measuring system, distance measuring method, and non-transitory storage medium | |
CN111726548B (en) | Image sensor pixel and image sensor | |
US20230134910A1 (en) | Readout architectures for motion blur reduction in indirect time-of-flight sensors | |
CN112437238B (en) | Phase-locked pixel structure of low kTC noise three-dimensional image sensor | |
CN107454316B (en) | Motion detection method and device | |
Sun et al. | A Build-in Pseudo-Event Driven Readout Circuit for 3D dToF Image Sensor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: TEXAS INSTRUMENTS INCORPORATED, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:METZ, WERNER ADAM;REEL/FRAME:030312/0537 Effective date: 20130420 |
|
AS | Assignment |
Owner name: TEXAS INSTRUMENTS INCORPORATED, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:METZ, WERNER ADAM;REEL/FRAME:031290/0594 Effective date: 20130310 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |