EP3940416A1 - Readout architecture for fmcw lidar - Google Patents
Readout architecture for fmcw lidar Download PDFInfo
- Publication number
- EP3940416A1 EP3940416A1 EP21173088.2A EP21173088A EP3940416A1 EP 3940416 A1 EP3940416 A1 EP 3940416A1 EP 21173088 A EP21173088 A EP 21173088A EP 3940416 A1 EP3940416 A1 EP 3940416A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- signal
- primary
- depth information
- receive
- counter
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000003287 optical effect Effects 0.000 claims abstract description 34
- 238000000034 method Methods 0.000 claims description 35
- 238000001514 detection method Methods 0.000 claims description 13
- 238000010586 diagram Methods 0.000 description 7
- 230000008569 process Effects 0.000 description 6
- 230000004044 response Effects 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 3
- 230000035559 beat frequency Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 238000004590 computer program Methods 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000003708 edge detection Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 230000003321 amplification Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000003750 conditioning effect Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/483—Details of pulse systems
- G01S7/486—Receivers
- G01S7/4861—Circuits for detection, sampling, integration or read-out
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/491—Details of non-pulse systems
- G01S7/4912—Receivers
- G01S7/4913—Circuits for detection, sampling, integration or read-out
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
- G01S17/32—Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated
- G01S17/34—Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated using transmission of continuous, frequency-modulated waves while heterodyning the received signal, or a signal derived therefrom, with a locally-generated signal related to the contemporaneously transmitted signal
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/50—Systems of measurement based on relative movement of target
- G01S17/58—Velocity or trajectory determination systems; Sense-of-movement determination systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
- G01S7/4816—Constructional features, e.g. arrangements of optical elements of receivers alone
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
- G01S7/4817—Constructional features, e.g. arrangements of optical elements relating to scanning
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/491—Details of non-pulse systems
- G01S7/4912—Receivers
- G01S7/4915—Time delay measurement, e.g. operational details for pixel components; Phase measurement
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/931—Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
Definitions
- One or more aspects of embodiments of the present disclosure relate generally to LiDAR systems.
- LiDAR is a portmanteau of light and radar and may refer to "light detection and ranging” or “laser imaging, detection, and ranging,” and refers to a method for measuring distances and/or velocity of a target or object by illuminating the target with light (e.g., laser light) from one or more sources, and by measuring a reflection of the light off of the target with one or more sensors.
- light e.g., laser light
- differences in return times e.g., an amount of time between emission of the light and detection of the reflection of portions of the light
- differences in wavelengths e.g., a difference between a wavelength of emitted laser light and a wavelength of the detected reflected laser light
- a distance of the target e.g., a distance from the light source or the sensor to the target
- Embodiments described herein provide improvements LiDAR technology, including improvements to a readout architecture, which includes a readout integrated chip, that is used in frequency-modulated continuous wave (FMCW) LiDAR systems.
- a readout architecture which includes a readout integrated chip, that is used in frequency-modulated continuous wave (FMCW) LiDAR systems.
- FMCW frequency-modulated continuous wave
- a readout integrated circuit of a FMCW LiDAR system including a balanced PIN photo diode configured to receive an optical signal, and convert the optical signal into a current signal, an optical mixer configured to receive the current signal from the balanced PIN photo diode, and convert the current signal into a mixed signal of high and low frequencies, a transimpedance amplifier configured to receive the converted current signal from the optical mixer; and convert the converted current signal into a voltage signal including a low frequency sinusoidal waveform after filtered high frequency, a comparator configured to receive the voltage signal from the transimpedance amplifier, and generate a pulse signal from the voltage signal, a primary counter configured to receive a primary signal corresponding to the pulse signal generated by the comparator, and count a number of pulses of the primary signal during a certain time period to generate primary counter data, a reference counter configured to receive a reference signal, and count a number of pulses of the reference signal during a time period to generate reference counter data, and an
- the readout integrated circuit may further include a primary phase-locked loop (PLL) configured to receive the pulse signal from the comparator, and multiply a frequency of the pulse signal to generate the primary signal, and a reference phase-locked loop configured to amplify a reference PLL signal to generate the reference signal having a reference frequency corresponding to a reference distance.
- PLL primary phase-locked loop
- the comparator may be further configured to compare the primary signal and the reference signal.
- the arithmetic logic unit may be configured to calculate the depth information by multiplying the reference distance by a ratio of the reference counter data to the primary counter data.
- the readout integrated circuit may further include a frame memory configured to receive the primary counter data, receive the depth information from the arithmetic logic unit, and store the depth information.
- the arithmetic logic unit may be further configured to calculate velocity information by receiving, from the frame memory, the depth information, the depth information including a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- the comparator may be further configured to compare the primary counter data and the reference counter data.
- a method to determine a distance or a velocity of an object including receiving a waveform signal corresponding to a target, generating a pulse signal from the received waveform signal, resetting a primary counter and a reference counter, beginning a primary count with the primary counter and a reference count with the reference counter after a first time period since the resetting of the primary counter and the reference counter has elapsed, continuing the primary count and the reference count for a second time period, retaining a last primary value of the primary counter and a last reference value of the reference counter, calculating a ratio of the last primary value to the last reference value, and multiplying a reference distance by the ratio.
- the method may further include receiving the waveform signal from an optical mixer output.
- the first time period may correspond to a detection range.
- the second time period may correspond to a flexible programmed time period.
- the method may further include generating a beam toward the target to cause the target to scatter the beam, and receiving the scattered beam as the waveform signal corresponding to the target.
- the method may further include calculating velocity information by receiving depth information including a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- the method may further include calculating the depth information by multiplying the reference distance by a ratio of the primary count to the reference count.
- a non-transitory computer readable medium (exemplarily implemented on a readout integrated circuit of a FMCW LiDAR system), the non-transitory computer readable medium having computer code that, when executed on a processor, implements a method of determining a distance or a velocity of an object, the method including receiving a waveform signal corresponding to a target, generating a pulse signal from the received waveform signal, resetting a primary counter and a reference counter, beginning a primary count with the primary counter and a reference count with the reference counter after a first time period since the resetting has elapsed, continuing the primary count and the reference count for a second time period, retaining a last primary value of the primary counter and a last reference value of the reference counter, calculating a ratio of the last primary value to the last reference value, and multiplying a reference distance by the ratio.
- the computer code when executed on the processor, may further implement the method of determining the distance or the velocity of the object by receiving the waveform signal from an optical mixer output.
- the first time period may correspond to a detection range, wherein the second time period corresponds to a flexible programmed time period.
- the computer code when executed on the processor, may further implement the method of determining the distance or the velocity of the object by generating a beam toward the target to cause the target to scatter the beam, and receiving the scattered beam as the waveform signal corresponding to the target.
- the computer code when executed on the processor, may further implement the method of determining the distance or the velocity of the object by calculating velocity information by receiving depth information including a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- the computer code when executed on the processor, may further implement the method of determining the distance or the velocity of the object by calculating the depth information by multiplying the reference distance by a ratio of the primary count to the reference count.
- system and method of some embodiments of the present disclosure are able to calculate a range of a target while omitting circuit elements used to perform transforms to process signals in calculating depth/distance/range.
- first,” “second,” “third,” etc. may be used herein to describe various elements, components, regions, layers and/or sections, these elements, components, regions, layers and/or sections should not be limited by these terms. These terms are used to distinguish one element, component, region, layer or section from another element, component, region, layer or section. Thus, a first element, component, region, layer or section described below could be termed a second element, component, region, layer or section, without departing from the scope of the present disclosure.
- the term “substantially,” “about,” “approximately,” and similar terms are used as terms of approximation and not as terms of degree, and are intended to account for the inherent deviations in measured or calculated values that would be recognized by those of ordinary skill in the art.
- “About” or “approximately,” as used herein, is inclusive of the stated value and means within an acceptable range of deviation for the particular value as determined by one of ordinary skill in the art, considering the measurement in question and the error associated with measurement of the particular quantity (i.e., the limitations of the measurement system). For example, “about” may mean within one or more standard deviations, or within ⁇ 30%, 20%, 10%, 5% of the stated value. Further, the use of “may” when describing embodiments of the present disclosure refers to "one or more embodiments of the present disclosure.”
- a specific process order may be performed differently from the described order.
- two consecutively described processes may be performed substantially at the same time or performed in an order opposite to the described order.
- the electronic or electric devices and/or any other relevant devices or components according to embodiments of the present disclosure described herein may be implemented utilizing any suitable hardware, firmware (e.g. an application-specific integrated circuit), software, or a combination of software, firmware, and hardware.
- firmware e.g. an application-specific integrated circuit
- the various components of these devices may be formed on one integrated circuit (IC) chip or on separate IC chips.
- the various components of these devices may be implemented on a flexible printed circuit film, a tape carrier package (TCP), a printed circuit board (PCB), or formed on one substrate.
- the various components of these devices may be a process or thread, running on one or more processors, in one or more computing devices, executing computer program instructions and interacting with other system components for performing the various functionalities described herein.
- the computer program instructions are stored in a memory which may be implemented in a computing device using a standard memory device, such as, for example, a random access memory (RAM).
- the computer program instructions may also be stored in other non-transitory computer readable media such as, for example, a CD-ROM, flash drive, or the like.
- a LiDAR system may be used to determine depth (e.g., the distance of a target from the LiDAR system) and/or a velocity of a target with respect thereto.
- Some embodiments of the present disclosure provide improvements to LiDAR technology by omitting components that otherwise may be used in a LiDAR system, thereby improving the speed and accuracy of the LiDAR system, and potentially reducing manufacturing costs associated therewith.
- FIG. 1 shows a block diagram depicting a LiDAR system, according to some embodiments of the present disclosure.
- a LiDAR system 100 includes a light source 110, a mirror 115, a scanner 120, a receiver 140, and a controller 150.
- the LiDAR system 100 may be referred to as a laser ranging system, a laser radar system, or a laser detection and ranging system.
- the LiDAR system 100 may be configured to sense, identify, or determine distances to one or more targets 130 within a field of regard.
- the LiDAR system 100 may determine a distance to a target 130, where all or part of the target 130 is contained within a field of regard (FOR) of the LiDAR system 100.
- All or part of the target 130 being contained within a FOR of the LiDAR system 100 may refer to the FOR overlapping, encompassing, or enclosing at least a portion of the target 130.
- target 130 may include all or part of an object that is moving or stationary relative to LiDAR system 100.
- the light source 110 may correspond to a laser for producing an output beam 125 having an operating wavelength (e.g., a wavelength in the electromagnetic spectrum).
- the output beam 125 may be referred to as an optical signal, a laser beam, a light beam, an optical beam, an emitted beam, an emitted light, or simply a beam.
- the LiDAR system 100 is a frequency-modulated continuous-wave (FMCW) LiDAR system.
- the output beam 125 may be an FMCW laser, although it should be noted that the output beam 125 may be pulsed or otherwise modulated in other embodiments.
- the output beam 125 may be directed toward the target 130 that is a distance D from the LiDAR system 100. Before being directed to the target 130, the output beam 125 may pass through, or by, the mirror 115 (e.g., the mirror 115 may include a hole, slot, or aperture through which the output beam 125 passes).
- the output beam 125 may also pass through the scanner 120.
- the scanner 120 may be configured to scan the output beam 125 across an area of interest (e.g., the FOR) of the LiDAR system 100.
- the scanner 120 may include one or more scanning mirrors configured to pivot, rotate, oscillate, or move in an angular manner about one or more rotation axes. Accordingly, the output beam 125 may be reflected by a scanning mirror, and as the scanning mirror pivots or rotates, the reflected output beam 125 may be scanned in a corresponding angular manner.
- a scanning mirror may be configured to periodically pivot back and forth, thereby causing the output beam 125 to scan back and forth.
- the target 130 may scatter or otherwise reflect at least a portion of the light of the output beam 125, and some of the scattered or reflected light may be returned toward the LiDAR system 100 as one or more input beams 135 as a received optical signal.
- the input beam 135 may then return while passing through the scanner 120, and may be then reflected by the mirror 115 to be directed to the receiver 140.
- the LiDAR system 100 may include one or more optical components (e.g., lenses, mirrors, or filters) that are configured to reflect, focus, filter, shape, modify, steer, direct, collimate, or combine light produced by, or received by, the LiDAR system 100 (e.g., to steer or focus the output beam 125 or the input beam 135).
- the LiDAR system 100 may include one or more lenses to focus the input beam 135 onto a photodetector of the receiver 140.
- the mirror 115 may provide for the output beam 125 and the input beam 135 to be substantially coaxial so that the two beams travel along approximately the same optical path (e.g., to travel substantially parallel to one another and in opposite directions). Accordingly, the mirror 115 and other corresponding components may direct the input beam 135 toward the receiver 140.
- the receiver 140 may be referred to as a photoreceiver, optical receiver, optical sensor, detector, photodetector, or optical detector.
- the receiver 140 may include one or more avalanche photodiodes (APDs), single-photon avalanche diodes (SPADs), PN photodiodes (PD), and/or PIN photodiodes (e.g., a balanced PIN PD 340, as shown in FIG. 3 ).
- the receiver 140 may include electronic circuitry that performs signal modification and analysis.
- the receiver 140 may include a transimpedance amplifier (TIA) (e.g., the TIA 265, 365 described below with respect to FIGS.
- TIA transimpedance amplifier
- the voltage signal may be used to determine one or more optical characteristics of the input beam 135, and may be used to produce a digital output signal/electrical signal 145 to be sent to the controller 150 for processing or analysis (e.g., to determine a time-of-flight value corresponding to a received optical pulse).
- the receiver 140 may receive and detect photons from the input beam 135.
- the receiver 140 may then produce and output an electrical current or voltage pulse as an electrical signal 145 that represents the input beam 135, and may send the electrical signal 145 to the controller 150 (e.g., a controller that may be part of, or that may include, the ROIC 250, 350 described below with respect to FIGS. 2 and 3 ).
- the controller 150 e.g., a controller that may be part of, or that may include, the ROIC 250, 350 described below with respect to FIGS. 2 and 3 ).
- the controller 150 may include a processor, computing system, or other suitable circuitry, and may be configured to analyze one or more characteristics of the electrical signal 145 from the receiver 140 to determine one or more characteristics of the target 130, such as a distance of, or velocity of, the target with respect to the LiDAR system 100. This may be done, for example, by analyzing a time of flight of the light that is common to the transmitted output beam 125 and the received input beam 135.
- a time of flight T represents a round-trip time of flight for an emitted beam or pulse of light to travel from the LiDAR system 100 to the target 130, and then from the target 130 back to the LiDAR system 100.
- the controller 150 may also be electrically or communicatively coupled to the light source 110, the mirror 115, and the scanner 120.
- the controller 150 may receive electrical trigger pulses or edges as a result of the light source 110, where each pulse or edge corresponds to the emission of an optical pulse by the light source 110.
- the controller 150 may also control the light source 110 to cause the light source 110 to produce and/or adjust the output beam 125.
- the controller 150 may determine a time-of-flight value for an optical pulse based on timing information associated with when the output beam 125 was emitted by the light source 110, and when the input beam 135 was detected by the receiver 140.
- the controller 150 may include circuitry that performs signal amplification, sampling, filtering, signal conditioning, analog-to-digital conversion, time-to-digital conversion, pulse detection, threshold detection, rising-edge detection, or falling-edge detection.
- one or more LiDAR systems 100 may be integrated into a vehicle, for example, as part of an advanced driver assistance system (ADAS) to assist a driver of the vehicle in operating the vehicle, or as part of an autonomous vehicle driving system.
- ADAS advanced driver assistance system
- the LiDAR system 100 may be part of an ADAS that provides information or feedback to a driver (e.g., to alert the driver to potential problems or hazards), or that automatically controls one or more functions of the vehicle (e.g., by controlling a braking system or a steering system) to avoid collisions or accidents.
- the LiDAR system 100 may provide information about the surrounding environment to a driving system of an autonomous vehicle.
- An autonomous-vehicle driving system may be configured to guide the autonomous vehicle through an environment surrounding the vehicle, and may include one or more computing systems that receive information from the LiDAR system 100 about the surrounding environment, that analyze the received information, and that provide control signals to the vehicle's driving systems and actuators (e.g., steering wheel, accelerator, brake, or turn signal).
- actuators e.g., steering wheel, accelerator, brake, or turn signal.
- FIG. 2 shows a block diagram depicting a readout architecture of a frequency-modulated continuous wave (FMCW) LiDAR system.
- FMCW frequency-modulated continuous wave
- the readout architecture 200 of a FMCW LiDAR system may include a balanced PIN PD 240 that is implemented as a receiver (e.g., the receiver 140 of FIG. 1 ) of the LiDAR system, and also may include a readout integrated circuit (ROIC) 250 that is implemented as a controller (e.g., the controller 150 of FIG. 1 ) of the LiDAR system.
- the ROIC 250 may be used in connection with a memory (PC) 260.
- PC memory
- the ROIC 250 may include, in relevant part, a TIA 265, an analog-to-digital controller (ADC) block 270 (e.g., an ultra-high speed high resolution ADC block 270), a Fast Fourier Transform (FFT) block 280 (e.g., a 2-dimensional FFT block), and a phase-locked loop (PLL) block 290.
- ADC analog-to-digital controller
- FFT Fast Fourier Transform
- PLL phase-locked loop
- the balanced PIN PD 240 may receive photons corresponding to an input beam (e.g., the input beam 135 of FIG. 1 ), and may then convert the photons to current as a signal 245 (e.g., the electrical signal 145 of FIG. 1 ). Then, the TIA 265 may receive the signal 245, and may transform the current of the signal 245 to a voltage signal (e.g., a TIA signal 267) to be sent to the ADC block 270.
- a voltage signal e.g., a TIA signal 267
- the ADC block 270 transforms the TIA signal 267 received from the TIA 265 into data flow (e.g., a data signal 285), such that each pixel, or voxel (e.g., a single sample, or data point, on a regularly spaced, three-dimensional grid), detected by the balanced PIN PD 240 may be converted to one or more corresponding data values.
- the FFT block 280 may run a transform (e.g., an FFT) on the data values of the data signal 285 received from the ADC block 270 to calculate depth information corresponding to the target reflecting or scattering the input beam 135.
- a transform e.g., an FFT
- depth values indicated in a depth signal 287 from the FFT block 280 are sent to the memory 260 via two respective lines to the memory 260 from the ADC block 270 and the FFT block 280, thereby enabling calculation of both depth and velocity of the target (e.g., the target 130 of FIG. 1 ) by using the data signal 285 and the depth signal 287.
- High-resolution and high-quality range/depth/distance information may be used by advanced driver assistance system (ADAS) and autonomous driving applications to enable the LiDAR system to "see” further (e.g., to have a longer target detection range), and to enable the LiDAR system to detect or recognize the target more quickly and easily, thereby reducing risk associated with the vehicle employing the LiDAR system either hitting or avoiding the target(s).
- ADAS advanced driver assistance system
- autonomous driving applications to enable the LiDAR system to "see” further (e.g., to have a longer target detection range), and to enable the LiDAR system to detect or recognize the target more quickly and easily, thereby reducing risk associated with the vehicle employing the LiDAR system either hitting or avoiding the target(s).
- a low-power-operation LiDAR system may be implemented with an ADAS or autonomous driving system.
- specifications of the ROIC 250 may consider receiver bandwidth (e.g., on the order of a few hundred MHz or more), ADC resolution and detection capabilities, low noise (e.g., a signal-to-noise ratio (SNR) of 12 dB or more for signal power, and/or a low-noise amplifier (LNA) used as the TIA), power consumption, performance of direct Fourier transforms (DFT) at chip level, and other RF processes.
- receiver bandwidth e.g., on the order of a few hundred MHz or more
- ADC resolution and detection capabilities e.g., low noise (e.g., a signal-to-noise ratio (SNR) of 12 dB or more for signal power, and/or a low-noise amplifier (LNA) used as the TIA
- SNR signal-to-noise ratio
- LNA low-noise amplifier
- DFT direct Fourier transforms
- a LiDAR system that is able to calculate target depth using a direct time-of-flight (dToF) method in conjunction with a single photo avalanche diode (SPAD) as the receiver may have difficulty due to the presence of various sources of ambient noise. Accordingly, implementation of the readout architecture 200 of FIG. 2 may present obstacles associated with the ADC block 270 and the FFT block 280.
- dToF direct time-of-flight
- SPAD single photo avalanche diode
- the 2D FFT block 280 may require a relatively large amount of space on the ROIC 250, and may consume a relatively large amount of power for several frames to enable detection of depth and velocity information, noting that the readout architecture 200 in the FMCW LiDAR system may concurrently determine depth and velocity by using the 2D FFT block 280. Further, there may be lag associated with operation of the FFT block 280 when determining depth and velocity information.
- the ADC block 270 may operate on the order of GHz (e.g., may have a sampling speed of a few GHz) to effectuate moderate-to-high resolution, and the readout architecture 200 may seek to separate signal noise and crosstalk associated with other blocks of the ROIC 250.
- This may be achieved by a high-speed, low-jitter phase-locked loop (PLL) block 290 and a highly sensitive bias circuit 292 to support high speed operation of the ADC block 270 (e.g., the PLL block 290 may be used as a clock for the ADC block 270 and the FFT block 280, and may operate on the order of GHz).
- PLL phase-locked loop
- the readout architecture 200 might not effectively integrate with an Optical PLL block (OPLL) due to a high level of sensitivity of each block.
- OPLL Optical PLL block
- FIG. 3 shows a block diagram depicting a readout architecture of a frequency-modulated continuous wave (FMCW) LiDAR system, according to some embodiments of the present disclosure.
- FMCW frequency-modulated continuous wave
- various issues associated with the readout architecture of a LiDAR system may be addressed or eliminated.
- ambient noise associated with the readout architecture 200 of FIG. 2 can be effectively eliminated by using a modulated signal (e.g., a FMCW signal) in accordance with an FMCW LiDAR system of embodiments of the present disclosure (e.g., the LiDAR system 100) by removing interference otherwise caused by other LiDAR systems and/or other light sources.
- a modulated signal e.g., a FMCW signal
- an FMCW LiDAR system of embodiments of the present disclosure e.g., the LiDAR system 100
- This may be achieved by providing a low power readout architecture 300 that omits the high-speed ADC block 270 and the 2D FFT block 280 of the readout architecture 200 of FIG. 2 , which are generally blocks that are relatively sensitive to noise, and by replacing the omitted blocks with a comparator 380 and one or more counters 395.
- the readout architecture 300 of a LiDAR system may include, in relevant part, a balanced PIN PD 340 as a receiver (e.g., the receiver 140 of FIG. 1 ), a ROIC 350 including a TIA 365, the comparator 380, a primary PLL block 390i, a reference PLL block 390r, a primary counter block 395i, a reference counter block 395r, a depth and/or depth/velocity block 375 for calculating depth and/or velocity, and a memory (PC) 360.
- the ROIC 350 of the present example omits overly sensitive blocks (e.g., the ADC block 270 and the FFT block 280), while still effectively calculating depth and velocity of the target.
- the LiDAR system of the present example may output an output beam (e.g., the output beam 125 of FIG. 1 ) by using a chirping signal that varies the frequency of the output beam 125.
- the target e.g., the target 130 of FIG. 1
- the target may then scatter or reflect the output beam 125 to thereby return an input beam (e.g., the input beam 135 of FIG. 1 ) to the LiDAR system (e.g., to the balanced PIN PD 340).
- the balanced PIN PD 340 may detect photons of the input beam 135 as a chirping signal. The balanced PIN PD 340 may then convert the photons into a current to deliver a signal (e.g., current as an electrical signal, or a digital output signal) 345 to the ROIC 350, which may operate, either by itself or in conjunction with the frame memory 360, as the controller 150 of FIG. 1 .
- the electrical signal 345 from the balanced PIN PD 340 may be thought of as having a beta frequency corresponding to the chirping frequency of the input beam 135.
- the electrical signal 345 may then be received directly by the TIA 365, although it should be noted that, in other embodiments, the readout architecture 300 may include optical mixers between the balanced PIN PD 340 and the TIA 365.
- the optical mixers may convert the frequency of the electrical signal 345 from the balanced PIN PD 340.
- the optical mixers may convert the electric signal 345 into a mixed signal of low frequency and high frequency.
- the TIA 365 is a current-to-voltage converter, and may be used with one or more operational amplifiers.
- the TIA 365 may be used to amplify the current output, which corresponds to the received electrical signal 345, to a usable voltage (e.g., to a suitable voltage to be used by the comparator 380).
- the TIA 365 may be used because the balanced PIN PD 340 may have a current response that is more linear than the voltage response (e.g., the current response of the balanced PIN PD 340 may be about 1% nonlinearity or better over a wide range of light input).
- the TIA 365 presents a low impedance to the balanced PIN PD 340, and may isolate the balanced PIN PD 340 from the output voltage of the TIA 365.
- the TIA 365 may be a large- valued feedback resistor.
- the gain, bandwidth, input offset current, and input offset voltages of the TIA 365 may vary depending on the type of receiver (e.g., the receiver 140 of FIG. 1 ) used in conjunction therewith.
- the TIA 365 may receive the electrical signal 345, and may transform the electrical signal 345 to a voltage. It should be noted that, unlike the TIA 265 of FIG. 2 , the TIA 365 of the present example does not require a low-noise amplifier, as the ROIC 350 is able to effectively separate noise from the signal various signals associated therewith by omitting ADC block and 2D FFT block that are sensitive to noise. After transforming the electrical signal 345 to a voltage and/or amplifying the current output of the electrical signal 345, the TIA 365 may deliver a TIA signal 367 to the comparator 380 (e.g., as a relatively small analog signal).
- the comparator 380 e.g., as a relatively small analog signal.
- the comparator 380 may then transform the TIA signal 367 from the TIA 365 to a pulse signal 385.
- the pulse signal 385 from the comparator 380 is then sent along as data effective for signal processing. Then, the data indicated by the pulse signal 385 may be sent to the primary PLL block 390i, which may be supported by a bias circuit 392.
- the primary PLL block 390i and/or the primary counter block 395i may count the frequency of the pulse signal 385, and may send the count along as a count signal 387a to the depth/velocity block 375.
- the primary PLL block 390i may generate a primary signal by multiplying the frequency of the pulse signal 385.
- the primary counter block 395i may count the number of pulses of the primary signal for a certain time period to generate primary counter data (e.g., count signal 387a).
- the depth/velocity block 375 is capable of simple multiplication and division to enable calculation of the depth and/or velocity of the target with respect to the LiDAR system based on the count signal 387a.
- the depth/velocity block 375 may be an arithmetic logic circuit or may be implemented as part of an arithmetic logic circuit.
- the readout architecture 300 may include the optical mixers between the balanced PIN PD 340 and the TIA 365
- the TIA signal 367 output by TIA 365 may include a low frequency sinusoidal waveform after filtered high frequency.
- the primary PLL block 390i in conjunction with the primary counter block 395i and the depth/velocity block 375, may be used to multiply beat frequency to enable calculation of the depth of the target (e.g., the target 130 of FIG. 1 ) using a direct time-of-flight (dToF) method.
- the primary PLL block 390i may be omitted, and the pulse signal 385 output by the comparator 380 may be sent to the primary counter block 395i and may be used, by itself, to produce a count for a given timeframe.
- the count frequency may be converted to a digital signal (e.g., by using a frequency-to-digital (F2D) converter) to determine depth/range/distance information associated with the target, wherein a smaller distance between the input frequency of the input beam 135 and the output frequency of the output beam 125 corresponds to a longer delay, thereby resulting in a high frequency and a higher count.
- F2D frequency-to-digital
- the readout architecture 300 may be suitable for high-resolution depth information. Accordingly, a count associated with the beat frequency of the pulse signal 385 may be longer, due to the increased distance of the target from the LiDAR system. Accordingly, according to some embodiments, a reference target at a reference distance that is associated with a refence count may be used to calculate depth information associated with the target.
- the reference PLL block 390r (e.g., in conjunction with the reference counter block 395r) may be used to count a reference target frequency (e.g., a frequency of a pulse signal associated with a reference target at a reference distance), and may send the count to the depth/velocity block 375 as a reference count signal 387b.
- the reference PLL block 390r may amplify a reference PLL signal to generate a reference signal having a reference frequency (e.g., reference target frequency) corresponding to a reference distance.
- the reference counter block 395r may count the number of pulses of the reference signal during a time period to receive the reference signal and generate reference counter data (e.g., reference count signal 387b). Thereafter, depth information associated with the target can be calculated as shown in Equation 1 below.
- D RD * CNT_i / CNT_r
- D is the calculated distance of the target
- RD is the reference distance of the reference target
- CNT_i is the count determined by the primary PLL block 390i and/or the primary counter block 395i (e.g., count signal 387a) in association with the distance of the target
- CNT_r is the count determined by the reference PLL block 390r and/or the reference counter block 395r (e.g., reference count signal 387b) in association with the reference distance of the reference target.
- resolution, accuracy, and noise of the readout architecture 300 may depend on a speed of the primary PLL block 390i and on a length of a counting time, wherein a longer counting time enables a more accurate and higher resolution readout architecture.
- a single frame memory 360 may be used to calculate velocity information.
- the depth/velocity block 375 may deliver depth information in a signal 393 delivered to the frame memory 360. Then, the frame memory 360 can store the depth information, and can thereafter calculate velocity of the target using the depth information corresponding to a distance to the target at a first time, new depth information corresponding to a distance to the target at a second time, and the time difference between the first time and the second time.
- the primary PLL block 390i, the reference PLL block 390r, the primary counter block 395i, the reference counter block 395r, the depth/velocity block 375, and/or the frame memory 360 may be implemented by an ALU, although the present disclosure is not limited thereto.
- velocity information associated with the target can be calculated as shown in Equation 2 below.
- V avg distance_i + 1 ⁇ distance_i / tunit or distance_i + 1 ⁇ distance_i when tunit equals to 1 second.
- V velocity
- distance_i is the distance of the target at a first time
- distance_i+1 is the distance of the target at a second time
- tunit depends on a timing of the chirping of the chirped signal corresponding to the input beam.
- tnuit is the time difference between the first time and the second time.
- FIG. 4 shows examples of various signals of a simulated ROIC of a readout architecture corresponding to a target at varying distances, according to some embodiments of the present disclosure.
- the ROIC may correspond to the ROIC 350 of FIG. 3 .
- a transmitted Tx Chirping signal 425 may correspond to a signal of an output beam (e.g., the output beam 125 of FIG. 1 ).
- the Tx Chirping signal 425 may have a frequency of about 2 microseconds, as indicated by the count, although the present disclosure is not limited thereto.
- the primary PLL block 390i and/or the primary counter block 395i may produce a first signal 487a1 corresponding to a first target distance of 150 meters, and may produce a second signal 487a2 corresponding to a second target distance of 15 meters, while the reference PLL block 390r and/or the reference counter block 395r may produce a reference signal 487b corresponding to a reference distance of a reference target.
- a count value corresponding to the first signal 487a1 may be about 3510
- the count value corresponding to the second signal 487a2 may be about 130
- the count value of the reference signal 487b may be about 352.
- FIG. 5 shows a flow chart depicting a method of determining a distance of a target from a LiDAR system, according to some embodiments of the present disclosure.
- a LiDAR system may receive a waveform signal corresponding to a target (e.g., the target 130 of FIG. 1 ).
- a target e.g., the target 130 of FIG. 1
- the balanced PIN PD 340 shown in FIG. 3 may receive the waveform signal.
- the LiDAR system may receive the waveform signal by generating a beam toward the target to cause the target to scatter the beam (e.g., may generate the output beam 125 by using the light source 110 and the scanner 120 of FIG.
- the scattered beam as the waveform signal corresponding to the target (e.g., may receive the input beam 135 by using the scanner 120, the mirror 115, and the receiver 140 of FIG. 1 ), and by receiving the waveform signal from an optical mixer output.
- the LiDAR system may generate a pulse signal from the received waveform signal (e.g., the TIA 365 and/or the comparator 380 may generate the pulse signal 385, as shown in FIG. 3 ).
- the TIA 365 may convert the waveform signal into a voltage signal
- the comparator 380 may convert the voltage signal into a pulse signal.
- the comparator 380 may compare the primary signal and the reference signal.
- the comparator 380 may compare the primary counter data and the reference counter data.
- the LiDAR system may reset a primary counter and a reference counter (e.g., the primary counter block 395i and the reference counter block 395r of FIG. 3 may be reset).
- a primary counter and a reference counter e.g., the primary counter block 395i and the reference counter block 395r of FIG. 3 may be reset.
- the LiDAR system may begin a primary count with the primary counter and a reference count with the reference counter after a first time period since the reset of the primary counter and the reference counter has elapsed.
- the first time period may correspond to a detection range.
- the LiDAR system may continue the primary count and the reference count for a second time period.
- the second time period may correspond to a flexible programmed time period.
- the LiDAR system may retain a last primary value of the primary counter and a last reference value of the reference counter (e.g., by using the frame memory 360 of FIG. 3 ).
- the LiDAR system may calculate a ratio of the last primary value to the last reference value (e.g., by using the depth and/or depth/velocity block 375 of FIG. 3 ).
- the LiDAR system may multiply a reference distance by the ratio of the last primary value to the last reference value (e.g., by using the depth and/or depth/velocity block 375 of FIG. 3 ).
- the LiDAR system may calculate velocity information (e.g., by using the depth and/or depth/velocity block 375 and the frame memory 360 of FIG. 3 ), by receiving depth information comprising a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- the LiDAR system may calculate the depth information by multiplying the reference distance by a ratio of the primary count to the reference count (e.g., by using the depth and/or depth/velocity block 375 of FIG. 3 ).
- the ROIC of various embodiments of the present disclosure may be integrated with an OPLL at a same chip die. Accordingly, a size of the LiDAR system employing the ROIC of the disclosed embodiments may be reduced. Further, the ROIC of the disclosed embodiments enables the readout architecture to achieve ultra-low power, high-speed operation, and can extract depth and velocity information without unacceptable system delay. Further, depth information may be obtained by using a relatively simple counter(s) while omitting use of a relatively complex FFT block (e.g., the FFT block 280 of the ROIC 250 of FIG. 2 ).
- a relatively simple counter(s) while omitting use of a relatively complex FFT block (e.g., the FFT block 280 of the ROIC 250 of FIG. 2 ).
- the readout architecture of the disclosed embodiments is also able to calculate velocity information by using the relatively simple counter(s) and a relatively small frame memory that is both able to store the depth information, and able to determine a difference of respective depths between two frames to determine velocity information.
- embodiments of the present disclosure provide a readout architecture used in conjunction with a LiDAR system to improve size, speed, and accuracy of the LiDAR system.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Electromagnetism (AREA)
- Optical Radar Systems And Details Thereof (AREA)
Abstract
Description
- One or more aspects of embodiments of the present disclosure relate generally to LiDAR systems.
- LiDAR is a portmanteau of light and radar and may refer to "light detection and ranging" or "laser imaging, detection, and ranging," and refers to a method for measuring distances and/or velocity of a target or object by illuminating the target with light (e.g., laser light) from one or more sources, and by measuring a reflection of the light off of the target with one or more sensors.
- For example, differences in return times (e.g., an amount of time between emission of the light and detection of the reflection of portions of the light) and/or differences in wavelengths (e.g., a difference between a wavelength of emitted laser light and a wavelength of the detected reflected laser light) may be used to determine a distance of the target (e.g., a distance from the light source or the sensor to the target), and may also be used to make a digital 3-dimensional representation of one or more portions of the target that reflect the light.
- The above information disclosed in this Background section is only for enhancement of understanding of the background of the disclosure, and therefore may contain information that does not form the prior art.
- Embodiments described herein provide improvements LiDAR technology, including improvements to a readout architecture, which includes a readout integrated chip, that is used in frequency-modulated continuous wave (FMCW) LiDAR systems.
- According to embodiments of the present disclosure, there is provided a readout integrated circuit of a FMCW LiDAR system, the readout integrated circuit including a balanced PIN photo diode configured to receive an optical signal, and convert the optical signal into a current signal, an optical mixer configured to receive the current signal from the balanced PIN photo diode, and convert the current signal into a mixed signal of high and low frequencies, a transimpedance amplifier configured to receive the converted current signal from the optical mixer; and convert the converted current signal into a voltage signal including a low frequency sinusoidal waveform after filtered high frequency, a comparator configured to receive the voltage signal from the transimpedance amplifier, and generate a pulse signal from the voltage signal, a primary counter configured to receive a primary signal corresponding to the pulse signal generated by the comparator, and count a number of pulses of the primary signal during a certain time period to generate primary counter data, a reference counter configured to receive a reference signal, and count a number of pulses of the reference signal during a time period to generate reference counter data, and an arithmetic logic unit configured to receive the primary counter data and the reference counter data, and calculate depth information corresponding to a distance of a target based on the primary counter data and the reference counter data.
- The readout integrated circuit may further include a primary phase-locked loop (PLL) configured to receive the pulse signal from the comparator, and multiply a frequency of the pulse signal to generate the primary signal, and a reference phase-locked loop configured to amplify a reference PLL signal to generate the reference signal having a reference frequency corresponding to a reference distance.
- The comparator may be further configured to compare the primary signal and the reference signal.
- The arithmetic logic unit may be configured to calculate the depth information by multiplying the reference distance by a ratio of the reference counter data to the primary counter data.
- The readout integrated circuit may further include a frame memory configured to receive the primary counter data, receive the depth information from the arithmetic logic unit, and store the depth information.
- The arithmetic logic unit may be further configured to calculate velocity information by receiving, from the frame memory, the depth information, the depth information including a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- The comparator may be further configured to compare the primary counter data and the reference counter data.
- According other embodiments of the present disclosure, there is provided a method to determine a distance or a velocity of an object, the method including receiving a waveform signal corresponding to a target, generating a pulse signal from the received waveform signal, resetting a primary counter and a reference counter, beginning a primary count with the primary counter and a reference count with the reference counter after a first time period since the resetting of the primary counter and the reference counter has elapsed, continuing the primary count and the reference count for a second time period, retaining a last primary value of the primary counter and a last reference value of the reference counter, calculating a ratio of the last primary value to the last reference value, and multiplying a reference distance by the ratio.
- The method may further include receiving the waveform signal from an optical mixer output.
- The first time period may correspond to a detection range.
- The second time period may correspond to a flexible programmed time period.
- The method may further include generating a beam toward the target to cause the target to scatter the beam, and receiving the scattered beam as the waveform signal corresponding to the target.
- The method may further include calculating velocity information by receiving depth information including a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- The method may further include calculating the depth information by multiplying the reference distance by a ratio of the primary count to the reference count.
- According to yet other embodiments of the present disclosure, there is provided a non-transitory computer readable medium (exemplarily implemented on a readout integrated circuit of a FMCW LiDAR system), the non-transitory computer readable medium having computer code that, when executed on a processor, implements a method of determining a distance or a velocity of an object, the method including receiving a waveform signal corresponding to a target, generating a pulse signal from the received waveform signal, resetting a primary counter and a reference counter, beginning a primary count with the primary counter and a reference count with the reference counter after a first time period since the resetting has elapsed, continuing the primary count and the reference count for a second time period, retaining a last primary value of the primary counter and a last reference value of the reference counter, calculating a ratio of the last primary value to the last reference value, and multiplying a reference distance by the ratio.
- The computer code, when executed on the processor, may further implement the method of determining the distance or the velocity of the object by receiving the waveform signal from an optical mixer output.
- The first time period may correspond to a detection range, wherein the second time period corresponds to a flexible programmed time period.
- The computer code, when executed on the processor, may further implement the method of determining the distance or the velocity of the object by generating a beam toward the target to cause the target to scatter the beam, and receiving the scattered beam as the waveform signal corresponding to the target.
- The computer code, when executed on the processor, may further implement the method of determining the distance or the velocity of the object by calculating velocity information by receiving depth information including a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- The computer code, when executed on the processor, may further implement the method of determining the distance or the velocity of the object by calculating the depth information by multiplying the reference distance by a ratio of the primary count to the reference count.
- Accordingly, the system and method of some embodiments of the present disclosure are able to calculate a range of a target while omitting circuit elements used to perform transforms to process signals in calculating depth/distance/range.
- Non-limiting and non-exhaustive embodiments of the present embodiments are described with reference to the following figures, wherein like reference numerals refer to like parts throughout the various views unless otherwise specified.
-
FIG. 1 shows a block diagram depicting a LiDAR system, according to some embodiments of the present disclosure; -
FIG. 2 shows a block diagram depicting a readout architecture of a frequency-modulated continuous wave (FMCW) LiDAR system; -
FIG. 3 shows a block diagram depicting a readout architecture of a frequency-modulated continuous wave (FMCW) LiDAR system, according to some embodiments of the present disclosure; -
FIG. 4 shows examples of various signals of a simulated ROIC of a readout architecture corresponding to a target at varying distances, according to some embodiments of the present disclosure; and -
FIG. 5 shows a flow chart depicting a method of determining a distance of a target from a LiDAR system, according to some embodiments of the present disclosure. - Corresponding reference characters indicate corresponding components throughout the several views of the drawings. Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity, and have not necessarily been drawn to scale. For example, the dimensions of some of the elements, layers, and regions in the figures may be exaggerated relative to other elements, layers, and regions to help to improve clarity and understanding of various embodiments. Also, common but well-understood elements and parts not related to the description of the embodiments might not be shown in order to facilitate a less obstructed view of these various embodiments and to make the description clear.
- Features of the inventive concept and methods of accomplishing the same may be understood more readily by reference to the detailed description of embodiments and the accompanying drawings. Hereinafter, embodiments will be described in more detail with reference to the accompanying drawings. The described embodiments, however, may be embodied in various different forms, and should not be construed as being limited to only the illustrated embodiments herein. Rather, these embodiments are provided as examples so that this disclosure will be thorough and complete, and will fully convey the aspects and features of the present inventive concept to those skilled in the art. Accordingly, processes, elements, and techniques that are not necessary to those having ordinary skill in the art for a complete understanding of the aspects and features of the present inventive concept may not be described.
- Unless otherwise noted, like reference numerals denote like elements throughout the attached drawings and the written description, and thus, descriptions thereof will not be repeated. Further, parts not related to the description of the embodiments might not be shown to make the description clear. In the drawings, the relative sizes of elements, layers, and regions may be exaggerated for clarity.
- In the detailed description, for the purposes of explanation, numerous specific details are set forth to provide a thorough understanding of various embodiments. It is apparent, however, that various embodiments may be practiced without these specific details or with one or more equivalent arrangements. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring various embodiments.
- It will be understood that, although the terms "first," "second," "third," etc., may be used herein to describe various elements, components, regions, layers and/or sections, these elements, components, regions, layers and/or sections should not be limited by these terms. These terms are used to distinguish one element, component, region, layer or section from another element, component, region, layer or section. Thus, a first element, component, region, layer or section described below could be termed a second element, component, region, layer or section, without departing from the scope of the present disclosure.
- As used herein, the singular forms "a" and "an" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises," "comprising," "have," "having," "includes," and "including," when used in this specification, specify the presence of the stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
- As used herein, the term "substantially," "about," "approximately," and similar terms are used as terms of approximation and not as terms of degree, and are intended to account for the inherent deviations in measured or calculated values that would be recognized by those of ordinary skill in the art. "About" or "approximately," as used herein, is inclusive of the stated value and means within an acceptable range of deviation for the particular value as determined by one of ordinary skill in the art, considering the measurement in question and the error associated with measurement of the particular quantity (i.e., the limitations of the measurement system). For example, "about" may mean within one or more standard deviations, or within ± 30%, 20%, 10%, 5% of the stated value. Further, the use of "may" when describing embodiments of the present disclosure refers to "one or more embodiments of the present disclosure."
- When a certain embodiment may be implemented differently, a specific process order may be performed differently from the described order. For example, two consecutively described processes may be performed substantially at the same time or performed in an order opposite to the described order.
- The electronic or electric devices and/or any other relevant devices or components according to embodiments of the present disclosure described herein may be implemented utilizing any suitable hardware, firmware (e.g. an application-specific integrated circuit), software, or a combination of software, firmware, and hardware. For example, the various components of these devices may be formed on one integrated circuit (IC) chip or on separate IC chips. Further, the various components of these devices may be implemented on a flexible printed circuit film, a tape carrier package (TCP), a printed circuit board (PCB), or formed on one substrate.
- Further, the various components of these devices may be a process or thread, running on one or more processors, in one or more computing devices, executing computer program instructions and interacting with other system components for performing the various functionalities described herein. The computer program instructions are stored in a memory which may be implemented in a computing device using a standard memory device, such as, for example, a random access memory (RAM). The computer program instructions may also be stored in other non-transitory computer readable media such as, for example, a CD-ROM, flash drive, or the like. Also, a person of skill in the art should recognize that the functionality of various computing devices may be combined or integrated into a single computing device, or the functionality of a particular computing device may be distributed across one or more other computing devices without departing from the scope of the embodiments of the present disclosure.
- As mentioned above, a LiDAR system may be used to determine depth (e.g., the distance of a target from the LiDAR system) and/or a velocity of a target with respect thereto. Some embodiments of the present disclosure provide improvements to LiDAR technology by omitting components that otherwise may be used in a LiDAR system, thereby improving the speed and accuracy of the LiDAR system, and potentially reducing manufacturing costs associated therewith.
-
FIG. 1 shows a block diagram depicting a LiDAR system, according to some embodiments of the present disclosure. - Referring to
FIG. 1 , aLiDAR system 100 includes alight source 110, amirror 115, ascanner 120, areceiver 140, and acontroller 150. TheLiDAR system 100 may be referred to as a laser ranging system, a laser radar system, or a laser detection and ranging system. - In some embodiments, the
LiDAR system 100 may be configured to sense, identify, or determine distances to one ormore targets 130 within a field of regard. As an example, theLiDAR system 100 may determine a distance to atarget 130, where all or part of thetarget 130 is contained within a field of regard (FOR) of theLiDAR system 100. All or part of thetarget 130 being contained within a FOR of theLiDAR system 100 may refer to the FOR overlapping, encompassing, or enclosing at least a portion of thetarget 130. In some embodiments,target 130 may include all or part of an object that is moving or stationary relative toLiDAR system 100. - The
light source 110 may correspond to a laser for producing anoutput beam 125 having an operating wavelength (e.g., a wavelength in the electromagnetic spectrum). Theoutput beam 125 may be referred to as an optical signal, a laser beam, a light beam, an optical beam, an emitted beam, an emitted light, or simply a beam. In some embodiments, theLiDAR system 100 is a frequency-modulated continuous-wave (FMCW) LiDAR system. Accordingly, theoutput beam 125 may be an FMCW laser, although it should be noted that theoutput beam 125 may be pulsed or otherwise modulated in other embodiments. Theoutput beam 125 may be directed toward thetarget 130 that is a distance D from theLiDAR system 100. Before being directed to thetarget 130, theoutput beam 125 may pass through, or by, the mirror 115 (e.g., themirror 115 may include a hole, slot, or aperture through which theoutput beam 125 passes). - After passing through or by the
mirror 115, theoutput beam 125 may also pass through thescanner 120. Thescanner 120 may be configured to scan theoutput beam 125 across an area of interest (e.g., the FOR) of theLiDAR system 100. Thescanner 120 may include one or more scanning mirrors configured to pivot, rotate, oscillate, or move in an angular manner about one or more rotation axes. Accordingly, theoutput beam 125 may be reflected by a scanning mirror, and as the scanning mirror pivots or rotates, the reflectedoutput beam 125 may be scanned in a corresponding angular manner. For example, a scanning mirror may be configured to periodically pivot back and forth, thereby causing theoutput beam 125 to scan back and forth. - After passing though the
scanner 120, once theoutput beam 125 reaches thetarget 130, thetarget 130 may scatter or otherwise reflect at least a portion of the light of theoutput beam 125, and some of the scattered or reflected light may be returned toward theLiDAR system 100 as one or more input beams 135 as a received optical signal. Generally, a relatively small fraction of the light from theoutput beam 125 is returned to theLiDAR system 100 as theinput beam 135. Theinput beam 135 may then return while passing through thescanner 120, and may be then reflected by themirror 115 to be directed to thereceiver 140. - In addition to the
mirror 115, theLiDAR system 100 may include one or more optical components (e.g., lenses, mirrors, or filters) that are configured to reflect, focus, filter, shape, modify, steer, direct, collimate, or combine light produced by, or received by, the LiDAR system 100 (e.g., to steer or focus theoutput beam 125 or the input beam 135). As an example, theLiDAR system 100 may include one or more lenses to focus theinput beam 135 onto a photodetector of thereceiver 140. In some embodiments, themirror 115 may provide for theoutput beam 125 and theinput beam 135 to be substantially coaxial so that the two beams travel along approximately the same optical path (e.g., to travel substantially parallel to one another and in opposite directions). Accordingly, themirror 115 and other corresponding components may direct theinput beam 135 toward thereceiver 140. - The
receiver 140 may be referred to as a photoreceiver, optical receiver, optical sensor, detector, photodetector, or optical detector. Thereceiver 140 may include one or more avalanche photodiodes (APDs), single-photon avalanche diodes (SPADs), PN photodiodes (PD), and/or PIN photodiodes (e.g., abalanced PIN PD 340, as shown inFIG. 3 ). Thereceiver 140 may include electronic circuitry that performs signal modification and analysis. Thereceiver 140 may include a transimpedance amplifier (TIA) (e.g., theTIA FIGS. 2 and3 ) that converts a received photocurrent, which is produced in response to a received optical signal, into a voltage signal. The voltage signal may be used to determine one or more optical characteristics of theinput beam 135, and may be used to produce a digital output signal/electrical signal 145 to be sent to thecontroller 150 for processing or analysis (e.g., to determine a time-of-flight value corresponding to a received optical pulse). - Accordingly, after being reflected by the
mirror 115, thereceiver 140 may receive and detect photons from theinput beam 135. Thereceiver 140 may then produce and output an electrical current or voltage pulse as anelectrical signal 145 that represents theinput beam 135, and may send theelectrical signal 145 to the controller 150 (e.g., a controller that may be part of, or that may include, theROIC FIGS. 2 and3 ). - The
controller 150 may include a processor, computing system, or other suitable circuitry, and may be configured to analyze one or more characteristics of theelectrical signal 145 from thereceiver 140 to determine one or more characteristics of thetarget 130, such as a distance of, or velocity of, the target with respect to theLiDAR system 100. This may be done, for example, by analyzing a time of flight of the light that is common to the transmittedoutput beam 125 and the receivedinput beam 135. - A time of flight T represents a round-trip time of flight for an emitted beam or pulse of light to travel from the
LiDAR system 100 to thetarget 130, and then from thetarget 130 back to theLiDAR system 100. The time of flight T may be used to calculate the distance/range/depth D from thetarget 130 to theLiDAR system 100, where the distance D may be expressed as D=c*T/2, where c is the speed of light. - The
controller 150 may also be electrically or communicatively coupled to thelight source 110, themirror 115, and thescanner 120. Thecontroller 150 may receive electrical trigger pulses or edges as a result of thelight source 110, where each pulse or edge corresponds to the emission of an optical pulse by thelight source 110. Thecontroller 150 may also control thelight source 110 to cause thelight source 110 to produce and/or adjust theoutput beam 125. As described further below, thecontroller 150 may determine a time-of-flight value for an optical pulse based on timing information associated with when theoutput beam 125 was emitted by thelight source 110, and when theinput beam 135 was detected by thereceiver 140. Accordingly, thecontroller 150 may include circuitry that performs signal amplification, sampling, filtering, signal conditioning, analog-to-digital conversion, time-to-digital conversion, pulse detection, threshold detection, rising-edge detection, or falling-edge detection. - In some embodiments, one or
more LiDAR systems 100 may be integrated into a vehicle, for example, as part of an advanced driver assistance system (ADAS) to assist a driver of the vehicle in operating the vehicle, or as part of an autonomous vehicle driving system. For example, theLiDAR system 100 may be part of an ADAS that provides information or feedback to a driver (e.g., to alert the driver to potential problems or hazards), or that automatically controls one or more functions of the vehicle (e.g., by controlling a braking system or a steering system) to avoid collisions or accidents. TheLiDAR system 100 may provide information about the surrounding environment to a driving system of an autonomous vehicle. An autonomous-vehicle driving system may be configured to guide the autonomous vehicle through an environment surrounding the vehicle, and may include one or more computing systems that receive information from theLiDAR system 100 about the surrounding environment, that analyze the received information, and that provide control signals to the vehicle's driving systems and actuators (e.g., steering wheel, accelerator, brake, or turn signal). -
FIG. 2 shows a block diagram depicting a readout architecture of a frequency-modulated continuous wave (FMCW) LiDAR system. - Referring to
FIG. 2 , thereadout architecture 200 of a FMCW LiDAR system (e.g., theLiDAR system 100 ofFIG. 1 ) may include abalanced PIN PD 240 that is implemented as a receiver (e.g., thereceiver 140 ofFIG. 1 ) of the LiDAR system, and also may include a readout integrated circuit (ROIC) 250 that is implemented as a controller (e.g., thecontroller 150 ofFIG. 1 ) of the LiDAR system. TheROIC 250 may be used in connection with a memory (PC) 260. TheROIC 250 may include, in relevant part, aTIA 265, an analog-to-digital controller (ADC) block 270 (e.g., an ultra-high speed high resolution ADC block 270), a Fast Fourier Transform (FFT) block 280 (e.g., a 2-dimensional FFT block), and a phase-locked loop (PLL)block 290. - The
balanced PIN PD 240 may receive photons corresponding to an input beam (e.g., theinput beam 135 ofFIG. 1 ), and may then convert the photons to current as a signal 245 (e.g., theelectrical signal 145 ofFIG. 1 ). Then, theTIA 265 may receive thesignal 245, and may transform the current of thesignal 245 to a voltage signal (e.g., a TIA signal 267) to be sent to theADC block 270. Then theADC block 270 transforms the TIA signal 267 received from theTIA 265 into data flow (e.g., a data signal 285), such that each pixel, or voxel (e.g., a single sample, or data point, on a regularly spaced, three-dimensional grid), detected by thebalanced PIN PD 240 may be converted to one or more corresponding data values. Thereafter, the FFT block 280 may run a transform (e.g., an FFT) on the data values of the data signal 285 received from the ADC block 270 to calculate depth information corresponding to the target reflecting or scattering theinput beam 135. Finally, depth values indicated in adepth signal 287 from theFFT block 280, along with the data values of the data signal 285 from theADC block 270, are sent to thememory 260 via two respective lines to thememory 260 from theADC block 270 and theFFT block 280, thereby enabling calculation of both depth and velocity of the target (e.g., thetarget 130 ofFIG. 1 ) by using the data signal 285 and thedepth signal 287. - High-resolution and high-quality range/depth/distance information may be used by advanced driver assistance system (ADAS) and autonomous driving applications to enable the LiDAR system to "see" further (e.g., to have a longer target detection range), and to enable the LiDAR system to detect or recognize the target more quickly and easily, thereby reducing risk associated with the vehicle employing the LiDAR system either hitting or avoiding the target(s). Further, a low-power-operation LiDAR system may be implemented with an ADAS or autonomous driving system. Accordingly, specifications of the
ROIC 250 may consider receiver bandwidth (e.g., on the order of a few hundred MHz or more), ADC resolution and detection capabilities, low noise (e.g., a signal-to-noise ratio (SNR) of 12 dB or more for signal power, and/or a low-noise amplifier (LNA) used as the TIA), power consumption, performance of direct Fourier transforms (DFT) at chip level, and other RF processes. - However, target analyses in a 3D system generally consume a relatively large amount of power. Furthermore, a LiDAR system that is able to calculate target depth using a direct time-of-flight (dToF) method in conjunction with a single photo avalanche diode (SPAD) as the receiver may have difficulty due to the presence of various sources of ambient noise. Accordingly, implementation of the
readout architecture 200 ofFIG. 2 may present obstacles associated with theADC block 270 and theFFT block 280. - For example, the
2D FFT block 280 may require a relatively large amount of space on theROIC 250, and may consume a relatively large amount of power for several frames to enable detection of depth and velocity information, noting that thereadout architecture 200 in the FMCW LiDAR system may concurrently determine depth and velocity by using the2D FFT block 280. Further, there may be lag associated with operation of the FFT block 280 when determining depth and velocity information. - As another example, the ADC block 270 may operate on the order of GHz (e.g., may have a sampling speed of a few GHz) to effectuate moderate-to-high resolution, and the
readout architecture 200 may seek to separate signal noise and crosstalk associated with other blocks of theROIC 250. This may be achieved by a high-speed, low-jitter phase-locked loop (PLL) block 290 and a highlysensitive bias circuit 292 to support high speed operation of the ADC block 270 (e.g., the PLL block 290 may be used as a clock for theADC block 270 and theFFT block 280, and may operate on the order of GHz). As a result, thereadout architecture 200 might not effectively integrate with an Optical PLL block (OPLL) due to a high level of sensitivity of each block. - Accordingly, as described below with respect to
FIGS. 3 and4 , there may be benefit associated with a novel readout architecture that addresses some of the above issues and limitations of thereadout architecture 200 ofFIG. 2 . -
FIG. 3 shows a block diagram depicting a readout architecture of a frequency-modulated continuous wave (FMCW) LiDAR system, according to some embodiments of the present disclosure. - Referring to
FIG. 3 , various issues associated with the readout architecture of a LiDAR system (e.g., thereadout architecture 200 ofFIG. 2 ) may be addressed or eliminated. For example, ambient noise associated with thereadout architecture 200 ofFIG. 2 can be effectively eliminated by using a modulated signal (e.g., a FMCW signal) in accordance with an FMCW LiDAR system of embodiments of the present disclosure (e.g., the LiDAR system 100) by removing interference otherwise caused by other LiDAR systems and/or other light sources. This may be achieved by providing a lowpower readout architecture 300 that omits the high-speed ADC block 270 and the 2D FFT block 280 of thereadout architecture 200 ofFIG. 2 , which are generally blocks that are relatively sensitive to noise, and by replacing the omitted blocks with acomparator 380 and one or more counters 395. - The
readout architecture 300 of a LiDAR system may include, in relevant part, abalanced PIN PD 340 as a receiver (e.g., thereceiver 140 ofFIG. 1 ), aROIC 350 including aTIA 365, thecomparator 380, aprimary PLL block 390i, areference PLL block 390r, a primary counter block 395i, areference counter block 395r, a depth and/or depth/velocity block 375 for calculating depth and/or velocity, and a memory (PC) 360. Unlike theROIC 250 ofFIG. 2 , theROIC 350 of the present example omits overly sensitive blocks (e.g., theADC block 270 and the FFT block 280), while still effectively calculating depth and velocity of the target. - The LiDAR system of the present example may output an output beam (e.g., the
output beam 125 ofFIG. 1 ) by using a chirping signal that varies the frequency of theoutput beam 125. The target (e.g., thetarget 130 ofFIG. 1 ) may then scatter or reflect theoutput beam 125 to thereby return an input beam (e.g., theinput beam 135 ofFIG. 1 ) to the LiDAR system (e.g., to the balanced PIN PD 340). - Accordingly, the
balanced PIN PD 340 may detect photons of theinput beam 135 as a chirping signal. Thebalanced PIN PD 340 may then convert the photons into a current to deliver a signal (e.g., current as an electrical signal, or a digital output signal) 345 to theROIC 350, which may operate, either by itself or in conjunction with theframe memory 360, as thecontroller 150 ofFIG. 1 . Theelectrical signal 345 from thebalanced PIN PD 340 may be thought of as having a beta frequency corresponding to the chirping frequency of theinput beam 135. Theelectrical signal 345 may then be received directly by theTIA 365, although it should be noted that, in other embodiments, thereadout architecture 300 may include optical mixers between thebalanced PIN PD 340 and theTIA 365. The optical mixers may convert the frequency of theelectrical signal 345 from thebalanced PIN PD 340. For example, the optical mixers may convert theelectric signal 345 into a mixed signal of low frequency and high frequency. - The
TIA 365 is a current-to-voltage converter, and may be used with one or more operational amplifiers. TheTIA 365 may be used to amplify the current output, which corresponds to the receivedelectrical signal 345, to a usable voltage (e.g., to a suitable voltage to be used by the comparator 380). TheTIA 365 may be used because thebalanced PIN PD 340 may have a current response that is more linear than the voltage response (e.g., the current response of thebalanced PIN PD 340 may be about 1% nonlinearity or better over a wide range of light input). TheTIA 365 presents a low impedance to thebalanced PIN PD 340, and may isolate thebalanced PIN PD 340 from the output voltage of theTIA 365. In some embodiments, theTIA 365 may be a large- valued feedback resistor. The gain, bandwidth, input offset current, and input offset voltages of theTIA 365 may vary depending on the type of receiver (e.g., thereceiver 140 ofFIG. 1 ) used in conjunction therewith. - Accordingly, the
TIA 365 may receive theelectrical signal 345, and may transform theelectrical signal 345 to a voltage. It should be noted that, unlike theTIA 265 ofFIG. 2 , theTIA 365 of the present example does not require a low-noise amplifier, as theROIC 350 is able to effectively separate noise from the signal various signals associated therewith by omitting ADC block and 2D FFT block that are sensitive to noise. After transforming theelectrical signal 345 to a voltage and/or amplifying the current output of theelectrical signal 345, theTIA 365 may deliver aTIA signal 367 to the comparator 380 (e.g., as a relatively small analog signal). - The
comparator 380 may then transform the TIA signal 367 from theTIA 365 to apulse signal 385. Thepulse signal 385 from thecomparator 380 is then sent along as data effective for signal processing. Then, the data indicated by thepulse signal 385 may be sent to theprimary PLL block 390i, which may be supported by abias circuit 392. - Thereafter, the
primary PLL block 390i and/or the primary counter block 395i, which may be implemented as a part of an arithmetic logic unit (ALU), may count the frequency of thepulse signal 385, and may send the count along as acount signal 387a to the depth/velocity block 375. For example, theprimary PLL block 390i may generate a primary signal by multiplying the frequency of thepulse signal 385. The primary counter block 395i may count the number of pulses of the primary signal for a certain time period to generate primary counter data (e.g.,count signal 387a). The depth/velocity block 375 is capable of simple multiplication and division to enable calculation of the depth and/or velocity of the target with respect to the LiDAR system based on thecount signal 387a. The depth/velocity block 375 may be an arithmetic logic circuit or may be implemented as part of an arithmetic logic circuit. In the embodiment where thereadout architecture 300 may include the optical mixers between thebalanced PIN PD 340 and theTIA 365, the TIA signal 367 output byTIA 365 may include a low frequency sinusoidal waveform after filtered high frequency. - Accordingly, the
primary PLL block 390i, in conjunction with the primary counter block 395i and the depth/velocity block 375, may be used to multiply beat frequency to enable calculation of the depth of the target (e.g., thetarget 130 ofFIG. 1 ) using a direct time-of-flight (dToF) method. However, it should be noted that, in other embodiments, theprimary PLL block 390i may be omitted, and thepulse signal 385 output by thecomparator 380 may be sent to the primary counter block 395i and may be used, by itself, to produce a count for a given timeframe. - For example, by counting the beat frequency of the
pulse signal 385, the count frequency may be converted to a digital signal (e.g., by using a frequency-to-digital (F2D) converter) to determine depth/range/distance information associated with the target, wherein a smaller distance between the input frequency of theinput beam 135 and the output frequency of theoutput beam 125 corresponds to a longer delay, thereby resulting in a high frequency and a higher count. Accordingly, by counting the beat pulse of thepulse signal 385 in the time domain, a distance from the LiDAR system to the target can be determined. - In some embodiments, the
readout architecture 300 may be suitable for high-resolution depth information. Accordingly, a count associated with the beat frequency of thepulse signal 385 may be longer, due to the increased distance of the target from the LiDAR system. Accordingly, according to some embodiments, a reference target at a reference distance that is associated with a refence count may be used to calculate depth information associated with the target. - For example, the
reference PLL block 390r (e.g., in conjunction with thereference counter block 395r) may be used to count a reference target frequency (e.g., a frequency of a pulse signal associated with a reference target at a reference distance), and may send the count to the depth/velocity block 375 as areference count signal 387b. For example, thereference PLL block 390r may amplify a reference PLL signal to generate a reference signal having a reference frequency (e.g., reference target frequency) corresponding to a reference distance. Thereference counter block 395r may count the number of pulses of the reference signal during a time period to receive the reference signal and generate reference counter data (e.g.,reference count signal 387b). Thereafter, depth information associated with the target can be calculated as shown inEquation 1 below. - Where D is the calculated distance of the target, RD is the reference distance of the reference target, CNT_i is the count determined by the
primary PLL block 390i and/or the primary counter block 395i (e.g.,count signal 387a) in association with the distance of the target, and CNT_r is the count determined by thereference PLL block 390r and/or thereference counter block 395r (e.g.,reference count signal 387b) in association with the reference distance of the reference target. - Accordingly, resolution, accuracy, and noise of the
readout architecture 300 may depend on a speed of theprimary PLL block 390i and on a length of a counting time, wherein a longer counting time enables a more accurate and higher resolution readout architecture. - Furthermore, a
single frame memory 360 may be used to calculate velocity information. For example, the depth/velocity block 375 may deliver depth information in a signal 393 delivered to theframe memory 360. Then, theframe memory 360 can store the depth information, and can thereafter calculate velocity of the target using the depth information corresponding to a distance to the target at a first time, new depth information corresponding to a distance to the target at a second time, and the time difference between the first time and the second time. It should be noted that theprimary PLL block 390i, thereference PLL block 390r, the primary counter block 395i, thereference counter block 395r, the depth/velocity block 375, and/or theframe memory 360 may be implemented by an ALU, although the present disclosure is not limited thereto. -
- Where V is velocity, distance_i is the distance of the target at a first time, distance_i+1 is the distance of the target at a second time, and "tunit" depends on a timing of the chirping of the chirped signal corresponding to the input beam. "tnuit" is the time difference between the first time and the second time.
-
FIG. 4 shows examples of various signals of a simulated ROIC of a readout architecture corresponding to a target at varying distances, according to some embodiments of the present disclosure. - Referring to
FIG. 4 , the ROIC may correspond to theROIC 350 ofFIG. 3 . For example, a transmitted Tx Chirping signal 425 may correspond to a signal of an output beam (e.g., theoutput beam 125 ofFIG. 1 ). The Tx Chirping signal 425 may have a frequency of about 2 microseconds, as indicated by the count, although the present disclosure is not limited thereto. Accordingly, theprimary PLL block 390i and/or the primary counter block 395i may produce a first signal 487a1 corresponding to a first target distance of 150 meters, and may produce a second signal 487a2 corresponding to a second target distance of 15 meters, while thereference PLL block 390r and/or thereference counter block 395r may produce areference signal 487b corresponding to a reference distance of a reference target. As shown inFIG. 4 , a count value corresponding to the first signal 487a1 may be about 3510, the count value corresponding to the second signal 487a2 may be about 130, and the count value of thereference signal 487b may be about 352. -
FIG. 5 shows a flow chart depicting a method of determining a distance of a target from a LiDAR system, according to some embodiments of the present disclosure. - Referring to
FIG. 5 , at S501 a LiDAR system (e.g., theLiDAR system 100 ofFIG. 1 ) may receive a waveform signal corresponding to a target (e.g., thetarget 130 ofFIG. 1 ). For example, thebalanced PIN PD 340 shown inFIG. 3 may receive the waveform signal. According to some embodiments, the LiDAR system may receive the waveform signal by generating a beam toward the target to cause the target to scatter the beam (e.g., may generate theoutput beam 125 by using thelight source 110 and thescanner 120 ofFIG. 1 ), by receiving the scattered beam as the waveform signal corresponding to the target (e.g., may receive theinput beam 135 by using thescanner 120, themirror 115, and thereceiver 140 ofFIG. 1 ), and by receiving the waveform signal from an optical mixer output. - At S502, the LiDAR system may generate a pulse signal from the received waveform signal (e.g., the
TIA 365 and/or thecomparator 380 may generate thepulse signal 385, as shown inFIG. 3 ). For example, theTIA 365 may convert the waveform signal into a voltage signal, and thecomparator 380 may convert the voltage signal into a pulse signal. Thecomparator 380 may compare the primary signal and the reference signal. Alternatively, thecomparator 380 may compare the primary counter data and the reference counter data. - At S503, the LiDAR system may reset a primary counter and a reference counter (e.g., the primary counter block 395i and the
reference counter block 395r ofFIG. 3 may be reset). - At S504, the LiDAR system may begin a primary count with the primary counter and a reference count with the reference counter after a first time period since the reset of the primary counter and the reference counter has elapsed. The first time period may correspond to a detection range.
- At S505, the LiDAR system may continue the primary count and the reference count for a second time period. The second time period may correspond to a flexible programmed time period.
- At S506, the LiDAR system may retain a last primary value of the primary counter and a last reference value of the reference counter (e.g., by using the
frame memory 360 ofFIG. 3 ). - At S507, the LiDAR system may calculate a ratio of the last primary value to the last reference value (e.g., by using the depth and/or depth/velocity block 375 of
FIG. 3 ). - At S508, the LiDAR system may multiply a reference distance by the ratio of the last primary value to the last reference value (e.g., by using the depth and/or depth/velocity block 375 of
FIG. 3 ). - According to some embodiments, at S509, the LiDAR system may calculate velocity information (e.g., by using the depth and/or depth/
velocity block 375 and theframe memory 360 ofFIG. 3 ), by receiving depth information comprising a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame, calculating a difference of depth corresponding to a difference between the first depth information and the second depth information, and dividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame. - According to some embodiments, at S510, the LiDAR system may calculate the depth information by multiplying the reference distance by a ratio of the primary count to the reference count (e.g., by using the depth and/or depth/velocity block 375 of
FIG. 3 ). - The ROIC of various embodiments of the present disclosure (e.g., the
ROIC 350 of thereadout architecture 300 ofFIG. 3 ), as described above, may be integrated with an OPLL at a same chip die. Accordingly, a size of the LiDAR system employing the ROIC of the disclosed embodiments may be reduced. Further, the ROIC of the disclosed embodiments enables the readout architecture to achieve ultra-low power, high-speed operation, and can extract depth and velocity information without unacceptable system delay. Further, depth information may be obtained by using a relatively simple counter(s) while omitting use of a relatively complex FFT block (e.g., the FFT block 280 of theROIC 250 ofFIG. 2 ). The readout architecture of the disclosed embodiments is also able to calculate velocity information by using the relatively simple counter(s) and a relatively small frame memory that is both able to store the depth information, and able to determine a difference of respective depths between two frames to determine velocity information. - Thus, embodiments of the present disclosure provide a readout architecture used in conjunction with a LiDAR system to improve size, speed, and accuracy of the LiDAR system.
- While the present disclosure has been particularly shown and described with reference to some example embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the scope of the present disclosure as set forth in the following claims.
Claims (15)
- A readout integrated circuit comprising:a balanced PIN photo diode (340) configured to:receive an optical signal; andconvert the optical signal into a current signal;an optical mixer configured to:receive the current signal from the balanced PIN photo diode; andconvert the current signal into a mixed signal of high and low frequencies;a transimpedance amplifier (365) configured to:receive the converted current signal from the optical mixer; andconvert the converted current signal into a voltage signal (367) including a low frequency sinusoidal waveform after filtered high frequency;a comparator (380) configured to:receive the voltage signal (367) from the transimpedance amplifier (365); andgenerate a pulse signal (385) from the voltage signal (367);a primary counter (395i) configured to:receive a primary signal corresponding to the pulse signal (385) generated by the comparator (380); andcount a number of pulses of the primary signal during a certain time period to generate primary counter data (387a);a reference counter (395r) configured to:receive a reference signal; andcount a number of pulses of the reference signal during a time period to generate reference counter data (387b); andan arithmetic logic unit (375) configured to:receive the primary counter data (387a) and the reference counter data (387b); andcalculate depth information corresponding to a distance of a target (130) based on the primary counter data (387a) and the reference counter data (387b).
- The readout integrated circuit of claim 1, further comprising:a primary phase-locked loop (390i)) configured to:receive the pulse signal (385) from the comparator (380); andmultiply a frequency of the pulse signal (385) to generate the primary signal; anda reference phase-locked loop (390r) configured to amplify a reference phase-locked loop signal to generate the reference signal having a reference frequency corresponding to a reference distance.
- The readout integrated circuit of claim 2, wherein the comparator (380) is further configured to compare the primary signal and the reference signal.
- The readout integrated circuit of claim 2 or 3, wherein the arithmetic logic unit (375) is configured to calculate the depth information by multiplying the reference distance by a ratio of the reference counter data (387b) to the primary counter data (387a).
- The readout integrated circuit of any one of claims 1 to 4, further comprising a frame memory (360) configured to:receive the primary counter data (387a);receive the depth information from the arithmetic logic unit (375); andstore the depth information.
- The readout integrated circuit of claim 5, wherein the arithmetic logic unit (375) is further configured to calculate velocity information by:receiving, from the frame memory (360), the depth information, the depth information comprising a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame;calculating a difference of depth corresponding to a difference between the first depth information and the second depth information; anddividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- The readout integrated circuit of claim 1, wherein the comparator (380) is further configured to compare the primary counter data and the reference counter data.
- A method to determine a distance or a velocity of an object, the method comprising:receiving (S501) a waveform signal corresponding to a target (130);generating (S502) a pulse signal (385) from the received waveform signal;resetting (S503) a primary counter (395i) and a reference counter (395r);beginning (S504) a primary count with the primary counter (395i) and a reference count with the reference counter (395r) after a first time period since the resetting of the primary counter (395i) and the reference counter (395r) has elapsed;continuing (S505) the primary count and the reference count for a second time period;retaining (S506) a last primary value of the primary counter and a last reference value of the reference counter;calculating (S507) a ratio of the last primary value to the last reference value; andmultiplying (S508) a reference distance by the ratio.
- The method of claim 8, further comprising receiving the waveform signal from an optical mixer output.
- The method of claim 8 or 9, wherein the first time period corresponds to a detection range.
- The method of any one of claims 8 to 10, wherein the second time period corresponds to a flexible programmed time period.
- The method of any one of claims 8 to 11, further comprising generating a beam (125) toward the target (130) to cause the target (130) to scatter the beam (125), and receiving the scattered beam (135) as the waveform signal corresponding to the target (130).
- The method of any one of claims 8 to 12, further comprising calculating (S509) velocity information by:receiving depth information comprising a first depth information calculated for a first memory frame and a second depth information calculated for a second memory frame;calculating a difference of depth corresponding to a difference between the first depth information and the second depth information; anddividing the difference of depth by a time period corresponding to a time between the first memory frame and the second memory frame.
- The method of claim 13, further comprising calculating the depth information by multiplying the reference distance by a ratio of the primary count to the reference count.
- A non-transitory computer readable medium, the non-transitory computer readable medium having computer code that, when executed on a processor, implements the method according to any one of claims 8 to 14.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202063051542P | 2020-07-14 | 2020-07-14 | |
US17/011,813 US11988779B2 (en) | 2020-07-14 | 2020-09-03 | Readout architecture for FMCW LiDAR |
Publications (2)
Publication Number | Publication Date |
---|---|
EP3940416A1 true EP3940416A1 (en) | 2022-01-19 |
EP3940416B1 EP3940416B1 (en) | 2023-10-04 |
Family
ID=75887946
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP21173088.2A Active EP3940416B1 (en) | 2020-07-14 | 2021-05-10 | Readout architecture for fmcw lidar |
Country Status (4)
Country | Link |
---|---|
US (1) | US11988779B2 (en) |
EP (1) | EP3940416B1 (en) |
KR (1) | KR20220008734A (en) |
CN (1) | CN113933816A (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11867814B1 (en) * | 2022-12-06 | 2024-01-09 | Aeva, Inc. | Techniques for driving a laser diode in a LIDAR system |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5563701A (en) * | 1993-12-29 | 1996-10-08 | Hyundai Electronics Industries Co., Ltd. | Apparatus and method for measuring distance optically using phase variation |
US20160313445A1 (en) * | 2012-03-16 | 2016-10-27 | Advanced Scientific Concepts, Inc. | Personal ladar sensor |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10207610A1 (en) | 2002-02-22 | 2003-09-25 | Rudolf Schwarte | Method and device for recording and processing electrical and optical signals |
US7986397B1 (en) | 2008-04-30 | 2011-07-26 | Lockheed Martin Coherent Technologies, Inc. | FMCW 3-D LADAR imaging systems and methods with reduced Doppler sensitivity |
EP3589974A2 (en) | 2017-03-01 | 2020-01-08 | Pointcloud Inc. | Modular three-dimensional optical sensing system |
CA3075736A1 (en) * | 2017-09-15 | 2019-11-14 | Aeye, Inc. | Intelligent ladar system with low latency motion planning updates |
US10401480B1 (en) | 2018-12-05 | 2019-09-03 | Luminar Technologies, Inc. | Lidar receiver with multiple detectors for range-ambiguity mitigation |
-
2020
- 2020-09-03 US US17/011,813 patent/US11988779B2/en active Active
-
2021
- 2021-03-03 KR KR1020210027986A patent/KR20220008734A/en active Search and Examination
- 2021-05-10 EP EP21173088.2A patent/EP3940416B1/en active Active
- 2021-05-17 CN CN202110532406.4A patent/CN113933816A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5563701A (en) * | 1993-12-29 | 1996-10-08 | Hyundai Electronics Industries Co., Ltd. | Apparatus and method for measuring distance optically using phase variation |
US20160313445A1 (en) * | 2012-03-16 | 2016-10-27 | Advanced Scientific Concepts, Inc. | Personal ladar sensor |
Also Published As
Publication number | Publication date |
---|---|
US11988779B2 (en) | 2024-05-21 |
KR20220008734A (en) | 2022-01-21 |
CN113933816A (en) | 2022-01-14 |
US20220018945A1 (en) | 2022-01-20 |
EP3940416B1 (en) | 2023-10-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102477195B1 (en) | Method and system for doppler detection and doppler correction of optical chirped range detection | |
KR102441163B1 (en) | Method and system for pitch-catch scanning of coherent lidar | |
CN102460209B (en) | Distance-measuring device | |
US11428786B2 (en) | Dual waveforms for three-dimensional imaging systems and methods thereof | |
JP2023103334A (en) | Providing spatial displacement of transmit and receive modes in lidar system | |
EP3679394B1 (en) | Ladar system supporting doublet waveform for sequential in-phase (i) and quadrature (q) processing | |
CN109116322A (en) | A kind of displacement and the light echo removing method apart from laser radar system | |
EP3940416A1 (en) | Readout architecture for fmcw lidar | |
US11187523B1 (en) | Three-dimensional shape estimation of objects via polarimetry | |
KR20210087349A (en) | LIDAR device and operating method for the same | |
US11971505B2 (en) | Methods and devices for peak signal detection | |
US20230204780A1 (en) | Lidar System Having A Shared Clock Source, And Methods Of Controlling Signal Processing Components Using The Same | |
US11513229B2 (en) | Multi-beam processing of lidar vibration signals | |
CN111527419A (en) | Sampling circuit, sampling method, distance measuring device and mobile platform | |
EP3835720B1 (en) | Method for multipath error compensation and multipath error-compensated indirect time of flight range calculation apparatus | |
KR20220112097A (en) | Lidar device and operating mothod for the same | |
US20220113414A1 (en) | Distance measuring device and method for measuring distance by using the same | |
US20240085537A1 (en) | Lidar device and operating method thereof | |
US20230213619A1 (en) | Lidar system having a linear focal plane, and related methods and apparatus | |
US20230204730A1 (en) | Multi-range lidar systems and methods | |
US20230213618A1 (en) | Lidar system having a linear focal plane, and related methods and apparatus | |
EP4206727A1 (en) | Lidar device and operating method thereof | |
US20230367014A1 (en) | Beam steering techniques for correcting scan line compression in lidar devices | |
KR20240109480A (en) | LiDAR device, and receiver and operating method therefor | |
CN115902922B (en) | Doppler laser radar based on electric frequency domain differential frequency discrimination and measuring method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20211122 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
INTG | Intention to grant announced |
Effective date: 20230503 |
|
P01 | Opt-out of the competence of the unified patent court (upc) registered |
Effective date: 20230520 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 602021005545 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: FP |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG9D |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: MK05 Ref document number: 1618284 Country of ref document: AT Kind code of ref document: T Effective date: 20231004 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240105 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240204 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: NL Payment date: 20240315 Year of fee payment: 4 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240204 Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240105 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240104 Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240205 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: RS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: NO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20240104 Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 602021005545 Country of ref document: DE |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20240313 Year of fee payment: 4 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SM Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20231004 |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20240705 |