WO2022159110A1 - Allongement d'impulsion lidar - Google Patents

Allongement d'impulsion lidar Download PDF

Info

Publication number
WO2022159110A1
WO2022159110A1 PCT/US2021/014865 US2021014865W WO2022159110A1 WO 2022159110 A1 WO2022159110 A1 WO 2022159110A1 US 2021014865 W US2021014865 W US 2021014865W WO 2022159110 A1 WO2022159110 A1 WO 2022159110A1
Authority
WO
WIPO (PCT)
Prior art keywords
pulse
pulse width
light
environment
determined
Prior art date
Application number
PCT/US2021/014865
Other languages
English (en)
Inventor
Luke Wachter
Pierre-Yves Droz
Clayton KUNZ
Original Assignee
Waymo Llc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Waymo Llc filed Critical Waymo Llc
Priority to PCT/US2021/014865 priority Critical patent/WO2022159110A1/fr
Publication of WO2022159110A1 publication Critical patent/WO2022159110A1/fr

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • G01S17/08Systems determining position data of a target for measuring distance only
    • G01S17/10Systems determining position data of a target for measuring distance only using transmission of interrupted, pulse-modulated waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/93Lidar systems specially adapted for specific applications for anti-collision purposes
    • G01S17/931Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/4802Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section

Definitions

  • This disclosure relates generally to LIDAR devices and systems, and more specifically to identifying features of an environment based, at least in part, on pulse elongation.
  • Vehicles may be configured to operate in an autonomous or semi- autonomous mode and navigate through an environment with little or no input from a driver.
  • These autonomous and semi-autonomous vehicles typically include a number of sensors configured to determine information about the environment in which the vehicle operates.
  • the sensors may include one or more light detection and ranging (LIDAR) devices that can detect multiple objects (such as other vehicles, pedestrians, traffic signals, obstacles, and so on) in the environment and determine distances between the autonomous vehicle and the multiple objects.
  • LIDAR light detection and ranging
  • Autonomous and semi-autonomous vehicles may also include other types of sensors including, for example, sonar devices, radar devices, cameras, and audio sensing devices. Data from one or more of these sensors may be used by a controller or computing device to determine various features and characteristics of the detected objects such as, for example, position, size, shape, type, movement, and orientation.
  • a LIDAR system which may include one or more LIDAR devices and one or more controllers or computing devices, may estimate distances to a number of objects and other features while scanning an environment to generate a point cloud of reflective surfaces in the environment.
  • Individual points in the point cloud may be determined by emitting light pulses into the environment and detecting return light pulses that were reflected from objects and other reflective surface in the environment. Timing information of the emitted light pulses and the return light pulses may provide continuous real-time or near real-time estimates of distances between a LIDAR device and objects detected in the environment, and characteristics of the return light pulses may be used to determine relative orientations of the detected objects.
  • the estimated distances and determined orientations of the detected objects may be used to generate a three-dimensional map of points indicative of locations of reflective surfaces in the environment.
  • the three-dimensional map may be represented by or associated with the point cloud.
  • the method may include emitting a light pulse into the environment, receiving a return light pulse corresponding to reflection of the emitted light pulse by a surface in the environment, determining a pulse width of the received light pulse, comparing the determined pulse width with a reference pulse width, determining an amount of pulse elongation of the received light pulse based on the comparing, and classifying the surface as either an object to be avoided or as air particulates associated with the volumetric medium based, at least in part, on the determined amount of pulse elongation.
  • the reference pulse width may be based, at least in part, on one or more reference light pulses reflected from a surface orthogonal to a direction of the emitted light pulses.
  • the volumetric medium may include air particulates from which photons associated with return light pulses are scattered.
  • the air particulates may include at least one of exhaust, aerosol, dust, pollution, smoke, rain, or fog.
  • the determined pulse width may be compared with the reference pulse width by determining a peak amplitude of the received light pulse, selecting one of a plurality of reference pulse widths based on the determined peak amplitude, and comparing the determined pulse width with the selected reference pulse width.
  • the method may include determining a peak amplitude of the received light pulse, and retrieving the reference pulse width from a look-up table based on the determined peak amplitude.
  • the look-up table may store a plurality of reference pulse widths each associated with a corresponding one of a plurality of different peak amplitudes.
  • the method may also include generating a corresponding point in a point cloud indicative of objects detected in the environment based on at least one of the classification of the surface and the determined amount of pulse elongation.
  • the method may include comparing the determined pulse width with a pulse width of light reflected from a road surface to determine a pulse elongation difference value, and identifying the surface as debris on the road surface based, at least in part, on the determined pulse elongation difference value.
  • the LIDAR system may include a light emitter configured to emit a light pulse into the environment, a photodetector configured to receive a return light pulse corresponding to reflection of the emitted light pulse by a surface in the environment, a number of processing channels, and a controller.
  • the processing channels may be configured to determine a pulse width of the received light pulse, to compare the determined pulse width with a reference pulse width, and to determine an amount of pulse elongation of the received light pulse based on the comparison.
  • the controller may be configured to classify the surface as either an object to be avoided or as air particulates associated with the volumetric medium based, at least in part, on the determined amount of pulse elongation.
  • the volumetric medium may include air particulates from which photons associated with return light pulses are scattered.
  • the air particulates may include at least one of exhaust, aerosol, dust, pollution, smoke, rain, or fog.
  • the determined pulse width may be compared with the reference pulse width by determining a peak amplitude of the received light pulse, selecting one of a plurality of reference pulse widths based on the determined peak amplitude, and comparing the determined pulse width with the selected reference pulse width.
  • the LIDAR system may determine the peak amplitude of the received light pulse, and may retrieve the reference pulse width from a look-up table based on the determined peak amplitude.
  • the look-up table may store a plurality of reference pulse widths each associated with a corresponding one of a plurality of different peak amplitudes.
  • the LIDAR system may also include circuitry configured to generate a corresponding point in a point cloud indicative of objects detected in the environment based on at least one of the classification of the surface and the determined amount of pulse elongation.
  • the processing channels may be configured to compare the determined pulse width with a pulse width of light reflected from a road surface to determine a pulse elongation difference value, and the controller may be further configured to identify the surface as debris on the road surface based, at least in part, on the determined pulse elongation difference value.
  • the LIDAR system may include means for emitting a light pulse into the environment, means for receiving a return light pulse corresponding to reflection of the emitted light pulse by a surface in the environment, means for measuring a pulse width of the received light pulse, means for comparing the measured pulse width with a reference pulse width, means for determining an amount of pulse elongation of the received light pulse based on the comparing, and means for classifying the surface as either an object to be avoided or as air particulates associated with the volumetric medium based, at least in part, on the determined amount of pulse elongation.
  • the volumetric medium may include air particulates from which photons associated with return light pulses are scattered.
  • the air particulates may include at least one of exhaust, aerosol, dust, pollution, smoke, rain, or fog.
  • the LIDAR system may include means for determining a peak amplitude of the received light pulse and means for retrieving the reference pulse width from a look-up table based on the determined peak amplitude, wherein the look-up table stores a plurality of reference pulse widths each associated with a corresponding one of a plurality of different peak amplitudes of respective reference light pulses.
  • the LIDAR system may also include means for emitting a plurality of light pulses during a calibration operation, means for receiving a plurality of reference light pulses reflected by an orthogonal surface relative to a direction of the emitted light pulse, means for determining a reference pulse width for each of the plurality of received reference light pulses, and means for storing the reference pulse widths in a memory associated with the LIDAR system.
  • the LIDAR system may also include means for generating a corresponding point in a point cloud indicative of objects detected in the environment based on at least one of the classification of the surface and the determined amount of pulse elongation.
  • the LIDAR system may include means for comparing the determined pulse width with a pulse width of light reflected from a road surface to determine a pulse elongation difference value, and means for identifying the surface as debris on the road surface based, at least in part, on the determined pulse elongation difference value.
  • the vehicle may include a light emitter configured to emit a light pulse into the environment, a photodetector configured to receive a return light pulse corresponding to reflection of the emitted light pulse by a surface in the environment, at least one processing channel, and a controller.
  • the at least one processing channel may be configured to determine a pulse width of the received light pulse, to compare the determined pulse width with a reference pulse width, and to determine an amount of pulse elongation of the received light pulse based on the comparison.
  • the controller may be configured to classify the surface as either an object to be avoided by the vehicle or as air particulates associated with the volumetric medium based, at least in part, on the determined amount of pulse elongation.
  • the volumetric medium may include air particulates from which photons associated with return light pulses are scattered.
  • the air particulates may include at least one of exhaust, aerosol, dust, pollution, smoke, rain, or fog.
  • the at least one processing channel may be further configured to determine a peak amplitude of the received light pulse, and to retrieve the reference pulse width from a look-up table based on the determined peak amplitude, wherein the look-up table stores a plurality of reference pulse widths each associated with a corresponding one of a plurality of different peak amplitudes of respective reference light pulses.
  • the at least one processing channel may be further configured to compare the determined pulse width with a pulse width of light reflected from a road surface to determine a pulse elongation difference value, and the controller may be further configured to identify the surface as debris on the road surface based, at least in part, on the determined pulse elongation difference value.
  • the controller may also be configured to generate a corresponding point in a point cloud indicative of objects detected in the environment based on at least one of the classification of the surface and the determined amount of pulse elongation.
  • the method may include emitting a light pulse into the environment associated with a LIDAR device; receiving a return light pulse corresponding to reflection of the emitted light pulse by a surface in the environment; determining a pulse width of the received light pulse; comparing the determined pulse width with a reference pulse width; determining an amount of pulse elongation of the received light pulse based on the comparing; and classifying the surface as road debris to be avoided by a vehicle based, at least in part, on the determined amount of pulse elongation.
  • the road debris may be pieces of a tire, for example.
  • the determined pulse elongation may differ from a pulse elongation associated with the road.
  • the pulse elongation associated with the road may be an expected pulse elongation based on previously obtained data such as from one or more of the following: the same sensor, other sensors, sensors on another car, and map data.
  • Figure 1 shows a block diagram of an example LIDAR system.
  • Figure 2A shows a LIDAR device in an environment.
  • Figure 2B shows a timing diagram of waveforms corresponding to emitted light pulses and received light pulses of a LIDAR device.
  • Figure 3 A shows an illustration depicting light pulses reflected by an orthogonal surface.
  • Figure 3B shows an illustration depicting light pulses reflected by an inclined surface.
  • Figure 3C shows an illustration depicting light pulses reflected in an environment containing a volumetric medium.
  • Figure 3D shows a graph depicting a light pulse including a number of individual reflected light pulses.
  • Figure 4A shows a block diagram of an example detector circuit suitable for use in a LIDAR device.
  • Figure 4B shows a block diagram of an example processing channel suitable for use in a LIDAR detector or a LIDAR device.
  • Figure 5 shows a graph depicting the pulse widths of a number of light pulses relative to an example relationship between amplitude and pulse width of a reference light pulse.
  • Figure 6 shows a graph depicting an example relationship between pulse width and peak amplitude of a number of light pulses.
  • Figure 7A shows a flow chart depicting an example operation for identifying a presence of a volumetric medium in an environment.
  • Figure 7B shows a flow chart depicting an example operation for identifying debris on a road surface.
  • Figure 8 A shows a flow chart depicting an example operation for comparing a determined pulse width of a received light pulse with a reference pulse width.
  • Figure 8B shows a flow chart depicting an example operation for determining the pulse width of a received light pulse.
  • Figure 8C shows a flow chart depicting an example operation for retrieving a reference pulse width from memory.
  • aspects of the present disclosure may be implemented within other vehicles, such as cars, trucks, motorcycles, buses, boats, airplanes, helicopters, robots, unmanned aerial vehicles, lawn mowers, recreational vehicles, amusement park vehicles, farm equipment, construction equipment, trams, golf carts, trains, and trolleys. Other vehicles are possible as well. Aspects of the present disclosure may also be implemented on stationary objects or structures (such as walls, buildings, roofs, fences, and the like) as part of security systems, inventory management systems, and other suitable monitoring systems.
  • a LIDAR device may emit or transmit light pulses into an environment, and may receive light pulses reflected by objects and other reflective surfaces in the environment to determine information about the objects and the other reflective surfaces.
  • the distance to a detected object may be determined based on a time between emission of a light pulse by the LIDAR device and reception of a corresponding reflected light pulse by the LIDAR device.
  • the size, shape, orientation, texture, and other features of the detected object may be determined based (at least in part) on the amplitudes, pulse widths, timing information, and other characteristics of a multitude of such received light pulses.
  • Information generated from the received light pulses may be used to generate a point cloud (or another suitable three-dimensional map of points) indicating the location, size, shape, movement, orientation, and other features of objects and other reflective surfaces detected in the environment.
  • An autonomous or semi-autonomous vehicle may use one or more of these point clouds to navigate through the environment without human input, and thus the accuracy with which these point clouds represent objects and other surfaces in the environment may be important for safe, comfortable, or other desirable operation of the vehicle.
  • the shape of light pulses reflected by a surface may be indicative of a number of features or characteristics of a corresponding object (such as size, shape, and orientation of the object).
  • the shape of a light pulse reflected by a surface orthogonal to the direction of the emitted light pulse may be the same (or nearly the same) as the shape of a corresponding emitted light pulse that illuminated the orthogonal surface
  • the shape of a light pulse reflected by a non-orthogonal surface (such as an inclined surface relative to the direction of the emitted light pulse) may be detectably different than the shape of the corresponding emitted light pulse that illuminated the non-orthogonal surface.
  • photons associated with the emitted light pulse may be reflected from different portions of the non-orthogonal surface and arrive at the LIDAR device at different times, thereby spreading the reflected light in time such that light pulses received by the LIDAR device are elongated relative to the emitted light pulses that illuminated the non-orthogonal surface.
  • the spreading of a reflected light pulse relative to a corresponding emitted light pulse may be indicative of the features or characteristics of an object detected by the LIDAR device. For example, as the angle of inclination of a reflective surface relative to the direction of the emitted light pulse increases, the range of distances that photons associated with a reflected light pulse travel to the LIDAR device increases and thereby elongates the pulse width of the reflected light pulse relative to the corresponding emitted light pulse.
  • LIDAR systems disclosed herein may use the amount of pulse elongation exhibited by reflected light pulses to determine the shapes, orientations, and other features of objects detected in a surrounding environment.
  • LIDAR systems disclosed herein may also use amounts of pulse elongation exhibited by reflected light pulses to identify debris and other small objects on a road or street.
  • a piece of tire or other debris on the road may be difficult to detect by a conventional LIDAR device because of the relatively small geometrical aspects of the piece of tire (as compared to the ground).
  • LIDAR systems disclosed herein may use the pulse elongation of received light pulses to distinguish between light pulses reflected by the piece of tire and light pulses reflected by the road.
  • Pulse elongation may also be caused by air particulates associated with a volumetric medium in the environment.
  • the air particulates may reflect different portions of the light pulse as many relatively small light pulses.
  • these relatively small light pulses may combine or constructively interfere with each other to form a relatively large light pulse that can be detected by the LIDAR device and misinterpreted as a light pulse reflected from a non-orthogonal surface (rather than from air particulates associated with a volumetric medium).
  • a LIDAR device configured to detect the relatively large light pulse but not the relatively small light pulses may not be able to distinguish between light pulses reflected by objects to be avoided by a vehicle and light pulses reflected by a volumetric medium through which the vehicle may traverse (such as pollution).
  • a volumetric medium through which the vehicle may traverse (such as pollution).
  • air particulates associated with the volumetric medium may cause errors in the perception of objects and other reflective surfaces in the environment if not properly identified and classified.
  • Implementations of the subject matter described herein may use a LIDAR device to detect or identify a presence of a volumetric medium in an environment, which may increase the accuracy with which objects and other reflective surfaces in the environment can be perceived by a controller or computing device associated with the LIDAR device.
  • a LIDAR system may include one or more LIDAR devices and one or more associated LIDAR controllers or computing devices. The LIDAR system may determine the pulse width of a light pulse reflected by a surface within the environment, may compare the determined pulse width with a reference pulse width, and may determine the amount of pulse elongation based on the comparison.
  • the amount of pulse elongation may be used to classify the surface as either an object to be avoided, e.g., by an autonomous vehicle, or as air particulates associated with a volumetric medium.
  • the amount of pulse elongation may be considered when determining a corresponding point in a point cloud indicative of objects detected in the environment. In this manner, the accuracy with which the LIDAR system may determine various features or characteristics of objects and other reflective surfaces detected in the environment may be improved over conventional LIDAR systems or devices.
  • a reference pulse width may be determined for each of a plurality of amplitude values of light pulses.
  • Each of the determined reference pulse widths may be stored, along with its amplitude and pulse characteristics, in a look-up table (LUT) or other memory provided within or otherwise accessible by the LIDAR device or system.
  • the LIDAR system may retrieve the reference pulse width and/or the pulse characteristics (such as pulse shape) corresponding to the amplitude of the received light pulse, and use the retrieved pulse width and/or pulse characteristics to determine an amount of pulse elongation of the received light pulse caused by air particulates associated with the volumetric medium.
  • the term “object” as used herein may include any physical object for which it may be desirable to detect and determine its location, speed, movement, size, shape, and other features, e.g., other vehicles, pedestrians, cyclists, signs, etc.
  • the term “volumetric medium” as used herein may include air particulates that may alter one or more characteristics of the reflected light pulses received by a LIDAR device (such as by elongating the reflected light pulses relative to their corresponding emitted light pulses). Air particulates that form a volumetric medium may include (but are not limited to) exhaust, aerosol, dust, pollution, smoke, rain, fog, or any combination thereof.
  • the detected objects may include objects and other environmental features to be mapped and/or to be avoided by an autonomous or semi-autonomous vehicle.
  • Each of the buses may alternatively be a single signal line, and each of the single signal lines may alternatively be buses, and a single line or bus might represent any one or more of a myriad of physical or logical mechanisms for communication between components.
  • the term “coupled” as used herein means coupled directly to or coupled through one or more intervening components or circuits.
  • FIG 1 shows a block diagram of an example light detection and ranging (LIDAR) system 100.
  • the LIDAR system 100 may be used to detect points or objects in an environment by emitting pulses of lights that illuminate the points or objects and then detecting light pulses reflected from the points or objects.
  • the LIDAR system 100 may determine the distance to a selected object based on the time delay between the emission of a light pulse and the reception of a corresponding light pulse reflected from the selected object. This time delay, which may also be referred to as the “time-of-flighf ’ or the round-trip time of the light pulse, may be multiplied by the speed of light to determine the distance between the LIDAR system 100 and the selected object.
  • the LIDAR system 100 may scan its surrounding environment and use distance information to generate a map of the environment.
  • the LIDAR system 100 may be used to control an autonomous or semi- autonomous vehicle, for example, so that the autonomous or semi-autonomous vehicle can navigate the environment to reach a destination while avoiding obstacles.
  • the LIDAR system 100 is shown to include a transmitter 110, a receiver 120, and a LIDAR controller 130.
  • the transmitter 110 and the receiver 120 may form (or be associated with) a LIDAR device 101.
  • the transmitter 110 may include a transmit controller 111, a number of light emitters 112, and a transmit aperture 113.
  • the light emitters 112 may emit one or more light pulses 125 that can be used to detect objects in a surrounding environment.
  • the light emitters 112 may include any number of suitable light sources such as (but not limited to) laser diodes, light emitting diodes (LEDs), vertical cavity surface emitting lasers (VCSELs), organic light emitting diodes (OLEDs), polymer light emitting diodes (PLEDs), light emitting polymers (LEPs), liquid crystal displays (LCDs), microelectromechanical systems (MEMS), or any other device configured to selectively transmit or emit light pulses 125 at a source wavelength.
  • the source wavelength may include, for example, the ultraviolet, visible, and/or infrared portions of the electromagnetic spectrum.
  • the light emitters 112 may be disposed on one or more substrates (such as printed circuit boards (PCB), flexible PCBs, and the like).
  • the light emitters 112 are described herein as emitting light pulses 115, one of ordinary skill in the art will readily understand that the light emitters 112 may transmit or emit light signals, light beams, photons, and the like. Thus, the terms light pulses, light signals, light beams, and photons may be used interchangeably herein.
  • the transmit aperture 113 is coupled to the light emitters 112, and may include any suitable components (such as mirrors, lenses, diffraction gratings, exit apertures, and the like) that can focus, direct, and/or condition the light pulses 115 for emission into the surrounding environment.
  • the transmit aperture 113 may be configured to steer the light pulses 115 in one or more specified directions relative to the LIDAR device 101.
  • the specified directions may span a range of directions, for example, so that distances between the LIDAR device 101 and a number of objects (such as cars, people, roads, traffic signals, traffic signs, obstacles, and so on) may be determined based on reflections of the light pulses 115 caused by the objects.
  • the transmit controller 111 may control operations of the light emitters 112 and the transmit aperture 113, may adjust a number of parameters or settings of the light emitters 112 and the transmit aperture 113, or both.
  • the transmit controller 111 may be responsive to one or more control signals provided by the LIDAR controller 130.
  • the transmit controller 111 may adjust the pulse width of the light pulses 115 emitted by the light emitters 112 based on the one or more control signals.
  • the transmit controller 111 may be omitted or may be included within the LIDAR controller 130.
  • the receiver 120 may include a number of photodetectors 121, a detector circuit 122, and an analog-to-digital converter (ADC) 123.
  • the photodetectors 121 may receive light pulses 125 (such as photons) from the surrounding environment.
  • the received light pulses 125 may include components of the emitted light pulses 115 reflected from one or more objects in the surrounding environment.
  • the received light pulses 125 may also include components of light that traversed a volumetric medium.
  • the photodetectors 121 may be configured to convert the received light pulses 125 into photodetector signals (such as analog current signals) indicative of intensity levels and/or pulse widths of the received light pulses 125.
  • the photodetectors 121 may be any suitable component or device that can receive or sense light including, for example, photodiodes, avalanche photodiodes, Silicon Photomultipliers (SiPMs), phototransistors, cameras, active pixel sensors (APS), charge coupled devices (CCDs), cryogenic detectors, or the like.
  • the photodetectors 121 may be reverse-biased photodiodes that generate a current in response to receiving light pulses, for example, such that the amount of current through each photodiode is proportional to the intensity of light pulses received by the photodiode.
  • the receiver 120 may include optics (e.g., lenses, mirrors, diffraction gratings) configured to emit light toward/receive light from a specific direction and/or to focus such light, to filter out one or more wavelengths, bands of wavelengths, polarizations, or other specified properties of such light, or to otherwise interact with or modify such light so that the photodetectors 121 primarily receive light corresponding to the wavelength of the light pulses 115 emitted by the transmitter 110 (and receive minimal light corresponding to other wavelengths).
  • optics e.g., lenses, mirrors, diffraction gratings
  • optics of the LIDAR device 101 may be configured to focus and/or collimate light emitted from the transmitter 110 into one or more respective beams of light directed in respective directions into an environment of the LIDAR device 101.
  • the optics may be configured to focus light reflected from respective regions of the environment located in respective directions from the LIDAR device 101.
  • the detector circuit 122 may use any suitable technique to sample photodetector signals provided by the photodetectors 121 to determine intensity levels and/or one or more characteristics of the received light pulses 125. In some implementations, the detector circuit 122 may sample the photodetector signals at a number of intervals or sampling times. In other implementations, the detector circuit 122 may continuously sample the photodetector signals. The detector circuit 122 may provide the determined intensity levels and characteristics of the received light pulses 125 to the ADC 123, for example, as analog signals containing magnitude information and timing information that may be used to generate a point cloud of a surrounding environment. In some implementations, the detector circuit 122 may amplify and/or filter the photodetector signals.
  • the detector circuit 122 may also include a look-up table (LUT) 124 storing reference pulse width information that may be used to determine an amount of pulse elongation of the received light pulses 125.
  • the LUT 124 may store a reference pulse width for each of a plurality of different amplitudes of light pulses.
  • each of the reference pulse widths stored in the LUT 124 may be indicative of an expected pulse width of a received light pulse 125 having a corresponding amplitude.
  • the detector circuit 122 may determine an amplitude of a received light pulse 125 and use the determined amplitude to retrieve a corresponding reference pulse width from the LUT 124.
  • the detector circuit 122 may compare the determined pulse width with the retrieved reference pulse width to determine an amount of pulse elongation of the received light pulse (which may be caused by the volumetric medium).
  • the ADC 123 may receive analog signals indicating intensity levels and/or characteristics of the received light pulses 125 from the detector circuit 122, and may convert the analog signals into digital data that can be processed by the LIDAR controller 130.
  • the ADC 123 may be any suitable ADC such as (but not limited to) a flash ADC, a successive- approximation-register (SAR) ADC, or a delta-sigma ADC.
  • each photodetector 121 may correspond to a respective ADC.
  • a plurality of photodetectors 121 may correspond to a single ADC (such as to reduce the size, cost, and/or power consumption of the LIDAR device 101).
  • the ADC 123 may be omitted.
  • the LIDAR controller 130 may include a processor 131, a memory 132, and a digital signal processor (DSP) 133.
  • the DSP 133 may process digital data provided by the ADC 123 to determine intensity information, range information, and pulse elongation information of light pulses received by any number of the photodetectors 121.
  • the determined intensity, range, and/or pulse elongation information may be used to determine the size, shape, location, movement, and orientation of a number of detected objects in the surrounding environment.
  • the DSP 133 may receive information indicative of the amount of pulse elongation of the received light pulse 125 (which may be caused by the volumetric medium), and may use the indicated amount of pulse elongation to classify the detected objects as either objects to be avoided (such as by an autonomous or semi-autonomous vehicle) or as air particulates associated with the volumetric medium.
  • the processor 131 may be any suitable one or more processors capable of executing scripts or instructions of one or more software programs stored in the LIDAR system 100 (e.g., within the memory 132).
  • the processor 131 may include one or more microprocessors and memory providing at least a portion of machine- readable media within which program instructions or scripts can be stored.
  • the processor 131 may be an Application Specific Integrated Circuit (ASIC).
  • the processor 131 may be or include one or more Field Programmable Gate Arrays (FPGAs) or Programmable Logic Devices (PLDs).
  • FPGAs Field Programmable Gate Arrays
  • PLDs Programmable Logic Devices
  • the memory 132 may store information pertaining to the transmitter 110, the receiver 120, the surrounding environment, or any combination thereof.
  • the memory 132 may also include a non-transitory computer-readable medium (e.g., one or more nonvolatile memory elements, such as EPROM, EEPROM, Flash memory, a hard drive, and so on) that may store a number of software programs each including instructions or scripts that, when executed by the processor 131, causes the LIDAR system 100 to perform all or a portion of the operations described with respect to Figures 7A-7B and 8A-8C.
  • the memory 132 may also store reference pulse width information that may be used to determine an amount of pulse elongation of the received light pulses 125 (which may be caused by a volumetric medium in the environment).
  • FIG. 2A illustrates the LIDAR device 101 in an example environment 200.
  • the example environment 200 is shown to include an automobile 201 and an overhang 202.
  • the LIDAR device 101 which may be attached to or otherwise associated with an autonomous or semi-autonomous vehicle (not shown for simplicity), is shown to include three light emitters 112A-112C and three photodetectors 121A-121C (other components of the LIDAR device 101 are not shown for simplicity).
  • the three light emitters 112A-112C may be configured to emit respective light pulses 115A-115C into the environment 200, and the three photodetectors 121A-121C may be configured to detect or receive respective light pulses 125A-125C reflected from objects in the environment 200.
  • a first light pulse 115A illuminates a portion 205 A of the overhang 202, and a first photodetector 121 A receives a corresponding reflected light pulse 125 A.
  • a second light pulse 115B illuminates a portion 205B of the automobile 201, and a second photodetector 121B receives a corresponding reflected light pulse 125B.
  • a third light pulse 115C illuminates another portion 205C of the automobile 201, and a third photodetector 121C receives a corresponding reflected light pulse 125C.
  • the LIDAR system 100 may use one or more properties of the received light pulses 125A-125C (e.g., timing, amplitude, pulse width, and so on) to determine the distance between the LIDAR device 101 and each of the portions 205A-205C in the environment 200.
  • properties of the received light pulses 125A-125C e.g., timing, amplitude, pulse width, and so on
  • Figure 2B shows an example timing diagram 210 of waveforms corresponding to emitted light pulses and received light pulses of a LIDAR device.
  • Transmit waveforms 215A-215C may be indicative of intensity levels of respective light pulses 11 SA- 115C emitted from the LIDAR device 101 of Figure 2A
  • receive waveforms 225A-225C may be indicative of intensity levels of respective light pulses 125A-125C received by the LIDAR device 101 of Figure 2A.
  • the light pulses 115A-115C are emitted from the LIDAR device 101 at the same time to (or at least substantially the same time), and the reflected light pulses 125A-125C are received by the LIDAR device 101 at different times tA-tc (e.g., due to different distances between the LIDAR device 101 and each of the portions 205A-205C of the environment 200).
  • the transmit waveforms 215A-215C include respective pulses 216A- 216C that represent the time to at which corresponding light pulses 115A-115C are emitted from the LIDAR device 100.
  • the receive waveforms 225A-225C include respective pulses 216A-216C that represent the times tA-tc at which corresponding reflected light pulses 125A-125C are received by the LIDAR device 101.
  • the reception times tA-tc may be determined from respective pulses 226A-226C using any suitable peak detection technique (e.g., determining a peak amplitude, determining a centroid, determining a mean time between threshold crossings, and the like).
  • the determined reception times tA-tc may be used to determine distances between the LIDAR device 101 and respective portions 205 A- 205C of the environment 200.
  • a LIDAR system it is important for a LIDAR system to distinguish between obstacles to be avoided by an autonomous vehicle and a volumetric medium of air particulates through which the autonomous vehicle may traverse.
  • obstacles to be avoided by the autonomous vehicle may include cars, trucks, cyclists, pedestrians, fallen trees, and the like, and some non-limiting examples of air particulates through which the autonomous vehicle may traverse may include exhaust, aerosol, dust, pollution, smoke, rain, fog, and the like.
  • a LIDAR system may detect objects in a surrounding environment by emitting (or transmitting) a plurality of light pulses into the environment, receiving return light pulses reflected by objects in the environment, and generating a point cloud (or some other suitable map of points) indicating the locations, movements, orientations, features, and other characteristics of the detected objects.
  • the point cloud may include a collection of depth-map points or pixels each representing a distance between a LIDAR device and a corresponding point or portion of an object from which an emitted light pulse was reflected and received as a return light pulse by the LIDAR device.
  • the collection of depth-map points that form the point cloud may be analyzed (or otherwise processed) to detect a presence of objects in the environment, to determine distances between the LIDAR device and the detected objects, to determine movement of the detected objects, to determine a number of features of the detected objects, to assist in the operation of autonomous vehicles, or any combination thereof. Because the autonomous vehicle may use one or more of these point clouds to navigate through the environment without human input, the accuracy with which the LIDAR system can detect the objects and determine their locations, movements, orientations, features, and other characteristics may be critical for the safe, comfortable, or otherwise desirable operation of an autonomous vehicle.
  • the accuracy with which a LIDAR system may determine the location, movement, orientation, features, and other characteristics of a detected object may be increased by distinguishing between light pulses reflected by an inclined surface (such as a road or street) and light pulses reflected by air particulates associated with a volumetric medium.
  • the amount of pulse elongation of reflected or return light pulses relative to their corresponding emitted light pulses may be used to determine the features and characteristics of objects detected by the LIDAR device.
  • the shape of a light pulse reflected from a surface that is orthogonal to the direction of a corresponding light pulse emitted from a LIDAR device may be the same as (or at least similar to) the shape of the corresponding emitted light pulse
  • the shape of a light pulse reflected from a surface that is inclined relative to the direction of a corresponding light pulse emitted by the LIDAR device may be different than (such as elongated as compared to) the shape of the corresponding emitted light pulse.
  • Figure 3A shows an illustration 300A depicting light pulses reflected by a reflective surface 320.
  • light pulses 311 emitted from a LIDAR device 310 are reflected by the reflective surface 320 as return light pulses 321 and received by the LIDAR device 310.
  • the reflective surface 320 is orthogonal to the direction of the emitted light pulses 311 such that the angle of incidence of the emitted light pulses 311 is the same (or at least substantially the same) as the angle of reflection of the return light pulses 321.
  • the shapes of the return light pulses 321 are the same (or substantially the same) as the shapes of the corresponding emitted light pulses 311.
  • the shape of a light pulse may be described using its rising edge timing, falling edge timing, peak amplitude, pulse width, or any combination thereof.
  • a first illustrative graph 312 depicts the shape of a return light pulse 321 received by the LIDAR device 310
  • a second illustrative graph 322 depicts the shape of a reference light pulse 325.
  • the pulse width (PWRX) of the received light pulse 321 may refer to a period of time during which the amplitude of the received light pulse 321 exceeds a threshold value ATH
  • the pulse width (PWREF) of the reference light pulse 325 may refer to a period of time during which the amplitude of the reference light pulse 325 exceeds the threshold value ATH.
  • the pulse width (PWRX) of the received light pulse 321 is the same (or at least substantially the same) as the reference pulse width (PWREF), which may indicate that the received light pulses 321 exhibit little (if any) pulse elongation relative to the corresponding emitted light pulses 311.
  • the LIDAR device 310 may sample portions of a respective light pulse corresponding to periods of time during which the amplitude of the respective light pulse exceeds a threshold value, and may store (at least temporarily) the sampled portions of the respective light pulse.
  • the LIDAR device 310 may trigger or enable sampling of the respective light pulse when its amplitude rises above a first threshold value, and may terminate or disable the sampling of the respective light pulse when its amplitude falls below a second threshold value.
  • the first and second threshold values may be any suitable values that allow the LIDAR device 310 to reduce the amount of data to be analyzed or processed by an associated controller or computing device (such as the LIDAR controller 130 of Figure 1) without reducing accuracy by more than a certain amount.
  • the first and second threshold values may be the same. In other aspects, the first and second threshold values may be different. In other implementations, the LIDAR device 310 may continuously sample the respective light pulse and store (at least temporarily) the entirety of the sampled light pulse.
  • Figure 3B shows an illustration 300B depicting light pulses reflected by another reflective surface 330.
  • light pulses 311 emitted from the LIDAR device 310 are reflected by the reflective surface 330 as return light pulses 331 and received by the LIDAR device 310.
  • the reflective surface 330 is inclined relative to the direction of the emitted light pulses 311 such that the angle of incidence of the emitted light pulses 311 is different than the angle of reflection of the return light pulses 331.
  • the pulse width of a respective return light pulse 331 may be longer than (such as elongated as compared to) the pulse width of the corresponding emitted light pulse 311).
  • a first illustrative graph 332 depicts the shape of a return light pulse 331 received by the LIDAR device 310
  • the second illustrative graph 322 depicts the shape of the reference light pulse 325.
  • the pulse width (PWRX) of the received light pulse 331 is longer than the reference pulse width (PWREF), which may indicate that the received light pulse 331 exhibits a detectable amount of pulse elongation relative to the corresponding emitted light pulse 311.
  • the amount of pulse elongation of the return light pulse 331 relative to the corresponding emitted light pulse 311 may be indicative of one or more features or characteristics of the inclined surface 330 (or other objects that reflect emitted light pulses as return light pulses that can be detected by the LIDAR device 310).
  • Pulse elongation may be caused by other environmental features.
  • light pulses reflected by multiple surfaces of the same object may experience pulse elongation. More specifically, first portions of a light pulse may be reflected from an edge of a first surface of an object, while second portions of the light pulse may be reflected from a second surface of the object that is a different distance from the LIDAR device than the first surface.
  • the first and second reflected portions of the light pulse may combine or overlap with each other to form a composite light pulse having a larger amplitude than either of the first and second reflected portions individually.
  • the composite light pulse may include first and second peaks corresponding to the first and second reflected portions of the light pulse, respectively.
  • the composite light pulse which may be associated with or cause a mixed pixel effect, may be incorrectly interpreted as indicating a presence of two objects in the environment (such as rather than indicating one object having first and second surfaces).
  • pulse elongation may be caused by a presence of a volumetric medium in the environment. More specifically, when a light pulse encounters a volumetric medium, photons may be reflected as the light pulse interacts with small air particulates that form the volumetric medium. These air particulates may cause returns such that the pulse width of the return light pulse is longer than the pulse width of the corresponding emitted light pulse.
  • pulse elongation caused by air particulates or other aspects of a volumetric medium may reduce the accuracy with which a LIDAR system can detect, identify, locate, and classify reflective surfaces in the surrounding environment, for example, because light pulses reflected from such air particulates may be incorrectly interpreted to indicate the presence of objects or obstacles to be avoided by autonomous vehicles (rather than interpreted to indicate the presence of a volumetric medium through which autonomous vehicles may traverse).
  • Figure 3C shows an illustration 300C depicting light pulses reflected in an environment containing a volumetric medium 350. Some portions of the emitted light pulses 311 may be reflected by the reflective surface 320 as first return light pulses 341, and other portions of the emitted light pulses 311 may be reflected by air particulates 351 A and 35 IB within the volumetric medium 350 as second return light pulses 360.
  • the second return light pulses 360 may include or refer to one or more light pulses 360A reflected by the first air particulate(s) 351 A, and may include or refer to one or more light pulses 360B reflected by the second air particulate(s) 35 IB.
  • the dashed lines referring to the one or more light pulses 360 A and the dashed lines referring to the one or more light pulses 360B are separated from the solid line referring to the corresponding emitted light pulse 311 in the depiction of Figure 3C for illustrative purposes only.
  • the direction of travel of the one or more light pulses 360A and the direction of travel of the one or more light pulses 360B are opposite of, yet collinear with, the direction of travel of the corresponding emitted light pulse 311 (e.g., the angle of arrival of the light pulses 360A and 360B may be the same as the angle of departure of the corresponding emitted light pulse 311).
  • air particulates 351 may collectively refer to air particulates 351 A and air particulates 35 IB
  • light pulses 360 may collectively refer to light pulses 360A and light pulses 360B.
  • the volumetric medium 350 may include additional air particulates from which the LIDAR device 310 may receive additional return light pulses 360.
  • the first return light pulses 341 may pass through the volumetric medium 350 and be reflected by reflective surface 320.
  • the second return light pulses 360 may be reflections of photons by the small air particulates that form the volumetric medium 350, and may be relatively weak light pulses as compared to the emitted light pulses 311 and the first return light pulses 341.
  • the relatively weak return light pulses 360 may not be individually detected by the LIDAR device 310 (such as because their peak amplitudes are less than a threshold value), a plurality of the relatively weak light pulses 360 may combine and/or constructively interfere with each other in a manner that results in a relatively strong “composite” light pulse that can be detected by the LIDAR device 310.
  • Figure 3D shows a graph 300D depicting a relatively strong composite light pulse 370 and a number of relatively weak light pulses 360.
  • the relatively weak light pulses 360 may be examples of the return light pulses 360 of Figure 3C.
  • the relatively strong composite light pulse 370 may include or represent the number of relatively weak light pulses 360, and each of the relatively weak light pulses 360 may correspond to a portion of an emitted light pulse reflected from air particulates within the volumetric medium 350.
  • the relatively strong composite light pulse 370 is depicted as having a detected pulse width (PWDET) corresponding to a period of time during which the amplitude of the light pulse 370 exceeds the threshold value ATH.
  • PWDET detected pulse width
  • the relatively strong composite light pulse 370 is elongated relative to the light pulses 311 emitted from the LIDAR device 310 and may also be elongated relative to the return light pulses 341 from the reflective surface 320, for example, such that the pulse width (PWDET) of the relatively strong composite light pulse 370 is longer than the reference pulse width (PWREF).
  • PWDET pulse width of the relatively strong composite light pulse 370
  • the relatively weak light pulses 360 are depicted in Figure 3D as having amplitudes that do not exceed the threshold value ATH, and may not be individually detectable by the LIDAR device 310.
  • the relatively strong composite light pulse 370 is depicted as having an amplitude that exceeds the threshold value ATH, and may be detectable by the LIDAR device 310.
  • the relatively strong composite light pulse 370 may be misinterpreted as a light pulse reflected from a solid object if pulse elongation was not considered, which may cause an associated controller or computing device (such as the LIDAR controller 130 of Figure 1) to incorrectly classify the volumetric medium 350 as an object in the surrounding environment to be avoided (e.g., by an autonomous vehicle).
  • the composite light pulse 370 may be interpreted as a light pulse reflected from an inclined surface or the ground, or as a multitude of relatively weak light pulses reflected from the volumetric medium 350. In such situations, other factors may also be considered in classifying or interpreting the signal, for example, to distinguish between objects to be avoided and air particulates such as exhaust.
  • Figure 4A shows a block diagram of an example detector circuit 400 suitable for use in a LIDAR device (such as the LIDAR device 101 of Figure 1).
  • the detector circuit 400 may be one example of the detector circuit 122 of Figure 1.
  • the detector circuit 400 is shown to include a comparator 410, an amplifier 415, a sample-and-hold circuit 420, and an output bus 430.
  • a photodiode 402 generates a photodetector signal 405 in response to receiving or detecting photons associated with received light pulses 125.
  • the photodetector signal 405 may include information indicative of at least the rising edge timing, the falling edge timing, the peak amplitude, and the pulse width of a respective return light pulse 125.
  • the amount of current generated by the photodiode 402 may be indicative of the intensity of photons detected by the photodiode 402
  • the shape of the signal waveform of the photodetector signal 405 may be indicative of timing and pulse width information of the respective return light pulse 125.
  • the photodetector signal 405 is generated by the photodiode 402 in the example of Figure 4A, other suitable circuits or devices may be used to detect photons associated with the return light pulses 125 and to generate electrical signals indicative of signal properties and timing information of the return light pulses 125.
  • the comparator 410 includes an input coupled to the photodiode 402, includes an output coupled to the amplifier 415, and may compare the amplitude of the photodetector signal 405 with the threshold value (ATH) to determine whether the corresponding light pulse is a valid signal to be sampled by the sample-and-hold circuit 420.
  • the comparator 410 may be configured to selectively pass the photodetector signal 405 to the sample-and-hold circuit 420 based on the comparison.
  • the comparator 410 may pass the photodetector signal 405 to the sample-and-hold circuit 420 for sampling when its amplitude exceeds the threshold value ATH, and may not pass the photodetector signal 405 to the sample-and-hold circuit 420 for sampling when its amplitude does not exceed the threshold value ATH.
  • the comparator 410 may be configured to selectively trigger or enable the sample-and-hold circuit 420 based on the comparison.
  • the sample-and-hold circuit 420 may be enabled to sample the photodetector signal 405 when its amplitude exceeds the threshold value ATH, and may not be enabled to sample the photodetector signal 405 when its amplitude does not exceed the threshold value ATH. In this manner, the comparator 410 may reduce the amount of information or data to be sampled, stored, and processed to generate point clouds of the surrounding environment, for example, by ignoring received signals having an amplitude less than the threshold value ATH.
  • the comparator 410 may include hysteresis to prevent chattering and inadvertent logic state transitions caused by spikes in the input signals.
  • the comparator 410 may be enabled when a magnitude of the photodetector signal 405 exceeds a first threshold level and may be disabled when the magnitude of the photodetector signal 405 falls below a second threshold level that is less than the first threshold level. In this manner, the comparator 410 may not only ignore spurious spikes in the input signal but also may determine a walk error associated with the received light pulses 125.
  • the first and second threshold levels may be dynamically adjusted based on reception of a number of return light pulses 125, for example, to improve the accuracy with which the comparator 410 identifies valid light pulses to be processed by the detector circuit 400.
  • the comparator 410 may be omitted, or may be modified to allow photodetector signals 405 to be continuously sampled by the sample-and- hold circuit 420.
  • the amplifier 415 is coupled between the comparator 410 and an input node N1 of the sample-and-hold circuit 420.
  • the amplifier 415 may be any suitable circuit (such as a low-noise amplifier) that amplifies the photodetector signal 405 to generate an input signal (IN) suitable for driving circuitry within or associated with the sample-and-hold circuit 420.
  • the amplifier 415 may convert a photocurrent generated by the photodiode 402 into an analog voltage signal indicative of signal properties and timing information of a respective return light pulse 125. More specifically, in some aspects, the amplifier 415 may be a transimpedance amplifier that converts a received photocurrent into a voltage signal. It is noted that although the amplifier 415 is shown to include power terminals coupled to VDD and ground, in other implementations, the power terminals of the amplifier 415 may be coupled to other suitable voltage supplies or voltage rails.
  • the sample-and-hold circuit 420 is coupled between node N1 and the output bus 430, and may sample the photodetector signal 405 to determine intensity and timing information of the received light pulses 125 at different times.
  • the sample-and-hold circuit 420 may provide a number of output signals (OUTi-OUTn) indicative of determined intensity and timing information of a corresponding number of photodetector signals 405 to the output bus 430.
  • the sample-and-hold circuit 420 may operate in a sample to sample the photodetector signal 405 to determine intensity and timing information, and may operate in a hold mode to hold or store the sampled signals for analysis and/or conversion to digital data that can be processed by the LIDAR controller 130 of Figure 1.
  • the sample-and-hold circuit 420 may include a number of processing channels 421(l)-421(n) coupled in parallel with each other between node N1 and the output bus 430.
  • the processing channels 421(l)-421(n) may independently sample the photodetector signal 405 (such as in a sequential manner) to determine intensity and timing information of a corresponding one of the received light pulses 125, and may generate the output signals OUTi-OUTn indicative of the intensity and timing information of the sampled photodetector signals 405.
  • the intensity information may include (but is not limited to) intensity levels, pulse shape, and the peak amplitude of a received light pulse 125
  • the timing information may include (but is not limited to) rising edge timing, falling edge timing, and pulse width information of the received light pulse 125.
  • the sample-and-hold circuit 420 may include 64 processing channels 421(1)-421(64). In other aspects, the sample-and-hold circuit 420 may include any suitable number of processing channels 421.
  • one or more of the processing channels 421(1)— 421(n) may be configured to determine a pulse width of a received light pulse reflected from a surface, to compare the determined pulse width with a selected one of a number of reference pulse widths, and to determine an amount of pulse elongation of the received light pulse (which may be caused by the volumetric medium) based on the comparison.
  • the determined amount of pulse elongation may be used (such as by the LIDAR controller 130 of Figure 1) to distinguish between light pulses reflected an inclined surface of an object and light pulses reflected by air particulates associated with a volumetric medium and/or to classify a reflective surface as either an object to be avoided by an autonomous vehicle or as air particulates associated with the volumetric medium.
  • the determined amount of pulse elongation may be used (such as by the LIDAR controller 130 of Figure 1) to distinguish between light pulses reflected by debris and other small foreign objects on a road and light pulses reflected by the road.
  • Information indicative of such distinctions and/or indicative of the determined amounts of pulse elongation may be used to identify such debris and other small foreign objects left on the road surface, which in turn may increase the accuracy with which point clouds or 3D maps generated by a LIDAR system represent objects, surfaces, and other features of the surrounding environment.
  • the output signals OUTi-OUTn may be provided to the output bus 430, converted to digital data by one or more ADCs (not shown for simplicity), and provided to one or more digital circuits or computing devices (such as the DSP 133 and/or the processor 131 of Figure 1) for processing and analyzing.
  • the output signals OUTi-OUTn may contain information indicative of timing and intensity values of received light pulses, indicative of determined amounts of pulse elongation, indicative of classifications of objects and surfaces in the environment, or any combination thereof.
  • one or more computing devices may process and analyze the digital data to detect objects in the surrounding environment, to determine distances to the detected objects, to determine various features (such as the size, shape, orientation, and movement) of the detected objects, to generate a 3D point cloud or map of the surrounding environment, and other suitable operations.
  • FIG. 4B shows a block diagram of an example processing channel 450 suitable for use in a LIDAR detector or a LIDAR device (such as the LIDAR device 101 of Figure 1).
  • the processing channel 450 may be one example of the processing channels 421 ( 1 )— 421 (n) of Figure 4A.
  • the processing channel 450 may include a gating circuit 452, a sampling circuit 453, a look-up table (LUT) 454, a compare circuit 455, and an output circuit 456.
  • the processing channel 450 may include one or more other suitable circuits or devices not shown in Figure 4B.
  • the gating circuit 452 includes an input to receive the input signal (IN) from the amplifier 415 via node Nl, and includes an output coupled to the sampling circuit 453. In some implementations, the gating circuit 452 may selectively connect (and disconnect) the processing channel 450 to node Nl, for example, to control when (and for how long) the photodetector signal 405 is provided to the processing channel 450 for sampling.
  • the gating circuit 452 may selectively route the photodetector signal 405 to the processing channel 450 based on one or more mode signals (not shown for simplicity), for example, so that the processing channels 421 ( 1 )— 421 (n) of Figure 4 A may sequentially sample the photodetector signal 405 to determine various signal properties and timing information of the return light pulses 125.
  • the sampling circuit 453 which may be any suitable circuit that can sample one or more values of an electrical signal, is shown to include an input coupled to the gating circuit 452, and to include one or more outputs coupled to the LUT 454, the compare circuit 455, and the output circuit 456.
  • the sampling circuit 453 may sample the photodetector signal 405 to determine signal properties and timing information of return light pulses 125 at different times, and may generate a number of output signals indicative of the determined signal properties and timing information.
  • the sampling circuit 453 may provide a first signal indicative of the peak amplitude (APEAK) of the received light pulse 125 to the LUT 454, may provide a second signal indicative of a determined pulse width (PWDET) of the received light pulse to the compare circuit 455, and may provide a third signal indicative of the sampled light pulse (INSAMPLED) to the output circuit 456.
  • APEAK peak amplitude
  • PWDET determined pulse width
  • INSAMPLED third signal indicative of the sampled light pulse
  • the sampling circuit 453 may determine the rising edge timing, the falling edge timing, the peak amplitude, and the pulse width of the received light pulse 125 based on the sampled photodetector signal 405. In some aspects, the sampling circuit 453 may determine a first time at which the amplitude of the received light pulse 125 rises above the threshold value (ATH), may determine a second time at which the amplitude of the received light pulse 125 falls below the threshold value (ATH), and may determine the pulse width (PWDET) of the received light pulse 125 based on a difference between the first and second times. In addition, or in the alternative, the sampling circuit 453 may use one or more of the determined rising edge timing, the determined falling edge timing, and the determined peak amplitude to determine the pulse width (PWDET) of the received light pulse 125.
  • ATH threshold value
  • PWDET pulse width
  • the LUT 454 may be any suitable memory circuit or device, may store a reference pulse width (PWREF) for each of a plurality of amplitude values of the return light pulses 125.
  • PWREF reference pulse width
  • APEAK determined peak amplitude
  • the sampling circuit 453 may be used as a look-up value or search key to retrieve a selected one of the plurality of reference pulse widths stored in the LUT 454.
  • the LUT 454 may provide the selected reference pulse width as a signal PWREF to the compare circuit 455.
  • the LUT 454 may store a number of pulse widths of light reflected from one or more road surfaces, and may provide one of these stored pulse widths for comparison with the pulse width of a received light pulse, for example, to distinguish between light pulses reflected by debris and other small foreign objects on a road and light pulses reflected by the road.
  • the pulse widths of light reflected from the one or more road surfaces may be determined during a calibration operation.
  • the pulse widths of light reflected from the one or more road surfaces may be determined or derived from one or more previously received light pulses that were classified as reflections from a particular road surface.
  • the compare circuit 455 includes a first input coupled to the sampling circuit 453, includes a second input coupled to the LUT 454, and includes an output coupled to the output circuit 456.
  • the compare circuit 455 may be configured or instructed to compare the determined pulse width (PWDET) provided by the sampling circuit 453 with the selected reference pulse width (PWREF) provided by the LUT 454 to determine a pulse width difference (PWDIFF).
  • PWDIFF PWDET - PWREF
  • the compare circuit 455 may be configured or instructed to compare the determined pulse width (PWDET) with a selected reference pulse width (PWREF) of light reflected from a road surface to determine an amount of pulse elongation indicative of whether the received light pulse was reflected by debris on the road or was reflected by the road itself.
  • PWDET determined pulse width
  • PWREF selected reference pulse width
  • the compare circuit 455 may include a control terminal to receive one or more enable signals that control or determine when the compare circuit 455 performs comparison operations to generate the pulse width difference signal PWDIFF. Further, although not shown in Figure 4B for simplicity, the compare circuit 455 may be independently trimmed to compensate for random offsets resulting from imperfections and defects inherent in semiconductor fabrication processes.
  • the output circuit 456 includes a first input coupled to the sampling circuit 453, a second input coupled to the compare circuit 455, and an output to generate the output signal OUT.
  • the output circuit 456 may be configured to provide the sampled light pulse (INSAMPLED) and/or the amount of pulse elongation (which may be caused by the volumetric medium) to the output bus 430, for example, for conversion to digital data.
  • one or more computing devices such as the LIDAR controller 130 may use information contained in the output signal OUT to distinguish between light pulses reflected by an object and light pulses reflected by air particulates associated with a volumetric medium. In this manner, the one or more computing devices may more accurately classify points in the environment corresponding to the volumetric medium as air particulates rather than objects to be avoided by an autonomous vehicle.
  • Figure 5 shows a graph 500 depicting the pulse widths of a number of light pulses relative to an example relationship between amplitude and pulse width of a reference light pulse.
  • the light pulses depicted in Figure 5 may be emitted or received by the LIDAR device 101 of Figure 1 or the LIDAR device 310 of Figures 3 A-3C.
  • the light pulses are depicted in Figure 5 as points 501-502 relative to an example plot 510 depicting a relationship between the pulse width (the y-axis) and the amplitude (the x-axis) of the reference light pulse.
  • the pulse width of the reference light pulse increases as the amplitude of the reference light pulse increases until a first amplitude value Ai, may remain constant (at least temporarily) as the amplitude of the reference light pulse increases from the first amplitude value Ai to a second amplitude value A2, and then begins to decrease as the amplitude of the reference light pulse further increases.
  • the first and second light pulses each have a pulse width that is greater than the pulse width of the reference light pulse, for example, as depicted by points 501 and 502 positioned above the plot 510.
  • the pulse width of the second light pulse is greater than the pulse width of the first light pulse, for example, as depicted by point 502 positioned above point 501.
  • the distance DI between point 501 and the plot 510 may be indicative of the pulse width of the first light pulse
  • the distance D2 between point 502 and the plot 510 may be indicative of the pulse width of the second light pulse.
  • Figure 6 shows a graph 600 depicting an example relationship between pulse width and peak amplitude of a number of light pulses.
  • the light pulses depicted in Figure 6 may be emitted or received by the LIDAR device 101 of Figure 1 or the LIDAR device 310 of Figures 3 A-3C.
  • the light pulses are depicted in Figure 6 as points 601 on the graph 600, with amplitude shown on the x-axis and pulse width shown on the y-axis.
  • the graph 600 also shows an example plot 610 depicting a relationship between averaged pulse width values and averaged amplitude values of the light pulses represented by the points 601.
  • the example plot 610 depicted in Figure 6 may be determined during manufacture of the LIDAR device. In other implementations, the example plot 610 depicted in Figure 6 may be determined after manufacture of the LIDAR device (such as during calibration or testing of the LIDAR device).
  • each of the reference pulse widths may be determined by emitting a light pulse at a corresponding power level towards an orthogonal surface, measuring the pulse width of the return light pulse reflected by the orthogonal surface, and storing the measured pulse width as the reference pulse width for a corresponding amplitude value in memory (such as in the LUT 454 of Figure 4B). The peak amplitude may be measured by the LIDAR device and/or may be determined or otherwise derived from the transmit power level of the emitted light pulse.
  • the LIDAR system 100 may be coupled to, mounted on, or otherwise associated with a vehicle.
  • the LIDAR system 100 may scan an environment by emitting light pulses into the environment and detecting return light pulses reflected by a number of objects in the environment.
  • the emitted light pulses and the received light pulses may be time-stamped to estimate distances between the LIDAR system 100 and the detected objects, and the received light pulses may be analyzed and/or processed to determine a number of characteristics of the received light pulses.
  • the estimated distances and determined characteristics may be used to generate a point cloud (or other suitable point mapping) indicative of objects and other reflective surfaces of the surrounding environment.
  • aspects of the present disclosure may increase the accuracy with which point clouds classify points and represent the location, features, and other characteristics of objects detected in an environment containing a volumetric medium by distinguishing between light pulses reflected by an inclined surface and light pulses reflected by air particulates associated with the volumetric medium.
  • implementations disclosed herein may also be used to detect and identify debris and other small foreign objects left on the road surface.
  • a number of points in a point cloud of a scene may be analyzed based on pulse elongation of a number of corresponding return light pulses.
  • a point on the point cloud may be represented by a certain color that may be indicative of an amount of pulse elongation of the corresponding return light pulse, such that points corresponding to return light pulses that were reflected by fog (or other air particulates) in the scene may be visually identified and distinguished from points corresponding to return light pulses that were reflected by an inclined surface of an object (e.g., rather than the fog or other air particulates).
  • points in the point cloud corresponding to light pulses having a relatively large amount of pulse elongation may be of a first color (such as red)
  • points in the point cloud corresponding to light pulses having a nominal amount of pulse elongation may be of a second color (such as green)
  • points in the point cloud corresponding to light pulses having a relatively small amount of pulse elongation may be of a third color (such as blue).
  • a greater number of colors may be used for the points in the point cloud, for example, to increase the granularity with which the point cloud may indicate different amounts of pulse elongation associated with the points in the point cloud.
  • points corresponding to return light pulses that were reflected by a volumetric medium may be distinguished, e.g., by their colors in this example or by other values in other implementations, from points corresponding to return light pulses that were not reflected by the volumetric medium.
  • Figure 7A shows a flow chart depicting an example operation 700 for identifying a presence of a volumetric medium in an environment associated with a LIDAR device.
  • the example operation 700 is described below with respect to the LIDAR system 100 of Figure 1 for illustrative purposes only.
  • One of ordinary skill in the art will recognize that the example operation 700 may be performed by any suitable LIDAR system or by any suitable LIDAR device and LIDAR controller according various implementations disclosed herein.
  • the example operation 700 may include one or more operations, functions, or actions depicted by one or more of blocks 701-707, and may be performed with additional steps, with fewer steps, with steps in a different order, with steps in parallel, or any combination thereof.
  • blocks 701-707 are illustrated in a sequential order, in other implementations, these blocks 701-707 may be performed in parallel, may be performed in a different order than those described herein, may be combined into fewer blocks, divided into additional blocks, removed from the example operation 700, or any combination thereof.
  • the LIDAR system 100 may emit a light pulse into the environment (701), and may receive a return light pulse corresponding to a reflection of the emitted light pulse by an object in the environment (702).
  • the received light pulse may include components of light pulses emitted from the LIDAR system 100 and reflected by one or more objects in the surrounding environment, and may also include components of reflected light pulses that propagated through a volumetric medium.
  • the LIDAR system 100 may generate a photodetector signal based on detecting photons associated with the received light pulse.
  • the LIDAR system 100 may determine a pulse width of the received light pulse (703). In some implementations, the LIDAR system 100 may determine a first time at which a magnitude of the received light pulse rises above a first value, may determine a second time at which the magnitude of the received light pulse falls below a second value, and determine the pulse width of the received light pulse based on a difference between the first and second times. In some aspects, the first and second values may be the same. In other aspects, the first and second values may be different. In addition, or in the alternative, the LIDAR system 100 may also use one or more of a determined rising edge time, a falling edge time, and a peak amplitude to determine the pulse width of the received light pulse. In other implementations, the LIDAR system 100 may use any other suitable technique to measure, determine, or estimate the pulse width of the received light pulse.
  • the LIDAR system 100 may compare the determined pulse width with a reference pulse width (704).
  • the LIDAR system 100 may select one of a plurality of reference pulse widths based on a peak amplitude of the received light pulse, and may use the selected reference pulse width for the comparison to generate a signal indicative of the difference between the reference pulse width and the determined pulse width.
  • the plurality of reference pulse widths may be stored in a look-up table (such as the LUT 124 of Figure 1 or the LUT 454 of Figure 4B), and an amplitude value of the received light pulse may be used as a search key to retrieve a corresponding reference pulse width from the look-up table.
  • the LIDAR system 100 may use any other suitable technique to compare the measured pulse width with the reference pulse width.
  • the LIDAR system 100 may determine an amount of pulse elongation of the received light pulse (705). In some implementations, the LIDAR system 100 may generate a signal indicating the difference between the reference pulse width and the determined pulse width, and use the signal to convey the amount of pulse elongation, which may be used as an indicator of the presence of a volumetric medium. In other implementations, the LIDAR system 100 may use any other suitable technique to determine the amount of pulse elongation of the received light pulse.
  • the LIDAR system 100 may classify the surface as either an object to be avoided (such as by an autonomous or semi-autonomous vehicle) or as air particulates associated with the volumetric medium based, at least in part, on the determined amount of pulse elongation (706). Information indicating the classification of the surface and/or the determined amount of pulse elongation may be used to increase the accuracy of mapping information (such as a point cloud) generated by the LIDAR system 100.
  • mapping information such as a point cloud
  • the LIDAR system 100 may generate a corresponding point in a point cloud indicative of objects detected in the environment based on at least one of the classification of the surface and the determined amount of pulse elongation (707).
  • the LIDAR system 100 may use any suitable technique to generate the point cloud.
  • the LIDAR system 100 may use the amplitudes, pulse widths, timing information, and other characteristics of the received light pulses along with object classification information and determined amounts of pulse elongation to generate the point cloud (or some other suitable three-dimensional map of points) indicating the location, size, shape, movement, orientation, and other features of objects and other reflective surfaces detected in the environment.
  • operations associated with blocks 706 and 707 of Figure 7A may be performed by any suitable computing device or hardware and data processing apparatus.
  • operations associated with blocks 706 and 707 of Figure 7A may be performed by the LIDAR controller 130 of Figure 1.
  • operations associated with blocks 706 and 707 of Figure 7A may be performed by a computing device external to the LIDAR system 100 of Figure 1.
  • operations associated with blocks 706 and 707 of Figure 7A may be performed by a computing device remote from the LIDAR device 101 of Figure 1.
  • the remote computing device may communicate wirelessly with the LIDAR device 101 (such as by using Wi-Fi communications, BLUETOOTH communications, Wi-Gig communications, cellular communications, satellite communications, or other wireless communication protocol). In other aspects, the remote computing device may communicate with the LIDAR device 101 using any suitable wired connection (such as by using Ethernet communications, Cable communications, POTS communications, powerline communications (PLC), or other wired communication protocol).
  • LIDAR systems disclosed herein may use amounts of pulse elongation exhibited by reflected light pulses to identify debris and other small foreign objects (such as a piece of tire) left on a road or street surface. Although debris and other small foreign objects may be difficult to detect because of their relatively small geometrical aspects, the elongation of light pulses reflected by such debris and other small foreign objects may be detectably different than the elongation of light pulses reflected by the road surface. Accordingly, LIDAR systems disclosed herein may use pulse elongation characteristics of received light pulses to distinguish between light pulses reflected by debris and other small foreign objects and light pulses reflected by the road, for example, to detect and/or identify such debris and other small foreign objects left on the road surface.
  • Figure 7B shows a flow chart depicting an example operation 710 for identifying debris on a road surface.
  • the example operation 710 is described below with respect to the LIDAR system 100 of Figure 1 for illustrative purposes only.
  • the example operation 710 may be performed by any suitable LIDAR system or by any suitable LIDAR device and LIDAR controller according various implementations disclosed herein.
  • the example operation 710 may include one or more operations, functions, or actions depicted by one or more of blocks 711-712, and may be performed with additional steps, with fewer steps, with steps in a different order, with steps in parallel, or any combination thereof.
  • blocks 711-712 are illustrated in a sequential order, in other implementations, these blocks 711-712 may be performed in parallel, may be performed in a different order than those described herein, may be combined into fewer blocks, divided into additional blocks, removed from the example operation 710, or any combination thereof.
  • the example operation 710 may be part of, or appended to, the example operation 700 of Figure 7A.
  • the LIDAR system 100 may compare the determined pulse width with a pulse width of light reflected from a road surface to determine a pulse elongation difference value (711).
  • the pulse width of light reflected from the road surface may be stored in a memory (such as the LUT 454 of Figure 4B) of the LIDAR system 100.
  • the pulse width of light reflected from the road surface may be determined during a calibration operation.
  • the pulse width of light reflected from the road surface may be determined or derived from one or more previously received light pulses classified as reflections from the road surface.
  • the LIDAR system 100 may identify the surface as debris on the road surface based, at least in part, on the determined pulse elongation difference value (712). As mentioned above, the elongation of light pulses reflected by debris and other small foreign objects may be detectably different than the elongation of light pulses reflected by the road surface. Thus, the LIDAR system 100 may use the determined pulse elongation difference value to distinguish between light pulses reflected by debris and other small foreign objects and light pulses reflected by the road, for example, to detect and/or identify such debris and other small foreign objects left on the road surface.
  • operations associated with block 712 of Figure 7B may be performed by any suitable computing device or hardware and data processing apparatus.
  • operations associated with block 712 of Figure 7B may be performed by the LIDAR controller 130 of Figure 1.
  • operations associated with block 712 of Figure 7B may be performed by a computing device external to the LIDAR system 100 of Figure 1.
  • operations associated with block 712 of Figure 7B may be performed by the remote computing device.
  • Figure 8 A shows a flow chart depicting an example operation 800 for comparing the determined pulse width of a received light pulse with a reference pulse width.
  • the example operation 800 is described below with respect to the LIDAR system 100 of Figure 1 for illustrative purposes only.
  • the example operation 800 may be performed by any suitable LIDAR system or by any suitable LIDAR device and LIDAR controller according various implementations disclosed herein.
  • the example operation 800 may include one or more operations, functions, or actions depicted by one or more of blocks 801-803, and may be performed with additional steps, with fewer steps, with steps in a different order, with steps in parallel, or any combination thereof.
  • blocks 801-803 are illustrated in a sequential order, in other implementations, these blocks 801-803 may be performed in parallel, may be performed in a different order than those described herein, may be combined into fewer blocks, divided into additional blocks, removed from the example operation 800, or any combination thereof.
  • the LIDAR system 100 may determine a peak amplitude of the received light pulse (801), may select one of a plurality of reference pulse widths based on the determined peak amplitude (802), and may compare the determined pulse width with the selected reference pulse width (803).
  • the plurality of reference pulse widths may be stored in a look-up table, and the determined peak amplitude may be used as a search key to retrieve a corresponding reference pulse width from the look-up table.
  • the LIDAR system 100 may use any other suitable technique to determine the peak amplitude and/or to select the reference pulse width.
  • Figure 8B shows a flow chart depicting an example operation 810 for determining the pulse width of a received light pulse.
  • the example operation 810 is described below with respect to the LIDAR system 100 of Figure 1 for illustrative purposes only.
  • the example operation 810 may be performed by any suitable LIDAR system or by any suitable LIDAR device and LIDAR controller according various implementations disclosed herein.
  • the example operation 810 may include one or more operations, functions, or actions depicted by one or more of blocks 811-812, and may be performed with additional steps, with fewer steps, with steps in a different order, with steps in parallel, or any combination thereof.
  • blocks 811-812 are illustrated in a sequential order, in other implementations, these blocks 811-812 may be performed in parallel, may be performed in a different order than those described herein, may be combined into fewer blocks, divided into additional blocks, removed from the example operation 810, or any combination thereof.
  • the LIDAR system 100 may determine a first time at which a magnitude of the received light pulse rises above a value (811), and may determine a second time at which the magnitude of the received light pulse falls below the value (812). In some implementations, the LIDAR system 100 may determine the pulse width of the received light pulse based on a difference between the first and second determined times. In other implementations, the LIDAR system 100 may use any other suitable technique to determine first and second times.
  • Figure 8C shows a flow chart depicting an example operation 820 for retrieving a reference pulse width.
  • the example operation 820 is described below with respect to the LIDAR system 100 of Figure 1 for illustrative purposes only.
  • the example operation 820 may be performed by any suitable LIDAR system or by any suitable LIDAR device and LIDAR controller according various implementations disclosed herein.
  • the example operation 820 may include one or more operations, functions, or actions depicted by one or more of blocks 821-822, and may be performed with additional steps, with fewer steps, with steps in a different order, with steps in parallel, or any combination thereof.
  • blocks 821-822 are illustrated in a sequential order, in other implementations, these blocks 821-822 may be performed in parallel, may be performed in a different order than those described herein, may be combined into fewer blocks, divided into additional blocks, removed from the example operation 820, or any combination thereof.
  • the LIDAR system 100 may determine a peak amplitude of the received light pulse (821), and may retrieve the reference pulse width from a look-up table based on the determined peak amplitude (822).
  • the LIDAR system 100 may use any suitable technique to determine the peak amplitude of the received light pulse.
  • the look-up table may store a plurality of reference pulse widths each for a corresponding one of a plurality of different amplitude values.
  • the reference pulse widths stored in the look-up table may be determined during a calibration operation (such as during part of a manufacturing process or quality control operation of the LIDAR system ).
  • each of the reference pulse widths may be determined by emitting a light pulse at a corresponding power level towards an orthogonal surface, determining the pulse width of the return light pulse reflected from the orthogonal surface, and storing the determined pulse width as the reference pulse width for a corresponding amplitude value in the look-up table.
  • the peak amplitude value may be measured by the LIDAR system and/or may be determined or otherwise derived from the transmit power level of the emitted light pulse.
  • a phrase referring to “at least one of’ a list of items refers to any combination of those items, including single members.
  • “at least one of a, b, or c” is intended to cover: a, b, c, a-b, a-c, b-c, and a-b-c.
  • the hardware and data processing apparatus used to implement the various illustrative logics, logical blocks, modules and circuits described in connection with the aspects disclosed herein may be implemented or performed with a general purpose single- or multi-chip processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein.
  • a general purpose processor may be a microprocessor, or, any conventional processor, controller, microcontroller, or state machine.
  • a processor also may be implemented as a combination of computing devices (such as a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration). In some implementations, particular processes and methods may be performed by circuitry that is specific to a given function.
  • the functions described may be implemented in hardware, digital electronic circuitry, computer software, firmware, including the structures disclosed in this specification and their structural equivalents thereof, or in any combination thereof. Implementations of the subject matter described in this specification also can be implemented as one or more computer programs (such as one or more modules of computer program instructions) encoded on a computer storage media for execution by, or to control the operation of, data processing apparatus.
  • Computer-readable media includes both computer storage media and communication media including any medium that can be enabled to transfer a computer program from one place to another.
  • a storage media may be any available media that may be accessed by a computer.
  • such computer-readable media may include RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that may be used to store desired program code in the form of instructions or data structures and that may be accessed by a computer.
  • any connection can be properly termed a computer-readable medium. Combinations of the above should also be included within the scope of computer-readable media. Additionally, the operations of a method or algorithm may reside as one or any combination or set of codes and instructions on a machine readable medium and computer-readable medium, which may be incorporated into a computer program product.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • General Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Electromagnetism (AREA)
  • Optical Radar Systems And Details Thereof (AREA)

Abstract

Des systèmes et des procédés sont divulgués pour identifier une présence d'un milieu volumétrique dans un environnement associé à un système LIDAR. Dans certains modes de réalisation, le système LIDAR peut émettre une impulsion lumineuse dans l'environnement, recevoir une impulsion lumineuse de retour correspondant à la réflexion de l'impulsion lumineuse émise par une surface dans l'environnement, et déterminer une largeur d'impulsion de l'impulsion lumineuse reçue. Le système LIDAR peut comparer la largeur d'impulsion déterminée avec une largeur d'impulsion d'une lumière réfléchie par une surface de route pour déterminer une valeur de différence d'allongement d'impulsion. Le système LIDAR peut identifier la présence de débris sur la surface de la route ou à proximité de celle-ci sur la base, au moins en partie, de la valeur de différence d'allongement d'impulsion déterminée.
PCT/US2021/014865 2021-01-25 2021-01-25 Allongement d'impulsion lidar WO2022159110A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/US2021/014865 WO2022159110A1 (fr) 2021-01-25 2021-01-25 Allongement d'impulsion lidar

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2021/014865 WO2022159110A1 (fr) 2021-01-25 2021-01-25 Allongement d'impulsion lidar

Publications (1)

Publication Number Publication Date
WO2022159110A1 true WO2022159110A1 (fr) 2022-07-28

Family

ID=82549158

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2021/014865 WO2022159110A1 (fr) 2021-01-25 2021-01-25 Allongement d'impulsion lidar

Country Status (1)

Country Link
WO (1) WO2022159110A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012087150A1 (fr) * 2010-12-22 2012-06-28 Edp Systems As Appareil de contrôle de condition de surface de route
US20140307247A1 (en) * 2013-04-11 2014-10-16 Google Inc. Methods and Systems for Detecting Weather Conditions Including Wet Surfaces Using Vehicle Onboard Sensors
KR20150049934A (ko) * 2013-10-31 2015-05-08 현대자동차주식회사 지면 데이터 필터링 방법 및 그 장치
US20180284226A1 (en) * 2017-03-28 2018-10-04 Luminar Technologies, Inc. Dynamically varying laser output in a vehicle in view of weather conditions
CN111208527A (zh) * 2018-11-16 2020-05-29 北京万集科技股份有限公司 激光雷达的测距方法、装置、激光雷达及存储介质

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012087150A1 (fr) * 2010-12-22 2012-06-28 Edp Systems As Appareil de contrôle de condition de surface de route
US20140307247A1 (en) * 2013-04-11 2014-10-16 Google Inc. Methods and Systems for Detecting Weather Conditions Including Wet Surfaces Using Vehicle Onboard Sensors
KR20150049934A (ko) * 2013-10-31 2015-05-08 현대자동차주식회사 지면 데이터 필터링 방법 및 그 장치
US20180284226A1 (en) * 2017-03-28 2018-10-04 Luminar Technologies, Inc. Dynamically varying laser output in a vehicle in view of weather conditions
CN111208527A (zh) * 2018-11-16 2020-05-29 北京万集科技股份有限公司 激光雷达的测距方法、装置、激光雷达及存储介质

Similar Documents

Publication Publication Date Title
US11892567B2 (en) LIDAR pulse elongation
USRE48763E1 (en) Multiple-field-of-view scannerless optical rangefinder in high ambient background light
CN111868560B (zh) 取决于脉冲类型选择lidar脉冲检测器
CN111868561B (zh) 使用本底噪声的自适应识别进行有效信号检测
CN111919138B (zh) 检测激光脉冲边沿以进行实时检测
CA2782180C (fr) Systeme de surveillance 3d actif pour une detection de trafic
US9378640B2 (en) System and method for traffic side detection and characterization
US20210278540A1 (en) Noise Filtering System and Method for Solid-State LiDAR
US20140211194A1 (en) Cost-effective lidar sensor for multi-signal detection, weak signal detection and signal disambiguation and method of using same
JP2017219502A (ja) 物体検出装置、センシング装置及び移動体装置
US11520019B2 (en) Light signal detection device, range finding device, and detection method
CN115720634A (zh) 具有雾检测和自适应响应的lidar系统
US20200271763A1 (en) Light detection and ranging signal correction methods and systems
WO2022159110A1 (fr) Allongement d'impulsion lidar
KR20230060923A (ko) 근거리의 반사 강도 측정을 위한 라이다 및 그 구동 방법
US20210396846A1 (en) Lidar system with detection sensitivity of photodetectors
US20230051395A1 (en) Scout pulsing
CN114556151A (zh) 测距装置、测距方法和可移动平台
WO2023282984A2 (fr) Normalisation de données de gmapd utilisant des épreuves de bernoulli
WO2024008550A1 (fr) Système lidar et procédé de fonctionnement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21921569

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21921569

Country of ref document: EP

Kind code of ref document: A1