US20200041619A1 - Fitting points to a surface - Google Patents
Fitting points to a surface Download PDFInfo
- Publication number
- US20200041619A1 US20200041619A1 US16/196,650 US201816196650A US2020041619A1 US 20200041619 A1 US20200041619 A1 US 20200041619A1 US 201816196650 A US201816196650 A US 201816196650A US 2020041619 A1 US2020041619 A1 US 2020041619A1
- Authority
- US
- United States
- Prior art keywords
- points
- point cloud
- relative velocity
- vehicle
- depth data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims abstract description 156
- 230000033001 locomotion Effects 0.000 claims description 19
- 230000001133 acceleration Effects 0.000 claims description 3
- 230000011218 segmentation Effects 0.000 description 32
- 230000008447 perception Effects 0.000 description 28
- 238000005096 rolling process Methods 0.000 description 24
- 230000000875 corresponding effect Effects 0.000 description 21
- 238000012937 correction Methods 0.000 description 15
- 238000010586 diagram Methods 0.000 description 15
- 238000010801 machine learning Methods 0.000 description 15
- 230000015654 memory Effects 0.000 description 13
- 238000013507 mapping Methods 0.000 description 11
- 230000003287 optical effect Effects 0.000 description 10
- 230000008569 process Effects 0.000 description 9
- 238000013528 artificial neural network Methods 0.000 description 6
- 238000001514 detection method Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 230000002787 reinforcement Effects 0.000 description 4
- 238000000638 solvent extraction Methods 0.000 description 4
- 230000008859 change Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 230000002085 persistent effect Effects 0.000 description 3
- 238000001931 thermography Methods 0.000 description 3
- 241001465754 Metazoa Species 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 230000001276 controlling effect Effects 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 241001025261 Neoraja caerulea Species 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000010813 municipal solid waste Substances 0.000 description 1
- 230000007935 neutral effect Effects 0.000 description 1
- 238000009828 non-uniform distribution Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/08—Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
- B60W30/095—Predicting travel path or likelihood of collision
- B60W30/0956—Predicting travel path or likelihood of collision the prediction being responsive to traffic or environmental parameters
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/42—Simultaneous measurement of distance and other co-ordinates
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/50—Systems of measurement based on relative movement of target
- G01S17/58—Velocity or trajectory determination systems; Sense-of-movement determination systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/86—Combinations of lidar systems with systems other than lidar, radar or sonar, e.g. with direction finders
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/87—Combinations of systems using electromagnetic waves other than radio waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/931—Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G01S17/936—
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/4808—Evaluating distance, position or velocity data
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
- G01S7/4817—Constructional features, e.g. arrangements of optical elements relating to scanning
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
- G01S7/4818—Constructional features, e.g. arrangements of optical elements using optical fibres
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/497—Means for monitoring or calibrating
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/60—Rotation of whole images or parts thereof
- G06T3/608—Rotation of whole images or parts thereof by skew deformation, e.g. two-pass or three-pass rotation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/80—Geometric correction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/521—Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/75—Determining position or orientation of objects or cameras using feature-based methods involving models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/689—Motion occurring during a rolling shutter mode
-
- H04N5/2329—
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/0088—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots characterized by the autonomous decision making process, e.g. artificial intelligence, predefined behaviours
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30168—Image quality inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
Definitions
- This disclosure generally relates to lidar systems, and, more particularly, to detecting, analyzing, correcting, and/or utilizing distortions in a point cloud produced by lidar systems.
- Self-driving or “autonomous” vehicles generally employ sensors, such as light detection and ranging (lidar) devices, to detect or “see” the surrounding environment as the vehicles move toward their destinations.
- lidar light detection and ranging
- a single lidar laser cannot scan the entire field of regard at once. Instead, lidar lasers may be configured to follow a scan pattern to scan the field of regard. As the lidar laser follows the scan pattern, a scanner determines depth data by measuring the travel time for light beams that are transmitted by the lidar system and scattered or reflected by objects in the field of regard. Each of these points along the scan pattern are referred to as “pixels” and are correlated to the determined depth data.
- the lidar system is capable of scanning the field of regard at high speeds, there is still a time a delay between when the scanner determines the depth data at the various pixels of the scan pattern. Thus, if an object in the field of regard is moving relative to the vehicle, the object may appear to be distorted.
- One example embodiment of the techniques of this disclosure is a computer-implemented method of determining a relative velocity between a vehicle and an object.
- the method includes receiving sensor data generated by one or more sensors of the vehicle.
- the one or more sensors are configured to sense an environment of the vehicle by following a scan pattern comprising component scan lines.
- the method also includes obtaining, based on the sensor data and by one or more processors, a point cloud frame representative of the environment.
- the point cloud frame comprises a plurality of points such that each point includes depth data associated with the point and a time at which the depth data was captured.
- the method includes selecting, by the one or more processors, two or more points of the scan pattern that overlap the object.
- the selected points are located on or near a two-dimensional surface corresponding to at least part of the object, and the depth data for two or more of the selected points are captured at different times. Further, the method includes calculating, by the one or more processors, the relative velocity between the vehicle and the object based at least in part on the depth data and capture times associated with the selected points.
- the system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines.
- the system also includes a computing system configured to receive the set of sensor data.
- the computing system is also configured to obtain, based on the sensor data, a point cloud frame representative of the environment.
- the point cloud frame includes a plurality of points such that each point includes depth data associated with the point and a time at which the depth data was captured by the set of sensors.
- the computing system is also configured to select two or more points of the scan pattern that overlap an object in the environment.
- the selected points are located on or near a two-dimensional surface corresponding to at least part of the object, and the depth data for two or more of the selected points are captured at different times. Additionally the computing system is configured to calculate a relative velocity between the autonomous vehicle and the object based at least in part on the depth data and capture times associated with the selected points.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of determining relative velocity between a vehicle and an object.
- the method includes receiving sensor data generated by one or more sensors of the vehicle.
- the one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern comprising component scan lines.
- the method includes obtaining, by one or more processors, a point cloud frame based on the sensor data and representative of the environment and identifying, by the one or more processors, a point cloud object within the point cloud frame.
- the method further includes determining, by the one or more processors, that the point cloud object is skewed relative to an expected configuration of the point cloud object, and determining, by the one or more processors, a relative velocity of the point cloud object by analyzing the skew of the object.
- the system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines.
- the system also includes a computing system configured to receive the set of sensor data and obtain a point cloud frame based on the sensor data and representative of the environment.
- the computing system is also configured to identify a point cloud object within the point cloud frame and determine that the point cloud object is skewed relative to an expected configuration of the point cloud object. Additionally, the computing system is configured to determine a relative velocity of the point cloud object by analyzing the skew of the object.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of determining relative velocity between a vehicle and an object.
- the method includes receiving sensor data generated by one or more sensors of the vehicle.
- the one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern comprising component scan lines.
- the method also includes obtaining, based on the sensor data and by one or more processors, a point cloud frame representative of the environment.
- the method includes identifying, by the one or more processors, a first pixel and a second pixel that are co-located within a field of regard of the one or more sensors and overlap a point cloud object within the point cloud frame and calculating, by the one or more processors, a difference between a depth associated with the first pixel and a depth associated with the second pixel.
- the method also includes determining, by the one or more processors, a relative velocity of the point cloud object by dividing the difference in depth data by a time difference between when the depth associated with the first pixel was sensed and the depth associated with the second pixel was sensed.
- the system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines.
- the system also includes a computing system configured to receive the set of sensor data and obtain, based on the set of sensor data, a point cloud frame representative of the environment.
- the computing system is also configured to identify a first pixel and a second pixel that are co-located within a field of regard of the one or more sensors and overlap a point cloud object within the point cloud frame and calculate a difference between a depth associated with the first pixel and a depth associated with the second pixel.
- the computing system is configured to determine a relative velocity of the point cloud object by dividing the difference in depth data by a time difference between when the depth associated with the first pixel was sensed and the depth associated with the second pixel was sensed.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of determining relative velocity between a vehicle and an object.
- the method includes receiving sensor data generated by one or more sensors of the vehicle.
- the one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern comprising component scan lines.
- the method also includes obtaining, based on the sensor data and by one or more processors, two or more point cloud frames representative of the environment and tracking, by the one or more processors, a point cloud object across the two or more point cloud frames.
- the method includes determining, based on the tracking and by the one or more processors, a relative velocity of the point cloud object and correcting, by the one or more processors, the point cloud object based on the relative velocity of the point cloud object.
- the system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines.
- the system also includes a computing system configured to receive the set of sensor data and obtain, based on the set of sensor data, two or more point cloud frames representative of the environment.
- the computing system is also configured to track a point cloud object across the two or more point cloud frames. Additionally, the computing system is configured to determine, based on the tracking, a relative velocity of the point cloud object and correct the point cloud object based on the relative velocity of the point cloud object.
- the method includes receiving sensor data generated by one or more sensors of the vehicle.
- the one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern.
- the method also includes obtaining, based on the sensor data and by one or more processors, a point cloud frame representative of the environment and identifying, by one or more processors, a point cloud object within the point cloud frame.
- the method includes analyzing, by the one or more processors, the point cloud object to identify a feature of the point cloud object that has an expected shape and comparing, by the one or more processors, the feature of the point cloud object to the expected shape.
- the method also includes identifying, by the one or more processors, that the point cloud object is distorted based on the feature of the point cloud object not matching the expected shape.
- the system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines.
- the system also includes a computing system configured to receive the set of sensor data and obtain, based on the sensor data, a point cloud frame representative of the environment.
- the computing system is also configured to identify a point cloud object within the point cloud frame and analyze the point cloud object to identify a feature of the point cloud object that has an expected shape. Additionally, the computing system is configured to compare the feature of the point cloud object to the expected shape and identify that the point cloud object is distorted based on the feature of the point cloud object not matching the expected shape.
- the method includes receiving sensor data generated by a plurality of sensors of the vehicle.
- the plurality of sensors includes a first set of one or more sensors, wherein each sensor of the first set of sensors is configured to sense an environment through which the vehicle is moving by, for each frame produced by the sensor, sequentially advancing through a plurality of points in a scan pattern.
- the plurality of sensors also include a second set of one or more sensors, wherein each sensor of the second set of sensors is configured to sense the environment by, for each frame produced by the sensor, capturing an entirety of the frame at a single time.
- the method also includes identifying, by one or more processors, an object within the environment.
- the method includes analyzing, by the one or more processors, the data sensed by the first set of sensors to identify a point cloud representation of the object and analyzing, by the one or more processors, the data sensed by the second set of sensors to identify the object within the data sensed by the second set of sensors.
- the method also includes comparing, by the one or more processors, a shape of the point cloud representation of the object to a shape of the object as sensed by the second set of sensors and identifying, by the one or more processors, that the object is distorted based on the shape of the point cloud representation of the object component not matching the shape of the object as sensed by the second set of sensors.
- the system includes a first set of one or more sensors configured to sense an environment through which the vehicle is moving by, for each frame produced by the sensor, sequentially advancing through a plurality of points in a scan pattern.
- the system also includes a second set of one or more sensors configured to sense the environment by, for each frame produced by the sensor, capturing an entirety of the frame at a single time.
- the system includes a computing system configured to receive data generated by the first and second sets of one or more sensors and identify an object within the environment.
- the computing system is also configured to analyze the data sensed by the first set of sensors to identify a point cloud representation of the object and analyze the data sensed by the second set of sensors to identify the object within the data sensed by the second set of sensors. Additionally, the computing system is configured to compare a shape of the point cloud representation of the object to a shape of the object as sensed by the second set of sensors and identify that the object is distorted based on the shape of the point cloud representation of the object component not matching the shape of the object as sensed by the second set of sensors.
- FIG. 1A illustrates an example, conventional “zigzag” scan pattern which a light detection and ranging (lidar) system can produce when identifying targets within a field of regard;
- FIG. 1B illustrates an example “looping” scan pattern which a lidar system can produce when identifying targets within a field of regard;
- FIG. 2 is a block diagram of an example lidar system which can implement the scan patterns of FIGS. 1A and/or 1B , as well as other techniques of this disclosure;
- FIG. 3A illustrates an example vehicle in which the lidar system of FIG. 2 may operate
- FIG. 3B illustrates another example vehicle in which the lidar system of FIG. 2 may operate
- FIG. 4 illustrates an example point cloud that may be generated by the lidar system of FIG. 2 or 3 ;
- FIG. 5 is a block diagram of a software architecture that may be used in the lidar system of FIG. 2 or 3 ;
- FIG. 6 is a block diagram of an example computing system for controlling an autonomous vehicle, which may be used in the lidar system of FIG. 2 or 3 ;
- FIG. 7 is a flow diagram of an example method for determining relative velocity of a point cloud object by fitting pixels onto a surface of the point cloud object;
- FIG. 8 is a flow diagram of an example method for determining relative velocity of a point cloud object by analyzing a skew of the point cloud object;
- FIG. 9 is a flow diagram of an example method for determining relative velocity of a point cloud object using co-located pixels
- FIG. 10 is a flow diagram of an example method for correcting point cloud distortion based on tracked object velocity
- FIG. 11 is a flow diagram of an example method for detecting that a point cloud object is distorted using features having a known shape.
- FIG. 12 is a flow diagram of an example method for method for detecting that a point cloud object is distorted by comparing to other sensor data.
- a lidar system when a lidar system produces a frame of depth data (e.g., a point cloud frame), the frame includes depth data captured at different times within the frame period. For instance, when the lidar system captures depth data indicative of a surface on which all points are equidistant from the lidar system, and if the surface is moving with respect to the lidar system, the surface will be at different positions for different pixels within the same point cloud frame. This may cause the surface to appear skewed within the point cloud representation. For example, if the surface is moving closer to or more distant from the lidar system, the surface may appear to be “tilted” or “skewed” toward or away from the lidar system.
- a surface moving laterally relative to the lidar system may appear to be tilted or skewed in a lateral direction, and a surface moving up or down relative to the lidar system may appear to be longer or shorter.
- the scan lines of the scan pattern are generally vertical (perpendicular to the ground)
- surfaces moving laterally relative to the lidar system may appear to be wider or narrower
- surfaces moving up or down relative to the lidar system may appear to be tilted or skewed.
- rolling shutter distortion is generally used herein to refer to this phenomenon.
- the lidar system is configured to implement various techniques to distinguish between objects that are experiencing rolling shutter distortion (e.g., a telephone pole that only appears to be tilted due to relative velocity) and objects that are actually shaped in a manner similar to a distorted object (e.g., a telephone pole that is actually tilted).
- the lidar system fits two or more points onto a two-dimensional surface to model the motion (including velocity) of the point cloud object.
- the lidar system may then use the depth data associated with the two or more points to model the motion of the surface to approximate the relative velocity of the point cloud object.
- the lidar system fits an additional two or more points associated with a different scan line to the two-dimensional surface. In these aspects, the lidar systems determines and compares a relative velocity of the surface using each set of two or more points.
- the lidar system can determine that the object is experiencing rolling shutter distortion and, if the object is distorted, the direction of the distortion (i.e., which direction the object is moving relative to the vehicle). Moreover, by determining a rate of change between the relative velocities, the lidar system can also determine the relative acceleration of the object.
- the lidar system analyzes co-located pixels (i.e., pixels at the same location in the field of regard but captured at different points of time within a point cloud frame) to detect rolling shutter distortion.
- the depth data for the co-located pixels are captured by two different lidar scanners.
- the depth data for the co-located pixels are captured by the same lidar scanner (i.e., the scan pattern crosses over the same location in the field of regard at multiple points in time during the point cloud frame). By comparing the depth data for the co-located pixels, the lidar system can determine whether the object is moving relative to the vehicle, and thus experiencing rolling shutter distortion.
- the lidar system compares the shape of a point cloud to reference image data captured by a RGB or stereo camera to detect rolling shutter distortion.
- the RGB or stereo camera doesn't include delays introduced by a scan pattern, and therefore indicates the “true” shape of an object. Accordingly, the lidar system can compare the point cloud object to corresponding RGB data to determine whether the shapes of the two objects match. When the shapes match, the point cloud object is not experiencing rolling shutter distortion (i.e., the object is actually shaped as detected by the lidar system).
- the rolling shutter distortion introduced by a lidar system scan pattern generally impacts objects that are closer to the vehicle more than objects that are further away.
- closer objects will typically be associated with more pixels than further away objects, and thus it takes more time to capture the depth data for all pixels corresponding to the nearby object. This allows the object and/or the vehicle to move further within the relevant portion of the scan pattern, which in turn causes the resulting depth data to distort the object more heavily, causing it to appear as if it is curved or skewed.
- the lidar system includes a database of objects (or features thereof) that have known, fixed shapes. For instance, license plates are rectangular, wheels are round, manufacturer logos have known, fixed shapes, etc.
- the lidar system can determine whether the feature is curved (and therefore not matching the known, expected shape) to determine that the object is distorted.
- disclosed techniques leverage the fact that the amount and nature of distortion detected is indicative of the relative velocity and direction between the object and the vehicle.
- the lidar system detects and analyzes the amount of the distortion to determine vehicle and/or object velocity.
- the lidar system can determine the vehicle and/or object velocity based on a single point cloud frame. This enables the lidar system to determine vehicle and/or object velocity faster than traditional techniques that track objects between multiple point cloud frames.
- the lidar system may still track objects across multiple point cloud frames. For example, in one embodiment, the lidar system determines relative velocity of an object using multiple point cloud frame tracking techniques. In some scenarios, a correlation between relative velocity and the magnitude of distortion is determined to develop a distortion factor indicative of the distortion introduced in each subsequent scan line and/or pixel. The lidar system then applies this distortion factor to other point cloud objects to determine the relative velocity of the other point cloud objects.
- the lidar system after detecting that a point cloud object is distorted, corrects the shape of the point cloud object before the point cloud frame and/or object data is analyzed by a classification module, a tracking module, or a prediction component of the lidar system.
- the distortion factor can be applied to remove the distortion introduced by the time difference between successive scan lines or pixels.
- the lidar system may apply an inverse operation to shift the point cloud data to remove rolling shutter distortion.
- FIGS. 1-4 An example lidar system in which the above techniques can be implemented is considered first with reference to FIGS. 1-4 .
- An example implementation of a vehicle controller that interacts with the example lidar systems is then discussed with reference to FIGS. 5-6 .
- example methods corresponding to distortion detection, analysis, and correction are discussed with reference to the flow diagrams of FIGS. 7-12 .
- FIG. 2 illustrates an example light detection and ranging (lidar) system 100 . While various lidar system components and characteristics are described herein, it is understood that any suitable lidar device(s) or system(s), and/or any other suitable types of sensors, may provide sensor data for processing using the software architectures described herein.
- the example lidar system 100 may include a light source 110 , a mirror 115 , a scanner 120 , a receiver 140 , and a controller 150 .
- the light source 110 may be, for example, a laser (e.g., a laser diode) that emits light having a particular operating wavelength in the infrared, visible, or ultraviolet portions of the electromagnetic spectrum.
- the light source 110 emits an output beam of light 125 which may be continuous-wave, pulsed, or modulated in any suitable manner for a given application.
- the output beam of light 125 is directed downrange toward a remote target 130 located a distance D from the lidar system 100 and at least partially contained within a field of regard of the system 100 .
- the target 130 may scatter or, in some cases, reflect at least a portion of light from the output beam 125 , and some of the scattered or reflected light may return toward the lidar system 100 .
- the scattered or reflected light is represented by input beam 135 , which passes through the scanner 120 , which may be referred to as a beam scanner, optical scanner, or laser scanner.
- the input beam 135 passes through the scanner 120 to the mirror 115 , which may be referred to as an overlap mirror, superposition mirror, or beam-combiner mirror.
- the mirror 115 in turn directs the input beam 135 to the receiver 140 .
- the input beam 135 may include light from the output beam 125 that is scattered by the target 130 , light from the output beam 125 that is reflected by the target 130 , or a combination of scattered and reflected light from target 130 .
- the lidar system 100 can include an “eye-safe” laser that present little or no possibility of causing damage to a person's eyes.
- the input beam 135 may contain only a relatively small fraction of the light from the output beam 125 .
- the receiver 140 may receive or detect photons from the input beam 135 and generate one or more representative signals. For example, the receiver 140 may generate an output electrical signal 145 that is representative of the input beam 135 . The receiver 140 may send the electrical signal 145 to the controller 150 .
- the controller 150 may include one or more instruction-executing processors, an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), and/or other suitable circuitry configured to analyze one or more characteristics of the electrical signal 145 in order to determine one or more characteristics of the target 130 , such as its distance downrange from the lidar system 100 . More particularly, the controller 150 may analyze the time of flight or phase modulation for the beam of light 125 transmitted by the light source 110 .
- the distance D from the lidar system 100 is less than or equal to a maximum range R MAX of the lidar system 100 .
- the maximum range R MAX (which also may be referred to as a maximum distance) of a lidar system 100 may correspond to the maximum distance over which the lidar system 100 is configured to sense or identify targets that appear in a field of regard of the lidar system 100 .
- the maximum range of lidar system 100 may be any suitable distance, such as 50 m, 200 m, 500 m, or 1 km, for example.
- the light source 110 , the scanner 120 , and the receiver 140 may be packaged together within a single housing 155 , which may be a box, case, or enclosure that holds or contains all or part of the lidar system 100 .
- the housing 155 includes a window 157 through which the beams 125 and 135 pass.
- the controller 150 may reside within the same housing 155 as the components 110 , 120 , and 140 , or the controller 150 may reside outside of the housing 155 . In one embodiment, for example, the controller 150 may instead reside within, or partially within, a perception component of a control architecture (described below).
- the housing 155 includes multiple lidar sensors, each including a respective scanner and a receiver.
- each of the multiple sensors can include a separate light source or a common light source.
- the multiple sensors can be configured to cover non-overlapping adjacent fields of regard or partially overlapping fields of regard, for example, depending on the implementation.
- the multiple sensors may direct an output beam 125 at the same location on the target 130 at different points of time to obtain co-located depth data.
- the output beam 125 and input beam 135 may be substantially coaxial.
- the output beam 125 and input beam 135 may at least partially overlap or share a common propagation axis, so that the input beam 135 and the output beam 125 travel along substantially the same optical path (albeit in opposite directions).
- the input beam 135 may follow along with the output beam 125 , so that the coaxial relationship between the two beams is maintained.
- the scanner 120 steers the output beam 125 in one or more directions downrange.
- the scanner 120 may include one or more scanning mirrors and one or more actuators driving the mirrors to rotate, tilt, pivot, or move the mirrors in an angular manner about one or more axes, for example.
- FIG. 2 depicts only a single mirror 115
- the lidar system 100 may include any suitable number of flat or curved mirrors (e.g., concave, convex, or parabolic mirrors) to steer or focus the output beam 125 or the input beam 135 .
- the first mirror of the scanner may scan the output beam 125 along a first direction
- the second mirror may scan the output beam 125 along a second direction that is substantially orthogonal to the first direction.
- a “field of regard” of the lidar system 100 may refer to an area, region, or angular range over which the lidar system 100 may be configured to scan or capture distance information.
- the lidar system 100 scans the output beam 125 within a 30-degree scanning range, for example, the lidar system 100 may be referred to as having a 30-degree angular field of regard.
- the scanner 120 may be configured to scan the output beam 125 horizontally and vertically, and the field of regard of the lidar system 100 may have a particular angular width along the horizontal direction and another particular angular width along the vertical direction.
- the lidar system 100 may have a horizontal field of regard of 10° to 120° and a vertical field of regard of 2° to 45°.
- the one or more scanning mirrors of the scanner 120 may be communicatively coupled to the controller 150 , which may control the scanning mirror(s) so as to guide the output beam 125 in a desired direction downrange or along a desired scan pattern.
- a scan (or scan line) pattern may refer to a pattern or path along which the output beam 125 is directed.
- the lidar system 100 can use the scan pattern to generate a point cloud with points or “pixels” that substantially cover the field of regard.
- the pixels may be approximately evenly distributed across the field of regard, or distributed according to a particular non-uniform distribution. In some embodiments, for example as shown in FIG. 1B , two or more of the pixels may be co-located within the field of regard.
- the light source 110 may emit pulses of light which the scanner 120 scans across a field of regard of the lidar system 100 .
- the target 130 may scatter one or more of the emitted pulses, and the receiver 140 may detect at least a portion of the pulses of light scattered by the target 130 .
- the receiver 140 may receive or detect at least a portion of the input beam 135 and produce an electrical signal that corresponds to the input beam 135 .
- the controller 150 may be electrically coupled or otherwise communicatively coupled to one or more of the light source 110 , the scanner 120 , and the receiver 140 .
- the controller 150 may provide instructions, a control signal, or a trigger signal to the light source 110 indicating when the light source 110 should produce optical pulses, and possibly characteristics (e.g., duration, period, peak power, wavelength, etc.) of the pulses.
- the controller 150 may also determine a time-of-flight value for an optical pulse based on timing information associated with when the pulse was emitted by light source 110 and when a portion of the pulse (e.g., the input beam 135 ) was detected or received by the receiver 140 .
- the lidar system 100 may be used to determine the distance to one or more downrange targets 130 .
- the system can be used to map the distance to a number of points (e.g., pixels) within the field of regard.
- a collection of pixels captured in succession (which may be referred to as a depth map, a point cloud, or a point cloud frame) may be rendered as an image or may be analyzed to identify or detect objects or to determine a shape or distance of objects within the field of regard.
- a depth map may cover a field of regard that extends 60° horizontally and 15° vertically, and the depth map may include a frame of 100-2000 pixels in the horizontal direction by 4-400 pixels in the vertical direction.
- the lidar system 100 may be configured to repeatedly capture or generate point cloud frames of a field of regard at any suitable frame rate between approximately 0.1 frames per second (FPS) and approximately 1,000 FPS, for example.
- the point cloud frame rate may be substantially fixed or dynamically adjustable, depending on the implementation.
- the lidar system 100 can use a slower frame rate (e.g., 1 Hz) to capture one or more high-resolution point clouds, and use a faster frame rate (e.g., 10 Hz) to rapidly capture multiple lower-resolution point clouds.
- the field of regard of the lidar system 100 can overlap, encompass, or enclose at least a portion of the target 130 , which may include all or part of an object that is moving or stationary relative to lidar system 100 .
- the target 130 may include all or a portion of a person, vehicle, motorcycle, truck, train, bicycle, wheelchair, pedestrian, animal, road sign, traffic light, lane marking, road-surface marking, parking space, pylon, guard rail, traffic barrier, pothole, railroad crossing, obstacle in or near a road, curb, stopped vehicle on or beside a road, utility pole, house, building, trash can, mailbox, tree, any other suitable object, or any suitable combination of all or part of two or more objects.
- FIGS. 1A and 1B illustrate example scan patterns 160 , either or both of which the lidar system 100 of FIG. 2 may produce.
- the lidar system 100 may be configured to scan the output optical beam 125 along the scan pattern 160 .
- the scan pattern 160 corresponds to a scan across any suitable field of regard having any suitable horizontal field of regard (FOR H ) and any suitable vertical field of regard (FOR V ).
- FOR H horizontal field of regard
- FOR V any suitable vertical field of regard
- a certain scan pattern may have a field of regard represented by angular dimensions (e.g., FOR H ⁇ FOR V ) 40° ⁇ 30°, 90° ⁇ 40°, or 60° ⁇ 15°. While FIG.
- the “looping” pattern 160 depicted in FIG. 1B includes co-located pixels 166 (that is, a pair of co-located pixels at 166 a , and a different pair co-located pixels at 166 b ) to capture depth data at the same horizontal and vertical angles in the field of regard at multiple points of time within the same point cloud frame.
- co-located is generally used to refer to pixels at the same location within the field of regard, in some implementations, co-located refers to pixels within a small, threshold distance of one another. In some implementations this threshold distance is defined as a difference in the vertical and/or horizontal angles.
- reference line 162 represents a center of the field of regard of scan pattern 160 . Accordingly, if the scan pattern 160 has a 60° ⁇ 15° field of regard, then the scan pattern 160 covers a ⁇ 30° horizontal range with respect to reference line 162 and a ⁇ 7.5° vertical range with respect to reference line 162 .
- An azimuth (which may be referred to as an azimuth angle) may represent a horizontal angle with respect to the reference line 162
- an altitude (which may be referred to as an altitude angle, elevation, or elevation angle) may represent a vertical angle with respect to the reference line 162 .
- the scan pattern 160 may include multiple points or pixels 164 , and each pixel 164 may be associated with one or more laser pulses and one or more corresponding distance measurements.
- a cycle of scan pattern 160 may include a total of P x ⁇ P y pixels 164 (e.g., a two-dimensional distribution of P x by P y pixels).
- the number of pixels 164 along a horizontal direction may be referred to as a horizontal resolution of the scan pattern 160
- the number of pixels 164 along a vertical direction may be referred to as a vertical resolution of the scan pattern 160 .
- Each pixel 164 may be associated with a distance (e.g., a distance to a portion of a target 130 from which the corresponding laser pulse was scattered), one or more angular values.
- the pixel 164 may be associated with a distance value and two angular values (e.g., an azimuth and altitude) that represent the angular location of the pixel 164 with respect to the lidar system 100 .
- a distance to a portion of the target 130 may be determined based at least in part on a time-of-flight measurement for a corresponding pulse.
- An angular value (e.g., an azimuth or altitude) may correspond to an angle (e.g., relative to reference line 162 ) of the output beam 125 (e.g., when a corresponding pulse is emitted from lidar system 100 ) or an angle of the input beam 135 (e.g., when an input signal is received by lidar system 100 ).
- the lidar system 100 determines an angular value based at least in part on a position of a component of the scanner 120 . For example, an azimuth or altitude value associated with the pixel 164 may be determined from an angular position of one or more corresponding scanning mirrors of the scanner 120 .
- each pixel 164 is also associated with a time value, such as a time stamp indicative of when an output beam (e.g., the output beam 125 ) is emitted or when an input beam (e.g., the input beam 135 ) is detected.
- a time value such as a time stamp indicative of when an output beam (e.g., the output beam 125 ) is emitted or when an input beam (e.g., the input beam 135 ) is detected.
- FIG. 3A illustrates an example vehicle 300 with a lidar system 302 .
- the lidar system 302 includes a laser 310 with multiple sensor heads 312 A-D coupled to the laser 310 via multiple laser-sensor links 314 .
- Each of the sensor heads 312 may include some or all of the components of the lidar system 100 illustrated in FIG. 2 .
- Each of the laser-sensor links 314 may include one or more optical links and/or one or more electrical links.
- the sensor heads 312 in FIG. 3A are positioned or oriented to provide a greater than 30-degree view of an environment around the vehicle. More generally, a lidar system with multiple sensor heads may provide a horizontal field of regard around a vehicle of approximately 30°, 45°, 60°, 90°, 120°, 180°, 270°, or 360°.
- Each of the sensor heads 312 may be attached to, or incorporated into, a bumper, fender, grill, side panel, spoiler, roof, headlight assembly, taillight assembly, rear-view mirror assembly, hood, trunk, window, or any other suitable part of the vehicle.
- each of the sensor heads 312 may be incorporated into a light assembly, side panel, bumper, or fender), and the laser 310 may be located within the vehicle 300 (e.g., in or near the trunk).
- the four sensor heads 312 may each provide a 90° to 120° horizontal field of regard (FOR), and the four sensor heads 312 may be oriented so that together they provide a complete 360-degree view around the vehicle.
- the lidar system 302 may include six sensor heads 312 positioned on or around the vehicle 300 , where each of the sensor heads 312 provides a 60° to 90° horizontal FOR.
- the lidar system 302 may include eight sensor heads 312 , and each of the sensor heads 312 may provide a 45° to 60° horizontal FOR. As yet another example, the lidar system 302 may include six sensor heads 312 , where each of the sensor heads 312 provides a 70° horizontal FOR with an overlap between adjacent FORs of approximately 10°. As another example, the lidar system 302 may include two sensor heads 312 which together provide a forward-facing horizontal FOR of greater than or equal to 30°.
- Data from each of the sensor heads 312 may be combined or stitched together to generate a point cloud that covers a greater than or equal to 30-degree horizontal view around a vehicle.
- the laser 310 may include a controller or processor that receives data from each of the sensor heads 312 (e.g., via a corresponding electrical link 320 ) and processes the received data to construct a point cloud covering a 360-degree horizontal view around a vehicle or to determine distances to one or more targets.
- the point cloud or information from the point cloud may be provided to a vehicle controller 322 via a corresponding electrical, optical, or radio link 320 .
- the vehicle controller 322 may include one or more CPUs, GPUs, and a non-transitory memory with persistent components (e.g., flash memory, an optical disk) and/or non-persistent components (e.g., RAM).
- the point cloud is generated by combining data from each of the multiple sensor heads 312 at a controller included within the laser 310 , and is provided to the vehicle controller 322 .
- each of the sensor heads 312 includes a controller or processor that constructs a point cloud for a portion of the 360-degree horizontal view around the vehicle and provides the respective point cloud to the vehicle controller 322 .
- the vehicle controller 322 then combines or stitches together the points clouds from the respective sensor heads 312 to construct a combined point cloud covering a 360-degree horizontal view.
- the vehicle controller 322 in some implementations communicates with a remote server to process point cloud data.
- respective controllers included within the multiple sensor heads 312 process their respectively constructed point clouds to detect, analyze, and/or correct for distortion, including rolling shutter distortion, prior to providing the constructed point cloud to the laser 310 or the vehicle controller 312 .
- a controller included within the laser 310 processes the point cloud to correct for detect, analyze, and/or correct distortion, including rolling shutter distortion, prior to providing the point cloud to the vehicle controller 322 .
- the laser 310 may analyze the constructed point clouds to detect and correct for distortion prior to combining or stitching the point clouds together.
- the laser 310 detects, analyzes, and/or corrects for the distortion while combining the raw data from the multiple sensor heads.
- the vehicle controller 322 analyzes point clouds data provided by the laser 310 or the multiple sensor heads 312 to detect, analyze, and/or correct for distortion, including rolling shutter distortion, prior to processing the point cloud via the various components of the vehicle controller 322 described herein.
- the vehicle 300 may be an autonomous vehicle where the vehicle controller 322 provides control signals to various components 330 within the vehicle 300 to maneuver and otherwise control operation of the vehicle 300 .
- the components 330 are depicted in an expanded view in FIG. 3A for ease of illustration only.
- the components 330 may include an accelerator 340 , brakes 342 , a vehicle engine 344 , a steering mechanism 346 , lights 348 such as brake lights, head lights, reverse lights, emergency lights, etc., a gear selector 350 , and/or other suitable components that effectuate and control movement of the vehicle 300 .
- the gear selector 350 may include the park, reverse, neutral, drive gears, etc.
- Each of the components 330 may include an interface via which the component receives commands from the vehicle controller 322 such as “increase speed,” “decrease speed,” “turn left 5 degrees,” “activate left turn signal,” etc. and, in some cases, provides feedback to the vehicle controller 322 .
- the vehicle controller 322 receives point cloud data from the sensor heads 312 or laser 310 via the link 320 and analyzes the received point cloud data to sense or identify targets 130 (see FIG. 1 ) and their respective locations, distances, speeds, shapes, sizes, type of target (e.g., vehicle, human, tree, animal), etc.
- the vehicle controller 322 then provides control signals via the link 320 to the components 330 to control operation of the vehicle based on the analyzed information.
- One, some or all of the components 330 may be the operational subsystems, or may be included within the operational subsystems, that receive the control signals.
- the vehicle 300 may also be equipped with other sensors such as a RGB or stereo camera, a thermal imager, a conventional radar (none illustrated to avoid clutter), etc.
- the sensors can provide additional data to the vehicle controller 322 via wired or wireless communication links.
- the vehicle 300 in an example implementation includes a microphone array operating as a part of an acoustic source localization system configured to determine sources of sounds.
- FIG. 3B illustrates a vehicle 360 in which a laser 370 is optically coupled to six sensor heads 372 , each of which may be similar to one of the sensor heads 312 of FIG. 3A .
- the sensor heads 372 A and 372 G are disposed at the front of the hood, the sensor heads 372 B and 372 F are disposed in the side view mirrors, and the sensor heads 372 C-E are disposed on the trunk.
- the sensor head 372 D is oriented to face backward relative to the orientation of the vehicle 360
- the sensor heads 372 C-E are oriented at approximately 45 degrees relative to the axis of orientation of the sensor head 372 D.
- FIG. 4 depicts an example point cloud 390 that is generated by a lidar system scanning the vehicle environment (e.g., the lidar system 100 of FIGS. 1 and 2 or the lidar system 302 of FIG. 3A ).
- the vehicle environment may include a divided highway that has a median separating the two directions of traffic, with multiple lanes in each direction.
- the vehicle environment may also include light poles on either side of the road and/or in the median.
- the point cloud 390 of FIG. 4 corresponds to an example embodiment in which a lidar device captures a roughly 60 degree horizontal field of regard in front of the vehicle 300 .
- the point cloud 390 may have been generated using the sensor heads 312 A or 312 D of FIG. 3A , or the sensor heads 372 A or 372 G of FIG. 3B , for example. While depicted as a visual image in FIG. 4 , it is understood that, in some embodiments, the point cloud 390 is not actually rendered or displayed at any time.
- the point cloud 390 depicts a ground plane 394 (here, the road surface) as a number of substantially continuous scan lines, and also depicts, above the ground plane 394 , a number of objects 396 (also referred to as “point cloud objects”). For clarity, only a small number of the objects shown in FIG. 4 are labeled with a reference number. Distinct ones of the objects 396 within the point cloud 390 may be identified by a segmentation module of the vehicle controller 322 of FIG. 3A . For example, the segmentation module may detect substantial gaps and/or other discontinuities in the scan lines of the ground plane 394 , and identify groups of points in the vicinity of those discontinuities as discrete objects.
- the segmentation module may determine which points belong to the same object using any suitable rules, algorithms or models. Once the objects 396 are identified, a classification module of the vehicle controller 322 of FIG. 3A may attempt to classify the objects, and a tracking module of the vehicle controller 322 of FIG. 3A may attempt to track the classified objects (and, in some embodiments/scenarios, unclassified objects) across future point clouds similar to point cloud 390 (i.e., across multiple point cloud frames).
- a vehicle 396 A may be relativity easy to identify and classify due to the high density of points, but more difficult to classify due to the lack of points within the lidar shadow (i.e., the absence or relative scarcity of points) along the lower right portions of the vehicle 396 C created by the relative positioning of the vehicle 396 C with respect to the vehicle 300 .
- a light pole 396 B may also be relatively easy to identify as an object, due to the high density of points and readily identifiable shape, but may not be as easy to classify due to the fact that large portions of the light pole 396 B are distorted (e.g., as shown by the middle portion leaning away from the vehicle 300 ).
- the van 396 C may be relatively easy to identify as an object, but more difficult to classify due to distortion causing the van 396 C to appears as though it is tilted away from the vehicle 300 .
- Still other objects may be difficult to identify, classify and/or track due to their small size and/or low profile.
- the segmentation module may identify (and the classification module may classify) lane markings within the point cloud 390 .
- the lane markings may appear as small but abrupt deviations in the path of the scan lines, for example, with those deviations collectively forming a line pattern that aligns with the direction of travel of the autonomous vehicle (e.g., approximately normal to the curve of the scan lines).
- the segmentation module, classification module and/or tracking module may use techniques that make object identification, classification and/or tracking highly accurate across a very wide range of scenarios, with scarce or otherwise suboptimal point cloud representations of objects.
- the segmentation module, classification module and/or tracking module may include neural networks that were trained using data (e.g., manually labeled scenes) corresponding to a very large number of diverse environments/scenarios (e.g., with various types of objects at different distances, in different orientations, with different degrees of concealment, in different weather and/or lighting conditions, and so on).
- a distortion correction module of the vehicle controller 322 analyzes the identified objects 396 to detect and correct distortion. For example, as depicted in FIG. 4 , the light pole 396 B and the van 396 C exhibit distortion causing the light pole 396 B and van 396 C to appear as though they are skewed away from the vehicle 300 . Accordingly, the distortion correction module may detect the distortion, modify the point cloud 390 to correct for the distortion, analyze the distortion to determine the velocity of one or more objects 396 , etc. Correcting the distortion of the objects 396 may improve the accuracy of the classification module and/or the tracking module (see description of FIG. 5 , below), for example. Ultimately, improving the correlation between the point cloud 390 and the actual vehicle environment may provide a smoother ride for passengers and/or reduce the likelihood of a collision.
- FIG. 5 illustrates an example software architecture 400 , which may be utilized in the lidar system 100 or 302 of FIG. 2 or 3 , respectively.
- the software architecture 400 receives as input M sets of sensor data 402 generated by M different sensors, with M being any suitable integer equal to or greater than one.
- the sensor data 402 may be data generated by the sensor heads 312 of FIG. 3 or any of the other non-illustrated sensors included in the lidar system 302 of FIG. 3 .
- the lidar system may include cameras, radar devices, thermal imaging units, IMUs, and/or other sensor types.
- “sensor data 1 ” may include frames of point cloud data generated by a first sensor head 312 A
- “sensor data 2 ” may include frames of point cloud data generated by a second sensor head 312 B
- “sensor data 3 ” (not shown in FIG. 3 ) may include frames of digital images generated by a RGB camera, and so on.
- the sensor data 402 is input to a perception component 406 of the software architecture 400 , and is processed by the perception component 406 to generate perception signals 408 descriptive of a current state of the autonomous vehicle's environment. It is understood that the term “current” may actually refer to a very short time prior to the generation of any given perception signals 408 , e.g., due to the short processing delay introduced by the perception component 406 and other factors.
- the perception component may include a segmentation module 410 , a classification module 412 , a tracking module 414 , and a distortion correction module 416 .
- the segmentation module 410 is generally configured to identify distinct objects within the sensor data 402 representing the sensed environment. Depending on the embodiment and/or scenario, the segmentation task may be performed separately for each of a number of different types of sensor data, or may be performed jointly on a fusion of multiple types of sensor data. In some embodiments where lidar devices are used, the segmentation module 410 analyzes point cloud frames to identify subsets of points within each frame that correspond to probable physical objects in the environment. In other embodiments, the segmentation module 410 jointly analyzes lidar point cloud frames in conjunction with camera image frames to identify objects in the environment. Other suitable techniques, and/or data from other suitable sensor types, may also be used to identify objects.
- references to different or distinct “objects” may encompass physical things that are entirely disconnected (e.g., with two vehicles being two different “objects”), as well as physical things that are connected or partially connected (e.g., with a vehicle being a first “object” and the vehicle's hitched trailer being a second “object”).
- the segmentation module 410 may use predetermined rules or algorithms to identify objects. For example, the segmentation module 410 may identify as distinct objects, within a point cloud, any clusters of points that meet certain criteria (e.g., having no more than a certain maximum distance between all points in the cluster, etc.). Alternatively, the segmentation module 410 may utilize a neural network that has been trained to identify distinct objects within the environment (e.g., using supervised learning with manually generated labels for different objects within test data point clouds, etc.), or another type of machine learning based model. For example, the segmentation module 410 may analyze the point cloud 390 of FIG. 4 to identify objects 396 .
- the classification module 412 is generally configured to determine classes (labels, categories, etc.) for different objects that have been identified by the segmentation module 410 . Like the segmentation module 410 , the classification module 412 may perform classification separately for different sets of the sensor data 402 , or may classify objects based on data from multiple sensors, etc. Moreover, and also similar to the segmentation module 410 , the classification module 412 may execute predetermined rules or algorithms to classify objects, or may utilize a neural network or other machine learning based model to classify objects. For example, the classification module 412 may classify the object 396 C of FIG. 4 as a truck and the object 396 A as a car.
- the classification module 412 communicates with an object database (not depicted) that stores information associated with object types.
- the object database may include information that indicates how an object of the corresponding object type should appear in a point cloud.
- this indication may be a generic model for an object of the particular type.
- an object database record for a particular model of car may include a three-dimensional model of the car, to which objects identified by the segmentation module 410 are compared during the classification process.
- the model includes indications of particular features of the object that have known shapes (e.g., a license plate, a tire, manufacturer emblem, etc.). As will be described below, this model of the object, including the particular features thereof, may be used to detect whether objects identified by the segmentation module 410 actually have a skewed shape, or instead have a distorted appearance in the point cloud frame due to rolling shutter distortion.
- the tracking module 414 is generally configured to track distinct objects over time (e.g., across multiple lidar point cloud or camera image frames).
- the tracked objects are generally objects that have been identified by the segmentation module 410 , but may or may not be objects that were classified by the classification module 412 , depending on the embodiment and/or scenario.
- the segmentation module 410 may assign identifiers to identified objects, and the tracking module 414 may associate existing identifiers with specific objects where appropriate (e.g., for lidar data, by associating the same identifier with different clusters of points, at different locations, in successive point cloud frames)
- the tracking module 414 may perform separate object tracking based on different sets of the sensor data 402 , or may track objects based on data from multiple sensors.
- the tracking module 414 may execute predetermined rules or algorithms to track objects, or may utilize a neural network or other machine learning model to track objects.
- the distortion correction module 416 is generally configured to detect that an object is distorted, analyze distortion to determine a relative velocity of the object, and/or correct distorted objects such that they no longer appear distorted.
- newly identified objects by the segmentation module 410 are analyzed by the distortion correction module 416 .
- the distortion correction module 416 jointly analyzes lidar point cloud frames in conjunction with camera image frames to determine whether objects are distorted.
- the newly identified objects are analyzed by the distortion correction module 416 prior to the objects being analyzed by the classification module 412 and/or the tracking module 414 . Additionally or alternatively, objects tracked by the tracking module 414 are analyzed by the distortion correction module 416 when the sensor data 402 indicates a new point cloud frame.
- the distortion correction module 416 analyzes the amount of distortion of an object to determine a velocity or relative velocity of at least one of the distorted object, another object, or the vehicle. Moreover, and also similar to the segmentation module 410 , the classification module 412 , and the tracking module 414 , the distortion correction module 416 may execute predetermined rules or algorithms, or may utilize a neural network or other machine learning model to detect distortion, correct distortion, and/or correlate an amount of distortion to relative velocity.
- the software architecture 400 also includes a prediction component 420 , which processes the perception signals 408 to generate prediction signals 422 descriptive of one or more predicted future states of the autonomous vehicle's environment.
- the prediction component 420 may analyze the type/class of the object (as determined by the classification module 412 ) along with the recent tracked movement of the object (as determined by the tracking module 414 ) to predict one or more future positions of the object.
- the prediction component 420 may assume that any moving objects will continue to travel on their current direction and with their current speed, possibly taking into account first- or higher-order derivatives to better track objects that have continuously changing directions, objects that are accelerating, and so on.
- the prediction component 420 also predicts movement of objects based on more complex behaviors. For example, the prediction component 420 may assume that an object that has been classified as another vehicle will follow rules of the road (e.g., stop when approaching a red light), and will react in a certain way to other dynamic objects (e.g., attempt to maintain some safe distance from other vehicles). The prediction component 420 may inherently account for such behaviors by utilizing a neural network or other machine learning model, for example. The prediction component 420 may be omitted from the software architecture 400 , in some embodiments.
- a mapping component 430 obtains map data (e.g., a digital map including the area currently being traversed by the autonomous vehicle) and/or navigation data (e.g., data indicating a route for the autonomous vehicle to reach the destination, such as turn-by-turn instructions), and outputs the data (possibly in a converted format) as mapping and navigation signals 432 .
- the mapping and navigation signals 432 include other map- or location-related information, such as speed limits, traffic indicators, and so on.
- the signals 432 may be obtained from a remote server (e.g., via a cellular or other communication network of the autonomous vehicle, or of a smartphone coupled to the autonomous vehicle, etc.), and/or may be locally stored in a persistent memory of the autonomous vehicle.
- a motion planner 440 processes the perception signals 408 , the prediction signals 422 , and the mapping and navigation signals 432 to generate decisions 442 regarding the next movements of the autonomous vehicle.
- the motion planner 440 may use any suitable architecture or technique to generate the decisions 442 , such as one or more trained machine learning models, or a model predictive control (MPC) technique, for example.
- the decisions 442 may be operational parameters (e.g., braking, speed and steering parameters) or particular maneuvers (e.g., turn left, move to right lane, move onto shoulder of road, etc.).
- the decisions 442 may be provided to one or more components of the autonomous vehicle (e.g., if the decisions 442 indicate specific operational parameters) such as the components 330 of FIG. 3 , or may be provided to one or more intermediate stages that convert the decisions 442 to operational parameters (e.g., if the decisions indicate specific maneuvers) prior to providing the operation parameters to the vehicle components 330 .
- FIG. 6 is a block diagram of an example computing system 500 for controlling an autonomous vehicle.
- the computing system 500 may be integrated within an autonomous vehicle in any suitable manner, and at any suitable location or locations within the vehicle.
- the computing system 500 may be included, or partially included, within the vehicle controller 322 of FIG. 3 , for example.
- the computing system 500 includes one or more processors 502 , and a memory 504 storing vehicle controller instructions 506 .
- the one or more processors 502 execute the vehicle controller instructions 506 to implement the disclosed techniques, including, in one embodiment, the software architecture 400 of FIG. 5 .
- each processor may be a different programmable microprocessor that executes software instructions stored in the memory 504 .
- each of the processor(s) 502 may be a different set of such microprocessors, or a set that includes one or more microprocessors and one or more other processor types (e.g., ASICs, FPGAs, etc.) for certain functions.
- the memory 504 may include one or more physical memory devices with non-volatile memory. Any suitable memory type or types may be used, such as ROM, solid-state drives (SSDs), hard disk drives (HDDs), and so on.
- the processor(s) 502 are coupled to the memory 504 via a bus or other network 508 .
- the network 508 may be a single wired network, or may include any suitable number of wired and/or wireless networks.
- the network 508 may be or include a controller area network (CAN) bus, a Local Interconnect Network (LNN) bus, and so on.
- CAN controller area network
- LNN Local Interconnect Network
- a vehicle control interface 510 Also coupled to the network 508 are a vehicle control interface 510 , a passenger interface 512 , a sensor interface 514 , and a network interface 516 .
- Each of the interfaces 510 , 512 , 514 and 516 may include one or more processors (e.g., ASICs, FPGAs, microprocessors, etc.) and/or other hardware, firmware and/or software to enable communication with systems, subsystems, devices, etc., that are external to the computing system 500 .
- processors e.g., ASICs, FPGAs, microprocessors, etc.
- the vehicle control interface 510 is generally configured to provide control data generated by the processor(s) 502 to the appropriate operational subsystems of the autonomous vehicle, such that the appropriate subsystems can effectuate driving decisions made by the processor(s) 502 .
- the vehicle control interface 510 may provide the control signal operational parameters associated with the decisions 442 to the appropriate subsystem(s) (e.g., accelerator 340 , brakes 342 , and steering mechanism 346 of FIG. 3A ).
- the vehicle control interface 510 includes separate interface hardware, firmware and/or software for different operational subsystems.
- the passenger interface 512 is generally configured to provide alerts, warnings, notifications, and/or other information to one or more passengers of the autonomous vehicle.
- the interface 512 may specifically provide such information to the driver (e.g., via dashboard indicators, etc.).
- the passenger interface 512 may cause a display and/or speaker in the vehicle to generate an alert when the processor(s) 502 (executing the vehicle controller instructions 506 ) determine that a collision with another object is likely.
- the passenger interface 502 may cause a display in the vehicle to show an estimated time of arrival (ETA) to passengers.
- ETA estimated time of arrival
- the passenger interface 512 also permits certain user inputs.
- the passenger interface 512 may cause a display to present a virtual control (e.g., button) that a passenger may activate (e.g., touch, scroll through, etc.) to select a particular driving style.
- a virtual control e.g., button
- the sensor interface 514 is generally configured to convert raw sensor data from one or more sensor devices (e.g., lidar, camera, microphones, thermal imaging units, IMUs, etc.) to a format that is consistent with a protocol of the network 508 and is recognized by one or more of the processor(s) 502 .
- the sensor interface 514 may be coupled to the lidar system 100 of FIG. 1 or 2 or the lidar system 302 of FIG. 3A , for example, with the sensor interface 514 converting point cloud data to an appropriate format.
- the sensor interface 514 includes separate interface hardware, firmware and/or software for each sensor device and/or each sensor type.
- the network interface 516 is generally configured to convert data received from one or more devices or systems external to the autonomous vehicle to a format that is consistent with a protocol of the network 508 and is recognized by one or more of the processor(s) 502 .
- the network interface 516 includes separate interface hardware, firmware and/or software for different external sources.
- a remote mapping/navigation server may send mapping and navigation/route data to the computing system 500 via a cellular network interface of the network interface 516
- one or more peer vehicles e.g., other autonomous vehicles
- data e.g., current positions of the other vehicles
- the computing system 500 may use the network interface 516 to receive data representing rules or regulations (e.g., speed limits), object positions (e.g., road rails, overhanging signage, etc.), and/or other information from various infrastructure devices or systems.
- rules or regulations e.g., speed limits
- object positions e.g., road rails, overhanging signage, etc.
- other information e.g., information from various infrastructure devices or systems.
- the vehicle controller instructions 506 utilize external data that is received by the network interface 516 from other vehicles, infrastructure, and/or other external devices/systems as an input to one or more components.
- the external data may include raw sensor data that is indicative of the vehicle environment (but was generated off-vehicle), and/or may include higher-level information that was generated externally using raw sensor data.
- the network 508 may also couple to other types of interfaces and/or components, and/or some of the interfaces shown in FIG. 6 may be omitted (e.g., the network interface 516 ).
- the computing system 500 represents just one possible configuration for supporting the software architectures, functions, features, etc., described herein, and that others are also within the scope of this disclosure.
- FIG. 7 depicts an example method 600 for determining the relative velocity of a point cloud object by fitting pixels onto a surface of the point cloud object, for example, one of the objects 396 of the point cloud 390 depicted in FIG. 4 .
- the method 600 may be implemented/performed by the computing system 500 of FIG. 6 (e.g., by the processor(s) 502 when executing the vehicle controller instructions 506 stored in memory 504 ), for example, by another suitable computing device or system (such as a controller of the laser 310 or the sensor heads 312 of FIG. 3A , or the software architecture 400 of FIG. 5 ), or by a combination thereof.
- sensor data is received.
- the sensor data may include lidar depth data (e.g., sensed by the lidar system 100 of FIGS. 1 and 2 , or the lidar system 302 of FIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and/or any other type of sensor data discussed above in connection with the sensor data 402 of FIG. 5 , for example.
- a point cloud frame based on the sensor data received at block 602 is obtained.
- the sensor data received at block 602 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud).
- block 604 may include extracting the point cloud frame from rest of the sensor data received at block 602 .
- block 604 includes analyzing depth or lidar data included in the sensor data received at block 602 to generate the point cloud frame, as discussed in connection to FIG. 3 , for example.
- two or more pixels that overlap a point cloud object are selected from the point cloud frame.
- the two points correspond to the same object as identified by a segmentation module.
- the two points are selected based on being within a threshold distance of one another.
- the selected points may be located on or near a two-dimensional surface corresponding to at least part of the point cloud object and captured at different times.
- Each of the pixels may be associated with depth data and time data.
- pixels 164 a and 164 b of FIG. 1A overlap the same point cloud object and are selected at block 606 .
- the pixels 164 a and 164 b are fitted to a two-dimensional plane used to model the object's two-dimensional surface.
- the general equation for a plane moving through an environment is represented by the following equation:
- “fitting” the pixels to a plane includes modeling the point cloud object's surface as a two-dimensional plane moving through the vehicle environment such that the pixels are included within (i.e., satisfy the equation for) the two-dimension plane.
- a relative velocity of the point cloud object is determined based on depth and time data associated with the selected pixels.
- the plane is not moving along the vertical axis with respect to the vehicle. Using this assumption, the motion of the two-dimensional plane may be represented using the following equation:
- a′, b′, and c′ are constants
- x, y, and z are indicative of position with respect to the x, y, and z axes (such as the longitudinal, lateral, and vertical axes), respectively
- v x and v y are the component of the plane velocity in the x and y directions respectively.
- the time and depth data associated with the first and second pixels can then be plugged into Equation 3 to solve for v x (as well as a′). That is, depth data x 1 captured at time t 1 and depth data x 2 captured at time t 2 associated with the first and second pixels is input into a best fit algorithm (such as a least square sum algorithm, a polynomial regression algorithm, etc.) to solve for v x . Because it is assumed that the plane is only moving in the longitudinal direction, v x is the relative velocity of the point cloud object.
- a best fit algorithm such as a least square sum algorithm, a polynomial regression algorithm, etc.
- three or more pixels are selected at block 606 .
- the equation of the plane determined with respect to the first two selected pixels is rotated about a single axis to fit the third pixel into the plane. Said another way, rotating the plane about a single axis to fit the third pixel solves for one of b′ or c′. Assuming that the other constant is 0, then Equation 1 can be solved for v x and v y or v x and v z (as well as a′) using another best fit algorithm that utilizes the time and depth data for the three selected pixels as an input.
- four or more pixels are selected at block 606 .
- the equation of the plane determined with respect to the first two selected pixels is rotated about two axes to fit the third and fourth pixels into the plane. Said another way, rotating the plane about two axes to fit the third and fourth pixels solves for both of b′ and c′. Accordingly, Equation 1 can be solved for v x , v y and v z (as well as a′) using another best fit algorithm that utilizes the time and depth data for the four selected pixels as an input.
- block 608 assumes that the point cloud object can have any shape.
- the relative velocity of the point cloud object can be calculated using the following equation:
- ⁇ d is the depth difference between the two selected pixel and ⁇ t is the time difference between when the depth data for the two pixels was captured.
- the first and second pixels are part of the same scan line that traverses the vehicle environment.
- the first pixel is part of a first scan line of pixels that are displaced from one another in a first direction and the second pixel is part of a second scan line of pixels that is substantially parallel to the first scan line and comprised of pixels that are displaced from one another in a second direction substantially orthogonal to the first direction.
- the method 600 includes one or more additional blocks not shown in FIG. 7 .
- another two or more pixels that overlap the point cloud object in a different region may be selected.
- This additional block may then calculate a relative velocity of this different region using the techniques described above with respect to blocks 606 and 608 . Accordingly, this additional block may include comparing the two different relative velocities for the two different regions to determine that the point cloud object is rotating with respect to the vehicle, for example, by determining that the relative velocities are not the same.
- the method 600 may include a block to detect that the point cloud object is distorted. Accordingly, at block 606 , two or more pixels that are part of a scan line that traverses the vehicle environment in a first direction (e.g., pixels 164 a and 164 b of FIG. 1A ) and two or more pixels that are part of a scan line that traverses the vehicle environment in a second direct (e.g., pixels 164 c and 164 d ) are selected.
- This additional block may include calculating a first relative velocity using the first two pixels and a second relative velocity using the second two pixels, for example by using Equation 4.
- this additional block includes comparing the magnitude of the two relative velocities.
- a rate of change between the relative velocities is analyzed to determine a relative acceleration.
- one of the first relative velocity, the second relative velocity, or an average relative velocity therebetween is associated with the point cloud object, for example, by the perception component 406 of FIG. 5 . It should be appreciated that if the magnitude of one of the relative velocities exceeds a threshold value, then it is likely that one of the pixels does not actually overlap the point cloud object. Therefore, relative velocities that exceed this threshold may be discarded.
- the method 600 may include correcting the rolling shutter distortion based on the relative velocity determined at block 608 .
- a machine learning module e.g., one that is trained using reinforcement learning
- the machine learning model may operate on (i.e., use as inputs) the relative velocity determined at block 608 and the depth data of the selected pixels to determine a correction factor to compensate for delay introduced by successive pixels and/or scan lines included in a scan pattern.
- Another additional block of the method 600 may include determining an object velocity (as opposed to relative velocity). For example, a relative velocity of a point cloud object that is known to be stationary (e.g., the point cloud object is classified as a light pole, a street sign, a tree, a bench, a power line, a building, a median, etc.) may be determined using the techniques of blocks 606 and 608 . Because the point cloud object is known to be stationary, the relative velocity is comprised solely of the vehicle's velocity. Said another way, by determining the relative velocity with respect to a known, stationary object, the vehicle velocity can be determined. Subsequently, the vehicle velocity can be subtracted from the relative velocity determined with respect to the identified point cloud object to determine object velocity.
- a relative velocity of a point cloud object that is known to be stationary e.g., the point cloud object is classified as a light pole, a street sign, a tree, a bench, a power line, a building, a median, etc.
- the point cloud object (either corrected or uncorrected) and the relative velocity may be analyzed, for example by a portion of the perception component 406 of FIG. 5 (e.g., classification module 412 and/or tracking module 414 ) and/or the prediction component 420 of FIG. 5 , to classify the object, track the object, and/or predict a future state of the vehicle environment.
- This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the current perceived state and/or the predicted future state.
- the block may include generate a control signal to apply the vehicle brakes.
- FIGS. 8, 9, and 10 depict example methods 700 , 720 , and 740 , respectively, for determining the relative velocity of a point cloud object, for example, one of the objects 396 of the point cloud 390 depicted in FIG. 4 .
- the method 700 of FIG. 8 relates to embodiments that determine relative velocity based on a scan pattern that includes successive scan lines that overlap the object
- the method 720 of FIG. 9 relates to embodiments that determine relative velocity based on depth data associated with co-located pixels. Both of the methods 700 and 720 rely on data captured during a single point cloud frame.
- the method 740 of FIG. 10 relates to embodiments that correct point cloud distortion based on tracked object velocity.
- the methods 700 , 720 , and 740 may be implemented/performed by the computing system 500 of FIG. 6 (e.g., by the processor(s) 502 when executing the vehicle controller instructions 506 stored in memory 504 ), for example, by another suitable computing device or system (such as a controller of the laser 310 or the sensor heads 312 of FIG. 3A , or the software architecture 400 of FIG. 5 ), or by a combination thereof.
- the computing system 500 of FIG. 6 e.g., by the processor(s) 502 when executing the vehicle controller instructions 506 stored in memory 504
- another suitable computing device or system such as a controller of the laser 310 or the sensor heads 312 of FIG. 3A , or the software architecture 400 of FIG. 5 , or by a combination thereof.
- sensor data is received at block 702 .
- the sensor data may include lidar depth data (e.g., sensed by the lidar system 100 of FIGS. 1 and 2 , or the lidar system 302 of FIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with the sensor data 402 of FIG. 5 , for example.
- a point cloud frame based on the sensor data received at block 702 is obtained.
- the sensor data received at block 702 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud).
- block 704 includes extracting the point cloud frame from the rest of the sensor data received at block 702 .
- block 704 includes analyzing depth or lidar data included in the sensor data received at block 702 to generate the point cloud frame, as discussed in connection to FIG. 3 , for example.
- a point cloud object within the point cloud frame obtained at block 704 is identified.
- block 706 includes partitioning the point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and/or tracking movement of classified objects over time, for example, by utilizing a perception component 406 of the vehicle controller 400 of FIG. 5 .
- a tracking module 414 of the perception component 406 tracks objects across multiple point cloud frames.
- the tracking module 414 identifies objects within the point cloud frame obtained at block 704 that are currently being tracked.
- the segmentation module 410 identifies a new object that is not being tracked by the tracking module 416 .
- the point cloud object identified at block 706 is determined to be skewed relative to an expected configuration of the point cloud object.
- the point cloud object may be skewed due to the above-described rolling shutter distortion.
- the point cloud object is classified (for example by a classification module 412 of the perception component 406 ) as an object that has a particular shape (e.g., an Acme Sedan, a stop sign, a person, etc.). Accordingly, determining that the object is skewed may include comparing a bound of the point cloud object to a bound of a reference object associated with the classification module.
- the expected configuration is determined based upon the relative position between the vehicle and the object.
- the classification module 412 may associate the point cloud object with a generic object of the same type of object.
- the generic object may be rotated and/or scaled based on the determined relative position to determine the expected configuration of the point cloud object. Skew may then be determined by comparing the expected configuration and the apparent/sensed configuration, and identifying a substantial difference (e.g., greater than a threshold difference) between the two.
- it is determined that the object is skewed because a bound of the point cloud object is determined to be curved when it is known that the point cloud object should instead have a straight bound.
- the classification module 412 may have classified the point cloud object as a type associated with a rule that side bounds should be approximately vertical (e.g., a truck container).
- the expected configuration of that bound is a line.
- the expected configuration is determined based upon map data of the environment through which the vehicle is moving.
- the vehicle controller may be configured to receive mapping data via the network interface 516 .
- the mapping data may include a plurality of images associated with location and/or geoposition data. Accordingly, for objects that are stationary and/or are generally permanent in the environment, the mapping data may include image data that is representative of the actual configuration of the object.
- the representation of the object as included in the mapping data can be identified.
- the point cloud object may be compared to the corresponding representation in the mapping data to determine a relative viewing angle between the point cloud frame and the representation in the map data. Based on this relative viewing angle, the point cloud object may be rotated and/or scaled such that the point cloud object and the map data are analyzed from the same perspective.
- the relative velocity of the point cloud object identified at block 706 is determined by analyzing the skew of the object.
- a machine learning model e.g., trained using reinforcement learning
- the machine learning model may operate on (i.e., use as inputs) the point cloud object identified at block 706 and the expected configuration.
- the point cloud object identified at block 706 is skewed along multiple axes. Accordingly, the skew with respect to each axis may be analyzed to determine the components of the velocity vector with respect to each of these axes. For example, a lateral skew is analyzed to determine a lateral component of the relative velocity and a longitudinal skew is analyzed to determine a longitudinal component of the relative velocity. Accordingly, the machine learning model may analyze lateral bounds of the point cloud object and the expected configuration to detect the lateral skew and longitudinal bounds of the point cloud object and the expected configuration to detect the longitudinal skew. In some implementations, the tracking module 414 , for example, associates the determined relative velocity with the point cloud object.
- the method 700 includes one or more additional blocks not shown in FIG. 8 .
- the sensor data received at block 702 includes velocity data indicative of the vehicle velocity (such as velocity data included in component reporting messages transmitted over a vehicle bus).
- the method 700 may include an additional block where the vehicle velocity is subtracted from the relative velocity determined at block 710 to determine the velocity of the point cloud object.
- the method 700 may include correcting the skew based on the relative velocity determined at block 710 .
- the point cloud object may be adjusted to align with the expected configuration.
- the point cloud object (either corrected or uncorrected) and the relative velocity may be analyzed, for example by the prediction component 420 of FIG.
- This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state.
- control signals e.g., operational parameters
- the block may include generate a control signal to apply the vehicle brakes.
- FIG. 9 depicts an example method 720 for determining relative velocity of a point cloud object based on depth data associated with co-located pixels.
- the lidar system (such as the lidar system 100 of FIGS. 1 and 2 ) is configured to follow a scan pattern that includes co-located pixels, for example, the co-located pixels 166 include in scan pattern 160 of FIG. 1B .
- the sensor heads 312 A and 312 D of the lidar system 302 of FIG. 3A are configured to implement complimentary, overlapping scan patterns wherein at least one pixel from the scan pattern implemented by the sensor head 312 A is co-located with at least one pixel from the scan pattern implemented by the sensor head 312 D.
- the scan patterns are configured such that the depth data associated with the co-located pixels are not obtained at the same time.
- the method 720 begins with receiving sensor data at block 722 .
- the sensor data may include lidar depth data (e.g., sensed by the lidar system 100 of FIGS. 1 and 2 , or the lidar system 302 of FIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with the sensor data 402 of FIG. 5 , for example.
- Block 722 may be the same or similar to block 702 of the method 700 .
- a point cloud frame based on the sensor data received at block 722 is obtained.
- the sensor data received at block 722 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud).
- block 724 may include extracting the point cloud frame from the rest of the sensor data received at block 722 .
- block 724 includes analyzing depth or lidar data included in the sensor data received at block 722 to generate the point cloud frame, as discussed in connection to FIG. 3 , for example.
- Block 724 may be the same or similar to block 704 of the method 700 .
- block 726 co-located pixels within the field of regard that overlap a point cloud object within the point cloud frame obtained at block 724 are identified.
- block 726 includes partitioning the point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and tracking movement of classified objects over time, for example, by utilizing a perception component 406 of the software architecture 400 of FIG. 5 .
- a tracking module 414 of the perception component 406 tracks objects across multiple point cloud frames.
- the tracking module 414 identifies objects within the point cloud frame obtained at block 724 that are currently being tracked.
- block 726 includes determining that a pair of co-located pixels overlaps one of the point cloud objects.
- depth data associated with the co-located pixels are analyzed to determine a depth difference ( ⁇ d).
- ⁇ t is indicative of the time span between capturing/sensing the depth data associated with the two co-located pixels
- ⁇ d is indicative of the distance the object moved, relative to the vehicle and either towards or away from the vehicle, during the time span ⁇ t.
- relative velocity of the object may be determined by based on ⁇ d and ⁇ t. More particularly, in some implementations, the relative velocity is determined by dividing ⁇ d by ⁇ t.
- the lidar system 100 includes a clock that time stamps the depth data when it is sensed at the receiver 140 .
- blocks 728 and 730 may implement Equation 4 using the depth and time data associated with the co-located pixels as inputs. Further, if ⁇ d is determined at block 728 by subtracting the depth data of the later-sensed co-located pixel from the depth data of the earlier-sensed co-located pixel, the polarity of ⁇ d is indicative of the direction of travel of the point cloud object. In particular, when ⁇ d is a positive number, the point cloud object is moving away from the vehicle and, conversely, when ⁇ d is a negative number, the point cloud object is moving towards the vehicle.
- the method 720 includes one or more additional blocks not shown in FIG. 9 .
- the sensor data received at block 722 includes velocity data indicative of the vehicle velocity (such as velocity data included in component reporting messages transmitted over a vehicle bus).
- the method 720 may include an additional block where the vehicle velocity is subtracted from the relative velocity determined at block 730 to determine the velocity of the point cloud object.
- an additional block may include identifying a second set of co-located pixels that overlap the point cloud object at a different region of the point cloud object. This block may include determining a relative velocity of the point cloud object based on the second set of co-located pixels. The two relative velocities may then be compared to one other to detect whether the point cloud object is turning or rotating.
- Another additional block may include correcting rolling shutter distortion based on the relative velocity determined at block 730 .
- the point cloud object may be adjusted to compensate for the motion of the point cloud object during the collection of depth data that comprises the point cloud frame.
- a machine learning module e.g., trained using reinforcement learning
- the machine learning model may operate on (i.e., use as inputs) the relative velocity determined at block 730 and the depth data of the co-located pixels to determine a correction factor to compensate for delay introduced by successive pixels and/or scan lines included in a scan pattern.
- Another additional block of the method 600 may include determining an object velocity (i.e. the object's absolute velocity, as opposed to its relative velocity). For example, a relative velocity of a point cloud object that is known to be stationary (e.g., the point cloud object is classified as a light pole, a street sign, a tree, a bench, a power line, a building, a median, etc.) may be determined using the techniques of blocks 606 and 608 . Because the point cloud object is known to be stationary, the relative velocity is comprised solely of the vehicle's velocity. Said another way, by determining the relative velocity with respect to a known, stationary object, the vehicle velocity can be determined. Subsequently, the vehicle velocity can be subtracted from the relative velocity determined with respect to the identified point cloud object to determine object velocity.
- an object velocity i.e. the object's absolute velocity, as opposed to its relative velocity.
- a relative velocity of a point cloud object that is known to be stationary e.g., the point cloud object is classified as
- the point cloud object (either corrected or uncorrected) and the relative velocity may be analyzed, for example by the prediction component 420 of FIG. 5 , to predict a future state of the vehicle environment.
- This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state.
- control signals e.g., operational parameters
- the block may include generate a control signal to apply the vehicle brakes.
- FIG. 10 depicts an example method 740 for correcting point cloud distortion based on tracked object velocity.
- the methods 700 and 720 determine the relative velocity of the point cloud object based on a single point cloud frame
- the method 740 determines the relative velocity of the point cloud object across multiple point cloud frames. Accordingly, while this increases the time it takes to determine a relative velocity of a newly detected point cloud object, already-tracked objects may be associated with historic position data. Thus, by comparing a point cloud object as represented in a new point cloud frame to the historic position data, the relative velocity of an already-tracked object can be determined based on a single successive point cloud frame.
- the method 740 begins when sensor data is received at block 742 .
- the sensor data may include lidar depth data (e.g., sensed by the lidar system 100 of FIGS. 1 and 2 , or the lidar system 302 of FIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with the sensor data 402 of FIG. 5 , for example.
- Block 742 may be the same or similar to performing block 702 of the method 700 , but with respect to subsequent sets of sensor data.
- two or more point cloud frames based on the sensor data received at block 742 are obtained.
- the sensor data received at block 742 includes the point cloud frames (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud).
- block 744 may include extracting the point cloud frames from rest of the sensor data received at block 742 .
- block 744 includes analyzing multiple sets of depth or lidar data included in the sensor data received at block 742 to generate the point cloud frames, as discussed in connection with FIG. 3 , for example.
- Block 744 may be the same or similar to performing block 704 of the method 700 , but with respect to subsequent sets of sensor data.
- a point cloud object is tracked across the two or more point cloud frames.
- block 746 includes partitioning a first point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and tracking movement of the classified objects over time, for example, by utilizing a perception component 406 of the vehicle controller 400 of FIG. 5 .
- a tracking module 414 of the perception component 406 tracks objects across multiple point cloud frames. More particularly, the tracking component 414 may associate tracked objects with position data (such as a position within the field of regard, a position in the vehicle environment, and/or a position relative to the vehicle), shape data, orientation data, or other information.
- the position data may be indicative of one or more of a position of the centroid of the point cloud object, a particular feature of the point cloud object, or a particular bound of the point cloud object.
- the point cloud object identified in the first point cloud object can be detected in the second point cloud frame (i.e., tracked across the first and second point cloud frames) using the associated shape or orientation information.
- a relative velocity of the point cloud object tracked at block 746 is determined.
- the position data of the point cloud object as represented in the first point cloud frame is subtracted from the position data of the point cloud object as represented in the second point cloud frame. This difference is divided by the time difference between the first and second point cloud frames to determine the relative velocity of the tracked object.
- rolling shutter distortion associated with the point cloud object tracked at block 746 is corrected based on the relative velocity determined at block 748 .
- a machine learning module e.g., trained using reinforcement learning
- the machine learning model may operate on (i.e., use as inputs) the relative velocity determined at block 748 and position information associated with the point cloud object to determine a correction factor to compensate for delay introduced by successive pixels and/or scan lines included in a scan pattern. Accordingly, based on the scan pattern utilized to capture the point cloud data, the correction factor is applied to the tracked point cloud object to correct for the rolling shutter distortion.
- a relative velocity of a point cloud object that is known to be stationary is determined using the techniques of blocks 746 and 748 and/or the techniques of blocks 606 and 608 of the method 600 .
- the relative velocity is comprised solely of the vehicle's velocity. Said another way, by determining the relative velocity with respect to a known, stationary object, the vehicle velocity can be determined. Subsequently, the vehicle velocity can be subtracted from the relative velocity determined with respect to the identified point cloud object to determine object velocity.
- the method 740 includes one or more additional blocks not shown in FIG. 10 .
- the sensor data received at block 742 includes velocity data indicative of the vehicle velocity (such as velocity data included in component reporting messages transmitted over a vehicle bus).
- the method 740 may include an additional block where the vehicle velocity is subtracted from the relative velocity determined at block 748 to determine the velocity of the point cloud object.
- the corrected point cloud object and the relative velocity may be analyzed, for example by the prediction component 420 of FIG. 5 , to predict a future state of the vehicle environment.
- This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state.
- FIGS. 11 and 12 depict example methods 800 and 820 , respectively, for detecting that a point cloud object, for example, one of the objects 396 of the point cloud 390 depicted in FIG. 4 , is distorted and not actually shaped in a manner that appears distorted.
- the method 800 of FIG. 11 relates to embodiments that detect distortion by analyzing known shapes, and the method 820 of FIG. 12 detects distortion by analyzing sensor data other than just lidar or depth data.
- the methods 800 and 820 may be implemented/performed by the computing system 500 of FIG.
- sensor data is received at block 802 .
- the sensor data may include lidar depth data (e.g., sensed by the lidar system 100 of FIGS. 1 and 2 , or the lidar system 302 of FIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with the sensor data 402 of FIG. 5 , for example.
- a point cloud frame based on the sensor data received at block 802 is obtained.
- the sensor data received at block 802 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud).
- block 804 may include extracting the point cloud frame from rest of the sensor data received at block 802 .
- block 804 includes analyzing depth or lidar data included in the sensor data received at block 802 to generate the point cloud frame, as discussed in connection to FIG. 3 , for example.
- a point cloud object within the point cloud frame obtained at block 804 is identified.
- block 806 includes partitioning the point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and tracking movement of classified objects over time, for example, by utilizing a perception component 406 of the vehicle controller 400 of FIG. 5 .
- a tracking module 414 of the perception component 406 tracks objects across multiple point cloud frames.
- the tracking module 414 identifies objects within the point cloud frame obtained at block 704 that are currently being tracked.
- the segmentation module 410 may identifies a new object that is not being tracked by the tracking module 416 .
- the point cloud object identified at block 806 is analyzed to identify a feature that has an expected shape.
- the classification module 412 of the perception component 406 classifies the identified point cloud object as a particular object type.
- the object type may be associated with a model that includes indications of particular expected features for point cloud objects classified as the particular object type.
- a car object type has wheels that are expected to be round, a license plate that is expected to be rectangular, and perhaps a manufacturer logo that has a known shape.
- a streetlight object may be expected to have round light bulbs.
- a stop sign object is expected to be octagonal.
- block 808 includes analyzing the point cloud object identified at block 806 to detect the region in the point cloud object associated with the feature that has the expected shape.
- the expected shape of the feature is compared to the shape of the feature of the point cloud object as represented in the point cloud frame obtained at block 804 .
- the expected shape of the feature may be scaled and/or rotated based on depth data associated with the point cloud object. Said another way, the expected shape of the feature is adjusted to account for how the feature is oriented in the point cloud frame. Accordingly, block 810 may include comparing the adjusted expected feature shape to the region on the point cloud object associated with the feature. Alternatively, the feature of the point cloud object may be scaled and/or rotated to match, as closely as possible, a fixed size and orientation corresponding to the expected feature shape.
- the point cloud object identified at block 806 is determined to be distorted based on the expected feature shape not matching the feature of the point cloud object corresponding to the feature.
- the feature of the point cloud object may be skewed with respect to the expected feature shape due to rolling shutter distortion.
- block 812 in addition to determining that point cloud object is distorted, block 812 includes determining a skew factor indicative of an amount of skew introduced by each scan line and/or component pixel thereof.
- the skew factor can be a three-dimensional vector representative of the direction via which the skew is introduced.
- block 812 may include determining an amount by which the point cloud feature is skewed with respect to the expected feature shape at two different pixels that overlap the point cloud feature. The difference in skews with respect to the expected configuration is then divided by a number of pixels and/or scan lines between when the pixels were captured to determine a skew factor indicative of an amount and direction of skew introduced by each pixel and/or scan line. Accordingly, in some implementations, block 812 includes correcting for the detected skew by utilizing the skew factor to compensate for the amount of skew introduced by each successive scan line and/or pixel.
- the method 800 includes one or more additional blocks not shown in FIG. 11 .
- the distortion of the point cloud object and/or the skew factor determined at block 812 is analyzed to determine the relative velocity of the point cloud object, for example, using the method 700 of FIG. 8 .
- the point cloud object for example by the prediction component 420 of FIG. 5 , to predict a future state of the vehicle environment.
- This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state.
- the sensor data includes a first set of sensor data captured by a first set of one or more sensors configured to sense the vehicle by sequentially advancing through a plurality of points in a scan pattern during a frame, such as lidar depth data sensed by the lidar system 100 of FIGS. 1 and 2 , or the lidar system 302 of FIG. 3A .
- the sensor data also includes a second set of sensor data captured by a second set of one or more sensors that are configured to sense the vehicle environment by capturing the entirety of a frame at a single time, such as camera data captured by a RGB or stereo camera or a thermal imaging sensor.
- the first set of sensor data is analyzed to obtain a point cloud representation of an object within the vehicle environment.
- the object may be a vehicle, a road, a street sign, a person, a tree, etc.
- each of the first and second sets of sensor may include a different digital representation of the same object.
- the first set of sensor data received at block 822 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud).
- block 824 includes extracting the point cloud frame from rest of the first set of sensor data received at block 822 .
- block 824 includes analyzing the first set of sensor data received at block 802 to generate the point cloud frame, as discussed in connection to FIG. 3 , for example.
- a perception component such as the perception component 406 of FIG. 5 , analyzes the point cloud frame to identify the object.
- a tracking module 414 of the perception component 414 may compare tracked objects to the obtained point cloud frame. Accordingly, the tracking component 414 may identify the point cloud representation of the object in the point cloud frame.
- the second set of sensor data is analyzed to obtain a second digital representation of the object.
- the second representation of the object may be a subset of the RGB data that corresponds to the object.
- the perception component 406 may associate the first and second sets of sensor data with one another. That is, the perception component 406 may analyze the viewing angles at which the first and second sets of sensors sensed the first and second sets of sensor data to identify regions in each set of sensor data representative of the same portion of the vehicle environment, for example, the object analyzed at block 824 .
- the tracking module 414 associates the point cloud object analyzed at block 826 with RGB data indicative of the object.
- the tracking module 414 may utilizes the associated RGB data to identify the object within the second set of sensor data.
- the perception component 406 includes separate segmentation, classification, and/or tracking modules optimized to analyze RGB data. In these implementations, the point cloud representation of the object and the RGB data representation of the object are tracked separately.
- the point cloud object and the representation of the object in the second set of sensor data are compared to each other. More particularly, the depth data associated with the first set of sensor data may be “flattened” to obtain a two-dimensional representation of the point cloud object. The shape of the flattened point cloud object may be compared to the shape of the object as represented in the second set of sensor data. In some implementations, the shape of the object as represented in the second set of sensor data is adjusted to account for the different viewing angles associated with the first and second set of sensors.
- the point cloud representation of the object is identified as distorted based on the comparison at block 828 .
- the second set of sensor data indicates the “true” shape of the object.
- the comparison at block 828 indicates that the shape of the point cloud representation of the object and the shape of the object as represented by the second set of sensor data do not match, it is determined that the point cloud object (i.e., the representation of the object, as opposed to the object itself) is distorted.
- the method 820 includes one or more additional blocks not shown in FIG. 12 .
- one additional block may include determining a skew factor indicative of an amount of skew introduced by each scan line and/or component pixel thereof.
- the skew factor can be a three-dimensional vector representative of the direction via which the skew is introduced.
- this additional block may include determining an amount of the point cloud feature is skewed with respect to an expected configuration as indicated by the second set of sensor data at two different pixels that overlap the point cloud feature based on the comparison at block 830 .
- the difference in skew with respect to the second set of sensor data is then divided by a number of pixels and/or scan lines between when the pixels were captured to determine the skew factor.
- this additional block includes correcting for the detected skew by utilizing the skew factor to compensate for the amount of skew introduced by each successive scan line and/or pixel
- the distortion of the point cloud object and/or the skew factor determined at the above additional block is analyzed to determine the relative velocity of the point cloud object, for example, using the method 700 of FIG. 8 .
- the point cloud object for example by the prediction component 420 of FIG. 5 , to predict a future state of the vehicle environment.
- This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state.
- a lidar system may be affixed to a stationary object (e.g., a light pole, a utility pole, a street sign, a traffic light, or a building). Any rolling shutter distortion of an object may be attributed to movement of that object since the stationary lidar system would not contribute to the rolling shutter distortion.
- a stationary lidar system may be configured to determine the absolute velocity of objects rather than a relative velocity.
- a computing device may be used to implement various modules, circuits, systems, methods, or algorithm steps disclosed herein.
- all or part of a module, circuit, system, method, or algorithm disclosed herein may be implemented or performed by a general-purpose single- or multi-chip processor, a digital signal processor (DSP), an ASIC, a FPGA, any other suitable programmable-logic device, discrete gate or transistor logic, discrete hardware components, or any suitable combination thereof.
- DSP digital signal processor
- a general-purpose processor may be a microprocessor, or, any conventional processor, controller, microcontroller, or state machine.
- a processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- one or more implementations of the subject matter described herein may be implemented as one or more computer programs (e.g., one or more modules of computer-program instructions encoded or stored on a computer-readable non-transitory storage medium).
- the steps of a method or algorithm disclosed herein may be implemented in a processor-executable software module which may reside on a computer-readable non-transitory storage medium.
- a computer-readable non-transitory storage medium may include any suitable storage medium that may be used to store or transfer computer software and that may be accessed by a computer system.
- a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs (e.g., compact discs (CDs), CD-ROM, digital versatile discs (DVDs), blue-ray discs, or laser discs), optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, flash memories, solid-state drives (SSDs), RAM, RAM-drives, ROM, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate.
- ICs such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (A
- FIG. 1 While operations may be depicted in the drawings as occurring in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all operations be performed.
- the drawings may schematically depict one more example processes or methods in the form of a flow diagram or a sequence diagram. However, other operations that are not depicted may be incorporated in the example processes or methods that are schematically illustrated. For example, one or more additional operations may be performed before, after, simultaneously with, or between any of the illustrated operations. Moreover, one or more operations depicted in a diagram may be repeated, where appropriate. Additionally, operations depicted in a diagram may be performed in any suitable order.
- words of approximation such as, without limitation, “approximately, “substantially,” or “about” refer to a condition that when so modified is understood to not necessarily be absolute or perfect but would be considered close enough to those of ordinary skill in the art to warrant designating the condition as being present.
- the extent to which the description may vary will depend on how great a change can be instituted and still have one of ordinary skill in the art recognize the modified feature as having the required characteristics or capabilities of the unmodified feature.
- a numerical value herein that is modified by a word of approximation such as “approximately” may vary from the stated value by ⁇ 0.5%, ⁇ 1%, ⁇ 2%, ⁇ 3%, ⁇ 4%, ⁇ 5%, ⁇ 10%, ⁇ 12%, or ⁇ 15%.
- first,” “second,” “third,” etc. may be used as labels for nouns that they precede, and these terms may not necessarily imply a particular ordering (e.g., a particular spatial, temporal, or logical ordering).
- a system may be described as determining a “first result” and a “second result,” and the terms “first” and “second” may not necessarily imply that the first result is determined before the second result.
- the terms “based on” and “based at least in part on” may be used to describe or present one or more factors that affect a determination, and these terms may not exclude additional factors that may affect a determination.
- a determination may be based solely on those factors which are presented or may be based at least in part on those factors.
- the phrase “determine A based on B” indicates that B is a factor that affects the determination of A. In some instances, other factors may also contribute to the determination of A. In other instances, A may be determined based solely on B.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Electromagnetism (AREA)
- Computer Networks & Wireless Communication (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Automation & Control Theory (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Optics & Photonics (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Aviation & Aerospace Engineering (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Optical Radar Systems And Details Thereof (AREA)
- Traffic Control Systems (AREA)
- Geometry (AREA)
Abstract
Description
- This claims the benefit of U.S. Provisional Patent Application No. 62/715,030, filed on Aug. 6, 2018 and entitled “Utilizing and Correcting Object Distortion in Point Cloud Frames,” the disclosure of which is hereby incorporated herein by reference in its entirety.
- This disclosure generally relates to lidar systems, and, more particularly, to detecting, analyzing, correcting, and/or utilizing distortions in a point cloud produced by lidar systems.
- The background description provided herein is for the purpose of generally presenting the context of the disclosure. Work of the presently named inventors, to the extent it is described in this background section, as well as aspects of the description that may not otherwise qualify as prior art at the time of filing, are neither expressly nor impliedly admitted as prior art against the present disclosure.
- Self-driving or “autonomous” vehicles generally employ sensors, such as light detection and ranging (lidar) devices, to detect or “see” the surrounding environment as the vehicles move toward their destinations. A single lidar laser cannot scan the entire field of regard at once. Instead, lidar lasers may be configured to follow a scan pattern to scan the field of regard. As the lidar laser follows the scan pattern, a scanner determines depth data by measuring the travel time for light beams that are transmitted by the lidar system and scattered or reflected by objects in the field of regard. Each of these points along the scan pattern are referred to as “pixels” and are correlated to the determined depth data. While the lidar system is capable of scanning the field of regard at high speeds, there is still a time a delay between when the scanner determines the depth data at the various pixels of the scan pattern. Thus, if an object in the field of regard is moving relative to the vehicle, the object may appear to be distorted.
- One example embodiment of the techniques of this disclosure is a computer-implemented method of determining a relative velocity between a vehicle and an object. The method includes receiving sensor data generated by one or more sensors of the vehicle. The one or more sensors are configured to sense an environment of the vehicle by following a scan pattern comprising component scan lines. The method also includes obtaining, based on the sensor data and by one or more processors, a point cloud frame representative of the environment. The point cloud frame comprises a plurality of points such that each point includes depth data associated with the point and a time at which the depth data was captured. Additionally, the method includes selecting, by the one or more processors, two or more points of the scan pattern that overlap the object. The selected points are located on or near a two-dimensional surface corresponding to at least part of the object, and the depth data for two or more of the selected points are captured at different times. Further, the method includes calculating, by the one or more processors, the relative velocity between the vehicle and the object based at least in part on the depth data and capture times associated with the selected points.
- Another example embodiment of the techniques of this disclosure is a system within an autonomous vehicle. The system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines. The system also includes a computing system configured to receive the set of sensor data. The computing system is also configured to obtain, based on the sensor data, a point cloud frame representative of the environment. The point cloud frame includes a plurality of points such that each point includes depth data associated with the point and a time at which the depth data was captured by the set of sensors. The computing system is also configured to select two or more points of the scan pattern that overlap an object in the environment. The selected points are located on or near a two-dimensional surface corresponding to at least part of the object, and the depth data for two or more of the selected points are captured at different times. Additionally the computing system is configured to calculate a relative velocity between the autonomous vehicle and the object based at least in part on the depth data and capture times associated with the selected points.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of determining relative velocity between a vehicle and an object. The method includes receiving sensor data generated by one or more sensors of the vehicle. The one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern comprising component scan lines. The method includes obtaining, by one or more processors, a point cloud frame based on the sensor data and representative of the environment and identifying, by the one or more processors, a point cloud object within the point cloud frame. The method further includes determining, by the one or more processors, that the point cloud object is skewed relative to an expected configuration of the point cloud object, and determining, by the one or more processors, a relative velocity of the point cloud object by analyzing the skew of the object.
- Another example embodiment of the techniques of this disclosure is a system within an autonomous vehicle. The system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines. The system also includes a computing system configured to receive the set of sensor data and obtain a point cloud frame based on the sensor data and representative of the environment. The computing system is also configured to identify a point cloud object within the point cloud frame and determine that the point cloud object is skewed relative to an expected configuration of the point cloud object. Additionally, the computing system is configured to determine a relative velocity of the point cloud object by analyzing the skew of the object.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of determining relative velocity between a vehicle and an object. The method includes receiving sensor data generated by one or more sensors of the vehicle. The one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern comprising component scan lines. The method also includes obtaining, based on the sensor data and by one or more processors, a point cloud frame representative of the environment. Additionally the method includes identifying, by the one or more processors, a first pixel and a second pixel that are co-located within a field of regard of the one or more sensors and overlap a point cloud object within the point cloud frame and calculating, by the one or more processors, a difference between a depth associated with the first pixel and a depth associated with the second pixel. The method also includes determining, by the one or more processors, a relative velocity of the point cloud object by dividing the difference in depth data by a time difference between when the depth associated with the first pixel was sensed and the depth associated with the second pixel was sensed.
- Another example embodiment of the techniques of this disclosure is a system within an autonomous vehicle. The system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines. The system also includes a computing system configured to receive the set of sensor data and obtain, based on the set of sensor data, a point cloud frame representative of the environment. The computing system is also configured to identify a first pixel and a second pixel that are co-located within a field of regard of the one or more sensors and overlap a point cloud object within the point cloud frame and calculate a difference between a depth associated with the first pixel and a depth associated with the second pixel. Additionally, the computing system is configured to determine a relative velocity of the point cloud object by dividing the difference in depth data by a time difference between when the depth associated with the first pixel was sensed and the depth associated with the second pixel was sensed.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of determining relative velocity between a vehicle and an object. The method includes receiving sensor data generated by one or more sensors of the vehicle. The one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern comprising component scan lines. The method also includes obtaining, based on the sensor data and by one or more processors, two or more point cloud frames representative of the environment and tracking, by the one or more processors, a point cloud object across the two or more point cloud frames. Additionally, the method includes determining, based on the tracking and by the one or more processors, a relative velocity of the point cloud object and correcting, by the one or more processors, the point cloud object based on the relative velocity of the point cloud object.
- Another example embodiment of the techniques of this disclosure is a system within an autonomous vehicle. The system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines. The system also includes a computing system configured to receive the set of sensor data and obtain, based on the set of sensor data, two or more point cloud frames representative of the environment. The computing system is also configured to track a point cloud object across the two or more point cloud frames. Additionally, the computing system is configured to determine, based on the tracking, a relative velocity of the point cloud object and correct the point cloud object based on the relative velocity of the point cloud object.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of detecting object distortion. The method includes receiving sensor data generated by one or more sensors of the vehicle. The one or more sensors are configured to sense an environment through which the vehicle is moving by following a scan pattern. The method also includes obtaining, based on the sensor data and by one or more processors, a point cloud frame representative of the environment and identifying, by one or more processors, a point cloud object within the point cloud frame. Additionally, the method includes analyzing, by the one or more processors, the point cloud object to identify a feature of the point cloud object that has an expected shape and comparing, by the one or more processors, the feature of the point cloud object to the expected shape. The method also includes identifying, by the one or more processors, that the point cloud object is distorted based on the feature of the point cloud object not matching the expected shape.
- Another example embodiment of the techniques of this disclosure is a system within an autonomous vehicle. The system includes a set of sensors configured to generate a set of sensor data by sensing an environment of the vehicle by following a scan pattern comprising component scan lines. The system also includes a computing system configured to receive the set of sensor data and obtain, based on the sensor data, a point cloud frame representative of the environment. The computing system is also configured to identify a point cloud object within the point cloud frame and analyze the point cloud object to identify a feature of the point cloud object that has an expected shape. Additionally, the computing system is configured to compare the feature of the point cloud object to the expected shape and identify that the point cloud object is distorted based on the feature of the point cloud object not matching the expected shape.
- Another example embodiment of the techniques of this disclosure is a computer-implemented method of detecting object distortion. The method includes receiving sensor data generated by a plurality of sensors of the vehicle. The plurality of sensors includes a first set of one or more sensors, wherein each sensor of the first set of sensors is configured to sense an environment through which the vehicle is moving by, for each frame produced by the sensor, sequentially advancing through a plurality of points in a scan pattern. The plurality of sensors also include a second set of one or more sensors, wherein each sensor of the second set of sensors is configured to sense the environment by, for each frame produced by the sensor, capturing an entirety of the frame at a single time. The method also includes identifying, by one or more processors, an object within the environment. Additionally, the method includes analyzing, by the one or more processors, the data sensed by the first set of sensors to identify a point cloud representation of the object and analyzing, by the one or more processors, the data sensed by the second set of sensors to identify the object within the data sensed by the second set of sensors. The method also includes comparing, by the one or more processors, a shape of the point cloud representation of the object to a shape of the object as sensed by the second set of sensors and identifying, by the one or more processors, that the object is distorted based on the shape of the point cloud representation of the object component not matching the shape of the object as sensed by the second set of sensors.
- Another example embodiment of the techniques of this disclosure is a system within an autonomous vehicle. The system includes a first set of one or more sensors configured to sense an environment through which the vehicle is moving by, for each frame produced by the sensor, sequentially advancing through a plurality of points in a scan pattern. The system also includes a second set of one or more sensors configured to sense the environment by, for each frame produced by the sensor, capturing an entirety of the frame at a single time. Additionally, the system includes a computing system configured to receive data generated by the first and second sets of one or more sensors and identify an object within the environment. The computing system is also configured to analyze the data sensed by the first set of sensors to identify a point cloud representation of the object and analyze the data sensed by the second set of sensors to identify the object within the data sensed by the second set of sensors. Additionally, the computing system is configured to compare a shape of the point cloud representation of the object to a shape of the object as sensed by the second set of sensors and identify that the object is distorted based on the shape of the point cloud representation of the object component not matching the shape of the object as sensed by the second set of sensors.
-
FIG. 1A illustrates an example, conventional “zigzag” scan pattern which a light detection and ranging (lidar) system can produce when identifying targets within a field of regard; -
FIG. 1B illustrates an example “looping” scan pattern which a lidar system can produce when identifying targets within a field of regard; -
FIG. 2 is a block diagram of an example lidar system which can implement the scan patterns ofFIGS. 1A and/or 1B , as well as other techniques of this disclosure; -
FIG. 3A illustrates an example vehicle in which the lidar system ofFIG. 2 may operate; -
FIG. 3B illustrates another example vehicle in which the lidar system ofFIG. 2 may operate; -
FIG. 4 illustrates an example point cloud that may be generated by the lidar system ofFIG. 2 or 3 ; -
FIG. 5 is a block diagram of a software architecture that may be used in the lidar system ofFIG. 2 or 3 ; -
FIG. 6 is a block diagram of an example computing system for controlling an autonomous vehicle, which may be used in the lidar system ofFIG. 2 or 3 ; -
FIG. 7 is a flow diagram of an example method for determining relative velocity of a point cloud object by fitting pixels onto a surface of the point cloud object; -
FIG. 8 is a flow diagram of an example method for determining relative velocity of a point cloud object by analyzing a skew of the point cloud object; -
FIG. 9 is a flow diagram of an example method for determining relative velocity of a point cloud object using co-located pixels; -
FIG. 10 is a flow diagram of an example method for correcting point cloud distortion based on tracked object velocity; -
FIG. 11 is a flow diagram of an example method for detecting that a point cloud object is distorted using features having a known shape; and -
FIG. 12 is a flow diagram of an example method for method for detecting that a point cloud object is distorted by comparing to other sensor data. - Generally speaking, when a lidar system produces a frame of depth data (e.g., a point cloud frame), the frame includes depth data captured at different times within the frame period. For instance, when the lidar system captures depth data indicative of a surface on which all points are equidistant from the lidar system, and if the surface is moving with respect to the lidar system, the surface will be at different positions for different pixels within the same point cloud frame. This may cause the surface to appear skewed within the point cloud representation. For example, if the surface is moving closer to or more distant from the lidar system, the surface may appear to be “tilted” or “skewed” toward or away from the lidar system. As another example, if the scan lines in the scan pattern are generally horizontal (parallel to the ground), a surface moving laterally relative to the lidar system may appear to be tilted or skewed in a lateral direction, and a surface moving up or down relative to the lidar system may appear to be longer or shorter. On the other hand, if the scan lines of the scan pattern are generally vertical (perpendicular to the ground), surfaces moving laterally relative to the lidar system may appear to be wider or narrower, and surfaces moving up or down relative to the lidar system may appear to be tilted or skewed. The term “rolling shutter distortion” is generally used herein to refer to this phenomenon.
- In some aspects, simply analyzing the overall shape of an object or surface may be inadequate to determine if an object is experiencing rolling shutter distortion; in some scenarios the object or surface is actually shaped in a manner that may appear distorted. Accordingly, the lidar system is configured to implement various techniques to distinguish between objects that are experiencing rolling shutter distortion (e.g., a telephone pole that only appears to be tilted due to relative velocity) and objects that are actually shaped in a manner similar to a distorted object (e.g., a telephone pole that is actually tilted).
- In one disclosed technique corresponding to one embodiment, the lidar system fits two or more points onto a two-dimensional surface to model the motion (including velocity) of the point cloud object. The lidar system may then use the depth data associated with the two or more points to model the motion of the surface to approximate the relative velocity of the point cloud object. In some aspects, the lidar system fits an additional two or more points associated with a different scan line to the two-dimensional surface. In these aspects, the lidar systems determines and compares a relative velocity of the surface using each set of two or more points. By comparing the polarity of the relative velocities of the object based on depth data on subsequent scan lines, the lidar system can determine that the object is experiencing rolling shutter distortion and, if the object is distorted, the direction of the distortion (i.e., which direction the object is moving relative to the vehicle). Moreover, by determining a rate of change between the relative velocities, the lidar system can also determine the relative acceleration of the object.
- In another disclosed technique that is implemented in one embodiment, the lidar system analyzes co-located pixels (i.e., pixels at the same location in the field of regard but captured at different points of time within a point cloud frame) to detect rolling shutter distortion. In some implementations, the depth data for the co-located pixels are captured by two different lidar scanners. In other implementations, the depth data for the co-located pixels are captured by the same lidar scanner (i.e., the scan pattern crosses over the same location in the field of regard at multiple points in time during the point cloud frame). By comparing the depth data for the co-located pixels, the lidar system can determine whether the object is moving relative to the vehicle, and thus experiencing rolling shutter distortion.
- In another disclosed technique corresponding to one embodiment, the lidar system compares the shape of a point cloud to reference image data captured by a RGB or stereo camera to detect rolling shutter distortion. The RGB or stereo camera doesn't include delays introduced by a scan pattern, and therefore indicates the “true” shape of an object. Accordingly, the lidar system can compare the point cloud object to corresponding RGB data to determine whether the shapes of the two objects match. When the shapes match, the point cloud object is not experiencing rolling shutter distortion (i.e., the object is actually shaped as detected by the lidar system).
- The rolling shutter distortion introduced by a lidar system scan pattern generally impacts objects that are closer to the vehicle more than objects that are further away. In particular, closer objects will typically be associated with more pixels than further away objects, and thus it takes more time to capture the depth data for all pixels corresponding to the nearby object. This allows the object and/or the vehicle to move further within the relevant portion of the scan pattern, which in turn causes the resulting depth data to distort the object more heavily, causing it to appear as if it is curved or skewed.
- Accordingly, in still another disclosed technique corresponding to one embodiment, the lidar system includes a database of objects (or features thereof) that have known, fixed shapes. For instance, license plates are rectangular, wheels are round, manufacturer logos have known, fixed shapes, etc. Thus, if the point cloud object in question includes a feature associated with one of these known, fixed shapes, the lidar system can determine whether the feature is curved (and therefore not matching the known, expected shape) to determine that the object is distorted.
- Additionally, disclosed techniques leverage the fact that the amount and nature of distortion detected is indicative of the relative velocity and direction between the object and the vehicle. In some disclosed embodiments, the lidar system detects and analyzes the amount of the distortion to determine vehicle and/or object velocity. Thus, utilizing the disclosed techniques, the lidar system can determine the vehicle and/or object velocity based on a single point cloud frame. This enables the lidar system to determine vehicle and/or object velocity faster than traditional techniques that track objects between multiple point cloud frames.
- That said, the lidar system may still track objects across multiple point cloud frames. For example, in one embodiment, the lidar system determines relative velocity of an object using multiple point cloud frame tracking techniques. In some scenarios, a correlation between relative velocity and the magnitude of distortion is determined to develop a distortion factor indicative of the distortion introduced in each subsequent scan line and/or pixel. The lidar system then applies this distortion factor to other point cloud objects to determine the relative velocity of the other point cloud objects.
- In some embodiments, after detecting that a point cloud object is distorted, the lidar system corrects the shape of the point cloud object before the point cloud frame and/or object data is analyzed by a classification module, a tracking module, or a prediction component of the lidar system. For example, the distortion factor can be applied to remove the distortion introduced by the time difference between successive scan lines or pixels. Accordingly, the lidar system may apply an inverse operation to shift the point cloud data to remove rolling shutter distortion.
- An example lidar system in which the above techniques can be implemented is considered first with reference to
FIGS. 1-4 . An example implementation of a vehicle controller that interacts with the example lidar systems is then discussed with reference toFIGS. 5-6 . Finally, example methods corresponding to distortion detection, analysis, and correction are discussed with reference to the flow diagrams ofFIGS. 7-12 . -
FIG. 2 illustrates an example light detection and ranging (lidar)system 100. While various lidar system components and characteristics are described herein, it is understood that any suitable lidar device(s) or system(s), and/or any other suitable types of sensors, may provide sensor data for processing using the software architectures described herein. - The
example lidar system 100 may include alight source 110, amirror 115, ascanner 120, areceiver 140, and acontroller 150. Thelight source 110 may be, for example, a laser (e.g., a laser diode) that emits light having a particular operating wavelength in the infrared, visible, or ultraviolet portions of the electromagnetic spectrum. In operation, thelight source 110 emits an output beam oflight 125 which may be continuous-wave, pulsed, or modulated in any suitable manner for a given application. The output beam oflight 125 is directed downrange toward aremote target 130 located a distance D from thelidar system 100 and at least partially contained within a field of regard of thesystem 100. - Once the
output beam 125 reaches thedownrange target 130, thetarget 130 may scatter or, in some cases, reflect at least a portion of light from theoutput beam 125, and some of the scattered or reflected light may return toward thelidar system 100. In the example ofFIG. 2 , the scattered or reflected light is represented byinput beam 135, which passes through thescanner 120, which may be referred to as a beam scanner, optical scanner, or laser scanner. Theinput beam 135 passes through thescanner 120 to themirror 115, which may be referred to as an overlap mirror, superposition mirror, or beam-combiner mirror. Themirror 115 in turn directs theinput beam 135 to thereceiver 140. - The
input beam 135 may include light from theoutput beam 125 that is scattered by thetarget 130, light from theoutput beam 125 that is reflected by thetarget 130, or a combination of scattered and reflected light fromtarget 130. According to some implementations, thelidar system 100 can include an “eye-safe” laser that present little or no possibility of causing damage to a person's eyes. Theinput beam 135 may contain only a relatively small fraction of the light from theoutput beam 125. - The
receiver 140 may receive or detect photons from theinput beam 135 and generate one or more representative signals. For example, thereceiver 140 may generate an outputelectrical signal 145 that is representative of theinput beam 135. Thereceiver 140 may send theelectrical signal 145 to thecontroller 150. Depending on the implementation, thecontroller 150 may include one or more instruction-executing processors, an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), and/or other suitable circuitry configured to analyze one or more characteristics of theelectrical signal 145 in order to determine one or more characteristics of thetarget 130, such as its distance downrange from thelidar system 100. More particularly, thecontroller 150 may analyze the time of flight or phase modulation for the beam oflight 125 transmitted by thelight source 110. If thelidar system 100 measures a time of flight of T (e.g., T representing a round-trip time of flight for an emitted pulse of light to travel from thelidar system 100 to thetarget 130 and back to the lidar system 100), then the distance D from thetarget 130 to thelidar system 100 may be expressed as D=c·T/2, where c is the speed of light (approximately 3.0×108 m/s). - The distance D from the
lidar system 100 is less than or equal to a maximum range RMAX of thelidar system 100. The maximum range RMAX (which also may be referred to as a maximum distance) of alidar system 100 may correspond to the maximum distance over which thelidar system 100 is configured to sense or identify targets that appear in a field of regard of thelidar system 100. The maximum range oflidar system 100 may be any suitable distance, such as 50 m, 200 m, 500 m, or 1 km, for example. - In some implementations, the
light source 110, thescanner 120, and thereceiver 140 may be packaged together within asingle housing 155, which may be a box, case, or enclosure that holds or contains all or part of thelidar system 100. Thehousing 155 includes awindow 157 through which thebeams controller 150 may reside within thesame housing 155 as thecomponents controller 150 may reside outside of thehousing 155. In one embodiment, for example, thecontroller 150 may instead reside within, or partially within, a perception component of a control architecture (described below). In some implementations, thehousing 155 includes multiple lidar sensors, each including a respective scanner and a receiver. Depending on the particular implementation, each of the multiple sensors can include a separate light source or a common light source. The multiple sensors can be configured to cover non-overlapping adjacent fields of regard or partially overlapping fields of regard, for example, depending on the implementation. In embodiments that include partially overlapping fields of regard, the multiple sensors may direct anoutput beam 125 at the same location on thetarget 130 at different points of time to obtain co-located depth data. - With continued reference to
FIG. 2 , theoutput beam 125 andinput beam 135 may be substantially coaxial. In other words, theoutput beam 125 andinput beam 135 may at least partially overlap or share a common propagation axis, so that theinput beam 135 and theoutput beam 125 travel along substantially the same optical path (albeit in opposite directions). As thelidar system 100 scans theoutput beam 125 across a field of regard, theinput beam 135 may follow along with theoutput beam 125, so that the coaxial relationship between the two beams is maintained. - Generally speaking, the
scanner 120 steers theoutput beam 125 in one or more directions downrange. To accomplish this, thescanner 120 may include one or more scanning mirrors and one or more actuators driving the mirrors to rotate, tilt, pivot, or move the mirrors in an angular manner about one or more axes, for example. WhileFIG. 2 depicts only asingle mirror 115, thelidar system 100 may include any suitable number of flat or curved mirrors (e.g., concave, convex, or parabolic mirrors) to steer or focus theoutput beam 125 or theinput beam 135. For example, the first mirror of the scanner may scan theoutput beam 125 along a first direction, and the second mirror may scan theoutput beam 125 along a second direction that is substantially orthogonal to the first direction. - A “field of regard” of the
lidar system 100 may refer to an area, region, or angular range over which thelidar system 100 may be configured to scan or capture distance information. When thelidar system 100 scans theoutput beam 125 within a 30-degree scanning range, for example, thelidar system 100 may be referred to as having a 30-degree angular field of regard. Thescanner 120 may be configured to scan theoutput beam 125 horizontally and vertically, and the field of regard of thelidar system 100 may have a particular angular width along the horizontal direction and another particular angular width along the vertical direction. For example, thelidar system 100 may have a horizontal field of regard of 10° to 120° and a vertical field of regard of 2° to 45°. - The one or more scanning mirrors of the
scanner 120 may be communicatively coupled to thecontroller 150, which may control the scanning mirror(s) so as to guide theoutput beam 125 in a desired direction downrange or along a desired scan pattern. In general, a scan (or scan line) pattern may refer to a pattern or path along which theoutput beam 125 is directed. Thelidar system 100 can use the scan pattern to generate a point cloud with points or “pixels” that substantially cover the field of regard. The pixels may be approximately evenly distributed across the field of regard, or distributed according to a particular non-uniform distribution. In some embodiments, for example as shown inFIG. 1B , two or more of the pixels may be co-located within the field of regard. - In operation, the
light source 110 may emit pulses of light which thescanner 120 scans across a field of regard of thelidar system 100. Thetarget 130 may scatter one or more of the emitted pulses, and thereceiver 140 may detect at least a portion of the pulses of light scattered by thetarget 130. Thereceiver 140 may receive or detect at least a portion of theinput beam 135 and produce an electrical signal that corresponds to theinput beam 135. Thecontroller 150 may be electrically coupled or otherwise communicatively coupled to one or more of thelight source 110, thescanner 120, and thereceiver 140. Thecontroller 150 may provide instructions, a control signal, or a trigger signal to thelight source 110 indicating when thelight source 110 should produce optical pulses, and possibly characteristics (e.g., duration, period, peak power, wavelength, etc.) of the pulses. Thecontroller 150 may also determine a time-of-flight value for an optical pulse based on timing information associated with when the pulse was emitted bylight source 110 and when a portion of the pulse (e.g., the input beam 135) was detected or received by thereceiver 140. - As indicated above, the
lidar system 100 may be used to determine the distance to one or moredownrange targets 130. By scanning thelidar system 100 across a field of regard, the system can be used to map the distance to a number of points (e.g., pixels) within the field of regard. A collection of pixels captured in succession (which may be referred to as a depth map, a point cloud, or a point cloud frame) may be rendered as an image or may be analyzed to identify or detect objects or to determine a shape or distance of objects within the field of regard. For example, a depth map may cover a field of regard that extends 60° horizontally and 15° vertically, and the depth map may include a frame of 100-2000 pixels in the horizontal direction by 4-400 pixels in the vertical direction. - The
lidar system 100 may be configured to repeatedly capture or generate point cloud frames of a field of regard at any suitable frame rate between approximately 0.1 frames per second (FPS) and approximately 1,000 FPS, for example. The point cloud frame rate may be substantially fixed or dynamically adjustable, depending on the implementation. In general, thelidar system 100 can use a slower frame rate (e.g., 1 Hz) to capture one or more high-resolution point clouds, and use a faster frame rate (e.g., 10 Hz) to rapidly capture multiple lower-resolution point clouds. - The field of regard of the
lidar system 100 can overlap, encompass, or enclose at least a portion of thetarget 130, which may include all or part of an object that is moving or stationary relative tolidar system 100. For example, thetarget 130 may include all or a portion of a person, vehicle, motorcycle, truck, train, bicycle, wheelchair, pedestrian, animal, road sign, traffic light, lane marking, road-surface marking, parking space, pylon, guard rail, traffic barrier, pothole, railroad crossing, obstacle in or near a road, curb, stopped vehicle on or beside a road, utility pole, house, building, trash can, mailbox, tree, any other suitable object, or any suitable combination of all or part of two or more objects. -
FIGS. 1A and 1B illustrateexample scan patterns 160, either or both of which thelidar system 100 ofFIG. 2 may produce. In particular, thelidar system 100 may be configured to scan the outputoptical beam 125 along thescan pattern 160. In some implementations, thescan pattern 160 corresponds to a scan across any suitable field of regard having any suitable horizontal field of regard (FORH) and any suitable vertical field of regard (FORV). For example, a certain scan pattern may have a field of regard represented by angular dimensions (e.g., FORH×FORV) 40°×30°, 90°×40°, or 60°×15°. WhileFIG. 1A depicts a conventional “zig-zag”pattern 160, other implementations may instead employ other patterns (e.g., parallel, horizontal scan lines), and/or other patterns may be employed in specific circumstances. For example, the “looping”pattern 160 depicted inFIG. 1B includes co-located pixels 166 (that is, a pair of co-located pixels at 166 a, and a different pair co-located pixels at 166 b) to capture depth data at the same horizontal and vertical angles in the field of regard at multiple points of time within the same point cloud frame. Although “co-located” is generally used to refer to pixels at the same location within the field of regard, in some implementations, co-located refers to pixels within a small, threshold distance of one another. In some implementations this threshold distance is defined as a difference in the vertical and/or horizontal angles. - In the example implementations and/or scenarios of
FIGS. 1A-1B ,reference line 162 represents a center of the field of regard ofscan pattern 160. Accordingly, if thescan pattern 160 has a 60°×15° field of regard, then thescan pattern 160 covers a ±30° horizontal range with respect toreference line 162 and a ±7.5° vertical range with respect toreference line 162. An azimuth (which may be referred to as an azimuth angle) may represent a horizontal angle with respect to thereference line 162, and an altitude (which may be referred to as an altitude angle, elevation, or elevation angle) may represent a vertical angle with respect to thereference line 162. - The
scan pattern 160 may include multiple points orpixels 164, and eachpixel 164 may be associated with one or more laser pulses and one or more corresponding distance measurements. A cycle ofscan pattern 160 may include a total of Px×Py pixels 164 (e.g., a two-dimensional distribution of Px by Py pixels). The number ofpixels 164 along a horizontal direction may be referred to as a horizontal resolution of thescan pattern 160, and the number ofpixels 164 along a vertical direction may be referred to as a vertical resolution of thescan pattern 160. - Each
pixel 164 may be associated with a distance (e.g., a distance to a portion of atarget 130 from which the corresponding laser pulse was scattered), one or more angular values. As an example, thepixel 164 may be associated with a distance value and two angular values (e.g., an azimuth and altitude) that represent the angular location of thepixel 164 with respect to thelidar system 100. A distance to a portion of thetarget 130 may be determined based at least in part on a time-of-flight measurement for a corresponding pulse. An angular value (e.g., an azimuth or altitude) may correspond to an angle (e.g., relative to reference line 162) of the output beam 125 (e.g., when a corresponding pulse is emitted from lidar system 100) or an angle of the input beam 135 (e.g., when an input signal is received by lidar system 100). In some implementations, thelidar system 100 determines an angular value based at least in part on a position of a component of thescanner 120. For example, an azimuth or altitude value associated with thepixel 164 may be determined from an angular position of one or more corresponding scanning mirrors of thescanner 120. In some implementations, eachpixel 164 is also associated with a time value, such as a time stamp indicative of when an output beam (e.g., the output beam 125) is emitted or when an input beam (e.g., the input beam 135) is detected. -
FIG. 3A illustrates anexample vehicle 300 with alidar system 302. Thelidar system 302 includes alaser 310 with multiple sensor heads 312A-D coupled to thelaser 310 via multiple laser-sensor links 314. Each of the sensor heads 312 may include some or all of the components of thelidar system 100 illustrated inFIG. 2 . - Each of the laser-
sensor links 314 may include one or more optical links and/or one or more electrical links. The sensor heads 312 inFIG. 3A are positioned or oriented to provide a greater than 30-degree view of an environment around the vehicle. More generally, a lidar system with multiple sensor heads may provide a horizontal field of regard around a vehicle of approximately 30°, 45°, 60°, 90°, 120°, 180°, 270°, or 360°. Each of the sensor heads 312 may be attached to, or incorporated into, a bumper, fender, grill, side panel, spoiler, roof, headlight assembly, taillight assembly, rear-view mirror assembly, hood, trunk, window, or any other suitable part of the vehicle. - In the example of
FIG. 3A , four sensor heads 312 are positioned at or near the four corners of the vehicle (e.g., each of the sensor heads 312 may be incorporated into a light assembly, side panel, bumper, or fender), and thelaser 310 may be located within the vehicle 300 (e.g., in or near the trunk). The four sensor heads 312 may each provide a 90° to 120° horizontal field of regard (FOR), and the four sensor heads 312 may be oriented so that together they provide a complete 360-degree view around the vehicle. As another example, thelidar system 302 may include six sensor heads 312 positioned on or around thevehicle 300, where each of the sensor heads 312 provides a 60° to 90° horizontal FOR. As another example, thelidar system 302 may include eight sensor heads 312, and each of the sensor heads 312 may provide a 45° to 60° horizontal FOR. As yet another example, thelidar system 302 may include six sensor heads 312, where each of the sensor heads 312 provides a 70° horizontal FOR with an overlap between adjacent FORs of approximately 10°. As another example, thelidar system 302 may include two sensor heads 312 which together provide a forward-facing horizontal FOR of greater than or equal to 30°. - Data from each of the sensor heads 312 may be combined or stitched together to generate a point cloud that covers a greater than or equal to 30-degree horizontal view around a vehicle. For example, the
laser 310 may include a controller or processor that receives data from each of the sensor heads 312 (e.g., via a corresponding electrical link 320) and processes the received data to construct a point cloud covering a 360-degree horizontal view around a vehicle or to determine distances to one or more targets. The point cloud or information from the point cloud may be provided to avehicle controller 322 via a corresponding electrical, optical, orradio link 320. Thevehicle controller 322 may include one or more CPUs, GPUs, and a non-transitory memory with persistent components (e.g., flash memory, an optical disk) and/or non-persistent components (e.g., RAM). - In some implementations, the point cloud is generated by combining data from each of the multiple sensor heads 312 at a controller included within the
laser 310, and is provided to thevehicle controller 322. In other implementations, each of the sensor heads 312 includes a controller or processor that constructs a point cloud for a portion of the 360-degree horizontal view around the vehicle and provides the respective point cloud to thevehicle controller 322. Thevehicle controller 322 then combines or stitches together the points clouds from the respective sensor heads 312 to construct a combined point cloud covering a 360-degree horizontal view. Still further, thevehicle controller 322 in some implementations communicates with a remote server to process point cloud data. - In some embodiments, respective controllers included within the multiple sensor heads 312 process their respectively constructed point clouds to detect, analyze, and/or correct for distortion, including rolling shutter distortion, prior to providing the constructed point cloud to the
laser 310 or the vehicle controller 312. In other embodiments, a controller included within thelaser 310 processes the point cloud to correct for detect, analyze, and/or correct distortion, including rolling shutter distortion, prior to providing the point cloud to thevehicle controller 322. For example, in implementations where the multiple sensor heads 312 construct respective point clouds, thelaser 310 may analyze the constructed point clouds to detect and correct for distortion prior to combining or stitching the point clouds together. As another example, in implementations where the multiple sensors heads 312 do not construct the point cloud, thelaser 310 detects, analyzes, and/or corrects for the distortion while combining the raw data from the multiple sensor heads. In still other embodiments, thevehicle controller 322 analyzes point clouds data provided by thelaser 310 or the multiple sensor heads 312 to detect, analyze, and/or correct for distortion, including rolling shutter distortion, prior to processing the point cloud via the various components of thevehicle controller 322 described herein. - In any event, the
vehicle 300 may be an autonomous vehicle where thevehicle controller 322 provides control signals tovarious components 330 within thevehicle 300 to maneuver and otherwise control operation of thevehicle 300. Thecomponents 330 are depicted in an expanded view inFIG. 3A for ease of illustration only. Thecomponents 330 may include anaccelerator 340,brakes 342, avehicle engine 344, asteering mechanism 346,lights 348 such as brake lights, head lights, reverse lights, emergency lights, etc., agear selector 350, and/or other suitable components that effectuate and control movement of thevehicle 300. Thegear selector 350 may include the park, reverse, neutral, drive gears, etc. Each of thecomponents 330 may include an interface via which the component receives commands from thevehicle controller 322 such as “increase speed,” “decrease speed,” “turn left 5 degrees,” “activate left turn signal,” etc. and, in some cases, provides feedback to thevehicle controller 322. - In some implementations, the
vehicle controller 322 receives point cloud data from the sensor heads 312 orlaser 310 via thelink 320 and analyzes the received point cloud data to sense or identify targets 130 (seeFIG. 1 ) and their respective locations, distances, speeds, shapes, sizes, type of target (e.g., vehicle, human, tree, animal), etc. Thevehicle controller 322 then provides control signals via thelink 320 to thecomponents 330 to control operation of the vehicle based on the analyzed information. One, some or all of thecomponents 330 may be the operational subsystems, or may be included within the operational subsystems, that receive the control signals. - In addition to the
lidar system 302, thevehicle 300 may also be equipped with other sensors such as a RGB or stereo camera, a thermal imager, a conventional radar (none illustrated to avoid clutter), etc. The sensors can provide additional data to thevehicle controller 322 via wired or wireless communication links. Further, thevehicle 300 in an example implementation includes a microphone array operating as a part of an acoustic source localization system configured to determine sources of sounds. - As another example,
FIG. 3B illustrates avehicle 360 in which alaser 370 is optically coupled to six sensor heads 372, each of which may be similar to one of the sensor heads 312 ofFIG. 3A . The sensor heads 372A and 372G are disposed at the front of the hood, the sensor heads 372B and 372F are disposed in the side view mirrors, and the sensor heads 372C-E are disposed on the trunk. In particular, thesensor head 372D is oriented to face backward relative to the orientation of thevehicle 360, and the sensor heads 372C-E are oriented at approximately 45 degrees relative to the axis of orientation of thesensor head 372D. -
FIG. 4 depicts anexample point cloud 390 that is generated by a lidar system scanning the vehicle environment (e.g., thelidar system 100 ofFIGS. 1 and 2 or thelidar system 302 ofFIG. 3A ). The vehicle environment may include a divided highway that has a median separating the two directions of traffic, with multiple lanes in each direction. The vehicle environment may also include light poles on either side of the road and/or in the median. Thepoint cloud 390 ofFIG. 4 corresponds to an example embodiment in which a lidar device captures a roughly 60 degree horizontal field of regard in front of thevehicle 300. Thepoint cloud 390 may have been generated using the sensor heads 312A or 312D ofFIG. 3A , or the sensor heads 372A or 372G ofFIG. 3B , for example. While depicted as a visual image inFIG. 4 , it is understood that, in some embodiments, thepoint cloud 390 is not actually rendered or displayed at any time. - As seen in
FIG. 4 , thepoint cloud 390 depicts a ground plane 394 (here, the road surface) as a number of substantially continuous scan lines, and also depicts, above theground plane 394, a number of objects 396 (also referred to as “point cloud objects”). For clarity, only a small number of the objects shown inFIG. 4 are labeled with a reference number. Distinct ones of the objects 396 within thepoint cloud 390 may be identified by a segmentation module of thevehicle controller 322 ofFIG. 3A . For example, the segmentation module may detect substantial gaps and/or other discontinuities in the scan lines of theground plane 394, and identify groups of points in the vicinity of those discontinuities as discrete objects. The segmentation module may determine which points belong to the same object using any suitable rules, algorithms or models. Once the objects 396 are identified, a classification module of thevehicle controller 322 ofFIG. 3A may attempt to classify the objects, and a tracking module of thevehicle controller 322 ofFIG. 3A may attempt to track the classified objects (and, in some embodiments/scenarios, unclassified objects) across future point clouds similar to point cloud 390 (i.e., across multiple point cloud frames). - For various reasons, it may be more difficult for the segmentation module to identify certain objects 396, and/or for the classification module to classify certain objects 396, within the
point cloud 390. As can also be seen inFIG. 4 , for example, avehicle 396A may be relativity easy to identify and classify due to the high density of points, but more difficult to classify due to the lack of points within the lidar shadow (i.e., the absence or relative scarcity of points) along the lower right portions of thevehicle 396C created by the relative positioning of thevehicle 396C with respect to thevehicle 300. Alight pole 396B may also be relatively easy to identify as an object, due to the high density of points and readily identifiable shape, but may not be as easy to classify due to the fact that large portions of thelight pole 396B are distorted (e.g., as shown by the middle portion leaning away from the vehicle 300). Thevan 396C may be relatively easy to identify as an object, but more difficult to classify due to distortion causing thevan 396C to appears as though it is tilted away from thevehicle 300. Still other objects may be difficult to identify, classify and/or track due to their small size and/or low profile. For example, while not shown inFIG. 4 , the segmentation module may identify (and the classification module may classify) lane markings within thepoint cloud 390. The lane markings may appear as small but abrupt deviations in the path of the scan lines, for example, with those deviations collectively forming a line pattern that aligns with the direction of travel of the autonomous vehicle (e.g., approximately normal to the curve of the scan lines). - Despite such difficulties, the segmentation module, classification module and/or tracking module may use techniques that make object identification, classification and/or tracking highly accurate across a very wide range of scenarios, with scarce or otherwise suboptimal point cloud representations of objects. For example, the segmentation module, classification module and/or tracking module may include neural networks that were trained using data (e.g., manually labeled scenes) corresponding to a very large number of diverse environments/scenarios (e.g., with various types of objects at different distances, in different orientations, with different degrees of concealment, in different weather and/or lighting conditions, and so on).
- As another example, in embodiments in which the
vehicle controller 322 ofFIG. 3A detects object distortion, a distortion correction module of thevehicle controller 322 analyzes the identified objects 396 to detect and correct distortion. For example, as depicted inFIG. 4 , thelight pole 396B and thevan 396C exhibit distortion causing thelight pole 396B andvan 396C to appear as though they are skewed away from thevehicle 300. Accordingly, the distortion correction module may detect the distortion, modify thepoint cloud 390 to correct for the distortion, analyze the distortion to determine the velocity of one or more objects 396, etc. Correcting the distortion of the objects 396 may improve the accuracy of the classification module and/or the tracking module (see description ofFIG. 5 , below), for example. Ultimately, improving the correlation between thepoint cloud 390 and the actual vehicle environment may provide a smoother ride for passengers and/or reduce the likelihood of a collision. -
FIG. 5 illustrates anexample software architecture 400, which may be utilized in thelidar system FIG. 2 or 3 , respectively. Thesoftware architecture 400 receives as input M sets ofsensor data 402 generated by M different sensors, with M being any suitable integer equal to or greater than one. Thesensor data 402 may be data generated by the sensor heads 312 ofFIG. 3 or any of the other non-illustrated sensors included in thelidar system 302 ofFIG. 3 . For example, in addition to the lidar sensor heads 312, the lidar system may include cameras, radar devices, thermal imaging units, IMUs, and/or other sensor types. In one example, “sensor data 1” may include frames of point cloud data generated by afirst sensor head 312A, “sensor data 2” may include frames of point cloud data generated by asecond sensor head 312B, “sensor data 3” (not shown inFIG. 3 ) may include frames of digital images generated by a RGB camera, and so on. - The
sensor data 402 is input to aperception component 406 of thesoftware architecture 400, and is processed by theperception component 406 to generate perception signals 408 descriptive of a current state of the autonomous vehicle's environment. It is understood that the term “current” may actually refer to a very short time prior to the generation of any given perception signals 408, e.g., due to the short processing delay introduced by theperception component 406 and other factors. To generate the perception signals, the perception component may include asegmentation module 410, aclassification module 412, atracking module 414, and adistortion correction module 416. - The
segmentation module 410 is generally configured to identify distinct objects within thesensor data 402 representing the sensed environment. Depending on the embodiment and/or scenario, the segmentation task may be performed separately for each of a number of different types of sensor data, or may be performed jointly on a fusion of multiple types of sensor data. In some embodiments where lidar devices are used, thesegmentation module 410 analyzes point cloud frames to identify subsets of points within each frame that correspond to probable physical objects in the environment. In other embodiments, thesegmentation module 410 jointly analyzes lidar point cloud frames in conjunction with camera image frames to identify objects in the environment. Other suitable techniques, and/or data from other suitable sensor types, may also be used to identify objects. It is noted that, as used herein, references to different or distinct “objects” may encompass physical things that are entirely disconnected (e.g., with two vehicles being two different “objects”), as well as physical things that are connected or partially connected (e.g., with a vehicle being a first “object” and the vehicle's hitched trailer being a second “object”). - The
segmentation module 410 may use predetermined rules or algorithms to identify objects. For example, thesegmentation module 410 may identify as distinct objects, within a point cloud, any clusters of points that meet certain criteria (e.g., having no more than a certain maximum distance between all points in the cluster, etc.). Alternatively, thesegmentation module 410 may utilize a neural network that has been trained to identify distinct objects within the environment (e.g., using supervised learning with manually generated labels for different objects within test data point clouds, etc.), or another type of machine learning based model. For example, thesegmentation module 410 may analyze thepoint cloud 390 ofFIG. 4 to identify objects 396. - The
classification module 412 is generally configured to determine classes (labels, categories, etc.) for different objects that have been identified by thesegmentation module 410. Like thesegmentation module 410, theclassification module 412 may perform classification separately for different sets of thesensor data 402, or may classify objects based on data from multiple sensors, etc. Moreover, and also similar to thesegmentation module 410, theclassification module 412 may execute predetermined rules or algorithms to classify objects, or may utilize a neural network or other machine learning based model to classify objects. For example, theclassification module 412 may classify theobject 396C ofFIG. 4 as a truck and theobject 396A as a car. - In some implementations, the
classification module 412 communicates with an object database (not depicted) that stores information associated with object types. For example, the object database may include information that indicates how an object of the corresponding object type should appear in a point cloud. In some implementations, this indication may be a generic model for an object of the particular type. As one example, an object database record for a particular model of car may include a three-dimensional model of the car, to which objects identified by thesegmentation module 410 are compared during the classification process. In some additional implementations, the model includes indications of particular features of the object that have known shapes (e.g., a license plate, a tire, manufacturer emblem, etc.). As will be described below, this model of the object, including the particular features thereof, may be used to detect whether objects identified by thesegmentation module 410 actually have a skewed shape, or instead have a distorted appearance in the point cloud frame due to rolling shutter distortion. - The
tracking module 414 is generally configured to track distinct objects over time (e.g., across multiple lidar point cloud or camera image frames). The tracked objects are generally objects that have been identified by thesegmentation module 410, but may or may not be objects that were classified by theclassification module 412, depending on the embodiment and/or scenario. Thesegmentation module 410 may assign identifiers to identified objects, and thetracking module 414 may associate existing identifiers with specific objects where appropriate (e.g., for lidar data, by associating the same identifier with different clusters of points, at different locations, in successive point cloud frames) Like thesegmentation module 410 and theclassification module 412, thetracking module 414 may perform separate object tracking based on different sets of thesensor data 402, or may track objects based on data from multiple sensors. Moreover, and also similar to thesegmentation module 410 and theclassification module 412, thetracking module 414 may execute predetermined rules or algorithms to track objects, or may utilize a neural network or other machine learning model to track objects. - The
distortion correction module 416 is generally configured to detect that an object is distorted, analyze distortion to determine a relative velocity of the object, and/or correct distorted objects such that they no longer appear distorted. In one implementation, newly identified objects by thesegmentation module 410 are analyzed by thedistortion correction module 416. In some embodiments, thedistortion correction module 416 jointly analyzes lidar point cloud frames in conjunction with camera image frames to determine whether objects are distorted. In some implementations, the newly identified objects are analyzed by thedistortion correction module 416 prior to the objects being analyzed by theclassification module 412 and/or thetracking module 414. Additionally or alternatively, objects tracked by thetracking module 414 are analyzed by thedistortion correction module 416 when thesensor data 402 indicates a new point cloud frame. In some embodiments, thedistortion correction module 416 analyzes the amount of distortion of an object to determine a velocity or relative velocity of at least one of the distorted object, another object, or the vehicle. Moreover, and also similar to thesegmentation module 410, theclassification module 412, and thetracking module 414, thedistortion correction module 416 may execute predetermined rules or algorithms, or may utilize a neural network or other machine learning model to detect distortion, correct distortion, and/or correlate an amount of distortion to relative velocity. - The
software architecture 400 also includes aprediction component 420, which processes the perception signals 408 to generateprediction signals 422 descriptive of one or more predicted future states of the autonomous vehicle's environment. For a given object, for example, theprediction component 420 may analyze the type/class of the object (as determined by the classification module 412) along with the recent tracked movement of the object (as determined by the tracking module 414) to predict one or more future positions of the object. As a relatively simple example, theprediction component 420 may assume that any moving objects will continue to travel on their current direction and with their current speed, possibly taking into account first- or higher-order derivatives to better track objects that have continuously changing directions, objects that are accelerating, and so on. In some embodiments, theprediction component 420 also predicts movement of objects based on more complex behaviors. For example, theprediction component 420 may assume that an object that has been classified as another vehicle will follow rules of the road (e.g., stop when approaching a red light), and will react in a certain way to other dynamic objects (e.g., attempt to maintain some safe distance from other vehicles). Theprediction component 420 may inherently account for such behaviors by utilizing a neural network or other machine learning model, for example. Theprediction component 420 may be omitted from thesoftware architecture 400, in some embodiments. - A
mapping component 430 obtains map data (e.g., a digital map including the area currently being traversed by the autonomous vehicle) and/or navigation data (e.g., data indicating a route for the autonomous vehicle to reach the destination, such as turn-by-turn instructions), and outputs the data (possibly in a converted format) as mapping and navigation signals 432. In some embodiments, the mapping andnavigation signals 432 include other map- or location-related information, such as speed limits, traffic indicators, and so on. Thesignals 432 may be obtained from a remote server (e.g., via a cellular or other communication network of the autonomous vehicle, or of a smartphone coupled to the autonomous vehicle, etc.), and/or may be locally stored in a persistent memory of the autonomous vehicle. - A
motion planner 440 processes the perception signals 408, the prediction signals 422, and the mapping andnavigation signals 432 to generatedecisions 442 regarding the next movements of the autonomous vehicle. Themotion planner 440 may use any suitable architecture or technique to generate thedecisions 442, such as one or more trained machine learning models, or a model predictive control (MPC) technique, for example. Depending on the type of themotion planner 440, thedecisions 442 may be operational parameters (e.g., braking, speed and steering parameters) or particular maneuvers (e.g., turn left, move to right lane, move onto shoulder of road, etc.). In some embodiments, thedecisions 442 may be provided to one or more components of the autonomous vehicle (e.g., if thedecisions 442 indicate specific operational parameters) such as thecomponents 330 ofFIG. 3 , or may be provided to one or more intermediate stages that convert thedecisions 442 to operational parameters (e.g., if the decisions indicate specific maneuvers) prior to providing the operation parameters to thevehicle components 330. -
FIG. 6 is a block diagram of anexample computing system 500 for controlling an autonomous vehicle. Thecomputing system 500 may be integrated within an autonomous vehicle in any suitable manner, and at any suitable location or locations within the vehicle. Thecomputing system 500 may be included, or partially included, within thevehicle controller 322 ofFIG. 3 , for example. Thecomputing system 500 includes one ormore processors 502, and amemory 504 storingvehicle controller instructions 506. The one ormore processors 502 execute thevehicle controller instructions 506 to implement the disclosed techniques, including, in one embodiment, thesoftware architecture 400 ofFIG. 5 . - In embodiments where the processor(s) 502 include more than a single processor, each processor may be a different programmable microprocessor that executes software instructions stored in the
memory 504. Alternatively, each of the processor(s) 502 may be a different set of such microprocessors, or a set that includes one or more microprocessors and one or more other processor types (e.g., ASICs, FPGAs, etc.) for certain functions. - The
memory 504 may include one or more physical memory devices with non-volatile memory. Any suitable memory type or types may be used, such as ROM, solid-state drives (SSDs), hard disk drives (HDDs), and so on. The processor(s) 502 are coupled to thememory 504 via a bus orother network 508. Thenetwork 508 may be a single wired network, or may include any suitable number of wired and/or wireless networks. For example, thenetwork 508 may be or include a controller area network (CAN) bus, a Local Interconnect Network (LNN) bus, and so on. - Also coupled to the
network 508 are avehicle control interface 510, apassenger interface 512, asensor interface 514, and anetwork interface 516. Each of theinterfaces computing system 500. - The
vehicle control interface 510 is generally configured to provide control data generated by the processor(s) 502 to the appropriate operational subsystems of the autonomous vehicle, such that the appropriate subsystems can effectuate driving decisions made by the processor(s) 502. Referring toFIG. 5 , for example, thevehicle control interface 510 may provide the control signal operational parameters associated with thedecisions 442 to the appropriate subsystem(s) (e.g.,accelerator 340,brakes 342, andsteering mechanism 346 ofFIG. 3A ). In some embodiments, thevehicle control interface 510 includes separate interface hardware, firmware and/or software for different operational subsystems. - The
passenger interface 512 is generally configured to provide alerts, warnings, notifications, and/or other information to one or more passengers of the autonomous vehicle. In some embodiments where the vehicle is not fully autonomous (e.g., allowing human driving in certain modes and/or situations), theinterface 512 may specifically provide such information to the driver (e.g., via dashboard indicators, etc.). As just one example, thepassenger interface 512 may cause a display and/or speaker in the vehicle to generate an alert when the processor(s) 502 (executing the vehicle controller instructions 506) determine that a collision with another object is likely. As another example, thepassenger interface 502 may cause a display in the vehicle to show an estimated time of arrival (ETA) to passengers. In some embodiments, thepassenger interface 512 also permits certain user inputs. If the vehicle supports passenger selection of specific driving styles, for example, thepassenger interface 512 may cause a display to present a virtual control (e.g., button) that a passenger may activate (e.g., touch, scroll through, etc.) to select a particular driving style. - The
sensor interface 514 is generally configured to convert raw sensor data from one or more sensor devices (e.g., lidar, camera, microphones, thermal imaging units, IMUs, etc.) to a format that is consistent with a protocol of thenetwork 508 and is recognized by one or more of the processor(s) 502. Thesensor interface 514 may be coupled to thelidar system 100 ofFIG. 1 or 2 or thelidar system 302 ofFIG. 3A , for example, with thesensor interface 514 converting point cloud data to an appropriate format. In some embodiments, thesensor interface 514 includes separate interface hardware, firmware and/or software for each sensor device and/or each sensor type. - The
network interface 516 is generally configured to convert data received from one or more devices or systems external to the autonomous vehicle to a format that is consistent with a protocol of thenetwork 508 and is recognized by one or more of the processor(s) 502. In some embodiments, thenetwork interface 516 includes separate interface hardware, firmware and/or software for different external sources. For example, a remote mapping/navigation server may send mapping and navigation/route data to thecomputing system 500 via a cellular network interface of thenetwork interface 516, while one or more peer vehicles (e.g., other autonomous vehicles) may send data (e.g., current positions of the other vehicles) to thecomputing system 500 via a WiFi network interface of thenetwork interface 516. Other types of external data may also, or instead, be received via thenetwork interface 516. For example, thecomputing system 500 may use thenetwork interface 516 to receive data representing rules or regulations (e.g., speed limits), object positions (e.g., road rails, overhanging signage, etc.), and/or other information from various infrastructure devices or systems. - In some embodiments, the
vehicle controller instructions 506 utilize external data that is received by thenetwork interface 516 from other vehicles, infrastructure, and/or other external devices/systems as an input to one or more components. In such an embodiment, the external data may include raw sensor data that is indicative of the vehicle environment (but was generated off-vehicle), and/or may include higher-level information that was generated externally using raw sensor data. - The
network 508 may also couple to other types of interfaces and/or components, and/or some of the interfaces shown inFIG. 6 may be omitted (e.g., the network interface 516). Moreover, it is understood that thecomputing system 500 represents just one possible configuration for supporting the software architectures, functions, features, etc., described herein, and that others are also within the scope of this disclosure. -
FIG. 7 depicts anexample method 600 for determining the relative velocity of a point cloud object by fitting pixels onto a surface of the point cloud object, for example, one of the objects 396 of thepoint cloud 390 depicted inFIG. 4 . Themethod 600 may be implemented/performed by thecomputing system 500 ofFIG. 6 (e.g., by the processor(s) 502 when executing thevehicle controller instructions 506 stored in memory 504), for example, by another suitable computing device or system (such as a controller of thelaser 310 or the sensor heads 312 ofFIG. 3A , or thesoftware architecture 400 ofFIG. 5 ), or by a combination thereof. - At
bock 602, sensor data is received. The sensor data may include lidar depth data (e.g., sensed by thelidar system 100 ofFIGS. 1 and 2 , or thelidar system 302 ofFIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and/or any other type of sensor data discussed above in connection with thesensor data 402 ofFIG. 5 , for example. - At
block 604, a point cloud frame based on the sensor data received atblock 602 is obtained. In some embodiments, the sensor data received atblock 602 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud). In these embodiments, block 604 may include extracting the point cloud frame from rest of the sensor data received atblock 602. In other embodiments, block 604 includes analyzing depth or lidar data included in the sensor data received atblock 602 to generate the point cloud frame, as discussed in connection toFIG. 3 , for example. - At
block 606, two or more pixels that overlap a point cloud object are selected from the point cloud frame. In some implementations, the two points correspond to the same object as identified by a segmentation module. In other implementations, the two points are selected based on being within a threshold distance of one another. The selected points may be located on or near a two-dimensional surface corresponding to at least part of the point cloud object and captured at different times. Each of the pixels may be associated with depth data and time data. In one scenario,pixels FIG. 1A overlap the same point cloud object and are selected atblock 606. In this scenario, thepixels -
a′(x−v x t)+b′(y−v y t)+c′(z−v z t)+1=0 (Eq. 1) - Accordingly, as it is generally used herein, “fitting” the pixels to a plane includes modeling the point cloud object's surface as a two-dimensional plane moving through the vehicle environment such that the pixels are included within (i.e., satisfy the equation for) the two-dimension plane.
- At
block 608, a relative velocity of the point cloud object is determined based on depth and time data associated with the selected pixels. In some implementations, it is assumed that the plane is not moving along the vertical axis with respect to the vehicle. Using this assumption, the motion of the two-dimensional plane may be represented using the following equation: -
a′(x−v x t)+b′(y−v y t)+c′(z)+1=0 (Eq. 2) - where a′, b′, and c′ are constants, x, y, and z are indicative of position with respect to the x, y, and z axes (such as the longitudinal, lateral, and vertical axes), respectively, and vx and vy are the component of the plane velocity in the x and y directions respectively. To calculate relative velocity based on only the two selected pixels, it may be further assumed that the plane is not moving along the y-axis (e.g., lateral axis) and that the plane is oriented to be perpendicular to the x-axis (e.g., longitudinal axis). That is, it is assumed the plane is fixed in a vertical orientation and moving solely along the longitudinal axis. Using these additional assumptions, the equation of the plane may be represented as:
-
a′(x−v x t)+1=0 (Eq. 3) - The time and depth data associated with the first and second pixels can then be plugged into Equation 3 to solve for vx (as well as a′). That is, depth data x1 captured at time t1 and depth data x2 captured at time t2 associated with the first and second pixels is input into a best fit algorithm (such as a least square sum algorithm, a polynomial regression algorithm, etc.) to solve for vx. Because it is assumed that the plane is only moving in the longitudinal direction, vx is the relative velocity of the point cloud object.
- In other implementations, three or more pixels (such as
pixels FIG. 1A ) are selected atblock 606. In these implementations, the equation of the plane determined with respect to the first two selected pixels is rotated about a single axis to fit the third pixel into the plane. Said another way, rotating the plane about a single axis to fit the third pixel solves for one of b′ or c′. Assuming that the other constant is 0, thenEquation 1 can be solved for vx and vy or vx and vz (as well as a′) using another best fit algorithm that utilizes the time and depth data for the three selected pixels as an input. - In still further implementations, four or more pixels (such as
pixels FIG. 1A ) are selected atblock 606. In these implementations, the equation of the plane determined with respect to the first two selected pixels is rotated about two axes to fit the third and fourth pixels into the plane. Said another way, rotating the plane about two axes to fit the third and fourth pixels solves for both of b′ and c′. Accordingly,Equation 1 can be solved for vx, vy and vz (as well as a′) using another best fit algorithm that utilizes the time and depth data for the four selected pixels as an input. - In yet further implementations, rather than assuming the point cloud object is a plane, block 608 assumes that the point cloud object can have any shape. Thus, using the two selected pixels, the relative velocity of the point cloud object can be calculated using the following equation:
-
- where Δd is the depth difference between the two selected pixel and Δt is the time difference between when the depth data for the two pixels was captured. In some scenarios, the first and second pixels are part of the same scan line that traverses the vehicle environment. In other scenarios, the first pixel is part of a first scan line of pixels that are displaced from one another in a first direction and the second pixel is part of a second scan line of pixels that is substantially parallel to the first scan line and comprised of pixels that are displaced from one another in a second direction substantially orthogonal to the first direction.
- In some embodiments, the
method 600 includes one or more additional blocks not shown inFIG. 7 . For example, in some embodiments, another two or more pixels that overlap the point cloud object in a different region may be selected. This additional block may then calculate a relative velocity of this different region using the techniques described above with respect toblocks - As another example, the
method 600 may include a block to detect that the point cloud object is distorted. Accordingly, atblock 606, two or more pixels that are part of a scan line that traverses the vehicle environment in a first direction (e.g.,pixels FIG. 1A ) and two or more pixels that are part of a scan line that traverses the vehicle environment in a second direct (e.g.,pixels pixel 164 a minus the depth atpixel 164 b has a different polarity than the depth atpixel 164 c minus the depth atpixel 164 d), then the point cloud object exhibits rolling shutter distortion. Additionally or alternatively, this additional block includes comparing the magnitude of the two relative velocities. In some aspects, a rate of change between the relative velocities is analyzed to determine a relative acceleration. In other aspects, when the two relative velocities have approximately the same magnitude, it is likely that all four pixels overlap the point cloud object. Therefore, one of the first relative velocity, the second relative velocity, or an average relative velocity therebetween is associated with the point cloud object, for example, by theperception component 406 ofFIG. 5 . It should be appreciated that if the magnitude of one of the relative velocities exceeds a threshold value, then it is likely that one of the pixels does not actually overlap the point cloud object. Therefore, relative velocities that exceed this threshold may be discarded. - As another additional block, the
method 600 may include correcting the rolling shutter distortion based on the relative velocity determined atblock 608. For example, a machine learning module (e.g., one that is trained using reinforcement learning) may associate an amount of distortion to relative velocity and point cloud object depth. In particular, the machine learning model may operate on (i.e., use as inputs) the relative velocity determined atblock 608 and the depth data of the selected pixels to determine a correction factor to compensate for delay introduced by successive pixels and/or scan lines included in a scan pattern. - Another additional block of the
method 600 may include determining an object velocity (as opposed to relative velocity). For example, a relative velocity of a point cloud object that is known to be stationary (e.g., the point cloud object is classified as a light pole, a street sign, a tree, a bench, a power line, a building, a median, etc.) may be determined using the techniques ofblocks - As still another example additional block, the point cloud object (either corrected or uncorrected) and the relative velocity may be analyzed, for example by a portion of the
perception component 406 ofFIG. 5 (e.g.,classification module 412 and/or tracking module 414) and/or theprediction component 420 ofFIG. 5 , to classify the object, track the object, and/or predict a future state of the vehicle environment. This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the current perceived state and/or the predicted future state. As one example, if the relative velocity indicates that the point cloud object is rapidly increasing its speed in the direction towards the vehicle along the longitudinal axis, then the block may include generate a control signal to apply the vehicle brakes. -
FIGS. 8, 9, and 10 depictexample methods point cloud 390 depicted inFIG. 4 . Themethod 700 ofFIG. 8 relates to embodiments that determine relative velocity based on a scan pattern that includes successive scan lines that overlap the object, while themethod 720 ofFIG. 9 relates to embodiments that determine relative velocity based on depth data associated with co-located pixels. Both of themethods method 740 ofFIG. 10 relates to embodiments that correct point cloud distortion based on tracked object velocity. Themethods computing system 500 ofFIG. 6 (e.g., by the processor(s) 502 when executing thevehicle controller instructions 506 stored in memory 504), for example, by another suitable computing device or system (such as a controller of thelaser 310 or the sensor heads 312 ofFIG. 3A , or thesoftware architecture 400 ofFIG. 5 ), or by a combination thereof. - Referring first to the
method 700 ofFIG. 8 , sensor data is received atblock 702. The sensor data may include lidar depth data (e.g., sensed by thelidar system 100 ofFIGS. 1 and 2 , or thelidar system 302 ofFIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with thesensor data 402 ofFIG. 5 , for example. - At
block 704, a point cloud frame based on the sensor data received atblock 702 is obtained. In some embodiments, the sensor data received atblock 702 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud). In these embodiments, block 704 includes extracting the point cloud frame from the rest of the sensor data received atblock 702. In other embodiments, block 704 includes analyzing depth or lidar data included in the sensor data received atblock 702 to generate the point cloud frame, as discussed in connection toFIG. 3 , for example. - At
block 706, a point cloud object within the point cloud frame obtained atblock 704 is identified. In some embodiments, block 706 includes partitioning the point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and/or tracking movement of classified objects over time, for example, by utilizing aperception component 406 of thevehicle controller 400 ofFIG. 5 . In one example, atracking module 414 of theperception component 406 tracks objects across multiple point cloud frames. In this example, thetracking module 414 identifies objects within the point cloud frame obtained atblock 704 that are currently being tracked. In another example, thesegmentation module 410 identifies a new object that is not being tracked by thetracking module 416. - At
block 708, the point cloud object identified atblock 706 is determined to be skewed relative to an expected configuration of the point cloud object. For example, the point cloud object may be skewed due to the above-described rolling shutter distortion. In some implementations, the point cloud object is classified (for example by aclassification module 412 of the perception component 406) as an object that has a particular shape (e.g., an Acme Sedan, a stop sign, a person, etc.). Accordingly, determining that the object is skewed may include comparing a bound of the point cloud object to a bound of a reference object associated with the classification module. - In some implementations, the expected configuration is determined based upon the relative position between the vehicle and the object. For example, the
classification module 412 may associate the point cloud object with a generic object of the same type of object. The generic object may be rotated and/or scaled based on the determined relative position to determine the expected configuration of the point cloud object. Skew may then be determined by comparing the expected configuration and the apparent/sensed configuration, and identifying a substantial difference (e.g., greater than a threshold difference) between the two. In other implementations and/or scenarios, it is determined that the object is skewed because a bound of the point cloud object is determined to be curved when it is known that the point cloud object should instead have a straight bound. For example, theclassification module 412 may have classified the point cloud object as a type associated with a rule that side bounds should be approximately vertical (e.g., a truck container). Thus, the expected configuration of that bound is a line. - In yet other implementations, the expected configuration is determined based upon map data of the environment through which the vehicle is moving. For example, as illustrated in
FIG. 6 , the vehicle controller may be configured to receive mapping data via thenetwork interface 516. The mapping data may include a plurality of images associated with location and/or geoposition data. Accordingly, for objects that are stationary and/or are generally permanent in the environment, the mapping data may include image data that is representative of the actual configuration of the object. Thus, based on positioning data received over a vehicle bus and a relative position of the point cloud object with respect to the vehicle, the representation of the object as included in the mapping data can be identified. As a result, instead of identifying and/or classifying the point cloud object as a particular type of object and rotating and/or scaling the generic object, the point cloud object may be compared to the corresponding representation in the mapping data to determine a relative viewing angle between the point cloud frame and the representation in the map data. Based on this relative viewing angle, the point cloud object may be rotated and/or scaled such that the point cloud object and the map data are analyzed from the same perspective. - At
block 710, the relative velocity of the point cloud object identified atblock 706 is determined by analyzing the skew of the object. For example, a machine learning model (e.g., trained using reinforcement learning) may associate an amount of skew to relative velocity. In particular, the machine learning model may operate on (i.e., use as inputs) the point cloud object identified atblock 706 and the expected configuration. - In some scenarios, the point cloud object identified at
block 706 is skewed along multiple axes. Accordingly, the skew with respect to each axis may be analyzed to determine the components of the velocity vector with respect to each of these axes. For example, a lateral skew is analyzed to determine a lateral component of the relative velocity and a longitudinal skew is analyzed to determine a longitudinal component of the relative velocity. Accordingly, the machine learning model may analyze lateral bounds of the point cloud object and the expected configuration to detect the lateral skew and longitudinal bounds of the point cloud object and the expected configuration to detect the longitudinal skew. In some implementations, thetracking module 414, for example, associates the determined relative velocity with the point cloud object. - In some embodiments, the
method 700 includes one or more additional blocks not shown inFIG. 8 . For example, in some embodiments, the sensor data received atblock 702 includes velocity data indicative of the vehicle velocity (such as velocity data included in component reporting messages transmitted over a vehicle bus). In this example, themethod 700 may include an additional block where the vehicle velocity is subtracted from the relative velocity determined atblock 710 to determine the velocity of the point cloud object. As another example, themethod 700 may include correcting the skew based on the relative velocity determined atblock 710. To this end, the point cloud object may be adjusted to align with the expected configuration. As yet another example, the point cloud object (either corrected or uncorrected) and the relative velocity may be analyzed, for example by theprediction component 420 ofFIG. 5 , to predict a future state of the vehicle environment. This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state. As one example, if the relative velocity indicates that the point cloud object is rapidly increasing its speed in the direction towards the vehicle along the longitudinal axis, then the block may include generate a control signal to apply the vehicle brakes. -
FIG. 9 depicts anexample method 720 for determining relative velocity of a point cloud object based on depth data associated with co-located pixels. In one implementation, the lidar system (such as thelidar system 100 ofFIGS. 1 and 2 ) is configured to follow a scan pattern that includes co-located pixels, for example, the co-located pixels 166 include inscan pattern 160 ofFIG. 1B . In additional or alternative implementations, the sensor heads 312A and 312D of thelidar system 302 ofFIG. 3A are configured to implement complimentary, overlapping scan patterns wherein at least one pixel from the scan pattern implemented by thesensor head 312A is co-located with at least one pixel from the scan pattern implemented by thesensor head 312D. In these implementations, the scan patterns are configured such that the depth data associated with the co-located pixels are not obtained at the same time. - The
method 720 begins with receiving sensor data atblock 722. The sensor data may include lidar depth data (e.g., sensed by thelidar system 100 ofFIGS. 1 and 2 , or thelidar system 302 ofFIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with thesensor data 402 ofFIG. 5 , for example.Block 722 may be the same or similar to block 702 of themethod 700. - At
block 724, a point cloud frame based on the sensor data received atblock 722 is obtained. In some embodiments, the sensor data received atblock 722 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud). In these embodiments, block 724 may include extracting the point cloud frame from the rest of the sensor data received atblock 722. In other embodiments, block 724 includes analyzing depth or lidar data included in the sensor data received atblock 722 to generate the point cloud frame, as discussed in connection toFIG. 3 , for example.Block 724 may be the same or similar to block 704 of themethod 700. - At
block 726, co-located pixels within the field of regard that overlap a point cloud object within the point cloud frame obtained atblock 724 are identified. In some embodiments, block 726 includes partitioning the point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and tracking movement of classified objects over time, for example, by utilizing aperception component 406 of thesoftware architecture 400 ofFIG. 5 . In one example, atracking module 414 of theperception component 406 tracks objects across multiple point cloud frames. In this example, thetracking module 414 identifies objects within the point cloud frame obtained atblock 724 that are currently being tracked. In another example, thesegmentation module 410 identifies a new object that is not being tracked by thetracking module 416. In any event, after any point cloud objects included in the cloud frame have been identified, block 726 includes determining that a pair of co-located pixels overlaps one of the point cloud objects. - At
block 728, depth data associated with the co-located pixels are analyzed to determine a depth difference (Δd). Generally, Δt is indicative of the time span between capturing/sensing the depth data associated with the two co-located pixels, and Δd is indicative of the distance the object moved, relative to the vehicle and either towards or away from the vehicle, during the time span Δt. Accordingly, atblock 730, relative velocity of the object may be determined by based on Δd and Δt. More particularly, in some implementations, the relative velocity is determined by dividing Δd by Δt. In some embodiments, thelidar system 100 includes a clock that time stamps the depth data when it is sensed at thereceiver 140. Said another way, blocks 728 and 730 may implement Equation 4 using the depth and time data associated with the co-located pixels as inputs. Further, if Δd is determined atblock 728 by subtracting the depth data of the later-sensed co-located pixel from the depth data of the earlier-sensed co-located pixel, the polarity of Δd is indicative of the direction of travel of the point cloud object. In particular, when Δd is a positive number, the point cloud object is moving away from the vehicle and, conversely, when Δd is a negative number, the point cloud object is moving towards the vehicle. - In some embodiments, the
method 720 includes one or more additional blocks not shown inFIG. 9 . For example, in some embodiments, the sensor data received atblock 722 includes velocity data indicative of the vehicle velocity (such as velocity data included in component reporting messages transmitted over a vehicle bus). In this example, themethod 720 may include an additional block where the vehicle velocity is subtracted from the relative velocity determined atblock 730 to determine the velocity of the point cloud object. - As another example, an additional block may include identifying a second set of co-located pixels that overlap the point cloud object at a different region of the point cloud object. This block may include determining a relative velocity of the point cloud object based on the second set of co-located pixels. The two relative velocities may then be compared to one other to detect whether the point cloud object is turning or rotating.
- Another additional block may include correcting rolling shutter distortion based on the relative velocity determined at
block 730. To this end, the point cloud object may be adjusted to compensate for the motion of the point cloud object during the collection of depth data that comprises the point cloud frame. For example, a machine learning module (e.g., trained using reinforcement learning) may associate an amount of distortion to relative velocity and point cloud object depth. In particular, the machine learning model may operate on (i.e., use as inputs) the relative velocity determined atblock 730 and the depth data of the co-located pixels to determine a correction factor to compensate for delay introduced by successive pixels and/or scan lines included in a scan pattern. - Another additional block of the
method 600 may include determining an object velocity (i.e. the object's absolute velocity, as opposed to its relative velocity). For example, a relative velocity of a point cloud object that is known to be stationary (e.g., the point cloud object is classified as a light pole, a street sign, a tree, a bench, a power line, a building, a median, etc.) may be determined using the techniques ofblocks - As yet another example, the point cloud object (either corrected or uncorrected) and the relative velocity may be analyzed, for example by the
prediction component 420 ofFIG. 5 , to predict a future state of the vehicle environment. This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state. As one example, if the relative velocity indicates that the point cloud object is rapidly increasing its speed in the direction towards the vehicle along the longitudinal axis, then the block may include generate a control signal to apply the vehicle brakes. -
FIG. 10 depicts anexample method 740 for correcting point cloud distortion based on tracked object velocity. Whereas themethods method 740 determines the relative velocity of the point cloud object across multiple point cloud frames. Accordingly, while this increases the time it takes to determine a relative velocity of a newly detected point cloud object, already-tracked objects may be associated with historic position data. Thus, by comparing a point cloud object as represented in a new point cloud frame to the historic position data, the relative velocity of an already-tracked object can be determined based on a single successive point cloud frame. - The
method 740 begins when sensor data is received atblock 742. The sensor data may include lidar depth data (e.g., sensed by thelidar system 100 ofFIGS. 1 and 2 , or thelidar system 302 ofFIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with thesensor data 402 ofFIG. 5 , for example.Block 742 may be the same or similar to performing block 702 of themethod 700, but with respect to subsequent sets of sensor data. - At
block 744, two or more point cloud frames based on the sensor data received atblock 742 are obtained. In some embodiments, the sensor data received atblock 742 includes the point cloud frames (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud). In these embodiments, block 744 may include extracting the point cloud frames from rest of the sensor data received atblock 742. In other embodiments, block 744 includes analyzing multiple sets of depth or lidar data included in the sensor data received atblock 742 to generate the point cloud frames, as discussed in connection withFIG. 3 , for example.Block 744 may be the same or similar to performing block 704 of themethod 700, but with respect to subsequent sets of sensor data. - At
block 746, a point cloud object is tracked across the two or more point cloud frames. In some embodiments, block 746 includes partitioning a first point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and tracking movement of the classified objects over time, for example, by utilizing aperception component 406 of thevehicle controller 400 ofFIG. 5 . In one example, atracking module 414 of theperception component 406 tracks objects across multiple point cloud frames. More particularly, thetracking component 414 may associate tracked objects with position data (such as a position within the field of regard, a position in the vehicle environment, and/or a position relative to the vehicle), shape data, orientation data, or other information. In various implementations, the position data may be indicative of one or more of a position of the centroid of the point cloud object, a particular feature of the point cloud object, or a particular bound of the point cloud object. In this example, the point cloud object identified in the first point cloud object can be detected in the second point cloud frame (i.e., tracked across the first and second point cloud frames) using the associated shape or orientation information. - At
block 748, a relative velocity of the point cloud object tracked atblock 746 is determined. In one example, the position data of the point cloud object as represented in the first point cloud frame is subtracted from the position data of the point cloud object as represented in the second point cloud frame. This difference is divided by the time difference between the first and second point cloud frames to determine the relative velocity of the tracked object. - At
block 750, rolling shutter distortion associated with the point cloud object tracked atblock 746 is corrected based on the relative velocity determined atblock 748. To correct the point cloud object, in one example, a machine learning module (e.g., trained using reinforcement learning) may associate an amount of distortion to relative velocity and point cloud object position information. In particular, the machine learning model may operate on (i.e., use as inputs) the relative velocity determined atblock 748 and position information associated with the point cloud object to determine a correction factor to compensate for delay introduced by successive pixels and/or scan lines included in a scan pattern. Accordingly, based on the scan pattern utilized to capture the point cloud data, the correction factor is applied to the tracked point cloud object to correct for the rolling shutter distortion. - As another example, a relative velocity of a point cloud object that is known to be stationary (e.g., the point cloud object is classified as a light post, a street sign, a tree, a bench, a power line, a building, a median, etc.) is determined using the techniques of
blocks blocks method 600. Because the point cloud object is known to be stationary, the relative velocity is comprised solely of the vehicle's velocity. Said another way, by determining the relative velocity with respect to a known, stationary object, the vehicle velocity can be determined. Subsequently, the vehicle velocity can be subtracted from the relative velocity determined with respect to the identified point cloud object to determine object velocity. - In some embodiments, the
method 740 includes one or more additional blocks not shown inFIG. 10 . For example, in some embodiments, the sensor data received atblock 742 includes velocity data indicative of the vehicle velocity (such as velocity data included in component reporting messages transmitted over a vehicle bus). In this example, themethod 740 may include an additional block where the vehicle velocity is subtracted from the relative velocity determined atblock 748 to determine the velocity of the point cloud object. As another example, the corrected point cloud object and the relative velocity may be analyzed, for example by theprediction component 420 ofFIG. 5 , to predict a future state of the vehicle environment. This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state. - Example Methods for Detecting that an Object is Distorted
-
FIGS. 11 and 12 depictexample methods point cloud 390 depicted inFIG. 4 , is distorted and not actually shaped in a manner that appears distorted. Themethod 800 ofFIG. 11 relates to embodiments that detect distortion by analyzing known shapes, and themethod 820 ofFIG. 12 detects distortion by analyzing sensor data other than just lidar or depth data. Themethods computing system 500 ofFIG. 6 (e.g., by the processor(s) 502 when executing thevehicle controller instructions 506 stored in memory 504), for example, by another suitable computing device or system (such as a controller of thelaser 310 or the sensor heads 312 ofFIG. 3A , or thesoftware architecture 400 ofFIG. 5 ), or by a combination thereof. - Referring first to the
method 800 ofFIG. 11 , sensor data is received atblock 802. The sensor data may include lidar depth data (e.g., sensed by thelidar system 100 ofFIGS. 1 and 2 , or thelidar system 302 ofFIG. 3A ), lidar point cloud data (e.g., stitched or compiled lidar depth data), camera data, and any other type of sensor data discussed above in connection with thesensor data 402 ofFIG. 5 , for example. - At
block 804, a point cloud frame based on the sensor data received atblock 802 is obtained. In some embodiments, the sensor data received atblock 802 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud). In these embodiments, block 804 may include extracting the point cloud frame from rest of the sensor data received atblock 802. In other embodiments, block 804 includes analyzing depth or lidar data included in the sensor data received atblock 802 to generate the point cloud frame, as discussed in connection toFIG. 3 , for example. - At
block 806, a point cloud object within the point cloud frame obtained atblock 804 is identified. In some embodiments, block 806 includes partitioning the point cloud frame into portions in accordance with probable boundaries between separate physical objects, determining object types/classes for at least some of those portions, and tracking movement of classified objects over time, for example, by utilizing aperception component 406 of thevehicle controller 400 ofFIG. 5 . In one example, atracking module 414 of theperception component 406 tracks objects across multiple point cloud frames. In this example, thetracking module 414 identifies objects within the point cloud frame obtained atblock 704 that are currently being tracked. In another example, thesegmentation module 410 may identifies a new object that is not being tracked by thetracking module 416. - At
block 808, the point cloud object identified atblock 806 is analyzed to identify a feature that has an expected shape. For example, theclassification module 412 of theperception component 406 classifies the identified point cloud object as a particular object type. The object type may be associated with a model that includes indications of particular expected features for point cloud objects classified as the particular object type. As one example, a car object type has wheels that are expected to be round, a license plate that is expected to be rectangular, and perhaps a manufacturer logo that has a known shape. As another example, a streetlight object may be expected to have round light bulbs. As yet another example, a stop sign object is expected to be octagonal. Accordingly, block 808 includes analyzing the point cloud object identified atblock 806 to detect the region in the point cloud object associated with the feature that has the expected shape. - At
block 810, the expected shape of the feature is compared to the shape of the feature of the point cloud object as represented in the point cloud frame obtained atblock 804. The expected shape of the feature may be scaled and/or rotated based on depth data associated with the point cloud object. Said another way, the expected shape of the feature is adjusted to account for how the feature is oriented in the point cloud frame. Accordingly, block 810 may include comparing the adjusted expected feature shape to the region on the point cloud object associated with the feature. Alternatively, the feature of the point cloud object may be scaled and/or rotated to match, as closely as possible, a fixed size and orientation corresponding to the expected feature shape. - At
block 812, the point cloud object identified atblock 806 is determined to be distorted based on the expected feature shape not matching the feature of the point cloud object corresponding to the feature. For example, the feature of the point cloud object may be skewed with respect to the expected feature shape due to rolling shutter distortion. Accordingly, in some implementations, in addition to determining that point cloud object is distorted, block 812 includes determining a skew factor indicative of an amount of skew introduced by each scan line and/or component pixel thereof. The skew factor can be a three-dimensional vector representative of the direction via which the skew is introduced. To determine the skew factor, block 812 may include determining an amount by which the point cloud feature is skewed with respect to the expected feature shape at two different pixels that overlap the point cloud feature. The difference in skews with respect to the expected configuration is then divided by a number of pixels and/or scan lines between when the pixels were captured to determine a skew factor indicative of an amount and direction of skew introduced by each pixel and/or scan line. Accordingly, in some implementations, block 812 includes correcting for the detected skew by utilizing the skew factor to compensate for the amount of skew introduced by each successive scan line and/or pixel. - In some embodiments, the
method 800 includes one or more additional blocks not shown inFIG. 11 . For example, in some embodiments, the distortion of the point cloud object and/or the skew factor determined atblock 812 is analyzed to determine the relative velocity of the point cloud object, for example, using themethod 700 ofFIG. 8 . As yet another example, the point cloud object, for example by theprediction component 420 ofFIG. 5 , to predict a future state of the vehicle environment. This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state. - Referring now to the
method 820 ofFIG. 12 , sensor data is received atblock 822. The sensor data includes a first set of sensor data captured by a first set of one or more sensors configured to sense the vehicle by sequentially advancing through a plurality of points in a scan pattern during a frame, such as lidar depth data sensed by thelidar system 100 ofFIGS. 1 and 2 , or thelidar system 302 ofFIG. 3A . The sensor data also includes a second set of sensor data captured by a second set of one or more sensors that are configured to sense the vehicle environment by capturing the entirety of a frame at a single time, such as camera data captured by a RGB or stereo camera or a thermal imaging sensor. - At
block 824, the first set of sensor data is analyzed to obtain a point cloud representation of an object within the vehicle environment. For example, the object may be a vehicle, a road, a street sign, a person, a tree, etc. Accordingly, each of the first and second sets of sensor may include a different digital representation of the same object. In some embodiments, the first set of sensor data received atblock 822 includes the point cloud frame (e.g., a laser controller or sensor head controller stitched depth data together to form the point cloud). In these embodiments, block 824 includes extracting the point cloud frame from rest of the first set of sensor data received atblock 822. In other embodiments, block 824 includes analyzing the first set of sensor data received atblock 802 to generate the point cloud frame, as discussed in connection toFIG. 3 , for example. In any event, a perception component, such as theperception component 406 ofFIG. 5 , analyzes the point cloud frame to identify the object. For example, atracking module 414 of theperception component 414 may compare tracked objects to the obtained point cloud frame. Accordingly, thetracking component 414 may identify the point cloud representation of the object in the point cloud frame. - At
block 826, the second set of sensor data is analyzed to obtain a second digital representation of the object. For example, if the second set of sensor is captured by a RGB camera, the second representation of the object may be a subset of the RGB data that corresponds to the object. For instance, theperception component 406 may associate the first and second sets of sensor data with one another. That is, theperception component 406 may analyze the viewing angles at which the first and second sets of sensors sensed the first and second sets of sensor data to identify regions in each set of sensor data representative of the same portion of the vehicle environment, for example, the object analyzed atblock 824. In some implementations, thetracking module 414 associates the point cloud object analyzed atblock 826 with RGB data indicative of the object. Thus, thetracking module 414 may utilizes the associated RGB data to identify the object within the second set of sensor data. In other implementations, theperception component 406 includes separate segmentation, classification, and/or tracking modules optimized to analyze RGB data. In these implementations, the point cloud representation of the object and the RGB data representation of the object are tracked separately. - At
block 828, the point cloud object and the representation of the object in the second set of sensor data are compared to each other. More particularly, the depth data associated with the first set of sensor data may be “flattened” to obtain a two-dimensional representation of the point cloud object. The shape of the flattened point cloud object may be compared to the shape of the object as represented in the second set of sensor data. In some implementations, the shape of the object as represented in the second set of sensor data is adjusted to account for the different viewing angles associated with the first and second set of sensors. - At
block 830, the point cloud representation of the object is identified as distorted based on the comparison atblock 828. As described above, because a frame based on the second set of sensor data is captured at a single time, there is no rolling shutter distortion that impacts the shape of the object as represented in the second set of sensor data. Said another way, the second set of sensor data indicates the “true” shape of the object. Thus, when the comparison atblock 828 indicates that the shape of the point cloud representation of the object and the shape of the object as represented by the second set of sensor data do not match, it is determined that the point cloud object (i.e., the representation of the object, as opposed to the object itself) is distorted. - In some embodiments, the
method 820 includes one or more additional blocks not shown inFIG. 12 . For example, one additional block may include determining a skew factor indicative of an amount of skew introduced by each scan line and/or component pixel thereof. The skew factor can be a three-dimensional vector representative of the direction via which the skew is introduced. To determine the skew factor, this additional block may include determining an amount of the point cloud feature is skewed with respect to an expected configuration as indicated by the second set of sensor data at two different pixels that overlap the point cloud feature based on the comparison atblock 830. The difference in skew with respect to the second set of sensor data is then divided by a number of pixels and/or scan lines between when the pixels were captured to determine the skew factor. Accordingly, in some implementations, this additional block includes correcting for the detected skew by utilizing the skew factor to compensate for the amount of skew introduced by each successive scan line and/or pixel - As another example, in some embodiments, the distortion of the point cloud object and/or the skew factor determined at the above additional block is analyzed to determine the relative velocity of the point cloud object, for example, using the
method 700 ofFIG. 8 . As yet another example, the point cloud object, for example by theprediction component 420 ofFIG. 5 , to predict a future state of the vehicle environment. This additional block may further include generating one or more control signals (e.g., operational parameters) that control operation of the vehicle and/or the particular components thereof based on the predicted future state. - Although this disclosure describes example embodiments where a lidar system is incorporated into a vehicle, the embodiments described herein may also be applied to stationary lidar systems. For example, a lidar system may be affixed to a stationary object (e.g., a light pole, a utility pole, a street sign, a traffic light, or a building). Any rolling shutter distortion of an object may be attributed to movement of that object since the stationary lidar system would not contribute to the rolling shutter distortion. As a result, a stationary lidar system may be configured to determine the absolute velocity of objects rather than a relative velocity.
- In some cases, a computing device may be used to implement various modules, circuits, systems, methods, or algorithm steps disclosed herein. As an example, all or part of a module, circuit, system, method, or algorithm disclosed herein may be implemented or performed by a general-purpose single- or multi-chip processor, a digital signal processor (DSP), an ASIC, a FPGA, any other suitable programmable-logic device, discrete gate or transistor logic, discrete hardware components, or any suitable combination thereof. A general-purpose processor may be a microprocessor, or, any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- In particular embodiments, one or more implementations of the subject matter described herein may be implemented as one or more computer programs (e.g., one or more modules of computer-program instructions encoded or stored on a computer-readable non-transitory storage medium). As an example, the steps of a method or algorithm disclosed herein may be implemented in a processor-executable software module which may reside on a computer-readable non-transitory storage medium. In particular embodiments, a computer-readable non-transitory storage medium may include any suitable storage medium that may be used to store or transfer computer software and that may be accessed by a computer system. Herein, a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs (e.g., compact discs (CDs), CD-ROM, digital versatile discs (DVDs), blue-ray discs, or laser discs), optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, flash memories, solid-state drives (SSDs), RAM, RAM-drives, ROM, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate. A computer-readable non-transitory storage medium may be volatile, non-volatile, or a combination of volatile and non-volatile, where appropriate.
- In some cases, certain features described herein in the context of separate implementations may also be combined and implemented in a single implementation.
- Conversely, various features that are described in the context of a single implementation may also be implemented in multiple implementations separately or in any suitable sub-combination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination may in some cases be excised from the combination, and the claimed combination may be directed to a sub-combination or variation of a sub-combination.
- While operations may be depicted in the drawings as occurring in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all operations be performed. Further, the drawings may schematically depict one more example processes or methods in the form of a flow diagram or a sequence diagram. However, other operations that are not depicted may be incorporated in the example processes or methods that are schematically illustrated. For example, one or more additional operations may be performed before, after, simultaneously with, or between any of the illustrated operations. Moreover, one or more operations depicted in a diagram may be repeated, where appropriate. Additionally, operations depicted in a diagram may be performed in any suitable order. Furthermore, although particular components, devices, or systems are described herein as carrying out particular operations, any suitable combination of any suitable components, devices, or systems may be used to carry out any suitable operation or combination of operations. In certain circumstances, multitasking or parallel processing operations may be performed. Moreover, the separation of various system components in the implementations described herein should not be understood as requiring such separation in all implementations, and it should be understood that the described program components and systems may be integrated together in a single software product or packaged into multiple software products.
- Various implementations have been described in connection with the accompanying drawings. However, it should be understood that the figures may not necessarily be drawn to scale. As an example, distances or angles depicted in the figures are illustrative and may not necessarily bear an exact relationship to actual dimensions or layout of the devices illustrated.
- The scope of this disclosure encompasses all changes, substitutions, variations, alterations, and modifications to the example embodiments described or illustrated herein that a person having ordinary skill in the art would comprehend. The scope of this disclosure is not limited to the example embodiments described or illustrated herein. Moreover, although this disclosure describes or illustrates respective embodiments herein as including particular components, elements, functions, operations, or steps, any of these embodiments may include any combination or permutation of any of the components, elements, functions, operations, or steps described or illustrated anywhere herein that a person having ordinary skill in the art would comprehend.
- The term “or” as used herein is to be interpreted as an inclusive or meaning any one or any combination, unless expressly indicated otherwise or indicated otherwise by context. Therefore, herein, the expression “A or B” means “A, B, or both A and B.” As another example, herein, “A, B or C” means at least one of the following: A; B; C; A and B; A and C; B and C; A, B and C. An exception to this definition will occur if a combination of elements, devices, steps, or operations is in some way inherently mutually exclusive.
- As used herein, words of approximation such as, without limitation, “approximately, “substantially,” or “about” refer to a condition that when so modified is understood to not necessarily be absolute or perfect but would be considered close enough to those of ordinary skill in the art to warrant designating the condition as being present. The extent to which the description may vary will depend on how great a change can be instituted and still have one of ordinary skill in the art recognize the modified feature as having the required characteristics or capabilities of the unmodified feature. In general, but subject to the preceding discussion, a numerical value herein that is modified by a word of approximation such as “approximately” may vary from the stated value by ±0.5%, ±1%, ±2%, ±3%, ±4%, ±5%, ±10%, ±12%, or ±15%.
- As used herein, the terms “first,” “second,” “third,” etc. may be used as labels for nouns that they precede, and these terms may not necessarily imply a particular ordering (e.g., a particular spatial, temporal, or logical ordering). As an example, a system may be described as determining a “first result” and a “second result,” and the terms “first” and “second” may not necessarily imply that the first result is determined before the second result.
- As used herein, the terms “based on” and “based at least in part on” may be used to describe or present one or more factors that affect a determination, and these terms may not exclude additional factors that may affect a determination. A determination may be based solely on those factors which are presented or may be based at least in part on those factors. The phrase “determine A based on B” indicates that B is a factor that affects the determination of A. In some instances, other factors may also contribute to the determination of A. In other instances, A may be determined based solely on B.
Claims (27)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/196,650 US10551485B1 (en) | 2018-08-06 | 2018-11-20 | Fitting points to a surface |
PCT/US2019/045240 WO2020033365A1 (en) | 2018-08-06 | 2019-08-06 | Determining relative velocity based on an expected configuration |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201862715030P | 2018-08-06 | 2018-08-06 | |
US16/196,650 US10551485B1 (en) | 2018-08-06 | 2018-11-20 | Fitting points to a surface |
Publications (2)
Publication Number | Publication Date |
---|---|
US10551485B1 US10551485B1 (en) | 2020-02-04 |
US20200041619A1 true US20200041619A1 (en) | 2020-02-06 |
Family
ID=67106255
Family Applications (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/196,641 Active US10345437B1 (en) | 2018-08-06 | 2018-11-20 | Detecting distortion using other sensors |
US16/196,632 Active US10677900B2 (en) | 2018-08-06 | 2018-11-20 | Detecting distortion using known shapes |
US16/196,650 Active US10551485B1 (en) | 2018-08-06 | 2018-11-20 | Fitting points to a surface |
US16/196,597 Active 2041-06-24 US11435479B2 (en) | 2018-08-06 | 2018-11-20 | Determining relative velocity based on an expected configuration |
US16/196,630 Active 2039-04-17 US10809364B2 (en) | 2018-08-06 | 2018-11-20 | Determining relative velocity using co-located pixels |
US16/196,618 Active US10539665B1 (en) | 2018-08-06 | 2018-11-20 | Determining distortion by tracking objects across successive frames |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/196,641 Active US10345437B1 (en) | 2018-08-06 | 2018-11-20 | Detecting distortion using other sensors |
US16/196,632 Active US10677900B2 (en) | 2018-08-06 | 2018-11-20 | Detecting distortion using known shapes |
Family Applications After (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/196,597 Active 2041-06-24 US11435479B2 (en) | 2018-08-06 | 2018-11-20 | Determining relative velocity based on an expected configuration |
US16/196,630 Active 2039-04-17 US10809364B2 (en) | 2018-08-06 | 2018-11-20 | Determining relative velocity using co-located pixels |
US16/196,618 Active US10539665B1 (en) | 2018-08-06 | 2018-11-20 | Determining distortion by tracking objects across successive frames |
Country Status (2)
Country | Link |
---|---|
US (6) | US10345437B1 (en) |
WO (1) | WO2020033365A1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3919930A1 (en) * | 2020-03-17 | 2021-12-08 | Beijing Baidu Netcom Science And Technology Co. Ltd. | Method and apparatus for determining velocity of obstacle, device, medium and computer program product |
US20220135074A1 (en) * | 2020-11-02 | 2022-05-05 | Waymo Llc | Classification of objects based on motion patterns for autonomous vehicle applications |
GB2602181A (en) * | 2020-07-22 | 2022-06-22 | Nuctech Co Ltd | Method and device for correcting a scanned image and image scanning system |
WO2022076242A3 (en) * | 2020-10-07 | 2022-06-23 | Luminar, Llc | Velocity determination with a scanned lidar system |
US11378396B2 (en) * | 2009-02-20 | 2022-07-05 | Aeva, Inc. | System and method for generating motion-stabilized images of a target using lidar and video measurements |
US12050267B2 (en) | 2020-11-09 | 2024-07-30 | Waymo Llc | Doppler-assisted object mapping for autonomous vehicle applications |
Families Citing this family (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11353559B2 (en) | 2017-10-09 | 2022-06-07 | Luminar, Llc | Adjustable scan patterns for lidar system |
US11415675B2 (en) | 2017-10-09 | 2022-08-16 | Luminar, Llc | Lidar system with adjustable pulse period |
US11035943B2 (en) * | 2018-07-19 | 2021-06-15 | Aptiv Technologies Limited | Radar based tracking of slow moving objects |
US10914813B2 (en) * | 2018-08-21 | 2021-02-09 | Aptiv Technologies Limited | Classifying potentially stationary objects tracked by radar |
US11520331B2 (en) * | 2018-12-28 | 2022-12-06 | Intel Corporation | Methods and apparatus to update autonomous vehicle perspectives |
US10829114B2 (en) * | 2019-02-06 | 2020-11-10 | Ford Global Technologies, Llc | Vehicle target tracking |
US10755571B1 (en) * | 2019-03-01 | 2020-08-25 | Amazon Technologies, Inc. | Identifying parking location using single camera reverse projection |
WO2020196536A1 (en) * | 2019-03-26 | 2020-10-01 | 株式会社小糸製作所 | Photographing system and image processing device |
US11016496B2 (en) * | 2019-04-10 | 2021-05-25 | Argo AI, LLC | Transferring synthetic LiDAR system data to real world domain for autonomous vehicle training applications |
US20220221558A1 (en) * | 2019-05-31 | 2022-07-14 | Volkswagen Aktiengesellschaft | System, method, and components utilizing live point cloud compression |
US11181640B2 (en) * | 2019-06-21 | 2021-11-23 | Blackmore Sensors & Analytics, Llc | Method and system for vehicle odometry using coherent range doppler optical sensors |
JP7316442B2 (en) * | 2019-07-31 | 2023-07-27 | ボルボトラックコーポレーション | Method, control system, vehicle, and computer program product for forming a vehicle travel path |
US11613201B2 (en) * | 2019-08-12 | 2023-03-28 | Nvidia Corporation | Automatic high beam control for autonomous machine applications |
US11144769B2 (en) * | 2019-09-30 | 2021-10-12 | Pony Ai Inc. | Variable resolution sensors |
CN110988848B (en) * | 2019-12-23 | 2022-04-26 | 潍柴动力股份有限公司 | Vehicle-mounted laser radar relative pose monitoring method and device |
EP4096901A4 (en) * | 2020-01-31 | 2023-10-11 | Hewlett-Packard Development Company, L.P. | Model prediction |
CN111461309B (en) * | 2020-04-17 | 2022-05-17 | 支付宝(杭州)信息技术有限公司 | Method and device for updating reinforcement learning system for realizing privacy protection |
WO2021231559A1 (en) | 2020-05-13 | 2021-11-18 | Luminar, Llc | Lidar system with high-resolution scan pattern |
KR20220041485A (en) * | 2020-09-25 | 2022-04-01 | 현대자동차주식회사 | Method and apparatus for tracking an object using LIDAR sensor, vehicle including the apparatus, and recording medium for recording program performing the method |
US20220126853A1 (en) * | 2020-10-26 | 2022-04-28 | GM Global Technology Operations LLC | Methods and systems for stiching of images into a virtual image |
NO20210151A1 (en) * | 2021-02-05 | 2022-08-08 | Toerring Invest As | High Speed Hybrid Scanning |
US11581697B2 (en) | 2021-03-10 | 2023-02-14 | Allegro Microsystems, Llc | Detector system comparing pixel response with photonic energy decay |
US11802945B2 (en) | 2021-03-10 | 2023-10-31 | Allegro Microsystems, Llc | Photonic ROIC having safety features |
WO2022217522A1 (en) * | 2021-04-14 | 2022-10-20 | 深圳市大疆创新科技有限公司 | Target sensing method and device, detection system, movable platform and storage medium |
US11815406B2 (en) | 2021-04-14 | 2023-11-14 | Allegro Microsystems, Llc | Temperature sensing of an array from temperature dependent properties of a PN junction |
US11770632B2 (en) | 2021-04-14 | 2023-09-26 | Allegro Microsystems, Llc | Determining a temperature of a pixel array by measuring voltage of a pixel |
US11601733B2 (en) | 2021-04-14 | 2023-03-07 | Allegro Microsystems, Llc | Temperature sensing of a photodetector array |
US11568634B2 (en) * | 2021-04-28 | 2023-01-31 | Intuit Inc. | Machine learning pipeline for document image quality detection and correction |
US11619914B2 (en) | 2021-06-03 | 2023-04-04 | Allegro Microsystems, Llc | Arrayed time to digital converter |
US11252359B1 (en) | 2021-06-21 | 2022-02-15 | Allegro Microsystems, Llc | Image compensation for sensor array having bad pixels |
US12032099B2 (en) * | 2021-07-12 | 2024-07-09 | Cyngn, Inc. | Adaptive motion compensation of perception channels |
US11600654B2 (en) | 2021-07-15 | 2023-03-07 | Allegro Microsystems, Llc | Detector array yield recovery |
US11885646B2 (en) | 2021-08-12 | 2024-01-30 | Allegro Microsystems, Llc | Programmable active pixel test injection |
US11585910B1 (en) | 2021-08-13 | 2023-02-21 | Allegro Microsystems, Llc | Non-uniformity correction of photodetector arrays |
US11659261B2 (en) | 2021-08-19 | 2023-05-23 | Ford Global Technologies, Llc | Vehicle optical sensor motion correction |
CN114387305A (en) * | 2021-12-30 | 2022-04-22 | 驭势科技(北京)有限公司 | Trailer queue tracking method and device, computer equipment and storage medium |
US11933669B2 (en) | 2022-02-16 | 2024-03-19 | Allegro Microsystems, Llc | Optical system for improved reliability and performance |
US20230288566A1 (en) * | 2022-03-09 | 2023-09-14 | Silc Technologies, Inc. | Adjusting imaging system data in response to edge effects |
US12049201B2 (en) | 2022-03-09 | 2024-07-30 | Ford Global Technologies, Llc | Vehicle sensor cleaning system |
US11722141B1 (en) | 2022-04-22 | 2023-08-08 | Allegro Microsystems, Llc | Delay-locked-loop timing error mitigation |
US12021107B2 (en) | 2022-06-03 | 2024-06-25 | Lg Innotek Co., Ltd. | Systems and methods for breakdown voltage correction in Geiger-mode avalanche photodiode (APD) focal plane arrays (FPA) |
DE102023100565A1 (en) * | 2023-01-12 | 2024-07-18 | Bayerische Motoren Werke Aktiengesellschaft | Method and device for operating multiple lidar sensors |
FR3146523A1 (en) * | 2023-03-08 | 2024-09-13 | Continental Autonomous Mobility Germany GmbH | Method of estimating the speed of a vehicle |
Family Cites Families (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10336638A1 (en) * | 2003-07-25 | 2005-02-10 | Robert Bosch Gmbh | Apparatus for classifying at least one object in a vehicle environment |
GB0411097D0 (en) | 2004-05-19 | 2004-06-23 | Qinetiq Ltd | Laser radar device and method |
US8098367B2 (en) | 2008-12-19 | 2012-01-17 | Trimble Navigation Limited | Laser slope adjustment |
US8717545B2 (en) * | 2009-02-20 | 2014-05-06 | Digital Signal Corporation | System and method for generating three dimensional images using lidar and video measurements |
WO2011127375A1 (en) | 2010-04-09 | 2011-10-13 | Pochiraju Kishore V | Adaptive mechanism control and scanner positioning for improved three-dimensional laser scanning |
US20130300740A1 (en) * | 2010-09-13 | 2013-11-14 | Alt Software (Us) Llc | System and Method for Displaying Data Having Spatial Coordinates |
US9424371B2 (en) | 2010-11-05 | 2016-08-23 | Autodesk, Inc. | Click to accept as built modeling |
US20120173185A1 (en) | 2010-12-30 | 2012-07-05 | Caterpillar Inc. | Systems and methods for evaluating range sensor calibration data |
GB201116959D0 (en) | 2011-09-30 | 2011-11-16 | Bae Systems Plc | Vehicle localisation with 2d laser scanner and 3d prior scans |
US9341740B1 (en) | 2012-02-13 | 2016-05-17 | See Scan, Inc. | Optical ground tracking apparatus, systems, and methods |
US9129211B2 (en) | 2012-03-15 | 2015-09-08 | GM Global Technology Operations LLC | Bayesian network to track objects using scan points using multiple LiDAR sensors |
US9329269B2 (en) | 2012-03-15 | 2016-05-03 | GM Global Technology Operations LLC | Method for registration of range images from multiple LiDARS |
US9128185B2 (en) * | 2012-03-15 | 2015-09-08 | GM Global Technology Operations LLC | Methods and apparatus of fusing radar/camera object data and LiDAR scan points |
US9448636B2 (en) | 2012-04-18 | 2016-09-20 | Arb Labs Inc. | Identifying gestures using gesture data compressed by PCA, principal joint variable analysis, and compressed feature matrices |
US9052721B1 (en) * | 2012-08-28 | 2015-06-09 | Google Inc. | Method for correcting alignment of vehicle mounted laser scans with an elevation map for obstacle detection |
US8948497B2 (en) | 2012-09-04 | 2015-02-03 | Digital Signal Corporation | System and method for increasing resolution of images obtained from a three-dimensional measurement system |
US9043069B1 (en) | 2012-11-07 | 2015-05-26 | Google Inc. | Methods and systems for scan matching approaches for vehicle heading estimation |
US20140313321A1 (en) | 2013-02-13 | 2014-10-23 | SeeScan, Inc. | Optical ground tracking apparatus, systems, and methods |
US9110163B2 (en) | 2013-06-14 | 2015-08-18 | Microsoft Technology Licensing, Llc | Lidar-based classification of object movement |
US8989944B1 (en) | 2013-11-26 | 2015-03-24 | Google Inc. | Methods and devices for determining movements of an object in an environment |
US9720072B2 (en) * | 2014-08-28 | 2017-08-01 | Waymo Llc | Methods and systems for vehicle radar coordination and interference reduction |
US9625582B2 (en) * | 2015-03-25 | 2017-04-18 | Google Inc. | Vehicle with multiple light detection and ranging devices (LIDARs) |
US9760996B2 (en) | 2015-08-11 | 2017-09-12 | Nokia Technologies Oy | Non-rigid registration for large-scale space-time 3D point cloud alignment |
US20170046865A1 (en) * | 2015-08-14 | 2017-02-16 | Lucasfilm Entertainment Company Ltd. | Animation motion capture using three-dimensional scanner data |
US9916703B2 (en) | 2015-11-04 | 2018-03-13 | Zoox, Inc. | Calibration for autonomous vehicle operation |
US20170263127A1 (en) | 2016-03-08 | 2017-09-14 | GM Global Technology Operations LLC | Vehicle collision system and method of using the same |
US10504241B2 (en) | 2016-12-19 | 2019-12-10 | Magna Electronics Inc. | Vehicle camera calibration system |
CN117310741A (en) | 2017-01-03 | 2023-12-29 | 应诺维思科技有限公司 | Lidar system and method for detecting and classifying objects |
US20180197139A1 (en) | 2017-01-06 | 2018-07-12 | Position Imaging, Inc. | Package delivery sharing systems and methods |
EP3545337B1 (en) | 2017-01-26 | 2024-07-03 | Mobileye Vision Technologies Ltd. | Vehicle navigation based on aligned image and lidar information |
US10401866B2 (en) * | 2017-05-03 | 2019-09-03 | GM Global Technology Operations LLC | Methods and systems for lidar point cloud anomalies |
US20180349522A1 (en) | 2017-06-05 | 2018-12-06 | Siteaware Systems Ltd. | Adaptive Modeling of Buildings During Construction |
US10670725B2 (en) | 2017-07-25 | 2020-06-02 | Waymo Llc | Determining yaw error from map data, lasers, and cameras |
US10988143B2 (en) | 2017-08-25 | 2021-04-27 | International Business Machines Corporation | Using cloud-based traffic policies to alleviate issues with cross geographic traffic in autonomous vehicles |
US10163017B2 (en) | 2017-09-01 | 2018-12-25 | GM Global Technology Operations LLC | Systems and methods for vehicle signal light detection |
US10788585B2 (en) | 2017-09-15 | 2020-09-29 | Toyota Research Institute, Inc. | System and method for object detection using a probabilistic observation model |
CA3075736A1 (en) * | 2017-09-15 | 2019-11-14 | Aeye, Inc. | Intelligent ladar system with low latency motion planning updates |
US10841496B2 (en) | 2017-10-19 | 2020-11-17 | DeepMap Inc. | Lidar to camera calibration based on edge detection |
US20180074200A1 (en) | 2017-11-21 | 2018-03-15 | GM Global Technology Operations LLC | Systems and methods for determining the velocity of lidar points |
US11270162B2 (en) * | 2018-10-30 | 2022-03-08 | Here Global B.V. | Method and apparatus for detecting objects of interest in an environment |
-
2018
- 2018-11-20 US US16/196,641 patent/US10345437B1/en active Active
- 2018-11-20 US US16/196,632 patent/US10677900B2/en active Active
- 2018-11-20 US US16/196,650 patent/US10551485B1/en active Active
- 2018-11-20 US US16/196,597 patent/US11435479B2/en active Active
- 2018-11-20 US US16/196,630 patent/US10809364B2/en active Active
- 2018-11-20 US US16/196,618 patent/US10539665B1/en active Active
-
2019
- 2019-08-06 WO PCT/US2019/045240 patent/WO2020033365A1/en active Application Filing
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11378396B2 (en) * | 2009-02-20 | 2022-07-05 | Aeva, Inc. | System and method for generating motion-stabilized images of a target using lidar and video measurements |
US20230168085A1 (en) * | 2009-02-20 | 2023-06-01 | Aeva, Inc, | System and Method for Generating Motion-Stabilized Images of a Target Using Lidar and Video Measurements |
US12025441B2 (en) * | 2009-02-20 | 2024-07-02 | Aeva, Inc. | System and method for generating motion-stabilized images of a target using lidar and video measurements |
EP3919930A1 (en) * | 2020-03-17 | 2021-12-08 | Beijing Baidu Netcom Science And Technology Co. Ltd. | Method and apparatus for determining velocity of obstacle, device, medium and computer program product |
US11353476B2 (en) | 2020-03-17 | 2022-06-07 | Apollo Intelligent Driving Technology (Beijing) Co., Ltd. | Method and apparatus for determining velocity of obstacle, device and medium |
GB2602181A (en) * | 2020-07-22 | 2022-06-22 | Nuctech Co Ltd | Method and device for correcting a scanned image and image scanning system |
US11812002B2 (en) | 2020-07-22 | 2023-11-07 | Nuctech Company Limited | Method and device for correcting a scanned image and image scanning system |
GB2602181B (en) * | 2020-07-22 | 2024-04-10 | Nuctech Co Ltd | Method and device for correcting a scanned image and image scanning system |
WO2022076242A3 (en) * | 2020-10-07 | 2022-06-23 | Luminar, Llc | Velocity determination with a scanned lidar system |
US20220135074A1 (en) * | 2020-11-02 | 2022-05-05 | Waymo Llc | Classification of objects based on motion patterns for autonomous vehicle applications |
US12050267B2 (en) | 2020-11-09 | 2024-07-30 | Waymo Llc | Doppler-assisted object mapping for autonomous vehicle applications |
Also Published As
Publication number | Publication date |
---|---|
US10539665B1 (en) | 2020-01-21 |
US10551485B1 (en) | 2020-02-04 |
US11435479B2 (en) | 2022-09-06 |
US20200041648A1 (en) | 2020-02-06 |
US10677900B2 (en) | 2020-06-09 |
US20200043146A1 (en) | 2020-02-06 |
US10809364B2 (en) | 2020-10-20 |
US20200043176A1 (en) | 2020-02-06 |
WO2020033365A1 (en) | 2020-02-13 |
US10345437B1 (en) | 2019-07-09 |
US20200041647A1 (en) | 2020-02-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10551485B1 (en) | Fitting points to a surface | |
US10338223B1 (en) | Processing point clouds of vehicle sensors having variable scan line distributions using two-dimensional interpolation and distance thresholding | |
US10394243B1 (en) | Autonomous vehicle technology for facilitating operation according to motion primitives | |
US10481605B1 (en) | Autonomous vehicle technology for facilitating safe stopping according to separate paths | |
US10275689B1 (en) | Object identification and labeling tool for training autonomous vehicle controllers | |
US10345447B1 (en) | Dynamic vision sensor to direct lidar scanning | |
US20190387216A1 (en) | Post-processing by lidar system guided by camera information | |
US20220107414A1 (en) | Velocity determination with a scanned lidar system | |
US20200097010A1 (en) | Autonomous vehicle technology for facilitating safe stopping according to hybrid paths | |
US11391842B2 (en) | Adaptive scan pattern with virtual horizon estimation | |
US20230152458A1 (en) | Lidar System with Gyroscope-Aided Focus Steering | |
US20230152466A1 (en) | Lidar System with Scene Dependent Focus Intensity |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
AS | Assignment |
Owner name: LUMINAR TECHNOLOGIES, INC., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MAILA, TOMI P.;REEL/FRAME:049028/0583 Effective date: 20190425 Owner name: LUMINAR TECHNOLOGIES, INC., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MAHESHWARI, PRANAV;ENGLARD, BENJAMIN;SIGNING DATES FROM 20190111 TO 20190114;REEL/FRAME:049028/0566 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: LUMINAR, LLC, FLORIDA Free format text: CHANGE OF NAME;ASSIGNOR:LUMINAR HOLDCO, LLC;REEL/FRAME:060755/0975 Effective date: 20210330 Owner name: LAZR, INC., FLORIDA Free format text: CHANGE OF NAME;ASSIGNOR:LUMINAR TECHNOLOGIES, INC.;REEL/FRAME:060755/0927 Effective date: 20201202 Owner name: LUMINAR HOLDCO, LLC, FLORIDA Free format text: MERGER;ASSIGNOR:LAZR, INC.;REEL/FRAME:060560/0896 Effective date: 20201202 |
|
AS | Assignment |
Owner name: LUMINAR TECHNOLOGIES, INC., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LUMINAR, LLC;REEL/FRAME:064371/0452 Effective date: 20230710 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |