US12372651B2 - Retrofit light detection and ranging (LIDAR)-based vehicle system to operate with vision-based sensor data - Google Patents
Retrofit light detection and ranging (LIDAR)-based vehicle system to operate with vision-based sensor dataInfo
- Publication number
- US12372651B2 US12372651B2 US17/830,582 US202217830582A US12372651B2 US 12372651 B2 US12372651 B2 US 12372651B2 US 202217830582 A US202217830582 A US 202217830582A US 12372651 B2 US12372651 B2 US 12372651B2
- Authority
- US
- United States
- Prior art keywords
- lidar
- data
- sensor
- vehicle
- sensor data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W60/00—Drive control systems specially adapted for autonomous road vehicles
- B60W60/001—Planning or execution of driving tasks
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/86—Combinations of lidar systems with systems other than lidar, radar or sonar, e.g. with direction finders
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/931—Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0475—Generative networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/094—Adversarial learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/778—Active pattern-learning, e.g. online learning of image or video features
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/40—Photo, light or radio wave sensitive means, e.g. infrared sensors
- B60W2420/403—Image sensing, e.g. optical camera
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/40—Photo, light or radio wave sensitive means, e.g. infrared sensors
- B60W2420/408—Radar; Laser, e.g. lidar
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2556/00—Input parameters relating to data
- B60W2556/45—External transmission of data to or from the vehicle
Definitions
- the present disclosure relates generally to autonomous vehicles, and more particularly, to retrofitting a light detection and ranging (LIDAR)-based vehicle computing system (e.g., for autonomous driving) to operate with vision-based sensor data.
- LIDAR light detection and ranging
- FIG. 1 provides an illustration of an exemplary autonomous driving scenario in which an autonomous vehicle (AV) having a light detection and ranging (LIDAR)-based computing system is retrofitted to make control decisions using vision-based sensor data, according to some embodiments of the present disclosure
- AV autonomous vehicle
- LIDAR light detection and ranging
- FIG. 2 provides an illustration of an exemplary implementation of a sensor data converter, according to some embodiments of the present disclosure
- FIG. 3 provides an illustration of an exemplary implementation of a sensor data converter, according to some embodiments of the present disclosure
- FIG. 6 provides an illustration of an exemplary generative adversarial network (GAN) for training a generator model to convert vision-based sensor data to LIDAR data, according to some embodiments of the present disclosure
- FIG. 7 provides an illustration of an exemplary GAN for training a generator model to convert vision-based sensor data to LIDAR data, according to some embodiments of the present disclosure
- FIG. 8 is a flow diagram illustrating a process for retrofitting a LIDAR-based vehicle computing system to operate with vision-based sensor data, according to some embodiments of the present disclosure
- FIG. 9 is a flow diagram illustrating a process for training an ML model for vision-based sensor data to LIDAR-based sensor data conversion, according to some embodiments of the present disclosure.
- LIDAR sensors may rely heavily on LIDAR sensors for perception, prediction, planning, and/or control.
- perception, prediction, planning, and/or control at an AV may use algorithms and/or ML models that are designed, developed, trained, optimized, and/or tested based on LIDAR data.
- an AV may also use other sensors such as vision camera sensors, radio detection and ranging (RADAR) sensors, and/or ultrasonic sensors to sense a surrounding environment, sensor data from these sensors may mostly be used to supplement and/or correct information extracted from the LIDAR data and not for main operations of perception, prediction, and/or planning.
- RADAR radio detection and ranging
- One approach to supporting both LIDAR-based computing and vision-based computing is to include a computing system designed, developed, trained, optimized, and/or tested for operating on LIDAR data and a separate computing system designed, developed, trained, optimized, and/or tested for operating on camera sensor data (e.g., vision-based sensor data).
- a computing system designed, developed, trained, optimized, and/or tested for operating on LIDAR data and a separate computing system designed, developed, trained, optimized, and/or tested for operating on camera sensor data (e.g., vision-based sensor data).
- camera sensor data e.g., vision-based sensor data
- building and/or maintaining both a LIDAR-based computing system and a vision-based computing system for a single vehicle can increase the cost for design, manufacture, and/or production, and thus may be undesirable.
- the point cloud may reproduce the data acquisition temporal characteristic (e.g., the same scan frequency as that particular LIDAR sensor) or improve the data acquisition temporal characteristic (e.g., a higher scan frequency, resolution or field of view than that particular LIDAR sensor).
- the generating the point cloud representative of the at least the portion of the scene may be further based on a limitation (e.g., scan range, reflectivity, behavior in weather conditions, etc.) of that particular LIDAR sensor.
- the vehicle may receive third sensor data from one or more sensors of the first sensing modality and fourth sensor data from one or more sensors of the second sensing modality.
- the vehicle may combine the third sensor data of the first sensing modality and the fourth sensor data of the second sensing modality to generate fifth sensor data (e.g., enhanced sensor data more informational or accurate than the fourth sensor data alone).
- the vehicle may determine an action for the vehicle based on the generated fifth sensor data.
- a computer-implemented system may include receiving input image (e.g., captured from vision-based sensors) and target LIDAR data associated with a geographical area.
- the input image data may include images of scenes in the geographical area while the target LIDAR data may include point cloud data representing the scene in the geographical area.
- the computer-implemented system may train an ML model using the input image data and the target LIDAR data. For example, as part of training, the computer-implemented system may process the input image data using the ML model to generate synthesized LIDAR data and update the ML model based on the synthesized LIDAR data and the target LIDAR data.
- the ML model may be a GAN model including a generator and a discriminator.
- the updating the ML model is further based on one or more criteria associated with a driving performance. For instance, as part of training the ML model, the computer-implemented system may perform at least one of perception, prediction, or planning operations associated with driving using a first driving performance and update the ML model further based on a comparison of the first driving performance to a target driving performance.
- the systems, schemes, and mechanisms described herein can advantageously enable a vehicle (e.g., AV) to utilize a processing system (e.g., an AV processing stack) designed, developed, trained, optimized, and/or tested for operations with LIDAR sensors to operate with vision-based sensors (e.g., camera sensors, video cameras).
- a processing system e.g., an AV processing stack
- LIDAR sensors e.g., LIDAR sensors
- vision-based sensors e.g., camera sensors, video cameras.
- the AV 110 may be a fully autonomous vehicle or a semi-autonomous vehicle.
- a fully autonomous vehicle may make driving decisions and drive the vehicle without human inputs.
- a semi-autonomous vehicle may make at least some driving decisions without human inputs.
- the AV 110 may be a vehicle that switches between a semi-autonomous state and a fully autonomous state and thus, the AV 110 may have attributes of both a semi-autonomous vehicle and a fully autonomous vehicle depending on the state of the vehicle.
- the AV 110 may include a sensor suite 150 and an onboard computer 160 .
- the sensor suite 150 may include a wide variety of sensors, which may broadly categorize into a computer vision (“CV”) system, localization sensors, and driving sensors.
- the sensor suite 150 may include one or more vision sensors 152 (e.g., camera sensors).
- the one or more vision sensors may capture images of the surrounding environment of the AV 110 .
- the one or more vision sensors may capture images of at least some of the trees 114 , the road sign 116 , the traffic light 117 , the buildings 118 , and the object 119 located around the roadway system 102 .
- the sensor suite 150 may include multiple vision sensors to capture different views, e.g., a front-facing camera, a back-facing camera, a wide-angle (surround) camera, and side-facing cameras.
- one or more vision sensors may be implemented using a high-resolution imager with a fixed mounting and field of view.
- One or more vision sensors may have adjustable field of views and/or adjustable zooms.
- the vision sensors may capture images continually or at some intervals during operation of the AV 110 .
- the vision sensors may transmit the captured images to the onboard computer 160 of the AV 110 for further processing, for example, to assist the AV 110 in determining certain action(s) to be carried out by the AV 110 .
- the point cloud may include data points representing at least some of the trees 114 , the road sign 116 , the traffic light 117 , the buildings 118 , and the object 119 located around the roadway system 102 .
- the one or more LIDAR sensors 154 may transmit the captured point cloud to the onboard computer 160 of the AV 110 for further processing, for example, to assist the AV 110 in determining certain action(s) to be carried out by the AV 110 .
- the perception module 142 may detect one or more of the vehicle 112 , the trees 114 , the road sign 116 , the traffic light 117 , the buildings 118 , and/or the objects 119 in the surroundings of the AV 110 .
- the perception module 142 may include one or more classifiers trained using ML to identify particular objects. For example, a multi-class classifier may be used to classify each object in the environment of the AV 110 as one of a set of potential objects, e.g., a vehicle, a pedestrian, or a cyclist. As another example, a pedestrian classifier may recognize pedestrians in the environment of the AV 110 , a vehicle classifier may recognize vehicles in the environment of the AV 110 , etc.
- the planning module 146 may determine a pathway for the AV 110 to follow.
- the planning module 146 may determine the pathway for the AV 110 based on predicted behaviors of the objects provided by the prediction module 144 and right-of-way rules that regulate behavior of vehicles, cyclists, pedestrians, or other objects.
- the pathway may include locations for the AV 110 to maneuver to, and timing and/or speed of the AV 110 in maneuvering to the locations.
- the control module 148 may send appropriate commands to instruct movement-related subsystems (e.g., actuators, steering wheel, throttle, brakes, etc.) of the AV 110 to maneuver according to the pathway determined by the planning module 146 .
- movement-related subsystems e.g., actuators, steering wheel, throttle, brakes, etc.
- the onboard computer 160 may further include a sensor data converter 130 .
- the sensor data converter 130 may be implemented using a combination of hardware and/or software components.
- the sensor data converter 130 may be a software component executed by the one or more processors of the onboard computer 160 .
- the sensor data converter 130 may convert vision-based or image data to LIDAR data (e.g., point cloud data). In this way, the same AV processing stack 140 that relies on the LIDAR data can be reused for processing the generated or synthesized LIDAR data output from the conversion.
- the sensor data converter 130 may generate the synthesized LIDAR data 122 by simulating and/or emulating characteristics of that particular certain LIDAR sensor using heuristic algorithms. In other aspects, the sensor data converter 130 may generate the synthesized LIDAR data 122 using ML (e.g., a GAN model). The synthesized LIDAR data 122 output by the sensor data converter 130 may be provided to the AV processing stack 140 . The AV processing stack 140 may determine an action (e.g., a driving decision) to be carried out by the AV 110 . The action may be associated with perception, prediction, planning, and/or control operations as discussed above.
- ML e.g., a GAN model
- the AV 110 may receive LIDAR data 124 captured by the LIDAR sensor(s) 154 in real-time. Because the AV processing stack 140 is configured for processing LIDAR data, the AV processing stack 140 may process the lived-captured LIDAR data 124 directly (without conversion). However, in some examples, it may be desirable to combine the real-time captured raw LIDAR data 124 with the synthesized LIDAR data 122 to generate enhanced LIDAR data (e.g., to provide more information or more accurate information about the surroundings of the AV 110 ).
- the enhanced LIDAR data may be provided to the AV processing stack 140 , which may then determine an action (e.g., a driving/control decision) for the AV 110 using the enhanced LIDAR data.
- an action e.g., a driving/control decision
- Mechanisms for converting or mapping vision-based sensor data or image data to LIDAR data and/or augmenting or enhancing LIDAR data will be discussed more fully below.
- FIGS. 2 - 4 are discussed in relation to FIG. 1 to illustrate various implementations for the sensor data converter 130 .
- FIG. 2 provides an illustration of an exemplary implementation of a sensor data converter 200 , according to some embodiments of the present disclosure.
- the AV 110 of FIG. 1 may implement the sensor data converter 200 in place of the sensor data converter 130 shown in FIG. 1 .
- the sensor data converter 200 can be implemented in software executed by the one or more processors of the onboard computer 160 .
- the sensor data converter 200 may receive image data 202 (e.g., from camera or vision sensor(s) 152 of the AV 110 in real-time).
- the image data 202 may include an image of a scene in a surrounding environment of the AV 110 .
- the AV 110 may have vision sensors 152 with different facings (e.g., front-facing, side-facing, rear-facing, etc.) and/or with the same facing but with a separation distance, the image data 202 can include images captured by different vision sensors 152 .
- the sensor data converter 200 may detect, from the image data 202 , one or more objects (e.g., the trees 114 , the road sign 116 , the traffic light 117 , the buildings 118 , the object 119 , etc.) in the surrounding environment of the AV 110 .
- the sensor data converter 200 may generate point cloud data representative of the detected one or more objects to provide synthesized LIDAR data 208 .
- the sensor data converter 200 may include an object detection sub-module 220 , an object library 224 , and a LIDAR sensor simulation/emulation sub-module 226 .
- the object detection sub-module 220 may process the image data 202 .
- the object detection sub-module 220 may implement any suitable object detection algorithms to accurately determine objects (e.g., traffic lights, road signs, road markings, buildings, trees, barriers, etc.) in the AV 110 's vicinity.
- the object detection sub-module 220 may implement one or more classifiers to differentiate cars from non-cars, pedestrians from non-pedestrians, or more generally identify particular object(s).
- the object detection sub-module 220 can interact with the perception module 142 in the AV processing stack 140 to detect and identify objects around the AV 110 .
- the object library 224 may include a collection of point cloud representations of various objects.
- the object library 224 may store one point cloud for each object or each type of objects, e.g., an image for a road sign, an image for a traffic light, an image for a building, an image for a tree, an image for a crosswalk, etc.
- the object library 224 may store multiple point clouds for one object or one object type, for example, including a 3D representation, a 2D representation, and/or representations of various orientations of the object or object type.
- the point clouds stored at the object library 224 may be captured using certain LIDAR sensors.
- the LIDAR sensor simulation/emulation sub-module 226 may simulate and/or emulate characteristics of a particular LIDAR sensor device (or LIDAR sensor device model) that was used to capture LIDAR data on which the design, development, training, optimization, and/or test the AV processing stack 140 was based. As shown, the LIDAR sensor simulation/emulation sub-module 226 may receive LIDAR sensor model characteristics 230 for the particular LIDAR sensor hardware and process or modify the selected point cloud (e.g., using heuristic algorithms that are based on interpretation and/or rules) so that the output synthesized LIDAR data 208 may have those characteristics 230 of the particular LIDAR sensor device.
- LIDAR sensor model characteristics 230 for the particular LIDAR sensor hardware and process or modify the selected point cloud (e.g., using heuristic algorithms that are based on interpretation and/or rules) so that the output synthesized LIDAR data 208 may have those characteristics 230 of the particular LIDAR sensor device.
- a beam characteristic may include a beam size of a laser beam emitted by the particular LIDAR sensor for the measurement.
- a vertical resolution characteristic may refer to the angular distance between the scan lines of the LIDAR sensor.
- a horizontal resolution characteristic may refer to the angular distance between each adjacent lidar point.
- a range characteristic may refer to farthest distance that the particular LIDAR sensor may detect an object. In some instances, the range can be dependent on the power of the laser source at the particular LIDAR sensor.
- a scan frequency characteristic may refer to how frequent the particular LIDAR sensor emit a light pulse or acquire measurement data in a scan cycle
- a scan angle characteristic may refer to a field of view or the angle covered by the particular LIDAR sensor (or the angle at which the light signals are emitted).
- a reflectivity characteristic may refer to an amount or an intensity of light that may be reflected from a certain target surface when using the particular LIDAR sensor.
- a blind spot characteristic may refer to an area in which the particular LIDAR sensor may failed or missed to detect.
- a behavior characteristic may include range, measurement accuracy, reflectivity, etc. of the particular LIDAR sensor when sensing under certain weather conditions. For instance, the performance of the particular LIDAR sensor may be impacted by wavelength stability and/or detector (receiver) sensitivity. As an example, the wavelength of the laser source at the particular LIDAR sensor may vary with temperatures while a poor signal-to-noise ratio (SNR) can degrade the LIDAR sensor receive.
- SNR signal-to-noise ratio
- the LIDAR sensor simulation/emulation sub-module 226 can reproduce or improve on any one or more of the characteristics 230 .
- the LIDAR sensor simulation/emulation sub-module 226 can reproduce a temporal characteristic (e.g., scan frequency) and/or a limitation (e.g., range, reflectivity, behaviors under weather conditions, blind spot, etc.) of the particular LIDAR sensor when generating the synthesized LIDAR data 208 .
- the LIDAR sensor simulation/emulation sub-module 226 can improve a temporal characteristic and/or a limitation of the particular LIDAR sensor when generating the synthesized LIDAR data 208 .
- FIG. 4 provides an illustration of an exemplary implementation of a sensor data converter 400 , according to some embodiments of the present disclosure.
- the AV 110 of FIG. 1 may implement the sensor data converter 400 in place of the sensor data converter 130 shown in FIG. 1 .
- the sensor data converter 400 can be implemented in software executed by the one or more processors of the onboard computer 160 .
- the sensor data converter 400 may be substantially similar to the sensor data converter 300 .
- the sensor data converter 400 may include an ML model 410 .
- the ML model 410 may have a substantially similar architecture as the ML model 310 .
- the ML model 410 's parameters e.g., weights and/or biases
- AV 110 may receive image data 402 captured by the vision sensor(s) 152 in real-time.
- the image data 402 may be substantially similar to the image data 120 , 202 , and/or 302 .
- the AV 110 may receive LIDAR data 404 (raw LIDAR data including point clouds) captured by the LIDAR sensor(s) 154 in real-time.
- the LIDAR data 404 may be substantially similar to the LIDAR data 124 . Both the image data 402 and the LIDAR data 404 may include information associated with a common scene in a surrounding of the AV 110 .
- the trained ML model 410 may process the image data 402 and the raw LIDAR data 404 (e.g., through each of the plurality of layers for computations using respective parameters for the layer) to generate enhanced LIDAR data 408 on the fly.
- camera sensors may capture visual data from optics in the lens while LIDAR sensors emit light pulse and use light signals reflected from objects in the surroundings to determine distances to those objects and/or attributes of those objects. Due to the different sensing modalities used by camera sensors versus LIDAR sensors, camera sensors and LIDAR sensors can have different strengths and weaknesses.
- the generator model 610 may be alternatively trained to generate enhanced LIDAR data from input image data and input LIDAR data
- the discriminator model 630 may be alternatively trained to distinguish between target enhanced LIDAR data (e.g., a real sample) and target enhanced LIDAR data (e.g., a fake sample) so that the trained generator model 610 may be used to enhance LIDAR data with vision data as discussed above with reference to FIG. 4 .
- the training/updating mechanisms for the generator model 610 and the discriminator model 630 may be substantially the same as for the sensor data conversion discussed above.
- the AV performance metric 720 may be based on latency and/or memory consumption of individual components ( 142 , 144 , 146 , 148 ). In other instances, the AV performance metric 720 may be based on latency and/or memory consumption of the end-to-end AV processing stack 140 . For example, images that lead to a very high latency, reaction time and/or memory may be penalized by including an appropriate term in the loss function 740 . Subsequently, the real/fake score 742 may be used to update the generator model 710 and the discriminator model 730 .
- the generating the second sensor data at 804 may include generating the point cloud representative of the at least the portion of the scene (captured by the image) in the image further based on a temporal characteristic (e.g., a scan frequency) of the particular LIDAR sensor.
- the generating the point cloud representative of the at least the portion of the scene may be further based on an improved temporal characteristic of the particular LIDAR sensor, the improved temporal characteristic including at least a scan frequency higher than a scan frequency of the particular LIDAR sensor.
- the generating the point cloud representative of the at least the portion of the scene may be further based on a limitation of the particular LIDAR sensor.
- the generating the second sensor data at 804 may include processing the first sensor data using an ML model to generate the second sensor data, for example, as discussed above with reference to FIG. 3 .
- the ML model is a generator model trained jointly with a discriminator model in a GAN model, for example, as discussed above with reference to FIGS. 6 - 7 .
- the process 800 may further include receiving third sensor data from the one or more sensors of the first sensing modality at the vehicle and receiving fourth sensor data from one or more sensors of the second sensing modality at the vehicle.
- the process 800 may further include generating enhanced sensor data of the second modality based on the third sensor data of the first sensing modality and the fourth sensor data of the second sensing modality.
- the process 800 may further include determining, by the vehicle controller, another action for the vehicle based on the enhanced fourth sensor data.
- FIG. 9 is a flow diagram illustrating an exemplary process 900 for training an ML model for vision-based sensor data to LIDAR-based sensor data conversion, according to some embodiments of the present disclosure.
- the process 900 can be implemented by a computed-implemented system (e.g., the computer system 1100 of FIG. 11 ).
- the process 900 may utilize similar mechanisms as discussed above with reference to FIGS. 5 - 7 . Operations are illustrated once each and in a particular order in FIG. 9 , but the operations may be performed in parallel, reordered, and/or repeated as desired.
- the computer-implement system may include memory storing instruction and one or more computer processors, where the instructions, when executed by the one or more computer processors, cause the one or more computer processors to perform the operations of the process 900 .
- the operations of the process 900 may be in the form of instructions encoded in a non-transitory computable-readable storage medium that, when executed by one or more computer processors of the computer-implemented system, cause the one or more computer processors to perform the process 900 .
- input image data associated with a geographical area may be received.
- target LIDAR data associated with the geographical area may be received.
- the input image data may correspond to the input image data 504
- the target LIDAR data may correspond to the target LIDAR data 502
- the input image data may correspond to the input image data 602
- the target LIDAR data may correspond to the target LIDAR data 604 .
- an ML model (e.g., the ML models 310 , 410 , 510 , 610 , 630 , 600 , 710 , 730 , and/or 700 ) may be trained.
- the training may include processing the input image data using the ML model to generate synthesized LIDAR data and updating the ML model based on the synthesized LIDAR data and the target LIDAR data.
- the ML model may be a GAN model including a generator model and a discriminator model, for example, as discussed above with reference to FIGS. 6 - 7 . Accordingly, the training may include processing the input image data using the generator model to generate the synthesized LIDAR data.
- the training may further include processing the synthesized LIDAR data and the target LIDAR data using the discriminator model.
- the training may further include updating at least one of the generator model or the discriminator model based on an output of the discriminator model.
- the updating the ML model may be further based on one or more criteria associated with a driving performance.
- the training the ML model may further include performing at least one of perception, prediction, or planning operations associated with driving using a first driving performance and updating the ML model further based on a comparison of the first driving performance to a target driving performance.
- FIG. 10 illustrates an exemplary AV 10 , according to some embodiments of the present disclosure.
- the AV 10 may correspond to the AV 110 of FIG. 1 .
- the AV 10 may correspond to a level four or level five automation system under the Society of Automotive Engineers (SAE) “J3016” standard taxonomy of automated driving levels.
- SAE Society of Automotive Engineers
- a level four system may indicate “high automation,” referring to a driving mode in which the automated driving system performs aspects of the dynamic driving task, even if a human driver does not respond appropriately to a request to intervene.
- a level five system may indicate “full automation,” referring to a driving mode in which the automated driving system performs aspects of the dynamic driving task under roadway and environmental conditions that can be managed by a human driver. Implementations in accordance with the present subject matter are not limited to any taxonomy or rubric of automation categories.
- systems in accordance with the present disclosure can be used in conjunction with any autonomous or other vehicle that utilizes a navigation system and/or other systems to provide route guidance.
- the AV 10 may generally include a propulsion system 20 , a transmission system 22 , a steering system 24 , a brake system 26 , a sensor system 28 , an actuator system 30 , data storage device 32 , controller 34 , and a communication system 36 .
- the propulsion system 20 can, in various embodiments, include an internal combustion engine, an electric machine such as a traction motor, and/or a fuel cell propulsion system.
- the transmission system 22 may be configured to transmit power from the propulsion system 20 to the front wheels 16 and rear wheels 18 according to selectable speed ratios.
- the transmission system 22 can include a step-ratio automatic transmission, a continuously-variable transmission, or other appropriate transmission.
- the brake system 26 may be configured to provide braking torque to the front wheels 16 and rear wheels 18 .
- Brake system 26 can, in various embodiments, include friction brakes, brake by wire, a regenerative braking system such as an electric machine, and/or other appropriate braking systems.
- the steering system 24 may influence a position of the front wheels 16 and/or rear wheels 18 . While depicted as including a steering wheel 25 for illustrative purposes, in some embodiments contemplated within the scope of the present disclosure, the steering system 24 may not include a steering wheel.
- the sensor system 28 may include one or more sensing devices 40 a - 40 n that sense observable conditions of the exterior environment and/or the interior environment of the AV 10 .
- the sensing devices 40 a - 40 n can include RADAR sensors, LIDAR sensors, GPSs, optical cameras, thermal cameras, time-of-flight (TOF) cameras, ultrasonic sensors, speedometers, compasses, and/or other sensors.
- the actuator system 30 may include one or more actuator devices 42 a - 42 n that control one or more vehicle features such as the propulsion system 20 , the transmission system 22 , the steering system 24 , and the brake system 26 .
- the AV 10 can also include interior and/or exterior vehicle features not illustrated in FIG. 10 , such as various doors, a trunk, and cabin features such as air conditioning, music players, lighting, touch-screen display components (such as those used in connection with navigation systems), and the like.
- the data storage device 32 may store data for use in automatically controlling the AV 10 .
- the data storage device 32 may store defined maps of the navigable environment.
- the defined maps may be predefined by and obtained from a remote system.
- the defined maps may be assembled by the remote system and communicated to the AV 10 (wirelessly and/or in a wired manner) and stored in the data storage device 32 .
- Route information can also be stored within the data storage device 32 —i.e., a set of road segments (associated geographically with one or more of the defined maps) that together define a route that the user might take to travel from a start location (e.g., the user's current location) to a target location.
- the data storage device 32 may store ML models 38 that are trained to facilitate autonomous driving.
- the ML models 38 may correspond to the ML models 310 and/or 410 discussed above with reference to FIGS. 3 and/or 4 , respectively.
- the data storage device 32 may include any suitable volatile or non-volatile memory technology, including double data rate (DDR) random access memory (RAM), synchronous RAM (SRAM), dynamic RAM (DRAM), flash, read-only memory (ROM), optical media, virtual memory regions, magnetic or tape memory, or any other suitable technology.
- DDR double data rate
- SRAM synchronous RAM
- DRAM dynamic RAM
- flash read-only memory
- ROM read-only memory
- any data storage devices or memory elements discussed herein should be construed as being encompassed within the broad term “memory.”
- the data storage device 32 can be part of the controller 34 , separate from the controller 34 , or part of the controller 34 and part of a separate system.
- the controller 34 may include a processor 44 and a computer-readable storage device or media 46 .
- the processor 44 can be any custom-made or commercially available processor, a central processing unit (CPU), a graphics processing unit (GPU), an auxiliary processor among several processors associated with the controller 34 , a semiconductor-based microprocessor (in the form of a microchip or chip set), any combination thereof, or generally any device for executing computer instructions.
- the computer-readable storage device or media 46 can include volatile and non-volatile storage in ROM, RAM, and keep-alive memory (KAM), for example. KAM may be a persistent or non-volatile memory that can store various operating variables while the processor 44 is powered down.
- the computer-readable storage device or media 46 can be implemented using any of a number of memory devices such as PROMs (programmable read-only memory), EPROMs (electrically PROM), EEPROMs (electrically erasable PROM), flash memory, or any other electric, magnetic, optical, resistive, or combination memory devices capable of storing data, some of which represent executable instructions, used by the controller 34 in controlling the AV 10 .
- PROMs programmable read-only memory
- EPROMs electrically PROM
- EEPROMs electrically erasable PROM
- flash memory or any other electric, magnetic, optical, resistive, or combination memory devices capable of storing data, some of which represent executable instructions, used by the controller 34 in controlling the AV 10 .
- any other electric, magnetic, optical, resistive, or combination memory devices capable of storing data, some of which represent executable instructions, used by the controller 34 in controlling the AV 10 .
- the computer-readable storage device or media 46 is depicted in FIG. 10 as part of the controller 34
- the instructions can include one or more separate programs that comprise an ordered listing of executable instructions for implementing logical functions.
- the instructions when executed by the processor 44 , can receive and process signals from the sensor system 28 , perform logic, calculations, methods and/or algorithms for automatically controlling the components of the AV 10 , and generate control signals transmitted to the actuator system 30 to control the components of the AV 10 based on the logic, calculations, methods, and/or algorithms.
- one controller 34 is shown in FIG. 10
- embodiments of the AV 10 can include any number of controllers 34 that communicate over any suitable communication medium or a combination of communication mediums and that cooperate to process the sensor signals, perform logic, calculations, methods, and/or algorithms, and generate control signals to control features of the AV 10 .
- the communication system 36 may wirelessly communicates information to and from other entities 48 , such as other vehicles (“V2V” communication), infrastructure (“V2I” communication), remote transportation systems, and/or user devices.
- the communication system 36 may be a wireless communication system configured to communicate via a wireless local area network (WLAN) using Institute of Electrical and Electronics Engineers (IEEE) 802.11 standards or by using cellular data communication (e.g., fifth-generation (5G) under the third Generation Partnership Project (3GPP)).
- 5G fifth-generation
- 3GPP third Generation Partnership Project
- DSRC channels may refer to one-way or two-way short-range to medium-range wireless communication channels specifically designed for automotive use and a corresponding set of protocols and standards.
- FIG. 11 illustrates components of a computing system 1100 used in implementations described herein.
- the components of FIG. 11 can be present in a vehicle or an AV (e.g., the AV 10 of FIG. 10 and/or the AV 110 of FIG. 1 ).
- the components of FIG. 11 can be present in an infrastructure system for AV.
- system 1100 can be implemented within one computing device or distributed across multiple computing devices or subsystems that cooperate in executing program instructions.
- the system 1100 can include one or more blade server devices, standalone server devices, personal computers, routers, hubs, switches, bridges, firewall devices, intrusion detection devices, mainframe computers, network-attached storage devices, smartphones and other mobile telephones, and other computing devices.
- the system hardware can be configured according to any suitable computer architectures such as a Symmetric Multi-Processing (SMP) architecture or a Non-Uniform Memory Access (NUMA) architecture.
- SMP Symmetric Multi-Processing
- NUMA Non-Uniform Memory Access
- the memory 1110 can include any computer-readable storage media readable by one or more processing unit(s) 1120 and that stores instructions 1112 .
- the memory 1110 can be implemented as one storage device and can also be implemented across multiple co-located or distributed storage devices or subsystems.
- the memory 1110 can include additional elements, such as a controller, that communicate with the one or more processing units 1120 .
- the memory 1110 can also include storage devices and/or subsystems on which data and/or instructions may be stored.
- System 1100 can access one or more storage resources to access information to carry out any of the processes indicated by instructions 1112 .
- the system 1100 may further include a sensor data converter 1114 and a vehicle controller 1116 , for example, when the system 1100 is part of an AV such as the AV 110 of FIG. 1 and/or the AV 10 of FIG. 10 .
- Each of the sensor data converter 1114 and the vehicle controller 1116 can include hardware and/or software components.
- the sensor data converter 1114 and the vehicle controller 1116 can be implemented as part of the one or more processing unit(s) 1120 .
- the sensor data converter 1114 may convert sensor data from one sensing modality (e.g., vision-based) to another sensing modality (e.g., LIDAR-based) as discussed herein, and the vehicle controller 1116 may be retrofitted to perform AV processing including perception, prediction, planning, and/or control as discussed herein.
- one sensing modality e.g., vision-based
- another sensing modality e.g., LIDAR-based
- the vehicle controller 1116 may be retrofitted to perform AV processing including perception, prediction, planning, and/or control as discussed herein.
- the server can use one or more communications networks that facilitate communication among the computing devices.
- the one or more communications networks can include or be a local or wide area network that facilitates communication among the computing devices.
- One or more direct communication links can be included between the computing devices.
- the computing devices can be installed at geographically distributed locations or at one geographic location, such as a server farm or an office.
- Example 2 the method of Example 1 can optionally include where the one or more sensors of the first sensing modality are vision-based sensors, and the second sensing modality is light detection and ranging (LIDAR).
- the one or more sensors of the first sensing modality are vision-based sensors
- the second sensing modality is light detection and ranging (LIDAR).
- Example 9 of any of Examples 1-8 can optionally include where the generating the point cloud representative of the at least the portion of the scene in the image is further based on an improved limitation of the particular LIDAR sensor, the improved limitation including at least one of a scan range longer than a scan range of the particular LIDAR sensor; a reflectivity higher than a reflectivity of the particular LIDAR sensor; or a visibility range under a weather condition longer than a visibility range of the particular LIDAR sensor under the weather condition.
- Example 12 the method of any of Examples 1-11 can optionally include where the machine learning model is a generator model trained jointly with a discriminator model in a generative adversarial network (GAN) model.
- GAN generative adversarial network
- Example 19 includes one or more non-transitory, computer-readable media encoded with instructions that, when executed by one or more processing units, perform a method including receiving input image data associated with a geographical area; receiving target light detection and ranging (LIDAR) data associated with the geographical area; and training a machine learning model, where the training includes processing the input image data using the machine learning model to generate synthesized LIDAR data; and updating the machine learning model based on the synthesized LIDAR data and the target LIDAR data.
- LIDAR target light detection and ranging
- Example 20 the one or more non-transitory, computer-readable media of Example 19 can optionally include where the machine learning model is a generative adversarial network (GAN) model including a generator model and a discriminator model, where the training the machine learning model includes processing the input image data using the generator model to generate the synthesized LIDAR data; processing the synthesized LIDAR data and the target LIDAR data using the discriminator model; and updating at least one of the generator model or the discriminator model based on an output of the discriminator model.
- GAN generative adversarial network
- Example 21 the one or more non-transitory, computer-readable media of any of Examples 19-20 can optionally include where the updating the machine learning model is further based on one or more criteria associated with a driving performance.
- Example 22 the one or more non-transitory, computer-readable media of any of Examples 19-21 can optionally include where the training the machine learning model further includes performing at least one of perception, prediction, or planning operations associated with driving using a first driving performance; and updating the machine learning model further based on a comparison of the first driving performance to a target driving performance.
- aspects of the present disclosure can take the form of a hardware implementation, a software implementation (including firmware, resident software, or micro-code) or an implementation combining software and hardware aspects that can generally be referred to herein as a “circuit,” “module,” “component” or “system.”
- Functions described in this disclosure can be implemented as an algorithm executed by one or more hardware processing units, e.g. one or more microprocessors of one or more computers.
- different steps and portions of the operations of the methods described herein can be performed by different processing units.
- aspects of the present disclosure can take the form of a computer program product embodied in one or more computer-readable media having computer-readable program code embodied, e.g., stored or encoded, thereon.
- such a computer program can, for example, be downloaded (or updated) to the existing devices and systems or be stored upon manufacturing of these devices and systems.
- the ‘means for’ in these instances can include (but is not limited to) using any suitable component discussed herein, along with any suitable software, circuitry, hub, computer code, logic, algorithms, hardware, controller, interface, link, bus, communication pathway, etc.
- the system includes memory that further comprises machine-readable instructions that when executed cause the system to perform any of the activities discussed above.
- storage media can refer to non-transitory storage media, such as a hard drive, a memory chip, and cache memory, and to transitory storage media, such as carrier waves or propagating signals.
- the terms “comprise,” “comprising,” “include,” “including,” “have,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion.
- a method, process, device, or system that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such method, process, device, or system.
- the term “or” refers to an inclusive or and not to an exclusive or.
- any number of electrical circuits of the FIGS. can be implemented on a board of an associated electronic device.
- the board can be a general circuit board that can hold various components of the internal electronic system of the electronic device and, further, provide connectors for other peripherals. More specifically, the board can provide the electrical connections by which the other components of the system can communicate electrically.
- Any suitable processors (inclusive of digital signal processors, microprocessors, supporting chipsets, etc.) and computer-readable, non-transitory memory elements can be coupled to the board based on particular configurations, processing demands, or computer designs.
- Other components such as external storage, additional sensors, controllers for audio/video display, and peripheral devices may be attached to the board as plug-in cards, via cables, or integrated into the board itself.
- the functionalities described herein may be implemented in emulation form as software or firmware running within one or more configurable (e.g., programmable) elements arranged in a structure that supports these functions.
- the software or firmware providing the emulation may be provided on non-transitory computer-readable storage medium comprising instructions to allow a processor to carry out those functionalities.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Remote Sensing (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- Radar, Positioning & Navigation (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computer Networks & Wireless Communication (AREA)
- Electromagnetism (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Molecular Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Human Computer Interaction (AREA)
- Automation & Control Theory (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
Claims (11)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/830,582 US12372651B2 (en) | 2022-06-02 | 2022-06-02 | Retrofit light detection and ranging (LIDAR)-based vehicle system to operate with vision-based sensor data |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/830,582 US12372651B2 (en) | 2022-06-02 | 2022-06-02 | Retrofit light detection and ranging (LIDAR)-based vehicle system to operate with vision-based sensor data |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20230393280A1 US20230393280A1 (en) | 2023-12-07 |
| US12372651B2 true US12372651B2 (en) | 2025-07-29 |
Family
ID=88977467
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/830,582 Active 2043-05-01 US12372651B2 (en) | 2022-06-02 | 2022-06-02 | Retrofit light detection and ranging (LIDAR)-based vehicle system to operate with vision-based sensor data |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US12372651B2 (en) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20250157204A1 (en) * | 2023-11-14 | 2025-05-15 | Qualcomm Incorporated | Asynchronous multimodal feature fusion |
Citations (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190012808A1 (en) * | 2017-07-06 | 2019-01-10 | GM Global Technology Operations LLC | Calibration verification methods for autonomous vehicle operations |
| US20190120948A1 (en) * | 2017-10-19 | 2019-04-25 | DeepMap Inc. | Lidar and camera synchronization |
| US20190180502A1 (en) * | 2017-12-13 | 2019-06-13 | Luminar Technologies, Inc. | Processing point clouds of vehicle sensors having variable scan line distributions using interpolation functions |
| US20190271767A1 (en) * | 2016-11-16 | 2019-09-05 | Innoviz Technologies Ltd. | Dynamically Allocating Detection Elements to Pixels in LIDAR Systems |
| US20200284883A1 (en) * | 2019-03-08 | 2020-09-10 | Osram Gmbh | Component for a lidar sensor system, lidar sensor system, lidar sensor device, method for a lidar sensor system and method for a lidar sensor device |
| US20200301799A1 (en) * | 2019-03-23 | 2020-09-24 | Uatc, Llc | Systems and Methods for Generating Synthetic Sensor Data via Machine Learning |
| US10841483B1 (en) * | 2019-07-11 | 2020-11-17 | Denso International America, Inc. | System and method for calibrating at least one camera and a light detection and ranging sensor |
| US11092690B1 (en) * | 2016-09-22 | 2021-08-17 | Apple Inc. | Predicting lidar data using machine learning |
| US20210255330A1 (en) * | 2018-09-06 | 2021-08-19 | Robert Bosch Gmbh | Method for selecting an image detail of a sensor |
| US11275673B1 (en) * | 2019-06-24 | 2022-03-15 | Zoox, Inc. | Simulated LiDAR data |
| US20220091266A1 (en) * | 2020-09-18 | 2022-03-24 | Denso International America, Inc. | Systems and methods for enhancing outputs of a lidar |
| US20220137218A1 (en) * | 2020-10-30 | 2022-05-05 | Waymo Llc | Detecting Retroreflectors in NIR Images to Control LIDAR Scan |
| US20220221585A1 (en) * | 2021-01-14 | 2022-07-14 | Argo AI, LLC | Systems and methods for monitoring lidar sensor health |
| US20230274462A1 (en) * | 2022-02-28 | 2023-08-31 | Basis Software, Inc. | System and method for camera calibration |
| US20250022143A1 (en) * | 2023-07-13 | 2025-01-16 | Waymo Llc | Object tracking across a wide range of distances for driving applications |
-
2022
- 2022-06-02 US US17/830,582 patent/US12372651B2/en active Active
Patent Citations (19)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11092690B1 (en) * | 2016-09-22 | 2021-08-17 | Apple Inc. | Predicting lidar data using machine learning |
| US20190271767A1 (en) * | 2016-11-16 | 2019-09-05 | Innoviz Technologies Ltd. | Dynamically Allocating Detection Elements to Pixels in LIDAR Systems |
| US20190012808A1 (en) * | 2017-07-06 | 2019-01-10 | GM Global Technology Operations LLC | Calibration verification methods for autonomous vehicle operations |
| US20190120948A1 (en) * | 2017-10-19 | 2019-04-25 | DeepMap Inc. | Lidar and camera synchronization |
| US20190180502A1 (en) * | 2017-12-13 | 2019-06-13 | Luminar Technologies, Inc. | Processing point clouds of vehicle sensors having variable scan line distributions using interpolation functions |
| US20190178988A1 (en) * | 2017-12-13 | 2019-06-13 | Luminar Technologies, Inc. | Training a machine learning based model of a vehicle perception component based on sensor settings |
| US20190176841A1 (en) * | 2017-12-13 | 2019-06-13 | Luminar Technologies, Inc. | Training multiple neural networks of a vehicle perception component based on sensor settings |
| US20190179024A1 (en) * | 2017-12-13 | 2019-06-13 | Luminar Technologies, Inc. | Processing point clouds of vehicle sensors having variable scan line distributions using voxel grids |
| US20190179317A1 (en) * | 2017-12-13 | 2019-06-13 | Luminar Technologies, Inc. | Controlling vehicle sensors using an attention model |
| US20210255330A1 (en) * | 2018-09-06 | 2021-08-19 | Robert Bosch Gmbh | Method for selecting an image detail of a sensor |
| US20200284883A1 (en) * | 2019-03-08 | 2020-09-10 | Osram Gmbh | Component for a lidar sensor system, lidar sensor system, lidar sensor device, method for a lidar sensor system and method for a lidar sensor device |
| US20200301799A1 (en) * | 2019-03-23 | 2020-09-24 | Uatc, Llc | Systems and Methods for Generating Synthetic Sensor Data via Machine Learning |
| US11275673B1 (en) * | 2019-06-24 | 2022-03-15 | Zoox, Inc. | Simulated LiDAR data |
| US10841483B1 (en) * | 2019-07-11 | 2020-11-17 | Denso International America, Inc. | System and method for calibrating at least one camera and a light detection and ranging sensor |
| US20220091266A1 (en) * | 2020-09-18 | 2022-03-24 | Denso International America, Inc. | Systems and methods for enhancing outputs of a lidar |
| US20220137218A1 (en) * | 2020-10-30 | 2022-05-05 | Waymo Llc | Detecting Retroreflectors in NIR Images to Control LIDAR Scan |
| US20220221585A1 (en) * | 2021-01-14 | 2022-07-14 | Argo AI, LLC | Systems and methods for monitoring lidar sensor health |
| US20230274462A1 (en) * | 2022-02-28 | 2023-08-31 | Basis Software, Inc. | System and method for camera calibration |
| US20250022143A1 (en) * | 2023-07-13 | 2025-01-16 | Waymo Llc | Object tracking across a wide range of distances for driving applications |
Also Published As
| Publication number | Publication date |
|---|---|
| US20230393280A1 (en) | 2023-12-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10754037B2 (en) | Processing point clouds of vehicle sensors having variable scan line distributions using voxel grids | |
| JP7440324B2 (en) | Vehicle control device, vehicle control method, and program | |
| US10394243B1 (en) | Autonomous vehicle technology for facilitating operation according to motion primitives | |
| US10481605B1 (en) | Autonomous vehicle technology for facilitating safe stopping according to separate paths | |
| JP2024045402A (en) | Vehicle control device, vehicle control method, and vehicle control program | |
| US11556127B2 (en) | Static obstacle map based perception system | |
| US20200097010A1 (en) | Autonomous vehicle technology for facilitating safe stopping according to hybrid paths | |
| US12479462B2 (en) | Retrofit vehicle computing system to operate with multiple types of maps | |
| US12545297B2 (en) | Methods and systems for generating a longitudinal plan for an autonomous vehicle based on behavior of uncertain road users | |
| JP7369078B2 (en) | Vehicle control device, vehicle control method, and program | |
| US11496707B1 (en) | Fleet dashcam system for event-based scenario generation | |
| US12441369B2 (en) | Automated cut-in identification and classification | |
| US11932242B1 (en) | Fleet dashcam system for autonomous vehicle operation | |
| CN116569070A (en) | Method and system for analyzing dynamic LiDAR point cloud data | |
| US12372651B2 (en) | Retrofit light detection and ranging (LIDAR)-based vehicle system to operate with vision-based sensor data | |
| CN116674591A (en) | Automatic driving vehicle operation method, electronic device, and computer-readable medium | |
| US20230399008A1 (en) | Multistatic radar point cloud formation using a sensor waveform encoding schema | |
| JP2021149464A (en) | Vehicle control apparatus, vehicle control method, and program | |
| CN115508841A (en) | Road edge detection method and device | |
| US20250110230A1 (en) | Multi-axis Radar Velocity Image from Stereo Doppler Radar | |
| JP2026008945A (en) | Method and system for mitigating the effects of weather-related attenuation on radar images | |
| JP2026502808A (en) | Multiframe Temporal Aggregation and Dense Motion Estimation for Autonomous Vehicles | |
| CN117242488A (en) | Autonomous vehicle system for object detection using logistic cylinder pedestrian model |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: GM CRUISE HOLDINGS LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DONDERICI, BURKAY;REEL/FRAME:060084/0138 Effective date: 20220527 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |