EP3515780A1 - Véhicule autonome à architecture modulaire - Google Patents

Véhicule autonome à architecture modulaire

Info

Publication number
EP3515780A1
EP3515780A1 EP16781619.8A EP16781619A EP3515780A1 EP 3515780 A1 EP3515780 A1 EP 3515780A1 EP 16781619 A EP16781619 A EP 16781619A EP 3515780 A1 EP3515780 A1 EP 3515780A1
Authority
EP
European Patent Office
Prior art keywords
controller
vehicle
autonomous vehicle
driving
corridor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP16781619.8A
Other languages
German (de)
English (en)
Inventor
Troy Jones
Scott Lennox
John Sgueglia
Nicholas Alexander Zervoglos
Jon Demerly
Hsin-Hsiang Yang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Charles Stark Draper Laboratory Inc
Veoneer US LLC
Original Assignee
Charles Stark Draper Laboratory Inc
Veoneer US LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Charles Stark Draper Laboratory Inc, Veoneer US LLC filed Critical Charles Stark Draper Laboratory Inc
Publication of EP3515780A1 publication Critical patent/EP3515780A1/fr
Withdrawn legal-status Critical Current

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/09Taking automatic action to avoid collision, e.g. braking and steering
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W60/00Drive control systems specially adapted for autonomous road vehicles
    • B60W60/001Planning or execution of driving tasks
    • B60W60/0011Planning or execution of driving tasks involving control alternatives for a single driving scenario, e.g. planning several paths to avoid obstacles
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/04Conjoint control of vehicle sub-units of different type or different function including control of propulsion units
    • B60W10/06Conjoint control of vehicle sub-units of different type or different function including control of propulsion units including control of combustion engines
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/18Conjoint control of vehicle sub-units of different type or different function including control of braking systems
    • B60W10/184Conjoint control of vehicle sub-units of different type or different function including control of braking systems with wheel brakes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/20Conjoint control of vehicle sub-units of different type or different function including control of steering systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/095Predicting travel path or likelihood of collision
    • B60W30/0953Predicting travel path or likelihood of collision the prediction being responsive to vehicle dynamic parameters
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/095Predicting travel path or likelihood of collision
    • B60W30/0956Predicting travel path or likelihood of collision the prediction being responsive to traffic or environmental parameters
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/10Path keeping
    • B60W30/12Lane keeping
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/14Adaptive cruise control
    • B60W30/16Control of distance between vehicles, e.g. keeping a distance to preceding vehicle
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/04Monitoring the functioning of the control system
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W60/00Drive control systems specially adapted for autonomous road vehicles
    • B60W60/001Planning or execution of driving tasks
    • B60W60/0027Planning or execution of driving tasks using trajectory prediction for other traffic participants
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W2050/0001Details of the control system
    • B60W2050/0002Automatic control, details of type of controller or control system architecture
    • B60W2050/0004In digital systems, e.g. discrete-time systems involving sampling
    • B60W2050/0006Digital architecture hierarchy
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2554/00Input parameters relating to objects
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2554/00Input parameters relating to objects
    • B60W2554/40Dynamic objects, e.g. animals, windblown objects
    • B60W2554/404Characteristics
    • B60W2554/4041Position
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2556/00Input parameters relating to data
    • B60W2556/45External transmission of data to or from the vehicle
    • B60W2556/50External transmission of data to or from the vehicle of positioning data, e.g. GPS [Global Positioning System] data
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2556/00Input parameters relating to data
    • B60W2556/45External transmission of data to or from the vehicle
    • B60W2556/65Data transmitted between vehicles
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/06Combustion engines, Gas turbines
    • B60W2710/0605Throttle position
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/18Braking system
    • B60W2710/182Brake pressure, e.g. of fluid or between pad and disc
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/20Steering systems
    • B60W2710/207Steering angle of wheels
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2756/00Output or target parameters relating to data
    • B60W2756/10Involving external transmission of data to or from the vehicle

Definitions

  • vehicles can employ automated systems such as lane assist, pre- collision breaking, and rear cross-track detection. These systems can assist a driver of the vehicle from making human error and to avoid crashes with other vehicles, moving objects, or pedestrians. However, these systems only automate certain vehicle functions, and still rely on the driver of the vehicle for other operations.
  • the method further includes determining, at a vehicle controller (VC), based on the one or more planned driving corridors, one or more vehicle trajectories which are predicted to avoid collisions with the surrounding vehicles and other objects.
  • the method further includes selecting, at the VC, one of the determined trajectories as active based on criteria of collision likelihood.
  • the method further includes sending steering, throttle, and braking commands from the VC to one or more respective actuator controllers within the vehicle to follow the active trajectory.
  • the method further includes, at a perception controller (PC), generating a stochastic prediction of free space available for driving based on locations of observed vehicles or other objects.
  • the method further includes, at the ADC, determining the driving corridors from the stochastic prediction of free space.
  • the method further includes, at the PC, generating a kinematic prediction of the free space available for driving by performing kinematic-based predictions of the locations of vehicles and objects within a threshold radius.
  • the method further includes, at the VC, determining one or more corridor trajectories in a given corridor that meet ride comfort design goals, determining an emergency trajectory from an updated kinematic free space prediction, selecting a nominal or emergency trajectory based on a collision-avoidance likelihood, the collision-avoidance likelihood being based on the updated kinematic free space prediction, generating updated steering, throttle, and braking commands to follow the selected trajectory corridor, and sending updated commands to the actuator controllers for execution.
  • the method includes, at the ADC, planning a route of roadways for the autonomous vehicle to travel to a destination, wherein determining the corridor within the route of roadways is based on information determined by a plurality of sensors about a driving surface and objects surrounding the autonomous vehicle.
  • the method further includes translating inputs from a plurality of sensor subsystems to a vendor-neutral format at a Sensor Interface Server (SIS).
  • SIS Sensor Interface Server
  • the method includes determining a position and an attitude of the autonomous vehicle from a plurality of sensor inputs at a Localization
  • the method includes determining objects around the autonomous vehicle and on drivable surfaces, detected by a plurality of sensor subsystems, at a Perception Controller (PC).
  • PC Perception Controller
  • the method includes providing an interface for interaction between an operator, passengers, and humans external to the vehicle at a Human Interaction Controller (HIC).
  • HIC Human Interaction Controller
  • the method includes interacting with other-self driving cars or automated systems through a Machine Interaction Controller (MC).
  • MC Machine Interaction Controller
  • the method includes coordinating exchanges, at a System Controller (SC), of control between an operator and elements of the autonomous vehicle.
  • SC System Controller
  • the system controller can further monitor the elements of the autonomous vehicle for failures or other abnormal behavior, managing corrective actions to resolve failures.
  • the method can include, at the VC, driving outside of the ADC determined corridor in response to determining a likelihood of collision within that corridor.
  • the method includes minimizing use of communication bandwidth in an architecture of the autonomous vehicle by providing a sensor interface server (SIS) configured to translate and filter sensor data sent to other elements of the architecture, a perception controller (PC) configured to fuse sensor measurements from a plurality of sensors into a single estimate of perceptions of the environment around the autonomous vehicle, and a localization controller (LC) configured to fuse outputs from multiple sensor measurements into a single position and attitude of the autonomous vehicle in the world.
  • SIS sensor interface server
  • PC perception controller
  • LC localization controller
  • the method includes measuring availability of the operator of the autonomous vehicle to assist in driving the vehicle and providing a variable level of automated function by the autonomous vehicle based on the measured availability of the operator at a human interface controller (HC).
  • HC human interface controller
  • the method includes, at a human interface controller (HC), translating internal status of the autonomous vehicle to a human-understandable format (e.g., visual, audio, haptic), and presenting the translated internal status in the human- understandable format to the operator.
  • a human interface controller e.g., HC
  • a human-understandable format e.g., visual, audio, haptic
  • a system for operating an autonomous vehicle includes an Automated Driving Controller (ADC) configured to determine, one or more planned driving corridors that are predicted to be drivable by the vehicle and safely separated from
  • ADC Automated Driving Controller
  • the system further includes a vehicle controller (VC) configured to determine, based on the one or more planned driving corridors, one or more vehicle trajectories which are predicted to avoid collisions with the surrounding vehicles and other objects, select one of the determined trajectories as active based on criteria of collision likelihood, and send steering, throttle, and braking commands to one or more respective actuator controllers within the vehicle to follow the active trajectory.
  • VC vehicle controller
  • the system includes a perception controller (PC) configured to generate a stochastic prediction of free space available for driving based on locations of observed vehicles or other objects.
  • the ADC is further configured to determine the driving corridors from the stochastic prediction of free space.
  • the PC is further configured to generate a kinematic prediction of the free space available for driving by performing kinematic-based predictions of the locations of vehicles and objects within a threshold radius.
  • the VC is further configured to determine one or more corridor trajectories in a given corridor that meet ride comfort design goals, determine an emergency trajectory corridor from an updated kinematic free space prediction, select a nominal or emergency trajectory based on a collision-avoidance likelihood, the collision-avoidance likelihood being based on the updated kinematic free space prediction, generate updated steering, throttle, and braking commands to follow the selected trajectory corridor, and send updated commands to the actuator controllers for execution.
  • Fig. 1 is a diagram illustrating steps in an embodiment of an automated control system of the Observe, Orient, Decide, and Act (OODA) model.
  • Fig. 2 is a block diagram of an embodiment of an autonomous vehicle high-level architecture.
  • Fig. 3 is a block diagram illustrating an embodiment of the sensor interaction controller (SIC), perception controller (PC), and localization controller (LC).
  • SIC sensor interaction controller
  • PC perception controller
  • LC localization controller
  • Fig. 4 is a block diagram illustrating an example embodiment of the automatic driving controller (ADC), vehicle controller (VC) and actuator controller.
  • ADC automatic driving controller
  • VC vehicle controller
  • actuator controller actuator controller
  • Fig. 5 is a diagram illustrating decision time scales of the ADC and VC.
  • Fig. 6 is a block diagram illustrating an example embodiment of the system controller, human interface controller (HC) and machine interface controller (MC).
  • HC human interface controller
  • MC machine interface controller
  • FIG. 7 illustrates a computer network or similar digital processing environment in which embodiments of the present invention may be implemented.
  • Fig. 8 is a diagram of an example internal structure of a computer (e.g., client processor/device or server computers in the computer system of Fig. 7.
  • Fig. 1 is a diagram illustrating steps in an embodiment of an automated control system of the Observe, Orient, Decide, and Act (OODA) model.
  • Automated systems such as highly-automated driving systems, or, self-driving cars, or autonomous vehicles, employ an OODA model.
  • the observe virtual layer 102 involves sensing features from the world using machine sensors, such as laser ranging, radar, infra-red, vision systems, or other systems.
  • the orientation virtual layer 104 involves perceiving situational awareness based on the sensed information. Examples of orientation virtual layer activities are Kalman filtering, model based matching, machine or deep learning, and Bayesian predictions.
  • the decide virtual layer 106 selects an action from multiple objects to a final decision.
  • the act virtual layer 108 provides guidance and control for executing the decision.
  • Fig. 2 is a block diagram 200 of an embodiment of an autonomous vehicle high-level architecture 206. The
  • architecture 206 is built using a top-down approach to enable fully automated driving.
  • the architecture 206 is preferably modular such that it can be adaptable with hardware from different vehicle manufacturers.
  • the architecture 206 therefore, has several modular elements functionally divided to maximize these properties.
  • the modular architecture 206 described herein can interface with sensor systems 202 of any vehicle 204. Further, the modular architecture 206 can receive vehicle information from and communicate with any vehicle 204.
  • Elements of the modular architecture 206 include sensors 202, Sensor Interface Controller (SIC) 208, localization controller (LC) 210, perception controller (PC) 212, automated driving controller 214 (ADC), vehicle controller 216 (VC), system controller 218 (SC), human interaction controller 220 (HC) and machine interaction controller 222 (MC).
  • SIC Sensor Interface Controller
  • LC localization controller
  • PC perception controller
  • ADC automated driving controller
  • VC vehicle controller
  • SC system controller 218
  • HC human interaction controller 220
  • MC machine interaction controller 222
  • the observation layer of the model includes gathering sensor readings, for example, from vision sensors, Radar (Radio Detection And Ranging), LIDAR (Light Detection And
  • the sensors 202 shown in Fig. 2 shows such an observation layer.
  • Examples of the orientation layer of the model can include determining where a car is relative to the world, relative to the road it is driving on, and relative to lane markings on the road, shown by Perception Controller (PC) 212 and
  • LC Localization Controller
  • ADC Automatic Driving Controller
  • VC Vehicle Controller
  • Act layer include converting that corridor into commands to the vehicle's driving systems (e.g., steering sub-system, acceleration sub-system, and breaking sub-system) that direct the car along the corridor, such as actuator control 410 of Fig. 4.
  • a person of ordinary skill in the art can recognize that the layers of the system are not strictly sequential, and as observations change, so do the results of the other layers.
  • changing conditions on the road such as detection of another object, may direct the car to modify its corridor, or enact emergency procedures to prevent a collision.
  • the commands of the vehicle controller may need to be adjusted dynamically to compensate for drift, skidding, or other changes to expected vehicle behavior.
  • the module architecture 206 receives measurements from sensors 202. While different sensors may output different sets of information in different formats, the modular architecture 206 includes Sensor Interface Controller (SIC) 208, sometimes also referred to as a Sensor Interface Server (SIS), configured to translate the sensor data into data having a vendor-neutral format that can be read by the modular architecture 206. Therefore, the modular architecture 206 learns about the environment around the vehicle 204 from the vehicle's sensors, no matter the vendor, manufacturer, or configuration of the sensors. The SIS 208 can further tag each sensor's data with a metadata tag having its location and orientation in the car, which can be used by the perception controller to determine the unique angle, perspective, and blind spot of each sensor.
  • SIC Sensor Interface Controller
  • SIS Sensor Interface Server
  • the modular architecture 206 includes vehicle controller 216 (VC).
  • the VC 216 is configured to send commands to the vehicle and receive status messages from the vehicle.
  • the vehicle controller 216 receives status messages from the vehicle 204 indicating the vehicle's status, such as information regarding the vehicle's speed, attitude, steering position, braking status, and fuel level, or any other information about the vehicle's subsystems that is relevant for autonomous driving.
  • the modular architecture 206 based on the information from the vehicle 204 and the sensors 202, therefore can calculate commands to send from the VC 216 to the vehicle 204 to implement self-driving.
  • the functions of the various modules within the modular architecture 206 are described in further detail below.
  • the modular architecture 206 when viewing the modular architecture 206 at a high level, it receives (a) sensor information from the sensors 202 and (b) vehicle status information from the vehicle 204, and in turn, provides the vehicle instructions to the vehicle 204.
  • Such an architecture allows the modular architecture to be employed for any vehicle with any sensor configuration.
  • any vehicle platform that includes a sensor subsystem (e.g., sensors 202) and an actuation subsystem having the ability to provide vehicle status and accept driving commands (e.g., actuator control 410 of Fig. 4) can integrate with the modular architecture 206.
  • a sensor subsystem e.g., sensors 202
  • an actuation subsystem having the ability to provide vehicle status and accept driving commands (e.g., actuator control 410 of Fig. 4) can integrate with the modular architecture 206.
  • the modular architecture 206 various modules work together to implement automated driving according to the OODA model.
  • the sensors 202 and SIC 208 reside in the "observe" virtual layer.
  • the SIC 208 receives measurements (e.g., sensor data) having various formats.
  • the SIC 208 is configured to convert vendor-specific data directly from the sensors to vendor-neutral data.
  • the set of sensors 202 can include any brand of Radar, LIDAR, image sensor, or other sensors, and the modular architecture 206 can use their perceptions of the environment effectively.
  • the measurements output by the sensor interface server are then processed by perception controller (PC) 212 and localization controller (LC) 210.
  • the PC 212 and LC 210 both reside in the "orient" virtual layer of the OODA model.
  • the LC 210 determines a robust world-location of the vehicle that can be more precise than a GPS signal, and still determines the world-location of the vehicle when there is no available or an inaccurate GPS signal.
  • the LC 210 determines the location based on GPS data and sensor data.
  • the PC 212 on the other hand, generates prediction models representing a state of the environment around the car, including objects around the car and state of the road.
  • Fig. 3 provides further details regarding the SIC 208, LC 210 and PC 212.
  • Automated driving controller 214 and vehicle controller 216 (VC) receive the outputs of the perception controller and localization controller.
  • the ADC 214 and VC 216 reside in the "decide" virtual layer of the OODA model.
  • the ADC 214 is responsible for destination selection, route and lane guidance, and high-level traffic surveillance.
  • the ADC 214 further is responsible for lane selection within the route, and identification of safe harbor areas to diver the vehicle in case of an emergency.
  • the ADC 214 selects a route to reach the destination, and a corridor within the route to direct the vehicle.
  • the ADC 214 passes this corridor onto the VC 216. Given the corridor, the VC 216 provides a trajectory and lower level driving functions to direct the vehicle through the corridor safely.
  • the VC 216 first determines the best trajectory to maneuver through the corridor while providing comfort to the driver, an ability to reach safe harbor, emergency maneuverability, and ability to follow the vehicle's current trajectory. In emergency situations, the VC 216 overrides the corridor provided by the ADC 214 and immediately guides the car into a safe harbor corridor, returning to the corridor provided by the ADC 214 when it is safe to do so.
  • the VC 216 after determining how to maneuver the vehicle, including safety maneuvers, then provides actuation commands to the vehicle 204, which executes the commands in its steering, throttle, and braking subsystems. This element of the VC 216 is therefore in the "act" virtual layer of the OODA model.
  • Fig. 4 describes the ADC 214 and VC 216 in further detail.
  • the modular architecture 206 further coordinates communication with various modules through system controller 218 (SC).
  • SC system controller 218
  • the SC 218 enables operation of human interaction controller 220 (HC) and machine interaction controller 222 (MC).
  • HC human interaction controller 220
  • MC machine interaction controller 222
  • the HC 220 provides information about the autonomous vehicle's operation in a human understandable format based on status messages coordinated by the system controller.
  • the HC 220 further allows for human input to be factored into the car's decisions.
  • the HC 220 enables the operator of the vehicle to enter or modify the destination or route of the vehicle, as one example.
  • the SC 218 interprets the operator's input and relays the information to the VC 216 or ADC 214 as necessary.
  • the MC 222 can coordinate messages with other machines or vehicles.
  • other vehicles can electronically and wirelessly transmit route intentions, intended corridors of travel, and sensed objects that may be in other vehicle's blind spot to autonomous vehicles, and the MC 222 can receive such information, and relay it to the VC 216 and ADC 214 via the SC 218.
  • the MC 222 can send information to other vehicles wirelessly.
  • the MC 222 can receive a notification that the vehicle intends to turn.
  • the MC 222 receives this information via the VC 216 sending a status message to the SC 218, which relays the status to the MC 222.
  • other examples of machine communication can also be implemented.
  • Fig. 3 is a block diagram 300 illustrating an embodiment of the sensor interaction controller 304 (SIC), perception controller (PC) 306, and localization controller (LC) 308.
  • a sensor array 302 of the vehicle can include various types of sensors, such as a camera 302a, radar 302b, LIDAR 302c, GPS 302d, IMU 302e, or vehicle-to-everything (V2X) 302f. Each sensor sends individual vendor defined data types to the SIC 304.
  • the camera 302a sends object lists and images
  • the radar 302b sends object lists, and in-phase/quadrature (IQ) data
  • the LIDAR 302c sends object lists and scan points
  • the GPS 302d sends position and velocity
  • the IMU 302e sends acceleration data
  • the V2X 302f controller sends tracks of other vehicles, turn signals, other sensor data, or traffic light data.
  • the SIC 304 monitors and diagnoses faults at each of the sensors 302a-f.
  • the SIC 304 isolates the data from each sensor from its vendor specific package and sends vendor neutral data types to the perception controller (PC) 306 and localization controller 308 (LC).
  • the SIC 304 forwards localization feature measurements and position and attitude measurements to the LC 308, and forwards tracked object measurements, driving surface measurements, and position & attitude measurements to the PC 306.
  • the SIC 304 can further be updated with firmware so that new sensors having different formats can be used with the same modular architecture.
  • the LC 308 fuses GPS and IMU data with Radar, Lidar, and Vision data to determine a vehicle location, velocity, and attitude with more precision than GPS can provide alone.
  • the LC 308 reports that robustly determined location, velocity, and attitude to the PC 306.
  • the LC 308 further monitors measurements representing position, velocity, and attitude data for accuracy relative to each other, such that if one sensor measurement fails or becomes degraded, such as a GPS signal in a city, the LC 308 can correct for it.
  • the PC 306 identifies and locates objects around the vehicle based on the sensed information.
  • the PC 306 further estimates drivable surface regions surrounding the vehicle, and further estimates other surfaces such as road shoulders or drivable terrain in the case of an emergency.
  • the PC 306 further provides a stochastic prediction of future locations of objects.
  • the PC 306 further stores a history of objects and drivable surfaces.
  • the PC 306 outputs two predictions, a strategic prediction, and a tactical prediction.
  • the tactical prediction represents the world around 2-4 seconds into the future, which only predicts the nearest traffic and road to the vehicle. This prediction includes a free space harbor on shoulder of the road or other location. This tactical prediction is based entirely on measurements from sensors on the vehicle of nearest traffic and road conditions.
  • the strategic prediction is a long term prediction that predicts areas of the car's visible environment beyond the visible range of the sensors. This prediction is for greater than four seconds into the future, but has a higher uncertainty than the tactical prediction because objects (e.g., cars and people) may change their currently observed behavior in an unanticipated manner.
  • objects e.g., cars and people
  • Such a prediction can also be based on sensor measurements from external sources including other autonomous vehicles, manual vehicles with a sensor system and sensor communication network, sensors positioned near or on the roadway or received over a network from transponders on the objects, and traffic lights, signs, or other signals configured to communicate wirelessly with the autonomous vehicle.
  • FIG. 4 is a block diagram 400 illustrating an example embodiment of the automatic driving controller (ADC) 402, vehicle controller (VC) 404 and actuator controller 410.
  • ADC automatic driving controller
  • VC vehicle controller
  • actuator controller 410 The ADC 402 and VC 404 execute the "decide" virtual layer of the OODA model.
  • the ADC 402 based on destination input by the operator and current position, first creates an overall route from the current position to the destination including a list of roads and junctions between roads in order to reach the destination.
  • This strategic route plan may be based on traffic conditions, and can change based on updating traffic conditions, however such changes are generally enforced for large changes in estimated time of arrival (ETA).
  • ETA estimated time of arrival
  • the ADC 402 plans a safe, collision-free, corridor for the autonomous vehicle to drive through based on the surrounding objects and permissible drivable surface - both supplied by the PC.
  • This corridor is continuously sent as a request to the VC 404 and is updated as traffic and other conditions change.
  • the VC 404 receives the updates to the corridor in real time.
  • the ADC 402 receives back from the VC 404 the current actual trajectory of the vehicle, which is also used to modify the next planned update to the driving corridor request.
  • the ADC 402 generates a strategic corridor for the vehicle to navigate.
  • the ADC 402 generates the corridor based on predictions of the free space on the road in the strategic/tactical prediction.
  • the ADC 402 further receives the vehicle position information and vehicle attitude information from the perception controller of Fig. 3.
  • the VC 404 further provides the ADC 402 with an actual trajectory of the vehicle from the vehicle's actuator control 410. Based on this information, the ADC 402 calculates feasible corridors to drive the road, or any drivable surface. In the example of being on an empty road, the corridor may follow the lane ahead of the car.
  • the ADC 402 can determine whether there is free space in a passing lane and in front of the car to safely execute the pass.
  • the ADC 402 can automatically calculate based on (a) the current distance to the car to be passed, (b) amount of drivable road space available in the passing lane, (c) amount of free space in front of the car to be passed, (d) speed of the vehicle to be passed, (e) current speed of the autonomous vehicle, and (f) known acceleration of the autonomous vehicle, a corridor for the vehicle to travel through to execute the pass maneuver.
  • the ADC 402 can determine a corridor to switch lanes when approaching a highway exit. In addition to all of the above factors, the ADC 402 monitors the planned route to the destination and, upon approaching a junction, calculates the best corridor to safely and legally continue on the planned route.
  • the ADC 402 the provides the requested corridor 406 to the VC 404, which works in tandem with the ADC 402 to allow the vehicle to navigate the corridor.
  • the requested corridor 406 places geometric and velocity constraints on any planned trajectories for a number of seconds into the future.
  • the VC 404 determines a trajectory to maneuver within the corridor 406.
  • the VC 404 bases its maneuvering decisions from the tactical / maneuvering prediction received from the perception controller and the position of the vehicle and the attitude of the vehicle. As described previously, the tactical / maneuvering prediction is for a shorter time period, but has less uncertainty. Therefore, for lower-level maneuvering and safety calculations, the VC 404 effectively uses the tactical / maneuvering prediction to plan collision-free trajectories within requested corridor 406. As needed in emergency situations, the VC 404 plans trajectories outside the corridor 406 to avoid collisions with other objects.
  • the VC 404 determines, based on the requested corridor 406, the current velocity and acceleration of the car, and the nearest objects, how to drive the car through that corridor 406 while avoiding collisions with objects and remain on the drivable surface.
  • the VC 404 calculates a tactical trajectory within the corridor, which allows the vehicle to maintain a safe separation between objects.
  • the tactical trajectory also includes a backup safe harbor trajectory in the case of an emergency, such as a vehicle unexpectedly
  • the VC 404 may be required to command a maneuver suddenly outside of the requested corridor from the ADC 402. This emergency maneuver can be initiated entirely by the VC 404 as it has faster response times than the ADC 402 to imminent collision threats. This capability isolates the safety critical collision avoidance responsibility within the VC 404.
  • the VC 404 sends maneuvering commands to the actuators that control steering, throttling, and braking of the vehicle platform.
  • the VC 404 executes its maneuvering strategy by sending a current vehicle trajectory 408 having driving commands (e.g., steering, throttle, braking) to the vehicle's actuator controls 410.
  • the vehicle's actuator controls 410 apply the commands to the car's respective steering, throttle, and braking systems.
  • the VC 404 sending the trajectory 408 to the actuator controls represent the "Act" virtual layer of the OODA model.
  • the VC is the only component needing configuration to control a specific model of car (e.g., format of each command, acceleration performance, turning performance, and braking performance), whereas the ADC remaining highly agnostic to the specific vehicle capacities.
  • the VC 404 can be updated with firmware configured to allow interfacing with particular vehicle's actuator control systems, or a fleet-wide firmware update for all vehicles.
  • Fig. 5 is a diagram 500 illustrating decision time scales of the ADC 402 and VC 404.
  • the ADC 402 implements higher-level, strategic 502 and tactical 504 decisions by generating the corridor.
  • the ADC 402 therefore implements the decisions having a longer range/ or time scale.
  • the estimate of world state used by the ADC 402 for planning strategic routes and tactical driving corridors for behaviors such as passing or making turns has higher uncertainty, but predicts longer into the future, which is necessary for planning these autonomous actions.
  • the strategic predictions have high uncertainty because they predict beyond the sensor's visible range, relying solely on non-vision technologies, such as Radar, for predictions of objects far away from the car, that events can change quickly due to, for example, a human suddenly changing his or her behavior, or the lack of visibility of objects beyond the visible range of the sensors.
  • Many tactical decisions such as passing a car at highway speed, require perception Beyond the Visible Range (BVR) of an autonomous vehicle (e.g., 100m or greater), whereas all maneuverability 506 decisions are made based on locally perceived objects to avoid collisions.
  • BVR Visible Range
  • the VC 404 generates maneuverability decisions 506 using maneuverability predictions that are short time frame/range predictions of object behaviors and the driving surface. These maneuverability predictions have a lower uncertainty because of the shorter time scale of the predictions, however, they rely solely on measurements taken within visible range of the sensors on the autonomous vehicle. Therefore, the VC 404 uses these maneuverability predictions (or estimates) of the state of the environment immediately around the car for fast response planning of collision-free trajectories for the autonomous vehicle.
  • the VC 402 issues actuation commands, on the lowest end of the time scale, representing the execution of the already planned corridor and maneuvering through the corridor.
  • Fig. 6 is a block diagram 600 illustrating an example embodiment of the system controller 602, human interface controller 604 (HC) and machine interface controller 606 (MC).
  • the human interaction controller 604 (HC) receives input command requests from the operator.
  • the HC 604 also provides outputs to the operator, passengers of the vehicle, and humans external to the autonomous vehicle.
  • the HC 604 provides the operator and passengers (via visual, audio, haptic, or other interfaces) a human-understandable
  • the HC 604 can display the vehicle's long-term route, or planned corridor and safe harbor areas. Additionally, the HC 604 reads sensor measurements about the state of the driver, allowing the HC 604 to monitor the availability of the driver to assist with operations of the car at any time. As one example, a sensor system within the vehicle could sense whether the operator has hands on the steering wheel. If so, the HC 604 can signal that a transition to operator steering can be allowed, but otherwise, the HC 604 can prevent a turnover of steering controls to the operator. In another example, the HC 604 can synthesize and summarize decision making rationale to the operator, such as reasons why it selected a particular route.
  • a sensor system within the vehicle can monitor the direction the driver is looking.
  • the HC 604 can signal that a transition to driver operation is allowed if the driver is looking at the road, but if the driver is looking elsewhere, the system does not allow operator control.
  • the HC 604 can take over control, or emergency only control, of the vehicle while the operator checks the vehicle's blind spot and looks away from the windshield.
  • the machine interaction controller 606 interacts with other autonomous vehicles or automated system to coordinate activities such as formation driving or traffic management.
  • the MC 606 reads the internal system status and generates an output data type that can be read by collaborating machine systems, such as the V2X data type. This status can be broadcast over a network by collaborating systems.
  • the MC 606 can translate any command requests from external machine systems (e.g., slow down, change route, merge request, traffic signal status) into commands requests routed to the SC for arbitration against the other command requests from the HC 604.
  • the MC 606 can further authenticate (e.g., using signed messages from other trusted manufacturers) messages from other systems to ensure that they are valid and represent the environment around the car. Such an
  • the system controller 602 serves as an overall manager of the elements within the architecture.
  • the SC 602 aggregates the status data from all of the system elements to determine total operational status, and sends commands to the elements to execute system functions. If elements of the system report failures, the SC 602 initiates diagnostic and recovery behaviors to ensure autonomous operation such that the vehicle remains safe. Any transitions of the vehicle to/from an automated state of driving are approved or denied by the SC 602 pending the internal evaluation of operational readiness for automated driving and the availability of the human driver.
  • FIG. 7 illustrates a computer network or similar digital processing environment in which embodiments of the present invention may be implemented.
  • Client computer(s)/devices 50 and server computer(s) 60 provide processing, storage, and input/output devices executing application programs and the like.
  • the client computer(s)/devices 50 can also be linked through communications network 70 to other computing devices, including other client devices/processes 50 and server computer(s) 60.
  • the communications network 70 can be part of a remote access network, a global network (e.g., the Internet), a worldwide collection of computers, local area or wide area networks, and gateways that currently use respective protocols (TCP/IP, Bluetooth®, etc.) to communicate with one another.
  • Other electronic device/computer network architectures are suitable.
  • Fig. 8 is a diagram of an example internal structure of a computer (e.g., client processor/device 50 or server computers 60) in the computer system of Fig. 7.
  • Each computer 50, 60 contains a system bus 79, where a bus is a set of hardware lines used for data transfer among the components of a computer or processing system.
  • the system bus 79 is essentially a shared conduit that connects different elements of a computer system (e.g., processor, disk storage, memory, input/output ports, network ports, etc.) that enables the transfer of information between the elements.
  • Attached to the system bus 79 is an I/O device interface 82 for connecting various input and output devices (e.g., keyboard, mouse, displays, printers, speakers, etc.) to the computer 50, 60.
  • a network interface 86 allows the computer to connect to various other devices attached to a network (e.g., network 70 of Fig. 7).
  • Memory 90 provides volatile storage for computer software instructions 92 and data 94 used to implement an embodiment of the present invention (e.g., sensor interface controller, perception controller, localization controller, automated driving controller, vehicle controller, system controller, human interaction controller, and machine interaction controller detailed above).
  • Disk storage 95 provides non-volatile storage for computer software instructions 92 and data 94 used to implement an embodiment of the present invention.
  • a central processor unit 84 is also attached to the system bus 79 and provides for the execution of computer instructions.
  • the processor routines 92 and data 94 are a computer program product (generally referenced 92), including a non-transitory computer-readable medium (e.g., a removable storage medium such as one or more DVD-ROM's, CD-ROM's, diskettes, tapes, etc.) that provides at least a portion of the software instructions for the invention system.
  • the computer program product 92 can be installed by any suitable software installation procedure, as is well known in the art.
  • at least a portion of the software instructions may also be downloaded over a cable communication and/or wireless connection.
  • the invention programs are a computer program propagated signal product embodied on a propagated signal on a propagation medium (e.g., a radio wave, an infrared wave, a laser wave, a sound wave, or an electrical wave propagated over a global network such as the Internet, or other network(s)).
  • a propagation medium e.g., a radio wave, an infrared wave, a laser wave, a sound wave, or an electrical wave propagated over a global network such as the Internet, or other network(s)
  • Such carrier medium or signals may be employed to provide at least a portion of the software instructions for the present invention routines/program 92.

Landscapes

  • Engineering & Computer Science (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Automation & Control Theory (AREA)
  • Chemical & Material Sciences (AREA)
  • Combustion & Propulsion (AREA)
  • Human Computer Interaction (AREA)
  • Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
  • Traffic Control Systems (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
  • Navigation (AREA)

Abstract

Une architecture (206) pour véhicule autonome utilise une approche de haut-bas pour permettre une conduite entièrement automatisée. L'architecture (206) est modulaire et compatible avec le matériel provenant de différent fabricants. Chaque composant modulaire peut être adapté pour des voitures particulières, qui ont différents sous-systèmes de commande de véhicule et différents sous-systèmes de capteur.
EP16781619.8A 2016-09-29 2016-09-29 Véhicule autonome à architecture modulaire Withdrawn EP3515780A1 (fr)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2016/054466 WO2018063250A1 (fr) 2016-09-29 2016-09-29 Véhicule autonome à architecture modulaire

Publications (1)

Publication Number Publication Date
EP3515780A1 true EP3515780A1 (fr) 2019-07-31

Family

ID=57133432

Family Applications (1)

Application Number Title Priority Date Filing Date
EP16781619.8A Withdrawn EP3515780A1 (fr) 2016-09-29 2016-09-29 Véhicule autonome à architecture modulaire

Country Status (3)

Country Link
EP (1) EP3515780A1 (fr)
JP (1) JP2019530609A (fr)
WO (1) WO2018063250A1 (fr)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11372403B2 (en) * 2018-11-05 2022-06-28 Tusimple, Inc. Systems and methods for dynamic predictive control of autonomous vehicles
CN109733411B (zh) * 2019-02-20 2020-10-13 苏州风图智能科技有限公司 一种车速控制方法和装置
WO2020233772A1 (fr) * 2019-05-17 2020-11-26 Volvo Truck Corporation Procédé permettant de faire fonctionner un véhicule autonome
KR102231503B1 (ko) * 2019-09-02 2021-03-24 엘지전자 주식회사 자율 주행 제어 방법과 이를 이용한 자율 주행 시스템
CN115734906A (zh) * 2020-07-21 2023-03-03 松下知识产权经营株式会社 移动控制系统、移动控制方法、程序及移动体
US11783178B2 (en) 2020-07-30 2023-10-10 Toyota Research Institute, Inc. Systems and methods for corridor intent prediction
DE102020215778A1 (de) * 2020-12-14 2022-06-15 Continental Automotive Gmbh Verfahren zur Planung eines zumindest teilweise automatisierten Fahrvorgangs mittels eines Fahrassistenzsystems
JP7468411B2 (ja) * 2021-03-05 2024-04-16 トヨタ自動車株式会社 自動運転車両、配車管理装置、及び端末機器
CN114162129B (zh) * 2021-12-16 2024-06-25 华人运通(上海)云计算科技有限公司 车辆碰撞责任判定方法、装置和系统
US20240034356A1 (en) * 2022-07-28 2024-02-01 Zoox, Inc. Reference trajectory validating and collision checking management

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5917920A (en) * 1997-05-29 1999-06-29 Humphries; Alan Safety vehicle communication system
US8949018B2 (en) * 2011-06-13 2015-02-03 Toyota Jidosha Kabushiki Kaisha Driving assistance device and driving assistance method
US9381916B1 (en) * 2012-02-06 2016-07-05 Google Inc. System and method for predicting behaviors of detected objects through environment representation
US9120485B1 (en) * 2012-09-14 2015-09-01 Google Inc. Methods and systems for smooth trajectory generation for a self-driving vehicle
DE102013212710A1 (de) * 2013-05-16 2014-11-20 Siemens Aktiengesellschaft Sensorprodukt, Simulator und Verfahren zur Simulation von Sensormessungen, zur Fusion von Sensormessungen, zur Validierung eines Sensormodells und zum Entwurf eines Fahrerassistenzsystems
JP6489632B2 (ja) * 2014-09-30 2019-03-27 株式会社Subaru 車両の走行支援装置
JP6354561B2 (ja) * 2014-12-15 2018-07-11 株式会社デンソー 軌道判定方法、軌道設定装置、自動運転システム
KR102029562B1 (ko) * 2015-01-05 2019-10-07 닛산 지도우샤 가부시키가이샤 목표 경로 생성 장치 및 주행 제어 장치
JP6573769B2 (ja) * 2015-02-10 2019-09-11 国立大学法人金沢大学 車両走行制御装置
DE102015001971A1 (de) * 2015-02-19 2016-08-25 Iav Gmbh Ingenieurgesellschaft Auto Und Verkehr Verfahren und Überwachungsvorrichtung zur Überwachung von Fahrerassistenzsystemen

Also Published As

Publication number Publication date
WO2018063250A8 (fr) 2018-06-07
JP2019530609A (ja) 2019-10-24
WO2018063250A1 (fr) 2018-04-05

Similar Documents

Publication Publication Date Title
US10377375B2 (en) Autonomous vehicle: modular architecture
US10599150B2 (en) Autonomous vehicle: object-level fusion
US10963462B2 (en) Enhancing autonomous vehicle perception with off-vehicle collected data
US20180087907A1 (en) Autonomous vehicle: vehicle localization
WO2018063250A1 (fr) Véhicule autonome à architecture modulaire
US11789461B2 (en) Autonomous vehicle collision mitigation systems and methods
EP3516422A1 (fr) Localisation de véhicule autonome
US10902165B2 (en) Deployable development platform for autonomous vehicle (DDPAV)
EP3665061B1 (fr) Commande à tolérance de panne d'un véhicule autonome à multiples voies de commande
US10503165B2 (en) Input from a plurality of teleoperators for decision making regarding a predetermined driving situation
US11679776B2 (en) Autonomous vehicle safety platform system and method
US11513518B2 (en) Avoidance of obscured roadway obstacles
WO2018063241A1 (fr) Véhicule autonome : combinaison de niveau objet
CN110998469A (zh) 对具有自主驾驶能力的车辆的操作进行干预
US11565717B2 (en) Method and system for remote assistance of an autonomous agent
WO2018199941A1 (fr) Amélioration de la perception de véhicule autonome à l'aide de données collectées hors véhicule
JP7517835B2 (ja) 不明瞭な路上障害物の回避
JP2024526037A (ja) 自律エージェントの遠隔支援のための方法及びシステム
KR20240055023A (ko) 자동차의 운행 환경을 모델링하는 방법

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: UNKNOWN

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20190425

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20200603