WO2012129437A2 - Driver assistance system - Google Patents

Driver assistance system Download PDF

Info

Publication number
WO2012129437A2
WO2012129437A2 PCT/US2012/030185 US2012030185W WO2012129437A2 WO 2012129437 A2 WO2012129437 A2 WO 2012129437A2 US 2012030185 W US2012030185 W US 2012030185W WO 2012129437 A2 WO2012129437 A2 WO 2012129437A2
Authority
WO
WIPO (PCT)
Prior art keywords
vehicle
data
driver assistance
driver
assistance system
Prior art date
Application number
PCT/US2012/030185
Other languages
French (fr)
Other versions
WO2012129437A3 (en
Inventor
Troy Otis Cooprider
Michael J. Schmidlin
Faroog Ibrahim
Pavan K. VEMPATY
Original Assignee
Tk Holdings Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tk Holdings Inc. filed Critical Tk Holdings Inc.
Publication of WO2012129437A2 publication Critical patent/WO2012129437A2/en
Publication of WO2012129437A3 publication Critical patent/WO2012129437A3/en

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/14Adaptive cruise control
    • B60W30/143Speed control
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/04Conjoint control of vehicle sub-units of different type or different function including control of propulsion units
    • B60W10/06Conjoint control of vehicle sub-units of different type or different function including control of propulsion units including control of combustion engines
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/18Conjoint control of vehicle sub-units of different type or different function including control of braking systems
    • B60W10/184Conjoint control of vehicle sub-units of different type or different function including control of braking systems with wheel brakes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/18Propelling the vehicle
    • B60W30/18009Propelling the vehicle related to particular drive situations
    • B60W30/18109Braking
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/0097Predicting future conditions
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • B60W2050/143Alarm means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • B60W2050/146Display means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2510/00Input parameters relating to a particular sub-units
    • B60W2510/10Change speed gearings
    • B60W2510/1005Transmission ratio engaged
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2520/00Input parameters relating to overall vehicle dynamics
    • B60W2520/10Longitudinal speed
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2555/00Input parameters relating to exterior conditions, not covered by groups B60W2552/00, B60W2554/00
    • B60W2555/60Traffic rules, e.g. speed limits or right of way
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2556/00Input parameters relating to data
    • B60W2556/45External transmission of data to or from the vehicle
    • B60W2556/50External transmission of data to or from the vehicle of positioning data, e.g. GPS [Global Positioning System] data
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2720/00Output or target parameters relating to overall vehicle dynamics
    • B60W2720/10Longitudinal speed
    • B60W2720/106Longitudinal acceleration
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • B60W50/16Tactile feedback to the driver, e.g. vibration or force feedback to the driver on the steering wheel or the accelerator pedal
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60YINDEXING SCHEME RELATING TO ASPECTS CROSS-CUTTING VEHICLE TECHNOLOGY
    • B60Y2300/00Purposes or special features of road vehicle drive control systems
    • B60Y2300/18Propelling the vehicle
    • B60Y2300/18008Propelling the vehicle related to particular drive situations
    • B60Y2300/18158Approaching intersection

Definitions

  • Driver assistance systems can help a driver deal with an upcoming road hazard condition, whether it be an upcoming acute curve in the road or an accident that has occurred in a portion of the road in which the driver is driving towards.
  • Navigation warning systems alert the driver when various driving events on a segment of road the vehicle is traveling on are encountered.
  • Optical sensors are the dominant technology to detect driving events.
  • Color cameras are typically used to help detect a traffic sign on the roadside and to distinguish between different types of traffic signs, and a classification algorithm is typically used to recognize the printed speed on the sign.
  • optical sensor based zone warning inevitably suffers from adverse illumination and weather conditions when the assistance is needed most.
  • a method of detecting speed or no-passing zone warning using visual sensors suffers from several limitations.
  • the visual sensors can fail to detect signs in complex environment (e.g., downtown streets).
  • the visual sensors can also fail to detect signs because of different sign shape and location.
  • the visual sensors can also incorrectly recognize speeds because of misclassification at high speeds.
  • the visual sensors can also suffer from degraded detection/recognition at night, in rain or snow, when facing low angle sunlight (e.g., at dawn or dusk).
  • a driver assistance system includes a map database comprising a map database comprising navigation characteristics related to road locations, a GPS unit that receives location data of the vehicle, a map matching module configured to receive the location data of the vehicle and retrieve navigation characteristics relevant to the location data using a processing circuit, a prediction module configured to generate a most probable future path for the vehicle and to determine a location of at least one navigation characteristic with respect to the most probable future path and the vehicle, at least one vehicle sensor unit configured to generate vehicle data, and a warning module configured to transmit a control signal to a vehicle control area network to warn the driver of an upcoming navigation characteristic on the most probable path.
  • a driver assistance method includes receiving location data of the vehicle from a GPS unit, receiving the location data of the vehicle and retrieving navigation characteristics relevant to the location data using a processing circuit, generating a most probable future path for the vehicle and determining a location of at least one navigation characteristic with respect to the most probable future path and the vehicle, generating vehicle data at least one vehicle sensor, and transmitting a control signal to a vehicle control area network to warn the driver of an upcoming navigation characteristic on the most probable path.
  • FIG. 1 is a schematic diagram of a vehicle control area network
  • FIG. 2 is a schematic diagram of various vehicle system components and a general driver assistance system
  • FIG. 3 is a schematic diagram of a driver assistance system
  • FIG. 4 depicts a graphical representation of a generated path tree
  • FIG. 5 depicts a graphical representation of a future most probable path
  • FIG. 6 is a general flow chart of a method for producing a control signal
  • FIG. 7 is a flow chart of a method for detecting stop sign data and producing a control signal in response to the intersection data
  • FIG. 8 is a flow chart of a method for detecting slope distribution for the most probable path and producing a control signal based on the detected slope.
  • a driver assistance system includes a digital map system, vehicle sensor input, vision system input, location input, such as global positioning system (GPS) input, and various driver assistance modules used to make vehicle related determinations based on driver assistance system input.
  • the various driver assistance modules may be used to provide indicators or warnings to a vehicle passenger or may be used to send a control signal to a vehicle system component such as a vehicle engine control unit, or a vehicle steering control unit, for example, by communicating a control signal through a vehicle control area network (CAN).
  • a vehicle system component such as a vehicle engine control unit, or a vehicle steering control unit, for example, by communicating a control signal through a vehicle control area network (CAN).
  • CAN vehicle control area network
  • Vehicle communication network 100 is located within a vehicle body and allows various vehicle sensors including a radar sensor 108, a speed sensor and/or accelerometer 114, a vehicle vision system 120 which may include a stereovision camera and/or a monovision camera.
  • vehicle vision system 120 which may include a stereovision camera and/or a monovision camera.
  • communication network 100 receives vehicle location data from GPS module 118.
  • communication network 100 communicates with various vehicle control modules including brake control modules 110 and 112, gear control module 116, engine control module 122, and warning mechanism module 124, for example.
  • Central controller 102 includes at least one memory 104 and at least one processing unit 106. According to one exemplary embodiment vehicle
  • communication network 100 is a control area network (CAN) communication system and prioritizes communications in the network using a CAN bus.
  • CAN control area network
  • driver assistance system 220 is stored in the memory 104 of central controller 102 according to one embodiment.
  • Driver assistance system 220 includes a map matching module 210.
  • the map matching module 201 includes a map matching algorithm that receives vehicle location data (e.g., latitude, longitude, elevation, etc.) from the GPS unit 202.
  • vehicle location data e.g., latitude, longitude, elevation, etc.
  • the vehicle location data is enhanced and made more accurate by combining the GPS vehicle location data with vehicle sensor data from at least one vehicle sensor 204 at a positioning engine 206.
  • vehicle sensor data such as vision data, speed sensor data, and yaw rate data can be combined with GPS data at positioning engine 206 to reduce the set of coordinates that the vehicle may be located to improve the accuracy of the location data.
  • cameras 222 and 224 my be included in vehicle sensors 204 and positioning engine 206 may receive vision data from a camera 222, 224 that has been processed by a lane detection algorithm.
  • the lane detection software can modify the received GPS data to indicate that the vehicle is located in a specific lane rather than a general path or road.
  • other vehicle sensor data such as vision data, speed data, yaw rate data, etc. can be used to further supplement the GPS location data to improve the accuracy of the vehicle location.
  • Driver assistance system 220 also includes a map database 208 which includes navigation characteristics associated with pathways and roadways that may be traveled on by a vehicle.
  • the map database includes data not included in the GPS location data such as road elevations, road slopes, degrees of curvature of various road segments, the location of intersections, the location of stop signs, the location of traffic lights, no passing zone locations, yield sign locations, speed limits at various road locations, and various other navigation characteristics, for example.
  • the enhanced vehicle location is forwarded to map matching module 210.
  • the map matching algorithm uses the enhanced location of the vehicle from positioning engine 206 or raw location data from the GPS 202 to extract all navigation characteristics associated with the vehicle location.
  • the navigation characteristics extracted from map database 208 may be used for a variety of application algorithms to add to or enhance a vehicle's active or passive electronic safety systems.
  • the application algorithms may be executed alone (i.e., only used with the map data).
  • warning detection module 214 including a traffic signal warning algorithm, an intersection warning algorithm, a railroad crossing algorithm, a school zone warning algorithm, a slope warning algorithm, an exit ramp warning algorithm, and a lane change control algorithm.
  • each algorithm has various thresholds that are monitored to determine if a control signal is monitored. In some cases multiple algorithms are used to determine of a control signal should be transmitted.
  • control logic module 232 can include further algorithms to determine how various sensor inputs will cause CAN connected vehicle modules to actuate according to a control signal.
  • the application algorithms may be used to inform the driver directly via human machine interface (HMI) indicators (e.g., audible indicators, visual indicators, tactile indicators) or a combination of HMI indicators.
  • HMI human machine interface
  • an audible indicator may alert a driver with a audible sound or message in the case that the speed limit warning algorithm determines the vehicle speed is above a speed limit or is about to exceed a speed limit threshold.
  • visual indicators may use a display such as an LCD screen or LED light to indicate a warning message and tactile indicators may use a vibration element in a vehicle steering wheel, for example, to alert the driver to a warning message output from the warning determination module 214.
  • the application algorithms may also be provided to a vehicle control module 238 to send a control signal to various vehicle actuators 110, 112, 116, and 122 for example, to directly change how the vehicle operates without human intervention. Additionally, a vehicle driver may be able to decide if they would like to allow vehicle control module 238 to automatically control vehicle modules or not based on the position of switch 270.
  • the driver assistance system 220 is used to provide a slope distribution warning or a stop sign warning.
  • the warning determination module 214 sends a control signal to CAN system 240 to convey a warning indication to driver of the vehicle via an HMI.
  • the HMI warning may also be based on known intersections, railroad crossings, school zones, road elevation levels, road lanes, and traffic signal coordinates stored in map database 208 for various geographic locations and provides reliable warnings in all illumination and environmental conditions.
  • GPS unit 320 provides the current vehicle location to positioning engine or dead reckoning module 350.
  • Module 350 also receives the vehicle speed from sensor 340, if available, the yaw rate of the vehicle from angular rate sensors, if available, and acceleration sensors (accelerometers, not shown), if available, at positioning engine 350 in order to calculate position with better accuracy and produce a higher update rate for map matching module 360, look ahead module 328, and most probable path build 390.
  • the resulting fused position map from module 350 allows the driver assistance system 220 to predict vehicle position points for more accurate vehicle route data.
  • the GPS and inertial fusion has the benefits of: 1) helping to eliminate GPS multipath and loss of signal in urban canyons, 2) providing significantly better dead reckoning when GPS signal is temporarily unavailable, especially while maneuvering, 3) providing mutual validation between GPS and inertial sensors, and 4) allows the accurate measurement of instantaneous host vehicle behavior due to high sample rate and relative accuracy of the inertial sensors 330, 340.
  • the driver assistance system 220 can handle GPS update rates of 5 Hz or greater.
  • map matching data produced at map matching module 360 provides an output location of a vehicle with respect to a road and navigation characteristics associated with the road.
  • the stereo vision or monocular vision system provides the forward looking image of the road environment.
  • Such vision system data may be provided directly to map matching module 360 or may be provided at a later step from sensor module 310, for example.
  • a lane detection and tracking algorithm using the stereo vision or monocular vision system calculates host lane position and lane horizontal curvature.
  • the stereo vision system can also calculate a 3D lane profile including vertical curvature, incline/decline angle, and bank angle information.
  • map matching module 360 may be performed at map matching module 360 or may alternatively be performed at various other modules including look ahead module 328, probable path module 390, slop distribution calculation module 32, distance calculation module 324, prediction module 212, fusion module 218, control logic module 232, or warning determination module 214, for example.
  • prediction module 200 as shown in FIG. 2 look ahead module 328 and probable path module 390 as shown in more detailed FIG. 3.
  • prediction module 200 receives the output of map matching module 210 to generate a path tree 400 comprising a set of forward paths or roads the vehicle 402 can take such as the path between node 420 and node 426 and the current path the vehicle 402 is on as shown in FIG. 4.
  • a most probable future path 500 of the vehicle 514 is generated based on the generated path tree, the vehicle data, and the navigation characteristics.
  • the look ahead module 328 organizes the links in a hierarchical fashion, providing quick access to link features important in path prediction, such as intersecting angles and travel direction.
  • Details of output of the map matching unit 360 that are provided to the most probable path building unit 390 according to one or more embodiments is described below.
  • the map matching unit 360 matches the GPS-processed position of the vehicle output by the GPS processing unit 350 (which takes into account the inertial sensor data as provided by the sensors 330, 340) to a position on a map in single path and branching road geometry scenarios.
  • map matching unit 360 provides navigation characteristics, as obtained from the map database 370 to various locations relevant to a vehicle.
  • a GPS position is used as an input to a look up table or software algorithm which is used to retrieve navigation characteristics stored in map database 370.
  • the map matching unit 360 finds the position on the map that is closest to the corrected GPS position provided by module 350, whereby this filtering to find the closest map position using an error vector based on the last time epoch.
  • GPS heading angle and history weights can used by the map matching unit 360 in some embodiments to eliminate irrelevant road links.
  • Map matching as performed by the map matching unit 360 can also utilize information regarding the vehicle's intention (e.g., it's destination), if available, and also the vehicle trajectory. In some embodiments, map matching can be performed by reducing history weight near branching (e.g., a first road intersection with a second road), and by keeping connectivity alive for a few seconds after branching.
  • the most probable path unit 390 uses the map-matched position as output by the map matching unit 360 as a reference to look ahead of the host vehicle position, extracts the possible road links, and constructs a MPP (Most Probable Path) from the extracted road links.
  • MPP Mobile Packet Probable Path
  • the MPP construction can be affected by the host vehicle speed.
  • angles between the connected branches making up the MPP are computed and are used with other attributes to determine the 'n' Most Probable Paths.
  • a path list is then constructed using the 'n' MPPs, whereby vehicle status signals as output by the vehicle status signals unit 310 can be used in the selection of the MPPs.
  • a vehicle imaging system can also be utilized in some embodiments to assist in the selection of the MPPs.
  • FIG. 4 is a diagrammatic representation of the n MPPs that can be output by the most probable path of a vehicle 402, as shown by way of path tree 400 with the various possible paths shown as branches of the tree 400.
  • the path between nodes 420 and 426 as well as the path between 420 and 422 are both possible future paths while subsection 450 between the vehicle location 402 and node 420 is the path tree root.
  • the various nodes on the generated path tree 400 are associated with navigation characteristics retrieved from the map database 370 such as road curve data, stop sign data, road elevation and slope data, and no passing zone data that may be used to determine if a control signal should be transmitted from the warning determination module 214 or the vehicle control module 238.
  • map database 370 may be used at map matching module 360 to identify certain nodes as having particular slope values in comparison with an adjacent node.
  • the MPP slope distribution calculation unit 324 and distance calculation unit 326 also can be made on one or more of the n MPPs output by the most probable path unit 390. Time and distance calculations can be performed on one or more of the MPPs output by the most probable path unit 390. In some embodiments, time and distance are calculated using vehicle speed and intermodal distances 502, 504, 522, and 524 as shown in Fig. 5 on a determined MPP 500. [0035] Furthermore, if vehicle 514 is traveling at a speed of 70 m.p.h. and based on distance 502, time and distance calculation module 326 will be able to determine how long it will take for vehicle 514 to enter zone 504 with a slope determined at module 324.
  • warning determination module 342 will send a control signal to at least one of an HMI indicator or a vehicle module. Similar calculations may be undertake to warn a driver or control a vehicle module if a stop sign is with a predetermined distance from the vehicle.
  • control signal may communicate a required deceleration to bring the determined threshold violation under the threshold speed value.
  • This required deceleration may be provided to a break control module 112 or engine control module 122, for example, to remove the determined threat.
  • warning determination module 214 may transmit a control signal to an HMI to convey a warning to a vehicle passenger if one of several thresholds is exceeded.
  • Each algorithm included in warning determination module 214 may have one or more thresholds that are monitored. For example, if the current vehicle speed is over the
  • DOT Department of Transportation
  • a control signal may be transmitted from module 214 to a CAN system 240 to be provided to an HMI.
  • warning control module 214 may use various vehicle data collected by vehicle sensors 204 including camera and radar input to calculate the distance and time to an upcoming curve, which, together with the targeted speed, can be provided to the an automatic control module 232 to produce a vehicle control signal at vehicle control module 238 to automatically adjust vehicle speed/deceleration for optimal fuel efficiency without human intervention.
  • Such automatic adjustments may be transmitted as control signals from vehicle control module 238 and provided to a CAN system 240 which distributes the control signal to an appropriate vehicle module such as an engine control module 122 or a brake control module 110, 112.
  • the driver assistance system 220 can accurately inform the operator of the vehicle 105 with suitable lead time about an upcoming road condition such as a declining or inclining slope that may pose a hazard or cause an undesirable reduction in fuel efficiency.
  • the driver assistance system 220 can warn the driver if the vehicle is moving too fast, whereby the driver assistance system can provide warnings through a HMI prior to entering a high slope or low slope zone thereby improving on previous warning systems and methods.
  • Process 600 may be carried out by several different driver assistance system embodiments 200 or 300 and may be a computer program stored in the memory 104 of central controller 104 and executed by at least one processor 106 in central controller 102.
  • Process 600 is merely exemplary and may include additional steps or may not include one or more steps displayed in FIG. 6.
  • driver assistance system 220 receives raw GPS data from GPS unit 202.
  • this raw GPS data may be enhanced at positioning engine 206 or dead reckoning module 350, for example.
  • the positioning engine improves the accuracy of raw GPS data provided by GPS unit 202 using vehicle sensor data 204 received at step 622 including data from camera units 222 and 224 as well as from other sensors such as an accelerometer, a vehicle speed sensor 340, or a vision system/ lane detection software sensor 330.
  • the vehicle location data which may comprise a set of coordinates, such as longitude and latitude, is provided to a map matching algorithm stored in map matching module 210 for example.
  • the map matching algorithm uses the vehicle position coordinates as a reference to look up navigation characteristics associated with the position coordinates in map database 208. For example, a given coordinate may have an associated elevation above sea level, slope value, road curve measurement, lane data, stop sign presence, no passing zone presence, or speed limit for example.
  • step 604 generates a series of relevant location coordinates within a road that are associated with various navigation characteristics, this data is provided to prediction module 212 to generate a path tree at step 606 and a most probable path at step 608.
  • the most probable path is segmented into a series of nodes, each of which are may be associated with a speed zone and/or a no passing zone as determined by navigation characteristics retrieved from map database 208.
  • prediction module 212 may calculate time and distance data for future nodes 510, 512 on the most probable path 500 at step 612 based on vehicle speed and/or lane detection data received at step 610.
  • the most probable path and associated navigation characteristics such as intersection locations, exit ramp locations, slope data, or school zones, for example, may then be provided to several other driver assistance modules 218, 232, 234, and 214 for further calculations or processing.
  • the most probable path and exit ramp locations are transmitted to warning determination module 214 and entered as input to an exit ramp warning algorithm.
  • FIG. 7 depicts one exemplary embodiment of a process carried out by as stop sign warning algorithm.
  • the zone warning algorithm will analyze the most probable path data and compare the vehicles speed or lane data with a threshold value associated with a most probable path node 506, 508, 510, and 512, for example.
  • process 600 determines if at least one or more thresholds for a given node have been exceeded. According to one embodiment, if a threshold value has been exceeded warning determination module 214 provides a control signal to CAN system 240, which in turn actuates an HMI to provide a warning or other indication to a vehicle passenger that a dangerous condition is approaching along the most probable path at step 620.
  • step 620 may take place at control logic module 232, eco optimization module 234, or vehicle control module 238 with additional algorithms providing various threshold determinations.
  • vehicle control module 238 may receive the most probable path data from prediction module 212 and determine based on a gear algorithm or braking algorithm whether to actuate a gear control module 116 or brake module 1 10, 112 by providing a control signal to CAN system 240.
  • FIG. 7 and FIG. 8 show processes carried out by various application algorithms stored in warning determination module 214 are shown. Referring to FIG. 7, a process for detecting stop signs and providing a warning to a driver or a control signal to a vehicle module in response to detecting the stop sign.
  • the driver assistance system may prevent or reduce the likelihood of accidents and intentional stop sign rolling.
  • the digital map system may identify stop signs locations that are in the vehicle path by mapping the vehicle location with the GPS device.
  • the driver assistance system may include electronics configured to combine the vehicle position with one or more of the vehicle speed, data from angular rate sensors (e.g. yaw rate) and acceleration sensors (e.g., accelerometers) to calculate position with better accuracy and a higher update rate.
  • the resulting vehicle position may be matched to a map using the digital map system.
  • the map includes stop sign attributes (e.g., stop sign identification, map location, etc.).
  • a distance to the upcoming stop sign(s) may be estimated, for example with a Kalman filtering technique.
  • a Kalman filtering technique advantageously provides accurate distance measurements from noisy GPS data. Also, because of the vehicle speed information, the aforementioned technique may be used even in the absence of a GPS signal.
  • the driver assistance system may also combine the calculated stop sign position with data from the vision system to more precisely recognize the stop sign.
  • a warning may be issued to driver ahead of the stop sign based on the vehicle speed/location.
  • the driver assistance system may also generate and/or execute a control algorithm to control the vehicle speed. Specifically, at step 702 in process 700 it is determined if a stop sign is on the future most probable path, such as path 500.
  • vision system data may be analyzed to confirm that a stop sign is present using object detection software, for example.
  • module 326 may determine the distance to the stop sign from the vehicle.
  • step 710 determines whether a speed threshold associated with the distance determined at step 708 has been exceeded. If the speed threshold has been exceeded, a control signal is transmitted to an HMI to alert the driver of the unsafe speed in view of the distance between the vehicle and the stop sign.
  • process 800 provides advance knowledge of roadway and terrain variations that may also be beneficial for autonomous vehicle functions.
  • slope distribution data for the most probable path determined at step 608 of process 600 as shown in FIG. 6 is retrieved from the map database 208 at various nodes 506, 508, 510, and 512 along the most probable path 500 at step 802.
  • slopes associated with more than one segment are added and averaged to determine a future slope distribution.
  • the slope distribution predicted to be encountered by vehicle 514 is compared with a speed threshold associated with a particular range of slope distributions.
  • the slope threshold is variable and depends on input factors such as vehicle location data or vehicle sensor input data.
  • a control signal may be sent to a vehicle module at step 808 or an HMI at step 810.
  • knowledge of extended downhill slopes allows hybrid vehicles to utilize regenerative braking to reenergize battery capacity.
  • Advance knowledge of problematic intersections allows vehicles to pre -prime braking pressure in silent anticipation of cross-traffic collision.
  • the map system, vision system, and GPS device of the driver assistance system can be used together to advise the driver regarding lane changes in order to minimize braking.
  • the driver assistance system 220 may provide the driver with lane change advice while nearing an exit ramp so that the vehicle has a smooth transition from high to low speed with minimal braking.
  • the lane change advice may be shown in an HMI display and be determined by an exit ramp algorithm stored in warning determination module 214.
  • the driver assistance system may use data from the digital map system, vision system and GPS device to generate and execute an algorithm to provide lane change recommendations and vehicle speed profiles to the driver.
  • the driver assistance system may also generate and execute a control algorithm for controlling the vehicle speed and steering angle.
  • the driver assistance system 220 may assist in improving gas mileage of the vehicle and aid in reducing gas consumption.
  • the driver assistance system may assist in optimal braking to increase the life of brakes/vehicle by providing a control signal to eco- optimization module 234, for example.
  • the driver assistance system may assist in avoiding last minute exit situations and thus reduce risk while driving.
  • the driver assistance system may provide optional speed information based on the vehicle parameters and road
  • the driver assistance system may assist in driver training for an optimal driving style.
  • the driver assistance system may assist in reducing insurance costs.
  • Exemplary embodiments may include program products comprising computer or machine-readable media for carrying or having machine-executable instructions or data structures stored thereon.
  • the driver monitoring system may be computer driven.
  • Exemplary embodiments illustrated in the methods of the figures may be controlled by program products comprising computer or machine-readable media for carrying or having machine-executable instructions or data structures stored thereon.
  • Such computer or machine-readable media can be any available media which can be accessed by a general purpose or special purpose computer or other machine with a processor.
  • Such computer or machine-readable media can comprise RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor.
  • Computer or machine-executable instructions comprise, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing machines to perform a certain function or group of functions.
  • Software implementations of the present invention could be accomplished with standard programming techniques with rule based logic and other logic to accomplish the various connection steps, processing steps, comparison steps and decision steps.
  • elements shown as integrally formed may be constructed of multiple parts or elements shown as multiple parts may be integrally formed, the operation of the assemblies may be reversed or otherwise varied, the length or width of the structures and/or members or connectors or other elements of the system may be varied, the nature or number of adjustment or attachment positions provided between the elements may be varied.
  • the elements and/or assemblies of the system may be constructed from any of a wide variety of materials that provide sufficient strength or durability. Accordingly, all such modifications are intended to be included within the scope of the present disclosure.
  • the order or sequence of any process or method steps may be varied or re-sequenced according to alternative embodiments.
  • Other substitutions, modifications, changes and omissions may be made in the design, operating conditions and arrangement of the preferred and other exemplary embodiments without departing from the spirit of the present subject matter

Landscapes

  • Engineering & Computer Science (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Automation & Control Theory (AREA)
  • Chemical & Material Sciences (AREA)
  • Combustion & Propulsion (AREA)
  • Human Computer Interaction (AREA)
  • Traffic Control Systems (AREA)
  • Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
  • Navigation (AREA)
  • Position Fixing By Use Of Radio Waves (AREA)

Abstract

A system and method of assisting a driver of a vehicle by providing driver and vehicle feedback control signals is disclosed. The system and method includes receiving location data of the vehicle from a GPS unit, receiving the location data of the vehicle and retrieving navigation characteristics relevant to the location data using a processing circuit, generating a most probable future path for the vehicle and determining a location of at least one navigation characteristic with respect to the most probable future path and the vehicle, generating vehicle data at least one vehicle sensor, and transmitting a control signal to a vehicle control area network to warn the driver of an upcoming navigation characteristic on the most probable path.

Description

DRIVER ASSISTANCE SYSTEM
CROSS-REFERENCE TO RELATED PATENT APPLICATIONS
[0001] This application claims priority from Provisional Application US Application 61/466,870, filed 3/23/2011, incorporated herein by reference in its entirety. This application also claims priority from Provisional Application US Application 61/466,873, filed
3/23/2011, incorporated herein by reference in its entirety. This application also claims priority from Provisional Application US Application 61/466,880, filed 3/23/2011, incorporated herein by reference in its entirety.
BACKGROUND OF THE INVENTION
[0002] Driver assistance systems are becoming more and more prevalent in vehicles.
Driver assistance systems can help a driver deal with an upcoming road hazard condition, whether it be an upcoming acute curve in the road or an accident that has occurred in a portion of the road in which the driver is driving towards.
[0003] Navigation warning systems alert the driver when various driving events on a segment of road the vehicle is traveling on are encountered. Optical sensors are the dominant technology to detect driving events. Color cameras are typically used to help detect a traffic sign on the roadside and to distinguish between different types of traffic signs, and a classification algorithm is typically used to recognize the printed speed on the sign.
[0004] Like most vision systems, optical sensor based zone warning inevitably suffers from adverse illumination and weather conditions when the assistance is needed most. A method of detecting speed or no-passing zone warning using visual sensors suffers from several limitations. The visual sensors can fail to detect signs in complex environment (e.g., downtown streets). The visual sensors can also fail to detect signs because of different sign shape and location. The visual sensors can also incorrectly recognize speeds because of misclassification at high speeds. The visual sensors can also suffer from degraded detection/recognition at night, in rain or snow, when facing low angle sunlight (e.g., at dawn or dusk).
SUMMARY OF THE INVENTION
[0005] According to an exemplary embodiment, a driver assistance system includes a map database comprising a map database comprising navigation characteristics related to road locations, a GPS unit that receives location data of the vehicle, a map matching module configured to receive the location data of the vehicle and retrieve navigation characteristics relevant to the location data using a processing circuit, a prediction module configured to generate a most probable future path for the vehicle and to determine a location of at least one navigation characteristic with respect to the most probable future path and the vehicle, at least one vehicle sensor unit configured to generate vehicle data, and a warning module configured to transmit a control signal to a vehicle control area network to warn the driver of an upcoming navigation characteristic on the most probable path.
[0006] According to yet another exemplary embodiment, a driver assistance method includes receiving location data of the vehicle from a GPS unit, receiving the location data of the vehicle and retrieving navigation characteristics relevant to the location data using a processing circuit, generating a most probable future path for the vehicle and determining a location of at least one navigation characteristic with respect to the most probable future path and the vehicle, generating vehicle data at least one vehicle sensor, and transmitting a control signal to a vehicle control area network to warn the driver of an upcoming navigation characteristic on the most probable path.
BRIEF DESCRIPTION OF THE DRAWINGS
[0007] These and other features, aspects, and advantages of the present invention will become apparent from the following description, appended claims, and the accompanying exemplary embodiments shown in the drawings, which are briefly described below.
[0008] FIG. 1 is a schematic diagram of a vehicle control area network;
[0009] FIG. 2 is a schematic diagram of various vehicle system components and a general driver assistance system;
[0010] FIG. 3 is a schematic diagram of a driver assistance system;
[0011] FIG. 4 depicts a graphical representation of a generated path tree;
[0012] FIG. 5 depicts a graphical representation of a future most probable path
determination;
[0013] FIG. 6 is a general flow chart of a method for producing a control signal;
[0014] FIG. 7 is a flow chart of a method for detecting stop sign data and producing a control signal in response to the intersection data; and
[0015] FIG. 8 is a flow chart of a method for detecting slope distribution for the most probable path and producing a control signal based on the detected slope.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
[0016] Before describing in detail the particular improved system and method, it should be observed that the several disclosed embodiments include, but are not limited to a novel structural combination of conventional data and/or signal processing components and communications circuits, and not in the particular detailed configurations thereof.
Accordingly, the structure, methods, functions, control and arrangement of conventional components and circuits have, for the most part, been illustrated in the drawings by readily understandable block representations and schematic diagrams, in order not to obscure the disclosure with structural details which will be readily apparent to those skilled in the art, having the benefit of the description herein. Further, the disclosed embodiments are not limited to the particular embodiments depicted in the exemplary diagrams, but should be construed in accordance with the language in the claims.
[0017] In general, according to various exemplary embodiments, a driver assistance system includes a digital map system, vehicle sensor input, vision system input, location input, such as global positioning system (GPS) input, and various driver assistance modules used to make vehicle related determinations based on driver assistance system input. The various driver assistance modules may be used to provide indicators or warnings to a vehicle passenger or may be used to send a control signal to a vehicle system component such as a vehicle engine control unit, or a vehicle steering control unit, for example, by communicating a control signal through a vehicle control area network (CAN).
[0018] Referring to FIG. 1, a block diagram of a vehicle communication network 100 is shown, according to an exemplary embodiment. Vehicle communication network 100 is located within a vehicle body and allows various vehicle sensors including a radar sensor 108, a speed sensor and/or accelerometer 114, a vehicle vision system 120 which may include a stereovision camera and/or a monovision camera. In addition, communication network 100 receives vehicle location data from GPS module 118. Furthermore, communication network 100 communicates with various vehicle control modules including brake control modules 110 and 112, gear control module 116, engine control module 122, and warning mechanism module 124, for example. Central controller 102 includes at least one memory 104 and at least one processing unit 106. According to one exemplary embodiment vehicle
communication network 100 is a control area network (CAN) communication system and prioritizes communications in the network using a CAN bus.
[0019] Referring now to FIG. 2, driver assistance system 220 is stored in the memory 104 of central controller 102 according to one embodiment. Driver assistance system 220 includes a map matching module 210. The map matching module 201 includes a map matching algorithm that receives vehicle location data (e.g., latitude, longitude, elevation, etc.) from the GPS unit 202. According to one embodiment, the vehicle location data is enhanced and made more accurate by combining the GPS vehicle location data with vehicle sensor data from at least one vehicle sensor 204 at a positioning engine 206.
[0020] According to one exemplary embodiment, vehicle sensor data such as vision data, speed sensor data, and yaw rate data can be combined with GPS data at positioning engine 206 to reduce the set of coordinates that the vehicle may be located to improve the accuracy of the location data. For example, cameras 222 and 224 my be included in vehicle sensors 204 and positioning engine 206 may receive vision data from a camera 222, 224 that has been processed by a lane detection algorithm. According to one embodiment, the lane detection software can modify the received GPS data to indicate that the vehicle is located in a specific lane rather than a general path or road. In addition, other vehicle sensor data such as vision data, speed data, yaw rate data, etc. can be used to further supplement the GPS location data to improve the accuracy of the vehicle location. [0021] Driver assistance system 220 also includes a map database 208 which includes navigation characteristics associated with pathways and roadways that may be traveled on by a vehicle. According to one embodiment, the map database includes data not included in the GPS location data such as road elevations, road slopes, degrees of curvature of various road segments, the location of intersections, the location of stop signs, the location of traffic lights, no passing zone locations, yield sign locations, speed limits at various road locations, and various other navigation characteristics, for example.
[0022] According to one exemplary embodiment, once the positioning engine 206 has determined an enhanced location of the vehicle, the enhanced vehicle location is forwarded to map matching module 210. The map matching algorithm uses the enhanced location of the vehicle from positioning engine 206 or raw location data from the GPS 202 to extract all navigation characteristics associated with the vehicle location. The navigation characteristics extracted from map database 208 may be used for a variety of application algorithms to add to or enhance a vehicle's active or passive electronic safety systems. The application algorithms may be executed alone (i.e., only used with the map data). Several application algorithms are shown in warning detection module 214 including a traffic signal warning algorithm, an intersection warning algorithm, a railroad crossing algorithm, a school zone warning algorithm, a slope warning algorithm, an exit ramp warning algorithm, and a lane change control algorithm. According to some embodiments, each algorithm has various thresholds that are monitored to determine if a control signal is monitored. In some cases multiple algorithms are used to determine of a control signal should be transmitted.
Furthermore, several algorithms are shown in flow chart form in FIG. 6-8. These application algorithms may also be executed in connection with a variety of vehicle sensors such as RADAR 226, LIDAR 228, monocular vision 224, stereo vision 224, and various other vehicle sensors 204 to add further functionality. Furthermore, control logic module 232 can include further algorithms to determine how various sensor inputs will cause CAN connected vehicle modules to actuate according to a control signal.
[0023] According to one exemplary embodiment, the application algorithms may be used to inform the driver directly via human machine interface (HMI) indicators (e.g., audible indicators, visual indicators, tactile indicators) or a combination of HMI indicators. For example, an audible indicator may alert a driver with a audible sound or message in the case that the speed limit warning algorithm determines the vehicle speed is above a speed limit or is about to exceed a speed limit threshold. In a similar manner, visual indicators may use a display such as an LCD screen or LED light to indicate a warning message and tactile indicators may use a vibration element in a vehicle steering wheel, for example, to alert the driver to a warning message output from the warning determination module 214.
Furthermore, the application algorithms may also be provided to a vehicle control module 238 to send a control signal to various vehicle actuators 110, 112, 116, and 122 for example, to directly change how the vehicle operates without human intervention. Additionally, a vehicle driver may be able to decide if they would like to allow vehicle control module 238 to automatically control vehicle modules or not based on the position of switch 270.
[0024] In one embodiment of the present disclosure, the driver assistance system 220 is used to provide a slope distribution warning or a stop sign warning. According to some embodiments, when a current or predicted vehicle speed is above a threshold speed and the vehicle is a predetermined distance from a stop sign on the road the vehicle is traveling on or is predicted to travel on, the warning determination module 214 sends a control signal to CAN system 240 to convey a warning indication to driver of the vehicle via an HMI.
According to one exemplary embodiment, the HMI warning may also be based on known intersections, railroad crossings, school zones, road elevation levels, road lanes, and traffic signal coordinates stored in map database 208 for various geographic locations and provides reliable warnings in all illumination and environmental conditions.
[0025] According to one embodiment as shown in FIG. 3, GPS unit 320 provides the current vehicle location to positioning engine or dead reckoning module 350. Module 350 also receives the vehicle speed from sensor 340, if available, the yaw rate of the vehicle from angular rate sensors, if available, and acceleration sensors (accelerometers, not shown), if available, at positioning engine 350 in order to calculate position with better accuracy and produce a higher update rate for map matching module 360, look ahead module 328, and most probable path build 390.
[0026] The resulting fused position map from module 350 allows the driver assistance system 220 to predict vehicle position points for more accurate vehicle route data. The GPS and inertial fusion has the benefits of: 1) helping to eliminate GPS multipath and loss of signal in urban canyons, 2) providing significantly better dead reckoning when GPS signal is temporarily unavailable, especially while maneuvering, 3) providing mutual validation between GPS and inertial sensors, and 4) allows the accurate measurement of instantaneous host vehicle behavior due to high sample rate and relative accuracy of the inertial sensors 330, 340. By way of example, the driver assistance system 220 can handle GPS update rates of 5 Hz or greater.
[0027] Referring again to FIG. 3, map matching data produced at map matching module 360, provides an output location of a vehicle with respect to a road and navigation characteristics associated with the road. In addition, the stereo vision or monocular vision system provides the forward looking image of the road environment. Such vision system data may be provided directly to map matching module 360 or may be provided at a later step from sensor module 310, for example. A lane detection and tracking algorithm using the stereo vision or monocular vision system calculates host lane position and lane horizontal curvature. The stereo vision system can also calculate a 3D lane profile including vertical curvature, incline/decline angle, and bank angle information. These calculations may be performed at map matching module 360 or may alternatively be performed at various other modules including look ahead module 328, probable path module 390, slop distribution calculation module 32, distance calculation module 324, prediction module 212, fusion module 218, control logic module 232, or warning determination module 214, for example.
[0028] According to one embodiment, prediction module 200 as shown in FIG. 2 look ahead module 328 and probable path module 390 as shown in more detailed FIG. 3.
Accordingly, prediction module 200 receives the output of map matching module 210 to generate a path tree 400 comprising a set of forward paths or roads the vehicle 402 can take such as the path between node 420 and node 426 and the current path the vehicle 402 is on as shown in FIG. 4.
[0029] Once path tree 400 has been generated, a most probable future path 500 of the vehicle 514 is generated based on the generated path tree, the vehicle data, and the navigation characteristics. In some embodiments, the look ahead module 328 organizes the links in a hierarchical fashion, providing quick access to link features important in path prediction, such as intersecting angles and travel direction. [0030] Details of output of the map matching unit 360 that are provided to the most probable path building unit 390 according to one or more embodiments is described below. The map matching unit 360 matches the GPS-processed position of the vehicle output by the GPS processing unit 350 (which takes into account the inertial sensor data as provided by the sensors 330, 340) to a position on a map in single path and branching road geometry scenarios. In this way, map matching unit 360 provides navigation characteristics, as obtained from the map database 370 to various locations relevant to a vehicle. According to one example, a GPS position is used as an input to a look up table or software algorithm which is used to retrieve navigation characteristics stored in map database 370.
[0031] Furthermore, the map matching unit 360 finds the position on the map that is closest to the corrected GPS position provided by module 350, whereby this filtering to find the closest map position using an error vector based on the last time epoch. GPS heading angle and history weights can used by the map matching unit 360 in some embodiments to eliminate irrelevant road links. Map matching as performed by the map matching unit 360 can also utilize information regarding the vehicle's intention (e.g., it's destination), if available, and also the vehicle trajectory. In some embodiments, map matching can be performed by reducing history weight near branching (e.g., a first road intersection with a second road), and by keeping connectivity alive for a few seconds after branching.
[0032] Details of the operation of the most probable path unit 390 according to one or more embodiments is described below. The most probable path unit 390 uses the map-matched position as output by the map matching unit 360 as a reference to look ahead of the host vehicle position, extracts the possible road links, and constructs a MPP (Most Probable Path) from the extracted road links. The MPP construction can be affected by the host vehicle speed. Also, angles between the connected branches making up the MPP are computed and are used with other attributes to determine the 'n' Most Probable Paths. A path list is then constructed using the 'n' MPPs, whereby vehicle status signals as output by the vehicle status signals unit 310 can be used in the selection of the MPPs. Further, a vehicle imaging system can also be utilized in some embodiments to assist in the selection of the MPPs.
[0033] FIG. 4 is a diagrammatic representation of the n MPPs that can be output by the most probable path of a vehicle 402, as shown by way of path tree 400 with the various possible paths shown as branches of the tree 400. For example, the path between nodes 420 and 426 as well as the path between 420 and 422 are both possible future paths while subsection 450 between the vehicle location 402 and node 420 is the path tree root.
According to one exemplary embodiment the various nodes on the generated path tree 400 are associated with navigation characteristics retrieved from the map database 370 such as road curve data, stop sign data, road elevation and slope data, and no passing zone data that may be used to determine if a control signal should be transmitted from the warning determination module 214 or the vehicle control module 238. In addition, map database 370 may be used at map matching module 360 to identify certain nodes as having particular slope values in comparison with an adjacent node.
[0034] As shown in FIG. 3, the MPP slope distribution calculation unit 324 and distance calculation unit 326 also can be made on one or more of the n MPPs output by the most probable path unit 390. Time and distance calculations can be performed on one or more of the MPPs output by the most probable path unit 390. In some embodiments, time and distance are calculated using vehicle speed and intermodal distances 502, 504, 522, and 524 as shown in Fig. 5 on a determined MPP 500. [0035] Furthermore, if vehicle 514 is traveling at a speed of 70 m.p.h. and based on distance 502, time and distance calculation module 326 will be able to determine how long it will take for vehicle 514 to enter zone 504 with a slope determined at module 324.
According to one example if the speed of the vehicle is above the speed threshold determined by the determined slope of zone 504 and the time until a vehicle reaches a zone is under a time threshold, warning determination module 342 will send a control signal to at least one of an HMI indicator or a vehicle module. Similar calculations may be undertake to warn a driver or control a vehicle module if a stop sign is with a predetermined distance from the vehicle.
[0036] In addition the control signal may communicate a required deceleration to bring the determined threshold violation under the threshold speed value. This required deceleration may be provided to a break control module 112 or engine control module 122, for example, to remove the determined threat.
[0037] Furthermore, warning determination module 214 may transmit a control signal to an HMI to convey a warning to a vehicle passenger if one of several thresholds is exceeded. Each algorithm included in warning determination module 214 may have one or more thresholds that are monitored. For example, if the current vehicle speed is over the
Department of Transportation (DOT) recommended safe speed for the current road curvature and bank angle as determined by a curve speed warning algorithm, or over the posted warning speed of this curve or if a predicted future vehicle speed is over the DOT
recommended safe speed for the upcoming lane curvature and bank angle (or over the posted warning speed of this upcoming curve) that the host vehicle is about to enter in a predefined time threshold (e.g., 10 seconds), a control signal may be transmitted from module 214 to a CAN system 240 to be provided to an HMI.
[0038] Additionally, the algorithms depicted in warning control module 214 may use various vehicle data collected by vehicle sensors 204 including camera and radar input to calculate the distance and time to an upcoming curve, which, together with the targeted speed, can be provided to the an automatic control module 232 to produce a vehicle control signal at vehicle control module 238 to automatically adjust vehicle speed/deceleration for optimal fuel efficiency without human intervention. Such automatic adjustments may be transmitted as control signals from vehicle control module 238 and provided to a CAN system 240 which distributes the control signal to an appropriate vehicle module such as an engine control module 122 or a brake control module 110, 112.
[0039] Based on the road path information as provided by the GPS 202 and the most probable future path as determined by the prediction module 212, the driver assistance system 220 can accurately inform the operator of the vehicle 105 with suitable lead time about an upcoming road condition such as a declining or inclining slope that may pose a hazard or cause an undesirable reduction in fuel efficiency. The driver assistance system 220, according to an embodiment of the invention, can warn the driver if the vehicle is moving too fast, whereby the driver assistance system can provide warnings through a HMI prior to entering a high slope or low slope zone thereby improving on previous warning systems and methods.
[0040] Referring to FIG. 6, a general flow chart of a method for producing a curve related control signal is disclosed. Process 600 may be carried out by several different driver assistance system embodiments 200 or 300 and may be a computer program stored in the memory 104 of central controller 104 and executed by at least one processor 106 in central controller 102. Process 600 is merely exemplary and may include additional steps or may not include one or more steps displayed in FIG. 6. According to one exemplary embodiment, at step 602 driver assistance system 220 receives raw GPS data from GPS unit 202. According to one embodiment, this raw GPS data may be enhanced at positioning engine 206 or dead reckoning module 350, for example. As stated previously, the positioning engine improves the accuracy of raw GPS data provided by GPS unit 202 using vehicle sensor data 204 received at step 622 including data from camera units 222 and 224 as well as from other sensors such as an accelerometer, a vehicle speed sensor 340, or a vision system/ lane detection software sensor 330.
[0041] Once vehicle location data or enhanced vehicle data is determined at step 602, the vehicle location data, which may comprise a set of coordinates, such as longitude and latitude, is provided to a map matching algorithm stored in map matching module 210 for example. According to one embodiment, the map matching algorithm uses the vehicle position coordinates as a reference to look up navigation characteristics associated with the position coordinates in map database 208. For example, a given coordinate may have an associated elevation above sea level, slope value, road curve measurement, lane data, stop sign presence, no passing zone presence, or speed limit for example. Once step 604 generates a series of relevant location coordinates within a road that are associated with various navigation characteristics, this data is provided to prediction module 212 to generate a path tree at step 606 and a most probable path at step 608. According to one embodiment the most probable path is segmented into a series of nodes, each of which are may be associated with a speed zone and/or a no passing zone as determined by navigation characteristics retrieved from map database 208. According to another embodiment, prediction module 212 may calculate time and distance data for future nodes 510, 512 on the most probable path 500 at step 612 based on vehicle speed and/or lane detection data received at step 610.
[0042] The most probable path and associated navigation characteristics such as intersection locations, exit ramp locations, slope data, or school zones, for example, may then be provided to several other driver assistance modules 218, 232, 234, and 214 for further calculations or processing. According to one embodiment, the most probable path and exit ramp locations are transmitted to warning determination module 214 and entered as input to an exit ramp warning algorithm. FIG. 7 depicts one exemplary embodiment of a process carried out by as stop sign warning algorithm. The zone warning algorithm will analyze the most probable path data and compare the vehicles speed or lane data with a threshold value associated with a most probable path node 506, 508, 510, and 512, for example.
[0043] At step 614, process 600 determines if at least one or more thresholds for a given node have been exceeded. According to one embodiment, if a threshold value has been exceeded warning determination module 214 provides a control signal to CAN system 240, which in turn actuates an HMI to provide a warning or other indication to a vehicle passenger that a dangerous condition is approaching along the most probable path at step 620.
Furthermore, step 620 may take place at control logic module 232, eco optimization module 234, or vehicle control module 238 with additional algorithms providing various threshold determinations. For example, vehicle control module 238 may receive the most probable path data from prediction module 212 and determine based on a gear algorithm or braking algorithm whether to actuate a gear control module 116 or brake module 1 10, 112 by providing a control signal to CAN system 240. [0044] FIG. 7 and FIG. 8 show processes carried out by various application algorithms stored in warning determination module 214 are shown. Referring to FIG. 7, a process for detecting stop signs and providing a warning to a driver or a control signal to a vehicle module in response to detecting the stop sign. In one exemplary embodiment, the driver assistance system may prevent or reduce the likelihood of accidents and intentional stop sign rolling. The digital map system may identify stop signs locations that are in the vehicle path by mapping the vehicle location with the GPS device.
[0045] The driver assistance system may include electronics configured to combine the vehicle position with one or more of the vehicle speed, data from angular rate sensors (e.g. yaw rate) and acceleration sensors (e.g., accelerometers) to calculate position with better accuracy and a higher update rate. The resulting vehicle position may be matched to a map using the digital map system. The map includes stop sign attributes (e.g., stop sign identification, map location, etc.). By combining the calculated vehicle position with the digital map system, a distance to the upcoming stop sign(s) may be estimated, for example with a Kalman filtering technique. A Kalman filtering technique advantageously provides accurate distance measurements from noisy GPS data. Also, because of the vehicle speed information, the aforementioned technique may be used even in the absence of a GPS signal.
[0046] The driver assistance system may also combine the calculated stop sign position with data from the vision system to more precisely recognize the stop sign. A warning may be issued to driver ahead of the stop sign based on the vehicle speed/location. The driver assistance system may also generate and/or execute a control algorithm to control the vehicle speed. Specifically, at step 702 in process 700 it is determined if a stop sign is on the future most probable path, such as path 500. Next, at step 704, vision system data may be analyzed to confirm that a stop sign is present using object detection software, for example. Next, at step 706 module 326 may determine the distance to the stop sign from the vehicle. In addition, step 710 determines whether a speed threshold associated with the distance determined at step 708 has been exceeded. If the speed threshold has been exceeded, a control signal is transmitted to an HMI to alert the driver of the unsafe speed in view of the distance between the vehicle and the stop sign.
[0047] With respect to FIG. 8, process 800 provides advance knowledge of roadway and terrain variations that may also be beneficial for autonomous vehicle functions. According to one embodiment, slope distribution data for the most probable path determined at step 608 of process 600 as shown in FIG. 6 is retrieved from the map database 208 at various nodes 506, 508, 510, and 512 along the most probable path 500 at step 802. According to one embodiment, slopes associated with more than one segment are added and averaged to determine a future slope distribution. At step 804, the slope distribution predicted to be encountered by vehicle 514 is compared with a speed threshold associated with a particular range of slope distributions. In addition there is also a slope distribution threshold set at step 804. According to one embodiment there is one slope threshold magnitude for positive and negative slopes. According to another embodiment, there are separate thresholds for positive and negative slopes. According to a another embodiment, the slope threshold is variable and depends on input factors such as vehicle location data or vehicle sensor input data. At step 806 and 808, if the speed threshold is exceeded for a particular slope distribution, a control signal may be sent to a vehicle module at step 808 or an HMI at step 810. For example, knowledge of extended downhill slopes allows hybrid vehicles to utilize regenerative braking to reenergize battery capacity. Advance knowledge of problematic intersections (hidden intersections or high incidence of accidents) allows vehicles to pre -prime braking pressure in silent anticipation of cross-traffic collision.
[0048] In one exemplary embodiment, the map system, vision system, and GPS device of the driver assistance system can be used together to advise the driver regarding lane changes in order to minimize braking. The driver assistance system 220 may provide the driver with lane change advice while nearing an exit ramp so that the vehicle has a smooth transition from high to low speed with minimal braking. The lane change advice may be shown in an HMI display and be determined by an exit ramp algorithm stored in warning determination module 214.
[0049] Accordingly, the driver assistance system may use data from the digital map system, vision system and GPS device to generate and execute an algorithm to provide lane change recommendations and vehicle speed profiles to the driver. The driver assistance system may also generate and execute a control algorithm for controlling the vehicle speed and steering angle.
[0050] The driver assistance system 220 may assist in improving gas mileage of the vehicle and aid in reducing gas consumption. The driver assistance system may assist in optimal braking to increase the life of brakes/vehicle by providing a control signal to eco- optimization module 234, for example. The driver assistance system may assist in avoiding last minute exit situations and thus reduce risk while driving. The driver assistance system may provide optional speed information based on the vehicle parameters and road
environment. The driver assistance system may assist in driver training for an optimal driving style. The driver assistance system may assist in reducing insurance costs. [0051] The present disclosure has been described with reference to example embodiments, however persons skilled in the art will recognize that changes may be made in form and detail without departing from the spirit and scope of the disclosed subject matter. For example, although different example embodiments may have been described as including one or more features providing one or more benefits, it is contemplated that the described features may be interchanged with one another or alternatively be combined with one another in the described example embodiments or in other alternative embodiments. Because the technology of the present disclosure is relatively complex, not all changes in the technology are foreseeable. The present disclosure described with reference to the exemplary embodiments is manifestly intended to be as broad as possible. For example, unless specifically otherwise noted, the exemplary embodiments reciting a single particular element also encompass a plurality of such particular elements.
[0052] Exemplary embodiments may include program products comprising computer or machine-readable media for carrying or having machine-executable instructions or data structures stored thereon. For example, the driver monitoring system may be computer driven. Exemplary embodiments illustrated in the methods of the figures may be controlled by program products comprising computer or machine-readable media for carrying or having machine-executable instructions or data structures stored thereon. Such computer or machine-readable media can be any available media which can be accessed by a general purpose or special purpose computer or other machine with a processor. By way of example, such computer or machine-readable media can comprise RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor.
Combinations of the above are also included within the scope of computer or machine- readable media. Computer or machine-executable instructions comprise, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing machines to perform a certain function or group of functions. Software implementations of the present invention could be accomplished with standard programming techniques with rule based logic and other logic to accomplish the various connection steps, processing steps, comparison steps and decision steps.
[0053] It is also important to note that the construction and arrangement of the elements of the system as shown in the preferred and other exemplary embodiments is illustrative only. Although only a certain number of embodiments have been described in detail in this disclosure, those skilled in the art who review this disclosure will readily appreciate that many modifications are possible (e.g., variations in sizes, dimensions, structures, shapes and proportions of the various elements, values of parameters, mounting arrangements, use of materials, colors, orientations, etc.) without materially departing from the novel teachings and advantages of the subject matter recited. For example, elements shown as integrally formed may be constructed of multiple parts or elements shown as multiple parts may be integrally formed, the operation of the assemblies may be reversed or otherwise varied, the length or width of the structures and/or members or connectors or other elements of the system may be varied, the nature or number of adjustment or attachment positions provided between the elements may be varied. It should be noted that the elements and/or assemblies of the system may be constructed from any of a wide variety of materials that provide sufficient strength or durability. Accordingly, all such modifications are intended to be included within the scope of the present disclosure. The order or sequence of any process or method steps may be varied or re-sequenced according to alternative embodiments. Other substitutions, modifications, changes and omissions may be made in the design, operating conditions and arrangement of the preferred and other exemplary embodiments without departing from the spirit of the present subject matter

Claims

WHAT IS CLAIMED IS:
1. A driver assistance system for providing driver and vehicle feedback control signals comprising:
a map database comprising navigation characteristics related to road locations;
a GPS unit that receives location data of the vehicle;
a map matching module configured to receive the location data of the vehicle and retrieve navigation characteristics relevant to the location data using a processing circuit; a prediction module configured to generate a most probable future path for the vehicle and to determine a location of at least one navigation characteristic with respect to the most probable future path and the vehicle;
at least one vehicle sensor unit configured to generate vehicle data; and
a warning module configured to transmit a control signal to a vehicle control area network to warn the driver of an upcoming navigation characteristic on the most probable path.
2. The driver assistance system of claim 1, wherein the at least one navigation characteristic comprises a stop sign location on the most probable path.
3. The driver assistance system of claim 2, wherein the generated vehicle data comprises vision system data.
4. The driver assistance system of claim 3, wherein the vision system data is analyzed using object detection software to recognize a stop sign
5. The driver assistance system of claim 3, wherein the warning module transmits the control signal to a human machine interface based on the location of the stop sign on the most probable path and the vision system data.
6. The driver assistance system of claim 5, wherein the generated vehicle data further comprises vehicle speed data, and the control signal comprises data used to advise the driver of optimal deceleration at the human machine interface.
7. The driver assistance system of claim 5, wherein the warning module transmits the control signal to at least one of a engine control module and a braking module to control vehicle speed or vehicle steering without human intervention.
8. The driver assistance system of claim 1, wherein the at least one navigation characteristic comprises slope distribution over a plurality of the future most probable path nodes.
9. The driver assistance system of claim 8, wherein, the generated vehicle data comprises vehicle speed data and the current gear position of the vehicle.
10. The driver assistance system of claim 8, wherein the warning module transmits the control signal to the engine control module and braking module to control vehicle speed or vehicle braking without human intervention.
11. The driver assistance system of claim 5, wherein the human machine interface comprises at least one of an audible indicator, a visual indictor, and a tactile indicator.
12. A method of assisting a driver of a vehicle by providing driver and vehicle feedback control signals, the method comprising:
receiving location data of the vehicle from a GPS unit;
receiving the location data of the vehicle and retrieving navigation characteristics relevant to the location data using a processing circuit;
generating a most probable future path for the vehicle and determining a location of at least one navigation characteristic with respect to the most probable future path and the vehicle;
generating vehicle data at least one vehicle sensor; and transmitting a control signal to a vehicle control area network to warn the driver of an upcoming navigation characteristic on the most probable path.
13. The method of claim 12, wherein the driver assistance system of claim 1, wherein the at least one navigation characteristic comprises a stop sign location on the most probable path.
14. The method of claim 13, wherein the generated vehicle data comprises vision system data.
15. The method of claim 14, wherein the vision system data is analyzed using object detection software to determine if a stop sign is located in proximity to the vehicle.
16. The method of claim 15, wherein the warning module transmits the control signal to a human machine interface based on the location of the stop sign on the most probable path and the vision system data.
17. The method of claim 15, wherein the generated vehicle data further comprises vehicle speed data, and the control signal comprises data used to advise the driver of optimal deceleration at the human machine interface.
18. The method of claim 15, wherein the warning module transmits the control signal to at least one of a engine control module and a braking module to control vehicle speed or vehicle steering without human intervention.
19. The method of claim 12, wherein the at least one navigation characteristic comprises slope distribution over a plurality of the future most probable path nodes.
20. The method of claim 19, wherein, the generated vehicle data comprises vehicle speed data and the current gear position of the vehicle.
21. The method of claim 19, wherein the warning module transmits the control signal to the engine control module and braking module to control vehicle speed or vehicle braking without human intervention.
22. The method of claim 17, wherein the human machine interface comprises at least one of an audible indicator, a visual indictor, and a tactile indicator.
PCT/US2012/030185 2011-03-23 2012-03-22 Driver assistance system WO2012129437A2 (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US201161466873P 2011-03-23 2011-03-23
US201161466870P 2011-03-23 2011-03-23
US201161466880P 2011-03-23 2011-03-23
US61/466,873 2011-03-23
US61/466,880 2011-03-23
US61/466,870 2011-03-23

Publications (2)

Publication Number Publication Date
WO2012129437A2 true WO2012129437A2 (en) 2012-09-27
WO2012129437A3 WO2012129437A3 (en) 2012-11-29

Family

ID=46880049

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2012/030185 WO2012129437A2 (en) 2011-03-23 2012-03-22 Driver assistance system

Country Status (2)

Country Link
US (1) US20120303222A1 (en)
WO (1) WO2012129437A2 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014074588A1 (en) * 2012-11-06 2014-05-15 Google Inc. Methods and systems to aid autonomous driving through a lane merge
EP2881297A3 (en) * 2013-11-21 2016-09-07 Robert Bosch Gmbh Method for predictive influence of a vehicle speed
WO2017178728A1 (en) * 2016-04-13 2017-10-19 Psa Automobiles S.A. Device for aiding the driving of a vehicle in an obligatory transit zone
CN107264529A (en) * 2017-06-28 2017-10-20 北京新能源汽车股份有限公司 Cruise method of controlling security, device and vehicle
US10017111B2 (en) 2015-05-12 2018-07-10 Elektrobit Automotive Gmbh Driver assistance system and method for avoiding collisions
FR3064098A1 (en) * 2017-03-20 2018-09-21 Robert Bosch Gmbh METHOD AND DEVICE FOR DETERMINING THE MOST PROBABLE PATHWAY OF A VEHICLE AND A DEVICE FOR ITS IMPLEMENTATION
CN108780606A (en) * 2016-03-18 2018-11-09 株式会社电装 Vehicle device
EP3334631A4 (en) * 2015-08-12 2019-04-24 Lytx Inc. Vehicle to vehicle and vehicle to infrastructure response monitoring
CN111746542A (en) * 2020-06-04 2020-10-09 重庆长安汽车股份有限公司 Method and system for reminding intelligent lane change of vehicle, vehicle and storage medium

Families Citing this family (57)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012145819A1 (en) 2011-04-25 2012-11-01 Magna International Inc. Image processing method for detecting objects using relative motion
US20140218535A1 (en) 2011-09-21 2014-08-07 Magna Electronics Inc. Vehicle vision system using image data transmission and power supply via a coaxial cable
US10099614B2 (en) 2011-11-28 2018-10-16 Magna Electronics Inc. Vision system for vehicle
US8694224B2 (en) 2012-03-01 2014-04-08 Magna Electronics Inc. Vehicle yaw rate correction
US10089537B2 (en) 2012-05-18 2018-10-02 Magna Electronics Inc. Vehicle vision system with front and rear camera integration
US8706417B2 (en) * 2012-07-30 2014-04-22 GM Global Technology Operations LLC Anchor lane selection method using navigation input in road change scenarios
DE102013217430A1 (en) 2012-09-04 2014-03-06 Magna Electronics, Inc. Driver assistance system for a motor vehicle
US9090234B2 (en) 2012-11-19 2015-07-28 Magna Electronics Inc. Braking control system for vehicle
US9092986B2 (en) 2013-02-04 2015-07-28 Magna Electronics Inc. Vehicular vision system
US10027930B2 (en) 2013-03-29 2018-07-17 Magna Electronics Inc. Spectral filtering for vehicular driver assistance systems
US9043072B1 (en) * 2013-04-04 2015-05-26 Google Inc. Methods and systems for correcting an estimated heading using a map
US9327693B2 (en) 2013-04-10 2016-05-03 Magna Electronics Inc. Rear collision avoidance system for vehicle
US9164511B1 (en) 2013-04-17 2015-10-20 Google Inc. Use of detected objects for image processing
US9260095B2 (en) 2013-06-19 2016-02-16 Magna Electronics Inc. Vehicle vision system with collision mitigation
US20140375476A1 (en) 2013-06-24 2014-12-25 Magna Electronics Inc. Vehicle alert system
US9988047B2 (en) 2013-12-12 2018-06-05 Magna Electronics Inc. Vehicle control system with traffic driving control
WO2015096878A1 (en) * 2013-12-24 2015-07-02 Volvo Truck Corporation Method and system for driver assistance for a vehicle
US10599155B1 (en) 2014-05-20 2020-03-24 State Farm Mutual Automobile Insurance Company Autonomous vehicle operation feature monitoring and evaluation of effectiveness
US9972054B1 (en) 2014-05-20 2018-05-15 State Farm Mutual Automobile Insurance Company Accident fault determination for autonomous vehicles
US11669090B2 (en) 2014-05-20 2023-06-06 State Farm Mutual Automobile Insurance Company Autonomous vehicle operation feature monitoring and evaluation of effectiveness
US9754325B1 (en) 2014-05-20 2017-09-05 State Farm Mutual Automobile Insurance Company Autonomous vehicle operation feature monitoring and evaluation of effectiveness
US10373259B1 (en) 2014-05-20 2019-08-06 State Farm Mutual Automobile Insurance Company Fully autonomous vehicle insurance pricing
US10387962B1 (en) 2014-07-21 2019-08-20 State Farm Mutual Automobile Insurance Company Methods of reconstructing an accident scene using telematics data
US9886040B1 (en) * 2014-09-24 2018-02-06 Rockwell Collins, Inc. System and method for platform alignment, navigation or targeting
US9463804B2 (en) * 2014-11-11 2016-10-11 Ford Global Tehnologies, LLC Vehicle cornering modes
US9946531B1 (en) 2014-11-13 2018-04-17 State Farm Mutual Automobile Insurance Company Autonomous vehicle software version assessment
US10049129B2 (en) * 2014-12-22 2018-08-14 Here Global B.V. Method and apparatus for providing map updates from distance based bucket processing
US20160231124A1 (en) * 2015-01-15 2016-08-11 GM Global Technology Operations LLC Horizon-based driver assistance systems and methods
US10553115B1 (en) * 2015-01-21 2020-02-04 Allstate Insurance Company System and method of vehicular collision avoidance
US10163350B1 (en) 2015-08-28 2018-12-25 State Farm Mutual Automobile Insurance Company Vehicular driver warnings
US9701244B2 (en) * 2015-09-29 2017-07-11 Toyota Motor Engineering & Manufacturing North America, Inc. Systems, methods, and vehicles for generating cues to drivers
US9796388B2 (en) 2015-12-17 2017-10-24 Ford Global Technologies, Llc Vehicle mode determination
US11242051B1 (en) 2016-01-22 2022-02-08 State Farm Mutual Automobile Insurance Company Autonomous vehicle action communications
US10134278B1 (en) 2016-01-22 2018-11-20 State Farm Mutual Automobile Insurance Company Autonomous vehicle application
US10395332B1 (en) 2016-01-22 2019-08-27 State Farm Mutual Automobile Insurance Company Coordinated autonomous vehicle automatic area scanning
US11719545B2 (en) 2016-01-22 2023-08-08 Hyundai Motor Company Autonomous vehicle component damage and salvage assessment
US11441916B1 (en) 2016-01-22 2022-09-13 State Farm Mutual Automobile Insurance Company Autonomous vehicle trip routing
US9940834B1 (en) 2016-01-22 2018-04-10 State Farm Mutual Automobile Insurance Company Autonomous vehicle application
US10324463B1 (en) 2016-01-22 2019-06-18 State Farm Mutual Automobile Insurance Company Autonomous vehicle operation adjustment based upon route
US10386845B1 (en) 2016-01-22 2019-08-20 State Farm Mutual Automobile Insurance Company Autonomous vehicle parking
US9903733B2 (en) 2016-03-17 2018-02-27 Honda Motor Co., Ltd. Vehicular communications network and methods of use and manufacture thereof
JP6414567B2 (en) * 2016-06-02 2018-10-31 トヨタ自動車株式会社 Speed limit display device for vehicle
US10029698B2 (en) * 2016-07-19 2018-07-24 Futurewei Technologies, Inc. Adaptive passenger comfort enhancement in autonomous vehicles
US10540895B2 (en) 2017-06-21 2020-01-21 International Business Machines Corporation Management of mobile objects
US10585180B2 (en) 2017-06-21 2020-03-10 International Business Machines Corporation Management of mobile objects
US10600322B2 (en) 2017-06-21 2020-03-24 International Business Machines Corporation Management of mobile objects
US10535266B2 (en) 2017-06-21 2020-01-14 International Business Machines Corporation Management of mobile objects
US10546488B2 (en) 2017-06-21 2020-01-28 International Business Machines Corporation Management of mobile objects
US10504368B2 (en) 2017-06-21 2019-12-10 International Business Machines Corporation Management of mobile objects
DE102018104987A1 (en) * 2018-03-05 2019-09-05 Jungheinrich Ag System for collision avoidance and method for collision avoidance
US10989553B2 (en) 2018-04-17 2021-04-27 Here Global B.V. Method, apparatus and computer program product for determining likelihood of a route
US10650623B2 (en) 2018-09-18 2020-05-12 Avinew, Inc. Detecting of automatic driving
CN110126817B (en) * 2018-12-16 2023-01-06 魔门塔(苏州)科技有限公司 Method and system for self-adapting to parking or recalling between any point and fixed point
GB2580388B (en) * 2019-01-09 2021-07-28 Jaguar Land Rover Ltd Control system for a vehicle
DE102019003430B3 (en) * 2019-05-15 2020-06-04 Daimler Ag Method for performing automated or autonomous driving of a vehicle
JP7263997B2 (en) * 2019-09-20 2023-04-25 トヨタ自動車株式会社 DRIVING ACTION EVALUATION DEVICE, METHOD AND PROGRAM
US11807221B2 (en) * 2020-04-29 2023-11-07 Gm Cruise Holdings Llc System for parking an autonomous vehicle

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100272912B1 (en) * 1996-11-19 2000-12-01 하나와 요시카즈 Vehicle drive force controller
JP2005182307A (en) * 2003-12-17 2005-07-07 Denso Corp Vehicle driving support device
US7260465B2 (en) * 2002-04-30 2007-08-21 Ford Global Technology, Llc Ramp identification in adaptive cruise control
JP2011000935A (en) * 2009-06-17 2011-01-06 Denso Corp Speed control device for vehicle

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7629899B2 (en) * 1997-10-22 2009-12-08 Intelligent Technologies International, Inc. Vehicular communication arrangement and method
US6311552B1 (en) * 1997-04-07 2001-11-06 Luk Leamington Limited Gear position sensor
KR100418763B1 (en) * 2000-12-28 2004-02-18 현대자동차주식회사 A method for measuring road slope and a system for controling vehicle speed upon road slope measured thereby
US7751973B2 (en) * 2004-05-04 2010-07-06 Visteon Global Technologies, Inc. Curve warning system
JP4631750B2 (en) * 2006-03-06 2011-02-16 トヨタ自動車株式会社 Image processing system
US7798578B2 (en) * 2006-08-17 2010-09-21 Ford Global Technologies, Llc Driver feedback to improve vehicle performance

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100272912B1 (en) * 1996-11-19 2000-12-01 하나와 요시카즈 Vehicle drive force controller
US7260465B2 (en) * 2002-04-30 2007-08-21 Ford Global Technology, Llc Ramp identification in adaptive cruise control
JP2005182307A (en) * 2003-12-17 2005-07-07 Denso Corp Vehicle driving support device
JP2011000935A (en) * 2009-06-17 2011-01-06 Denso Corp Speed control device for vehicle

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9026300B2 (en) 2012-11-06 2015-05-05 Google Inc. Methods and systems to aid autonomous vehicles driving through a lane merge
CN104870288A (en) * 2012-11-06 2015-08-26 谷歌公司 Methods and systems to aid autonomous driving through a lane merge
WO2014074588A1 (en) * 2012-11-06 2014-05-15 Google Inc. Methods and systems to aid autonomous driving through a lane merge
EP2881297A3 (en) * 2013-11-21 2016-09-07 Robert Bosch Gmbh Method for predictive influence of a vehicle speed
US10017111B2 (en) 2015-05-12 2018-07-10 Elektrobit Automotive Gmbh Driver assistance system and method for avoiding collisions
EP3334631A4 (en) * 2015-08-12 2019-04-24 Lytx Inc. Vehicle to vehicle and vehicle to infrastructure response monitoring
CN108780606A (en) * 2016-03-18 2018-11-09 株式会社电装 Vehicle device
FR3050162A1 (en) * 2016-04-13 2017-10-20 Peugeot Citroen Automobiles Sa DEVICE FOR ASSISTING THE DRIVING OF A VEHICLE IN A PASSAGE AREA OBLIGED
CN108885829A (en) * 2016-04-13 2018-11-23 标致雪铁龙汽车股份有限公司 Auxiliary driving device for the auxiliary driving means of transport in forcing traffic areas
WO2017178728A1 (en) * 2016-04-13 2017-10-19 Psa Automobiles S.A. Device for aiding the driving of a vehicle in an obligatory transit zone
FR3064098A1 (en) * 2017-03-20 2018-09-21 Robert Bosch Gmbh METHOD AND DEVICE FOR DETERMINING THE MOST PROBABLE PATHWAY OF A VEHICLE AND A DEVICE FOR ITS IMPLEMENTATION
CN107264529A (en) * 2017-06-28 2017-10-20 北京新能源汽车股份有限公司 Cruise method of controlling security, device and vehicle
CN111746542A (en) * 2020-06-04 2020-10-09 重庆长安汽车股份有限公司 Method and system for reminding intelligent lane change of vehicle, vehicle and storage medium
CN111746542B (en) * 2020-06-04 2023-04-14 重庆长安汽车股份有限公司 Method and system for vehicle intelligent lane change reminding, vehicle and storage medium

Also Published As

Publication number Publication date
US20120303222A1 (en) 2012-11-29
WO2012129437A3 (en) 2012-11-29

Similar Documents

Publication Publication Date Title
US20120303222A1 (en) Driver assistance system
US20120296539A1 (en) Driver assistance system
US20120245756A1 (en) Driver assistance system
US20120245817A1 (en) Driver assistance system
US11989666B1 (en) Predicting trajectory intersection by another road user
JP6380274B2 (en) Navigation device for autonomous vehicles
JP6462328B2 (en) Travel control system
CN110471415B (en) Vehicle with automatic driving mode and control method and system thereof
CN102449672B (en) Vehicular peripheral surveillance device
US9440647B1 (en) Safely navigating crosswalks
US8620571B2 (en) Driving assistance apparatus, driving assistance method, and driving assistance program
JP7205154B2 (en) Display device
US20150153184A1 (en) System and method for dynamically focusing vehicle sensors
EP3211374B1 (en) Travel route calculation device
WO2015052577A1 (en) Traffic lane guidance system for vehicle and traffic lane guidance method for vehicle
CN112689588A (en) Control method and device for automatically driving vehicle
CN110329250A (en) Method for exchanging information between at least two automobiles
US20170151945A1 (en) Information processing device, information processing method, control device for vehicle, and control method for vehicle
US11423780B2 (en) Traffic control system
CN114475648A (en) Autonomous vehicle control based on behavior of ambient contributing factors and limited environmental observations
WO2022150836A1 (en) Methods and systems for monitoring vehicle motion with driver safety alerts
US20230118619A1 (en) Parking-stopping point management device, parking-stopping point management method, and vehicle device
CN110827575A (en) Cooperative vehicle safety system and method
US20220121216A1 (en) Railroad Light Detection
Manichandra et al. Advanced Driver Assistance Systems

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12760373

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12760373

Country of ref document: EP

Kind code of ref document: A2