US20180339713A1 - Method and device for supporting a vehicle occupant in a vehicle - Google Patents

Method and device for supporting a vehicle occupant in a vehicle Download PDF

Info

Publication number
US20180339713A1
US20180339713A1 US15/976,985 US201815976985A US2018339713A1 US 20180339713 A1 US20180339713 A1 US 20180339713A1 US 201815976985 A US201815976985 A US 201815976985A US 2018339713 A1 US2018339713 A1 US 2018339713A1
Authority
US
United States
Prior art keywords
transportation vehicle
sensor
behavior pattern
user information
vehicle occupant
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/976,985
Other versions
US10745025B2 (en
Inventor
Esref Köse
Ahmed Ali
Sebastian GEHRLING
Jens Kampermann
Sarah Strygulec
Sven Klomp
Jessica Apfel
Sandra Jürgensmeier
Wolfgang Theimer
Ernst Zielinski
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Volkswagen AG
Original Assignee
Volkswagen AG
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Volkswagen AG filed Critical Volkswagen AG
Assigned to VOLKSWAGEN AG reassignment VOLKSWAGEN AG ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Apfel, Jessica, GEHRLING, SEBASTIAN, Strygulec, Sarah, THEIMER, WOLFGANG, ZIELINSKI, ERNST, ALI, AHMED, Jürgensmeier, Sandra, Kampermann, Jens, Köse, Esref, KLOMP, SVEN
Publication of US20180339713A1 publication Critical patent/US20180339713A1/en
Application granted granted Critical
Publication of US10745025B2 publication Critical patent/US10745025B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W40/00Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
    • B60W40/08Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/10Interpretation of driver requests or demands
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W40/00Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W40/00Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
    • B60W40/12Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to parameters of the vehicle itself, e.g. tyre models
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/085Changing the parameters of the control units, e.g. changing limit values, working points by control input
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W40/00Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
    • B60W40/08Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
    • B60W2040/0881Seat occupation; Driver or passenger presence
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W2050/0062Adapting control system settings
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W2050/0062Adapting control system settings
    • B60W2050/0075Automatic parameter input, automatic initialising or calibrating means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2400/00Indexing codes relating to detected, measured or calculated conditions or factors
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2540/00Input parameters relating to occupants
    • B60W2540/22Psychological state; Stress level or workload
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2540/00Input parameters relating to occupants
    • B60W2540/225Direction of gaze
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2556/00Input parameters relating to data
    • B60W2556/10Historical data
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/18Braking system
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/20Steering systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/30Auxiliary equipments
    • B60W2710/305Auxiliary equipments target power to auxiliaries
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60YINDEXING SCHEME RELATING TO ASPECTS CROSS-CUTTING VEHICLE TECHNOLOGY
    • B60Y2400/00Special features of vehicle units
    • B60Y2400/30Sensors

Definitions

  • Illustrative embodiments relate to a method for assisting a transportation vehicle occupant located in a transportation vehicle. Illustrative embodiments also relate to a device for carrying out the disclosed method and to a transportation vehicle which is configured for carrying out the disclosed method or which comprises the disclosed device.
  • FIG. 2 schematically shows a block diagram of a device.
  • a plurality of different types of assistance systems which can assist the driver in motorized transportation vehicles have become known.
  • the assistance systems are limited only to signaling or warning functions, and in some cases the assistance systems intervene into the drive or the control of the transportation vehicle.
  • Knowledge of the present transportation vehicle situation is required in both cases, which is obtained on the basis of data from suitable sensors.
  • the sensors can ascertain, in this case, technical parameters of the transportation vehicle itself, such as for a driving stabilization system (ESP) or anti-lock system (ABS), as well as information regarding the transportation vehicle surroundings.
  • ESP driving stabilization system
  • ABS anti-lock system
  • Various techniques are utilized in this case, frequently also in combination, for observing the transportation vehicle surroundings and assistance functions based thereon, such as parking assistance systems, automatic distance warnings and distance regulating functions, traffic sign recognition, blind spot monitoring, or emergency braking systems for pedestrian protection.
  • ultrasonic, radar, LIDAR, and camera systems are utilized for identifying objects in the surroundings of the transportation vehicle, such as pedestrians, other transportation vehicles, traffic signs, etc.
  • the sensor data from multiple sensors can be combined or linked, to improve the quality of the gathered information. If the information gathered from one or several sensors is not reliable, for example, this information can be partially or completely compensated for by data from further sensors. Likewise, however, information which is already available can also be confirmed by data from further sensors.
  • sensors can be provided in the transportation vehicle, which gather information regarding the behavior and/or the state of a transportation vehicle occupant, in particular, the driver.
  • the head and eye position of the driver can be detected and, by suitable image-analysis methods, the viewing direction of the driver can be ascertained or it can be detected that the driver is becoming fatigued.
  • GB 2532457 A describes determining, with the aid of an image sensor, the attentiveness of the driver depending on the amount of time the driver looks at the road or looks away from the road and, based thereon, controlling the sensitivity of a driver assistance system.
  • DE 10 2014 201 036 A1 describes an image-based classification method for ascertaining the driver state and/or the driver behavior, to ascertain the capability of a driver to take over.
  • cameras are arranged in such a way that the eyes of the driver can be detected given any orientation of the head, to ascertain the viewing direction of the driver and statistically evaluate the viewing direction over time.
  • Disclosed embodiments provide an improved method and an improved device for assisting a transportation vehicle occupant located in a transportation vehicle.
  • Disclosed embodiments provide a method and a corresponding device.
  • user information regarding the behavior and/or the state of a transportation vehicle occupant as well as transportation vehicle parameters of the transportation vehicle is gathered.
  • the gathered user information and transportation vehicle parameters are combined and jointly analyzed.
  • At least one behavior pattern of the transportation vehicle occupant is determined from the jointly analyzed user information and transportation vehicle parameters.
  • the behavior pattern is stored together with the user information.
  • this user information is compared with the user information regarding the at least one stored behavior pattern.
  • a transportation vehicle function is automatically carried out.
  • the disclosed method makes it possible to detect or learn individual behavior patterns and, based thereon, to assist the transportation vehicle occupant more efficiently, rapidly, or more comfortably in situations in which a behavior pattern is recognized.
  • the user information and transportation vehicle parameters gathered before the operation are analyzed.
  • the analysis and memory outlay can be reduced in this way.
  • the user information includes the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
  • the method can be applied when user information regarding the driver of the transportation vehicle is gathered.
  • a disclosed device for assisting a transportation vehicle occupant located in a transportation vehicle comprises a first sensor for gathering user information regarding the behavior and/or the state of a transportation vehicle occupant and a second sensor for gathering transportation vehicle parameters.
  • Sensor data from the first and second sensors are fed to an analysis and control unit.
  • the analysis and control unit analyzes the sensor data, determines at least one behavior pattern of the transportation vehicle occupant, compares subsequent sensor data from the first sensor with the at least one stored behavior pattern and, if the behavior pattern is recognized, automatically activates a transportation vehicle function.
  • the at least one behavior pattern is stored in a memory unit.
  • the first sensor is a camera sensor in the transportation vehicle, which is oriented toward the transportation vehicle occupant and detects the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
  • the second sensor ascertains one or several of the following transportation vehicle parameters:
  • an adaptation of a user interface of the transportation vehicle takes place.
  • a function of a driver assistance system is carried out.
  • the disclosed method or the disclosed device is utilized in a motorized transportation vehicle.
  • FIG. 1 schematically shows the method for assisting a transportation vehicle occupant located in a transportation vehicle.
  • a gathering of user information takes place.
  • arbitrary information regarding the user behavior and/or the user state of a transportation vehicle occupant can be ascertained in this case, provided this information is associated with the use of the transportation vehicle by the transportation vehicle occupant and, in the case that the same operation is carried out multiple times, is identical or at least similar and, therefore, after a learning phase, makes it possible to predict intentions of the transportation vehicle occupant when operations are carried out in the future. It is significant, in this case, to detect the viewing direction, the facial expression and/or the orientation of the head of a transportation vehicle occupant, in particular, of the driver of the transportation vehicle.
  • transportation vehicle parameters are gathered in a method operation at 2 .
  • these can be different types of parameters, such as the actuation of an operating element of the transportation vehicle, such as the turn-signal lamp, the steering wheel, the gas or brake pedal, the trajectory of the transportation vehicle, or the position of the transportation vehicle.
  • method operation at 3 the user information and the transportation vehicle parameters are combined and jointly evaluated.
  • method operation at 4 a behavior pattern is determined therefrom. Provided no behavior patterns were ascertained, for example, in an initialization phase, the behavior pattern is immediately stored in method operation at 5 , to allow for a recognition at a later time and, based thereon, to assist the transportation vehicle occupant. However, if one or several behavior patterns have already been learned and stored by the system, a comparison of the presently ascertained behavior pattern with the behavior patterns already present can take place in method operation at 6 . If a behavior pattern is recognized in this case, a transportation vehicle function can be automatically carried out in method operation at 7 .
  • a function of a driver assistance system can be carried out or a user interface can be suitably adapted.
  • this behavior pattern is also stored, to enable access thereto in the future.
  • the driver In the case of a lane change, for example, on the highway, the driver normally glances over his/her shoulder and repeatedly looks in the side-view mirror before he/she carries out the lane change. According to the disclosure, the driver is then observed, for example, by a camera-based eye-tracking system. The glance over the shoulder and the repeated glances into the side-view mirror are detected by the eye-tracking system and are analyzed by suitable image-processing methods.
  • transportation vehicle sensors monitor parameters of the transportation vehicle and detect when the turn signal is activated by the driver.
  • a lane change of the transportation vehicle can be ascertained, for example, by sensors which detect the steering wheel or tire position, or by a compass function of a navigation system. If it is determined via the sensor data that the transportation vehicle has initially been traveling in a constant direction for a certain period of time, then a brief change in direction associated with the lane change takes place and the transportation vehicle then continues traveling in the original direction, it can be assumed that a lane change has taken place.
  • a behavior pattern for the lane change can then be ascertained by a combination and a joint analysis of this user information and transportation vehicle parameters. If it is known via the transportation vehicle parameters that a lane change has taken place and the user information yields the finding that the driver has glanced over his/her shoulder and glanced in the side-view mirror before the lane change, this can be stored as a behavior pattern. If it is then ascertained at a later point in time that the driver is glancing over his/her shoulder and in the side-view mirror, this behavior pattern is recognized as being one of the known, stored behavior patterns.
  • the driver assistance system can then respond to the present situation at an early point in time, for example, by way of automatically actuating the turn signal or warning the driver, by a display, for example, on or in the side-view mirror, about a transportation vehicle that is difficult to see or is approaching at a high rate of speed.
  • the disclosed method can even assist the driver of a transportation vehicle when the driver wants to park the transportation vehicle into a parking space. If the behavior of the driver, transportation vehicle parameters, and, optionally, the transportation vehicle surroundings have been detected in earlier parking maneuvers, the corresponding behavior pattern can be recognized when parking maneuvers are carried out again.
  • a camera-based eye-tracking system that the driver is looking at the transportation vehicle surroundings through the side windows, the windshield, or the rear window, or, optionally, a side-view mirror.
  • suitable sensors for example, ultrasonic sensors, objects in the transportation vehicle surroundings as well as the distance of these objects from the transportation vehicle can be simultaneously detected and provide an indication that the driver has initially placed the transportation vehicle next to a parking space and has then maneuvered into the parking space.
  • it can be evaluated whether the objects are located in the direction of the detected viewing direction of the driver.
  • transportation vehicle sensors can gather parameters of the transportation vehicle such as the actuation of the turn signal, a large steering angle of the steering wheel, typically first in one direction and then in the opposite direction, and whether the transportation vehicle was moved forward or backward as a result. If the behavior pattern which is typical for a parking maneuver is then recognized at the beginning of a new parking maneuver, the parking assistance system can be automatically activated, for example.
  • Yet another example of an application which is not limited to the assistance of the driver, but rather can also be utilized for other transportation vehicle occupants, in particular, for the passenger, relates to an automatic orientation of a sun visor.
  • the pupil size can also be ascertained by suitable image-evaluation methods. A suddenly occurring, considerable increase in the pupil size can indicate that the transportation vehicle occupant is blinded. In addition to or instead thereof, a squinting of the eyes can also be registered by the image-evaluation method. If necessary, this information can also be combined with sensor data from a light sensor. If sensors are then additionally provided on the sun visor, which detect a lowering or tilting of the sun visor by the transportation vehicle occupant, the present information can be combined and a corresponding behavior pattern can be determined and stored.
  • the sun visor can be automatically oriented, provided suitable actuators are present, without the need for the transportation vehicle occupant to be active therefor.
  • the behavior patterns of different persons in the above-described examples of an application are usually similar, but are not identical.
  • a decision is made depending on the transportation vehicle occupant during the learning of the behavior pattern and the assistance of the transportation vehicle occupant based thereon. This can take place, for example, by methods for facial recognition, but also on the basis of other biometric methods for personal identification in the transportation vehicle.
  • FIG. 2 schematically shows a block diagram of a disclosed device.
  • a camera 8 arranged in the transportation vehicle interior detects the head position of the driver and ascertains, by suitable image-evaluation methods, the eye position and viewing direction, but also, optionally, the facial expressions of the driver.
  • This may be a near-infrared (NIR) camera which detects short-wave infrared radiation immediately adjacent to the visible range.
  • NIR near-infrared
  • an illumination in particular, in the dark, can take place by one or several NIR LEDs (not shown), without this being disturbing to the driver or being capable of blinding the driver.
  • the NIR camera and the NIR LEDs can be installed separately or also in one shared sensor component.
  • At least one sensor 9 is installed in the transportation vehicle for gathering the transportation vehicle parameters.
  • at least one sensor 10 is provided for detecting the transportation vehicle surroundings.
  • the sensor data are fed to an analysis and control unit 14 , in which the sensor data are evaluated, to determine behavior patterns, carry out a comparison with known behavior patterns and, depending thereon, to be capable of controlling, for example, an HMI 12 or a driver assistance system 13 .
  • a memory 11 is provided for storing the known behavior patterns.
  • the disclosed embodiments can be utilized in arbitrary areas of automotive engineering.

Landscapes

  • Engineering & Computer Science (AREA)
  • Automation & Control Theory (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Traffic Control Systems (AREA)
  • Control Of Driving Devices And Active Controlling Of Vehicle (AREA)

Abstract

A method for assisting a transportation vehicle occupant located in a transportation vehicle wherein user information regarding the behavior and/or the state of a transportation vehicle occupant and transportation vehicle parameters of the transportation vehicle are gathered. The gathered user information and transportation vehicle parameters are combined and jointly analyzed and at least one behavior pattern of the transportation vehicle occupant is determined from the jointly analyzed user information and transportation vehicle parameters. The behavior pattern is stored together with the user information. In a subsequent gathering of user information, this user information is compared with the user information behavior pattern previously stored. When there is conformance with the behavior pattern, a transportation vehicle function is automatically carried out.

Description

    PRIORITY CLAIM
  • This patent application claims priority to German Patent Application No. 10 2017 208 971.7, filed 29 May 2017, the disclosure of which is incorporated herein by reference in its entirety.
  • SUMMARY
  • Illustrative embodiments relate to a method for assisting a transportation vehicle occupant located in a transportation vehicle. Illustrative embodiments also relate to a device for carrying out the disclosed method and to a transportation vehicle which is configured for carrying out the disclosed method or which comprises the disclosed device.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Further features of the disclosure are apparent from the following description and the claims in conjunction with the figures, in which:
  • FIG. 1 schematically shows a method for assisting a transportation vehicle occupant located in a transportation vehicle; and
  • FIG. 2 schematically shows a block diagram of a device.
  • DETAILED DESCRIPTION
  • A plurality of different types of assistance systems which can assist the driver in motorized transportation vehicles have become known. In some cases, the assistance systems are limited only to signaling or warning functions, and in some cases the assistance systems intervene into the drive or the control of the transportation vehicle. Knowledge of the present transportation vehicle situation is required in both cases, which is obtained on the basis of data from suitable sensors. The sensors can ascertain, in this case, technical parameters of the transportation vehicle itself, such as for a driving stabilization system (ESP) or anti-lock system (ABS), as well as information regarding the transportation vehicle surroundings. Various techniques are utilized in this case, frequently also in combination, for observing the transportation vehicle surroundings and assistance functions based thereon, such as parking assistance systems, automatic distance warnings and distance regulating functions, traffic sign recognition, blind spot monitoring, or emergency braking systems for pedestrian protection. Frequently, ultrasonic, radar, LIDAR, and camera systems are utilized for identifying objects in the surroundings of the transportation vehicle, such as pedestrians, other transportation vehicles, traffic signs, etc. In this case, the sensor data from multiple sensors can be combined or linked, to improve the quality of the gathered information. If the information gathered from one or several sensors is not reliable, for example, this information can be partially or completely compensated for by data from further sensors. Likewise, however, information which is already available can also be confirmed by data from further sensors.
  • In addition, sensors can be provided in the transportation vehicle, which gather information regarding the behavior and/or the state of a transportation vehicle occupant, in particular, the driver. By using a camera situated in the transportation vehicle interior, the head and eye position of the driver can be detected and, by suitable image-analysis methods, the viewing direction of the driver can be ascertained or it can be detected that the driver is becoming fatigued.
  • DE 10 2015 122 603 A1 discloses a driver assistance system comprising a camera directed at the driver, to record the face of the driver, especially the eyes of the driver, and, based on the viewing direction, to monitor whether the driver is distracted or is looking away from the road. If the driver is looking away from the road, a driver status module detects an abnormal situation and activates the driver assistance system. Likewise, the driver status module can establish whether the driver is sleepy or awake, based on how wide and for how long the eyes of the driver are open.
  • Similarly, GB 2532457 A describes determining, with the aid of an image sensor, the attentiveness of the driver depending on the amount of time the driver looks at the road or looks away from the road and, based thereon, controlling the sensitivity of a driver assistance system.
  • DE 10 2014 201 036 A1 describes an image-based classification method for ascertaining the driver state and/or the driver behavior, to ascertain the capability of a driver to take over. For this purpose, cameras are arranged in such a way that the eyes of the driver can be detected given any orientation of the head, to ascertain the viewing direction of the driver and statistically evaluate the viewing direction over time.
  • Disclosed embodiments provide an improved method and an improved device for assisting a transportation vehicle occupant located in a transportation vehicle.
  • Disclosed embodiments provide a method and a corresponding device.
  • In the disclosed method for assisting a transportation vehicle occupant located in a transportation vehicle, user information regarding the behavior and/or the state of a transportation vehicle occupant as well as transportation vehicle parameters of the transportation vehicle is gathered. The gathered user information and transportation vehicle parameters are combined and jointly analyzed. At least one behavior pattern of the transportation vehicle occupant is determined from the jointly analyzed user information and transportation vehicle parameters. The behavior pattern is stored together with the user information. In a subsequent gathering of user information, this user information is compared with the user information regarding the at least one stored behavior pattern. When there is conformance with the behavior pattern, a transportation vehicle function is automatically carried out.
  • Contrary to the conventional assistance systems which respond in a standardized manner to the input from a transportation vehicle occupant, in particular, the driver, the disclosed method makes it possible to detect or learn individual behavior patterns and, based thereon, to assist the transportation vehicle occupant more efficiently, rapidly, or more comfortably in situations in which a behavior pattern is recognized.
  • According to at least one disclosed embodiment, when there is a change to a transportation vehicle parameter due to an operation carried out by the transportation vehicle occupant, the user information and transportation vehicle parameters gathered before the operation are analyzed. The analysis and memory outlay can be reduced in this way.
  • The user information includes the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
  • The method can be applied when user information regarding the driver of the transportation vehicle is gathered.
  • Correspondingly, a disclosed device for assisting a transportation vehicle occupant located in a transportation vehicle comprises a first sensor for gathering user information regarding the behavior and/or the state of a transportation vehicle occupant and a second sensor for gathering transportation vehicle parameters. Sensor data from the first and second sensors are fed to an analysis and control unit. The analysis and control unit analyzes the sensor data, determines at least one behavior pattern of the transportation vehicle occupant, compares subsequent sensor data from the first sensor with the at least one stored behavior pattern and, if the behavior pattern is recognized, automatically activates a transportation vehicle function. The at least one behavior pattern is stored in a memory unit.
  • Optionally, the first sensor is a camera sensor in the transportation vehicle, which is oriented toward the transportation vehicle occupant and detects the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
  • The second sensor ascertains one or several of the following transportation vehicle parameters:
      • the actuation of an operating element of the transportation vehicle, such as the turn-signal lamp, the steering wheel, the gas or brake pedal;
      • the position and/or trajectory of the transportation vehicle.
  • Likewise, optionally, a third sensor may be provided for detecting transportation vehicle-external objects in the transportation vehicle surroundings and the sensor data from the third sensor are likewise fed to the analysis and control unit and are analyzed thereby.
  • According to at least one disclosed embodiment, when there is conformance with the behavior pattern, an adaptation of a user interface of the transportation vehicle takes place.
  • According to yet another disclosed embodiment, when there is conformance with the behavior pattern, a function of a driver assistance system is carried out.
  • Optionally, the disclosed method or the disclosed device is utilized in a motorized transportation vehicle.
  • Disclosed embodiments are described in greater detail in the following with reference to the figures, to improve the understanding of the principles of the present disclosure. It is understood that the disclosed is not limited to these embodiments and that the described features can also be combined or modified without departing from the scope of protection of the disclosure as it is defined in the claims.
  • FIG. 1 schematically shows the method for assisting a transportation vehicle occupant located in a transportation vehicle. According to method operation at 1, a gathering of user information takes place. In principle, arbitrary information regarding the user behavior and/or the user state of a transportation vehicle occupant can be ascertained in this case, provided this information is associated with the use of the transportation vehicle by the transportation vehicle occupant and, in the case that the same operation is carried out multiple times, is identical or at least similar and, therefore, after a learning phase, makes it possible to predict intentions of the transportation vehicle occupant when operations are carried out in the future. It is significant, in this case, to detect the viewing direction, the facial expression and/or the orientation of the head of a transportation vehicle occupant, in particular, of the driver of the transportation vehicle.
  • In parallel to the gathering of the user information, transportation vehicle parameters are gathered in a method operation at 2. Depending on the application, these can be different types of parameters, such as the actuation of an operating element of the transportation vehicle, such as the turn-signal lamp, the steering wheel, the gas or brake pedal, the trajectory of the transportation vehicle, or the position of the transportation vehicle.
  • In method operation at 3, the user information and the transportation vehicle parameters are combined and jointly evaluated. In method operation at 4, a behavior pattern is determined therefrom. Provided no behavior patterns were ascertained, for example, in an initialization phase, the behavior pattern is immediately stored in method operation at 5, to allow for a recognition at a later time and, based thereon, to assist the transportation vehicle occupant. However, if one or several behavior patterns have already been learned and stored by the system, a comparison of the presently ascertained behavior pattern with the behavior patterns already present can take place in method operation at 6. If a behavior pattern is recognized in this case, a transportation vehicle function can be automatically carried out in method operation at 7. In this case, for example, a function of a driver assistance system can be carried out or a user interface can be suitably adapted. However, if there is a new behavior pattern that has not been known so far, this behavior pattern is also stored, to enable access thereto in the future.
  • Exemplary embodiments are mentioned in the following for further illustration.
  • In the case of a lane change, for example, on the highway, the driver normally glances over his/her shoulder and repeatedly looks in the side-view mirror before he/she carries out the lane change. According to the disclosure, the driver is then observed, for example, by a camera-based eye-tracking system. The glance over the shoulder and the repeated glances into the side-view mirror are detected by the eye-tracking system and are analyzed by suitable image-processing methods.
  • At the same time, transportation vehicle sensors monitor parameters of the transportation vehicle and detect when the turn signal is activated by the driver. In addition, a lane change of the transportation vehicle can be ascertained, for example, by sensors which detect the steering wheel or tire position, or by a compass function of a navigation system. If it is determined via the sensor data that the transportation vehicle has initially been traveling in a constant direction for a certain period of time, then a brief change in direction associated with the lane change takes place and the transportation vehicle then continues traveling in the original direction, it can be assumed that a lane change has taken place.
  • A behavior pattern for the lane change can then be ascertained by a combination and a joint analysis of this user information and transportation vehicle parameters. If it is known via the transportation vehicle parameters that a lane change has taken place and the user information yields the finding that the driver has glanced over his/her shoulder and glanced in the side-view mirror before the lane change, this can be stored as a behavior pattern. If it is then ascertained at a later point in time that the driver is glancing over his/her shoulder and in the side-view mirror, this behavior pattern is recognized as being one of the known, stored behavior patterns. The driver assistance system can then respond to the present situation at an early point in time, for example, by way of automatically actuating the turn signal or warning the driver, by a display, for example, on or in the side-view mirror, about a transportation vehicle that is difficult to see or is approaching at a high rate of speed.
  • The disclosed method can even assist the driver of a transportation vehicle when the driver wants to park the transportation vehicle into a parking space. If the behavior of the driver, transportation vehicle parameters, and, optionally, the transportation vehicle surroundings have been detected in earlier parking maneuvers, the corresponding behavior pattern can be recognized when parking maneuvers are carried out again.
  • For example, it can be ascertained with the aid of a camera-based eye-tracking system that the driver is looking at the transportation vehicle surroundings through the side windows, the windshield, or the rear window, or, optionally, a side-view mirror. By suitable sensors, for example, ultrasonic sensors, objects in the transportation vehicle surroundings as well as the distance of these objects from the transportation vehicle can be simultaneously detected and provide an indication that the driver has initially placed the transportation vehicle next to a parking space and has then maneuvered into the parking space. In addition, it can be evaluated whether the objects are located in the direction of the detected viewing direction of the driver. In addition, as in the aforementioned example of an application, transportation vehicle sensors can gather parameters of the transportation vehicle such as the actuation of the turn signal, a large steering angle of the steering wheel, typically first in one direction and then in the opposite direction, and whether the transportation vehicle was moved forward or backward as a result. If the behavior pattern which is typical for a parking maneuver is then recognized at the beginning of a new parking maneuver, the parking assistance system can be automatically activated, for example.
  • Yet another example of an application which is not limited to the assistance of the driver, but rather can also be utilized for other transportation vehicle occupants, in particular, for the passenger, relates to an automatic orientation of a sun visor.
  • If the eyes of the transportation vehicle occupant are detected, the pupil size can also be ascertained by suitable image-evaluation methods. A suddenly occurring, considerable increase in the pupil size can indicate that the transportation vehicle occupant is blinded. In addition to or instead thereof, a squinting of the eyes can also be registered by the image-evaluation method. If necessary, this information can also be combined with sensor data from a light sensor. If sensors are then additionally provided on the sun visor, which detect a lowering or tilting of the sun visor by the transportation vehicle occupant, the present information can be combined and a corresponding behavior pattern can be determined and stored. If it is established at a later point in time, on the basis of the observation of the eyes of the transportation vehicle occupant, that the transportation vehicle occupant is being blinded again, the sun visor can be automatically oriented, provided suitable actuators are present, without the need for the transportation vehicle occupant to be active therefor.
  • The behavior patterns of different persons in the above-described examples of an application are usually similar, but are not identical. To increase the accuracy of the assistance system, it can therefore be provided that a decision is made depending on the transportation vehicle occupant during the learning of the behavior pattern and the assistance of the transportation vehicle occupant based thereon. This can take place, for example, by methods for facial recognition, but also on the basis of other biometric methods for personal identification in the transportation vehicle.
  • FIG. 2 schematically shows a block diagram of a disclosed device. A camera 8 arranged in the transportation vehicle interior, for example, in the instrument panel or in the area of the rear-view mirror, detects the head position of the driver and ascertains, by suitable image-evaluation methods, the eye position and viewing direction, but also, optionally, the facial expressions of the driver. This may be a near-infrared (NIR) camera which detects short-wave infrared radiation immediately adjacent to the visible range. In this case, an illumination, in particular, in the dark, can take place by one or several NIR LEDs (not shown), without this being disturbing to the driver or being capable of blinding the driver. The NIR camera and the NIR LEDs can be installed separately or also in one shared sensor component. At least one sensor 9 is installed in the transportation vehicle for gathering the transportation vehicle parameters. In addition, at least one sensor 10 is provided for detecting the transportation vehicle surroundings. The sensor data are fed to an analysis and control unit 14, in which the sensor data are evaluated, to determine behavior patterns, carry out a comparison with known behavior patterns and, depending thereon, to be capable of controlling, for example, an HMI 12 or a driver assistance system 13. A memory 11 is provided for storing the known behavior patterns.
  • The disclosed embodiments can be utilized in arbitrary areas of automotive engineering.
  • LIST OF REFERENCE SIGNS
    • 1 method operation including gathering of user information
    • 2 method operation including gathering of transportation vehicle parameters
    • 3 method operation including analysis of the user information and transportation vehicle parameters
    • 4 method operation including determination of a behavior pattern
    • 5 method operation including storing the behavior pattern
    • 6 method operation including checking whether a known behavior pattern is present
    • 7 method operation including carrying out a transportation vehicle function
    • 8 camera for observing the transportation vehicle occupant
    • 9 sensor for transportation vehicle parameters
    • 10 sensor for detecting the transportation vehicle surroundings
    • 11 memory
    • 12 HMI
    • 13 assistance system
    • 14 analysis and control unit

Claims (21)

1. A method for assisting a transportation vehicle occupant located in a transportation vehicle, the method comprising:
gathering user information regarding the behavior and/or the state of a transportation vehicle occupant;
gathering transportation vehicle parameters of the transportation vehicle; wherein the detected user information and transportation vehicle parameters are combined and jointly analyzed;
determining at least one behavior pattern of the transportation vehicle occupant from the analyzed user information and transportation vehicle parameters;
storing the behavior pattern together with the user information;
comparing the user information, in a subsequent gathering of user information, with the user information of the at least one stored behavior pattern; and
in response to the comparison indicating conformance with the behavior pattern, automatically carrying out a transportation vehicle function.
2. The method of claim 1, wherein the user information and transportation vehicle parameters gathered before the operation are analyzed in response to there being a change to a transportation vehicle parameter due to an operation carried out by the transportation vehicle occupant.
3. The method of claim 1, wherein the user information includes the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
4. The method of claim 1, wherein user information regarding the driver of the transportation vehicle is gathered.
5. The method of claim 1, wherein a third sensor detects transportation vehicle-external objects in the transportation vehicle surroundings and the sensor data from the third sensor are fed to the analysis and control unit and are analyzed.
6. The method of claim 1, wherein an adaptation of a user interface of the transportation vehicle takes place when there is conformance with the behavior pattern.
7. The method of claim 1, wherein a function of a driver assistance system is carried out when there is conformance with the behavior pattern.
8. A device for assisting a transportation vehicle occupant located in a transportation vehicle, the device comprising:
a first sensor for gathering user information regarding the behavior and/or the state of a transportation vehicle occupant;
a second sensor for gathering transportation vehicle parameters, wherein an analysis and control unit, to which sensor data from the first and second sensors are fed and which analyzes the sensor data and determines at least one behavior pattern of the transportation vehicle occupant and compares subsequent sensor data from the first sensor with the at least one stored behavior pattern and, in response to recognition of the behavior pattern, automatically activates a transportation vehicle function; and
a memory unit for storing the at least one behavior pattern.
9. The device of claim 8, wherein the first sensor is a camera sensor in the transportation vehicle which is oriented toward the transportation vehicle occupant and detects the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
10. The device of claim 8, wherein the second sensor ascertains one or several of the following transportation vehicle parameters:
the actuation of an operating element of the transportation vehicle; and
the position and/or trajectory of the transportation vehicle.
11. The device of claim 10, wherein the operating element of the transportation vehicle is one of a turn-signal lamp, a steering wheel, a gas pedal or a brake pedal.
12. The device of claim 8, wherein a third sensor detects transportation vehicle-external objects in the transportation vehicle surroundings and the sensor data from the third sensor are fed to the analysis and control unit and are analyzed.
13. The device of claim 8, wherein an adaptation of a user interface of the transportation vehicle takes place when there is conformance with the behavior pattern.
14. The device of claim 8, wherein a function of a driver assistance system is carried out when there is conformance with the behavior pattern.
15. A transportation vehicle which comprises a device for assisting a transportation vehicle occupant located in the transportation vehicle, the device comprising:
a first sensor for gathering user information regarding the behavior and/or the state of a transportation vehicle occupant;
a second sensor for gathering transportation vehicle parameters, wherein an analysis and control unit, to which sensor data from the first and second sensors are fed and which analyzes the sensor data and determines at least one behavior pattern of the transportation vehicle occupant and compares subsequent sensor data from the first sensor with the at least one stored behavior pattern and, in response to recognition of the behavior pattern, automatically activates a transportation vehicle function; and
a memory unit for storing the at least one behavior pattern.
16. The vehicle of claim 15, wherein the first sensor is a camera sensor in the transportation vehicle which is oriented toward the transportation vehicle occupant and detects the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
17. The vehicle of claim 15, wherein the second sensor ascertains one or several of the following transportation vehicle parameters:
the actuation of an operating element of the transportation vehicle; and
the position and/or trajectory of the transportation vehicle.
18. The vehicle of claim 15, wherein the operating element of the transportation vehicle is one of a turn-signal lamp, a steering wheel, a gas pedal or a brake pedal.
19. The vehicle of claim 15, wherein a third sensor detects transportation vehicle-external objects in the transportation vehicle surroundings and the sensor data from the third sensor are fed to the analysis and control unit and are analyzed.
20. The vehicle of claim 15, wherein an adaptation of a user interface of the transportation vehicle takes place when there is conformance with the behavior pattern.
21. The vehicle of claim 15, wherein a function of a driver assistance system is carried out when there is conformance with the behavior pattern.
US15/976,985 2017-05-29 2018-05-11 Method and device for supporting a vehicle occupant in a vehicle Active 2039-01-01 US10745025B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
DE102017208971.7A DE102017208971A1 (en) 2017-05-29 2017-05-29 METHOD AND DEVICE FOR SUPPORTING A VEHICLE IN A VEHICLE
DE102017208971 2017-05-29
DE102017208971.7 2017-05-29

Publications (2)

Publication Number Publication Date
US20180339713A1 true US20180339713A1 (en) 2018-11-29
US10745025B2 US10745025B2 (en) 2020-08-18

Family

ID=61189267

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/976,985 Active 2039-01-01 US10745025B2 (en) 2017-05-29 2018-05-11 Method and device for supporting a vehicle occupant in a vehicle

Country Status (5)

Country Link
US (1) US10745025B2 (en)
EP (1) EP3409551B1 (en)
KR (1) KR102050466B1 (en)
CN (1) CN108944938A (en)
DE (1) DE102017208971A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190171211A1 (en) * 2016-08-09 2019-06-06 Nissan Motor Co., Ltd. Control Method and Control Device of Automatic Driving Vehicle

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102017219365B4 (en) 2017-10-27 2023-03-30 Zf Friedrichshafen Ag Apparatus and method for adaptive vehicle control
CN111319626B (en) * 2020-03-31 2021-07-20 深圳市英威腾电动汽车驱动技术有限公司 Vehicle control parameter adjusting method and device, electronic equipment and storage medium
DE102020124910A1 (en) 2020-09-24 2022-03-24 Bayerische Motoren Werke Aktiengesellschaft Method for supporting a driver when driving a motor vehicle manually, support device and motor vehicle
DE102022103912A1 (en) 2022-02-18 2023-08-24 Bayerische Motoren Werke Aktiengesellschaft Method for monitoring a condition of a driver of a vehicle through a camera in an interior of the vehicle, computer-readable medium, system and vehicle
DE102022132377A1 (en) 2022-12-06 2024-06-06 Bayerische Motoren Werke Aktiengesellschaft Method for operating an automated driving function, method for training an artificial intelligence and processing device

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102012221036A1 (en) * 2012-11-19 2014-05-22 Bayerische Motoren Werke Aktiengesellschaft Method for automatic switching on of parking- and/or maneuvering assistance systems e.g. park distance control system in motor car, involves switching on parking- and/or maneuvering assistance systems based on evaluation of driver behavior

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10255016A (en) 1997-03-13 1998-09-25 Nissan Motor Co Ltd Device for predicting driver's operation by blink
DE10126224A1 (en) 2001-05-30 2002-12-12 Bosch Gmbh Robert Method and device for characterizing the condition of the driver of a motor vehicle
EP1284432A1 (en) 2001-08-14 2003-02-19 Ford Global Technologies, Inc., A subsidiary of Ford Motor Company Device and process for operating a driver assistance system
DE10343683A1 (en) 2003-09-20 2005-04-21 Daimler Chrysler Ag Information system for motor vehicles
JP5286035B2 (en) * 2008-11-06 2013-09-11 本田技研工業株式会社 Vehicle speed control device
DE102011055685A1 (en) * 2011-11-24 2013-05-29 Continental Teves Ag & Co. Ohg Method for operating system influencing driving behavior of vehicle, involves carrying out automatic adjustment of adjustable control parameter of system manually by vehicle occupant
TWI447039B (en) * 2011-11-25 2014-08-01 Driving behavior analysis and warning system and method thereof
DE102012004601A1 (en) 2012-03-07 2012-10-11 Daimler Ag Assistance system for rider for offering and/or signaling information to passenger of motor car, has control device for controlling output device such that occupants offer and/or signal information in context-sensitive manner
US20130325202A1 (en) 2012-06-01 2013-12-05 GM Global Technology Operations LLC Neuro-cognitive driver state processing
KR20140002373A (en) * 2012-06-29 2014-01-08 현대자동차주식회사 Apparatus and method for monitoring driver state by driving pattern learning
BR112015025117A2 (en) * 2013-04-12 2017-07-18 Dana Ltd method for vehicle and operator orientation through pattern recognition
US9103688B2 (en) 2013-08-21 2015-08-11 Continental Automotive Systems, Inc. Adapting vehicle personality using analyzed driver performance metrics
DE102013217552A1 (en) 2013-09-03 2015-03-05 Bayerische Motoren Werke Aktiengesellschaft Situation-dependent activation of driver assistance functions
KR20150066308A (en) * 2013-12-06 2015-06-16 한국전자통신연구원 Apparatus and method for determining driving condition of deiver
DE102014201036A1 (en) 2014-01-21 2015-07-23 Bayerische Motoren Werke Aktiengesellschaft Image-based classification of driver status and / or driver behavior
EP2923912B1 (en) * 2014-03-24 2018-12-26 Volvo Car Corporation Driver intention estimation arrangement
JP2016020177A (en) * 2014-07-15 2016-02-04 株式会社デンソー Driving control device
DE102014216168A1 (en) 2014-08-14 2016-02-18 Bayerische Motoren Werke Aktiengesellschaft Control of a communication output of a service device of a motor vehicle
GB2532457B (en) 2014-11-19 2018-04-18 Jaguar Land Rover Ltd Dynamic control apparatus and related method
DE102014018913A1 (en) * 2014-12-17 2016-06-23 Daimler Ag Driver assistance device for a motor vehicle and method for operating such
US9308914B1 (en) 2015-01-23 2016-04-12 Denso International America, Inc. Advanced driver assistance system for vehicle
DE102015208570A1 (en) 2015-05-08 2016-11-10 Volkswagen Aktiengesellschaft Method for controlling vehicle functions
CN108137052B (en) * 2015-09-30 2021-09-07 索尼公司 Driving control device, driving control method, and computer-readable medium

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102012221036A1 (en) * 2012-11-19 2014-05-22 Bayerische Motoren Werke Aktiengesellschaft Method for automatic switching on of parking- and/or maneuvering assistance systems e.g. park distance control system in motor car, involves switching on parking- and/or maneuvering assistance systems based on evaluation of driver behavior

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190171211A1 (en) * 2016-08-09 2019-06-06 Nissan Motor Co., Ltd. Control Method and Control Device of Automatic Driving Vehicle
US10671071B2 (en) * 2016-08-09 2020-06-02 Nissan Motor Co., Ltd. Control method and control device of automatic driving vehicle

Also Published As

Publication number Publication date
DE102017208971A1 (en) 2018-11-29
CN108944938A (en) 2018-12-07
EP3409551A1 (en) 2018-12-05
US10745025B2 (en) 2020-08-18
KR20180130433A (en) 2018-12-07
EP3409551B1 (en) 2023-07-12
KR102050466B1 (en) 2019-11-29

Similar Documents

Publication Publication Date Title
US10745025B2 (en) Method and device for supporting a vehicle occupant in a vehicle
US9950707B2 (en) Method for controlling a vehicle in accordance with parameters preferred by an identified driver
CN107176165B (en) Vehicle control device
US10800428B2 (en) Vehicle driving assistance method and vehicle
US9283963B2 (en) Method for operating a driver assist system of an automobile providing a recommendation relating to a passing maneuver, and an automobile
US9650041B2 (en) Predictive human-machine interface using eye gaze technology, blind spot indicators and driver experience
EP2431225B1 (en) Method for an automotive hazardous detection and information system
EP2168815B1 (en) Method and device for detecting possibly colliding objects in a blind spot area
US9576208B2 (en) Emergency vehicle detection with digital image sensor
US11223775B2 (en) Method and apparatus for the spatially resolved detection of an object outside a transportation vehicle with the aid of a sensor installed in a transportation vehicle
US20170043720A1 (en) Camera system for displaying an area exterior to a vehicle
US10759334B2 (en) System for exchanging information between vehicles and control method thereof
JP7215228B2 (en) Control device, control method, control program
CN107472137B (en) Method and device for representing the environment of a motor vehicle
US20190361533A1 (en) Automated Activation of a Vision Support System
EP3892489B1 (en) Vehicle display device
GB2521274A (en) Emergency vehicle detection with digital image sensor
US20230141584A1 (en) Apparatus for displaying at least one virtual lane line based on environmental condition and method of controlling same
JP7432198B2 (en) Situation awareness estimation system and driving support system
EP4239598A1 (en) Method for determining an attentiveness of a driver of an automated vehicle
KR20230046798A (en) Method and apparatus for detecting blind spots of vehicle
KR20230168060A (en) Method and apparatus for supporting safety exit of a vehicle
WO2020167110A1 (en) System and method based on the correlation of interior and exterior visual information in a vehicle for improving safety in manual or semi-automatic driving
KR20230133994A (en) Method and Vehicle for recognizing traffic sign
KR20230155147A (en) Method and system for driver notification to respond to dangerous objects in the blind spot

Legal Events

Date Code Title Description
AS Assignment

Owner name: VOLKSWAGEN AG, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOESE, ESREF;ALI, AHMED;GEHRLING, SEBASTIAN;AND OTHERS;SIGNING DATES FROM 20180213 TO 20180216;REEL/FRAME:045776/0486

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4