US20180339713A1 - Method and device for supporting a vehicle occupant in a vehicle - Google Patents
Method and device for supporting a vehicle occupant in a vehicle Download PDFInfo
- Publication number
- US20180339713A1 US20180339713A1 US15/976,985 US201815976985A US2018339713A1 US 20180339713 A1 US20180339713 A1 US 20180339713A1 US 201815976985 A US201815976985 A US 201815976985A US 2018339713 A1 US2018339713 A1 US 2018339713A1
- Authority
- US
- United States
- Prior art keywords
- transportation vehicle
- sensor
- behavior pattern
- user information
- vehicle occupant
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 40
- 230000006399 behavior Effects 0.000 claims description 63
- 230000006870 function Effects 0.000 claims description 16
- 238000004458 analytical method Methods 0.000 claims description 13
- 230000008921 facial expression Effects 0.000 claims description 7
- 230000006978 adaptation Effects 0.000 claims description 4
- 230000004044 response Effects 0.000 claims 4
- 210000003128 head Anatomy 0.000 description 6
- 238000011156 evaluation Methods 0.000 description 3
- 230000001276 controlling effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 210000001747 pupil Anatomy 0.000 description 2
- 206010041349 Somnolence Diseases 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000003703 image analysis method Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
- 230000006641 stabilisation Effects 0.000 description 1
- 238000011105 stabilization Methods 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/08—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/10—Interpretation of driver requests or demands
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/12—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to parameters of the vehicle itself, e.g. tyre models
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/085—Changing the parameters of the control units, e.g. changing limit values, working points by control input
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/08—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
- B60W2040/0881—Seat occupation; Driver or passenger presence
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W2050/0062—Adapting control system settings
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W2050/0062—Adapting control system settings
- B60W2050/0075—Automatic parameter input, automatic initialising or calibrating means
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2400/00—Indexing codes relating to detected, measured or calculated conditions or factors
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2540/00—Input parameters relating to occupants
- B60W2540/22—Psychological state; Stress level or workload
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2540/00—Input parameters relating to occupants
- B60W2540/225—Direction of gaze
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2556/00—Input parameters relating to data
- B60W2556/10—Historical data
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2710/00—Output or target parameters relating to a particular sub-units
- B60W2710/18—Braking system
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2710/00—Output or target parameters relating to a particular sub-units
- B60W2710/20—Steering systems
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2710/00—Output or target parameters relating to a particular sub-units
- B60W2710/30—Auxiliary equipments
- B60W2710/305—Auxiliary equipments target power to auxiliaries
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Y—INDEXING SCHEME RELATING TO ASPECTS CROSS-CUTTING VEHICLE TECHNOLOGY
- B60Y2400/00—Special features of vehicle units
- B60Y2400/30—Sensors
Definitions
- Illustrative embodiments relate to a method for assisting a transportation vehicle occupant located in a transportation vehicle. Illustrative embodiments also relate to a device for carrying out the disclosed method and to a transportation vehicle which is configured for carrying out the disclosed method or which comprises the disclosed device.
- FIG. 2 schematically shows a block diagram of a device.
- a plurality of different types of assistance systems which can assist the driver in motorized transportation vehicles have become known.
- the assistance systems are limited only to signaling or warning functions, and in some cases the assistance systems intervene into the drive or the control of the transportation vehicle.
- Knowledge of the present transportation vehicle situation is required in both cases, which is obtained on the basis of data from suitable sensors.
- the sensors can ascertain, in this case, technical parameters of the transportation vehicle itself, such as for a driving stabilization system (ESP) or anti-lock system (ABS), as well as information regarding the transportation vehicle surroundings.
- ESP driving stabilization system
- ABS anti-lock system
- Various techniques are utilized in this case, frequently also in combination, for observing the transportation vehicle surroundings and assistance functions based thereon, such as parking assistance systems, automatic distance warnings and distance regulating functions, traffic sign recognition, blind spot monitoring, or emergency braking systems for pedestrian protection.
- ultrasonic, radar, LIDAR, and camera systems are utilized for identifying objects in the surroundings of the transportation vehicle, such as pedestrians, other transportation vehicles, traffic signs, etc.
- the sensor data from multiple sensors can be combined or linked, to improve the quality of the gathered information. If the information gathered from one or several sensors is not reliable, for example, this information can be partially or completely compensated for by data from further sensors. Likewise, however, information which is already available can also be confirmed by data from further sensors.
- sensors can be provided in the transportation vehicle, which gather information regarding the behavior and/or the state of a transportation vehicle occupant, in particular, the driver.
- the head and eye position of the driver can be detected and, by suitable image-analysis methods, the viewing direction of the driver can be ascertained or it can be detected that the driver is becoming fatigued.
- GB 2532457 A describes determining, with the aid of an image sensor, the attentiveness of the driver depending on the amount of time the driver looks at the road or looks away from the road and, based thereon, controlling the sensitivity of a driver assistance system.
- DE 10 2014 201 036 A1 describes an image-based classification method for ascertaining the driver state and/or the driver behavior, to ascertain the capability of a driver to take over.
- cameras are arranged in such a way that the eyes of the driver can be detected given any orientation of the head, to ascertain the viewing direction of the driver and statistically evaluate the viewing direction over time.
- Disclosed embodiments provide an improved method and an improved device for assisting a transportation vehicle occupant located in a transportation vehicle.
- Disclosed embodiments provide a method and a corresponding device.
- user information regarding the behavior and/or the state of a transportation vehicle occupant as well as transportation vehicle parameters of the transportation vehicle is gathered.
- the gathered user information and transportation vehicle parameters are combined and jointly analyzed.
- At least one behavior pattern of the transportation vehicle occupant is determined from the jointly analyzed user information and transportation vehicle parameters.
- the behavior pattern is stored together with the user information.
- this user information is compared with the user information regarding the at least one stored behavior pattern.
- a transportation vehicle function is automatically carried out.
- the disclosed method makes it possible to detect or learn individual behavior patterns and, based thereon, to assist the transportation vehicle occupant more efficiently, rapidly, or more comfortably in situations in which a behavior pattern is recognized.
- the user information and transportation vehicle parameters gathered before the operation are analyzed.
- the analysis and memory outlay can be reduced in this way.
- the user information includes the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
- the method can be applied when user information regarding the driver of the transportation vehicle is gathered.
- a disclosed device for assisting a transportation vehicle occupant located in a transportation vehicle comprises a first sensor for gathering user information regarding the behavior and/or the state of a transportation vehicle occupant and a second sensor for gathering transportation vehicle parameters.
- Sensor data from the first and second sensors are fed to an analysis and control unit.
- the analysis and control unit analyzes the sensor data, determines at least one behavior pattern of the transportation vehicle occupant, compares subsequent sensor data from the first sensor with the at least one stored behavior pattern and, if the behavior pattern is recognized, automatically activates a transportation vehicle function.
- the at least one behavior pattern is stored in a memory unit.
- the first sensor is a camera sensor in the transportation vehicle, which is oriented toward the transportation vehicle occupant and detects the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
- the second sensor ascertains one or several of the following transportation vehicle parameters:
- an adaptation of a user interface of the transportation vehicle takes place.
- a function of a driver assistance system is carried out.
- the disclosed method or the disclosed device is utilized in a motorized transportation vehicle.
- FIG. 1 schematically shows the method for assisting a transportation vehicle occupant located in a transportation vehicle.
- a gathering of user information takes place.
- arbitrary information regarding the user behavior and/or the user state of a transportation vehicle occupant can be ascertained in this case, provided this information is associated with the use of the transportation vehicle by the transportation vehicle occupant and, in the case that the same operation is carried out multiple times, is identical or at least similar and, therefore, after a learning phase, makes it possible to predict intentions of the transportation vehicle occupant when operations are carried out in the future. It is significant, in this case, to detect the viewing direction, the facial expression and/or the orientation of the head of a transportation vehicle occupant, in particular, of the driver of the transportation vehicle.
- transportation vehicle parameters are gathered in a method operation at 2 .
- these can be different types of parameters, such as the actuation of an operating element of the transportation vehicle, such as the turn-signal lamp, the steering wheel, the gas or brake pedal, the trajectory of the transportation vehicle, or the position of the transportation vehicle.
- method operation at 3 the user information and the transportation vehicle parameters are combined and jointly evaluated.
- method operation at 4 a behavior pattern is determined therefrom. Provided no behavior patterns were ascertained, for example, in an initialization phase, the behavior pattern is immediately stored in method operation at 5 , to allow for a recognition at a later time and, based thereon, to assist the transportation vehicle occupant. However, if one or several behavior patterns have already been learned and stored by the system, a comparison of the presently ascertained behavior pattern with the behavior patterns already present can take place in method operation at 6 . If a behavior pattern is recognized in this case, a transportation vehicle function can be automatically carried out in method operation at 7 .
- a function of a driver assistance system can be carried out or a user interface can be suitably adapted.
- this behavior pattern is also stored, to enable access thereto in the future.
- the driver In the case of a lane change, for example, on the highway, the driver normally glances over his/her shoulder and repeatedly looks in the side-view mirror before he/she carries out the lane change. According to the disclosure, the driver is then observed, for example, by a camera-based eye-tracking system. The glance over the shoulder and the repeated glances into the side-view mirror are detected by the eye-tracking system and are analyzed by suitable image-processing methods.
- transportation vehicle sensors monitor parameters of the transportation vehicle and detect when the turn signal is activated by the driver.
- a lane change of the transportation vehicle can be ascertained, for example, by sensors which detect the steering wheel or tire position, or by a compass function of a navigation system. If it is determined via the sensor data that the transportation vehicle has initially been traveling in a constant direction for a certain period of time, then a brief change in direction associated with the lane change takes place and the transportation vehicle then continues traveling in the original direction, it can be assumed that a lane change has taken place.
- a behavior pattern for the lane change can then be ascertained by a combination and a joint analysis of this user information and transportation vehicle parameters. If it is known via the transportation vehicle parameters that a lane change has taken place and the user information yields the finding that the driver has glanced over his/her shoulder and glanced in the side-view mirror before the lane change, this can be stored as a behavior pattern. If it is then ascertained at a later point in time that the driver is glancing over his/her shoulder and in the side-view mirror, this behavior pattern is recognized as being one of the known, stored behavior patterns.
- the driver assistance system can then respond to the present situation at an early point in time, for example, by way of automatically actuating the turn signal or warning the driver, by a display, for example, on or in the side-view mirror, about a transportation vehicle that is difficult to see or is approaching at a high rate of speed.
- the disclosed method can even assist the driver of a transportation vehicle when the driver wants to park the transportation vehicle into a parking space. If the behavior of the driver, transportation vehicle parameters, and, optionally, the transportation vehicle surroundings have been detected in earlier parking maneuvers, the corresponding behavior pattern can be recognized when parking maneuvers are carried out again.
- a camera-based eye-tracking system that the driver is looking at the transportation vehicle surroundings through the side windows, the windshield, or the rear window, or, optionally, a side-view mirror.
- suitable sensors for example, ultrasonic sensors, objects in the transportation vehicle surroundings as well as the distance of these objects from the transportation vehicle can be simultaneously detected and provide an indication that the driver has initially placed the transportation vehicle next to a parking space and has then maneuvered into the parking space.
- it can be evaluated whether the objects are located in the direction of the detected viewing direction of the driver.
- transportation vehicle sensors can gather parameters of the transportation vehicle such as the actuation of the turn signal, a large steering angle of the steering wheel, typically first in one direction and then in the opposite direction, and whether the transportation vehicle was moved forward or backward as a result. If the behavior pattern which is typical for a parking maneuver is then recognized at the beginning of a new parking maneuver, the parking assistance system can be automatically activated, for example.
- Yet another example of an application which is not limited to the assistance of the driver, but rather can also be utilized for other transportation vehicle occupants, in particular, for the passenger, relates to an automatic orientation of a sun visor.
- the pupil size can also be ascertained by suitable image-evaluation methods. A suddenly occurring, considerable increase in the pupil size can indicate that the transportation vehicle occupant is blinded. In addition to or instead thereof, a squinting of the eyes can also be registered by the image-evaluation method. If necessary, this information can also be combined with sensor data from a light sensor. If sensors are then additionally provided on the sun visor, which detect a lowering or tilting of the sun visor by the transportation vehicle occupant, the present information can be combined and a corresponding behavior pattern can be determined and stored.
- the sun visor can be automatically oriented, provided suitable actuators are present, without the need for the transportation vehicle occupant to be active therefor.
- the behavior patterns of different persons in the above-described examples of an application are usually similar, but are not identical.
- a decision is made depending on the transportation vehicle occupant during the learning of the behavior pattern and the assistance of the transportation vehicle occupant based thereon. This can take place, for example, by methods for facial recognition, but also on the basis of other biometric methods for personal identification in the transportation vehicle.
- FIG. 2 schematically shows a block diagram of a disclosed device.
- a camera 8 arranged in the transportation vehicle interior detects the head position of the driver and ascertains, by suitable image-evaluation methods, the eye position and viewing direction, but also, optionally, the facial expressions of the driver.
- This may be a near-infrared (NIR) camera which detects short-wave infrared radiation immediately adjacent to the visible range.
- NIR near-infrared
- an illumination in particular, in the dark, can take place by one or several NIR LEDs (not shown), without this being disturbing to the driver or being capable of blinding the driver.
- the NIR camera and the NIR LEDs can be installed separately or also in one shared sensor component.
- At least one sensor 9 is installed in the transportation vehicle for gathering the transportation vehicle parameters.
- at least one sensor 10 is provided for detecting the transportation vehicle surroundings.
- the sensor data are fed to an analysis and control unit 14 , in which the sensor data are evaluated, to determine behavior patterns, carry out a comparison with known behavior patterns and, depending thereon, to be capable of controlling, for example, an HMI 12 or a driver assistance system 13 .
- a memory 11 is provided for storing the known behavior patterns.
- the disclosed embodiments can be utilized in arbitrary areas of automotive engineering.
Landscapes
- Engineering & Computer Science (AREA)
- Automation & Control Theory (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Traffic Control Systems (AREA)
- Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
Abstract
Description
- This patent application claims priority to German Patent Application No. 10 2017 208 971.7, filed 29 May 2017, the disclosure of which is incorporated herein by reference in its entirety.
- Illustrative embodiments relate to a method for assisting a transportation vehicle occupant located in a transportation vehicle. Illustrative embodiments also relate to a device for carrying out the disclosed method and to a transportation vehicle which is configured for carrying out the disclosed method or which comprises the disclosed device.
- Further features of the disclosure are apparent from the following description and the claims in conjunction with the figures, in which:
-
FIG. 1 schematically shows a method for assisting a transportation vehicle occupant located in a transportation vehicle; and -
FIG. 2 schematically shows a block diagram of a device. - A plurality of different types of assistance systems which can assist the driver in motorized transportation vehicles have become known. In some cases, the assistance systems are limited only to signaling or warning functions, and in some cases the assistance systems intervene into the drive or the control of the transportation vehicle. Knowledge of the present transportation vehicle situation is required in both cases, which is obtained on the basis of data from suitable sensors. The sensors can ascertain, in this case, technical parameters of the transportation vehicle itself, such as for a driving stabilization system (ESP) or anti-lock system (ABS), as well as information regarding the transportation vehicle surroundings. Various techniques are utilized in this case, frequently also in combination, for observing the transportation vehicle surroundings and assistance functions based thereon, such as parking assistance systems, automatic distance warnings and distance regulating functions, traffic sign recognition, blind spot monitoring, or emergency braking systems for pedestrian protection. Frequently, ultrasonic, radar, LIDAR, and camera systems are utilized for identifying objects in the surroundings of the transportation vehicle, such as pedestrians, other transportation vehicles, traffic signs, etc. In this case, the sensor data from multiple sensors can be combined or linked, to improve the quality of the gathered information. If the information gathered from one or several sensors is not reliable, for example, this information can be partially or completely compensated for by data from further sensors. Likewise, however, information which is already available can also be confirmed by data from further sensors.
- In addition, sensors can be provided in the transportation vehicle, which gather information regarding the behavior and/or the state of a transportation vehicle occupant, in particular, the driver. By using a camera situated in the transportation vehicle interior, the head and eye position of the driver can be detected and, by suitable image-analysis methods, the viewing direction of the driver can be ascertained or it can be detected that the driver is becoming fatigued.
- DE 10 2015 122 603 A1 discloses a driver assistance system comprising a camera directed at the driver, to record the face of the driver, especially the eyes of the driver, and, based on the viewing direction, to monitor whether the driver is distracted or is looking away from the road. If the driver is looking away from the road, a driver status module detects an abnormal situation and activates the driver assistance system. Likewise, the driver status module can establish whether the driver is sleepy or awake, based on how wide and for how long the eyes of the driver are open.
- Similarly, GB 2532457 A describes determining, with the aid of an image sensor, the attentiveness of the driver depending on the amount of time the driver looks at the road or looks away from the road and, based thereon, controlling the sensitivity of a driver assistance system.
- DE 10 2014 201 036 A1 describes an image-based classification method for ascertaining the driver state and/or the driver behavior, to ascertain the capability of a driver to take over. For this purpose, cameras are arranged in such a way that the eyes of the driver can be detected given any orientation of the head, to ascertain the viewing direction of the driver and statistically evaluate the viewing direction over time.
- Disclosed embodiments provide an improved method and an improved device for assisting a transportation vehicle occupant located in a transportation vehicle.
- Disclosed embodiments provide a method and a corresponding device.
- In the disclosed method for assisting a transportation vehicle occupant located in a transportation vehicle, user information regarding the behavior and/or the state of a transportation vehicle occupant as well as transportation vehicle parameters of the transportation vehicle is gathered. The gathered user information and transportation vehicle parameters are combined and jointly analyzed. At least one behavior pattern of the transportation vehicle occupant is determined from the jointly analyzed user information and transportation vehicle parameters. The behavior pattern is stored together with the user information. In a subsequent gathering of user information, this user information is compared with the user information regarding the at least one stored behavior pattern. When there is conformance with the behavior pattern, a transportation vehicle function is automatically carried out.
- Contrary to the conventional assistance systems which respond in a standardized manner to the input from a transportation vehicle occupant, in particular, the driver, the disclosed method makes it possible to detect or learn individual behavior patterns and, based thereon, to assist the transportation vehicle occupant more efficiently, rapidly, or more comfortably in situations in which a behavior pattern is recognized.
- According to at least one disclosed embodiment, when there is a change to a transportation vehicle parameter due to an operation carried out by the transportation vehicle occupant, the user information and transportation vehicle parameters gathered before the operation are analyzed. The analysis and memory outlay can be reduced in this way.
- The user information includes the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
- The method can be applied when user information regarding the driver of the transportation vehicle is gathered.
- Correspondingly, a disclosed device for assisting a transportation vehicle occupant located in a transportation vehicle comprises a first sensor for gathering user information regarding the behavior and/or the state of a transportation vehicle occupant and a second sensor for gathering transportation vehicle parameters. Sensor data from the first and second sensors are fed to an analysis and control unit. The analysis and control unit analyzes the sensor data, determines at least one behavior pattern of the transportation vehicle occupant, compares subsequent sensor data from the first sensor with the at least one stored behavior pattern and, if the behavior pattern is recognized, automatically activates a transportation vehicle function. The at least one behavior pattern is stored in a memory unit.
- Optionally, the first sensor is a camera sensor in the transportation vehicle, which is oriented toward the transportation vehicle occupant and detects the viewing direction and/or the facial expression and/or the orientation of the head of a transportation vehicle occupant.
- The second sensor ascertains one or several of the following transportation vehicle parameters:
-
- the actuation of an operating element of the transportation vehicle, such as the turn-signal lamp, the steering wheel, the gas or brake pedal;
- the position and/or trajectory of the transportation vehicle.
- Likewise, optionally, a third sensor may be provided for detecting transportation vehicle-external objects in the transportation vehicle surroundings and the sensor data from the third sensor are likewise fed to the analysis and control unit and are analyzed thereby.
- According to at least one disclosed embodiment, when there is conformance with the behavior pattern, an adaptation of a user interface of the transportation vehicle takes place.
- According to yet another disclosed embodiment, when there is conformance with the behavior pattern, a function of a driver assistance system is carried out.
- Optionally, the disclosed method or the disclosed device is utilized in a motorized transportation vehicle.
- Disclosed embodiments are described in greater detail in the following with reference to the figures, to improve the understanding of the principles of the present disclosure. It is understood that the disclosed is not limited to these embodiments and that the described features can also be combined or modified without departing from the scope of protection of the disclosure as it is defined in the claims.
-
FIG. 1 schematically shows the method for assisting a transportation vehicle occupant located in a transportation vehicle. According to method operation at 1, a gathering of user information takes place. In principle, arbitrary information regarding the user behavior and/or the user state of a transportation vehicle occupant can be ascertained in this case, provided this information is associated with the use of the transportation vehicle by the transportation vehicle occupant and, in the case that the same operation is carried out multiple times, is identical or at least similar and, therefore, after a learning phase, makes it possible to predict intentions of the transportation vehicle occupant when operations are carried out in the future. It is significant, in this case, to detect the viewing direction, the facial expression and/or the orientation of the head of a transportation vehicle occupant, in particular, of the driver of the transportation vehicle. - In parallel to the gathering of the user information, transportation vehicle parameters are gathered in a method operation at 2. Depending on the application, these can be different types of parameters, such as the actuation of an operating element of the transportation vehicle, such as the turn-signal lamp, the steering wheel, the gas or brake pedal, the trajectory of the transportation vehicle, or the position of the transportation vehicle.
- In method operation at 3, the user information and the transportation vehicle parameters are combined and jointly evaluated. In method operation at 4, a behavior pattern is determined therefrom. Provided no behavior patterns were ascertained, for example, in an initialization phase, the behavior pattern is immediately stored in method operation at 5, to allow for a recognition at a later time and, based thereon, to assist the transportation vehicle occupant. However, if one or several behavior patterns have already been learned and stored by the system, a comparison of the presently ascertained behavior pattern with the behavior patterns already present can take place in method operation at 6. If a behavior pattern is recognized in this case, a transportation vehicle function can be automatically carried out in method operation at 7. In this case, for example, a function of a driver assistance system can be carried out or a user interface can be suitably adapted. However, if there is a new behavior pattern that has not been known so far, this behavior pattern is also stored, to enable access thereto in the future.
- Exemplary embodiments are mentioned in the following for further illustration.
- In the case of a lane change, for example, on the highway, the driver normally glances over his/her shoulder and repeatedly looks in the side-view mirror before he/she carries out the lane change. According to the disclosure, the driver is then observed, for example, by a camera-based eye-tracking system. The glance over the shoulder and the repeated glances into the side-view mirror are detected by the eye-tracking system and are analyzed by suitable image-processing methods.
- At the same time, transportation vehicle sensors monitor parameters of the transportation vehicle and detect when the turn signal is activated by the driver. In addition, a lane change of the transportation vehicle can be ascertained, for example, by sensors which detect the steering wheel or tire position, or by a compass function of a navigation system. If it is determined via the sensor data that the transportation vehicle has initially been traveling in a constant direction for a certain period of time, then a brief change in direction associated with the lane change takes place and the transportation vehicle then continues traveling in the original direction, it can be assumed that a lane change has taken place.
- A behavior pattern for the lane change can then be ascertained by a combination and a joint analysis of this user information and transportation vehicle parameters. If it is known via the transportation vehicle parameters that a lane change has taken place and the user information yields the finding that the driver has glanced over his/her shoulder and glanced in the side-view mirror before the lane change, this can be stored as a behavior pattern. If it is then ascertained at a later point in time that the driver is glancing over his/her shoulder and in the side-view mirror, this behavior pattern is recognized as being one of the known, stored behavior patterns. The driver assistance system can then respond to the present situation at an early point in time, for example, by way of automatically actuating the turn signal or warning the driver, by a display, for example, on or in the side-view mirror, about a transportation vehicle that is difficult to see or is approaching at a high rate of speed.
- The disclosed method can even assist the driver of a transportation vehicle when the driver wants to park the transportation vehicle into a parking space. If the behavior of the driver, transportation vehicle parameters, and, optionally, the transportation vehicle surroundings have been detected in earlier parking maneuvers, the corresponding behavior pattern can be recognized when parking maneuvers are carried out again.
- For example, it can be ascertained with the aid of a camera-based eye-tracking system that the driver is looking at the transportation vehicle surroundings through the side windows, the windshield, or the rear window, or, optionally, a side-view mirror. By suitable sensors, for example, ultrasonic sensors, objects in the transportation vehicle surroundings as well as the distance of these objects from the transportation vehicle can be simultaneously detected and provide an indication that the driver has initially placed the transportation vehicle next to a parking space and has then maneuvered into the parking space. In addition, it can be evaluated whether the objects are located in the direction of the detected viewing direction of the driver. In addition, as in the aforementioned example of an application, transportation vehicle sensors can gather parameters of the transportation vehicle such as the actuation of the turn signal, a large steering angle of the steering wheel, typically first in one direction and then in the opposite direction, and whether the transportation vehicle was moved forward or backward as a result. If the behavior pattern which is typical for a parking maneuver is then recognized at the beginning of a new parking maneuver, the parking assistance system can be automatically activated, for example.
- Yet another example of an application which is not limited to the assistance of the driver, but rather can also be utilized for other transportation vehicle occupants, in particular, for the passenger, relates to an automatic orientation of a sun visor.
- If the eyes of the transportation vehicle occupant are detected, the pupil size can also be ascertained by suitable image-evaluation methods. A suddenly occurring, considerable increase in the pupil size can indicate that the transportation vehicle occupant is blinded. In addition to or instead thereof, a squinting of the eyes can also be registered by the image-evaluation method. If necessary, this information can also be combined with sensor data from a light sensor. If sensors are then additionally provided on the sun visor, which detect a lowering or tilting of the sun visor by the transportation vehicle occupant, the present information can be combined and a corresponding behavior pattern can be determined and stored. If it is established at a later point in time, on the basis of the observation of the eyes of the transportation vehicle occupant, that the transportation vehicle occupant is being blinded again, the sun visor can be automatically oriented, provided suitable actuators are present, without the need for the transportation vehicle occupant to be active therefor.
- The behavior patterns of different persons in the above-described examples of an application are usually similar, but are not identical. To increase the accuracy of the assistance system, it can therefore be provided that a decision is made depending on the transportation vehicle occupant during the learning of the behavior pattern and the assistance of the transportation vehicle occupant based thereon. This can take place, for example, by methods for facial recognition, but also on the basis of other biometric methods for personal identification in the transportation vehicle.
-
FIG. 2 schematically shows a block diagram of a disclosed device. Acamera 8 arranged in the transportation vehicle interior, for example, in the instrument panel or in the area of the rear-view mirror, detects the head position of the driver and ascertains, by suitable image-evaluation methods, the eye position and viewing direction, but also, optionally, the facial expressions of the driver. This may be a near-infrared (NIR) camera which detects short-wave infrared radiation immediately adjacent to the visible range. In this case, an illumination, in particular, in the dark, can take place by one or several NIR LEDs (not shown), without this being disturbing to the driver or being capable of blinding the driver. The NIR camera and the NIR LEDs can be installed separately or also in one shared sensor component. At least one sensor 9 is installed in the transportation vehicle for gathering the transportation vehicle parameters. In addition, at least onesensor 10 is provided for detecting the transportation vehicle surroundings. The sensor data are fed to an analysis andcontrol unit 14, in which the sensor data are evaluated, to determine behavior patterns, carry out a comparison with known behavior patterns and, depending thereon, to be capable of controlling, for example, anHMI 12 or adriver assistance system 13. Amemory 11 is provided for storing the known behavior patterns. - The disclosed embodiments can be utilized in arbitrary areas of automotive engineering.
-
- 1 method operation including gathering of user information
- 2 method operation including gathering of transportation vehicle parameters
- 3 method operation including analysis of the user information and transportation vehicle parameters
- 4 method operation including determination of a behavior pattern
- 5 method operation including storing the behavior pattern
- 6 method operation including checking whether a known behavior pattern is present
- 7 method operation including carrying out a transportation vehicle function
- 8 camera for observing the transportation vehicle occupant
- 9 sensor for transportation vehicle parameters
- 10 sensor for detecting the transportation vehicle surroundings
- 11 memory
- 12 HMI
- 13 assistance system
- 14 analysis and control unit
Claims (21)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102017208971.7A DE102017208971A1 (en) | 2017-05-29 | 2017-05-29 | METHOD AND DEVICE FOR SUPPORTING A VEHICLE IN A VEHICLE |
DE102017208971 | 2017-05-29 | ||
DE102017208971.7 | 2017-05-29 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20180339713A1 true US20180339713A1 (en) | 2018-11-29 |
US10745025B2 US10745025B2 (en) | 2020-08-18 |
Family
ID=61189267
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/976,985 Active 2039-01-01 US10745025B2 (en) | 2017-05-29 | 2018-05-11 | Method and device for supporting a vehicle occupant in a vehicle |
Country Status (5)
Country | Link |
---|---|
US (1) | US10745025B2 (en) |
EP (1) | EP3409551B1 (en) |
KR (1) | KR102050466B1 (en) |
CN (1) | CN108944938A (en) |
DE (1) | DE102017208971A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190171211A1 (en) * | 2016-08-09 | 2019-06-06 | Nissan Motor Co., Ltd. | Control Method and Control Device of Automatic Driving Vehicle |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102017219365B4 (en) | 2017-10-27 | 2023-03-30 | Zf Friedrichshafen Ag | Apparatus and method for adaptive vehicle control |
CN111319626B (en) * | 2020-03-31 | 2021-07-20 | 深圳市英威腾电动汽车驱动技术有限公司 | Vehicle control parameter adjusting method and device, electronic equipment and storage medium |
DE102020124910A1 (en) | 2020-09-24 | 2022-03-24 | Bayerische Motoren Werke Aktiengesellschaft | Method for supporting a driver when driving a motor vehicle manually, support device and motor vehicle |
DE102022103912A1 (en) | 2022-02-18 | 2023-08-24 | Bayerische Motoren Werke Aktiengesellschaft | Method for monitoring a condition of a driver of a vehicle through a camera in an interior of the vehicle, computer-readable medium, system and vehicle |
DE102022132377A1 (en) | 2022-12-06 | 2024-06-06 | Bayerische Motoren Werke Aktiengesellschaft | Method for operating an automated driving function, method for training an artificial intelligence and processing device |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102012221036A1 (en) * | 2012-11-19 | 2014-05-22 | Bayerische Motoren Werke Aktiengesellschaft | Method for automatic switching on of parking- and/or maneuvering assistance systems e.g. park distance control system in motor car, involves switching on parking- and/or maneuvering assistance systems based on evaluation of driver behavior |
Family Cites Families (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10255016A (en) | 1997-03-13 | 1998-09-25 | Nissan Motor Co Ltd | Device for predicting driver's operation by blink |
DE10126224A1 (en) | 2001-05-30 | 2002-12-12 | Bosch Gmbh Robert | Method and device for characterizing the condition of the driver of a motor vehicle |
EP1284432A1 (en) | 2001-08-14 | 2003-02-19 | Ford Global Technologies, Inc., A subsidiary of Ford Motor Company | Device and process for operating a driver assistance system |
DE10343683A1 (en) | 2003-09-20 | 2005-04-21 | Daimler Chrysler Ag | Information system for motor vehicles |
JP5286035B2 (en) * | 2008-11-06 | 2013-09-11 | 本田技研工業株式会社 | Vehicle speed control device |
DE102011055685A1 (en) * | 2011-11-24 | 2013-05-29 | Continental Teves Ag & Co. Ohg | Method for operating system influencing driving behavior of vehicle, involves carrying out automatic adjustment of adjustable control parameter of system manually by vehicle occupant |
TWI447039B (en) * | 2011-11-25 | 2014-08-01 | Driving behavior analysis and warning system and method thereof | |
DE102012004601A1 (en) | 2012-03-07 | 2012-10-11 | Daimler Ag | Assistance system for rider for offering and/or signaling information to passenger of motor car, has control device for controlling output device such that occupants offer and/or signal information in context-sensitive manner |
US20130325202A1 (en) | 2012-06-01 | 2013-12-05 | GM Global Technology Operations LLC | Neuro-cognitive driver state processing |
KR20140002373A (en) * | 2012-06-29 | 2014-01-08 | 현대자동차주식회사 | Apparatus and method for monitoring driver state by driving pattern learning |
BR112015025117A2 (en) * | 2013-04-12 | 2017-07-18 | Dana Ltd | method for vehicle and operator orientation through pattern recognition |
US9103688B2 (en) | 2013-08-21 | 2015-08-11 | Continental Automotive Systems, Inc. | Adapting vehicle personality using analyzed driver performance metrics |
DE102013217552A1 (en) | 2013-09-03 | 2015-03-05 | Bayerische Motoren Werke Aktiengesellschaft | Situation-dependent activation of driver assistance functions |
KR20150066308A (en) * | 2013-12-06 | 2015-06-16 | 한국전자통신연구원 | Apparatus and method for determining driving condition of deiver |
DE102014201036A1 (en) | 2014-01-21 | 2015-07-23 | Bayerische Motoren Werke Aktiengesellschaft | Image-based classification of driver status and / or driver behavior |
EP2923912B1 (en) * | 2014-03-24 | 2018-12-26 | Volvo Car Corporation | Driver intention estimation arrangement |
JP2016020177A (en) * | 2014-07-15 | 2016-02-04 | 株式会社デンソー | Driving control device |
DE102014216168A1 (en) | 2014-08-14 | 2016-02-18 | Bayerische Motoren Werke Aktiengesellschaft | Control of a communication output of a service device of a motor vehicle |
GB2532457B (en) | 2014-11-19 | 2018-04-18 | Jaguar Land Rover Ltd | Dynamic control apparatus and related method |
DE102014018913A1 (en) * | 2014-12-17 | 2016-06-23 | Daimler Ag | Driver assistance device for a motor vehicle and method for operating such |
US9308914B1 (en) | 2015-01-23 | 2016-04-12 | Denso International America, Inc. | Advanced driver assistance system for vehicle |
DE102015208570A1 (en) | 2015-05-08 | 2016-11-10 | Volkswagen Aktiengesellschaft | Method for controlling vehicle functions |
CN108137052B (en) * | 2015-09-30 | 2021-09-07 | 索尼公司 | Driving control device, driving control method, and computer-readable medium |
-
2017
- 2017-05-29 DE DE102017208971.7A patent/DE102017208971A1/en active Pending
-
2018
- 2018-02-09 EP EP18155937.8A patent/EP3409551B1/en active Active
- 2018-03-30 KR KR1020180037360A patent/KR102050466B1/en active IP Right Grant
- 2018-05-11 US US15/976,985 patent/US10745025B2/en active Active
- 2018-05-29 CN CN201810529250.2A patent/CN108944938A/en active Pending
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102012221036A1 (en) * | 2012-11-19 | 2014-05-22 | Bayerische Motoren Werke Aktiengesellschaft | Method for automatic switching on of parking- and/or maneuvering assistance systems e.g. park distance control system in motor car, involves switching on parking- and/or maneuvering assistance systems based on evaluation of driver behavior |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190171211A1 (en) * | 2016-08-09 | 2019-06-06 | Nissan Motor Co., Ltd. | Control Method and Control Device of Automatic Driving Vehicle |
US10671071B2 (en) * | 2016-08-09 | 2020-06-02 | Nissan Motor Co., Ltd. | Control method and control device of automatic driving vehicle |
Also Published As
Publication number | Publication date |
---|---|
DE102017208971A1 (en) | 2018-11-29 |
CN108944938A (en) | 2018-12-07 |
EP3409551A1 (en) | 2018-12-05 |
US10745025B2 (en) | 2020-08-18 |
KR20180130433A (en) | 2018-12-07 |
EP3409551B1 (en) | 2023-07-12 |
KR102050466B1 (en) | 2019-11-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10745025B2 (en) | Method and device for supporting a vehicle occupant in a vehicle | |
US9950707B2 (en) | Method for controlling a vehicle in accordance with parameters preferred by an identified driver | |
CN107176165B (en) | Vehicle control device | |
US10800428B2 (en) | Vehicle driving assistance method and vehicle | |
US9283963B2 (en) | Method for operating a driver assist system of an automobile providing a recommendation relating to a passing maneuver, and an automobile | |
US9650041B2 (en) | Predictive human-machine interface using eye gaze technology, blind spot indicators and driver experience | |
EP2431225B1 (en) | Method for an automotive hazardous detection and information system | |
EP2168815B1 (en) | Method and device for detecting possibly colliding objects in a blind spot area | |
US9576208B2 (en) | Emergency vehicle detection with digital image sensor | |
US11223775B2 (en) | Method and apparatus for the spatially resolved detection of an object outside a transportation vehicle with the aid of a sensor installed in a transportation vehicle | |
US20170043720A1 (en) | Camera system for displaying an area exterior to a vehicle | |
US10759334B2 (en) | System for exchanging information between vehicles and control method thereof | |
JP7215228B2 (en) | Control device, control method, control program | |
CN107472137B (en) | Method and device for representing the environment of a motor vehicle | |
US20190361533A1 (en) | Automated Activation of a Vision Support System | |
EP3892489B1 (en) | Vehicle display device | |
GB2521274A (en) | Emergency vehicle detection with digital image sensor | |
US20230141584A1 (en) | Apparatus for displaying at least one virtual lane line based on environmental condition and method of controlling same | |
JP7432198B2 (en) | Situation awareness estimation system and driving support system | |
EP4239598A1 (en) | Method for determining an attentiveness of a driver of an automated vehicle | |
KR20230046798A (en) | Method and apparatus for detecting blind spots of vehicle | |
KR20230168060A (en) | Method and apparatus for supporting safety exit of a vehicle | |
WO2020167110A1 (en) | System and method based on the correlation of interior and exterior visual information in a vehicle for improving safety in manual or semi-automatic driving | |
KR20230133994A (en) | Method and Vehicle for recognizing traffic sign | |
KR20230155147A (en) | Method and system for driver notification to respond to dangerous objects in the blind spot |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: VOLKSWAGEN AG, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOESE, ESREF;ALI, AHMED;GEHRLING, SEBASTIAN;AND OTHERS;SIGNING DATES FROM 20180213 TO 20180216;REEL/FRAME:045776/0486 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |