US20160039285A1 - Scene awareness system for a vehicle - Google Patents
Scene awareness system for a vehicle Download PDFInfo
- Publication number
- US20160039285A1 US20160039285A1 US14/920,420 US201514920420A US2016039285A1 US 20160039285 A1 US20160039285 A1 US 20160039285A1 US 201514920420 A US201514920420 A US 201514920420A US 2016039285 A1 US2016039285 A1 US 2016039285A1
- Authority
- US
- United States
- Prior art keywords
- user
- vehicle
- scene
- data pertaining
- gaze
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 claims abstract description 37
- 230000000007 visual effect Effects 0.000 claims abstract description 23
- 238000004891 communication Methods 0.000 claims description 12
- 210000003128 head Anatomy 0.000 claims description 5
- 230000033001 locomotion Effects 0.000 description 7
- 238000003384 imaging method Methods 0.000 description 4
- 230000003068 static effect Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000007613 environmental effect Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 239000000872 buffer Substances 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000003750 conditioning effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 239000000446 fuel Substances 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000010399 physical interaction Effects 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 239000012780 transparent material Substances 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K28/00—Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions
- B60K28/02—Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions responsive to conditions relating to the driver
- B60K28/06—Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions responsive to conditions relating to the driver responsive to incapacity of driver
- B60K28/066—Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions responsive to conditions relating to the driver responsive to incapacity of driver actuating a signalling device
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/113—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for determining or recording eye movement
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/0205—Simultaneously evaluating both cardiovascular conditions and different types of body conditions, e.g. heart and respiratory condition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6887—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient mounted on external non-worn devices, e.g. non-medical devices
- A61B5/6893—Cars
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60C—VEHICLE TYRES; TYRE INFLATION; TYRE CHANGING; CONNECTING VALVES TO INFLATABLE ELASTIC BODIES IN GENERAL; DEVICES OR ARRANGEMENTS RELATED TO TYRES
- B60C9/00—Reinforcements or ply arrangement of pneumatic tyres
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
- B60K35/20—Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor
- B60K35/21—Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor using visual output, e.g. blinking lights or matrix displays
- B60K35/23—Head-up displays [HUD]
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
- B60K35/20—Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor
- B60K35/28—Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor characterised by the type of the output information, e.g. video entertainment or vehicle dynamics information; characterised by the purpose of the output information, e.g. for attracting the attention of the driver
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
- B60K35/20—Output arrangements, i.e. from vehicle to user, associated with vehicle functions or specially adapted therefor
- B60K35/29—Instruments characterised by the way in which information is handled, e.g. showing information on plural displays or prioritising information according to driving conditions
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
- B60K35/60—Instruments characterised by their location or relative disposition in or on vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
- B60K35/80—Arrangements for controlling instruments
- B60K35/81—Arrangements for controlling instruments for controlling displays
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q9/00—Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/09—Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
- G02B27/0927—Systems for changing the beam intensity distribution, e.g. Gaussian to top-hat
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/09—Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
- G02B27/0938—Using specific optical elements
- G02B27/095—Refractive optical elements
- G02B27/0955—Lenses
- G02B27/0961—Lens arrays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/0001—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings specially adapted for lighting devices or systems
- G02B6/0011—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings specially adapted for lighting devices or systems the light guides being planar or of plate-like form
- G02B6/0033—Means for improving the coupling-out of light from the light guide
- G02B6/005—Means for improving the coupling-out of light from the light guide provided by one optical element, or plurality thereof, placed on the light output side of the light guide
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D3/00—Control of position or direction
-
- G06K9/00845—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/59—Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
- G06V20/597—Recognising the driver's state or behaviour, e.g. attention or drowsiness
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C5/00—Registering or indicating the working of vehicles
- G07C5/02—Registering or indicating driving, working, idle, or waiting time only
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C5/00—Registering or indicating the working of vehicles
- G07C5/08—Registering or indicating performance data other than driving, working, idle, or waiting time, with or without registering driving, working, idle or waiting time
-
- G—PHYSICS
- G08—SIGNALLING
- G08C—TRANSMISSION SYSTEMS FOR MEASURED VALUES, CONTROL OR SIMILAR SIGNALS
- G08C17/00—Arrangements for transmitting signals characterised by the use of a wireless electrical link
- G08C17/02—Arrangements for transmitting signals characterised by the use of a wireless electrical link using a radio link
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/024—Detecting, measuring or recording pulse rate or heart rate
- A61B5/02405—Determining heart rate variability
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/053—Measuring electrical impedance or conductance of a portion of the body
- A61B5/0531—Measuring skin impedance
- A61B5/0533—Measuring galvanic skin response
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K2360/00—Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
- B60K2360/16—Type of output information
- B60K2360/177—Augmented reality
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K2360/00—Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
- B60K2360/18—Information management
- B60K2360/197—Blocking or enabling of input functions
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K2360/00—Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
- B60K2360/20—Optical features of instruments
- B60K2360/21—Optical features of instruments using cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K2360/00—Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
- B60K2360/20—Optical features of instruments
- B60K2360/25—Optical features of instruments using filters
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K2360/00—Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
- B60K2360/20—Optical features of instruments
- B60K2360/33—Illumination features
- B60K2360/334—Projection means
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K2360/00—Indexing scheme associated with groups B60K35/00 or B60K37/00 relating to details of instruments or dashboards
- B60K2360/77—Instrument locations other than the dashboard
- B60K2360/785—Instrument locations other than the dashboard on or in relation to the windshield or windows
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/8006—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for monitoring and displaying scenes of vehicle interior, e.g. for monitoring passengers or cargo
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/8093—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for obstacle warning
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
- B60W2050/146—Display means
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Y—INDEXING SCHEME RELATING TO ASPECTS CROSS-CUTTING VEHICLE TECHNOLOGY
- B60Y2302/00—Responses or measures related to driver conditions
- B60Y2302/03—Actuating a signal or alarm device
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0112—Head-up displays characterised by optical features comprising device for genereting colour display
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0118—Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0141—Head-up displays characterised by optical features characterised by the informative content of the display
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B2027/0192—Supplementary details
- G02B2027/0194—Supplementary details with combiner of laminated type, for optical or mechanical aspects
Definitions
- the present disclosure is related to an adaptive external road scene transfer to improve driver awareness.
- Vehicles such as cars, typically include displays or indicators to provide information to the vehicle user.
- Such displays or indicators may, for example, provide information regarding mileage, fuel consumption, and vehicle speed.
- the vehicle user usually has to shift his eye gaze away from the road scene and onto an in-vehicle display in order to visually process the information presented by these displays or indicators.
- One possible aspect of the disclosure provides a method of alerting a user of a vehicle as to a scene, external to the vehicle, which includes capturing data pertaining to the scene, external to the vehicle.
- the captured data pertaining to the scene external to the vehicle, is transmitted to a processor.
- the determination is made, in the processor that a characteristic is in the scene that requires the user's visual attention.
- Data pertaining to a gaze of the user is captured.
- Captured data, pertaining to the gaze of the user is transmitted to the processor.
- a determination is made, in the processor that the user is gazing toward a user device.
- a signal is transmitted to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.
- a scene awareness system for a vehicle that includes at least one exterior camera, at least one interior camera, and a vehicle controller.
- the exterior camera is configured to capture data pertaining to a scene, external to the vehicle.
- the interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle.
- the vehicle controller is in communication with each of the at least one exterior camera and the at least one interior camera.
- the vehicle controller is configured to determine whether at least one characteristic in the scene, exterior to the vehicle, requires the user's visual attention.
- the vehicle controller is also configured to determine whether the user is gazing at a user device.
- the vehicle controller is configured to transmit a signal to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention when the controller determines there is at least one characteristic in the scene requiring the user's attention, simultaneous with the controller determining the user is gazing at the user device.
- a vehicle in yet another aspect of the disclosure, includes a body and a scene awareness system.
- the scene awareness system includes at least one exterior camera, at least one interior camera, and a vehicle controller.
- the cameras are operatively attached to the body.
- the exterior camera is configured to capture data pertaining to a scene, external to the vehicle.
- the interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle.
- the vehicle controller is in communication with each of the cameras.
- the vehicle controller is operable for receiving data pertaining to the scene, external to the vehicle, from the at least one exterior camera.
- the vehicle controller includes a process and a characteristic in the scene requiring the user's visual attention that is determined in the processor.
- Data, pertaining to a gaze of the user is received from the interior camera.
- a determination is made that the user is gazing toward a user device.
- a signal is transmitted to the user device, from the controller that alerts the user as to the characteristic in the scene requiring the user's
- FIG. 1 is a schematic illustrative side view of a vehicle.
- FIG. 2 is a schematic diagrammatic view of a scene awareness system of the vehicle of FIG. 1 .
- FIG. 3A is schematic fragmentary view of an interior of the vehicle having a scene awareness system, illustrating the user of the vehicle gazing at a scene, external to the vehicle.
- FIG. 3B is a schematic fragmentary view of the interior of the vehicle having the scene awareness system, illustrating the user of the vehicle gazing at a user device.
- FIG. 4 is a schematic flow chart diagram of a method of alerting the user of the vehicle as to a scene, external to the vehicle, requiring the user's attention.
- FIG. 1 schematically illustrates a vehicle 10 including a body 12 .
- the vehicle 10 may be a land vehicle, such as a car, or any other type of vehicle such as an airplane, farm equipment, construction equipment, a boat, etc.
- the vehicle 10 may be configured to be operated by a user 18 , operated autonomously, and/or operated semi-autonomously.
- the vehicle 10 includes a windshield 14 coupled to the body 12 .
- the windshield 14 is wholly or partly made of a substantially transparent material.
- the vehicle 10 includes an interior 16 .
- a vehicle user 18 can be in position within the interior 16 and see through the windshield 14 , external to the vehicle 10 .
- the body 12 extends between a forward end 21 and a rearward end 23 .
- the windshield 14 faces the forward end 21 of the vehicle 10 , opposite the rearward end 23 .
- a scene awareness system 25 having one or more exterior cameras 26 , one or more internal cameras 28 , and a vehicle controller (C) 20 .
- the vehicle controller 20 which is in communication with the internal and the exterior camera(s) 28 , 26 , is programmed to use the exterior camera(s) 26 to automatically locate and/or identify one or more characteristics in a scene 32 , exterior to the vehicle 10 , and to use the interior camera(s) 28 to simultaneously locate and identify where the user's 18 eyes are gazing within the vehicle 10 via execution of instructions embodying a method 100 , an example of which is described below with reference to FIG. 4 .
- the characteristics may include, but should not be limited to, identifying one or more objects 30 , colors, brightness, darkness, temperatures, temperature gradients, graphical patterns, motion, and other like characteristics.
- the exterior camera(s) 26 is configured to monitor a scene 32 , external to the vehicle 10 .
- the exterior cameras 26 may be operatively attached to the vehicle 10 to view the scenery 32 in front of, beside, above, and/or behind the vehicle 10 .
- at least one exterior camera 26 may be resident within a user device 42 located within the interior 16 of the vehicle 10 through the windshield 14 or some other window within the vehicle 10 .
- the user device 42 would be positioned within the vehicle such that the exterior camera 26 is directed to view the scenery 32 around the vehicle.
- the scenery 32 may be of a landscape, a billboard, a landmark, and the like.
- the exterior camera(s) 26 may include a camera, a sensor, and the like.
- the exterior camera(s) 26 is configured to capture images, in static or in motion, external to the vehicle 10 , in real-time.
- the exterior camera(s) 26 is in operative communication with a processor 22 configured to determine if the scenery 32 captured in real-time is sufficient to require the user's 18 visual attention, i.e., the scenery 32 is visually salient.
- the processor 22 may be resident within the vehicle controller 20 .
- the exterior camera(s) 26 may be configured to use machine vision (MV) to recognize characteristics, such as objects 30 , in the scenery 32 , in real-time.
- MV machine vision
- the captured images may be transmitted as data (arrow 38 ) from the exterior camera(s) 26 to the processor 22 within the controller to make an imaging-based determination of whether the scenery 32 is visually salient.
- Machine vision may make an imaging-based determination based on edge detection of objects 30 , color analysis to identify objects 30 , pattern recognition, feature detection, motion analysis, and the like.
- the processor 22 may be pre-programmed with criteria.
- scenery 32 sufficient to require the user's 18 visual attention may include an upcoming curve in the road, stop sign, landmark, destination, and the like.
- such scenery 32 may include objects 30 which are static and/or objects which are in motion.
- the interior camera(s) 28 is also in operative communication with the processor 22 .
- the interior camera(s) 28 may be used to capture visual information with respect to the user 18 of the vehicle 10 . More specifically, the interior camera(s) 28 may capture information about the users 18 head, including their eyes, in real-time.
- the interior camera(s) 28 may be positioned within the vehicle 10 , e.g., attached to a rearview mirror, and the like. Alternatively, the interior camera(s) 28 may be positioned within the vehicle, where the interior camera 28 is resident within the user device 42 . In such an example, the user device 42 may be configured with both the interior camera 28 and the exterior camera 26 .
- the captured visual information may be transmitted as data (arrow 40 ) to the processor 22 within the vehicle controller 20 .
- the interior camera(s) 28 is configured to determine where the eyes of the user 18 are looking and/or to determine an orientation of the user's 18 head, relative to the windshield 14 and/or the forward end 21 of the vehicle 10 .
- the interior camera(s) 28 and/or the user device 42 may be configured to determine whether the user 18 is gazing out the windshield 14 , gazing at the user device 42 , or gazing in any other direction, using MV for facial detection.
- the interior camera(s) 28 may be a camera, an infrared (IR) sensor, and the like.
- the vehicle controller 20 is also configured to be in operative communication with the user device 42 .
- the user device 42 includes a device controller 44 and a display screen 46 .
- the user device 42 is configured to display information, including text and graphics, on the display screen 46 .
- the user device 42 may be a portable personal device, such as a cell phone, a tablet, a computer, and the like. Alternatively, the user device 42 may be integrated into the vehicle 10 , e.g., as an integrated infotainment device.
- the user device 42 is configured to be in operative communication with the vehicle controller 20 .
- the vehicle controller 20 may selectively transmit a signal (arrow 48 ) to the device controller 44 to clear the display screen 46 and/or to replace the displayed image of the display screen 46 with the captured scenery 32 .
- the vehicle 10 may be configured to be in selective communication with an off-board system 49 .
- the vehicle controller 20 may be configured to receive data relating to traffic conditions, traffic signals, weather conditions, and the like, from the off-board system 49 .
- An example of an off-board system 49 may include a service provider, which may be configured as a server located off-board the vehicle 10 , e.g., at a location remote from the vehicle 10 .
- the off-board system 49 may be a vehicle integrated service provider, such as the OnStar® service system, which may be selectively linked to the vehicle interface device and/or in communication with the portable device.
- the service provider may be configured to provide data to the vehicle controller 20 via Wi-Fi, a mobile telecommunications network, satellite, Bluetooth, and the like.
- the off-board system 49 may include environmental sensors 52 disposed in conjunction with landmarks, such as a tree, and the like. The environmental sensor 52 may be configured to transmit a signal (arrow 55 ) to the vehicle controller 20 to alert the vehicle controller 20 as to the proximity of such a landmark to the vehicle 10 .
- the vehicle controller 20 is programmed with, or has access to, the algorithm or method 100 , the execution of which provides a method of improving awareness of a user 18 of the vehicle 10 as to the scene 32 , external to a vehicle 10 , with the algorithm 100 explained in detail below and as shown in FIG. 4 .
- the vehicle controller 20 of FIG. 1 may be embodied as one or more computer devices having a processor (P) 22 and tangible, non-transitory memory (M) 24 on which is recorded instructions for executing the method 100 .
- the memory 24 may include magnetic or optical memory, electrically-erasable programmable read only memory (EEPROM), and the like. Additional transitory memory may be included as needed, e.g., random access memory (RAM), memory for internal signal buffers, etc.
- Other hardware of the vehicle controller 20 may include a high-speed clock, analog-to-digital (A/D) circuitry, digital-to-analog (D/A) circuitry, and any required input/output (I/O) circuitry and devices, as well as signal conditioning and buffer electronics. Individual control algorithms resident in the vehicle controller 20 or readily accessible by the vehicle controller 20 may be stored in memory 24 and/or other suitable memory, and automatically executed via the processor 22 to provide the required control functionality.
- A/D analog-to-digital
- D/A digital-to-
- the exterior camera(s) 26 are operable for imaging the scene 32 exterior to the vehicle 10 , in real-time
- the interior camera(s) 28 are operable for imaging the user's 18 head including their eyes, inside of the vehicle 10 , in real-time.
- Other sensors may also be used to determine or supplement the interior camera's determination of whether the user 18 is gazing at the user device 42 . More specifically, such sensors may sense a physical interaction between the device and the user 18 of the vehicle 10 .
- the collected data (arrow 38 ) of the exterior camera(s) 26 and the collected data (arrow 40 ) of the interior camera(s) 28 are transmitted to the vehicle controller 20 for processing according to the method 100 .
- the exterior camera(s) 26 and/or the interior camera(s) 28 may be three-dimensional (3D) point cloud cameras.
- a 3D point cloud is a set of data points in a 3D coordinate system, such as the X, Y, Z Cartesian coordinate system.
- Such cameras are able to capture any number of data points describing the surface contour of a target object 30 , and to output the collected data (arrow 38 ) as a depth data file with synchronized color data.
- the vehicle controller 20 may be preprogrammed with predetermined target data and/or to receive predetermined target information from the off-board system 49 (arrow 50 ) and/or the environmental sensors 52 (arrow 55 ) which define the known size, shape, color, movements, and/or other descriptive parameters of the particular objects 30 to be located by the external and/or interior cameras 28 .
- the vehicle controller 20 may be made aware of the dimensions, color, and/or other identifying features of the objects 30 it is attempting to identify. While one exterior camera 26 and one interior camera 28 are shown in FIG. 2 for illustrative simplicity, the present method 100 may work with more than one exterior camera 26 and/or more than one interior camera 28 .
- an example embodiment of the method 100 begins with step 102 , where the vehicle controller 20 receives data (arrow 56 ) pertaining to operation of the vehicle 10 .
- the data may be received from a plurality of sensors 54 .
- some of the sensors may be configured to determine revolutions per minute (RPM) of a transmission and/or engine of the vehicle 10 , i.e., via a transmission sensor 54 A and/or an engine sensor 54 B; determine, via a brake pedal sensor 54 C, whether a brake pedal is being actuated; determine, via a steering wheel sensor 54 D, whether the user's 18 hand is touching a steering wheel; determine a wheel speed of the vehicle 10 , i.e., via a wheel speed sensor 54 E; determine whether a clutch switch has been actuated, i.e., via a clutch sensor 54 F; determine, via an accelerometer 54 G, a longitudinal acceleration of the vehicle 10 ; and the like.
- RPM revolutions per minute
- determining whether the vehicle 10 is being operated may include receiving geographic data into the vehicle controller 20 regarding vehicle 10 position, e.g., via a global positioning system (GPS), and the like.
- This geographic data may, in turn, be used as a prompt to determine upcoming traffic conditions in order to suggest a driving route and/or suggest a traffic lane to be used.
- This geographic data may also be used to prevent a preview of an upcoming scene, e.g., a hidden intersection, a hidden driveway, a curve, and other geographic data.
- the method proceeds to step 104 .
- the vehicle controller 20 determines, based on the data, received by the vehicle controller 20 , whether the vehicle 10 is being operated by the user 18 . If the vehicle controller 20 determines the vehicle 10 is not being operated, the method returns to step 102 . However, if the vehicle controller 20 determines the vehicle 10 is being operated, the method proceeds to step 106 .
- the exterior camera(s) 26 captures data pertaining to the scene 32 , external to the vehicle 10 .
- the method then proceeds to step 108 .
- step 108 the captured data, pertaining to the scene 32 exterior to the vehicle 10 , is transmitted to the processor 22 , in real-time. The method then proceeds to step 110 .
- the processor 22 determines whether the scene 32 is visually salient. More specifically, the processor 22 determines whether there are important objects 30 or characteristic within the captured data of the scene 32 . If a determination is made that an object 30 considered to be important is detected in the scene 32 , the method proceeds to step 112 . However, if a determination is made that no visually salient objects 30 or characteristics are detected in the scene 32 , the method returns to step 102 .
- the interior camera(s) 28 captures data, in real-time, data pertaining to a gaze 29 of the user 18 , including but not limited to, an orientation of the head and/or eyes of the user 18 .
- the method then proceeds to step 114 .
- step 114 the captured data is transmitted to the processor 22 in the vehicle controller 20 , in real-time. The method then proceeds to step 116 .
- the processor 22 determines where the eyes of the user 18 are gazing.
- the vehicle controller 20 receives the captured data and processes the captured data using a processor 22 . If the processor 22 determines at step 116 that the user 18 is not gazing at the user device 42 , the method returns to step 118 .
- the vehicle controller 20 may transmit a signal to the user device 42 such that the user device 42 displays content on the display screen 46 , not pertaining to the scenery 32 , i.e., the display screen 46 returns to a nominal condition or to an originally displayed condition. It should be appreciated that, in one embodiment, such a signal may only be transmitted to the user device 42 after certain criteria are satisfied, e.g., the passage of a pre-defined period of time where the user 18 is not gazing at the user device 42 , etc. The method then returns to step 102 .
- step 116 determines at step 116 that the user 18 is gazing at the user device 42 .
- the method proceeds to step 118 .
- the controller may, in turn, transmit a signal to the user device 42 .
- a signal may instruct the controller of the user device 42 to clear the display screen 46 .
- a signal may instruct the controller of the user device 42 to display imagery corresponding to the visually salient scenery 32 .
- the user 18 is alerted to scenery 32 external to the vehicle 10 , requiring the user's 18 attention.
- the signal transmitted by the vehicle controller 20 may not be limited to clearing the display screen 46 and/or displaying scenery 32 external to the vehicle 10 .
- the signal may be also be transmitted to the user device 42 to instruct the user device 42 to output an audible sound and/or cause the user device 42 to vibrate.
- the signal may be transmitted to change displays of the driving displays, e.g., instrument cluster, driver information center (DIC), navigation screen, heads-up display, inside rearview mirror, outside rearview mirror, and the like when a determination is made that the user 18 may be gazing at such driving displays, instead of at the scenery 32 external to the vehicle 10 .
- the driving displays e.g., instrument cluster, driver information center (DIC), navigation screen, heads-up display, inside rearview mirror, outside rearview mirror, and the like when a determination is made that the user 18 may be gazing at such driving displays, instead of at the scenery 32 external to the vehicle 10 .
- DIC driver information center
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Mechanical Engineering (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Transportation (AREA)
- Combustion & Propulsion (AREA)
- Chemical & Material Sciences (AREA)
- Optics & Photonics (AREA)
- Automation & Control Theory (AREA)
- General Health & Medical Sciences (AREA)
- Veterinary Medicine (AREA)
- Biophysics (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Human Computer Interaction (AREA)
- Pathology (AREA)
- Cardiology (AREA)
- Physiology (AREA)
- Computer Networks & Wireless Communication (AREA)
- Ophthalmology & Optometry (AREA)
- Pulmonology (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Traffic Control Systems (AREA)
- Instrument Panels (AREA)
- User Interface Of Digital Computer (AREA)
- Fittings On The Vehicle Exterior For Carrying Loads, And Devices For Holding Or Mounting Articles (AREA)
Abstract
A method of alerting a user of a vehicle as to a scene, external to the vehicle, includes capturing data pertaining to the scene, external to the vehicle. The captured data, pertaining to the scene external to the vehicle, is transmitted to a processor. The determination is made, in the processor that a characteristic is in the scene that requires the user's visual attention. Data pertaining to a gaze of the user is captured. Captured data, pertaining to the gaze of the user, is transmitted to the processor. A determination is made, in the processor that the drive is gazing toward a user device. A signal is transmitted to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.
Description
- This application is a continuation of International Patent Application No. PCT/US2014/035398, filed on Apr. 25, 2014, which claims the benefit of, U.S. Provisional Application No. 61/816,089, filed Apr. 25, 2013, each of which are hereby incorporated by reference in their entirety.
- The present disclosure is related to an adaptive external road scene transfer to improve driver awareness.
- Vehicles, such as cars, typically include displays or indicators to provide information to the vehicle user. Such displays or indicators may, for example, provide information regarding mileage, fuel consumption, and vehicle speed. The vehicle user usually has to shift his eye gaze away from the road scene and onto an in-vehicle display in order to visually process the information presented by these displays or indicators.
- One possible aspect of the disclosure provides a method of alerting a user of a vehicle as to a scene, external to the vehicle, which includes capturing data pertaining to the scene, external to the vehicle. The captured data, pertaining to the scene external to the vehicle, is transmitted to a processor. The determination is made, in the processor that a characteristic is in the scene that requires the user's visual attention. Data pertaining to a gaze of the user is captured. Captured data, pertaining to the gaze of the user, is transmitted to the processor. A determination is made, in the processor that the user is gazing toward a user device. A signal is transmitted to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.
- Another aspect of the disclosure provides a scene awareness system for a vehicle that includes at least one exterior camera, at least one interior camera, and a vehicle controller. The exterior camera is configured to capture data pertaining to a scene, external to the vehicle. The interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle. The vehicle controller is in communication with each of the at least one exterior camera and the at least one interior camera. The vehicle controller is configured to determine whether at least one characteristic in the scene, exterior to the vehicle, requires the user's visual attention. The vehicle controller is also configured to determine whether the user is gazing at a user device. The vehicle controller is configured to transmit a signal to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention when the controller determines there is at least one characteristic in the scene requiring the user's attention, simultaneous with the controller determining the user is gazing at the user device.
- In yet another aspect of the disclosure, a vehicle includes a body and a scene awareness system. The scene awareness system includes at least one exterior camera, at least one interior camera, and a vehicle controller. The cameras are operatively attached to the body. The exterior camera is configured to capture data pertaining to a scene, external to the vehicle. The interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle. The vehicle controller is in communication with each of the cameras. The vehicle controller is operable for receiving data pertaining to the scene, external to the vehicle, from the at least one exterior camera. The vehicle controller includes a process and a characteristic in the scene requiring the user's visual attention that is determined in the processor. Data, pertaining to a gaze of the user, is received from the interior camera. A determination is made that the user is gazing toward a user device. A signal is transmitted to the user device, from the controller that alerts the user as to the characteristic in the scene requiring the user's visual attention.
- The above features and advantages and other features and advantages of the present teachings are readily apparent from the following detailed description of the best modes for carrying out the present teachings when taken in connection with the accompanying drawings.
-
FIG. 1 is a schematic illustrative side view of a vehicle. -
FIG. 2 is a schematic diagrammatic view of a scene awareness system of the vehicle ofFIG. 1 . -
FIG. 3A is schematic fragmentary view of an interior of the vehicle having a scene awareness system, illustrating the user of the vehicle gazing at a scene, external to the vehicle. -
FIG. 3B is a schematic fragmentary view of the interior of the vehicle having the scene awareness system, illustrating the user of the vehicle gazing at a user device. -
FIG. 4 is a schematic flow chart diagram of a method of alerting the user of the vehicle as to a scene, external to the vehicle, requiring the user's attention. - Those having ordinary skill in the art will recognize that terms such as “above,” “below,” “upward,” “downward,” “top,” “bottom,” etc., are used descriptively for the figures, and do not represent limitations on the scope of the invention, as defined by the appended claims. Furthermore, the invention may be described herein in terms of functional and/or logical block components and/or various processing steps. It should be realized that such block components may be comprised of any number of hardware, software, and/or firmware components configured to perform the specified functions.
- Referring now to the drawings, wherein the like numerals indicate corresponding parts throughout the several views,
FIG. 1 schematically illustrates avehicle 10 including abody 12. Thevehicle 10 may be a land vehicle, such as a car, or any other type of vehicle such as an airplane, farm equipment, construction equipment, a boat, etc. Thevehicle 10 may be configured to be operated by auser 18, operated autonomously, and/or operated semi-autonomously. Regardless of the specific kind of vehicle, thevehicle 10 includes awindshield 14 coupled to thebody 12. Thewindshield 14 is wholly or partly made of a substantially transparent material. Referring now toFIGS. 3A and 3B , thevehicle 10 includes aninterior 16. Accordingly, avehicle user 18 can be in position within theinterior 16 and see through thewindshield 14, external to thevehicle 10. Referring again toFIG. 1 , thebody 12 extends between aforward end 21 and arearward end 23. Thewindshield 14 faces theforward end 21 of thevehicle 10, opposite therearward end 23. - With reference to
FIG. 2 , ascene awareness system 25 is shown having one or moreexterior cameras 26, one or moreinternal cameras 28, and a vehicle controller (C) 20. The vehicle controller 20, which is in communication with the internal and the exterior camera(s) 28, 26, is programmed to use the exterior camera(s) 26 to automatically locate and/or identify one or more characteristics in ascene 32, exterior to thevehicle 10, and to use the interior camera(s) 28 to simultaneously locate and identify where the user's 18 eyes are gazing within thevehicle 10 via execution of instructions embodying amethod 100, an example of which is described below with reference toFIG. 4 . The characteristics may include, but should not be limited to, identifying one ormore objects 30, colors, brightness, darkness, temperatures, temperature gradients, graphical patterns, motion, and other like characteristics. - Referring to
FIG. 2 , the exterior camera(s) 26 is configured to monitor ascene 32, external to thevehicle 10. Theexterior cameras 26 may be operatively attached to thevehicle 10 to view thescenery 32 in front of, beside, above, and/or behind thevehicle 10. Alternatively, at least oneexterior camera 26 may be resident within auser device 42 located within theinterior 16 of thevehicle 10 through thewindshield 14 or some other window within thevehicle 10. In such an embodiment, theuser device 42 would be positioned within the vehicle such that theexterior camera 26 is directed to view thescenery 32 around the vehicle. Thescenery 32 may be of a landscape, a billboard, a landmark, and the like. It should be appreciated that the portions of thescenery 32 may be static, in motion, and/or a combination of in static or in motion. The exterior camera(s) 26 may include a camera, a sensor, and the like. The exterior camera(s) 26 is configured to capture images, in static or in motion, external to thevehicle 10, in real-time. The exterior camera(s) 26 is in operative communication with aprocessor 22 configured to determine if thescenery 32 captured in real-time is sufficient to require the user's 18 visual attention, i.e., thescenery 32 is visually salient. Theprocessor 22 may be resident within the vehicle controller 20. The exterior camera(s) 26 may be configured to use machine vision (MV) to recognize characteristics, such asobjects 30, in thescenery 32, in real-time. The captured images may be transmitted as data (arrow 38) from the exterior camera(s) 26 to theprocessor 22 within the controller to make an imaging-based determination of whether thescenery 32 is visually salient. Machine vision may make an imaging-based determination based on edge detection ofobjects 30, color analysis to identifyobjects 30, pattern recognition, feature detection, motion analysis, and the like. Theprocessor 22 may be pre-programmed with criteria. By way of a non-limiting example,scenery 32 sufficient to require the user's 18 visual attention may include an upcoming curve in the road, stop sign, landmark, destination, and the like. As previously mentioned,such scenery 32 may includeobjects 30 which are static and/or objects which are in motion. - The interior camera(s) 28 is also in operative communication with the
processor 22. The interior camera(s) 28 may be used to capture visual information with respect to theuser 18 of thevehicle 10. More specifically, the interior camera(s) 28 may capture information about theusers 18 head, including their eyes, in real-time. The interior camera(s) 28 may be positioned within thevehicle 10, e.g., attached to a rearview mirror, and the like. Alternatively, the interior camera(s) 28 may be positioned within the vehicle, where theinterior camera 28 is resident within theuser device 42. In such an example, theuser device 42 may be configured with both theinterior camera 28 and theexterior camera 26. The captured visual information may be transmitted as data (arrow 40) to theprocessor 22 within the vehicle controller 20. The interior camera(s) 28 is configured to determine where the eyes of theuser 18 are looking and/or to determine an orientation of the user's 18 head, relative to thewindshield 14 and/or theforward end 21 of thevehicle 10. The interior camera(s) 28 and/or theuser device 42 may be configured to determine whether theuser 18 is gazing out thewindshield 14, gazing at theuser device 42, or gazing in any other direction, using MV for facial detection. The interior camera(s) 28 may be a camera, an infrared (IR) sensor, and the like. - The vehicle controller 20 is also configured to be in operative communication with the
user device 42. Theuser device 42 includes adevice controller 44 and adisplay screen 46. Theuser device 42 is configured to display information, including text and graphics, on thedisplay screen 46. Theuser device 42 may be a portable personal device, such as a cell phone, a tablet, a computer, and the like. Alternatively, theuser device 42 may be integrated into thevehicle 10, e.g., as an integrated infotainment device. - As described in more detail below, the
user device 42 is configured to be in operative communication with the vehicle controller 20. As such, the vehicle controller 20 may selectively transmit a signal (arrow 48) to thedevice controller 44 to clear thedisplay screen 46 and/or to replace the displayed image of thedisplay screen 46 with the capturedscenery 32. - Referring again to
FIG. 2 , thevehicle 10 may be configured to be in selective communication with an off-board system 49. More specifically, the vehicle controller 20 may be configured to receive data relating to traffic conditions, traffic signals, weather conditions, and the like, from the off-board system 49. An example of an off-board system 49 may include a service provider, which may be configured as a server located off-board thevehicle 10, e.g., at a location remote from thevehicle 10. The off-board system 49 may be a vehicle integrated service provider, such as the OnStar® service system, which may be selectively linked to the vehicle interface device and/or in communication with the portable device. By way of another non-limiting example, the service provider may be configured to provide data to the vehicle controller 20 via Wi-Fi, a mobile telecommunications network, satellite, Bluetooth, and the like. In yet another non-limiting example, as illustrated inFIG. 2 , the off-board system 49 may includeenvironmental sensors 52 disposed in conjunction with landmarks, such as a tree, and the like. Theenvironmental sensor 52 may be configured to transmit a signal (arrow 55) to the vehicle controller 20 to alert the vehicle controller 20 as to the proximity of such a landmark to thevehicle 10. - The vehicle controller 20 is programmed with, or has access to, the algorithm or
method 100, the execution of which provides a method of improving awareness of auser 18 of thevehicle 10 as to thescene 32, external to avehicle 10, with thealgorithm 100 explained in detail below and as shown inFIG. 4 . - The vehicle controller 20 of
FIG. 1 may be embodied as one or more computer devices having a processor (P) 22 and tangible, non-transitory memory (M) 24 on which is recorded instructions for executing themethod 100. Thememory 24 may include magnetic or optical memory, electrically-erasable programmable read only memory (EEPROM), and the like. Additional transitory memory may be included as needed, e.g., random access memory (RAM), memory for internal signal buffers, etc. Other hardware of the vehicle controller 20 may include a high-speed clock, analog-to-digital (A/D) circuitry, digital-to-analog (D/A) circuitry, and any required input/output (I/O) circuitry and devices, as well as signal conditioning and buffer electronics. Individual control algorithms resident in the vehicle controller 20 or readily accessible by the vehicle controller 20 may be stored inmemory 24 and/or other suitable memory, and automatically executed via theprocessor 22 to provide the required control functionality. - Still referring to
FIG. 2 , when thevehicle 10 is being operated, the exterior camera(s) 26 are operable for imaging thescene 32 exterior to thevehicle 10, in real-time, and the interior camera(s) 28 are operable for imaging the user's 18 head including their eyes, inside of thevehicle 10, in real-time. Other sensors may also be used to determine or supplement the interior camera's determination of whether theuser 18 is gazing at theuser device 42. More specifically, such sensors may sense a physical interaction between the device and theuser 18 of thevehicle 10. The collected data (arrow 38) of the exterior camera(s) 26 and the collected data (arrow 40) of the interior camera(s) 28 are transmitted to the vehicle controller 20 for processing according to themethod 100. The exterior camera(s) 26 and/or the interior camera(s) 28 may be three-dimensional (3D) point cloud cameras. As is known in the art, a 3D point cloud is a set of data points in a 3D coordinate system, such as the X, Y, Z Cartesian coordinate system. Such cameras are able to capture any number of data points describing the surface contour of atarget object 30, and to output the collected data (arrow 38) as a depth data file with synchronized color data. The vehicle controller 20 may be preprogrammed with predetermined target data and/or to receive predetermined target information from the off-board system 49 (arrow 50) and/or the environmental sensors 52 (arrow 55) which define the known size, shape, color, movements, and/or other descriptive parameters of theparticular objects 30 to be located by the external and/orinterior cameras 28. - The vehicle controller 20, shown in
FIG. 2 , may be made aware of the dimensions, color, and/or other identifying features of theobjects 30 it is attempting to identify. While oneexterior camera 26 and oneinterior camera 28 are shown inFIG. 2 for illustrative simplicity, thepresent method 100 may work with more than oneexterior camera 26 and/or more than oneinterior camera 28. - Referring to
FIG. 4 , an example embodiment of themethod 100 begins withstep 102, where the vehicle controller 20 receives data (arrow 56) pertaining to operation of thevehicle 10. The data may be received from a plurality ofsensors 54. By way of a non-limiting example, some of the sensors may be configured to determine revolutions per minute (RPM) of a transmission and/or engine of thevehicle 10, i.e., via atransmission sensor 54A and/or anengine sensor 54B; determine, via abrake pedal sensor 54C, whether a brake pedal is being actuated; determine, via asteering wheel sensor 54D, whether the user's 18 hand is touching a steering wheel; determine a wheel speed of thevehicle 10, i.e., via a wheel speed sensor 54E; determine whether a clutch switch has been actuated, i.e., via aclutch sensor 54F; determine, via anaccelerometer 54G, a longitudinal acceleration of thevehicle 10; and the like. - Additionally, determining whether the
vehicle 10 is being operated may include receiving geographic data into the vehicle controller 20 regardingvehicle 10 position, e.g., via a global positioning system (GPS), and the like. This geographic data may, in turn, be used as a prompt to determine upcoming traffic conditions in order to suggest a driving route and/or suggest a traffic lane to be used. This geographic data may also be used to prevent a preview of an upcoming scene, e.g., a hidden intersection, a hidden driveway, a curve, and other geographic data. Next, the method proceeds to step 104. - At
step 104, the vehicle controller 20 determines, based on the data, received by the vehicle controller 20, whether thevehicle 10 is being operated by theuser 18. If the vehicle controller 20 determines thevehicle 10 is not being operated, the method returns to step 102. However, if the vehicle controller 20 determines thevehicle 10 is being operated, the method proceeds to step 106. - At
step 106, the exterior camera(s) 26 captures data pertaining to thescene 32, external to thevehicle 10. The method then proceeds to step 108. - At
step 108, the captured data, pertaining to thescene 32 exterior to thevehicle 10, is transmitted to theprocessor 22, in real-time. The method then proceeds to step 110. - At
step 110, theprocessor 22 determines whether thescene 32 is visually salient. More specifically, theprocessor 22 determines whether there areimportant objects 30 or characteristic within the captured data of thescene 32. If a determination is made that anobject 30 considered to be important is detected in thescene 32, the method proceeds to step 112. However, if a determination is made that no visuallysalient objects 30 or characteristics are detected in thescene 32, the method returns to step 102. - At
step 112, with theuser 18 seated within theinterior 16 of thevehicle 10, the interior camera(s) 28 captures data, in real-time, data pertaining to agaze 29 of theuser 18, including but not limited to, an orientation of the head and/or eyes of theuser 18. The method then proceeds to step 114. - At
step 114, the captured data is transmitted to theprocessor 22 in the vehicle controller 20, in real-time. The method then proceeds to step 116. - At
step 116, theprocessor 22 determines where the eyes of theuser 18 are gazing. The vehicle controller 20 receives the captured data and processes the captured data using aprocessor 22. If theprocessor 22 determines atstep 116 that theuser 18 is not gazing at theuser device 42, the method returns to step 118. - At
step 118, the vehicle controller 20 may transmit a signal to theuser device 42 such that theuser device 42 displays content on thedisplay screen 46, not pertaining to thescenery 32, i.e., thedisplay screen 46 returns to a nominal condition or to an originally displayed condition. It should be appreciated that, in one embodiment, such a signal may only be transmitted to theuser device 42 after certain criteria are satisfied, e.g., the passage of a pre-defined period of time where theuser 18 is not gazing at theuser device 42, etc. The method then returns to step 102. - However, if the
processor 22 determines atstep 116 that theuser 18 is gazing at theuser device 42, the method proceeds to step 118. - At
step 118, the controller may, in turn, transmit a signal to theuser device 42. Such a signal may instruct the controller of theuser device 42 to clear thedisplay screen 46. Additionally, such a signal may instruct the controller of theuser device 42 to display imagery corresponding to the visuallysalient scenery 32. By changing thedisplay screen 46 of theuser device 42, theuser 18 is alerted toscenery 32 external to thevehicle 10, requiring the user's 18 attention. - It should be appreciated that the signal transmitted by the vehicle controller 20 may not be limited to clearing the
display screen 46 and/or displayingscenery 32 external to thevehicle 10. By way of a non-limiting example, the signal may be also be transmitted to theuser device 42 to instruct theuser device 42 to output an audible sound and/or cause theuser device 42 to vibrate. In another non-limiting example, the signal may be transmitted to change displays of the driving displays, e.g., instrument cluster, driver information center (DIC), navigation screen, heads-up display, inside rearview mirror, outside rearview mirror, and the like when a determination is made that theuser 18 may be gazing at such driving displays, instead of at thescenery 32 external to thevehicle 10. - While the best modes for carrying out the many aspects of the present teachings have been described in detail, those familiar with the art to which these teachings relate will recognize various alternative aspects for practicing the present teachings that are within the scope of the appended claims.
Claims (18)
1. A method of alerting a user of a vehicle as to a scene, external to the vehicle, the method comprising:
capturing data pertaining to the scene, external to the vehicle;
transmitting the captured data pertaining to the scene, external to the vehicle, to a processor;
determining, in the processor, a characteristic of the scene requiring the user's visual attention;
capturing data pertaining to a gaze of the user;
transmitting the captured data pertaining to the gaze of the user to the processor;
determining, in the processor, the driver is gazing toward a user device; and
transmitting a signal to the user device such that the user device alerts the user as to the characteristic of the scene requiring the user's visual attention.
2. A method, as set forth in claim 1 , wherein transmitting is further defined as transmitting a signal to the user device such that the user device alerts the user as to the characteristic of the scene requiring the user's visual attention by clearing the screen of the user device.
3. A method, as set forth in claim 2 , wherein transmitting is further defined as transmitting a signal to the user device such that the user device alerts the user as to the characteristic of the scene requiring the user's visual attention by displaying a visual depiction of the characteristic of the scene requiring the user's visual attention on the display screen of the user device.
4. A method, as set forth in claim 1 , further comprising:
receiving data pertaining to operation of the vehicle; and
determining the vehicle is being operated by the user.
5. A method, as set forth in claim 4 , wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user when a determination is made that the vehicle is being operated by the user.
6. A method, as set forth in claim 4 , wherein receiving data is further defined as receiving geographic data regarding a vehicle position.
7. A method, as set forth in claim 1 , wherein capturing data pertaining to the scene, external to the vehicle, is further defined as capturing data pertaining to the scene, external to the vehicle, with at least one external camera.
8. A method, as set forth in claim 1 , wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user when a determination is made that a characteristic of the scene requiring the user's attention is detected in the scene.
9. A method, as set forth in claim 1 , wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user with at least one internal camera.
10. A method, as set forth in claim 9 , wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to an orientation of at least one of the head and the eyes of the user.
11. A method, as set forth in claim 1 , wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user after determining, in the processor, a characteristic of the scene requiring the user's visual attention.
12. A scene awareness system for a vehicle, the scene awareness system comprising:
at least one exterior camera configured to capture data pertaining to a scene, external to the vehicle;
at least one interior camera configured to capture data pertaining to an orientation of a gaze of a user of the vehicle;
a vehicle controller in communication with each of the at least one exterior camera and the at least one interior camera;
wherein the vehicle controller is configured to determine whether at least one characteristic in the scene, exterior to the vehicle, requires the user's visual attention;
wherein the vehicle controller is configured to determine whether the user is gazing at a user device; and
wherein the vehicle controller is configured to transmit a signal to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention when the controller determines there is at least one characteristic in the scene requiring the user's attention simultaneous with the controller determining the user is gazing at the user device.
13. A scene awareness system, as set forth in claim 12 , further comprising at least one sensor configured for transmitting data to the vehicle controller;
wherein the data pertains to operation of the vehicle.
14. A vehicle comprising:
a body;
a scene awareness system including:
at least one exterior camera operatively attached to the body;
wherein the at least one exterior camera is configured to capture data pertaining to a scene, external to the vehicle;
at least one interior camera operatively attached to the body;
wherein the at least one interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle;
a vehicle controller in communication with each of the at least one exterior camera and the at least one interior camera, the vehicle controller operable for:
receiving data pertaining to the scene, external to the vehicle, from the at least one exterior camera;
determining, in the processor, a characteristic in the scene requiring the user's visual attention;
receiving data pertaining to a gaze of the user from the at least one interior camera;
determining the driver is gazing toward a user device; and
transmitting a signal to a user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.
15. A vehicle, as set forth in claim 14 , wherein the body extends between a forward end and a rearward end; and
wherein the at least one exterior camera is operatively attached to the body, proximate the forward end.
16. A vehicle, as set forth in claim 15 , wherein the body defines an interior configured for receiving the user of the vehicle therein;
wherein the at least one interior camera is operatively disposed in the interior of the body.
17. A vehicle, as set forth in claim 16 , further comprising at least one sensor configured to be in operative communication with the vehicle controller;
wherein the vehicle controller is further operable for receiving data, pertaining to operation of the vehicle, from the at least one sensor.
18. A vehicle, as set forth in claim 14 , wherein the vehicle controller is further operable for:
receiving data pertaining to operation of the vehicle; and
determining the vehicle is being operated by the user.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/920,420 US20160039285A1 (en) | 2013-04-25 | 2015-10-22 | Scene awareness system for a vehicle |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361816089P | 2013-04-25 | 2013-04-25 | |
PCT/US2014/035398 WO2014176478A1 (en) | 2013-04-25 | 2014-04-25 | Scene awareness system for a vehicle |
US14/920,420 US20160039285A1 (en) | 2013-04-25 | 2015-10-22 | Scene awareness system for a vehicle |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2014/035398 Continuation WO2014176478A1 (en) | 2013-04-25 | 2014-04-25 | Scene awareness system for a vehicle |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160039285A1 true US20160039285A1 (en) | 2016-02-11 |
Family
ID=51792397
Family Applications (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/918,073 Active 2034-06-28 US9688287B2 (en) | 2013-04-25 | 2015-10-20 | Situation awareness system and method |
US14/920,420 Abandoned US20160039285A1 (en) | 2013-04-25 | 2015-10-22 | Scene awareness system for a vehicle |
US14/920,413 Abandoned US20160041562A1 (en) | 2013-04-25 | 2015-10-22 | Method of controlling a component of a vehicle with a user device |
US14/920,499 Active 2035-06-10 US10131364B2 (en) | 2013-04-25 | 2015-10-22 | Ambient display |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/918,073 Active 2034-06-28 US9688287B2 (en) | 2013-04-25 | 2015-10-20 | Situation awareness system and method |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/920,413 Abandoned US20160041562A1 (en) | 2013-04-25 | 2015-10-22 | Method of controlling a component of a vehicle with a user device |
US14/920,499 Active 2035-06-10 US10131364B2 (en) | 2013-04-25 | 2015-10-22 | Ambient display |
Country Status (4)
Country | Link |
---|---|
US (4) | US9688287B2 (en) |
CN (1) | CN105324268A (en) |
DE (1) | DE112014001607B4 (en) |
WO (4) | WO2014176474A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170054837A1 (en) * | 2014-05-09 | 2017-02-23 | Samsung Electronics Co., Ltd. | Terminal and method for displaying caller information |
US20170294065A1 (en) * | 2014-04-18 | 2017-10-12 | Gentex Corporation | Trainable transceiver and cloud computing system architecture systems and methods |
US10166996B2 (en) * | 2017-02-09 | 2019-01-01 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods for adaptively communicating notices in a vehicle |
US11312300B1 (en) | 2021-01-29 | 2022-04-26 | Toyota Motor Engineering & Manufacturing North America, Inc. | Object notification systems for identifying and notifying vehicles of relevant objects |
Families Citing this family (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150321604A1 (en) * | 2014-05-07 | 2015-11-12 | Ford Global Technologies, Llc | In-vehicle micro-interactions |
DE102015212676A1 (en) * | 2015-07-07 | 2017-01-12 | Bayerische Motoren Werke Aktiengesellschaft | Determining the driving ability of the driver of a first motor vehicle |
CN105631977A (en) * | 2016-02-18 | 2016-06-01 | 广东百事泰电子商务股份有限公司 | Intelligent monitoring and recording instrument |
TW201836890A (en) * | 2017-03-31 | 2018-10-16 | 育全 李 | Method of showing the inside status of a vehicle via a plurality of first icons |
JP6325154B1 (en) * | 2017-06-07 | 2018-05-16 | スマート ビート プロフィッツ リミテッド | Information processing system |
US11267340B2 (en) * | 2017-10-10 | 2022-03-08 | Sekisui Chemical Co., Ltd. | Passenger vehicle window glass and warning display method |
US10709386B2 (en) | 2017-12-12 | 2020-07-14 | Lear Corporation | Electrocardiogram waveform identification and diagnostics via electrophysiological sensor system fusion |
CN118386996A (en) * | 2018-01-04 | 2024-07-26 | 哈曼国际工业有限公司 | Contextual sunroof for enhanced media experience in a car |
JPWO2020071169A1 (en) * | 2018-10-01 | 2021-09-02 | 富士フイルム株式会社 | display |
US11151810B2 (en) * | 2018-10-12 | 2021-10-19 | Aurora Flight Sciences Corporation | Adaptable vehicle monitoring system |
US20200376937A1 (en) * | 2019-05-29 | 2020-12-03 | Toyota Boshoku Kabushiki Kaisha | Light-adjusting system and vehicle light-adjusting system |
CN112918381B (en) * | 2019-12-06 | 2023-10-27 | 广州汽车集团股份有限公司 | Vehicle-mounted robot welcome method, device and system |
US11467401B2 (en) * | 2021-03-02 | 2022-10-11 | GM Global Technology Operations LLC | Display and light blocking screens |
US11506892B1 (en) | 2021-05-03 | 2022-11-22 | GM Global Technology Operations LLC | Holographic display system for a motor vehicle |
US11762195B2 (en) | 2021-05-06 | 2023-09-19 | GM Global Technology Operations LLC | Holographic display system with conjugate image removal for a motor vehicle |
US20240054528A1 (en) * | 2022-08-10 | 2024-02-15 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods for measuring a reaction of a user to an advertisement |
GB2624974A (en) * | 2022-11-29 | 2024-06-05 | E Lead Electronic Co Ltd | Shield for helmet, helmet, and head-up display device |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6320497B1 (en) * | 1999-07-27 | 2001-11-20 | Kubota Corporation | Display control system for a working vehicle |
US20040017300A1 (en) * | 2002-07-25 | 2004-01-29 | Kotzin Michael D. | Portable communication device and corresponding method of operation |
US20040150514A1 (en) * | 2003-02-05 | 2004-08-05 | Newman Timothy J. | Vehicle situation alert system with eye gaze controlled alert signal generation |
US20060071766A1 (en) * | 2000-03-02 | 2006-04-06 | O'brien Frank | Tire pressure alert system |
US7209833B2 (en) * | 2004-01-19 | 2007-04-24 | Denso Corporation | Collision possibility determination device |
US7400233B2 (en) * | 2005-05-30 | 2008-07-15 | Honda Motor Co., Ltd. | Travel safety apparatus for vehicle |
US7643737B2 (en) * | 2006-03-27 | 2010-01-05 | Honda Motor Co., Ltd. | Line of sight detection apparatus |
US20100079270A1 (en) * | 2008-10-01 | 2010-04-01 | Wolfgang Krautter | Method for displaying a visual warning signal |
US8085140B2 (en) * | 2006-09-04 | 2011-12-27 | Panasonic Corporation | Travel information providing device |
US20120050068A1 (en) * | 1997-08-25 | 2012-03-01 | Donnelly Corporation | Interior rearview mirror system |
US8144002B2 (en) * | 2008-03-31 | 2012-03-27 | Hyundai Motor Company | Alarm system for alerting driver to presence of objects |
US8599027B2 (en) * | 2010-10-19 | 2013-12-03 | Deere & Company | Apparatus and method for alerting machine operator responsive to the gaze zone |
US20160023604A1 (en) * | 2013-07-08 | 2016-01-28 | LightSpeed Automotive Technology | Head-Up Display Controller |
Family Cites Families (55)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5883605A (en) * | 1992-02-25 | 1999-03-16 | Gentex Corporation | Automatic electrochromic control of light level of vacuum fluorescent display |
DE4211728A1 (en) | 1992-04-08 | 1993-10-14 | Zeiss Carl Fa | Holographic display device e.g. for vehicle or aircraft head=up display - uses curved windscreen incorporating monomode waveguide for supplied light and holographic gratings |
JPH08272321A (en) * | 1995-03-31 | 1996-10-18 | Toyoda Gosei Co Ltd | External display device of vehicle |
US7346374B2 (en) * | 1999-05-26 | 2008-03-18 | Johnson Controls Technology Company | Wireless communications system and method |
US7449081B2 (en) * | 2000-06-21 | 2008-11-11 | E. I. Du Pont De Nemours And Company | Process for improving the emission of electron field emitters |
US6580973B2 (en) * | 2000-10-14 | 2003-06-17 | Robert H. Leivian | Method of response synthesis in a driver assistance system |
US7565230B2 (en) * | 2000-10-14 | 2009-07-21 | Temic Automotive Of North America, Inc. | Method and apparatus for improving vehicle operator performance |
JP3846872B2 (en) * | 2002-06-27 | 2006-11-15 | パイオニア株式会社 | Driver mental state information provision system |
DE102004004302A1 (en) * | 2003-02-03 | 2004-08-12 | Denso Corp., Kariya | Vehicle remote control air conditioning system has a control unit that activates only a ventilation or climate control component of the air conditioning system in order to reduce battery power consumption |
US20050084659A1 (en) * | 2003-10-20 | 2005-04-21 | General Atomics | Vehicle windshield head-up display |
US7801283B2 (en) * | 2003-12-22 | 2010-09-21 | Lear Corporation | Method of operating vehicular, hands-free telephone system |
DE102004005816B4 (en) | 2004-02-06 | 2007-02-08 | Audi Ag | motor vehicle |
US7413328B2 (en) * | 2004-12-30 | 2008-08-19 | Honeywell International Inc. | Remotely coupled hybrid HUD backlight |
DE102005059216A1 (en) | 2005-07-16 | 2007-01-25 | Ralf Michel | Supervision system in particular for motorcycle, comprises units for determination and evaluation of driving performance |
JP4617226B2 (en) * | 2005-08-30 | 2011-01-19 | 本田技研工業株式会社 | Vehicle display device |
KR100828965B1 (en) * | 2006-07-31 | 2008-05-13 | 삼성전자주식회사 | Method and apparatus for setting environment of cars in portable terminal |
WO2008061098A1 (en) * | 2006-11-14 | 2008-05-22 | Johnson Controls Technology Company | System and method of synchronizing an in-vehicle control system with a remote source |
US20080158510A1 (en) * | 2007-01-02 | 2008-07-03 | Gm Global Technology Operations, Inc. | Apparatus And Method For Displaying Information Within A Vehicle Interior |
WO2008086216A2 (en) * | 2007-01-05 | 2008-07-17 | Johnson Controls Technology Company | System and method for customized audio prompting |
KR101117912B1 (en) * | 2007-05-17 | 2012-03-13 | 프리즘, 인코포레이티드 | Multilayered screens with light-emitting stripes for scanning beam display systems |
US7908060B2 (en) * | 2007-07-31 | 2011-03-15 | International Business Machines Corporation | Method and system for blind spot identification and warning utilizing portable and wearable devices |
JP2009156898A (en) * | 2007-12-25 | 2009-07-16 | Seiko Epson Corp | Display device |
US8311734B2 (en) * | 2008-07-01 | 2012-11-13 | Sony Corporation | Automatic speed limit adjust for road conditions |
DE102009010623A1 (en) | 2009-02-26 | 2010-09-02 | Hella Kgaa Hueck & Co. | Device for issuing visual warning information to driver of vehicle, particularly motor vehicle, has vehicle window pane, through which vehicle driver visually captures area of environment of vehicle |
US8564502B2 (en) * | 2009-04-02 | 2013-10-22 | GM Global Technology Operations LLC | Distortion and perspective correction of vector projection display |
FR2946336B1 (en) * | 2009-06-03 | 2011-05-20 | Saint Gobain | LAMINATED GLAZING FOR HIGH HEAD VISUALIZATION SYSTEM |
US20110012720A1 (en) * | 2009-07-15 | 2011-01-20 | Hirschfeld Robert A | Integration of Vehicle On-Board Diagnostics and Smart Phone Sensors |
US20110025584A1 (en) * | 2009-07-29 | 2011-02-03 | Gm Global Technology Operations, Inc. | Light-emitting diode heads-up display for a vehicle |
KR20110038563A (en) * | 2009-10-08 | 2011-04-14 | 최운호 | Method, vehicle terminal, biometrics card and system for controlling vehicle through authenticating driver, and method for providing passenger protecting/tracking function using biometrics card and terminal |
US8498757B2 (en) * | 2009-10-09 | 2013-07-30 | Visteon Global Technologies, Inc. | Portable and personal vehicle presets |
CN201525262U (en) * | 2009-11-25 | 2010-07-14 | 王辉 | Automobile front windshield glass with transparent LCD (liquid crystal display) device |
US8384534B2 (en) * | 2010-01-14 | 2013-02-26 | Toyota Motor Engineering & Manufacturing North America, Inc. | Combining driver and environment sensing for vehicular safety systems |
KR101177660B1 (en) * | 2010-01-20 | 2012-08-27 | 최인주 | Vehicle management system through radio relay of vehicle remote controller |
KR20120075672A (en) * | 2010-12-29 | 2012-07-09 | 성균관대학교산학협력단 | System and method for safe driving induction using detection of gaze direction |
DE102011009840A1 (en) * | 2011-01-31 | 2012-08-02 | Continental Automotive Gmbh | operating device |
WO2012135018A2 (en) * | 2011-03-25 | 2012-10-04 | Tk Holdings Inc. | System and method for determining driver alertness |
KR20120113579A (en) * | 2011-04-05 | 2012-10-15 | 현대자동차주식회사 | Apparatus and method for displaying road guide information on the windshield |
US9348492B1 (en) * | 2011-04-22 | 2016-05-24 | Angel A. Penilla | Methods and systems for providing access to specific vehicle controls, functions, environment and applications to guests/passengers via personal mobile devices |
US20120303178A1 (en) * | 2011-05-26 | 2012-11-29 | Hendry Jeffrey C | Method and system for establishing user settings of vehicle components |
KR20120136721A (en) * | 2011-06-09 | 2012-12-20 | 현대자동차주식회사 | Apparatus and method for alarming use of mobile phone on driving |
KR101337014B1 (en) * | 2011-07-12 | 2013-12-05 | 주식회사 팬택 | Mobile terminal, system and method for controlling ecu of vehicle using the same |
EP2564776B1 (en) * | 2011-09-02 | 2019-08-28 | Volvo Car Corporation | Method, system and computer readable medium embodying a computer program product for determining a vehicle operator's expectation of a state of an object |
CN202357886U (en) * | 2011-12-09 | 2012-08-01 | 常州永旭车辆配件厂 | Dashboard of electric vehicle |
FR2985042B1 (en) * | 2011-12-22 | 2014-01-17 | Saint Gobain | DEVICE FOR VISUALIZING AN IMAGE ON A SHEET SUPPORT |
US8989961B2 (en) * | 2012-04-13 | 2015-03-24 | Htc Corporation | Method of controlling interaction between mobile electronic device and in-vehicle electronic system and devices using the same |
CN202806308U (en) * | 2012-08-21 | 2013-03-20 | 惠州市德赛西威汽车电子有限公司 | Automotive windshield |
US9426274B2 (en) * | 2012-09-27 | 2016-08-23 | Intel Corporation | Device, method, and system for portable configuration of vehicle controls |
JP2014069592A (en) * | 2012-09-27 | 2014-04-21 | Mitsubishi Motors Corp | Remote control system for on-vehicle equipment |
DE102012219924A1 (en) * | 2012-10-31 | 2014-04-30 | Bayerische Motoren Werke Aktiengesellschaft | Vehicle assistance device |
US20140142948A1 (en) * | 2012-11-21 | 2014-05-22 | Somya Rathi | Systems and methods for in-vehicle context formation |
US20140163771A1 (en) * | 2012-12-10 | 2014-06-12 | Ford Global Technologies, Llc | Occupant interaction with vehicle system using brought-in devices |
US8751065B1 (en) * | 2012-12-14 | 2014-06-10 | Denso Corporation | Smartphone controller of vehicle settings |
KR101555444B1 (en) * | 2014-07-10 | 2015-10-06 | 현대모비스 주식회사 | An apparatus mounted in vehicle for situational awareness and a method thereof |
CN105313898B (en) * | 2014-07-23 | 2018-03-20 | 现代摩比斯株式会社 | Driver status induction installation and its method |
US20160109701A1 (en) * | 2014-10-15 | 2016-04-21 | GM Global Technology Operations LLC | Systems and methods for adjusting features within a head-up display |
-
2014
- 2014-04-25 WO PCT/US2014/035387 patent/WO2014176474A1/en active Application Filing
- 2014-04-25 WO PCT/US2014/035396 patent/WO2014176476A1/en active Application Filing
- 2014-04-25 WO PCT/US2014/035398 patent/WO2014176478A1/en active Application Filing
- 2014-04-25 WO PCT/US2014/035385 patent/WO2014176473A1/en active Application Filing
- 2014-04-25 CN CN201480035937.0A patent/CN105324268A/en active Pending
- 2014-04-25 DE DE112014001607.1T patent/DE112014001607B4/en active Active
-
2015
- 2015-10-20 US US14/918,073 patent/US9688287B2/en active Active
- 2015-10-22 US US14/920,420 patent/US20160039285A1/en not_active Abandoned
- 2015-10-22 US US14/920,413 patent/US20160041562A1/en not_active Abandoned
- 2015-10-22 US US14/920,499 patent/US10131364B2/en active Active
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8779910B2 (en) * | 1997-08-25 | 2014-07-15 | Donnelly Corporation | Interior rearview mirror system |
US20120050068A1 (en) * | 1997-08-25 | 2012-03-01 | Donnelly Corporation | Interior rearview mirror system |
US6320497B1 (en) * | 1999-07-27 | 2001-11-20 | Kubota Corporation | Display control system for a working vehicle |
US20060071766A1 (en) * | 2000-03-02 | 2006-04-06 | O'brien Frank | Tire pressure alert system |
US7423522B2 (en) * | 2000-03-02 | 2008-09-09 | Donnelly Corporation | Tire pressure alert system |
US20040017300A1 (en) * | 2002-07-25 | 2004-01-29 | Kotzin Michael D. | Portable communication device and corresponding method of operation |
US20040150514A1 (en) * | 2003-02-05 | 2004-08-05 | Newman Timothy J. | Vehicle situation alert system with eye gaze controlled alert signal generation |
US7209833B2 (en) * | 2004-01-19 | 2007-04-24 | Denso Corporation | Collision possibility determination device |
US7400233B2 (en) * | 2005-05-30 | 2008-07-15 | Honda Motor Co., Ltd. | Travel safety apparatus for vehicle |
US7643737B2 (en) * | 2006-03-27 | 2010-01-05 | Honda Motor Co., Ltd. | Line of sight detection apparatus |
US8085140B2 (en) * | 2006-09-04 | 2011-12-27 | Panasonic Corporation | Travel information providing device |
US8144002B2 (en) * | 2008-03-31 | 2012-03-27 | Hyundai Motor Company | Alarm system for alerting driver to presence of objects |
US20100079270A1 (en) * | 2008-10-01 | 2010-04-01 | Wolfgang Krautter | Method for displaying a visual warning signal |
US8390440B2 (en) * | 2008-10-01 | 2013-03-05 | Robert Bosch Gmbh | Method for displaying a visual warning signal |
US8599027B2 (en) * | 2010-10-19 | 2013-12-03 | Deere & Company | Apparatus and method for alerting machine operator responsive to the gaze zone |
US20160023604A1 (en) * | 2013-07-08 | 2016-01-28 | LightSpeed Automotive Technology | Head-Up Display Controller |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170294065A1 (en) * | 2014-04-18 | 2017-10-12 | Gentex Corporation | Trainable transceiver and cloud computing system architecture systems and methods |
US10096186B2 (en) * | 2014-04-18 | 2018-10-09 | Gentex Corporation | Trainable transceiver and cloud computing system architecture systems and methods |
US10127804B2 (en) | 2014-04-18 | 2018-11-13 | Gentex Corporation | Trainable transceiver and cloud computing system architecture systems and methods |
US20170054837A1 (en) * | 2014-05-09 | 2017-02-23 | Samsung Electronics Co., Ltd. | Terminal and method for displaying caller information |
US10166996B2 (en) * | 2017-02-09 | 2019-01-01 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods for adaptively communicating notices in a vehicle |
US11312300B1 (en) | 2021-01-29 | 2022-04-26 | Toyota Motor Engineering & Manufacturing North America, Inc. | Object notification systems for identifying and notifying vehicles of relevant objects |
Also Published As
Publication number | Publication date |
---|---|
WO2014176476A1 (en) | 2014-10-30 |
CN105324268A (en) | 2016-02-10 |
US20160041562A1 (en) | 2016-02-11 |
DE112014001607T5 (en) | 2015-12-24 |
US9688287B2 (en) | 2017-06-27 |
DE112014001607B4 (en) | 2021-09-02 |
US20160082979A1 (en) | 2016-03-24 |
US10131364B2 (en) | 2018-11-20 |
WO2014176474A1 (en) | 2014-10-30 |
US20160085070A1 (en) | 2016-03-24 |
WO2014176478A1 (en) | 2014-10-30 |
WO2014176473A1 (en) | 2014-10-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20160039285A1 (en) | Scene awareness system for a vehicle | |
CN110678371B (en) | Vehicle control system, vehicle control method, and storage medium | |
CN106218506B (en) | Vehicle display device and vehicle including the vehicle display device | |
US9507345B2 (en) | Vehicle control system and method | |
JP5811804B2 (en) | Vehicle periphery monitoring device | |
CN108621943B (en) | System and method for dynamically displaying images on a vehicle electronic display | |
US20160185293A1 (en) | Vehicle vision system | |
EP2857886A2 (en) | Display control apparatus, computer-implemented method, storage medium, and projection apparatus | |
KR102531888B1 (en) | How to operate a display device in a car | |
CN106394553A (en) | Driver assistance apparatus and control method for the same | |
US20070124071A1 (en) | System for providing 3-dimensional vehicle information with predetermined viewpoint, and method thereof | |
CN106564432A (en) | Apparatus and method for controlling viewing angle for vehicle, and vehicle including the apparatus | |
US10488658B2 (en) | Dynamic information system capable of providing reference information according to driving scenarios in real time | |
CN110462702B (en) | Travel route providing system, control method thereof, and medium | |
CN109415018A (en) | Method and control unit for digital rearview mirror | |
JP6186905B2 (en) | In-vehicle display device and program | |
WO2020105685A1 (en) | Display control device, method, and computer program | |
JP6589775B2 (en) | Vehicle display control device and vehicle display system | |
US11828947B2 (en) | Vehicle and control method thereof | |
CN114103982A (en) | Display control device for vehicle, display control method for vehicle, computer-readable storage medium, and display system for vehicle | |
JP2008040667A (en) | Image processor and program | |
JP2024124855A (en) | Vehicle control device | |
CN116868251A (en) | Display control device and display control program | |
WO2023213416A1 (en) | Method and user device for detecting an environment of the user device | |
CN118613382A (en) | Display control device, display system, and display control method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GM GLOBAL TECHNOLOGY OPERATIONS LLC, MICHIGAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MATHIEU, ROY J.;SZCZERBA, JOSEPH F.;JONES, MICAH R.;AND OTHERS;SIGNING DATES FROM 20151021 TO 20151022;REEL/FRAME:036882/0669 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |