US20210347298A1 - Dynamic lighting and sensor adjustment for occupant monitoring - Google Patents

Dynamic lighting and sensor adjustment for occupant monitoring Download PDF

Info

Publication number
US20210347298A1
US20210347298A1 US16/869,293 US202016869293A US2021347298A1 US 20210347298 A1 US20210347298 A1 US 20210347298A1 US 202016869293 A US202016869293 A US 202016869293A US 2021347298 A1 US2021347298 A1 US 2021347298A1
Authority
US
United States
Prior art keywords
environmental
occupant
sensors
prospective
vehicle
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/869,293
Other versions
US11584295B2 (en
Inventor
Thomas Balch
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Toyota Motor Corp
Original Assignee
Toyota Research Institute Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Toyota Research Institute Inc filed Critical Toyota Research Institute Inc
Priority to US16/869,293 priority Critical patent/US11584295B2/en
Assigned to Toyota Research Institute, Inc. reassignment Toyota Research Institute, Inc. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BALACH, THOMAS
Publication of US20210347298A1 publication Critical patent/US20210347298A1/en
Application granted granted Critical
Publication of US11584295B2 publication Critical patent/US11584295B2/en
Assigned to TOYOTA JIDOSHA KABUSHIKI KAISHA reassignment TOYOTA JIDOSHA KABUSHIKI KAISHA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Toyota Research Institute, Inc.
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q3/00Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors
    • B60Q3/80Circuits; Control arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q3/00Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors
    • B60Q3/10Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors for dashboards
    • B60Q3/16Circuits; Control arrangements
    • B60Q3/18Circuits; Control arrangements for varying the light intensity
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q3/00Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors
    • B60Q3/60Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors characterised by optical aspects
    • B60Q3/62Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors characterised by optical aspects using light guides
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q3/00Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors
    • B60Q3/70Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors characterised by the purpose
    • B60Q3/72Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors characterised by the purpose for preventing the driver from being dazzled
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q3/00Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors
    • B60Q3/70Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors characterised by the purpose
    • B60Q3/74Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors characterised by the purpose for overall compartment lighting; for overall compartment lighting in combination with specific lighting, e.g. room lamps with reading lamps
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q9/00Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling
    • B60Q9/002Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling for parking purposes, e.g. for warning the driver that his vehicle has contacted or is about to contact an obstacle
    • B60Q9/004Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling for parking purposes, e.g. for warning the driver that his vehicle has contacted or is about to contact an obstacle using wave sensors
    • B60Q9/005Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling for parking purposes, e.g. for warning the driver that his vehicle has contacted or is about to contact an obstacle using wave sensors using a video camera
    • FMECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
    • F21LIGHTING
    • F21VFUNCTIONAL FEATURES OR DETAILS OF LIGHTING DEVICES OR SYSTEMS THEREOF; STRUCTURAL COMBINATIONS OF LIGHTING DEVICES WITH OTHER ARTICLES, NOT OTHERWISE PROVIDED FOR
    • F21V23/00Arrangement of electric circuit elements in or on lighting devices
    • F21V23/04Arrangement of electric circuit elements in or on lighting devices the elements being switches
    • F21V23/0442Arrangement of electric circuit elements in or on lighting devices the elements being switches activated by means of a sensor, e.g. motion or photodetectors
    • F21V23/0464Arrangement of electric circuit elements in or on lighting devices the elements being switches activated by means of a sensor, e.g. motion or photodetectors the sensor sensing the level of ambient illumination, e.g. dawn or dusk sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/141Control of illumination
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • HELECTRICITY
    • H05ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
    • H05BELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
    • H05B47/00Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
    • H05B47/10Controlling the light source
    • H05B47/105Controlling the light source in response to determined parameters
    • H05B47/115Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings
    • H05B47/125Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings by using cameras
    • HELECTRICITY
    • H05ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
    • H05BELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
    • H05B47/00Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
    • H05B47/10Controlling the light source
    • H05B47/105Controlling the light source in response to determined parameters
    • H05B47/115Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings
    • H05B47/13Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings by using passive infrared detectors
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q2300/00Indexing codes for automatically adjustable headlamps or automatically dimmable headlamps
    • B60Q2300/30Indexing codes relating to the vehicle environment
    • B60Q2300/31Atmospheric conditions
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02BCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
    • Y02B20/00Energy efficient lighting technologies, e.g. halogen lamps or gas discharge lamps
    • Y02B20/40Control techniques providing energy savings, e.g. smart controller or presence detection

Definitions

  • the present disclosure relates to a system and method for monitoring the occupant of a vehicle, more particularly to an occupant monitoring system to proactively adjust monitoring sensors as ambient light changes.
  • Autonomous and semi-autonomous vehicles rely on occupant monitoring systems and applications to supervise or otherwise track the activities and movements of a vehicle occupant.
  • the ability to determine and track the position, pose, posture, attentiveness, and other characteristics of an occupant assist the safe and efficient operation of the vehicle.
  • an occupant monitoring system may require identifying a driver in the driver seat prior to operating the vehicle.
  • the system may further monitor the occupant in the driver seat to ensure the driver is awake, alert and in a proper pose to take over operation of the vehicle should manual intervention be needed.
  • Traditional driver monitoring systems feature one or more sensors and/or cameras that may identify and track the driver's eyes, head position, pose and the like. Those sensors and cameras are highly affected by lighting conditions in and around the vehicle. Infrared sensors used to capture the driver's image, for example, may be highly sensitive to changing lighting conditions in the vehicle cabin due to ambient changes in light outside of the vehicle. Fast ambient lighting changes may cause the sensed image to be too dark or too washed out. For example, when the car makes a turn and the sun is on the driver's face, the image provided by the sensor may be washed out for a second or two. This may temporarily disable the driver monitoring system, causing it not function properly. Traditional driver monitoring systems, and adjustments to the image sensing components are reactive to changes in ambient lighting and lag behind the real-time changes to the environment.
  • aspects of the present disclosure provide for systems and methods to proactively adapt image sensor settings of an occupant monitoring system to accommodate ambient lighting changes.
  • the system may track current ambient light, track the driver, and predict upcoming changing lighting conditions. Sensors are used to predict when the ambient light will change. and preemptively adjust to prevent dark or washout images.
  • the sensor adjustment may be timed so that the adjustment is made just in time or concurrently with the change in ambient light conditions.
  • the system may predict when the light on the occupant's face will change and proactively readjust the settings of the sensor to accommodate the changed lighting conditions.
  • the imaging system may include one or more environmental sensors, a camera including one or more image sensors, and a predictive lighting module in communication with the one or more environmental sensors and the camera.
  • the predictive lighting module may be configured to receive environmental data from the one or more environmental sensors and adjust the one or more image sensors according to a first in-cabin lighting level.
  • a prospective environmental change may be determined based on the environmental data.
  • a second in-cabin lighting may be predicted level based upon the prospective environmental change and the one or more image sensors may be adjusted according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
  • a method of monitoring a vehicle occupant is provided.
  • environmental data may be received from the one or more environmental sensors and one or more image sensors may be adjusted according to a first in-cabin lighting level.
  • a prospective environmental change may be determined based on the environmental data and a second in-cabin lighting level may be predicted based upon the prospective environmental change.
  • the one or more image sensors may be adjusted according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
  • an occupant monitoring system for a vehicle may include an occupant facing camera comprising one or more image sensors, a communications module configured to receive navigational data, one or more environmental sensors configured to receive environmental data and a predictive lighting module in communication with the one or more environmental sensors and the occupant-facing camera.
  • the predictive lighting module may be configured to receive environmental data from the one or more environmental sensors and navigation information from the communications module.
  • One or more image sensors may be adjusted according to a first in-cabin lighting level.
  • a prospective environmental change may be determined based on the environmental data and the navigational data.
  • a second in-cabin lighting level may be predicted based upon the prospective environmental change.
  • the one or more image sensors may be adjusted according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
  • FIG. 1 is a diagram of a dynamic lighting environment and vehicle according to one aspect of the present disclosure.
  • FIG. 2 is a schematic diagram of a vehicle with an occupant monitoring system according to one aspect of the present disclosure.
  • FIG. 3 is a hardware implementation of an occupant monitoring system according to one aspect of the present disclosure.
  • FIG. 4 is a flow diagram of a method of monitoring a vehicle occupant according to one aspect of the present disclosure.
  • FIG. 1 depicts a dynamic lighting environment 100 and vehicle 101 according to one aspect of the present disclosure.
  • a vehicle 101 equipped with an occupant monitoring system 102 may experience a number of environmental light changes as it travels along a roadway 104 .
  • the vehicle 101 may enter a tunnel 112 or other structure, or shadowed area. While longer tunnels may have lighting or other illumination to assist drivers, the ambient light level may not match that outside of the tunnel 112 . Also, a shorter tunnel 112 may not offer any artificial illumination, relying instead on surrounding lighting conditions outside of the tunnel 112 to illuminate the roadway 104 .
  • the sun may shine light on or around the vehicle 101 from a number of angles and elevations (represented as positions 106 ′, 106 ′′, and 106 ′′′).
  • positions 106 ′, 106 ′′, and 106 ′′′ When the sun is in a first position 106 ′, as depicted in the exemplary environment of FIG. 1 , the sun's rays may be directly incident on the front of the vehicle 101 and the driver's face.
  • the occupant monitoring system's sensors may detect a brightness level in the cabin of the vehicle and adjust the image sensors accordingly to ensure accurate and consistent capturing of the driver.
  • the sun When the sun is in a second position 106 ′′, it may be behind a cloud or some other obstacle obscuring the vehicle 101 from direct sunlight. Additionally, as the sun may be at a different angle, the level of sunlight entering the cabin of the vehicle may decrease. In order to maintain a consistent image of the driver, the driver monitoring system may adjust the sensor settings to adjust for the change in cabin lighting. Further yet, when the sun is in a third position 106 ′′′, for example, there may be little or no direct sunlight incident on the driver's face, and the image sensors may be facing into the sun. Again, the in-cabin lighting conditions may change, necessitating an adjustment to the image sensors to ensure an accurate capture of the driver.
  • an occupant monitoring system may track current ambient lighting conditions and proactively determine upcoming changes to the ambient lighting.
  • the system may adjust the on-board image sensors according to a prospective change such that the sensors are adjusts substantially concurrently with the ambient lighting change.
  • the vehicle 101 equipped with an occupant monitoring system 102 may include a predictive lighting module to estimate and predict the amount of in-cabin lighting based on environmental and other external data.
  • Environmental data may include current ambient light levels as sensed by on-board sensors and other equipment, including cameras and other light sensing equipment. Cameras and other image sensing equipment may be configured to sense or perceive light levels at a distance from the vehicle 101 along the roadway 104 or the vehicle's projected or programmed path.
  • Such imaging devices may be part of or include a trajectory and navigation system configured to assist in operating and moving a vehicle through the environment.
  • Sensors and imaging devices may include a forward-facing camera, inertial measurement unit (IMU), driver facing camera, light intensity sensor, GPS, and ambient light sensor, LiDAR, RADAR, SONOR, or the like.
  • a sensor may sense or detect upcoming changes in ambient light by detecting large shadows or other structures that may occlude the vehicle's cabin from the light's source.
  • on-board sensors may track, over time, the shifting light levels in and around the vehicle 101 . Based on the vehicle's trajectory or route, and the trending changes to the ambient light, the predictive lighting module may predict a prospective environmental change, for example, an increase or decrease in ambient light.
  • the system may then proactively adjust the image sensors to account for the approaching change such that the sensors adjust at about the same time as the change in light level, i.e., substantially concurrently.
  • a LiDAR- or other camera-equipped vehicle may obtain images in front of the vehicle 101 that the on-board trajectory and navigation system may identify as a tunnel 112 .
  • the occupant monitoring system 102 in conjunction with the on-board trajectory and navigation system may determine a distance and time until the vehicle may enter the tunnel.
  • the occupant monitoring system may use this information to predict the in-cabin light levels when the vehicle 101 enters the tunnel.
  • the occupant monitoring system may then adjust the image sensors of the occupant-facing camera simultaneously with the change in ambient lighting.
  • the real-time, or near-real-time adjustment of the occupant monitoring system's image sensors eliminates any gap or transition period between when the lighting condition changes and the occupant monitoring system adjusts to that change.
  • external data may also inform the predictive lighting module of upcoming changes to ambient light levels.
  • the occupant monitoring system may communicate with external data sources, including for example, a satellite 110 , or other wireless communication sources to determine a potential location and severity of a prospective light change.
  • the trajectory and navigation system of the vehicle 101 may be in communication with the satellite 110 and receive data relevant to determining up-coming lighting conditions.
  • the satellite 110 , or other wireless communication source may transmit weather data, map information, or other information that may assist the predictive lighting module in determining a future change in lighting condition.
  • the predictive lighting module of a vehicle 101 traveling according to a pre-determined route guided by GPS or other navigation may obtain map information that indicates the tunnel 112 or other light-obstruction, such as buildings, mountains, or other obstacles.
  • the predictive lighting module may correlate that information with the on-board systems tracking ambient light levels to determine when the vehicle will reach the predicted lighting change and configure the settings of the occupant monitoring system accordingly and substantially concurrently with the light change. As such, any adjustment period between before, during, and after the lighting change may be eliminated and the occupant monitoring system may maintain a consistent and accurate image of the vehicle's occupants.
  • FIG. 2A is a schematic side-view 200 of a vehicle 101 equipped with an occupant monitoring system according to one aspect of the present disclosure.
  • the occupant 201 of the vehicle 101 depicted here as a driver facing the front of the vehicle 101 through a windshield 204 , may be monitored by a camera 202 of an occupant monitoring system.
  • the schematic of FIG. 2A depicts varying lighting conditions incident on the vehicle 101 and cabin. For example, when the vehicle 101 is facing the sun, direct sunlight may pass through the windshield 204 and illuminate the cabin and the face of the occupant 201 .
  • the sensor settings of the camera 202 may adjust according to the in-cabin light levels created by the direct sunlight to provide an accurate image of the occupant and enable precise monitoring and tracking of the occupant's behavior.
  • the moon 207 may be a source of ambient light and may shine little to no light on the vehicle 101 and the vehicle's cabin.
  • the darkness inside the cabin may require the camera 202 and image sensors to adjust accordingly to account for the low light levels to ensure the camera and image sensors can sense, monitor and track the occupant.
  • an occupant monitoring system including a predictive lighting module that tracks and detect changing ambient light levels.
  • the system further proactively adjusts image sensors so that the occupant monitoring system consistently and persistently images a vehicle occupant before, during, and after ambient light changes.
  • the predictive lighting module may predict the changes and adjust the settings of the camera 202 and image sensors accordingly.
  • FIG. 2B depicts a top view of a vehicle 101 equipped with an occupant monitoring system 102 according to aspects described herein.
  • the vehicle 101 may include one or more environmental sensors 236 to assist the vehicle 101 with its autonomous and semi-autonomous functions and sense changing environmental conditions.
  • the environmental sensors 236 are defined in a rear section and a front section of the vehicle 101 , however the environmental sensors 236 may be located elsewhere on the vehicle.
  • the environmental sensors 236 may be used to detect the driving environment from a 360° perspective.
  • the environmental sensors 236 may provide information related to the operation of the vehicle 101 , including the occupant monitoring system 102 and predictive lighting module.
  • the environmental sensors 236 may include RADAR, LiDAR, SONAR, or another type of sensor. Information received by the occupant monitoring system 102 from the sensor system may be used to determine the current driving environment of the vehicle as well as a future or approaching environment.
  • the occupant monitoring system 102 may include an occupant-facing camera 202 that is configured to obtain images, or a sequence of images, of one or more occupants of the vehicle, such as a driver D, or a passenger P.
  • the camera 202 may include infrared sensors to obtain high-resolution images of the occupants' eyes, face, hands, posture, pose, or the like.
  • the occupant monitoring system 102 may interface with other components and systems of the vehicle 101 to enhance and assist autonomous and semi-autonomous functions.
  • FIG. 3 is a diagram illustrating an example of a hardware implementation for an occupant monitoring system 300 , according to aspects of the present disclosure.
  • the occupant monitoring system 300 may be part of a passenger vehicle, a carrier vehicle, or other device.
  • the occupant monitoring system 300 may be a component of an autonomous or semi-autonomous car 328 .
  • Aspects of the present disclosure are not limited to the occupant monitoring system 300 being a component of the car 328 , as other devices, including, but not limited to, autonomous, semi-autonomous, or other vehicles may also include and use the occupant monitoring system 300 .
  • the occupant monitoring system 300 may be implemented with a bus architecture, represented generally by a bus 330 .
  • the bus 330 may include any number of interconnecting buses and bridges depending on the specific application of the occupant monitoring system 300 and the overall design constraints.
  • the bus 330 may link together various circuits including one or more processors and/or hardware modules, represented by a processor 320 , a communication module 322 , a location module 318 , a sensor module 302 , an actuation module 326 , a planning module 324 , and a computer-readable medium 314 .
  • the bus 330 may also link various other circuits such as timing sources, peripherals, voltage regulators, and power management circuits, which are well known in the art, and therefore, will not be described any further.
  • the occupant monitoring system 300 may include a transceiver 316 coupled to the processor 320 , the sensor module 302 , a predictive lighting module 308 , the communication module 322 , the location module 318 , the actuation module 326 , the planning module 324 , and the computer-readable medium 314 .
  • the transceiver 316 is coupled to an antenna 334 .
  • the transceiver 316 communicates with various other devices over a transmission medium. For example, the transceiver 316 may send and receive commands via transmissions to and from a server or a remote device, such as remote device or server (not shown).
  • the predictive lighting module 308 may include the processor 320 coupled to the computer-readable medium 314 .
  • the processor 320 may perform processing, including the execution of software stored on the computer-readable medium 314 providing functionality according to the disclosure.
  • the software when executed by the processor 320 , causes the occupant monitoring system 300 to perform the various functions described for a particular device, such as car 328 , or any of the modules 302 , 308 , 314 , 316 , 318 , 320 , 322 , 324 , 326 .
  • the computer-readable medium 314 may also be used for storing data that is manipulated by the processor 320 when executing the software.
  • the sensor module 302 may be used to obtain measurements via different sensors, such as a first sensor 306 , a second sensor 304 , and a third sensor 310 .
  • the first sensor 306 may be a motion sensor, such as an accelerometer, gyroscope, inertial measurement unit, or the like.
  • the second sensor may include a visual sensor, such as a stereoscopic camera, a red-green-blue (RGB) camera, LIDAR or RADAR.
  • the third sensor 304 may be an in-cabin sensor, such as a camera, CCD, infrared sensor, or the like, configured to obtain images of an occupant of the car 328 .
  • aspects of the present disclosure are not limited to the aforementioned sensors as other types of sensors, such as, for example, thermal, sonar, and/or lasers are also contemplated for either of the sensors 304 , 306 , 310 .
  • the measurements of the sensors 304 , 306 , 310 , 306 may be processed by one or more of the processor 320 , the sensor module 302 , the object tracking module 308 , the communication module 322 , the location module 318 , the actuation module 326 , the planning module 324 , in conjunction with the computer-readable medium 314 to implement the functionality described herein.
  • the data captured by the first sensor 306 , the second sensor 304 , and the third sensor 306 may be transmitted to an external device via the transceiver 316 .
  • the sensors 304 , 306 , 310 may be coupled to the car 328 or may be in communication with the car 328 .
  • the location module 318 may be used to determine a location of the car 328 .
  • the location module 318 may use a global positioning system (GPS) to determine the location of the car 328 .
  • GPS global positioning system
  • the occupant monitoring system 300 may be able to communicate with a remote monitoring service, such as mapping/navigation service, a weather service, or other environmental information provider.
  • Information obtained through the location module may assist in determining approaching changes in environmental conditions and ambient lighting conditions.
  • the information received through and generated by the location module 318 may inform the predictive lighting module 308 of approaching environmental conditions. If a perceived or preprogrammed route indicates the car 328 will be entering a tunnel, or urban area with tall buildings, or other environmental change, the predictive lighting module 308 may use such information to determine when and to what extent the in-cabin image sensors should be adjusted.
  • the communication module 322 may be used to facilitate communications via the transceiver 316 .
  • the communication module 322 may be configured to provide communication capabilities via different wireless protocols, such as Bluetooth, Wi-Fi, long term evolution (LTE), 3G, 5G, or the like.
  • the communications module may also be configured to establish a communication channel between the car 328 and an information provider.
  • the communication module 322 may also be used to communicate with other components of the car 328 that are not modules of the predictive lighting module 308 .
  • the occupant monitoring system 300 may also include the planning module 324 for planning a response to a detected or predicted change in environmental conditions.
  • the planning module may interface with or be a part of the predictive lighting module 308 .
  • the planning module 324 may include a set of instructions or settings that dictate how the occupant monitoring system 300 may respond when triggered by a change. For example, depending on the signals from any of the sensors 304 , 306 , 310 , detailing the direction, speed, and severity of a lighting change in the approaching environment, the planning module may respond with information necessary for the predictive lighting module to adjust the in-cabin sensors ensuring a consistent image capture of the occupant.
  • the planning module 324 may be software modules running in the processor 320 , resident/stored in the computer-readable medium 314 , one or more hardware modules coupled to the processor 320 , or some combination thereof.
  • the predictive lighting module 308 may be in communication with the sensor module 302 , the transceiver 316 , the processor 320 , the communication module 322 , the location module 318 , the actuation module 326 , the planning module 324 , and the computer-readable medium 314 .
  • the predictive lighting module 308 may receive sensor data from the sensor module 302 .
  • the sensor module 302 may receive the sensor data from the sensors 304 , 306 , 310 .
  • the sensor module 302 may filter the data to remove noise, encode the data, decode the data, merge the data, or perform other functions.
  • the predictive lighting module 308 may receive sensor data directly from the sensors 304 , 306 , 310 .
  • the predictive lighting module 308 may include or be in communication with the planning module 324 and/or the location module.
  • the predictive lighting module 308 based on the environmental information obtained from the sensor module 302 , communication module, and others, may estimate the in-cabin lighting conditions and levels for an approaching change in environmental conditions. Given, for example, current and recent light level data, GPS data, weather data, time-of-day data, and the like, the predictive lighting module 308 may proactively determine appropriate sensor settings for the in-cabin sensors before, during, and after the environmental change to ensure a consistent and persistent image capture of the occupant.
  • the ability to track and predict ambient light levels. proactively determine sensor settings, and adjust the in-cabin sensors in substantial real-time as the environmental changes occur may eliminate temporary image washouts or blackouts that can lead to errors or malfunctions in the occupant tracking system.
  • FIG. 4 depicts a method 400 of monitoring a vehicle occupant according to an aspect of the present disclosure.
  • an occupant monitoring system may receive environmental data, as described herein, from one or more environmental sensors, cameras, or other information sources.
  • Environmental data may include information relating to, without limitation, traffic, route navigation, weather, ambient temperatures, ambient light conditions, buildings, structures, terrain, or the like.
  • the system may, as shown in block 404 , adjust one or more image sensors of the occupant monitoring system according to a first in-cabin light level.
  • the image sensors may require adjustment based on the ambient light level in the vehicle to ensure capturing a consistent and clear image of the occupant.
  • the first in-cabin light level may be set or adjusted according to the environmental data received from the environmental sensors.
  • the system may continually monitor the environment as well as in-cabin ambient lighting.
  • the environmental sensors may continuously update local and remote environmental information that may be relevant to the system, including environmental data on a programmed navigational route.
  • the system may also continually monitor the ambient light levels in the cabin of the vehicle. Such information may be monitored to assist in adjusting the image sensors capturing images of the vehicle occupant.
  • the system may determine a prospective environmental change.
  • the prospective environmental change may include changes to the local environment or the environment into which the vehicle may enter in the future.
  • the prospective environmental change may lead to a potential change in in-cabin lighting levels that may affect the image capturing capabilities of the system.
  • the system may use environmental data obtained from the environmental sensors to determine an upcoming change in environmental conditions that may change the ambient light level in the cabin and therefore necessitate an adjustment to the image sensors.
  • the system may determine a time, T, when the environmental change will occur. For example, if the system determines that the vehicle may be approaching a tunnel or other light-obscuring structure on a navigational route, the system may estimate, based on current operating conditions, when the vehicle may arrive at the tunnel. Estimating the time, T, at which the environmental change may occur may allow the system to proactively determine adjustments to the image sensors needed to maintain consistent image capture of the occupant. Further, such an estimation may allow the system to make the determined adjustments at same time as, or substantially concurrently with, the change in the environment.
  • the system may determine a time range, T T ⁇ X ⁇ T ⁇ T T+X , where X is a predetermined time before and after the time, T.
  • the system may define the range such that changes to the image sensors occur within a range of time around when the environmental change occurs.
  • the system may predict a second in-cabin lighting level according to the prospective environmental change.
  • the system may use the environmental data to estimate the lighting conditions in the vehicle cabin at or after the environmental change occurs. Based on the estimated lighting conditions, the system may define the second in-cabin lighting level accordingly and prepare to adjust the image sensors to account for the change in lighting.
  • the system may determine if the time, T, of the environmental change is nearing or occurring. According to one aspect, the system may determine whether the time, T, is within a time range triggering the adjustment of the image sensors. If the system determines that T is within the determined range, the system may make the adjustments to the image sensors according to the second in-cabin lighting level, as shown in block 418 . The system determines that T is not within the determined range, indicating that the environmental change has not yet occurred, the system may continue to monitor the environment and in-cabin lighting conditions until T occurs or falls within the time range.
  • the system may continue to monitor the environment and in-cabin lighting for other prospective changes.
  • aspects of the present disclosure describe an occupant monitoring system focused on a driver or a car, one skilled on the art will appreciate that the system may be implemented to monitor any occupant or multiple occupants in any vehicle type. Further, while the systems disclosed herein describe an occupant monitoring system in communication with GPS, weather and on-board light sensors, one skilled in the art will appreciate that the information on which the system may predict approaching environmental changes is not limited to those specific sources and use of any source or information detailing lighting and environmental conditions is within the scope of the present disclosure.
  • determining encompasses a wide variety of actions. For example, “determining” may include calculating, computing, processing, deriving, investigating, looking up (e.g., looking up in a table, a database or another data structure), ascertaining and the like. Additionally, “determining” may include receiving (e.g., receiving information), accessing (e.g., accessing data in a memory) and the like. Furthermore, “determining” may include resolving, selecting, choosing, establishing, and the like.
  • a phrase referring to “at least one of” a list of items refers to any combination of those items, including single members.
  • “at least one of: a, b, or c” is intended to cover: a, b, c, a-b, a-c, b-c, and a-b-c.
  • the various illustrative logical blocks, modules and circuits described in connection with the present disclosure may be implemented or performed with a processor specially configured to perform the functions discussed in the present disclosure.
  • the processor may be a neural network processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array signal (FPGA) or other programmable logic device (PLD), discrete gate or transistor logic, discrete hardware components or any combination thereof designed to perform the functions described herein.
  • the processing system may comprise one or more neuromorphic processors for implementing the neuron models and models of neural systems described herein.
  • the processor may be a microprocessor, controller, microcontroller, or state machine specially configured as described herein.
  • a processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or such other special configuration, as described herein.
  • a software module may reside in storage or machine readable medium, including random access memory (RAM), read only memory (ROM), flash memory, erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), registers, a hard disk, a removable disk, a CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer.
  • RAM random access memory
  • ROM read only memory
  • EPROM erasable programmable read-only memory
  • EEPROM electrically erasable programmable read-only memory
  • registers a hard disk, a removable disk, a CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer.
  • a software module may comprise a single instruction, or many instructions, and may be distributed over several different code segments, among different programs, and across multiple storage media.
  • a storage medium may be coupled to a processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor.
  • the methods disclosed herein comprise one or more steps or actions for achieving the described method.
  • the method steps and/or actions may be interchanged with one another without departing from the scope of the claims.
  • the order and/or use of specific steps and/or actions may be modified without departing from the scope of the claims.
  • an example hardware configuration may comprise a processing system in a device.
  • the processing system may be implemented with a bus architecture.
  • the bus may include any number of interconnecting buses and bridges depending on the specific application of the processing system and the overall design constraints.
  • the bus may link together various circuits including a processor, machine-readable media, and a bus interface.
  • the bus interface may be used to connect a network adapter, among other things, to the processing system via the bus.
  • the network adapter may be used to implement signal processing functions.
  • a user interface e.g., keypad, display, mouse, joystick, etc.
  • the bus may also link various other circuits such as timing sources, peripherals, voltage regulators, power management circuits, and the like, which are well known in the art, and therefore, will not be described any further.
  • the processor may be responsible for managing the bus and processing, including the execution of software stored on the machine-readable media.
  • Software shall be construed to mean instructions, data, or any combination thereof, whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise.
  • the machine-readable media may be part of the processing system separate from the processor.
  • the machine-readable media, or any portion thereof may be external to the processing system.
  • the machine-readable media may include a transmission line, a carrier wave modulated by data, and/or a computer product separate from the device, all which may be accessed by the processor through the bus interface.
  • the machine-readable media, or any portion thereof may be integrated into the processor, such as the case may be with cache and/or specialized register files.
  • the various components discussed may be described as having a specific location, such as a local component, they may also be configured in various ways, such as certain components being configured as part of a distributed computing system.
  • the machine-readable media may comprise a number of software modules.
  • the software modules may include a transmission module and a receiving module. Each software module may reside in a single storage device or be distributed across multiple storage devices.
  • a software module may be loaded into RANI from a hard drive when a triggering event occurs.
  • the processor may load some of the instructions into cache to increase access speed.
  • One or more cache lines may then be loaded into a special purpose register file for execution by the processor.
  • Computer-readable media include both computer storage media and communication media including any storage medium that facilitates transfer of a computer program from one place to another.
  • modules and/or other appropriate means for performing the methods and techniques described herein can be downloaded and/or otherwise obtained by a user terminal and/or base station as applicable.
  • a user terminal and/or base station can be coupled to a server to facilitate the transfer of means for performing the methods described herein.
  • various methods described herein can be provided via storage means, such that a user terminal and/or base station can obtain the various methods upon coupling or providing the storage means to the device.
  • any other suitable technique for providing the methods and techniques described herein to a device can be utilized.

Abstract

Systems and methods to proactively adapt image sensor settings of an occupant monitoring system to accommodate ambient lighting changes are provided. The system may track current ambient light, track the driver, and predict upcoming changing lighting conditions. Sensors are used to predict when the ambient light will change and preemptively adjust to prevent dark or washout images. The sensor adjustment may be timed so that the adjustment is made just in time or concurrently with the change in ambient light conditions. The system may predict when the light on the occupant's face will change and proactively readjust the settings of the sensor to accommodate the changed lighting conditions.

Description

    FIELD OF TECHNOLOGY
  • The present disclosure relates to a system and method for monitoring the occupant of a vehicle, more particularly to an occupant monitoring system to proactively adjust monitoring sensors as ambient light changes.
  • BACKGROUND
  • Autonomous and semi-autonomous vehicles rely on occupant monitoring systems and applications to supervise or otherwise track the activities and movements of a vehicle occupant. The ability to determine and track the position, pose, posture, attentiveness, and other characteristics of an occupant assist the safe and efficient operation of the vehicle. For example, an occupant monitoring system may require identifying a driver in the driver seat prior to operating the vehicle. The system may further monitor the occupant in the driver seat to ensure the driver is awake, alert and in a proper pose to take over operation of the vehicle should manual intervention be needed.
  • Traditional driver monitoring systems feature one or more sensors and/or cameras that may identify and track the driver's eyes, head position, pose and the like. Those sensors and cameras are highly affected by lighting conditions in and around the vehicle. Infrared sensors used to capture the driver's image, for example, may be highly sensitive to changing lighting conditions in the vehicle cabin due to ambient changes in light outside of the vehicle. Fast ambient lighting changes may cause the sensed image to be too dark or too washed out. For example, when the car makes a turn and the sun is on the driver's face, the image provided by the sensor may be washed out for a second or two. This may temporarily disable the driver monitoring system, causing it not function properly. Traditional driver monitoring systems, and adjustments to the image sensing components are reactive to changes in ambient lighting and lag behind the real-time changes to the environment.
  • SUMMARY
  • Aspects of the present disclosure provide for systems and methods to proactively adapt image sensor settings of an occupant monitoring system to accommodate ambient lighting changes. The system may track current ambient light, track the driver, and predict upcoming changing lighting conditions. Sensors are used to predict when the ambient light will change. and preemptively adjust to prevent dark or washout images. The sensor adjustment may be timed so that the adjustment is made just in time or concurrently with the change in ambient light conditions. The system may predict when the light on the occupant's face will change and proactively readjust the settings of the sensor to accommodate the changed lighting conditions.
  • According to one aspect an imaging system is disclosed. The imaging system may include one or more environmental sensors, a camera including one or more image sensors, and a predictive lighting module in communication with the one or more environmental sensors and the camera. The predictive lighting module may be configured to receive environmental data from the one or more environmental sensors and adjust the one or more image sensors according to a first in-cabin lighting level. A prospective environmental change may be determined based on the environmental data. A second in-cabin lighting may be predicted level based upon the prospective environmental change and the one or more image sensors may be adjusted according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
  • According to another aspect of the present disclosure, a method of monitoring a vehicle occupant is provided. According to the method, environmental data may be received from the one or more environmental sensors and one or more image sensors may be adjusted according to a first in-cabin lighting level. A prospective environmental change may be determined based on the environmental data and a second in-cabin lighting level may be predicted based upon the prospective environmental change. The one or more image sensors may be adjusted according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
  • According to another aspect of the present disclosure, an occupant monitoring system for a vehicle may include an occupant facing camera comprising one or more image sensors, a communications module configured to receive navigational data, one or more environmental sensors configured to receive environmental data and a predictive lighting module in communication with the one or more environmental sensors and the occupant-facing camera. The predictive lighting module may be configured to receive environmental data from the one or more environmental sensors and navigation information from the communications module. One or more image sensors may be adjusted according to a first in-cabin lighting level. A prospective environmental change may be determined based on the environmental data and the navigational data. A second in-cabin lighting level may be predicted based upon the prospective environmental change. The one or more image sensors may be adjusted according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
  • This has outlined, rather broadly, the features and technical advantages of the present disclosure in order that the detailed description that follows may be better understood. Additional features and advantages of the present disclosure will be described below. It should be appreciated by those skilled in the art that this present disclosure may be readily utilized as a basis for modifying or designing other structures for carrying out the same purposes of the present disclosure. It should also be realized by those skilled in the art that such equivalent constructions do not depart from the teachings of the present disclosure as set forth in the appended claims. The novel features, which are believed to be characteristic of the present disclosure, both as to its organization and method of operation, together with further objects and advantages, will be better understood from the following description when considered in connection with the accompanying figures. It is to be expressly understood, however, that each of the figures is provided for the purpose of illustration and description only and is not intended as a definition of the limits of the present disclosure.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The features, nature, and advantages of the present disclosure will become more apparent from the detailed description set forth below when taken in conjunction with the drawings in which like reference characters identify correspondingly throughout.
  • FIG. 1 is a diagram of a dynamic lighting environment and vehicle according to one aspect of the present disclosure.
  • FIG. 2 is a schematic diagram of a vehicle with an occupant monitoring system according to one aspect of the present disclosure.
  • FIG. 3 is a hardware implementation of an occupant monitoring system according to one aspect of the present disclosure.
  • FIG. 4 is a flow diagram of a method of monitoring a vehicle occupant according to one aspect of the present disclosure.
  • DETAILED DESCRIPTION
  • The detailed description set forth below, in connection with the appended drawings, is intended as a description of various configurations and is not intended to represent the only configurations in which the concepts described herein may be practiced. The detailed description includes specific details for providing a thorough understanding of the various concepts. It will be apparent to those skilled in the art, however, that these concepts may be practiced without these specific details. In some instances, well-known structures and components are shown in block diagram form in order to avoid obscuring such concepts.
  • Aspects of the present disclosure provide an occupant monitoring system with proactive lighting adjustment to accommodate pending changes to ambient lighting. FIG. 1 depicts a dynamic lighting environment 100 and vehicle 101 according to one aspect of the present disclosure. A vehicle 101 equipped with an occupant monitoring system 102 may experience a number of environmental light changes as it travels along a roadway 104. For example, the vehicle 101 may enter a tunnel 112 or other structure, or shadowed area. While longer tunnels may have lighting or other illumination to assist drivers, the ambient light level may not match that outside of the tunnel 112. Also, a shorter tunnel 112 may not offer any artificial illumination, relying instead on surrounding lighting conditions outside of the tunnel 112 to illuminate the roadway 104.
  • Additionally, as the vehicle 101 travels the environment may change. For example, the sun may shine light on or around the vehicle 101 from a number of angles and elevations (represented as positions 106′, 106″, and 106′″). When the sun is in a first position 106′, as depicted in the exemplary environment of FIG. 1, the sun's rays may be directly incident on the front of the vehicle 101 and the driver's face. The occupant monitoring system's sensors may detect a brightness level in the cabin of the vehicle and adjust the image sensors accordingly to ensure accurate and consistent capturing of the driver.
  • When the sun is in a second position 106″, it may be behind a cloud or some other obstacle obscuring the vehicle 101 from direct sunlight. Additionally, as the sun may be at a different angle, the level of sunlight entering the cabin of the vehicle may decrease. In order to maintain a consistent image of the driver, the driver monitoring system may adjust the sensor settings to adjust for the change in cabin lighting. Further yet, when the sun is in a third position 106′″, for example, there may be little or no direct sunlight incident on the driver's face, and the image sensors may be facing into the sun. Again, the in-cabin lighting conditions may change, necessitating an adjustment to the image sensors to ensure an accurate capture of the driver.
  • According to one aspect of the present disclosure, an occupant monitoring system may track current ambient lighting conditions and proactively determine upcoming changes to the ambient lighting. The system may adjust the on-board image sensors according to a prospective change such that the sensors are adjusts substantially concurrently with the ambient lighting change. For example, the vehicle 101 equipped with an occupant monitoring system 102 may include a predictive lighting module to estimate and predict the amount of in-cabin lighting based on environmental and other external data. Environmental data may include current ambient light levels as sensed by on-board sensors and other equipment, including cameras and other light sensing equipment. Cameras and other image sensing equipment may be configured to sense or perceive light levels at a distance from the vehicle 101 along the roadway 104 or the vehicle's projected or programmed path. Such imaging devices may be part of or include a trajectory and navigation system configured to assist in operating and moving a vehicle through the environment.
  • Sensors and imaging devices may include a forward-facing camera, inertial measurement unit (IMU), driver facing camera, light intensity sensor, GPS, and ambient light sensor, LiDAR, RADAR, SONOR, or the like. A sensor may sense or detect upcoming changes in ambient light by detecting large shadows or other structures that may occlude the vehicle's cabin from the light's source. For example, on-board sensors may track, over time, the shifting light levels in and around the vehicle 101. Based on the vehicle's trajectory or route, and the trending changes to the ambient light, the predictive lighting module may predict a prospective environmental change, for example, an increase or decrease in ambient light. The system may then proactively adjust the image sensors to account for the approaching change such that the sensors adjust at about the same time as the change in light level, i.e., substantially concurrently.
  • As another example, a LiDAR- or other camera-equipped vehicle may obtain images in front of the vehicle 101 that the on-board trajectory and navigation system may identify as a tunnel 112. The occupant monitoring system 102, in conjunction with the on-board trajectory and navigation system may determine a distance and time until the vehicle may enter the tunnel. The occupant monitoring system may use this information to predict the in-cabin light levels when the vehicle 101 enters the tunnel. The occupant monitoring system may then adjust the image sensors of the occupant-facing camera simultaneously with the change in ambient lighting. The real-time, or near-real-time adjustment of the occupant monitoring system's image sensors eliminates any gap or transition period between when the lighting condition changes and the occupant monitoring system adjusts to that change.
  • According to another aspect, external data may also inform the predictive lighting module of upcoming changes to ambient light levels. The occupant monitoring system may communicate with external data sources, including for example, a satellite 110, or other wireless communication sources to determine a potential location and severity of a prospective light change. The trajectory and navigation system of the vehicle 101 may be in communication with the satellite 110 and receive data relevant to determining up-coming lighting conditions. The satellite 110, or other wireless communication source may transmit weather data, map information, or other information that may assist the predictive lighting module in determining a future change in lighting condition. For example, the predictive lighting module of a vehicle 101 traveling according to a pre-determined route guided by GPS or other navigation, may obtain map information that indicates the tunnel 112 or other light-obstruction, such as buildings, mountains, or other obstacles. The predictive lighting module may correlate that information with the on-board systems tracking ambient light levels to determine when the vehicle will reach the predicted lighting change and configure the settings of the occupant monitoring system accordingly and substantially concurrently with the light change. As such, any adjustment period between before, during, and after the lighting change may be eliminated and the occupant monitoring system may maintain a consistent and accurate image of the vehicle's occupants.
  • FIG. 2A is a schematic side-view 200 of a vehicle 101 equipped with an occupant monitoring system according to one aspect of the present disclosure. The occupant 201 of the vehicle 101, depicted here as a driver facing the front of the vehicle 101 through a windshield 204, may be monitored by a camera 202 of an occupant monitoring system. The schematic of FIG. 2A depicts varying lighting conditions incident on the vehicle 101 and cabin. For example, when the vehicle 101 is facing the sun, direct sunlight may pass through the windshield 204 and illuminate the cabin and the face of the occupant 201. The sensor settings of the camera 202 may adjust according to the in-cabin light levels created by the direct sunlight to provide an accurate image of the occupant and enable precise monitoring and tracking of the occupant's behavior. Conversely, at a different time of day, the moon 207 may be a source of ambient light and may shine little to no light on the vehicle 101 and the vehicle's cabin. The darkness inside the cabin may require the camera 202 and image sensors to adjust accordingly to account for the low light levels to ensure the camera and image sensors can sense, monitor and track the occupant.
  • As described herein, aspects of the present disclosure provide for an occupant monitoring system including a predictive lighting module that tracks and detect changing ambient light levels. The system further proactively adjusts image sensors so that the occupant monitoring system consistently and persistently images a vehicle occupant before, during, and after ambient light changes. For example, as depicted in FIG. 2A, as the light from the sun 206 changes due to movement across the sky, sunset, the vehicle's changing trajectory, occluding clouds or buildings, or the like, the predictive lighting module may predict the changes and adjust the settings of the camera 202 and image sensors accordingly.
  • FIG. 2B depicts a top view of a vehicle 101 equipped with an occupant monitoring system 102 according to aspects described herein. The vehicle 101 may include one or more environmental sensors 236 to assist the vehicle 101 with its autonomous and semi-autonomous functions and sense changing environmental conditions. For simplicity, in FIG. 2B, the environmental sensors 236 are defined in a rear section and a front section of the vehicle 101, however the environmental sensors 236 may be located elsewhere on the vehicle. The environmental sensors 236 may be used to detect the driving environment from a 360° perspective. The environmental sensors 236 may provide information related to the operation of the vehicle 101, including the occupant monitoring system 102 and predictive lighting module. The environmental sensors 236 may include RADAR, LiDAR, SONAR, or another type of sensor. Information received by the occupant monitoring system 102 from the sensor system may be used to determine the current driving environment of the vehicle as well as a future or approaching environment.
  • According to one aspect of the present disclosure, the occupant monitoring system 102 may include an occupant-facing camera 202 that is configured to obtain images, or a sequence of images, of one or more occupants of the vehicle, such as a driver D, or a passenger P. The camera 202 may include infrared sensors to obtain high-resolution images of the occupants' eyes, face, hands, posture, pose, or the like. The occupant monitoring system 102 may interface with other components and systems of the vehicle 101 to enhance and assist autonomous and semi-autonomous functions.
  • FIG. 3 is a diagram illustrating an example of a hardware implementation for an occupant monitoring system 300, according to aspects of the present disclosure. The occupant monitoring system 300 may be part of a passenger vehicle, a carrier vehicle, or other device. For example, as shown in FIG. 3, the occupant monitoring system 300 may be a component of an autonomous or semi-autonomous car 328. Aspects of the present disclosure are not limited to the occupant monitoring system 300 being a component of the car 328, as other devices, including, but not limited to, autonomous, semi-autonomous, or other vehicles may also include and use the occupant monitoring system 300.
  • The occupant monitoring system 300 may be implemented with a bus architecture, represented generally by a bus 330. The bus 330 may include any number of interconnecting buses and bridges depending on the specific application of the occupant monitoring system 300 and the overall design constraints. The bus 330 may link together various circuits including one or more processors and/or hardware modules, represented by a processor 320, a communication module 322, a location module 318, a sensor module 302, an actuation module 326, a planning module 324, and a computer-readable medium 314. The bus 330 may also link various other circuits such as timing sources, peripherals, voltage regulators, and power management circuits, which are well known in the art, and therefore, will not be described any further.
  • The occupant monitoring system 300 may include a transceiver 316 coupled to the processor 320, the sensor module 302, a predictive lighting module 308, the communication module 322, the location module 318, the actuation module 326, the planning module 324, and the computer-readable medium 314. The transceiver 316 is coupled to an antenna 334. The transceiver 316 communicates with various other devices over a transmission medium. For example, the transceiver 316 may send and receive commands via transmissions to and from a server or a remote device, such as remote device or server (not shown).
  • The predictive lighting module 308 may include the processor 320 coupled to the computer-readable medium 314. The processor 320 may perform processing, including the execution of software stored on the computer-readable medium 314 providing functionality according to the disclosure. The software, when executed by the processor 320, causes the occupant monitoring system 300 to perform the various functions described for a particular device, such as car 328, or any of the modules 302, 308, 314, 316, 318, 320, 322, 324, 326. The computer-readable medium 314 may also be used for storing data that is manipulated by the processor 320 when executing the software.
  • The sensor module 302 may be used to obtain measurements via different sensors, such as a first sensor 306, a second sensor 304, and a third sensor 310. The first sensor 306 may be a motion sensor, such as an accelerometer, gyroscope, inertial measurement unit, or the like. The second sensor may include a visual sensor, such as a stereoscopic camera, a red-green-blue (RGB) camera, LIDAR or RADAR. The third sensor 304 may be an in-cabin sensor, such as a camera, CCD, infrared sensor, or the like, configured to obtain images of an occupant of the car 328. Of course, aspects of the present disclosure are not limited to the aforementioned sensors as other types of sensors, such as, for example, thermal, sonar, and/or lasers are also contemplated for either of the sensors 304, 306, 310. The measurements of the sensors 304, 306, 310, 306 may be processed by one or more of the processor 320, the sensor module 302, the object tracking module 308, the communication module 322, the location module 318, the actuation module 326, the planning module 324, in conjunction with the computer-readable medium 314 to implement the functionality described herein. In one configuration, the data captured by the first sensor 306, the second sensor 304, and the third sensor 306 may be transmitted to an external device via the transceiver 316. The sensors 304, 306, 310 may be coupled to the car 328 or may be in communication with the car 328.
  • The location module 318 may be used to determine a location of the car 328. For example, the location module 318 may use a global positioning system (GPS) to determine the location of the car 328. For example, the occupant monitoring system 300 may be able to communicate with a remote monitoring service, such as mapping/navigation service, a weather service, or other environmental information provider. Information obtained through the location module may assist in determining approaching changes in environmental conditions and ambient lighting conditions. The information received through and generated by the location module 318 may inform the predictive lighting module 308 of approaching environmental conditions. If a perceived or preprogrammed route indicates the car 328 will be entering a tunnel, or urban area with tall buildings, or other environmental change, the predictive lighting module 308 may use such information to determine when and to what extent the in-cabin image sensors should be adjusted.
  • The communication module 322 may be used to facilitate communications via the transceiver 316. For example, the communication module 322 may be configured to provide communication capabilities via different wireless protocols, such as Bluetooth, Wi-Fi, long term evolution (LTE), 3G, 5G, or the like. The communications module may also be configured to establish a communication channel between the car 328 and an information provider. The communication module 322 may also be used to communicate with other components of the car 328 that are not modules of the predictive lighting module 308.
  • The occupant monitoring system 300 may also include the planning module 324 for planning a response to a detected or predicted change in environmental conditions. The planning module may interface with or be a part of the predictive lighting module 308. The planning module 324 may include a set of instructions or settings that dictate how the occupant monitoring system 300 may respond when triggered by a change. For example, depending on the signals from any of the sensors 304, 306, 310, detailing the direction, speed, and severity of a lighting change in the approaching environment, the planning module may respond with information necessary for the predictive lighting module to adjust the in-cabin sensors ensuring a consistent image capture of the occupant. The planning module 324, as well as other modules described herein, may be software modules running in the processor 320, resident/stored in the computer-readable medium 314, one or more hardware modules coupled to the processor 320, or some combination thereof.
  • The predictive lighting module 308 may be in communication with the sensor module 302, the transceiver 316, the processor 320, the communication module 322, the location module 318, the actuation module 326, the planning module 324, and the computer-readable medium 314. In one configuration, the predictive lighting module 308 may receive sensor data from the sensor module 302. The sensor module 302 may receive the sensor data from the sensors 304, 306, 310. According to aspects of the disclosure, the sensor module 302 may filter the data to remove noise, encode the data, decode the data, merge the data, or perform other functions. In an alternate configuration, the predictive lighting module 308 may receive sensor data directly from the sensors 304, 306, 310.
  • As shown in FIG. 3, the predictive lighting module 308 may include or be in communication with the planning module 324 and/or the location module. The predictive lighting module 308, based on the environmental information obtained from the sensor module 302, communication module, and others, may estimate the in-cabin lighting conditions and levels for an approaching change in environmental conditions. Given, for example, current and recent light level data, GPS data, weather data, time-of-day data, and the like, the predictive lighting module 308 may proactively determine appropriate sensor settings for the in-cabin sensors before, during, and after the environmental change to ensure a consistent and persistent image capture of the occupant. The ability to track and predict ambient light levels. proactively determine sensor settings, and adjust the in-cabin sensors in substantial real-time as the environmental changes occur may eliminate temporary image washouts or blackouts that can lead to errors or malfunctions in the occupant tracking system.
  • FIG. 4 depicts a method 400 of monitoring a vehicle occupant according to an aspect of the present disclosure. As shown in block 402, an occupant monitoring system may receive environmental data, as described herein, from one or more environmental sensors, cameras, or other information sources. Environmental data may include information relating to, without limitation, traffic, route navigation, weather, ambient temperatures, ambient light conditions, buildings, structures, terrain, or the like.
  • The system may, as shown in block 404, adjust one or more image sensors of the occupant monitoring system according to a first in-cabin light level. As described herein, the image sensors may require adjustment based on the ambient light level in the vehicle to ensure capturing a consistent and clear image of the occupant. The first in-cabin light level may be set or adjusted according to the environmental data received from the environmental sensors.
  • As shown in block 406, the system may continually monitor the environment as well as in-cabin ambient lighting. The environmental sensors may continuously update local and remote environmental information that may be relevant to the system, including environmental data on a programmed navigational route. The system may also continually monitor the ambient light levels in the cabin of the vehicle. Such information may be monitored to assist in adjusting the image sensors capturing images of the vehicle occupant.
  • According to one aspect, as shown in block 408, the system may determine a prospective environmental change. The prospective environmental change may include changes to the local environment or the environment into which the vehicle may enter in the future. As described herein, the prospective environmental change may lead to a potential change in in-cabin lighting levels that may affect the image capturing capabilities of the system. As such, the system may use environmental data obtained from the environmental sensors to determine an upcoming change in environmental conditions that may change the ambient light level in the cabin and therefore necessitate an adjustment to the image sensors.
  • In addition to determining a prospective environmental change, as shown in block 410, the system may determine a time, T, when the environmental change will occur. For example, if the system determines that the vehicle may be approaching a tunnel or other light-obscuring structure on a navigational route, the system may estimate, based on current operating conditions, when the vehicle may arrive at the tunnel. Estimating the time, T, at which the environmental change may occur may allow the system to proactively determine adjustments to the image sensors needed to maintain consistent image capture of the occupant. Further, such an estimation may allow the system to make the determined adjustments at same time as, or substantially concurrently with, the change in the environment.
  • As shown in block 412, the system may determine a time range, TT−X<T<TT+X, where X is a predetermined time before and after the time, T. The system may define the range such that changes to the image sensors occur within a range of time around when the environmental change occurs. Given a moving vehicle, and perhaps moving environmental objects impacting the in-cabin light levels, a time range may be implemented to ensure that any adjustments made to the image sensors are executed as closely to the change in environmental conditions as possible. For example, if the system determines the vehicle may be entering an approaching tunnel in T=60 seconds, the system may define the time range to adjust the image sensors within 1 second (X=1 second). Defining a narrow time range allows the system to make adjustments to the image sensors substantially concurrently with the environmental change.
  • As shown in block 414, the system may predict a second in-cabin lighting level according to the prospective environmental change. The system may use the environmental data to estimate the lighting conditions in the vehicle cabin at or after the environmental change occurs. Based on the estimated lighting conditions, the system may define the second in-cabin lighting level accordingly and prepare to adjust the image sensors to account for the change in lighting.
  • As shown in block 416, the system may determine if the time, T, of the environmental change is nearing or occurring. According to one aspect, the system may determine whether the time, T, is within a time range triggering the adjustment of the image sensors. If the system determines that T is within the determined range, the system may make the adjustments to the image sensors according to the second in-cabin lighting level, as shown in block 418. The system determines that T is not within the determined range, indicating that the environmental change has not yet occurred, the system may continue to monitor the environment and in-cabin lighting conditions until T occurs or falls within the time range.
  • After the image sensors are adjusted substantially concurrently with the environmental change, the system may continue to monitor the environment and in-cabin lighting for other prospective changes.
  • While aspects of the present disclosure describe an occupant monitoring system focused on a driver or a car, one skilled on the art will appreciate that the system may be implemented to monitor any occupant or multiple occupants in any vehicle type. Further, while the systems disclosed herein describe an occupant monitoring system in communication with GPS, weather and on-board light sensors, one skilled in the art will appreciate that the information on which the system may predict approaching environmental changes is not limited to those specific sources and use of any source or information detailing lighting and environmental conditions is within the scope of the present disclosure.
  • Based on the teachings, one skilled in the art should appreciate that the scope of the present disclosure is intended to cover any aspect of the present disclosure, whether implemented independently of or combined with any other aspect of the present disclosure. For example, an apparatus may be implemented or a method may be practiced using any number of the aspects set forth. In addition, the scope of the present disclosure is intended to cover such an apparatus or method practiced using other structure, functionality, or structure and functionality in addition to, or other than the various aspects of the present disclosure set forth. It should be understood that any aspect of the present disclosure may be embodied by one or more elements of a claim.
  • The word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any aspect described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects.
  • Although particular aspects are described herein, many variations and permutations of these aspects fall within the scope of the present disclosure. Although some benefits and advantages of the preferred aspects are mentioned, the scope of the present disclosure is not intended to be limited to particular benefits, uses or objectives. Rather, aspects of the present disclosure are intended to be broadly applicable to different technologies, system configurations, networks and protocols, some of which are illustrated by way of example in the figures and in the following description of the preferred aspects. The detailed description and drawings are merely illustrative of the present disclosure rather than limiting, the scope of the present disclosure being defined by the appended claims and equivalents thereof.
  • As used herein, the term “determining” encompasses a wide variety of actions. For example, “determining” may include calculating, computing, processing, deriving, investigating, looking up (e.g., looking up in a table, a database or another data structure), ascertaining and the like. Additionally, “determining” may include receiving (e.g., receiving information), accessing (e.g., accessing data in a memory) and the like. Furthermore, “determining” may include resolving, selecting, choosing, establishing, and the like.
  • As used herein, a phrase referring to “at least one of” a list of items refers to any combination of those items, including single members. As an example, “at least one of: a, b, or c” is intended to cover: a, b, c, a-b, a-c, b-c, and a-b-c.
  • The various illustrative logical blocks, modules and circuits described in connection with the present disclosure may be implemented or performed with a processor specially configured to perform the functions discussed in the present disclosure. The processor may be a neural network processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array signal (FPGA) or other programmable logic device (PLD), discrete gate or transistor logic, discrete hardware components or any combination thereof designed to perform the functions described herein. Alternatively, the processing system may comprise one or more neuromorphic processors for implementing the neuron models and models of neural systems described herein. The processor may be a microprocessor, controller, microcontroller, or state machine specially configured as described herein. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or such other special configuration, as described herein.
  • The steps of a method or algorithm described in connection with the present disclosure may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in storage or machine readable medium, including random access memory (RAM), read only memory (ROM), flash memory, erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), registers, a hard disk, a removable disk, a CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer. A software module may comprise a single instruction, or many instructions, and may be distributed over several different code segments, among different programs, and across multiple storage media. A storage medium may be coupled to a processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor.
  • The methods disclosed herein comprise one or more steps or actions for achieving the described method. The method steps and/or actions may be interchanged with one another without departing from the scope of the claims. In other words, unless a specific order of steps or actions is specified, the order and/or use of specific steps and/or actions may be modified without departing from the scope of the claims.
  • The functions described may be implemented in hardware, software, firmware, or any combination thereof. If implemented in hardware, an example hardware configuration may comprise a processing system in a device. The processing system may be implemented with a bus architecture. The bus may include any number of interconnecting buses and bridges depending on the specific application of the processing system and the overall design constraints. The bus may link together various circuits including a processor, machine-readable media, and a bus interface. The bus interface may be used to connect a network adapter, among other things, to the processing system via the bus. The network adapter may be used to implement signal processing functions. For certain aspects, a user interface (e.g., keypad, display, mouse, joystick, etc.) may also be connected to the bus. The bus may also link various other circuits such as timing sources, peripherals, voltage regulators, power management circuits, and the like, which are well known in the art, and therefore, will not be described any further.
  • The processor may be responsible for managing the bus and processing, including the execution of software stored on the machine-readable media. Software shall be construed to mean instructions, data, or any combination thereof, whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise.
  • In a hardware implementation, the machine-readable media may be part of the processing system separate from the processor. However, as those skilled in the art will readily appreciate, the machine-readable media, or any portion thereof, may be external to the processing system. By way of example, the machine-readable media may include a transmission line, a carrier wave modulated by data, and/or a computer product separate from the device, all which may be accessed by the processor through the bus interface. Alternatively, or in addition, the machine-readable media, or any portion thereof, may be integrated into the processor, such as the case may be with cache and/or specialized register files. Although the various components discussed may be described as having a specific location, such as a local component, they may also be configured in various ways, such as certain components being configured as part of a distributed computing system.
  • The machine-readable media may comprise a number of software modules. The software modules may include a transmission module and a receiving module. Each software module may reside in a single storage device or be distributed across multiple storage devices. By way of example, a software module may be loaded into RANI from a hard drive when a triggering event occurs. During execution of the software module, the processor may load some of the instructions into cache to increase access speed. One or more cache lines may then be loaded into a special purpose register file for execution by the processor. When referring to the functionality of a software module below, it will be understood that such functionality is implemented by the processor when executing instructions from that software module. Furthermore, it should be appreciated that aspects of the present disclosure result in improvements to the functioning of the processor, computer, machine, or other system implementing such aspects.
  • If implemented in software, the functions may be stored or transmitted over as one or more instructions or code on a computer-readable medium. Computer-readable media include both computer storage media and communication media including any storage medium that facilitates transfer of a computer program from one place to another.
  • Further, it should be appreciated that modules and/or other appropriate means for performing the methods and techniques described herein can be downloaded and/or otherwise obtained by a user terminal and/or base station as applicable. For example, such a device can be coupled to a server to facilitate the transfer of means for performing the methods described herein. Alternatively, various methods described herein can be provided via storage means, such that a user terminal and/or base station can obtain the various methods upon coupling or providing the storage means to the device. Moreover, any other suitable technique for providing the methods and techniques described herein to a device can be utilized.
  • It is to be understood that the claims are not limited to the precise configuration and components illustrated above. Various modifications, changes, and variations may be made in the arrangement, operation, and details of the methods and apparatus described above without departing from the scope of the claims.

Claims (20)

1. An imaging system comprising:
one or more environmental sensors;
a camera including one or more image sensors;
a predictive lighting module in communication with the one or more environmental sensors and the camera, the predictive lighting module configured to:
receive environmental data from the one or more environmental sensors;
adjust the one or more image sensors according to a first in-cabin lighting level;
determine a prospective environmental change based on the environmental data;
predict a second in-cabin lighting level based upon the prospective environmental change;
adjust the one or more image sensors according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
2. The imaging system of claim 1 wherein the one or more image sensors include an infrared sensor.
3. The imaging system of claim 1 wherein the camera includes an occupant-facing camera disposed in a vehicle.
4. The imaging system of claim 1 wherein the one or more environmental sensors include a LiDAR sensor.
5. The imaging system of claim 1 wherein the one or more environmental sensors include a RADAR sensor.
6. The imaging system of claim 1 wherein the one or more environmental sensors include an infrared sensor.
7. The imaging system of claim 1 further comprising a communications module configured to receive the environmental data, the predictive lighting module further receiving the environmental data from the communications module.
8. The imaging system of claim 7 wherein the communications module includes a GPS receiver.
9. An occupant monitoring system of claim 1 wherein the predictive lighting module is further configured to determine a time (T), at which the prospective environmental change will occur, wherein substantially concurrently comprises a time range before and after the time (T).
10. A method of monitoring a vehicle occupant comprising:
receiving environmental data from the one or more environmental sensors;
adjusting one or more image sensors according to a first in-cabin lighting level;
determining a prospective environmental change based on the environmental data;
predicting a second in-cabin lighting level based upon the prospective environmental change; and
adjusting the one or more image sensors according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
11. The method of claim 10 further comprising:
determining a time (T), at which the prospective environmental change will occur, wherein substantially concurrently comprises a time range before and after the time (T).
12. The method of claim 10 further comprising:
obtaining a first sequence of images from the one or more image sensors according to the first in-cabin lighting level; and
obtaining a second sequence of images from the one or more image sensors according to the second in-cabin lighting level.
13. The method of claim 10 wherein the one or more image sensors include an infrared sensor.
14. The method of claim 10 wherein the camera includes an occupant-facing camera disposed in a vehicle.
15. The method of claim 10 wherein the one or more environmental sensors include a LiDAR sensor.
16. The method of claim 10 wherein the one or more environmental sensors include a RADAR sensor.
17. The method of claim 10 wherein the one or more environmental sensors include an infrared sensor.
18. The method of claim 10 wherein the environmental data includes GPS data.
19. An occupant monitoring system for a vehicle comprising:
an occupant facing camera, the occupant facing camera comprising one or more image sensors;
a communications module configured to receive navigational data;
one or more environmental sensors configured to receive environmental data;
a predictive lighting module in communication with the one or more environmental sensors and the occupant-facing camera, the predictive lighting module configured to:
receive environmental data from the one or more environmental sensors and navigation information from the communications module;
adjust the one or more image sensors according to a first in-cabin lighting level;
determine a prospective environmental change based on the environmental data and the navigational data;
predict a second in-cabin lighting level based upon the prospective environmental change; and
adjust the one or more image sensors according to the second in-cabin lighting level substantially concurrently with the prospective environmental change.
20. The occupant monitoring system of claim 19 wherein the predictive lighting module is further configured to determine a time (T), at which the prospective environmental change will occur, wherein substantially concurrently comprises a time range before and after the time (T).
US16/869,293 2020-05-07 2020-05-07 Dynamic lighting and sensor adjustment for occupant monitoring Active US11584295B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/869,293 US11584295B2 (en) 2020-05-07 2020-05-07 Dynamic lighting and sensor adjustment for occupant monitoring

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US16/869,293 US11584295B2 (en) 2020-05-07 2020-05-07 Dynamic lighting and sensor adjustment for occupant monitoring

Publications (2)

Publication Number Publication Date
US20210347298A1 true US20210347298A1 (en) 2021-11-11
US11584295B2 US11584295B2 (en) 2023-02-21

Family

ID=78412176

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/869,293 Active US11584295B2 (en) 2020-05-07 2020-05-07 Dynamic lighting and sensor adjustment for occupant monitoring

Country Status (1)

Country Link
US (1) US11584295B2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210370847A1 (en) * 2018-09-28 2021-12-02 Panasonic I-Pro Sensing Solutions Co., Ltd. Capturing camera
CN114872617A (en) * 2022-05-20 2022-08-09 苏州汉瑞森光电科技股份有限公司 Intelligent regulation and control method and system for lamp illumination light source of vehicle
US11753024B1 (en) * 2022-07-15 2023-09-12 Ghost Autonomy Inc. Anticipatory vehicle headlight actuation

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2003214595A1 (en) 2002-03-07 2003-09-16 Yechezkal Evan Spero Enhanced vision for driving
US7881496B2 (en) * 2004-09-30 2011-02-01 Donnelly Corporation Vision system for vehicle
SE529653C2 (en) 2005-09-06 2007-10-16 Gm Global Tech Operations Inc Systems and procedure for adaptive and self-learning control of vehicle lighting
US9633320B2 (en) * 2011-03-15 2017-04-25 Kabushiki Kaisha Toshiba Energy demand prediction apparatus and method
DE102012017276B4 (en) 2012-08-31 2023-03-30 Volkswagen Aktiengesellschaft Method and device for automatically controlling a lighting unit in the interior of a vehicle
US9409464B2 (en) 2013-04-25 2016-08-09 Globalfoundries Inc. Dynamically managing vehicle glass dimming
US9464887B2 (en) * 2013-11-21 2016-10-11 Ford Global Technologies, Llc Illuminated hitch angle detection component
US10137830B2 (en) * 2014-12-02 2018-11-27 Lenovo (Singapore) Pte. Ltd. Self-adjusting lighting based on viewing location
US9718405B1 (en) * 2015-03-23 2017-08-01 Rosco, Inc. Collision avoidance and/or pedestrian detection system
DE102015219111B4 (en) 2015-10-02 2024-01-25 Robert Bosch Gmbh Driver assistance system and method for adjusting the lighting of control elements in motor vehicles
IL250081B (en) * 2017-01-11 2019-07-31 Livneh Ofer Augmented reality display reflective of visibility affcting features in real-world environment
US10220675B1 (en) * 2017-08-30 2019-03-05 Toyota Motor Engineering & Manufacturing North America, Inc. Vehicle windshield defrosting control based on frozen precipitation accumulation
US10757320B2 (en) * 2017-12-28 2020-08-25 Waymo Llc Multiple operating modes to expand dynamic range
US10499180B1 (en) * 2018-05-17 2019-12-03 Zoox, Inc. Three-dimensional sound for passenger notification
US20200231111A1 (en) * 2019-01-17 2020-07-23 Ford Global Technologies, Llc Vehicle footwell reflector
US10767997B1 (en) * 2019-02-25 2020-09-08 Qualcomm Incorporated Systems and methods for providing immersive extended reality experiences on moving platforms
WO2021012153A1 (en) * 2019-07-22 2021-01-28 Baidu.Com Times Technology (Beijing) Co., Ltd. System for sensor synchronization data analysis in autonomous driving vehicle

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210370847A1 (en) * 2018-09-28 2021-12-02 Panasonic I-Pro Sensing Solutions Co., Ltd. Capturing camera
US11640718B2 (en) * 2018-09-28 2023-05-02 i-PRO Co., Ltd. Capturing camera
CN114872617A (en) * 2022-05-20 2022-08-09 苏州汉瑞森光电科技股份有限公司 Intelligent regulation and control method and system for lamp illumination light source of vehicle
US11753024B1 (en) * 2022-07-15 2023-09-12 Ghost Autonomy Inc. Anticipatory vehicle headlight actuation

Also Published As

Publication number Publication date
US11584295B2 (en) 2023-02-21

Similar Documents

Publication Publication Date Title
US10489222B2 (en) Distributed computing resource management
US11584295B2 (en) Dynamic lighting and sensor adjustment for occupant monitoring
US11891023B1 (en) Using camera data to manage a vehicle parked outside in cold climates
CN109212542B (en) Calibration method for autonomous vehicle operation
US10604156B2 (en) System and method for adjusting a road boundary
CN108454631B (en) Information processing apparatus, information processing method, and recording medium
CN103069153B (en) For controlling the system and method for the motor of vehicle
US10339400B1 (en) Traffic light detection using multiple cameras
US11192430B2 (en) Controlling sunshades in an autonomous vehicle
US20210302165A1 (en) Imu data offset compensation for an autonomous vehicle
US11144053B2 (en) Controlling driving condition components of an autonomous vehicle based on a current driving mode and current conditions
CN111201787B (en) Imaging apparatus, image processing apparatus, and image processing method
US10229505B2 (en) Motion determination system and method thereof
US10635915B1 (en) Method and device for warning blind spot cooperatively based on V2V communication with fault tolerance and fluctuation robustness in extreme situation
US11643082B2 (en) Systems and methods for determining real-time lane level snow accumulation
US20230209206A1 (en) Vehicle camera dynamics
CN116448128A (en) Travel route creation device, travel route creation method, and non-transitory computer-readable medium
JP7015665B2 (en) Information processing equipment, information processing methods and programs
US11851088B2 (en) Method for determining capability boundary and associated risk of a safety redundancy autonomous system in real-time
US20240075962A1 (en) Vehicle controller, method, and computer program for vehicle control
WO2024065173A1 (en) Cloud based scanning for detection of sensors malfunction for autonomous vehicles
US20220348195A1 (en) Driver temporary blindness early warning and avoidance recommendation system
US11836964B2 (en) Devices, systems and methods for scene image acquisition
US20240077619A1 (en) Sensor configuration for autonomous vehicles
EP3726494A1 (en) Method for assisting with the driving of vehicles, associated computer program and system

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: TOYOTA RESEARCH INSTITUTE, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BALACH, THOMAS;REEL/FRAME:053432/0073

Effective date: 20200501

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: TOYOTA JIDOSHA KABUSHIKI KAISHA, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TOYOTA RESEARCH INSTITUTE, INC.;REEL/FRAME:063066/0694

Effective date: 20230316