US20190051185A1 - Road surface image-drawing system for vehicle - Google Patents

Road surface image-drawing system for vehicle Download PDF

Info

Publication number
US20190051185A1
US20190051185A1 US16/161,670 US201816161670A US2019051185A1 US 20190051185 A1 US20190051185 A1 US 20190051185A1 US 201816161670 A US201816161670 A US 201816161670A US 2019051185 A1 US2019051185 A1 US 2019051185A1
Authority
US
United States
Prior art keywords
display
vehicle
image
driver
person
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/161,670
Other versions
US10741083B2 (en
Inventor
Takeshi Masuda
Michihiko Hayakawa
Yuichi Shibata
Takayuki Yagi
Misako Nakazawa
Toshiaki Tsuda
Hidetada Tanaka
Satoshi Yamamura
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Koito Manufacturing Co Ltd
Original Assignee
Koito Manufacturing Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koito Manufacturing Co Ltd filed Critical Koito Manufacturing Co Ltd
Priority to US16/161,670 priority Critical patent/US10741083B2/en
Publication of US20190051185A1 publication Critical patent/US20190051185A1/en
Application granted granted Critical
Publication of US10741083B2 publication Critical patent/US10741083B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/166Anti-collision systems for active traffic, e.g. moving vehicles, pedestrians, bikes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q1/00Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q1/00Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
    • B60Q1/02Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to illuminate the way ahead or to illuminate other areas of way or environments
    • B60Q1/04Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to illuminate the way ahead or to illuminate other areas of way or environments the devices being headlights
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q1/00Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
    • B60Q1/26Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q1/00Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
    • B60Q1/26Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic
    • B60Q1/50Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic for indicating other intentions or conditions, e.g. request for waiting or overtaking
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q1/00Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
    • B60Q1/26Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic
    • B60Q1/50Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic for indicating other intentions or conditions, e.g. request for waiting or overtaking
    • B60Q1/525Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic for indicating other intentions or conditions, e.g. request for waiting or overtaking automatically indicating risk of collision between vehicles in traffic or with pedestrians, e.g. after risk assessment using the vehicle sensor data
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q9/00Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling
    • B60Q9/008Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling for anti-collision purposes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R21/00Arrangements or fittings on vehicles for protecting or preventing injuries to occupants or pedestrians in case of accidents or other traffic risks
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/09Arrangements for giving variable traffic instructions
    • G08G1/0962Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3179Video signal processing therefor
    • H04N9/3185Geometric adjustment, e.g. keystone or convergence
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q2400/00Special features or arrangements of exterior signal lamps for vehicles
    • B60Q2400/50Projected symbol or information, e.g. onto the road or car body
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3626Details of the output of route guidance instructions
    • G01C21/365Guidance using head up displays or projectors, e.g. virtual vehicles or arrows projected on the windscreen or on the road itself

Definitions

  • the present invention relates to a road surface image-drawing system which is mounted on a vehicle and draws information regarding the vehicle or the like on a road surface.
  • a vehicle display system for notifying/warning information of a vehicle to a driver himself or other person such as a pedestrian/oncoming vehicle by illuminating (drawing) a figure or a character or the like on a road surface.
  • a vehicle display system disclosed in Patent Document 1 is configured to notify/warn the approach of an own vehicle to other person trying to enter an intersection by displaying a warning mark of a figure on a road surface at the intersection.
  • Patent Document 1 Japanese Patent Publication No. 2009-184428
  • the viewpoints of the driver of the own vehicle and other person receiving the warning are different from each other. Therefore, the display made by the system disclosed in Patent Document 1 may be distorted to the driver even when it is appropriate for other person receiving the warning. As a result, the driving comfort may be impaired.
  • the present invention aims to provide a suitable display to both a driver of a vehicle providing information such as a warning and other person receiving the information.
  • one aspect of the present invention may provide a road surface image-drawing system mounted on a vehicle.
  • the road surface image-drawing system includes:
  • an image-drawing device configured to draw a first display to be visible to a driver of the vehicle at a first position and draw a second display to be visible to a person other than the driver at a second position different from the first position
  • control device configured to correct a shape of the first display according to a positional relationship between the first position and the driver and correct a shape of the second display according to a positional relationship between the second position and the person other than the driver.
  • the first position and the second position are on a road.
  • the first position may be on the road located between an intersection and the vehicle.
  • the second position may be on the intersection.
  • the first position may be on a windshield of the vehicle.
  • the first display can be drawn so as to have a distortion-free shape, as seen from the driver of the vehicle on which the system is mounted.
  • the second display can be drawn so as to have a distortion-free shape, as seen from the person other than the driver. Accordingly, it is possible to provide a display including information such as a warning in a suitable manner to both the driver of the vehicle who provides the information and the person who receives the information. Particularly, the driver who views the first display can continue driving without diverting his attention to the second display.
  • the information included in the first display and the information included in the second display are related to each other.
  • the first display may include information or warning notifying the approach of other person to the driver.
  • the second display may include information or warning notifying the approach of the vehicle to the other person. That is, the first display and the second display which are related to each other for the purpose of avoiding collision between the vehicle and the other person may be presented in a manner corresponding to the situation of each of the driver of the vehicle and the other person. Therefore, it is possible to properly call attention of both the driver and the other person.
  • FIG. 1 is a front view showing a headlamp device including a road surface image-drawing system according to one embodiment.
  • FIG. 2 is a longitudinal sectional view showing the headlamp device.
  • FIG. 3A is a view showing a configuration of an image-drawing device in the road surface image-drawing system.
  • FIG. 3B is a view showing a configuration of the image-drawing device in the road surface image-drawing system.
  • FIG. 4 is a functional block diagram showing a configuration of a control device in the road surface image-drawing system.
  • FIG. 5 is a view for explaining an image-drawing operation by the road surface image-drawing system.
  • FIG. 6 is a flow chart showing a control process executed by the road surface image-drawing system.
  • FIG. 7 is a flow chart showing a control process executed by the road surface image-drawing system.
  • FIG. 8A is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 8B is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 8C is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 8D is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 9 is a view showing another example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 10 is a view for explaining a modified example of the road surface image-drawing system.
  • FIG. 1 is a front view of a headlamp device 1 according to one embodiment.
  • FIG. 2 shows a longitudinal section of the headlamp device 1 , when seeing a line II-II in FIG. 1 from an arrow direction.
  • the headlamp device 1 is a right headlamp device mounted on the right front part of a vehicle.
  • a left headlamp device mounted on the left front part of the vehicle has a configuration bilaterally-symmetric to the headlamp device 1 shown in FIG. 1 .
  • the headlamp device 1 includes a housing 2 and a translucent cover 3 .
  • the housing 2 has an opening portion.
  • the translucent cover 3 is formed of a light-transmitting resin or glass or the like.
  • the translucent cover 3 is attached to the opening portion of the housing 2 and defines a lamp chamber.
  • the headlamp device 1 further includes a high-beam light source unit 5 and a low-beam light source unit 6 .
  • the high-beam light source unit 5 and the low-beam light source unit 6 are accommodated in the lamp chamber.
  • the high-beam light source unit 5 in the present example is a reflection type and is configured to illuminate the front of the vehicle with a high-beam light distribution pattern.
  • the low-beam light source unit 6 in the present example is a projector type and is configured to illuminate the front of the vehicle with a low-beam light distribution pattern.
  • the types of the high-beam light source unit 5 and the low-beam light source unit 6 can be appropriately determined, so long as a desired illumination can be obtained.
  • the headlamp device 1 includes a bracket 4 and aiming screws 14 .
  • the bracket 4 is accommodated in the lamp chamber.
  • the high-beam light source unit 5 and the low-beam light source unit 6 are supported by the bracket 4 .
  • the bracket 4 is fixed to the housing 2 via the aiming screws 14 .
  • the aiming screws 14 are provided at three corner portions of the bracket 4 . By operating the aiming screws 14 to rotate, the direction of an optical axis of the headlamp device 1 can be adjusted in an up-down direction and a left-right direction.
  • the headlamp device 1 includes an extension 12 .
  • the extension 12 covers each component disposed in the lamp chamber so that these components cannot be visible from the outside while allowing the passage of the light emitted from the high-beam light source unit 5 and the low-beam light source unit 6 .
  • the headlamp device 1 includes a road surface image-drawing system 7 .
  • the road surface image-drawing system 7 is accommodated in the lamp chamber.
  • the road surface image-drawing system 7 includes an image-drawing device 8 and a control device 9 .
  • the image-drawing device 8 includes a laser light source unit 10 and a scanning mechanism 11 .
  • the control device 9 is fixed to the housing 2 .
  • FIG. 3A schematically shows a configuration of the laser light source unit 10 .
  • the laser light source unit 10 includes a first light source 15 , a second light source 16 , a third light source 17 and a support stand 18 .
  • the first light source 15 , the second light source 16 and the third light source 17 are supported on the support stand 18 .
  • the first light source 15 is configured to emit a red laser light.
  • the second light source 16 is configured to emit a green laser light.
  • the third light source 17 is configured to emit a blue laser light.
  • each of the first light source 15 , the second light source 16 and the third light source 17 is a laser diode.
  • the laser light source unit 10 includes a first condensing lens 20 , a second condensing lens 21 , a third condensing lens 22 , a first dichroic mirror 25 , a second dichroic mirror 26 and a third dichroic mirror 27 .
  • the red laser light emitted from the first light source 15 is condensed by the first condensing lens 20 and is incident on the first dichroic mirror 25 .
  • the green laser light emitted from the second light source 16 is condensed by the second condensing lens 21 and is incident on the second dichroic mirror 26 .
  • the blue laser light emitted from the third light source 17 is condensed by the third condensing lens 22 and is incident on the third dichroic mirror 27 .
  • the laser light source unit 10 includes a condensing lens 19 .
  • the light reflected by the first dichroic mirror 25 , the second dichroic mirror 26 and the third dichroic mirror 27 passes through the condensing lens 19 as an output light B 2 .
  • the output light B 2 can be a monochromatic light, a mixed color light or a white light.
  • the laser light source unit 10 includes a monitor 28 .
  • the monitor 28 is configured to be able to monitor the intensity of the red laser light, the green laser light, the blue laser light and the output light B 2 . Further, the monitor 28 is configured to be able to control the intensity of the red laser light emitted from the first light source 15 , the intensity of the green laser light emitted from the second light source 16 and the intensity of the blue laser light emitted from the third light source 17 .
  • FIG. 3B shows a configuration of the scanning mechanism 11 .
  • the scanning mechanism 11 is configured as an MEMS (Micro Electro Mechanical Systems) mirror.
  • the scanning mechanism 11 includes a base part 37 , a first pivoting body 38 , a second pivoting body 39 , a pair of first torsion bars 40 and a pair of second torsion bars 41 .
  • the base part 37 has a first opening portion.
  • the first pivoting body 38 is disposed in the opening portion and is supported on the base part 37 via the pair of first torsion bars 40 .
  • the first pivoting body 38 is pivotable in a left-right (horizontal) direction.
  • the first pivoting body 38 has a second opening portion.
  • the second pivoting body 39 is disposed in the second opening portion and is supported on the first pivoting body 38 via the pair of second torsion bars 41 .
  • the second pivoting body 39 is pivotable in an up-down (vertical) direction.
  • the second pivoting body 39 has a reflective surface 36 .
  • the scanning mechanism 11 includes a pair of first permanent magnets 42 , a pair of second permanent magnets 43 and a terminal portion 44 .
  • the pair of first torsion bars 40 is disposed between the pair of first permanent magnets 42 .
  • the direction of a straight line connecting the pair of first permanent magnets 42 is orthogonal to the extending direction of the pair of first torsion bars 40 .
  • the pair of second torsion bars 41 is disposed between the pair of second permanent magnets 43 .
  • the direction of a straight line connecting the pair of second permanent magnets 43 is orthogonal to the extending direction of the pair of second torsion bars 41 .
  • the first pivoting body 38 includes a first coil (not shown).
  • the second pivoting body 39 includes a second coil (not shown). The first coil and the second coil are electrically connected to the control device 9 via the terminal portion 44 .
  • the laser light source unit 10 is fixed to the bracket 4 so that the output light B 2 is emitted downward.
  • the scanning mechanism 11 is disposed so that the output light B 2 is reflected to the front of the headlamp device 1 by the reflective surface 36 .
  • the scanning mechanism 11 includes a scanning actuator 58 .
  • the scanning actuator 58 is composed of the pair of first permanent magnets 42 , the pair of second permanent magnets 43 , the first coil and the second coil.
  • the pivoting direction and amount of the first pivoting body 38 and the second pivoting body 39 are changed. In this way, the direction of the reflective surface 36 is changed in the up-down direction and the left-right direction.
  • the control device 9 is communicatively connected to a vehicle speed sensor 60 , a steering angle sensor 61 , an acceleration sensor 62 , a direction indicator detection sensor 63 , a brake sensor 64 , a GPS 65 , a navigation system 66 , an on-vehicle camera 67 , a preceding vehicle detection sensor 68 , a human body detection sensor 69 , an illuminance sensor 70 , an external system 71 , and an image processing device 72 .
  • the vehicle speed sensor 60 is configured to detect a travelling distance and a vehicle speed by the rotation of vehicle wheels.
  • the steering angle sensor 61 is configured to detect a steering angle of a steering wheel.
  • the acceleration sensor 62 is configured to detect an operating amount of an accelerator.
  • the direction indicator detection sensor 63 is configured to detect a switch operation of a direction indicator.
  • the brake system 64 is configured to detect an operating amount of a brake.
  • the navigation system 66 is configured to detect a current position of the own vehicle and perform a route guidance to the destination based on a data received from the GPS 65 and a map data.
  • the on-vehicle camera 67 is configured to capture and record an image in front of the own vehicle.
  • the preceding vehicle detection sensor 68 is configured to detect a distance to a preceding vehicle and a relative speed by using millimeter waves or the like.
  • the human body detection sensor 69 is configured to detect a human body located in front of the own vehicle by using infrared rays or the like.
  • the illuminance sensor 70 is configured to detect an illuminance of the surrounding environment of the own vehicle by using the photoelectric effect.
  • an intersection camera or a monitoring camera installed at intersection signals can be exemplified.
  • the image processing device 72 is configured to perform an image processing on an image data acquired by the on-vehicle camera 67 and the external system 71 and thus to recognize a specific object.
  • the control device 9 includes a lamp ECU 51 , a ROM 52 and a RAM 53 .
  • the lamp ECU 51 executes various programs stored in advance in the ROM 52 on the RAM 53 , thereby generating various control signals.
  • the lamp ECU 51 includes an own vehicle data acquisition unit 83 .
  • the own vehicle data acquisition unit 83 is configured to receive data from at least one of the vehicle speed sensor 60 , the steering angle sensor 61 , the acceleration sensor 62 , the direction indicator detection sensor 63 , the brake sensor 64 , the GPS 65 , the navigation system 66 and the illuminance sensor 70 and to acquire information on the travelling conditions including the current position of the own vehicle and the surrounding environments.
  • the lamp ECU 51 includes a travelling environment determination unit 84 .
  • the travelling environment determination unit 84 is configured to determine whether image-drawing by the image-drawing device 8 is possible based on the information on the travelling conditions and surrounding environments of the own vehicle acquired by the own vehicle data acquisition unit 83 . When it is determined that the image-drawing is possible, the travelling environment determination unit 84 is configured to permit the image-drawing by the image-drawing device 8 . When it is determined that the image-drawing is not possible, the travelling environment determination unit 84 is configured to prohibit the image-drawing by the image-drawing device 8 .
  • the travelling environment determination unit 84 determines that the image-drawing is unnecessary, thereby prohibiting the image-drawing.
  • the travelling environment determination unit 84 permits the image-drawing.
  • the travelling environment determination unit 84 determines that effective image-drawing cannot be performed, thereby prohibiting the image-drawing.
  • the travelling environment determination unit 84 prohibits the image-drawing.
  • the travelling environment determination unit 84 permits the image-drawing.
  • the travelling environment determination unit 84 prohibits the image-drawing.
  • the travelling environment determination unit 84 may determine that the image-drawing is unnecessary in order to make driving easier.
  • the lamp ECU 51 includes an other vehicle data acquisition unit 85 .
  • the other vehicle data acquisition unit 85 is configured to acquire data from at least one of the on-vehicle camera 67 , the preceding vehicle detection sensor 68 and the external system 71 and to acquire information on a distance to the other vehicle and a relative speed.
  • the lamp ECU 51 includes an other vehicle determination unit 86 .
  • the other vehicle determination unit 86 is configured to determine whether the image-drawing by the image-drawing device 8 is possible based on the information acquired by the other vehicle data acquisition unit 85 and to determine whether the image-drawing by the image-drawing device 8 is possible. When it is determined that the image-drawing is possible, the other vehicle determination unit 86 is configured to permit the image-drawing by the image-drawing device 8 . When it is determined that the image-drawing is not possible, the other vehicle determination unit 86 is configured to prohibit the image-drawing by the image-drawing device 8 .
  • the other vehicle determination unit 86 permits the image-drawing.
  • the other vehicle determination unit 86 permits the image-drawing.
  • the other vehicle determination unit 86 permits the image-drawing.
  • the other vehicle determination unit 86 prohibits the image-drawing.
  • the other vehicle determination unit 86 permits the image-drawing.
  • the lamp ECU 51 includes a pedestrian data acquisition unit 87 .
  • the pedestrian data acquisition unit 87 is configured to acquire data from at least one of the on-vehicle camera 67 , the human body detection sensor 69 and the external system 71 and to acquire information on the presence of a pedestrian (including a bicycle) and the state (traveling direction, etc.) thereof.
  • the lamp ECU 51 includes a pedestrian determination unit 88 .
  • the pedestrian determination unit 88 is configured to determine the presence/absence of a pedestrian based on the information acquired by the pedestrian data acquisition unit 87 .
  • the lamp ECU 51 includes an intersection data acquisition unit 89 .
  • the intersection data acquisition unit 89 is configured to acquire data from at least one of the on-vehicle camera 67 and the navigation system 66 and thus to acquire information on the presence of an intersection and the situations (the presence/absence of a signal the presence/absence of a road illumination, the presence of obstacles around the intersection, etc.) of the intersection.
  • the lamp ECU 51 includes an intersection determination unit 90 .
  • the intersection determination unit 90 is configured to determine the presence/absence of an intersection based on the information acquired by the intersection data acquisition unit 89 .
  • the intersection determination unit 90 can be also configured to determine whether it is an intersection in which the necessity of performing the image-drawing is high, based on the information on the situations of the intersection acquired by the intersection data acquisition unit 89 .
  • the intersection determination unit 90 permits the image-drawing in the normal mode.
  • the lamp ECU 51 includes a pedestrian risk determination unit 91 .
  • the pedestrian risk determination unit 91 is configured to acquire a distance (hereinafter, referred to as a “pedestrian distance”) to a pedestrian based on the position, travelling direction and walking speed or the like of a pedestrian. Further, the pedestrian risk determination unit 91 is configured to acquire a braking distance of the own vehicle based on the position, travelling direction and driving speed or the like of the own vehicle. When the pedestrian distance is equal to or greater than the braking distance, the pedestrian risk determination unit 91 permits the image-drawing in the normal mode. On the other hand, when the pedestrian distance is less than the braking distance, there is a danger of collision, and thus, the pedestrian risk determination unit 91 permits the image-drawing in the emergency mode.
  • a distance hereinafter, referred to as a “pedestrian distance”
  • the pedestrian risk determination unit 91 is configured to acquire a braking distance of the own vehicle based on the position, travelling direction and driving speed or the like of the own vehicle.
  • the lamp ECU 51 includes a display content determination unit 92 .
  • the display content determination unit 92 is configured to determine display contents when the image-drawing is permitted by the travelling environment determination unit 84 , the other vehicle determination unit 86 , the pedestrian determination unit 88 or the intersection determination unit 90 ). Specifically, the display content determination unit 92 is configured to determine a figure or a character to be drawn together with its display form in accordance with the determined situations (including a risk). Further, the display content determination unit 92 is configured to control the image-drawing device 8 , thereby image-drawing the display contents.
  • the lamp ECU 51 includes a laser light source control unit 81 .
  • the laser light source control unit 81 is configured to control the turn on/off and emission light intensity (i.e., color and brightness of output light B 2 ) of each of the first light source 15 , the second light source 16 and the third light source 17 , based on the display contents determined by the display content determination unit 92 .
  • the lamp ECU 51 includes an actuator control unit 82 .
  • the actuator control unit 82 is configured to control an operation of the scanning actuator 58 based on the display contents determined by the display content determination unit 92 .
  • the output light B 2 emitted from the laser light source unit 10 is reflected in a desired direction by the scanning mechanism 11 .
  • the image-drawing device 8 causes desired display contents to be drawn on a road surface located in front of the headlamp device 1 by using the output light B 2 .
  • FIG. 5 shows a virtual vertical screen positioned in front of the headlamp device 1 .
  • a reference numeral “CR” indicates a travelling lane of the own vehicle.
  • the scanning mechanism 11 of the image-drawing device 8 is configured so that the output light B 2 can scan the inside of a rectangular scanning area SA shown in FIG. 5 .
  • the scanning area SA is preferably set such that it can cover a road shoulder of the travelling lane CR, a lane adjacent to the travelling lane CR and the range of 5 m to 50 m in front of the own vehicle.
  • the scanning area SA is set to the left-right direction range of ⁇ 1° to ⁇ 8° on a horizontal line H-H shown in FIG. 5 and the up-down direction range of 20° to ⁇ 20° on a vertical line V-V shown in FIG. 5 .
  • the control device 9 controls an operation of the image-drawing device 8 to cause a display MM for a driver and a display MO for other person to be drawn at different places in the scanning area SA, respectively.
  • the display MM (an example of the first display) for the driver is a display to be visible to the driver of the own vehicle.
  • the display MO (an example of the second display) for the other person is a display to be visible to a person other than the driver of the own vehicle, such as a pedestrian or a driver of a preceding vehicle.
  • the display MM for the driver and the display MO for the other person are based on the display contents determined by the display content determination unit 92 of the control device 9 .
  • the display MM for the driver and the display MO for the other person may be drawn in a partially overlapping manner, as long as each of them includes a different point in the scanning area SA.
  • the scanning mechanism 11 of the image-drawing device 8 changes the orientation of the reflective surface 36 based on a control signal from the actuator control unit 82 of the control device 9 .
  • the scanning point can be horizontally reciprocated in the scanning area SA while shifting its position in the vertical direction by dl.
  • a reference numeral “SI” in FIG. 5 indicates a movement trajectory of the scanning point.
  • the scanning point reaches the image-drawing positions of the display MM for the driver and the display MO for the other person
  • at least one of the first light source 15 , the second light source 16 and the third light source 17 is turned on (pan indicated by a solid line, out of the movement trajectory SI of the scanning point shown in FIG. 5 ) on the basis of the control signal from the laser light source control unit 81 of the control device 9 .
  • the display MM for the driver and the display MO for the other person are drawn at desired positions by using the output light B 2 .
  • the road surface image-drawing system 7 repeatedly performs the scanning of one cycle shown in FIG. 5 and appropriately determines the drawing positions of the display MM for the driver and the display MO for the other person in each cycle. In this way, the display MM for the driver and the display MO for the other person with an arbitrary shape can be continuously or intermittently displayed in the scanning area SA.
  • the road surface image-drawing system 7 appropriately changes the drawing positions of the display MM for the driver and the display MO for the other person in each cycle. In this way, the display form of the display MM for the driver and the display MO for the other person can be dynamically changed.
  • Step S 10 information on the travelling conditions and surrounding environments of the own vehicle is acquired by the own vehicle data acquisition unit 83 . Then, when it is determined by the travelling environment determination unit 84 based on the information that the image-drawing is possible (Step S 11 : YES), the process proceeds to Step S 12 . When it is determined by the travelling environment determination unit 84 that the image-drawing is not possible (Step S 11 : NO), the image-drawing is not performed.
  • Step S 12 information on a distance to a preceding vehicle and a relative speed is acquired by the other vehicle data acquisition unit 85 .
  • Step S 13 information on a distance to a preceding vehicle and a relative speed is acquired by the other vehicle data acquisition unit 85 .
  • Step S 13 information on a distance to a preceding vehicle and a relative speed is acquired by the other vehicle data acquisition unit 85 .
  • Step S 14 information on a pedestrian is acquired by the pedestrian data acquisition unit 87 (Step S 14 ). Then, when it is determined by the pedestrian determination unit 88 based on the information that a pedestrian is present (Step S 15 : YES), the process proceeds to Step S 17 . When it is determined by the pedestrian determination unit 88 that a pedestrian is not present (Step S 15 : NO), the process proceeds to Step S 16 .
  • Step S 16 information on an intersection is acquired by the intersection data acquisition unit 89 .
  • Step S 16 determines whether an intersection is present.
  • Step S 18 determines whether an intersection is not present.
  • Step S 17 a pedestrian distance and a braking distance are acquired by the pedestrian risk determination unit 91 .
  • the pedestrian distance is equal to or greater than the braking distance
  • an instruction to perform image-drawing in the normal mode is issued and the process proceeds to Step S 18 .
  • an instruction to perform image-drawing in the emergency mode is issued and the process proceeds to Step S 18 .
  • Step S 18 is performed when image-drawing permission is made in previous Step S 11 and Step S 13 and image-drawing permission is made in Step S 15 or S 16 .
  • Step S 18 “display contents” are determined by the display content determination unit 92 .
  • FIG. 7 shows details of the process performed by the display content determination unit 92 in Step S 18 .
  • Step S 21 a figure or character or the like to be drawn as the display MO for the other person and an image-drawing area thereof and a figure or character or the like to be drawn as the display MM for the driver and an image-drawing area thereof are determined in accordance with the situations corresponding to the data acquired in previous steps (Step S 10 to Step S 16 ).
  • the display content determination unit 92 performs a trapezoidal distortion correction on the figure or character or the like to be drawn as the display MO for the other person so that a distortion-free shape is obtained as seen from the detected pedestrian.
  • the display content determination unit 92 performs a trapezoidal distortion correction on the figure or character or the like to be drawn as the display MO for the other person so that a distortion-free shape is obtained as seen from the driver of the preceding vehicle detected.
  • the display content determination unit 92 corrects the shape of the display MO for the other person based on a positional relationship between the detected other person (person other than the driver of the own vehicle) and the image-drawing area of the display MO for the other person.
  • the display content determination unit 92 performs a trapezoidal distortion correction on the figure or character or the like to be drawn as the display MM for the driver so that a distortion-free shape is obtained as seen from the driver of the own vehicle.
  • the trapezoidal distortion correction is performed on the basis of an installation height of the headlamp device 1 or a distance to the image-drawing area or the like.
  • the display content determination unit 92 corrects the shape of the display MM for the driver based on a positional relationship between the driver of the own vehicle and the drawing area of the display MM for the driver.
  • the display content determination unit 92 sequentially determines whether an instruction to perform the image-drawing in the normal mode in previous steps (Step S 16 to Step S 17 ) is issued (Step S 22 ) and whether an instruction to perform the image-drawing in the emergency mode is issued (Step S 24 ).
  • Step S 22 and Step S 23 NO
  • image-drawing is not performed and the process is repeated from the Step S 10 again.
  • Step S 22 When the instruction to perform the image-drawing in the normal mode is issued (Step S 22 : YES), the image-drawing is performed such that the figure or character or the like determined in the image-drawing area determined in Step S 21 is constantly displayed or is displayed with slow blinking (Step S 23 ).
  • Step S 24 When the instruction to perform the image-drawing in the emergency mode is issued (Step S 24 : YES), the image-drawing is performed such that the figure or character or the like determined in the image-drawing area determined in Step S 21 is displayed with fast blinking.
  • the blinking frequency of the display in the emergency mode image-drawing is faster than that of the display in the normal mode image-drawing.
  • the blinking frequency of the display in the emergency mode image-drawing is, for example, 1 Hz.
  • Step S 26 it is determined whether the instruction to perform the image-drawing in the emergency mode is continued for a predetermined time or more.
  • the blinking frequency is changed so as to rise (Step S 27 ). For example, when the initial fast blinking (1 Hz) is continued for the predetermined time or more, the blinking frequency is increased by 1 Hz. For example, the image-drawing is performed such that this processing is repeated until the blinking frequency becomes 4 Hz. This is repeatedly performed, for example, up to 4 Hz.
  • the instruction to perform the image-drawing in the emergency mode is not continued for the predetermined time or more (Step S 26 : NO)
  • the image-drawing in the initial fast blinking state is continued.
  • FIGS. 8A to 8D show an example where the road surface image-drawing is performed at an intersection by the road surface image-drawing system 7 .
  • FIGS. 8A to 8D schematically show a state of an intersection 103 seen from above.
  • the travelling environment determination unit 84 determines that the travelling conditions and the surrounding environments are in the state of capable of performing the image-drawing (Step S 11 : YES).
  • the other vehicle data acquisition unit 85 determines that an image-drawable distance (e.g., 50 m) is secured although a preceding vehicle 102 is present (Step S 13 : YES).
  • the pedestrian data acquisition unit 87 determines that a pedestrian is not present (Step S 15 : NO).
  • the intersection determination unit 90 determines that an intersection is present (e.g., at 30 m) in front of the own vehicle C (Step S 16 : YES). Therefore, an instruction to perform the image-drawing in the normal mode is issued.
  • Step S 21 display contents are determined. Specifically, it is determined that a “rectilinear arrow” for informing the other person that the own vehicle C goes straight through the intersection 103 is drawn at the intersection 103 (an example of the second position) as the display for the other person. Further, it is determined that a “rectilinear arrow” for informing the driver of the own vehicle C of the same information as the display for the other person is drawn on a road 104 (an example of the first position) as the display for the driver. The road 104 is positioned between the own vehicle C and the intersection 103 .
  • a straight arrow MO 1 as the display for the other person is drawn in a left end region of the intersection 103 .
  • the shape of the straight arrow MO 1 is corrected so that it can be seen without distortion, as viewed from a driver of other vehicle or a pedestrian who is likely to enter the intersection 103 from the left side.
  • a straight arrow MO 2 as the display for the other person is drawn in a right end region of the intersection 103 .
  • the shape of the straight arrow MO 2 is corrected so that it can be seen without distortion, as viewed from a driver of other vehicle or a pedestrian who is likely to enter the intersection 103 from the right side.
  • a straight arrow MO 3 as the display for the other person is drawn in a front end region of the intersection 103 .
  • the shape of the straight arrow MO 3 is corrected so that it can be seen without distortion, as viewed from a driver of other vehicle or a pedestrian who is likely to enter the intersection 103 from the front side.
  • a straight arrow MM 1 as the display for the driver is drawn on the road 104 .
  • the shape of the straight arrow MM 1 is corrected so that it can be seen without distortion, as viewed from the driver of the own vehicle C.
  • the straight arrows MO 1 , MO 2 , MO 3 as the displays for the other person are presented without distortion to the other person approaching the intersection 103 . Therefore, the other person can accurately recognize the fact that the own vehicle C is approaching the intersection 103 .
  • the display for the driver related to the display for the other person is presented without distortion to the driver of the own vehicle C. Therefore, the driver of the own vehicle C can continue driving without diverting his attention to the display for the other person.
  • the straight arrows MO 1 , MO 2 , MO 3 as the displays for the other person are preferably displayed to be large in the full width of the intersection 103 in order to enhance the effect of calling attention.
  • the straight arrow MM 1 as the display for the driver is preferably spot-displayed on the tip of line of sight of the driver of the own vehicle C.
  • the line of sight of the driver is detected by, for example, an eye camera connected to the lamp ECU 51 .
  • the timing at which the straight arrows MO 1 , MO 2 , MO 3 as the displays for the other person and the straight arrow MM 1 as the display for the driver are drawn can be appropriately determined.
  • These displays may be sequentially drawn one by one, or, at least two of these displays may be simultaneously drawn when the image-drawing areas thereof are not overlapped.
  • the image-drawing time of each display and the number of scanning cycles described with reference to FIG. 5 can be appropriately determined.
  • the display to be drawn per one scanning cycle may be switched or a specific display may be drawn over a plurality of scanning cycles.
  • the order in which each display is drawn can be appropriately determined.
  • FIG. 9 shows another example in which the road surface image-drawing is performed at the intersection by the road surface image-drawing system 7 .
  • FIG. 9 schematically shows a state of the intersection 103 seen from above.
  • the pedestrian determination unit 88 determines that a pedestrian 101 tying to enter the intersection 103 is present (Step S 15 : YES).
  • the pedestrian risk determination unit 91 determines that the distance (e.g., 20 m) between the own vehicle C and the pedestrian 101 is equal to or greater than the braking distance (e.g., 13 m) of the own vehicle C (Step S 17 ). Therefore, an instruction to perform the image-drawing in the normal mode is issued.
  • Step S 21 display contents are determined. Specifically, it is determined that the character of “STOP” for urging not to enter the intersection 103 is drawn at the intersection 103 as the display for the other person. Further, it is determined that a mark for informing the driver of the own vehicle C of the presence of the pedestrian 101 trying to enter the intersection 103 is drawn on the road 104 as the display for the driver.
  • a character MO 4 (an example of the second display) of “STOP” as the display for the other person is drawn at the intersection 103 (an example of the second position).
  • the shape of the character M 04 is corrected so that it can be seen without distortion, as viewed from the pedestrian 101 trying to enter the intersection 103 from the left side.
  • a mark MM 2 (an example of the first display) indicating the presence of the pedestrian 101 , the pedestrian distance (20 m) and the approaching direction to the intersection 103 is drawn on the road 104 (an example of the first position) as the display for the driver.
  • the shape of the mark MM 2 is corrected so that it can be seen without distortion, as viewed from the driver of the own vehicle C.
  • the character MO 4 as the display for the other person is presented without distortion to the pedestrian 101 approaching the intersection 103 . Therefore, the pedestrian 101 can accurately recognize the fact that the own vehicle C is approaching the intersection 103 .
  • the mark MM 2 related to the display for the other person is presented without distortion to the driver of the own vehicle C. Therefore, the driver of the own vehicle C can continue driving without diverting his attention to the display for the other person and can accurately recognize the fact that the pedestrian 101 is approaching the intersection 103 .
  • the character MO 4 as the display for the other person is intended to inform the pedestrian 101 of the approach of the own vehicle C to the intersection 103 and the mark MM 2 as the display for the driver is intended to inform the driver of the own vehicle C of the approach of the pedestrian 101 to the intersection 103 . That is, information for avoiding collision at the intersection 103 is presented in a manner corresponding to the situation of each of the driver of the own vehicle C and the pedestrian 101 . Therefore, it is possible to properly call attention of both the driver of the own vehicle C and the pedestrian 101 .
  • the pedestrian risk determination unit 91 determines that the distance (e.g., 10 m) to the pedestrian 101 is less than the braking distance (e.g., 13 m) of the own vehicle C (Step S 17 ). Therefore, an instruction to perform the image-drawing in the emergency mode is issued.
  • Step S 21 display contents are determined. Specifically, it is determined that the character MO 4 and the mark MM 2 are blinked at a high speed. Display contents with higher warning than the character M 04 and the mark MM 2 may be selected.
  • Step S 24 and S 25 Based on the determined display contents, the image-drawing in the emergency mode is performed (Steps S 24 and S 25 ).
  • the instruction to perform the image-drawing in the emergency mode is continued for a predetermined time (e.g., one second) (Step S 26 : YES)
  • the blinking frequency is increased (Step S 27 ).
  • the forms of the displays MO 1 , MO 2 , MO 3 , MO 4 for the other person are not limited to the examples described for the above embodiment. Any suitable form can be adopted according to the situations, so long as it can notify the other person of the approach or behavior of the own vehicle. For example, more detailed information such as an operation of a brake or a steering wheel by the driver of the own vehicle, acceleration of the own vehicle and driver's attribute information (sex or whether he is a child or an elderly person) may be included in the display for the other person.
  • the own vehicle data acquisition unit 83 determines that the driver of the own vehicle recognizes the other person such as a pedestrian based on the brake operation, travelling speed and information from the eye camera or the like, the content notifying the fact may be included in the display for the other person.
  • the forms of the displays MM 1 , MM 2 for the driver are not limited to the examples described for the above embodiment. Any suitable form can be adopted according to the situations, so long as it can notify the driver of the own vehicle of the presence or behavior of the other person. For example, more detailed information such as the number and attribute (sex or whether he is a child or an elderly person) of the pedestrian acquired by the pedestrian data acquisition unit 87 may be included in the display for the driver. Alternatively, when the pedestrian data acquisition unit 87 determines that the pedestrian recognizes the own vehicle, the content notifying the fact may be included in the display for the driver.
  • the display for the other person and the display for the driver are based on mutually related information. In order to emphasize this relevance, a line connecting the display for the other person and the display for the driver may be drawn.
  • a region that looks relatively dark may be drawn so as to border the display for the other person and the display for the driver.
  • the region may be drawn with a color which has brightness or saturation lower than the colors of the display for the other person and the display for the driver or a color which makes the colors of the display for the other person and the display for the driver stand out. In this way, the display for the other person and the display for the driver appear to emerge, and thus, visibility is improved.
  • the necessity of the image-drawing operation in the emergency mode is determined from a relationship between the distance from the own vehicle to the pedestrian and the braking distance of the own vehicle.
  • suitable conditions such as when the own vehicle accelerates or suddenly decelerates, when the pedestrian jumps ahead of the own vehicle and when the distance to the intersection is less than a predetermined value may be adopted.
  • the blinking of the display for the other person and the display for the driver do not need to be synchronized.
  • the blinking frequency of the display for the other person may be higher than that of the display for the driver.
  • the brightness of the display for the other person and the display for the driver does not need to be the same.
  • the display for the other person may be brighter than the display for the driver.
  • the change of the display mode is not limited to the change of blinking frequency of the display. As the emergency increases, the brightness of the display may be increased or the color sense of the display may be changed.
  • the display for the other person and the display for the driver described with reference to FIGS. 8A to 9 do not need to be continued throughout the execution period of the image-drawing operation by the image-drawing device 8 . It is also possible to adopt a configuration in which only a spotlight-like display is drawn on the road 104 when the distance to the intersection 103 is equal to or greater than a predetermined value and a display of the form illustrated in FIGS. 8A to 9 is drawn when the distance to the intersection 103 is less than the predetermined value.
  • the display for the other person and the display for the driver are drawn on the road in the travelling direction of the own vehicle C.
  • the own vehicle data acquisition unit 83 detects, through the steering angle sensor 61 or the direction indicator detection sensor 6 , that the own vehicle C performs its course change, at least one of the display for the other person and the display for the driver can be drawn on the road after the course change.
  • the display for the other person and the display for the driver may be individually stopped at the discretion of the driver.
  • a switch for releasing the image-drawing operation can be provided in the own vehicle.
  • the image-drawing device 8 can include a head up display (HUD) device 73 .
  • the HUD device 73 can be connected to the lamp ECU 51 .
  • the HUD device 73 is configured to project the display MM (an example of the first display) for the driver on the windshield W (an example of the first position) of the own vehicle C.
  • an operation of projecting a predetermined display on an object is also considered as an example of the “image-drawing”.
  • the driver of the own vehicle C can visually recognize the display MM for the driver in a state of being overlaid on the front landscape seen through the windshield W. Therefore, the driver of the own vehicle C can visually recognize the display MM for the driver in a reliable manner, regardless of the state of the road surface.
  • the road surface image-drawing system 7 may be provided in any one of the left and right headlamp devices.
  • the road surface image-drawing systems 7 on the left and right sides can perform different image-drawing operations.
  • the character M 04 may be drawn by the road surface image-drawing system 7 provided in the left headlamp device and the mark MM 2 may be drawn by the road surface image-drawing system 7 provided in the right headlamp device.
  • the character M 04 may be drawn by the road surface image-drawing system 7 provided in the right headlamp device and the mark MM 2 may be drawn by the road surface image-drawing system 7 provided in the left headlamp device.
  • the whole of the road surface image-drawing system 7 may be disposed outside the headlamp device 1 .
  • the road surface image-drawing system 7 can be disposed on a roof of the own vehicle such that an emission direction of the output light B 2 is directed forward.
  • the configuration of the laser light source unit 10 can be appropriately changed in accordance with the specifications of the headlamp device 1 .
  • a single light source for emitting a white light can be used.
  • a configuration that a white light is emitted through excitation by making a blue laser light incident on a yellow phosphor may be adopted.
  • a fourth light source for emitting an orange light may be added.
  • the first light source 15 , the second light source 16 and the third light source 17 may be laser light sources other than laser diodes.
  • a scanning optical system such as a galvanometer mirror or a DMMD (Digital Micro Mirror Device) can be adopted as the scanning mechanism 11 .
  • the scanning performed by the scanning mechanism 11 is not limited to the example shown in FIG. 5 .
  • a configuration that the scanning point is vertically reciprocated in the scanning area SA while shifting its position in the horizontal direction can be adopted.
  • the function of at least one of the laser light source control unit 81 , the actuator control unit 82 , the own vehicle data acquisition unit 83 , the travelling environment determination unit 84 , the other vehicle data acquisition unit 85 , the other vehicle determination unit 86 , the pedestrian data acquisition unit 87 , the pedestrian determination unit 88 , the intersection data acquisition unit 89 , the intersection determination unit 90 , the pedestrian risk determination unit 91 and the display content determination unit 92 in the lamp ECU 51 is realized by a software executed by the cooperation of the lamp ECU 51 , the ROM 52 and the RAM 53 .
  • the function of at least one of the laser light source control unit 81 , the actuator control unit 82 , the own vehicle data acquisition unit 83 , the travelling environment determination unit 84 , the other vehicle data acquisition unit 85 , the other vehicle determination unit 86 , the pedestrian data acquisition unit 87 , the pedestrian determination unit 88 , the intersection data acquisition unit 89 , the intersection determination unit 90 , the pedestrian risk determination unit 91 and the display content determination unit 92 can be realized by a hardware such as a circuit element or a combination of a hardware and a software.

Abstract

In this invention, letters that can be seen by a pedestrian (101) are drawn on an intersection (103). A mark (MM2) that can be seen by the driver of the host vehicle (C) is drawn on the road (104) between the intersection (103) and the host-vehicle (C). The shapes of the letters (MO4) are corrected in accordance with the positional relationship between the pedestrian (101) and the intersection (103). The shape of the mark (MM2) is corrected in accordance with the positional relationship between the driver of the host-vehicle (C) and the road (104).

Description

  • This application is a Continuation of U.S. application Ser. No. 15/509,259 filed Jul. 17, 2017, that is a National Stage Entry of PCT/JP2015/075320 filed Sep. 7, 2015, that claims priority from Japanese Application No. 2014-181881 filed Sep. 8, 2014, each of which is incorporated herein by reference in its entirety.
  • TECHNICAL FIELD
  • The present invention relates to a road surface image-drawing system which is mounted on a vehicle and draws information regarding the vehicle or the like on a road surface.
  • BACKGROUND ART
  • There is known a vehicle display system for notifying/warning information of a vehicle to a driver himself or other person such as a pedestrian/oncoming vehicle by illuminating (drawing) a figure or a character or the like on a road surface. For example, a vehicle display system disclosed in Patent Document 1 is configured to notify/warn the approach of an own vehicle to other person trying to enter an intersection by displaying a warning mark of a figure on a road surface at the intersection.
  • CITATION LIST Patent Document
  • Patent Document 1: Japanese Patent Publication No. 2009-184428
  • DISCLOSURE OF INVENTION Problems to be Solved by Invention
  • However, the viewpoints of the driver of the own vehicle and other person receiving the warning are different from each other. Therefore, the display made by the system disclosed in Patent Document 1 may be distorted to the driver even when it is appropriate for other person receiving the warning. As a result, the driving comfort may be impaired.
  • The present invention aims to provide a suitable display to both a driver of a vehicle providing information such as a warning and other person receiving the information.
  • Means for Solving the Problems
  • In order to solve the above problem, one aspect of the present invention may provide a road surface image-drawing system mounted on a vehicle. The road surface image-drawing system includes:
  • an image-drawing device configured to draw a first display to be visible to a driver of the vehicle at a first position and draw a second display to be visible to a person other than the driver at a second position different from the first position, and
  • a control device configured to correct a shape of the first display according to a positional relationship between the first position and the driver and correct a shape of the second display according to a positional relationship between the second position and the person other than the driver.
  • For example, the first position and the second position are on a road. In this case, the first position may be on the road located between an intersection and the vehicle. The second position may be on the intersection. Alternatively, the first position may be on a windshield of the vehicle.
  • According to such a configuration, the first display can be drawn so as to have a distortion-free shape, as seen from the driver of the vehicle on which the system is mounted. On the other hand, the second display can be drawn so as to have a distortion-free shape, as seen from the person other than the driver. Accordingly, it is possible to provide a display including information such as a warning in a suitable manner to both the driver of the vehicle who provides the information and the person who receives the information. Particularly, the driver who views the first display can continue driving without diverting his attention to the second display.
  • It is desirable that the information included in the first display and the information included in the second display are related to each other. For example, the first display may include information or warning notifying the approach of other person to the driver. On the other hand, the second display may include information or warning notifying the approach of the vehicle to the other person. That is, the first display and the second display which are related to each other for the purpose of avoiding collision between the vehicle and the other person may be presented in a manner corresponding to the situation of each of the driver of the vehicle and the other person. Therefore, it is possible to properly call attention of both the driver and the other person.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a front view showing a headlamp device including a road surface image-drawing system according to one embodiment.
  • FIG. 2 is a longitudinal sectional view showing the headlamp device.
  • FIG. 3A is a view showing a configuration of an image-drawing device in the road surface image-drawing system.
  • FIG. 3B is a view showing a configuration of the image-drawing device in the road surface image-drawing system.
  • FIG. 4 is a functional block diagram showing a configuration of a control device in the road surface image-drawing system.
  • FIG. 5 is a view for explaining an image-drawing operation by the road surface image-drawing system.
  • FIG. 6 is a flow chart showing a control process executed by the road surface image-drawing system.
  • FIG. 7 is a flow chart showing a control process executed by the road surface image-drawing system.
  • FIG. 8A is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 8B is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 8C is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 8D is a view showing an example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 9 is a view showing another example of the image-drawing operation executed by the road surface image-drawing system.
  • FIG. 10 is a view for explaining a modified example of the road surface image-drawing system.
  • EMBODIMENT FOR CARRYING OUT INVENTION
  • Hereinafter, examples of preferred embodiments will be described in detail with reference to the accompanying figures. In each figure used in the following description, the scale is appropriately changed in order to make each member have recognizable size. The “right” and “left” used in the following description indicate the left and right directions as seen from a driver's seat.
  • FIG. 1 is a front view of a headlamp device 1 according to one embodiment. FIG. 2 shows a longitudinal section of the headlamp device 1, when seeing a line II-II in FIG. 1 from an arrow direction. The headlamp device 1 is a right headlamp device mounted on the right front part of a vehicle. Although not shown, a left headlamp device mounted on the left front part of the vehicle has a configuration bilaterally-symmetric to the headlamp device 1 shown in FIG. 1.
  • The headlamp device 1 includes a housing 2 and a translucent cover 3. The housing 2 has an opening portion. The translucent cover 3 is formed of a light-transmitting resin or glass or the like. The translucent cover 3 is attached to the opening portion of the housing 2 and defines a lamp chamber.
  • The headlamp device 1 further includes a high-beam light source unit 5 and a low-beam light source unit 6. The high-beam light source unit 5 and the low-beam light source unit 6 are accommodated in the lamp chamber. The high-beam light source unit 5 in the present example is a reflection type and is configured to illuminate the front of the vehicle with a high-beam light distribution pattern. The low-beam light source unit 6 in the present example is a projector type and is configured to illuminate the front of the vehicle with a low-beam light distribution pattern. The types of the high-beam light source unit 5 and the low-beam light source unit 6 can be appropriately determined, so long as a desired illumination can be obtained.
  • The headlamp device 1 includes a bracket 4 and aiming screws 14. The bracket 4 is accommodated in the lamp chamber. The high-beam light source unit 5 and the low-beam light source unit 6 are supported by the bracket 4. The bracket 4 is fixed to the housing 2 via the aiming screws 14. The aiming screws 14 are provided at three corner portions of the bracket 4. By operating the aiming screws 14 to rotate, the direction of an optical axis of the headlamp device 1 can be adjusted in an up-down direction and a left-right direction.
  • The headlamp device 1 includes an extension 12. The extension 12 covers each component disposed in the lamp chamber so that these components cannot be visible from the outside while allowing the passage of the light emitted from the high-beam light source unit 5 and the low-beam light source unit 6.
  • The headlamp device 1 includes a road surface image-drawing system 7. The road surface image-drawing system 7 is accommodated in the lamp chamber. The road surface image-drawing system 7 includes an image-drawing device 8 and a control device 9. The image-drawing device 8 includes a laser light source unit 10 and a scanning mechanism 11. The control device 9 is fixed to the housing 2.
  • FIG. 3A schematically shows a configuration of the laser light source unit 10. The laser light source unit 10 includes a first light source 15, a second light source 16, a third light source 17 and a support stand 18. The first light source 15, the second light source 16 and the third light source 17 are supported on the support stand 18.
  • The first light source 15 is configured to emit a red laser light. The second light source 16 is configured to emit a green laser light. The third light source 17 is configured to emit a blue laser light. In the present embodiment, each of the first light source 15, the second light source 16 and the third light source 17 is a laser diode.
  • The laser light source unit 10 includes a first condensing lens 20, a second condensing lens 21, a third condensing lens 22, a first dichroic mirror 25, a second dichroic mirror 26 and a third dichroic mirror 27. The red laser light emitted from the first light source 15 is condensed by the first condensing lens 20 and is incident on the first dichroic mirror 25. The green laser light emitted from the second light source 16 is condensed by the second condensing lens 21 and is incident on the second dichroic mirror 26. The blue laser light emitted from the third light source 17 is condensed by the third condensing lens 22 and is incident on the third dichroic mirror 27.
  • The laser light source unit 10 includes a condensing lens 19. The light reflected by the first dichroic mirror 25, the second dichroic mirror 26 and the third dichroic mirror 27 passes through the condensing lens 19 as an output light B2. By controlling the turn on/off of the first light source 15, the second light source 16 and the third light source 17, the output light B2 can be a monochromatic light, a mixed color light or a white light.
  • The laser light source unit 10 includes a monitor 28. The monitor 28 is configured to be able to monitor the intensity of the red laser light, the green laser light, the blue laser light and the output light B2. Further, the monitor 28 is configured to be able to control the intensity of the red laser light emitted from the first light source 15, the intensity of the green laser light emitted from the second light source 16 and the intensity of the blue laser light emitted from the third light source 17.
  • FIG. 3B shows a configuration of the scanning mechanism 11. The scanning mechanism 11 is configured as an MEMS (Micro Electro Mechanical Systems) mirror. The scanning mechanism 11 includes a base part 37, a first pivoting body 38, a second pivoting body 39, a pair of first torsion bars 40 and a pair of second torsion bars 41.
  • The base part 37 has a first opening portion. The first pivoting body 38 is disposed in the opening portion and is supported on the base part 37 via the pair of first torsion bars 40. The first pivoting body 38 is pivotable in a left-right (horizontal) direction.
  • The first pivoting body 38 has a second opening portion. The second pivoting body 39 is disposed in the second opening portion and is supported on the first pivoting body 38 via the pair of second torsion bars 41. The second pivoting body 39 is pivotable in an up-down (vertical) direction. The second pivoting body 39 has a reflective surface 36.
  • The scanning mechanism 11 includes a pair of first permanent magnets 42, a pair of second permanent magnets 43 and a terminal portion 44.
  • The pair of first torsion bars 40 is disposed between the pair of first permanent magnets 42. The direction of a straight line connecting the pair of first permanent magnets 42 is orthogonal to the extending direction of the pair of first torsion bars 40.
  • The pair of second torsion bars 41 is disposed between the pair of second permanent magnets 43. The direction of a straight line connecting the pair of second permanent magnets 43 is orthogonal to the extending direction of the pair of second torsion bars 41.
  • The first pivoting body 38 includes a first coil (not shown). The second pivoting body 39 includes a second coil (not shown). The first coil and the second coil are electrically connected to the control device 9 via the terminal portion 44.
  • As shown in FIG. 2, the laser light source unit 10 is fixed to the bracket 4 so that the output light B2 is emitted downward. The scanning mechanism 11 is disposed so that the output light B2 is reflected to the front of the headlamp device 1 by the reflective surface 36.
  • As shown in FIG. 4, the scanning mechanism 11 includes a scanning actuator 58. Specifically, the scanning actuator 58 is composed of the pair of first permanent magnets 42, the pair of second permanent magnets 43, the first coil and the second coil. When the magnitude and direction of current flowing in the first coil and the second coil are changed, the pivoting direction and amount of the first pivoting body 38 and the second pivoting body 39 are changed. In this way, the direction of the reflective surface 36 is changed in the up-down direction and the left-right direction.
  • As shown in FIG. 4, the control device 9 is communicatively connected to a vehicle speed sensor 60, a steering angle sensor 61, an acceleration sensor 62, a direction indicator detection sensor 63, a brake sensor 64, a GPS 65, a navigation system 66, an on-vehicle camera 67, a preceding vehicle detection sensor 68, a human body detection sensor 69, an illuminance sensor 70, an external system 71, and an image processing device 72.
  • The vehicle speed sensor 60 is configured to detect a travelling distance and a vehicle speed by the rotation of vehicle wheels. The steering angle sensor 61 is configured to detect a steering angle of a steering wheel. The acceleration sensor 62 is configured to detect an operating amount of an accelerator. The direction indicator detection sensor 63 is configured to detect a switch operation of a direction indicator. The brake system 64 is configured to detect an operating amount of a brake. The navigation system 66 is configured to detect a current position of the own vehicle and perform a route guidance to the destination based on a data received from the GPS 65 and a map data. The on-vehicle camera 67 is configured to capture and record an image in front of the own vehicle. The preceding vehicle detection sensor 68 is configured to detect a distance to a preceding vehicle and a relative speed by using millimeter waves or the like. The human body detection sensor 69 is configured to detect a human body located in front of the own vehicle by using infrared rays or the like. The illuminance sensor 70 is configured to detect an illuminance of the surrounding environment of the own vehicle by using the photoelectric effect. As an example of the external system 71, an intersection camera or a monitoring camera installed at intersection signals can be exemplified. The image processing device 72 is configured to perform an image processing on an image data acquired by the on-vehicle camera 67 and the external system 71 and thus to recognize a specific object.
  • The control device 9 includes a lamp ECU 51, a ROM 52 and a RAM 53. The lamp ECU 51 executes various programs stored in advance in the ROM 52 on the RAM 53, thereby generating various control signals.
  • The lamp ECU 51 includes an own vehicle data acquisition unit 83. The own vehicle data acquisition unit 83 is configured to receive data from at least one of the vehicle speed sensor 60, the steering angle sensor 61, the acceleration sensor 62, the direction indicator detection sensor 63, the brake sensor 64, the GPS 65, the navigation system 66 and the illuminance sensor 70 and to acquire information on the travelling conditions including the current position of the own vehicle and the surrounding environments.
  • The lamp ECU 51 includes a travelling environment determination unit 84. The travelling environment determination unit 84 is configured to determine whether image-drawing by the image-drawing device 8 is possible based on the information on the travelling conditions and surrounding environments of the own vehicle acquired by the own vehicle data acquisition unit 83. When it is determined that the image-drawing is possible, the travelling environment determination unit 84 is configured to permit the image-drawing by the image-drawing device 8. When it is determined that the image-drawing is not possible, the travelling environment determination unit 84 is configured to prohibit the image-drawing by the image-drawing device 8.
  • For example, when the own vehicle is stopped, the travelling environment determination unit 84 determines that the image-drawing is unnecessary, thereby prohibiting the image-drawing. When the own vehicle is travelling at a speed within a predetermined range, the travelling environment determination unit 84 permits the image-drawing. When the speed of the own vehicle exceeds a predetermined value, the travelling environment determination unit 84 determines that effective image-drawing cannot be performed, thereby prohibiting the image-drawing.
  • As another example, when it is determined that the own vehicle is travelling around a curve to the extent that it cannot perform the effective image-drawing, the travelling environment determination unit 84 prohibits the image-drawing. When it is determined that the own vehicle is travelling on the priority road, the travelling environment determination unit 84 permits the image-drawing. When it is determined that the own vehicle is in an environment in which the image-drawing cannot be effectively performed on the road surface due to a tunnel or rainy weather or the like, the travelling environment determination unit 84 prohibits the image-drawing. When the presence of a signal is detected, the possibility of the run-out by other person is low, and thus, the travelling environment determination unit 84 may determine that the image-drawing is unnecessary in order to make driving easier.
  • The lamp ECU 51 includes an other vehicle data acquisition unit 85. The other vehicle data acquisition unit 85 is configured to acquire data from at least one of the on-vehicle camera 67, the preceding vehicle detection sensor 68 and the external system 71 and to acquire information on a distance to the other vehicle and a relative speed.
  • The lamp ECU 51 includes an other vehicle determination unit 86. The other vehicle determination unit 86 is configured to determine whether the image-drawing by the image-drawing device 8 is possible based on the information acquired by the other vehicle data acquisition unit 85 and to determine whether the image-drawing by the image-drawing device 8 is possible. When it is determined that the image-drawing is possible, the other vehicle determination unit 86 is configured to permit the image-drawing by the image-drawing device 8. When it is determined that the image-drawing is not possible, the other vehicle determination unit 86 is configured to prohibit the image-drawing by the image-drawing device 8.
  • For example, when it is determined that a preceding vehicle is not present, the other vehicle determination unit 86 permits the image-drawing. When it is determined that a preceding vehicle is present, the other vehicle determination unit 86 permits the image-drawing. When it is determined that an image-drawable inter-vehicle distance is secured even if a preceding vehicle is present, the other vehicle determination unit 86 permits the image-drawing. When it is determined that effective image-drawing cannot be performed in consideration of the relative speed even if the image-drawable inter-vehicle distance is secured, the other vehicle determination unit 86 prohibits the image-drawing. When it is determined that the image-drawing can be performed on the road surface on the side of the preceding vehicle even if the inter-vehicle distance is too short to perform effective image-drawing, the other vehicle determination unit 86 permits the image-drawing.
  • The lamp ECU 51 includes a pedestrian data acquisition unit 87. The pedestrian data acquisition unit 87 is configured to acquire data from at least one of the on-vehicle camera 67, the human body detection sensor 69 and the external system 71 and to acquire information on the presence of a pedestrian (including a bicycle) and the state (traveling direction, etc.) thereof.
  • The lamp ECU 51 includes a pedestrian determination unit 88. The pedestrian determination unit 88 is configured to determine the presence/absence of a pedestrian based on the information acquired by the pedestrian data acquisition unit 87.
  • The lamp ECU 51 includes an intersection data acquisition unit 89. The intersection data acquisition unit 89 is configured to acquire data from at least one of the on-vehicle camera 67 and the navigation system 66 and thus to acquire information on the presence of an intersection and the situations (the presence/absence of a signal the presence/absence of a road illumination, the presence of obstacles around the intersection, etc.) of the intersection.
  • The lamp ECU 51 includes an intersection determination unit 90. The intersection determination unit 90 is configured to determine the presence/absence of an intersection based on the information acquired by the intersection data acquisition unit 89. The intersection determination unit 90 can be also configured to determine whether it is an intersection in which the necessity of performing the image-drawing is high, based on the information on the situations of the intersection acquired by the intersection data acquisition unit 89. When it is determined that an intersection is present (when it is determined that the notification or warning by the image-drawing is necessary in consideration of the situations of the intersection), the intersection determination unit 90 permits the image-drawing in the normal mode.
  • The lamp ECU 51 includes a pedestrian risk determination unit 91. The pedestrian risk determination unit 91 is configured to acquire a distance (hereinafter, referred to as a “pedestrian distance”) to a pedestrian based on the position, travelling direction and walking speed or the like of a pedestrian. Further, the pedestrian risk determination unit 91 is configured to acquire a braking distance of the own vehicle based on the position, travelling direction and driving speed or the like of the own vehicle. When the pedestrian distance is equal to or greater than the braking distance, the pedestrian risk determination unit 91 permits the image-drawing in the normal mode. On the other hand, when the pedestrian distance is less than the braking distance, there is a danger of collision, and thus, the pedestrian risk determination unit 91 permits the image-drawing in the emergency mode.
  • The lamp ECU 51 includes a display content determination unit 92. The display content determination unit 92 is configured to determine display contents when the image-drawing is permitted by the travelling environment determination unit 84, the other vehicle determination unit 86, the pedestrian determination unit 88 or the intersection determination unit 90). Specifically, the display content determination unit 92 is configured to determine a figure or a character to be drawn together with its display form in accordance with the determined situations (including a risk). Further, the display content determination unit 92 is configured to control the image-drawing device 8, thereby image-drawing the display contents.
  • The lamp ECU 51 includes a laser light source control unit 81. The laser light source control unit 81 is configured to control the turn on/off and emission light intensity (i.e., color and brightness of output light B2) of each of the first light source 15, the second light source 16 and the third light source 17, based on the display contents determined by the display content determination unit 92.
  • The lamp ECU 51 includes an actuator control unit 82. The actuator control unit 82 is configured to control an operation of the scanning actuator 58 based on the display contents determined by the display content determination unit 92.
  • Therefore, through the control by the laser light source control unit 81 and the actuator control unit 82, the output light B2 emitted from the laser light source unit 10 is reflected in a desired direction by the scanning mechanism 11. In this way, the image-drawing device 8 causes desired display contents to be drawn on a road surface located in front of the headlamp device 1 by using the output light B2.
  • Next, a road surface image-drawing operation performed by the road surface image-drawing system 7 will be described with reference to FIG. 5. FIG. 5 shows a virtual vertical screen positioned in front of the headlamp device 1. A reference numeral “CR” indicates a travelling lane of the own vehicle.
  • The scanning mechanism 11 of the image-drawing device 8 is configured so that the output light B2 can scan the inside of a rectangular scanning area SA shown in FIG. 5. The scanning area SA is preferably set such that it can cover a road shoulder of the travelling lane CR, a lane adjacent to the travelling lane CR and the range of 5 m to 50 m in front of the own vehicle. For example, it is preferable that the scanning area SA is set to the left-right direction range of −1° to −8° on a horizontal line H-H shown in FIG. 5 and the up-down direction range of 20° to −20° on a vertical line V-V shown in FIG. 5.
  • The control device 9 controls an operation of the image-drawing device 8 to cause a display MM for a driver and a display MO for other person to be drawn at different places in the scanning area SA, respectively. The display MM (an example of the first display) for the driver is a display to be visible to the driver of the own vehicle. The display MO (an example of the second display) for the other person is a display to be visible to a person other than the driver of the own vehicle, such as a pedestrian or a driver of a preceding vehicle. The display MM for the driver and the display MO for the other person are based on the display contents determined by the display content determination unit 92 of the control device 9. The display MM for the driver and the display MO for the other person may be drawn in a partially overlapping manner, as long as each of them includes a different point in the scanning area SA.
  • The scanning mechanism 11 of the image-drawing device 8 changes the orientation of the reflective surface 36 based on a control signal from the actuator control unit 82 of the control device 9. In this way, the scanning point can be horizontally reciprocated in the scanning area SA while shifting its position in the vertical direction by dl. A reference numeral “SI” in FIG. 5 indicates a movement trajectory of the scanning point.
  • When the scanning point reaches the image-drawing positions of the display MM for the driver and the display MO for the other person, at least one of the first light source 15, the second light source 16 and the third light source 17 is turned on (pan indicated by a solid line, out of the movement trajectory SI of the scanning point shown in FIG. 5) on the basis of the control signal from the laser light source control unit 81 of the control device 9. In this way, the display MM for the driver and the display MO for the other person are drawn at desired positions by using the output light B2. When the scanning point is located at positions other than the drawing positions of the display MM for the driver and the display MO for the other person, all of the first light source 15, the second light source 16 and the third light source 17 are turned off (part indicated by a broken line, out of the movement trajectory SI of the scanning point shown in FIG. 5).
  • The road surface image-drawing system 7 repeatedly performs the scanning of one cycle shown in FIG. 5 and appropriately determines the drawing positions of the display MM for the driver and the display MO for the other person in each cycle. In this way, the display MM for the driver and the display MO for the other person with an arbitrary shape can be continuously or intermittently displayed in the scanning area SA. The road surface image-drawing system 7 appropriately changes the drawing positions of the display MM for the driver and the display MO for the other person in each cycle. In this way, the display form of the display MM for the driver and the display MO for the other person can be dynamically changed.
  • Next, an operation of the road surface image-drawing system 7 will be described in more detail with reference to the flow charts shown in FIGS. 6 and 7.
  • When the road surface image-drawing system 7 is activated at a predetermined timing, information on the travelling conditions and surrounding environments of the own vehicle is acquired by the own vehicle data acquisition unit 83 (Step S10). Then, when it is determined by the travelling environment determination unit 84 based on the information that the image-drawing is possible (Step S11: YES), the process proceeds to Step S12. When it is determined by the travelling environment determination unit 84 that the image-drawing is not possible (Step S11: NO), the image-drawing is not performed.
  • Subsequently, information on a distance to a preceding vehicle and a relative speed is acquired by the other vehicle data acquisition unit 85 (Step S12). Then, when it is determined by the other vehicle determination unit 86 based on the information that image-drawing is possible or partial image-drawing is possible (Step S13: YES), the process proceeds to Step S14. When it is determined by the other vehicle determination unit 86 that the image-drawing is not possible (Step S13: NO), the image-drawing is not performed.
  • Subsequently, information on a pedestrian is acquired by the pedestrian data acquisition unit 87 (Step S14). Then, when it is determined by the pedestrian determination unit 88 based on the information that a pedestrian is present (Step S15: YES), the process proceeds to Step S17. When it is determined by the pedestrian determination unit 88 that a pedestrian is not present (Step S15: NO), the process proceeds to Step S16.
  • In Step S16, information on an intersection is acquired by the intersection data acquisition unit 89. When it is determined by the intersection determination unit 90 based on the information that an intersection is present (Step S16: YES), an instruction to perform image-drawing in the normal mode is issued and the process proceeds to Step S18. When it is determined by the intersection determination unit 90 that an intersection is not present (Step S16: NO), image-drawing is not performed and the process is repeated from the Step S10 again.
  • In Step S17, a pedestrian distance and a braking distance are acquired by the pedestrian risk determination unit 91. When the pedestrian distance is equal to or greater than the braking distance, an instruction to perform image-drawing in the normal mode is issued and the process proceeds to Step S18. When the pedestrian distance is less than the braking distance, an instruction to perform image-drawing in the emergency mode is issued and the process proceeds to Step S18.
  • The Step S18 is performed when image-drawing permission is made in previous Step S11 and Step S13 and image-drawing permission is made in Step S15 or S16. In Step S18. “display contents” are determined by the display content determination unit 92. FIG. 7 shows details of the process performed by the display content determination unit 92 in Step S18.
  • In Step S21, a figure or character or the like to be drawn as the display MO for the other person and an image-drawing area thereof and a figure or character or the like to be drawn as the display MM for the driver and an image-drawing area thereof are determined in accordance with the situations corresponding to the data acquired in previous steps (Step S10 to Step S16).
  • When the display MO for the other person is presented to a pedestrian, the display content determination unit 92 performs a trapezoidal distortion correction on the figure or character or the like to be drawn as the display MO for the other person so that a distortion-free shape is obtained as seen from the detected pedestrian. When the display MO for the other person is presented to a driver of a preceding vehicle, the display content determination unit 92 performs a trapezoidal distortion correction on the figure or character or the like to be drawn as the display MO for the other person so that a distortion-free shape is obtained as seen from the driver of the preceding vehicle detected. In other words, the display content determination unit 92 corrects the shape of the display MO for the other person based on a positional relationship between the detected other person (person other than the driver of the own vehicle) and the image-drawing area of the display MO for the other person.
  • Further, the display content determination unit 92 performs a trapezoidal distortion correction on the figure or character or the like to be drawn as the display MM for the driver so that a distortion-free shape is obtained as seen from the driver of the own vehicle. The trapezoidal distortion correction is performed on the basis of an installation height of the headlamp device 1 or a distance to the image-drawing area or the like. In other words, the display content determination unit 92 corrects the shape of the display MM for the driver based on a positional relationship between the driver of the own vehicle and the drawing area of the display MM for the driver.
  • Subsequently, the display content determination unit 92 sequentially determines whether an instruction to perform the image-drawing in the normal mode in previous steps (Step S16 to Step S17) is issued (Step S22) and whether an instruction to perform the image-drawing in the emergency mode is issued (Step S24). When the instruction to perform the image-drawing is not issued (Step S22 and Step S23: NO), image-drawing is not performed and the process is repeated from the Step S10 again.
  • When the instruction to perform the image-drawing in the normal mode is issued (Step S22: YES), the image-drawing is performed such that the figure or character or the like determined in the image-drawing area determined in Step S21 is constantly displayed or is displayed with slow blinking (Step S23).
  • When the instruction to perform the image-drawing in the emergency mode is issued (Step S24: YES), the image-drawing is performed such that the figure or character or the like determined in the image-drawing area determined in Step S21 is displayed with fast blinking. The blinking frequency of the display in the emergency mode image-drawing is faster than that of the display in the normal mode image-drawing. The blinking frequency of the display in the emergency mode image-drawing is, for example, 1 Hz.
  • Subsequently, it is determined whether the instruction to perform the image-drawing in the emergency mode is continued for a predetermined time or more (Step S26). When it is determined that the instruction to perform the image-drawing in the emergency mode is continued for the predetermined time or more (Step S26: YES), the blinking frequency is changed so as to rise (Step S27). For example, when the initial fast blinking (1 Hz) is continued for the predetermined time or more, the blinking frequency is increased by 1 Hz. For example, the image-drawing is performed such that this processing is repeated until the blinking frequency becomes 4 Hz. This is repeatedly performed, for example, up to 4 Hz. When the instruction to perform the image-drawing in the emergency mode is not continued for the predetermined time or more (Step S26: NO), the image-drawing in the initial fast blinking state is continued.
  • A preferred example of display by the road surface image-drawing system 7 having the above configuration will be described with reference to FIGS. 8A to 10.
  • FIGS. 8A to 8D show an example where the road surface image-drawing is performed at an intersection by the road surface image-drawing system 7. FIGS. 8A to 8D schematically show a state of an intersection 103 seen from above. In the present example, the travelling environment determination unit 84 determines that the travelling conditions and the surrounding environments are in the state of capable of performing the image-drawing (Step S11: YES). The other vehicle data acquisition unit 85 determines that an image-drawable distance (e.g., 50 m) is secured although a preceding vehicle 102 is present (Step S13: YES). The pedestrian data acquisition unit 87 determines that a pedestrian is not present (Step S15: NO). The intersection determination unit 90 determines that an intersection is present (e.g., at 30 m) in front of the own vehicle C (Step S16: YES). Therefore, an instruction to perform the image-drawing in the normal mode is issued.
  • Based on these conditions, display contents are determined (Step S21). Specifically, it is determined that a “rectilinear arrow” for informing the other person that the own vehicle C goes straight through the intersection 103 is drawn at the intersection 103 (an example of the second position) as the display for the other person. Further, it is determined that a “rectilinear arrow” for informing the driver of the own vehicle C of the same information as the display for the other person is drawn on a road 104 (an example of the first position) as the display for the driver. The road 104 is positioned between the own vehicle C and the intersection 103.
  • As shown in FIG. 8A, a straight arrow MO1 as the display for the other person is drawn in a left end region of the intersection 103. The shape of the straight arrow MO1 is corrected so that it can be seen without distortion, as viewed from a driver of other vehicle or a pedestrian who is likely to enter the intersection 103 from the left side.
  • As shown in FIG. 8B, a straight arrow MO2 as the display for the other person is drawn in a right end region of the intersection 103. The shape of the straight arrow MO2 is corrected so that it can be seen without distortion, as viewed from a driver of other vehicle or a pedestrian who is likely to enter the intersection 103 from the right side.
  • As shown in FIG. 8C, a straight arrow MO3 as the display for the other person is drawn in a front end region of the intersection 103. The shape of the straight arrow MO3 is corrected so that it can be seen without distortion, as viewed from a driver of other vehicle or a pedestrian who is likely to enter the intersection 103 from the front side.
  • As shown in FIG. 8D, a straight arrow MM1 as the display for the driver is drawn on the road 104. The shape of the straight arrow MM1 is corrected so that it can be seen without distortion, as viewed from the driver of the own vehicle C.
  • According to these displays, the straight arrows MO1, MO2, MO3 as the displays for the other person are presented without distortion to the other person approaching the intersection 103. Therefore, the other person can accurately recognize the fact that the own vehicle C is approaching the intersection 103. On the other hand, the display for the driver related to the display for the other person is presented without distortion to the driver of the own vehicle C. Therefore, the driver of the own vehicle C can continue driving without diverting his attention to the display for the other person.
  • The straight arrows MO1, MO2, MO3 as the displays for the other person are preferably displayed to be large in the full width of the intersection 103 in order to enhance the effect of calling attention. On the other hand, the straight arrow MM1 as the display for the driver is preferably spot-displayed on the tip of line of sight of the driver of the own vehicle C. The line of sight of the driver is detected by, for example, an eye camera connected to the lamp ECU 51.
  • The timing at which the straight arrows MO1, MO2, MO3 as the displays for the other person and the straight arrow MM1 as the display for the driver are drawn can be appropriately determined. These displays may be sequentially drawn one by one, or, at least two of these displays may be simultaneously drawn when the image-drawing areas thereof are not overlapped. When the displays are sequentially drawn, the image-drawing time of each display and the number of scanning cycles described with reference to FIG. 5 can be appropriately determined. For example, the display to be drawn per one scanning cycle may be switched or a specific display may be drawn over a plurality of scanning cycles. When the displays are sequentially drawn, the order in which each display is drawn can be appropriately determined.
  • FIG. 9 shows another example in which the road surface image-drawing is performed at the intersection by the road surface image-drawing system 7. FIG. 9 schematically shows a state of the intersection 103 seen from above. In this example, the pedestrian determination unit 88 determines that a pedestrian 101 tying to enter the intersection 103 is present (Step S15: YES). The pedestrian risk determination unit 91 determines that the distance (e.g., 20 m) between the own vehicle C and the pedestrian 101 is equal to or greater than the braking distance (e.g., 13 m) of the own vehicle C (Step S17). Therefore, an instruction to perform the image-drawing in the normal mode is issued.
  • Based on these conditions, display contents are determined (Step S21). Specifically, it is determined that the character of “STOP” for urging not to enter the intersection 103 is drawn at the intersection 103 as the display for the other person. Further, it is determined that a mark for informing the driver of the own vehicle C of the presence of the pedestrian 101 trying to enter the intersection 103 is drawn on the road 104 as the display for the driver.
  • As shown in FIG. 9, a character MO4 (an example of the second display) of “STOP” as the display for the other person is drawn at the intersection 103 (an example of the second position). The shape of the character M04 is corrected so that it can be seen without distortion, as viewed from the pedestrian 101 trying to enter the intersection 103 from the left side.
  • On the other hand, a mark MM2 (an example of the first display) indicating the presence of the pedestrian 101, the pedestrian distance (20 m) and the approaching direction to the intersection 103 is drawn on the road 104 (an example of the first position) as the display for the driver. The shape of the mark MM2 is corrected so that it can be seen without distortion, as viewed from the driver of the own vehicle C.
  • According to these displays, the character MO4 as the display for the other person is presented without distortion to the pedestrian 101 approaching the intersection 103. Therefore, the pedestrian 101 can accurately recognize the fact that the own vehicle C is approaching the intersection 103. On the other hand, the mark MM2 related to the display for the other person is presented without distortion to the driver of the own vehicle C. Therefore, the driver of the own vehicle C can continue driving without diverting his attention to the display for the other person and can accurately recognize the fact that the pedestrian 101 is approaching the intersection 103.
  • Further, in the present example, the character MO4 as the display for the other person is intended to inform the pedestrian 101 of the approach of the own vehicle C to the intersection 103 and the mark MM2 as the display for the driver is intended to inform the driver of the own vehicle C of the approach of the pedestrian 101 to the intersection 103. That is, information for avoiding collision at the intersection 103 is presented in a manner corresponding to the situation of each of the driver of the own vehicle C and the pedestrian 101. Therefore, it is possible to properly call attention of both the driver of the own vehicle C and the pedestrian 101.
  • A case in which the pedestrian 101 overlooks the character MO4 as the display for the other person and further approaches the intersection 103 from the situation shown in FIG. 9 will be described as an example.
  • The pedestrian risk determination unit 91 determines that the distance (e.g., 10 m) to the pedestrian 101 is less than the braking distance (e.g., 13 m) of the own vehicle C (Step S17). Therefore, an instruction to perform the image-drawing in the emergency mode is issued.
  • Based on these conditions, display contents are determined (Step S21). Specifically, it is determined that the character MO4 and the mark MM2 are blinked at a high speed. Display contents with higher warning than the character M04 and the mark MM2 may be selected.
  • Based on the determined display contents, the image-drawing in the emergency mode is performed (Steps S24 and S25). In the case where the instruction to perform the image-drawing in the emergency mode is continued for a predetermined time (e.g., one second) (Step S26: YES), the blinking frequency is increased (Step S27).
  • The above description is intended to facilitate the understanding of the present invention and does not limit the present invention. It is obvious that the present invention can be changed/enhanced without departing from the spirit thereof and equivalents thereof are included in the present invention.
  • The forms of the displays MO1, MO2, MO3, MO4 for the other person are not limited to the examples described for the above embodiment. Any suitable form can be adopted according to the situations, so long as it can notify the other person of the approach or behavior of the own vehicle. For example, more detailed information such as an operation of a brake or a steering wheel by the driver of the own vehicle, acceleration of the own vehicle and driver's attribute information (sex or whether he is a child or an elderly person) may be included in the display for the other person. Alternatively, when the own vehicle data acquisition unit 83 determines that the driver of the own vehicle recognizes the other person such as a pedestrian based on the brake operation, travelling speed and information from the eye camera or the like, the content notifying the fact may be included in the display for the other person.
  • The forms of the displays MM1, MM2 for the driver are not limited to the examples described for the above embodiment. Any suitable form can be adopted according to the situations, so long as it can notify the driver of the own vehicle of the presence or behavior of the other person. For example, more detailed information such as the number and attribute (sex or whether he is a child or an elderly person) of the pedestrian acquired by the pedestrian data acquisition unit 87 may be included in the display for the driver. Alternatively, when the pedestrian data acquisition unit 87 determines that the pedestrian recognizes the own vehicle, the content notifying the fact may be included in the display for the driver.
  • The display for the other person and the display for the driver are based on mutually related information. In order to emphasize this relevance, a line connecting the display for the other person and the display for the driver may be drawn.
  • A region that looks relatively dark may be drawn so as to border the display for the other person and the display for the driver. Specifically, the region may be drawn with a color which has brightness or saturation lower than the colors of the display for the other person and the display for the driver or a color which makes the colors of the display for the other person and the display for the driver stand out. In this way, the display for the other person and the display for the driver appear to emerge, and thus, visibility is improved.
  • In the above embodiment, the necessity of the image-drawing operation in the emergency mode is determined from a relationship between the distance from the own vehicle to the pedestrian and the braking distance of the own vehicle. However, as a condition for performing the image-drawing operation in the emergency mode, suitable conditions such as when the own vehicle accelerates or suddenly decelerates, when the pedestrian jumps ahead of the own vehicle and when the distance to the intersection is less than a predetermined value may be adopted.
  • In the image-drawing operation in the emergency mode, the blinking of the display for the other person and the display for the driver do not need to be synchronized. From the viewpoint of calling attention to other person, the blinking frequency of the display for the other person may be higher than that of the display for the driver.
  • In the image-drawing operation in the emergency mode, the brightness of the display for the other person and the display for the driver does not need to be the same. From the viewpoint of calling attention to other person, the display for the other person may be brighter than the display for the driver.
  • In the image-drawing operation in the emergency mode, the change of the display mode is not limited to the change of blinking frequency of the display. As the emergency increases, the brightness of the display may be increased or the color sense of the display may be changed.
  • The display for the other person and the display for the driver described with reference to FIGS. 8A to 9 do not need to be continued throughout the execution period of the image-drawing operation by the image-drawing device 8. It is also possible to adopt a configuration in which only a spotlight-like display is drawn on the road 104 when the distance to the intersection 103 is equal to or greater than a predetermined value and a display of the form illustrated in FIGS. 8A to 9 is drawn when the distance to the intersection 103 is less than the predetermined value.
  • In the above embodiment, the display for the other person and the display for the driver are drawn on the road in the travelling direction of the own vehicle C. However, when the own vehicle data acquisition unit 83 detects, through the steering angle sensor 61 or the direction indicator detection sensor 6, that the own vehicle C performs its course change, at least one of the display for the other person and the display for the driver can be drawn on the road after the course change.
  • The display for the other person and the display for the driver may be individually stopped at the discretion of the driver. For example, a switch for releasing the image-drawing operation can be provided in the own vehicle. Alternatively, it is also possible to adopt a configuration in which the image-drawing operation is automatically canceled when a passing operation, a winker operation, a hazard lamp operation, a switching operation to a first speed in a manual transmission vehicle or an off operation of a first speed switch, or the like is performed.
  • The whole of the road surface image-drawing system 7 is not necessarily provided in the headlamp device 1. For example, as shown in FIG. 10, the image-drawing device 8 can include a head up display (HUD) device 73. As shown in FIG. 4, the HUD device 73 can be connected to the lamp ECU 51. The HUD device 73 is configured to project the display MM (an example of the first display) for the driver on the windshield W (an example of the first position) of the own vehicle C. In the present specification, an operation of projecting a predetermined display on an object is also considered as an example of the “image-drawing”.
  • According to this configuration, the driver of the own vehicle C can visually recognize the display MM for the driver in a state of being overlaid on the front landscape seen through the windshield W. Therefore, the driver of the own vehicle C can visually recognize the display MM for the driver in a reliable manner, regardless of the state of the road surface.
  • The road surface image-drawing system 7 may be provided in any one of the left and right headlamp devices. When the road surface image-drawing system 7 is provided in each of the left and right headlamp devices, the road surface image-drawing systems 7 on the left and right sides can perform different image-drawing operations. For example, as shown in FIG. 9, when a pedestrian is present on the left side of the intersection 103 or is supposed to be on the left side, the character M04 may be drawn by the road surface image-drawing system 7 provided in the left headlamp device and the mark MM2 may be drawn by the road surface image-drawing system 7 provided in the right headlamp device. On the contrary, when a pedestrian is present on the right side of the intersection 103 or is supposed to be on the right side, the character M04 may be drawn by the road surface image-drawing system 7 provided in the right headlamp device and the mark MM2 may be drawn by the road surface image-drawing system 7 provided in the left headlamp device.
  • The whole of the road surface image-drawing system 7 may be disposed outside the headlamp device 1. For example, the road surface image-drawing system 7 can be disposed on a roof of the own vehicle such that an emission direction of the output light B2 is directed forward.
  • The configuration of the laser light source unit 10 can be appropriately changed in accordance with the specifications of the headlamp device 1. Instead of the first light source 15, the second light source 16 and the third light source 17 for emitting light of three colors in the above embodiment, a single light source for emitting a white light can be used. Alternatively, a configuration that a white light is emitted through excitation by making a blue laser light incident on a yellow phosphor may be adopted. Further, a fourth light source for emitting an orange light may be added. Meanwhile, the first light source 15, the second light source 16 and the third light source 17 may be laser light sources other than laser diodes.
  • Instead of the MEMS mechanism, a scanning optical system such as a galvanometer mirror or a DMMD (Digital Micro Mirror Device) can be adopted as the scanning mechanism 11. The scanning performed by the scanning mechanism 11 is not limited to the example shown in FIG. 5. For example, a configuration that the scanning point is vertically reciprocated in the scanning area SA while shifting its position in the horizontal direction can be adopted.
  • The function of at least one of the laser light source control unit 81, the actuator control unit 82, the own vehicle data acquisition unit 83, the travelling environment determination unit 84, the other vehicle data acquisition unit 85, the other vehicle determination unit 86, the pedestrian data acquisition unit 87, the pedestrian determination unit 88, the intersection data acquisition unit 89, the intersection determination unit 90, the pedestrian risk determination unit 91 and the display content determination unit 92 in the lamp ECU 51 is realized by a software executed by the cooperation of the lamp ECU 51, the ROM 52 and the RAM 53. However, the function of at least one of the laser light source control unit 81, the actuator control unit 82, the own vehicle data acquisition unit 83, the travelling environment determination unit 84, the other vehicle data acquisition unit 85, the other vehicle determination unit 86, the pedestrian data acquisition unit 87, the pedestrian determination unit 88, the intersection data acquisition unit 89, the intersection determination unit 90, the pedestrian risk determination unit 91 and the display content determination unit 92 can be realized by a hardware such as a circuit element or a combination of a hardware and a software.
  • The contents of Japanese Patent Application No. 2014-181881 filed on Sep. 8, 2014 are incorporated as a pan of the description of the present application.

Claims (1)

1. A display system for a vehicle configured to display information at a plurality of places, wherein
the display system is mounted on the vehicle,
the display system is configured to display a first display on a windshield of the vehicle for a driver of the vehicle and to display a second display on a road for a person outside the vehicle other than the driver, and
the information displayed in the plurality of places are related to each other.
US16/161,670 2014-09-08 2018-10-16 Road surface image-drawing system for vehicle Active US10741083B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/161,670 US10741083B2 (en) 2014-09-08 2018-10-16 Road surface image-drawing system for vehicle

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
JP2014181881A JP6746270B2 (en) 2014-09-08 2014-09-08 Vehicle display system
JP2014-181881 2014-09-08
PCT/JP2015/075320 WO2016039288A1 (en) 2014-09-08 2015-09-07 Road surface image-drawing system for vehicle
US201715509259A 2017-07-17 2017-07-17
US16/161,670 US10741083B2 (en) 2014-09-08 2018-10-16 Road surface image-drawing system for vehicle

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/JP2015/075320 Continuation WO2016039288A1 (en) 2014-09-08 2015-09-07 Road surface image-drawing system for vehicle
US15/509,259 Continuation US10134283B2 (en) 2014-09-08 2015-09-07 Road surface image-drawing system for vehicle

Publications (2)

Publication Number Publication Date
US20190051185A1 true US20190051185A1 (en) 2019-02-14
US10741083B2 US10741083B2 (en) 2020-08-11

Family

ID=55459036

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/509,259 Active US10134283B2 (en) 2014-09-08 2015-09-07 Road surface image-drawing system for vehicle
US16/161,670 Active US10741083B2 (en) 2014-09-08 2018-10-16 Road surface image-drawing system for vehicle

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/509,259 Active US10134283B2 (en) 2014-09-08 2015-09-07 Road surface image-drawing system for vehicle

Country Status (5)

Country Link
US (2) US10134283B2 (en)
EP (1) EP3192698B1 (en)
JP (1) JP6746270B2 (en)
CN (2) CN106794792B (en)
WO (1) WO2016039288A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210206316A1 (en) * 2018-09-28 2021-07-08 Koito Manufacturing Co., Ltd. Lamp system
US20220381415A1 (en) * 2020-02-17 2022-12-01 Koito Manufacturing Co., Ltd. Lamp system
US11584298B2 (en) 2018-08-15 2023-02-21 Koito Manufacturing Co., Ltd. Vehicle display system and vehicle
US11597316B2 (en) 2018-08-06 2023-03-07 Koito Manufacturing Co., Ltd. Vehicle display system and vehicle
US11639138B2 (en) 2018-08-06 2023-05-02 Koito Manufacturing Co., Ltd. Vehicle display system and vehicle
US11796807B2 (en) 2019-06-11 2023-10-24 Koito Manufacturing Co., Ltd. Head-up display device and head-up display system
US11805587B2 (en) 2017-09-01 2023-10-31 Koito Manufacturing Co., Ltd. Vehicle illumination system, vehicle system, lamp unit and vehicle lamp

Families Citing this family (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6319350B2 (en) * 2015-04-03 2018-05-09 株式会社デンソー Information presentation device
JP6418182B2 (en) * 2016-03-07 2018-11-07 トヨタ自動車株式会社 Vehicle lighting device
JP2017165178A (en) * 2016-03-15 2017-09-21 トヨタ自動車株式会社 Illuminating device for vehicle
JP6791644B2 (en) * 2016-03-24 2020-11-25 株式会社小糸製作所 Vehicle headlights
US10676023B2 (en) * 2016-07-29 2020-06-09 Koito Manufacturing Co., Ltd. Vehicle lighting system, vehicle system, and vehicle
JP6680136B2 (en) * 2016-08-08 2020-04-15 株式会社デンソー Exterior display processing device and exterior display system
JP6429413B2 (en) 2017-03-10 2018-11-28 株式会社Subaru Image display device
JP6465317B2 (en) 2017-03-10 2019-02-06 株式会社Subaru Image display device
JP6860385B2 (en) * 2017-03-10 2021-04-14 株式会社Subaru Image display device
JP6465318B2 (en) 2017-03-10 2019-02-06 株式会社Subaru Image display device
JP6593803B2 (en) * 2017-03-10 2019-10-23 株式会社Subaru Image display device
JP6515125B2 (en) 2017-03-10 2019-05-15 株式会社Subaru Image display device
JP6860384B2 (en) * 2017-03-10 2021-04-14 株式会社Subaru Image display device
JP6497818B2 (en) 2017-03-10 2019-04-10 株式会社Subaru Image display device
JP6497819B2 (en) 2017-03-10 2019-04-10 株式会社Subaru Image display device
DE102018102575B4 (en) * 2017-05-25 2024-03-07 Hasco Vision Technology Co., Ltd. Intelligent multi-function automotive headlight module
JP6690612B2 (en) * 2017-08-03 2020-04-28 トヨタ自動車株式会社 Out-of-vehicle alarm device
JP2019048524A (en) * 2017-09-08 2019-03-28 スタンレー電気株式会社 Device for controlling vehicle headlight
JP7089853B2 (en) * 2017-09-14 2022-06-23 株式会社Subaru Vehicle road surface drawing device
US10885772B2 (en) * 2017-12-21 2021-01-05 Lumileds Llc Road lighting
WO2019156244A1 (en) * 2018-02-09 2019-08-15 大日本印刷株式会社 Illumination device, method of designing illumination device, and designing device
US10525871B2 (en) * 2018-02-22 2020-01-07 Toyota Motor Engineering & Manufacturing North America, Inc. Vehicle lamp assemblies including light diverting system and apparatus
US10829036B2 (en) * 2018-04-18 2020-11-10 Maxell, Ltd. Road sign projector and vehicle lamp
WO2020008560A1 (en) * 2018-07-04 2020-01-09 三菱電機株式会社 Information display apparatus and information display method
EP3835128A4 (en) * 2018-08-06 2021-09-08 Koito Manufacturing Co., Ltd. Vehicle display system and vehicle
US20210256933A1 (en) * 2018-09-05 2021-08-19 Koito Manufacturing Co., Ltd. Vehicle display system, vehicle system, and vehicle
CN112770939A (en) 2018-09-25 2021-05-07 株式会社小糸制作所 Lamp system and vehicle lamp
WO2020064012A1 (en) * 2018-09-30 2020-04-02 长城汽车股份有限公司 Control method and system for vehicles and vehicle
CN109263541A (en) * 2018-09-30 2019-01-25 京东方科技集团股份有限公司 A kind of vehicle-mounted early warning system, vehicle-mounted method for early warning and computer storage medium
WO2020128548A1 (en) * 2018-12-21 2020-06-25 Micron Technology, Inc. Improved head lights of a vehicle
CN109649266A (en) * 2019-01-21 2019-04-19 北京百度网讯科技有限公司 Control method for vehicle, device, computer equipment and storage medium
CN109878513A (en) * 2019-03-13 2019-06-14 百度在线网络技术(北京)有限公司 Defensive driving strategy generation method, device, equipment and storage medium
JP2020152246A (en) 2019-03-20 2020-09-24 株式会社リコー Information provision system, movable body, information provision method, and information provision program
WO2020218498A1 (en) * 2019-04-25 2020-10-29 株式会社小糸製作所 Sensing system and vehicle
CN111907402A (en) * 2019-05-10 2020-11-10 阿里巴巴集团控股有限公司 Vehicle and information control method and device thereof
JP7422148B2 (en) * 2019-06-28 2024-01-25 株式会社小糸製作所 lighting system
JPWO2021065617A1 (en) 2019-09-30 2021-04-08
KR20210053385A (en) * 2019-11-01 2021-05-12 엘지전자 주식회사 Vehicle Having a Dangerous Situation Notification Function and Control Method thereof
JP7357284B2 (en) * 2020-02-12 2023-10-06 パナソニックIpマネジメント株式会社 Drawing system, display system, moving object, drawing method and program
CN113085899A (en) * 2021-04-28 2021-07-09 安徽江淮汽车集团股份有限公司 Visual field blind area early warning method, device, equipment and storage medium
JP2023137474A (en) * 2022-03-18 2023-09-29 トヨタ自動車株式会社 Notification method, notification device, and notification program

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030147247A1 (en) * 2002-02-07 2003-08-07 Toyota Jidosha Kabushiki Kaisha Vehicle operation supporting device and vehicle operation supporting system
US20100017111A1 (en) * 2006-04-13 2010-01-21 Ferrari S.P.A. Road vehicle motoring aid method and system
US20140062685A1 (en) * 2012-08-31 2014-03-06 Denso Corporation Pedestrian notification apparatus
US20140236483A1 (en) * 2013-02-19 2014-08-21 Navteq B.V. Method and apparatus for determining travel path geometry based on mapping information
US20150179063A1 (en) * 2013-12-23 2015-06-25 Basem Albert Aziz Pedestrian-warning light system for vehicles
US10214143B2 (en) * 2007-06-16 2019-02-26 Bayerische Motoren Werke Aktiengesellschaft Method for assisting a motor vehicle driver when driving through a narrow passage and/or for maintaining a safe distance from a vehicle in front

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004306894A (en) * 2003-04-10 2004-11-04 Suzuki Motor Corp Lighting system for vehicle
CA2441644A1 (en) * 2003-09-17 2005-03-17 Michael Perri Pedestrian presence indicator
JP2006284458A (en) * 2005-04-01 2006-10-19 Denso Corp System for displaying drive support information
JP4207060B2 (en) * 2006-05-31 2009-01-14 アイシン・エィ・ダブリュ株式会社 Drawing system
JP2008247334A (en) * 2007-03-30 2008-10-16 Aisin Aw Co Ltd Attention-attracting control device and attention-attractive control method
JP4985095B2 (en) * 2007-05-21 2012-07-25 アイシン・エィ・ダブリュ株式会社 Safe driving support device, safe driving support method and program
JP5151452B2 (en) * 2007-12-19 2013-02-27 株式会社豊田中央研究所 Information display device
DE102007061723A1 (en) * 2007-12-20 2009-06-25 Robert Bosch Gmbh Method for displaying warnings in a motor vehicle and corresponding display arrangement
JP5056445B2 (en) * 2008-02-04 2012-10-24 株式会社豊田中央研究所 Vehicle lighting device
JP2009298360A (en) * 2008-06-17 2009-12-24 Mazda Motor Corp Driving assistance system of vehicle
US8305444B2 (en) * 2008-11-14 2012-11-06 Toyota Motor Engineering & Manufacturing North America, Inc. Integrated visual display system
JP5582008B2 (en) * 2010-12-08 2014-09-03 トヨタ自動車株式会社 Vehicle information transmission device
JP5761119B2 (en) * 2012-05-17 2015-08-12 株式会社デンソー In-vehicle lighting system
JP2014013524A (en) * 2012-07-05 2014-01-23 Mitsubishi Motors Corp Vehicle notification device
JP5492962B2 (en) * 2012-09-28 2014-05-14 富士重工業株式会社 Gaze guidance system
JP5924223B2 (en) 2012-10-01 2016-05-25 株式会社デンソー Pressure sensor for tire pressure
US10195980B2 (en) * 2013-09-03 2019-02-05 Jaguar Land Rover Limited System for imaging
JP6214389B2 (en) 2013-12-26 2017-10-18 株式会社小糸製作所 Vehicle lighting
US10446021B2 (en) * 2014-07-25 2019-10-15 Moasis Inc. Onboard traffic and pedestrian warning systems and methods having optical and audio signal feedback and control

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030147247A1 (en) * 2002-02-07 2003-08-07 Toyota Jidosha Kabushiki Kaisha Vehicle operation supporting device and vehicle operation supporting system
US20100017111A1 (en) * 2006-04-13 2010-01-21 Ferrari S.P.A. Road vehicle motoring aid method and system
US10214143B2 (en) * 2007-06-16 2019-02-26 Bayerische Motoren Werke Aktiengesellschaft Method for assisting a motor vehicle driver when driving through a narrow passage and/or for maintaining a safe distance from a vehicle in front
US20140062685A1 (en) * 2012-08-31 2014-03-06 Denso Corporation Pedestrian notification apparatus
US20140236483A1 (en) * 2013-02-19 2014-08-21 Navteq B.V. Method and apparatus for determining travel path geometry based on mapping information
US20150179063A1 (en) * 2013-12-23 2015-06-25 Basem Albert Aziz Pedestrian-warning light system for vehicles

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11805587B2 (en) 2017-09-01 2023-10-31 Koito Manufacturing Co., Ltd. Vehicle illumination system, vehicle system, lamp unit and vehicle lamp
US11597316B2 (en) 2018-08-06 2023-03-07 Koito Manufacturing Co., Ltd. Vehicle display system and vehicle
US11639138B2 (en) 2018-08-06 2023-05-02 Koito Manufacturing Co., Ltd. Vehicle display system and vehicle
US11584298B2 (en) 2018-08-15 2023-02-21 Koito Manufacturing Co., Ltd. Vehicle display system and vehicle
US20210206316A1 (en) * 2018-09-28 2021-07-08 Koito Manufacturing Co., Ltd. Lamp system
US11820280B2 (en) * 2018-09-28 2023-11-21 Koito Manufacturing Co., Ltd. Lamp system
US11796807B2 (en) 2019-06-11 2023-10-24 Koito Manufacturing Co., Ltd. Head-up display device and head-up display system
US20220381415A1 (en) * 2020-02-17 2022-12-01 Koito Manufacturing Co., Ltd. Lamp system

Also Published As

Publication number Publication date
US10741083B2 (en) 2020-08-11
EP3192698A1 (en) 2017-07-19
EP3192698A4 (en) 2018-05-23
CN106794792A (en) 2017-05-31
US10134283B2 (en) 2018-11-20
JP2016055691A (en) 2016-04-21
CN111634226A (en) 2020-09-08
CN111634226B (en) 2023-09-12
JP6746270B2 (en) 2020-08-26
US20170337821A1 (en) 2017-11-23
WO2016039288A1 (en) 2016-03-17
CN106794792B (en) 2020-06-12
EP3192698B1 (en) 2022-05-04

Similar Documents

Publication Publication Date Title
US10741083B2 (en) Road surface image-drawing system for vehicle
JP5392470B2 (en) Vehicle display device
US11351912B2 (en) Vehicle headlamp system and vehicle lamp system
US10794719B2 (en) Vehicle image projection apparatus that projects an image in the front and back of a vehicle based on sensor information
KR101760995B1 (en) Driving support apparatus for vehicle
US20180118109A1 (en) Information presentation apparatus
CN105539269B (en) Vehicle headlamp system with adaptive light distribution
CN109311480B (en) Method for operating a driver assistance system and driver assistance system
US9776559B2 (en) Vehicular display system
JP2019217790A (en) Head-up display device
CN105976641A (en) Method for generating light distribution to send driving instruction to first vehicle
EP2653346A1 (en) Vehicle safety illumination arrangement and method
KR20180132668A (en) Method for controlling automatic display of a pictogram indicating the presence of a distress situation in front of a vehicle
WO2017134861A1 (en) Head-up display device
JP6391347B2 (en) Vehicle display system
US10730427B2 (en) Lighting device
JP5542501B2 (en) Vehicle lamp system
US20210188163A1 (en) U-turn signal indicator
CN110382296B (en) Lighting device
JP7002612B2 (en) Vehicle display system
JP2017007581A (en) Vehicular safe control apparatus
US9994146B2 (en) Blind spot indicator
US20170061221A1 (en) Method and device for detecting objects in the dark using a vehicle camera and a vehicle lighting system
WO2020230523A1 (en) Transportation system and transportation infrastructure
KR20190054262A (en) Apparatus and method for inducing head lamp turn-on of rear vehicle

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4