US11518300B2 - Presentation device - Google Patents
Presentation device Download PDFInfo
- Publication number
- US11518300B2 US11518300B2 US16/613,004 US201716613004A US11518300B2 US 11518300 B2 US11518300 B2 US 11518300B2 US 201716613004 A US201716613004 A US 201716613004A US 11518300 B2 US11518300 B2 US 11518300B2
- Authority
- US
- United States
- Prior art keywords
- vehicle
- behavior
- projection pattern
- pattern
- presentation device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000012545 processing Methods 0.000 claims description 34
- 206010000117 Abnormal behaviour Diseases 0.000 claims description 30
- 230000008859 change Effects 0.000 claims description 16
- 230000001133 acceleration Effects 0.000 claims description 6
- 238000005286 illumination Methods 0.000 abstract description 75
- 230000006399 behavior Effects 0.000 description 113
- 238000010586 diagram Methods 0.000 description 20
- 238000004891 communication Methods 0.000 description 19
- 230000006870 function Effects 0.000 description 19
- 238000000034 method Methods 0.000 description 18
- 230000008569 process Effects 0.000 description 14
- 238000012544 monitoring process Methods 0.000 description 11
- 230000002159 abnormal effect Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 230000001154 acute effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q1/00—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
- B60Q1/26—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic
- B60Q1/50—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic for indicating other intentions or conditions, e.g. request for waiting or overtaking
- B60Q1/525—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic for indicating other intentions or conditions, e.g. request for waiting or overtaking automatically indicating risk of collision between vehicles in traffic or with pedestrians, e.g. after risk assessment using the vehicle sensor data
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q1/00—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
- B60Q1/02—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to illuminate the way ahead or to illuminate other areas of way or environments
- B60Q1/24—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to illuminate the way ahead or to illuminate other areas of way or environments for lighting other areas than only the way ahead
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R21/00—Arrangements or fittings on vehicles for protecting or preventing injuries to occupants or pedestrians in case of accidents or other traffic risks
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62J—CYCLE SADDLES OR SEATS; AUXILIARY DEVICES OR ACCESSORIES SPECIALLY ADAPTED TO CYCLES AND NOT OTHERWISE PROVIDED FOR, e.g. ARTICLE CARRIERS OR CYCLE PROTECTORS
- B62J27/00—Safety equipment
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62J—CYCLE SADDLES OR SEATS; AUXILIARY DEVICES OR ACCESSORIES SPECIALLY ADAPTED TO CYCLES AND NOT OTHERWISE PROVIDED FOR, e.g. ARTICLE CARRIERS OR CYCLE PROTECTORS
- B62J45/00—Electrical equipment arrangements specially adapted for use as accessories on cycles, not otherwise provided for
- B62J45/40—Sensor arrangements; Mounting thereof
- B62J45/41—Sensor arrangements; Mounting thereof characterised by the type of sensor
- B62J45/415—Inclination sensors
- B62J45/4152—Inclination sensors for sensing longitudinal inclination of the cycle
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62J—CYCLE SADDLES OR SEATS; AUXILIARY DEVICES OR ACCESSORIES SPECIALLY ADAPTED TO CYCLES AND NOT OTHERWISE PROVIDED FOR, e.g. ARTICLE CARRIERS OR CYCLE PROTECTORS
- B62J6/00—Arrangement of optical signalling or lighting devices on cycles; Mounting or supporting thereof; Circuits therefor
- B62J6/22—Warning or information lights
- B62J6/26—Warning or information lights warning or informing other road users, e.g. police flash lights
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62K—CYCLES; CYCLE FRAMES; CYCLE STEERING DEVICES; RIDER-OPERATED TERMINAL CONTROLS SPECIALLY ADAPTED FOR CYCLES; CYCLE AXLE SUSPENSIONS; CYCLE SIDE-CARS, FORECARS, OR THE LIKE
- B62K11/00—Motorcycles, engine-assisted cycles or motor scooters with one or two wheels
- B62K11/007—Automatic balancing machines with single main ground engaging wheel or coaxial wheels supporting a rider
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q2400/00—Special features or arrangements of exterior signal lamps for vehicles
- B60Q2400/50—Projected symbol or information, e.g. onto the road or car body
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/0088—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots characterized by the autonomous decision making process, e.g. artificial intelligence, predefined behaviours
Definitions
- the invention relates to a presentation device that illuminates a projection pattern in accordance with behavior of a vehicle.
- an autonomous mobile device described in Patent Literature 1 includes a monitor that displays a current point of gaze forward in a proceeding direction. A person located in a moving path of the autonomous mobile device can recognize the moving direction of the autonomous mobile device, on the basis of the point of gaze displayed on the display means.
- Patent Literature 1 JP 2004-118469 A
- Patent Literature 1 has a problem that behavior of a vehicle cannot be grasped unless a person located around the vehicle pays attention to display means of the vehicle.
- the warning sound By using a warning sound, behavior of the vehicle can be grasped even if a person around the vehicle does not pay attention to the vehicle. However, as described above, the warning sound becomes noise to other people than the waring target person.
- the present invention is devised to solve the above-described problem, and an object of the invention is to obtain a presentation device that allows a person to grasp behavior of a vehicle, without generating noise, even if the person does not pay attention to the vehicle.
- a presentation device includes: an estimating unit estimating behavior of a vehicle; an illuminating unit illuminating a projection pattern onto ground around the vehicle; and a control unit controlling the illuminating unit to illuminate the projection pattern in accordance with the behavior of the vehicle estimated by the estimating unit.
- the behavior of the vehicle can be grasped, without generating noise, by a person who does not pay attention to the vehicle.
- FIG. 1 is a plan view of a main part showing a vehicle control system including a presentation device according to a first embodiment of the invention.
- FIG. 2 is a side view of the main part showing the vehicle control system including the presentation device according to the first embodiment.
- FIG. 3 is a block diagram showing a configuration of the vehicle control system including the presentation device according to the first embodiment.
- FIG. 4 A is a block diagram showing a hardware configuration that implements functions of the presentation device according to the first embodiment
- FIG. 4 B is a block diagram showing a hardware configuration that executes software that implements the functions of the presentation device according to the first embodiment.
- FIG. 5 is a flowchart showing an operation of the presentation device according to the first embodiment.
- FIG. 6 is a diagram showing exemplary illumination of a projection pattern of the first embodiment.
- FIG. 7 is a diagram showing exemplary illumination of another projection pattern of the first embodiment.
- FIG. 8 is a block diagram showing another configuration of a vehicle control system including a presentation device according to the first embodiment.
- FIG. 9 is a block diagram showing a configuration of a vehicle control system including a presentation device according to a second embodiment of the invention.
- FIG. 10 is a flowchart showing an operation of the presentation device according to the second embodiment.
- FIG. 11 is a diagram showing exemplary illumination of a projection pattern of the second embodiment.
- FIG. 12 is a block diagram showing a configuration of a vehicle control system including a presentation device according to a third embodiment of the invention.
- FIG. 13 is a flowchart showing an operation of the vehicle control system including the presentation device according to the third embodiment.
- FIG. 14 is a diagram showing exemplary illumination of a projection pattern of the third embodiment.
- FIG. 1 is a plan view of a main part showing a vehicle control system including a presentation device according to a first embodiment of the invention.
- FIG. 2 is a side view of the main part showing the vehicle control system including the presentation device according to the first embodiment.
- a vehicle 1 travels autonomously toward a point set in advance, and includes driving wheels 2 a and 2 b and driven wheels 3 a and 3 b .
- the driving wheels 2 a and 2 b are driven to rotate by driving power generated by a motor (not shown) and transmitted to the driving wheels 2 a and 2 b .
- the driven wheels 3 a and 3 b are supported on the bottom of the vehicle 1 to be turnable about an axis in a vertical direction and so as not to hinder the turning operation of the driving wheels 2 a and 2 b.
- Pattern illuminating units 4 a to 4 d each illuminate a projection pattern onto ground G.
- the pattern illuminating units 4 a to 4 d are each formed using projection lamps or lasers.
- the projection pattern is a pattern of visible light illuminated onto the ground around the vehicle 1 , and is implemented by, for example, a graphic pattern of any shape, a pattern of characters, or a pattern of a combination of a graphic pattern and a pattern of characters.
- the pattern illuminating unit 4 a illuminates a projection pattern onto the ground G on the right side of the vehicle 1
- the pattern illuminating unit 4 b illuminates a projection pattern onto the ground G on the left side of the vehicle 1
- the pattern illuminating unit 4 c illuminates a projection pattern onto the ground G ahead of the vehicle 1
- the pattern illuminating unit 4 d illuminates a projection pattern onto the ground G behind the vehicle 1
- An external sensor 5 detects conditions around the vehicle 1 , and is, for example, a camera that detects three-dimensional shapes around the vehicle 1 or captures images of an area around the vehicle 1 . By the external sensor 5 , obstacles ahead of the traveling vehicle 1 are detected.
- FIG. 3 is a block diagram showing a configuration of the vehicle control system including a presentation device A according to the first embodiment.
- the vehicle control system shown in FIG. 3 includes the external sensor 5 , a travel route setting unit 6 , an input unit 7 , a travel control unit 8 , an internal sensor 9 , and the presentation device A.
- the travel route setting unit 6 sets a route on which the vehicle 1 is controlled to travel thereon. For example, the travel route setting unit 6 calculates a route to a destination point through a transit point with avoiding obstacles, on the basis of the transit point, the destination point, the locations of known obstacles, and map information which are received by the input unit 7 .
- the travel control unit 8 controls the rotating drive of the driving wheels 2 a and 2 b in such a manner that the vehicle 1 travels along the route set by the travel route setting unit 6 .
- the internal sensor 9 detects behavior of the vehicle 1 , and detects, for example, the travel speed, acceleration, angular velocity, and angle of inclination of the vehicle 1 .
- the presentation device A includes the pattern illuminating units 4 a to 4 d , a vehicle behavior estimating unit 10 , and an illumination control unit 11 .
- the vehicle behavior estimating unit 10 estimates behavior of the vehicle 1 on the basis of at least one of control information for the vehicle 1 used by the travel control unit 8 , detected information obtained by the external sensor 5 , and detected information obtained by the internal sensor 9 .
- the behavior of the vehicle 1 includes, for example, a collision of the vehicle 1 with an obstacle, inclination of the vehicle 1 , and the rollover, slide-down, skid, and slip of the vehicle 1 associated with acceleration or deceleration.
- the illumination control unit 11 controls the pattern illuminating units 4 a to 4 d to illuminate a projection pattern on a basis of the behavior estimated by the vehicle behavior estimating unit 10 .
- FIG. 4 A is a block diagram showing a hardware configuration that implements the functions of the presentation device A.
- a processing circuit 100 is connected to a light emitting device 101 .
- FIG. 4 B is a block diagram showing a hardware configuration that executes software that implements the functions of the presentation device A.
- a processor 102 and a memory 103 are connected to the light emitting device 101 .
- the pattern illuminating units 4 a to 4 d of the presentation device A are the light emitting device 101 , and are, for example, projection lamps or lasers.
- the presentation device A includes a processing circuit for performing processes at step ST 1 to ST 8 shown in FIG. 5 .
- the processing circuit may be dedicated hardware or may be a Central Processing Unit (CPU) that executes programs stored in a memory.
- CPU Central Processing Unit
- the processing circuit 100 corresponds, for example, to a single circuit, a combined circuit, a programmed processor, a parallel programmed processor, an Application Specific Integrated Circuit (ASIC), a Field-Programmable Gate Array (FPGA), or a combination thereof.
- ASIC Application Specific Integrated Circuit
- FPGA Field-Programmable Gate Array
- the functions of the vehicle behavior estimating unit 10 and the illumination control unit 11 may be implemented by different processing circuits, or the functions may be collectively implemented by a single processing circuit.
- each function of the vehicle behavior estimating unit 10 and the illumination control unit 11 is implemented by software, firmware, or a combination of software and firmware.
- the software or firmware is described as a program and stored in the memory 103 .
- the processor 102 implements the functions of the respective units by reading and executing the programs stored in the memory 103 .
- the presentation device A includes the memory 103 for storing programs that cause the processes at step ST 1 to ST 8 shown in FIG. 5 to be consequently performed when executed by the processor 102 .
- the programs cause a computer to perform procedures or methods for the vehicle behavior estimating unit 10 and the illumination control unit 11 .
- the memory 103 corresponds, for example, to a nonvolatile or volatile semiconductor memory such as a Random Access Memory (RAM), a Read Only Memory (ROM), a flash memory, an Erasable Programmable Read Only Memory (EPROM), or an Electrically-EPROM (EEPROM), a magnetic disk, a flexible disk, an optical disc, a compact disc, a MiniDisc, or a DVD.
- RAM Random Access Memory
- ROM Read Only Memory
- EPROM Erasable Programmable Read Only Memory
- EEPROM Electrically-EPROM
- One of the functions of the vehicle behavior estimating unit 10 and the illumination control unit 11 may be implemented by dedicated hardware, and the other may be implemented by software or firmware.
- the processing circuit 100 serving as dedicated hardware may implement the function of the vehicle behavior estimating unit 10
- the processor 102 may implement the function of the illumination control unit 11 by reading and executing a program stored in the memory 103 .
- the processing circuit can implement each of the above-described functions by hardware, software, firmware, or a combination thereof.
- FIG. 5 is a flowchart showing the operation of the presentation device A, and shows a specific example of a process of illuminating a projection pattern. Note that, it is assumed that, before the operation shown in FIG. 5 , the input unit 7 accepted information required to calculate a travel route, the travel route setting unit 6 calculated a travel route on the basis of the information accepted by the input unit 7 , and the travel control unit 8 caused the vehicle 1 to travel in accordance with the travel route.
- the vehicle behavior estimating unit 10 obtains vehicle's internal and external information detected by the external sensor 5 and the internal sensor 9 (step ST 1 ).
- the vehicle's internal and external information are also outputted to the travel control unit 8 from the external sensor 5 and the internal sensor 9 .
- the travel control unit 8 controls the travel of the vehicle 1 on the basis of the vehicle's internal and external information to avoid collisions of the vehicle 1 with obstacles, inclination of the vehicle 1 , the rollover, slide-down, skid, and slip of the vehicle 1 associated with acceleration or deceleration, etc.
- the vehicle behavior estimating unit 10 estimates behavior of the vehicle 1 to be exhibited from a current time to any certain point in time after the current time (hereinafter, such an estimated behavior is referred to as future behavior of the vehicle 1 ), on the basis of the vehicle's internal and external information (step ST 2 ).
- the vehicle behavior estimating unit 10 estimates behavior of the vehicle 1 controlled by the travel control unit 8 , such as acceleration, deceleration, or a turn, as future behavior of the vehicle 1 .
- the illumination control unit 11 determines whether the future behavior of the vehicle 1 estimated by the vehicle behavior estimating unit 10 is abnormal behavior such as rollover of the vehicle 1 (step ST 3 ).
- abnormal behavior in addition to rollover of the vehicle 1 , a slide-down on a slope or a skid of the vehicle 1 , and a slip of the driving wheels 2 a and 2 b can be considered.
- abnormal behavior of the vehicle 1 is not limited to the above examples, and may be any behavior as long as the behavior has potential to affect an area around the vehicle 1 .
- the illumination control unit 11 sets a projection pattern in accordance with the future behavior of the vehicle 1 , on the pattern illuminating units 4 a to 4 d (step ST 4 ). For example, the illumination control unit 11 selects a projection pattern in accordance with the future behavior of the vehicle 1 estimated by the vehicle behavior estimating unit 10 from table data in which details of behavior are associated with projection patterns, and sets the selected projection pattern on the pattern illuminating units 4 a to 4 d .
- table data information is set that indicates that a first projection pattern corresponds rollover and a second projection pattern corresponds inclination, and such table data is stored in a memory (not shown) included in the presentation device A.
- the pattern illuminating units 4 a to 4 d illuminate the projection pattern set by the illumination control unit 11 onto the ground G (step ST 5 ).
- a person around the vehicle 1 can grasp the future behavior of the vehicle 1 from the projection pattern illuminated onto the ground G, without paying attention to the vehicle 1 .
- the vehicle behavior estimating unit 10 checks whether there is a change in the future behavior of the vehicle 1 , by sequentially obtaining vehicle's internal and external information from the external sensor 5 and the internal sensor 9 and estimating future behavior of the vehicle 1 every time the vehicle's internal and external information are obtained (step ST 6 ).
- step ST 6 If there is no change in the future behavior of the vehicle 1 (step ST 6 ; NO), processing returns to the process at step ST 5 , and the illumination of the projection pattern by the pattern illuminating units 4 a to 4 d continues.
- the illumination control unit 11 determines whether the change in the future behavior of the vehicle 1 has settled (step ST 7 ).
- the illumination control unit 11 determines that the change in the future behavior of the vehicle 1 does not settle. In addition, when the future behavior estimated at step ST 2 has eased, the illumination control unit 11 determines that the change in the future behavior of the vehicle 1 has settled.
- step ST 7 If it is determined that the change in the future behavior of the vehicle 1 does not settle (step ST 7 ; NO), the illumination control unit 11 returns to the process at step ST 4 , and sets a projection pattern on the pattern illuminating units 4 a to 4 d in accordance with the future behavior of the vehicle 1 estimated at step ST 6 .
- the illumination control unit 11 controls the pattern illuminating units 4 a to 4 d to terminate the illumination of the projection pattern.
- the pattern illuminating units 4 a to 4 d terminate the illumination of the projection pattern (step ST 8 ).
- the illumination control unit 11 may allow the pattern illuminating units 4 a to 4 d to continue the illumination of the projection pattern having been illuminated so far, for a preset period of time without changing the projection pattern.
- the vehicle behavior estimating unit 10 determines whether the travel of the vehicle 1 is to be terminated, on the basis of vehicle's internal and external information (step ST 9 ). For example, when the vehicle behavior estimating unit 10 estimates that the travel control of the vehicle 1 is to be terminated on the basis of control information for the vehicle 1 inputted from the travel control unit 8 , the vehicle behavior estimating unit 10 determines that the travel of the vehicle 1 is to be terminated.
- step ST 9 If it is determined by the vehicle behavior estimating unit 10 that the travel of the vehicle 1 is to be terminated (step ST 9 ; YES), the process of FIG. 5 ends.
- step ST 9 If it is not determined by the vehicle behavior estimating unit 10 that the travel of the vehicle 1 is terminated (step ST 9 ; NO), a series of processes from step ST 1 are repeated.
- FIG. 6 is a diagram showing exemplary illumination of a projection pattern.
- there are a person 12 and a person 13 around the vehicle 1 and the person 12 faces in a direction in which he/she can view the vehicle 1 , but the person 13 faces in a direction in which the vehicle 1 is out of his/her sight.
- the pattern illuminating units 4 b and 4 d illuminates a projection pattern 14 onto the ground on the left side of the vehicle 1 and the ground behind the vehicle 1 .
- Contour lines 20 a to 20 h indicates the elevation of the ground, and the position of the contour line 20 a is the lowest and the position of the contour line 20 h is the highest, and the ground gets higher as it gets closer to the contour line 20 h from the contour line 20 a.
- the vehicle behavior estimating unit 10 estimates “rollover of the vehicle 1 to rear-left” as future behavior of the vehicle 1 , on the basis of vehicle's internal and external information.
- the illumination control unit 11 identifies an area around the vehicle 1 expected to be affected by the abnormal behavior of the vehicle 1 which is estimated by the vehicle behavior estimating unit 10 , and controls the pattern illuminating units 4 a to 4 d to illuminate a projection pattern onto the ground in the identified area around the vehicle 1 .
- the illumination control unit 11 selects a projection pattern on the basis of the future behavior “rollover of the vehicle 1 to rear-left”, and further identifies an area on the rear left side of the vehicle 1 expected to be affected by the future behavior, and selects the pattern illuminating units 4 b and 4 d whose illumination regions correspond to the identified area.
- the pattern illuminating unit 4 b illuminates a projection pattern onto the ground on the left side of the vehicle 1
- the pattern illuminating unit 4 d illuminates the projection pattern onto a ground behind the vehicle 1 .
- the person 12 can intuitively recognize the risk of rollover of the vehicle 1 from the projection pattern illuminated onto the ground on the left side of the vehicle 1 .
- the person 13 can, though he/she is facing in the direction in which the vehicle 1 is out of his/her sight, intuitively recognize the risk of rollover of the vehicle 1 from the projection pattern illuminated onto the ground behind the vehicle 1 . By this, it becomes possible for the person 12 and the person 13 to evacuate an area around the vehicle 1 .
- FIG. 6 shows rollover of the vehicle 1 associated with changes in the elevation of the ground
- no limitation is intended by this example.
- various abnormal behavior such as a slide-down or skid of the vehicle 1 and uncontrollable travel due to a slip of the vehicle 1 on flat ground which are caused by a reduction in the friction coefficient of the ground, too, by illuminating a projection pattern onto the ground around the vehicle 1 in the same manner as described above, the behavior of the vehicle can be grasped, without generating noise, by a person who does not pay attention to the vehicle.
- the illumination control unit 11 may control the pattern illuminating units 4 a to 4 d to change the display mode of a projection pattern in accordance with the tendency.
- the illumination control unit 11 changes the flashing interval of a projection pattern. Further, when the amount of increase in the inclination of the vehicle 1 at a set interval is greater than or equal to a threshold value, the illumination control unit 11 may control the pattern illuminating units 4 a to 4 d to increase the intensity of light of a projection pattern step by step or change color to alert color step by step.
- a person present around the vehicle 1 can intuitively recognize the risk of being affected by abnormal behavior of the vehicle 1 , by such a projection pattern.
- the illumination control unit 11 may divide an area around the vehicle 1 expected to be affected by abnormal behavior of the vehicle 1 into a plurality of areas in accordance with a distance from the vehicle 1 , and control the pattern illuminating units 4 a to 4 d to illuminate the plurality of areas with different shapes of the projection pattern, respectively, such that the different shapes of the projection pattern form a gradual pattern.
- the illumination control unit 11 makes the pattern illuminating units 4 a to 4 d illuminate a projection pattern of an acute shaped graphic onto an area close to the vehicle 1 , and continuously alters the shape of the projection pattern to a rounder graphic as it goes away from the vehicle 1 .
- a person present around the vehicle 1 can intuitively recognize risk of being affected by abnormal behavior of the vehicle 1 also by such a projection pattern.
- the projection pattern is a graphic pattern.
- the projection pattern may be a character pattern or may be a combination of a character pattern and a graphic pattern.
- the display mode of the projection pattern corresponds to at least one of the speed of flashing, color, brightness, chroma, intensity, shape, size, addition of character information, and addition of graphic information.
- FIG. 7 is a diagram showing an example of illumination of another projection pattern.
- the pattern illuminating units 4 b and 4 d illuminate projection patterns 14 a to 14 d onto the ground on the left side of the vehicle 1 and the ground behind the vehicle 1 .
- the projection patterns 14 a to 14 d are illuminated onto a plurality of areas set on the basis of the distance from the vehicle 1 .
- the projection pattern 14 a with the highest brightness is illuminated onto an area which is closest to the vehicle 1 and expected to be affected most by abnormal behavior of the vehicle 1 .
- the projection pattern 14 b and the projection pattern 14 c whose brightness decreases step by step as the area goes away from the vehicle 1 are illuminated, and the projection pattern 14 d with the lowest brightness is illuminated onto an area which is farthest from the vehicle 1 and expected to be affected the least by the abnormal behavior of the vehicle 1 .
- a person present around the vehicle 1 can intuitively recognize whether he/she is present in an area that is likely to be affected by the abnormal behavior of the vehicle 1 , by such projection patterns.
- the display mode of a projection pattern to be changed is change of at least one of the speed of flashing, color, brightness, chroma, intensity, shape, size, addition of character information, and addition of graphic information.
- the vehicle 1 is an autonomously moving vehicle without a person on board. Further, the vehicle 1 may be a vehicle whose travel is controlled by remote operations.
- FIG. 8 is a block diagram showing a configuration of a vehicle control system including a presentation device A 1 .
- the presentation device A 1 is mounted on a vehicle 1 A that is remotely operated by a remote control terminal 22 .
- the same components as those of FIG. 3 are denoted by the same reference signs and description thereof is omitted.
- the vehicle control system shown in FIG. 8 includes the external sensor 5 , the travel control unit 8 , the internal sensor 9 , a transmitting and receiving unit 21 , and the presentation device A 1 .
- the transmitting and receiving unit 21 is a component that performs wireless communication or wired communication with the remote control terminal 22 , and receives control information from the remote control terminal 22 .
- the remote control terminal 22 transmits control information in accordance with an operation by an operator 23 to the transmitting and receiving unit 21 .
- the travel control unit 8 controls the rotating drive of the driving wheels 2 a and 2 b in accordance with the control information received by the transmitting and receiving unit 21 , by which the vehicle 1 A travels.
- Vehicle's external information detected by the external sensor 5 is outputted to the travel control unit 8 and the transmitting and receiving unit 21 .
- the transmitting and receiving unit 21 transmits the vehicle's external information inputted from the external sensor 5 to the remote control terminal 22 .
- the operator 23 can perform remote control of the vehicle 1 A with reference to the vehicle's external information received by the remote control terminal 22 . For example, when an obstacle is detected in a proceeding direction of the vehicle 1 A by the external sensor 5 , the vehicle 1 A can be remotely controlled so as to avoid the detected obstacle.
- the presentation device A 1 includes the pattern illuminating units 4 a to 4 d , a vehicle behavior estimating unit 10 A, and the illumination control unit 11 .
- the vehicle behavior estimating unit 10 A is an estimating unit that estimates behavior of the vehicle 1 A, and estimates behavior of the vehicle 1 A on the basis of at least one of the control information for the vehicle 1 obtained from the travel control unit 8 , the vehicle's external information detected by the external sensor 5 , and vehicle's internal information detected by the internal sensor 9 .
- the illumination control unit 11 controls the pattern illuminating units 4 a to 4 d to illuminate a projection pattern in accordance with the behavior of the vehicle 1 A estimated by the vehicle behavior estimating unit 10 A.
- Behavior estimation information of the vehicle 1 A estimated by the vehicle behavior estimating unit 10 A is outputted to the travel control unit 8 , and the travel control unit 8 controls the transmitting and receiving unit 21 to transmit the behavior estimation information of the vehicle 1 A to the remote control terminal 22 .
- the operator 23 can perform remote control of the vehicle 1 A with reference to the behavior estimation information of the vehicle 1 A received by the remote control terminal 22 .
- a person around the vehicle 1 A can intuitively recognize the risk of being affected by abnormal behavior of the vehicle 1 A, by a projection pattern illuminated onto the ground around him/her.
- Each function of the vehicle behavior estimating unit 10 A and the illumination control unit 11 of the presentation device A 1 is implemented by a processing circuit.
- the presentation device A 1 includes a processing circuit for performing these functions.
- the processing circuit may be dedicated hardware or may be a processor that executes programs stored in a memory.
- the presentation devices A and A 1 include the vehicle behavior estimating units 10 and 10 A estimating behavior of the vehicles 1 and 1 A; the pattern illuminating units 4 a to 4 d illuminating a projection pattern onto the ground around the vehicles 1 and 1 A; and the illumination control unit 11 controlling the pattern illuminating units 4 a to 4 d to illuminate a projection pattern in accordance with the behavior of the vehicles 1 and 1 A estimated by the vehicle behavior estimating unit 10 .
- a projection pattern in accordance with behavior of the vehicles 1 and 1 A is illuminated onto the ground around the vehicles 1 and 1 A, and thus, without generating noise, the behavior of the vehicles 1 and 1 A can be grasped by a person who does not pay attention to the vehicles 1 and 1 A.
- the vehicle behavior estimating units 10 and 10 A estimate abnormal behavior of the vehicles 1 and 1 A, on the basis of detected information obtained from the internal sensor 9 detecting the behavior of the vehicles 1 and 1 A. Furthermore, the vehicle behavior estimating units 10 and 10 A estimate abnormal behavior of the vehicles 1 and 1 A, on the basis of detected information obtained from the external sensor 5 detecting conditions around the vehicles 1 and 1 A. By thus using vehicle's internal and external information obtained from the internal sensor 9 and the external sensor 5 , the vehicle behavior estimating units 10 and 10 A can estimate abnormal behavior of the vehicles 1 and 1 A in high accuracy.
- the illumination control unit 11 controls the pattern illuminating units 4 a to 4 d to change the display mode of a projection pattern in accordance with the tendency.
- the display mode of the projection pattern includes at least one of the speed of flashing, color, brightness, chroma, intensity, shape, size, addition of character information, and addition of graphic information.
- the illumination control unit 11 identifies an area around the vehicles 1 and 1 A expected to be affected by abnormal behavior of the vehicles 1 and 1 A, and controls the pattern illuminating units 4 a to 4 d to illuminate a projection pattern onto ground in the identified area around the vehicles 1 and 1 A, the abnormal behavior being estimated by the estimating unit.
- the illumination control unit 11 further divides the area around the vehicles 1 and 1 A expected to be affected by the abnormal behavior of the vehicles 1 and 1 A into a plurality of areas in accordance with a distance from the vehicle 1 , and controls the pattern illuminating units 4 a to 4 d to illuminate the plurality of areas with different modes of the projection pattern, respectively, such that the different modes of the projection pattern form a gradual pattern.
- FIG. 9 is a block diagram showing a configuration of a vehicle control system including a presentation device B according to a second embodiment of the invention.
- the vehicle control system shown in FIG. 9 is mounted on the vehicle 1 A that is remotely controlled using the remote control terminal 22 , and includes the external sensor 5 , the travel control unit 8 , the internal sensor 9 , the transmitting and receiving unit 21 , a communication state monitoring unit 24 , and the presentation device B.
- the communication state monitoring unit 24 is a component that monitors a communication state between the transmitting and receiving unit 21 and the remote control terminal 22 .
- the communication state monitoring unit 24 monitors a communication state between the transmitting and receiving unit 21 and the remote control terminal 22 every certain period of time.
- the presentation device B includes the pattern illuminating units 4 a to 4 d , the vehicle behavior estimating unit 10 A, and an illumination control unit 11 A.
- the illumination control unit 11 A controls, as in the first embodiment, the pattern illuminating units 4 a to 4 d to illuminate a projection pattern in accordance with behavior estimated by the vehicle behavior estimating unit 10 A.
- the illumination control unit 11 A controls the pattern illuminating units 4 a to 4 d to illuminate a projection pattern.
- Each function of the vehicle behavior estimating unit 10 A and the illumination control unit 11 A of the presentation device B is implemented by a processing circuit.
- the presentation device B includes a processing circuit for performing these functions.
- the processing circuit may be dedicated hardware or may be a processor that executes programs stored in a memory.
- FIG. 10 is a flowchart showing an operation of the presentation device B, and shows a specific example of a process of illuminating a projection pattern. Note that prior to the operation shown in FIG. 10 , travel of the vehicle 1 A is controlled by the remote control terminal 22 .
- the illumination control unit 11 A obtains a result of monitoring the communication state from the communication state monitoring unit 24 (step ST 1 a ). For example, every time the communication state monitoring unit 24 monitors a communication state every certain period of time, the illumination control unit 11 A obtains a result of the monitoring.
- the illumination control unit 11 A determines whether normal communication can be performed between the transmitting and receiving unit 21 and the remote control terminal 22 , on the basis of the result of monitoring a communication state obtained from the communication state monitoring unit 24 (step ST 2 a ). If normal communication can be performed (step ST 2 a ; YES), processing returns to the process at step ST 1 a , and monitoring of the communication state continues.
- the illumination control unit 11 A determines that control of the vehicle 1 A by the remote control terminal 22 has become impossible, and sets a projection pattern corresponding to the state that remote control is impossible, on the pattern illuminating units 4 a to 4 d (step ST 3 a ).
- the pattern illuminating units 4 a to 4 d illuminate the projection pattern set by the illumination control unit 11 A onto the ground around the vehicle 1 A (step ST 4 a ).
- a person around the vehicle 1 A can grasp the fact that remote control of the vehicle 1 A is impossible by a projection pattern, so that it is possible to prompt the person to evacuate an area around the vehicle 1 A.
- FIG. 11 is a diagram showing an example of illumination of a projection pattern.
- the person 12 and the person 13 around the vehicle 1 A there are the person 12 and the person 13 around the vehicle 1 A, and the person 12 faces in a direction in which he/she can view the vehicle 1 A, but the person 13 faces in a direction in which the vehicle 1 A is out of his/her sight.
- the pattern illuminating units 4 a to 4 d illuminate a projection pattern 14 onto the ground around the vehicle 1 A. Since the vehicle 1 A is proceeding in the direction of the arrow A, the illumination control unit 11 A may control the pattern illuminating unit 4 c to highlight a projection pattern illuminated onto the ground ahead of the vehicle 1 A.
- the person 12 can intuitively recognize the fact that the vehicle 1 A is uncontrollable, by the illumination of the projection pattern onto the ground ahead of the vehicle 1 A.
- the person 13 can, though he/she faces in the direction in which the vehicle 1 A itself is out of his/her sight, intuitively recognize the fact that the vehicle 1 A is uncontrollable, by the illumination of the projection pattern onto the ground ahead of the vehicle 1 A.
- the illumination control unit 11 A makes the pattern illuminating units 4 a to 4 d illuminate a projection pattern.
- FIG. 12 is a block diagram showing a configuration of a vehicle control system including a presentation device C according to a third embodiment of the invention.
- the vehicle control system shown in FIG. 12 is mounted on the vehicle 1 , and includes the external sensor 5 , the travel route setting unit 6 , the input unit 7 , the travel control unit 8 , the internal sensor 9 , and the presentation device C.
- the presentation device C includes the pattern illuminating units 4 a to 4 d , the vehicle behavior estimating unit 10 , and an illumination control unit 11 B.
- the illumination control unit 11 B controls, as in the first embodiment, the pattern illuminating units 4 a to 4 d to illuminate a projection pattern in accordance with behavior estimated by the vehicle behavior estimating unit 10 . Furthermore, when the vehicle behavior estimating unit 10 estimates that the vehicle 1 goes into a deadlock state in which the vehicle 1 cannot move in any direction, the illumination control unit 11 B makes the pattern illuminating units 4 a to 4 d illuminate a projection pattern in accordance with the result of the estimation.
- Each function of the vehicle behavior estimating unit 10 and the illumination control unit 11 B of the presentation device C is implemented by a processing circuit.
- the presentation device C includes a processing circuit for performing these functions.
- the processing circuit may be dedicated hardware or may be a processor that executes programs stored in a memory.
- FIG. 13 is a flowchart showing the operation of the vehicle control system including the presentation device C. Note that prior to the operation shown in FIG. 13 , the input unit 7 accepts information required to calculate a travel route, the travel route setting unit 6 calculates a travel route on the basis of the information accepted by the input unit 7 , and the travel control unit 8 makes the vehicle 1 travel in accordance with the travel route.
- the travel control unit 8 and the vehicle behavior estimating unit 10 obtain vehicle's internal and external information detected by the external sensor 5 and the internal sensor 9 (step ST 1 b ).
- the travel control unit 8 determines whether there is an obstacle around the vehicle 1 , on the basis of the vehicle's external information detected by the external sensor 5 (step ST 2 b ).
- step ST 2 b If the travel control unit 8 determines that there is an obstacle on the travel route of the vehicle 1 (step ST 2 b ; YES), the travel control unit 8 instructs the travel route setting unit 6 to calculate a route to avoid the obstacle.
- the travel control unit 8 checks whether the route can be changed to a route to avoid the obstacle (step ST 3 b ).
- the travel control unit 8 changes the travel route of the vehicle 1 to the above-described route calculated by the travel route setting unit 6 (step ST 4 b ).
- step ST 2 b If it is determined that there is no obstacle on the travel route of the vehicle 1 (step ST 2 b ; NO) or if the process at step ST 4 b is completed, the vehicle behavior estimating unit 10 determines whether the travel of the vehicle 1 is to be terminated, on the basis of vehicle's internal and external information (step ST 5 b ).
- the vehicle behavior estimating unit 10 estimates that the vehicle 1 approaches a destination point and the travel control by the travel control unit 8 is terminated, the vehicle behavior estimating unit 10 determines that the travel of the vehicle 1 is to be terminated.
- step ST 5 b If it is determined by the vehicle behavior estimating unit 10 that the travel of the vehicle 1 is to be terminated (step ST 5 b ; YES), the process of FIG. 13 ends.
- step ST 5 b If it is not determined by the vehicle behavior estimating unit 10 that the travel of the vehicle 1 is to be terminated (step ST 5 b ; NO), a series of processes from step ST 1 b is repeated.
- the travel control unit 8 determines that the vehicle 1 is in a so-called deadlock state in which the vehicle 1 cannot move in any direction, and stops the vehicle 1 (step ST 6 b ).
- the vehicle behavior estimating unit 10 estimates that the vehicle 1 is to be in a deadlock state, on the basis of control information for the vehicle 1 inputted from the travel control unit 8 .
- the illumination control unit 11 B sets a projection pattern corresponding to the future behavior that the vehicle 1 is to be in a deadlock state, on the pattern illuminating units 4 a to 4 d (step ST 7 b ).
- the pattern illuminating units 4 a to 4 d illuminate the projection pattern set by the illumination control unit 11 B onto the ground around the vehicle 1 (step ST 8 b ).
- a person around the vehicle 1 can grasp that the vehicle 1 is in a deadlock state, by visually recognizing the projection pattern illuminated onto the ground.
- the travel control unit 8 determines whether condition around the vehicle 1 is changed, on the basis of vehicle's external information (step ST 9 b ). For example, when an obstacle is detected around the vehicle 1 by the external sensor 5 , it is determined that the condition around the vehicle 1 is changed.
- step ST 9 b If it is determined that the condition around the vehicle 1 is not changed (step ST 9 b ; NO), processing returns to step ST 8 b , and the illumination of the projection pattern continues.
- step ST 9 b If it is determined that the condition around the vehicle is changed (step ST 9 b ; YES), the travel control unit 8 instructs the travel route setting unit 6 to calculate a route on which the vehicle 1 can move. Namely, the travel control unit 8 checks whether the route can be changed (step ST 10 b ).
- step ST 10 b If the travel route setting unit 6 was able to calculate a route, i.e., the route can be changed (step ST 10 b ; YES), the illumination control unit 11 B controls the pattern illuminating units 4 a to 4 d to terminate the illumination of the projection pattern (step ST 11 b ). Thereafter, processing transitions to the process at step ST 4 b .
- step ST 10 b if the travel route setting unit 6 was not able to calculate a route, i.e. the route cannot be changed (step ST 10 b ; NO), processing returns to step ST 7 b , and the illumination of the projection pattern continues.
- FIG. 14 is a diagram showing an example of illumination of a projection pattern.
- there are obstacles 31 and 32 such as walls, around the vehicle 1 , and furthermore, there are persons 33 a to 33 d .
- the person 33 a and the person 33 c each face in a direction in which the vehicle 1 is out of their sights.
- the pattern illuminating units 4 a to 4 d are illuminating projection patterns 34 a to 34 c onto the ground around the vehicle 1 .
- the projection patterns 34 a to 34 c are illuminated onto a plurality of areas which respectively correspond to the distances from the vehicle 1 .
- the projection pattern 34 a with the highest brightness is illuminated onto an area closest to the vehicle 1 .
- the projection pattern 34 b whose brightness decreases step by step as the area goes away from the vehicle 1 is illuminated, and the projection pattern 34 c with the lowest brightness is illuminated onto an area farthest from the vehicle 1 .
- the persons 33 a to 33 d present around the vehicle 1 can intuitively recognize the deadlock state of the vehicle 1 by these projection patterns, and as a result, the persons 33 a to 33 d are prompted to evacuate an area around the vehicle 1 . If the person 33 c and the person 33 d evacuate, then the vehicle 1 can move forward, and if the person 33 a and the person 33 b evacuate, then the vehicle 1 can move backward, by which the deadlock state is eliminated.
- the illumination control unit 11 B makes the pattern illuminating units 4 a to 4 d illuminate a projection pattern corresponding to a deadlock state of the vehicle 1 estimated by the vehicle behavior estimating unit 10 .
- the presentation device C By configuring the presentation device C in this manner, a person around the vehicle 1 can intuitively recognize the deadlock state of the vehicle 1 by the projection pattern.
- the presentation device allow to grasp behavior of a vehicle, without generating noise, by a person who does not pay attention to the vehicle, and thus, is applicable to, for example, vehicle control systems.
- 1 and 1 A Vehicle, 2 a and 2 b : Driving wheel, 3 a and 3 b : Driven wheel, 4 a to 4 d : Pattern illuminating unit, 5 : External sensor, 6 : Travel route setting unit, 7 : Input unit, 8 : Travel control unit, 9 : Internal sensor, 10 and 10 A: Vehicle behavior estimating unit, 11 , 11 A, and 11 B: Illumination control unit, 12 , 13 , and 33 a to 33 d : Person, 14 , 14 a to 14 d , and 34 a to 34 c : Projection pattern, 20 a to 20 h : Contour line, 21 : Transmitting and receiving unit, 22 : Remote control terminal, 23 : Operator, 24 : Communication state monitoring unit, 31 and 32 : Obstacle, A, A 1 , B, and C: Presentation device, 100 : Processing circuit, 101 : Light emitting device, 102 : Processor, and 103 : Memory.
Landscapes
- Engineering & Computer Science (AREA)
- Mechanical Engineering (AREA)
- Lighting Device Outwards From Vehicle And Optical Signal (AREA)
Abstract
Description
Claims (7)
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017-111276 | 2017-06-06 | ||
JPJP2017-111276 | 2017-06-06 | ||
JP2017111276A JP6877247B2 (en) | 2017-06-06 | 2017-06-06 | Presentation device |
PCT/JP2017/036977 WO2018225281A1 (en) | 2017-06-06 | 2017-10-12 | Presentation device |
Publications (2)
Publication Number | Publication Date |
---|---|
US20200164794A1 US20200164794A1 (en) | 2020-05-28 |
US11518300B2 true US11518300B2 (en) | 2022-12-06 |
Family
ID=64567058
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/613,004 Active 2037-10-25 US11518300B2 (en) | 2017-06-06 | 2017-10-12 | Presentation device |
Country Status (5)
Country | Link |
---|---|
US (1) | US11518300B2 (en) |
JP (1) | JP6877247B2 (en) |
CN (1) | CN110691716B (en) |
DE (1) | DE112017007499B4 (en) |
WO (1) | WO2018225281A1 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10676022B2 (en) * | 2017-12-27 | 2020-06-09 | X Development Llc | Visually indicating vehicle caution regions |
WO2021039612A1 (en) * | 2019-08-26 | 2021-03-04 | 株式会社小糸製作所 | Area marker lamp |
DE102020126251A1 (en) | 2020-10-07 | 2022-04-07 | Audi Aktiengesellschaft | Motor vehicle and method for operating an illumination device of a motor vehicle |
GB2614060B (en) * | 2021-12-17 | 2023-12-20 | Hilo Ev Ltd | Vehicle with lighting system |
Citations (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS59206533A (en) | 1983-05-11 | 1984-11-22 | Caterpillar Mitsubishi Ltd | Preventer for falling down of remote-operable vehicle |
JPH04316415A (en) | 1991-04-17 | 1992-11-06 | Ishikawajima Shibaura Mach Co Ltd | Controlling apparatus for automatic traveling working vehicle |
US5529138A (en) * | 1993-01-22 | 1996-06-25 | Shaw; David C. H. | Vehicle collision avoidance system |
JPH09160639A (en) | 1995-12-12 | 1997-06-20 | Mitsubishi Electric Corp | Traveling controller for vehicle |
US6023220A (en) * | 1997-05-30 | 2000-02-08 | Daimlerchrysler Ag | Vehicle having a scanning system |
JP2002079896A (en) | 2000-09-06 | 2002-03-19 | Auto Network Gijutsu Kenkyusho:Kk | Traveling state announcing device |
US20030147247A1 (en) * | 2002-02-07 | 2003-08-07 | Toyota Jidosha Kabushiki Kaisha | Vehicle operation supporting device and vehicle operation supporting system |
JP2003285685A (en) | 2002-03-27 | 2003-10-07 | Toyota Motor Corp | Vehicle operation-supporting system |
JP2004118469A (en) | 2002-09-25 | 2004-04-15 | Matsushita Electric Works Ltd | Autonomous moving device |
JP2007295033A (en) | 2006-04-20 | 2007-11-08 | Toyota Motor Corp | Remote operation controller and operation terminal thereof |
EP1916154A1 (en) | 2006-10-26 | 2008-04-30 | Bayerische Motoren Werke Aktiengesellschaft | Method for displaying information |
US20080175012A1 (en) * | 2006-11-17 | 2008-07-24 | Kabushiki Kaisha Toyota Chuo Kenkyusho | Alerting illumination device |
JP2010018165A (en) | 2008-07-10 | 2010-01-28 | Panasonic Corp | Vehicular illumination device and method of lighting |
US20100253594A1 (en) * | 2009-04-02 | 2010-10-07 | Gm Global Technology Operations, Inc. | Peripheral salient feature enhancement on full-windshield head-up display |
JP2013147083A (en) | 2012-01-18 | 2013-08-01 | Hino Motors Ltd | Behavior transmitting device for vehicle |
WO2016027312A1 (en) | 2014-08-19 | 2016-02-25 | 三菱電機株式会社 | Rear road surface irradiation device |
WO2016027314A1 (en) | 2014-08-19 | 2016-02-25 | 三菱電機株式会社 | Road surface illumination device |
US20160059771A1 (en) * | 2013-06-24 | 2016-03-03 | Bayerische Motoren Werke Aktiengesellschaft | Motor Vehicle |
US20160090025A1 (en) * | 2014-09-30 | 2016-03-31 | Fuji Jukogyo Kabushiki Kaisha | Vehicle headlight device |
WO2016114048A1 (en) | 2015-01-13 | 2016-07-21 | 日立マクセル株式会社 | Image projection device |
US20160257243A1 (en) * | 2011-12-07 | 2016-09-08 | Sl Corporation | Automotive lamp |
US20160259034A1 (en) * | 2015-03-04 | 2016-09-08 | Panasonic Intellectual Property Management Co., Ltd. | Position estimation device and position estimation method |
US20170021768A1 (en) * | 2015-07-22 | 2017-01-26 | Lg Electronics Inc. | Vehicle control device and vehicle control method thereof |
DE102015115242A1 (en) | 2015-09-10 | 2017-03-16 | Hella Kgaa Hueck & Co. | Vehicle with light projection system and method for creating a safety area on a ground surface |
US20170101147A1 (en) | 2015-10-13 | 2017-04-13 | Suzuki Motor Corporation | Vehicle conspicuity/signaling system |
WO2017073635A1 (en) | 2015-10-27 | 2017-05-04 | 株式会社小糸製作所 | Vehicle component, and vehicle |
US20170151906A1 (en) * | 2014-03-27 | 2017-06-01 | Nissan Motor Co., Ltd. | Vehicle Behavior Transmission Device |
US20170203685A1 (en) * | 2014-08-19 | 2017-07-20 | Mitsubishi Electric Corporation | Road surface illumination apparatus |
US20170259731A1 (en) * | 2011-07-08 | 2017-09-14 | Sl Corporation | Automotive lamp |
US20180029641A1 (en) * | 2016-08-01 | 2018-02-01 | Magna Electronics Inc. | Parking assist system using light projections |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007216747A (en) * | 2006-02-15 | 2007-08-30 | Fujitsu Ten Ltd | Vehicle rolling detection device and occupant crash protection system |
-
2017
- 2017-06-06 JP JP2017111276A patent/JP6877247B2/en active Active
- 2017-10-12 US US16/613,004 patent/US11518300B2/en active Active
- 2017-10-12 WO PCT/JP2017/036977 patent/WO2018225281A1/en active Application Filing
- 2017-10-12 DE DE112017007499.1T patent/DE112017007499B4/en active Active
- 2017-10-12 CN CN201780091340.1A patent/CN110691716B/en active Active
Patent Citations (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS59206533A (en) | 1983-05-11 | 1984-11-22 | Caterpillar Mitsubishi Ltd | Preventer for falling down of remote-operable vehicle |
JPH04316415A (en) | 1991-04-17 | 1992-11-06 | Ishikawajima Shibaura Mach Co Ltd | Controlling apparatus for automatic traveling working vehicle |
US5529138A (en) * | 1993-01-22 | 1996-06-25 | Shaw; David C. H. | Vehicle collision avoidance system |
JPH09160639A (en) | 1995-12-12 | 1997-06-20 | Mitsubishi Electric Corp | Traveling controller for vehicle |
US6023220A (en) * | 1997-05-30 | 2000-02-08 | Daimlerchrysler Ag | Vehicle having a scanning system |
JP2002079896A (en) | 2000-09-06 | 2002-03-19 | Auto Network Gijutsu Kenkyusho:Kk | Traveling state announcing device |
US20030147247A1 (en) * | 2002-02-07 | 2003-08-07 | Toyota Jidosha Kabushiki Kaisha | Vehicle operation supporting device and vehicle operation supporting system |
JP2003285685A (en) | 2002-03-27 | 2003-10-07 | Toyota Motor Corp | Vehicle operation-supporting system |
JP2004118469A (en) | 2002-09-25 | 2004-04-15 | Matsushita Electric Works Ltd | Autonomous moving device |
JP2007295033A (en) | 2006-04-20 | 2007-11-08 | Toyota Motor Corp | Remote operation controller and operation terminal thereof |
EP1916154A1 (en) | 2006-10-26 | 2008-04-30 | Bayerische Motoren Werke Aktiengesellschaft | Method for displaying information |
US20080175012A1 (en) * | 2006-11-17 | 2008-07-24 | Kabushiki Kaisha Toyota Chuo Kenkyusho | Alerting illumination device |
JP2010018165A (en) | 2008-07-10 | 2010-01-28 | Panasonic Corp | Vehicular illumination device and method of lighting |
US20100253594A1 (en) * | 2009-04-02 | 2010-10-07 | Gm Global Technology Operations, Inc. | Peripheral salient feature enhancement on full-windshield head-up display |
US20170259731A1 (en) * | 2011-07-08 | 2017-09-14 | Sl Corporation | Automotive lamp |
US20160257243A1 (en) * | 2011-12-07 | 2016-09-08 | Sl Corporation | Automotive lamp |
JP2013147083A (en) | 2012-01-18 | 2013-08-01 | Hino Motors Ltd | Behavior transmitting device for vehicle |
US20160059771A1 (en) * | 2013-06-24 | 2016-03-03 | Bayerische Motoren Werke Aktiengesellschaft | Motor Vehicle |
US20170151906A1 (en) * | 2014-03-27 | 2017-06-01 | Nissan Motor Co., Ltd. | Vehicle Behavior Transmission Device |
US20170267167A1 (en) * | 2014-08-19 | 2017-09-21 | Mitsubishi Electric Corporation | Road surface illumination apparatus |
US20170259734A1 (en) | 2014-08-19 | 2017-09-14 | Mitsubishi Electric Corporation | Rear-road surface illumination apparatus |
US20170203685A1 (en) * | 2014-08-19 | 2017-07-20 | Mitsubishi Electric Corporation | Road surface illumination apparatus |
WO2016027312A1 (en) | 2014-08-19 | 2016-02-25 | 三菱電機株式会社 | Rear road surface irradiation device |
WO2016027314A1 (en) | 2014-08-19 | 2016-02-25 | 三菱電機株式会社 | Road surface illumination device |
JP6300938B2 (en) | 2014-08-19 | 2018-03-28 | 三菱電機株式会社 | Road surface irradiation device |
JP6214775B2 (en) | 2014-08-19 | 2017-10-18 | 三菱電機株式会社 | Rear road surface irradiation device |
US20160090025A1 (en) * | 2014-09-30 | 2016-03-31 | Fuji Jukogyo Kabushiki Kaisha | Vehicle headlight device |
US20180004020A1 (en) | 2015-01-13 | 2018-01-04 | Hitachi Maxell, Ltd. | Image projection apparatus |
WO2016114048A1 (en) | 2015-01-13 | 2016-07-21 | 日立マクセル株式会社 | Image projection device |
US20160259034A1 (en) * | 2015-03-04 | 2016-09-08 | Panasonic Intellectual Property Management Co., Ltd. | Position estimation device and position estimation method |
US20170021768A1 (en) * | 2015-07-22 | 2017-01-26 | Lg Electronics Inc. | Vehicle control device and vehicle control method thereof |
DE102015115242A1 (en) | 2015-09-10 | 2017-03-16 | Hella Kgaa Hueck & Co. | Vehicle with light projection system and method for creating a safety area on a ground surface |
JP2017074820A (en) | 2015-10-13 | 2017-04-20 | スズキ株式会社 | Notification system for vehicle |
US20170101147A1 (en) | 2015-10-13 | 2017-04-13 | Suzuki Motor Corporation | Vehicle conspicuity/signaling system |
US20180319320A1 (en) | 2015-10-27 | 2018-11-08 | Koito Manufacturing Co., Ltd. | Vehicle component, and vehicle |
WO2017073635A1 (en) | 2015-10-27 | 2017-05-04 | 株式会社小糸製作所 | Vehicle component, and vehicle |
US20180029641A1 (en) * | 2016-08-01 | 2018-02-01 | Magna Electronics Inc. | Parking assist system using light projections |
Non-Patent Citations (5)
Title |
---|
Chinese Office Action dated Aug. 1, 2022 issued in corresponding Chinese Patent Application No. 201780091340.1 with an English Translation. |
German Office Action issued in corresponding German Patent Application No. 11 2017 007 499.1 dated Jul. 31, 2020. |
Indian Office Action dated Mar. 31, 2021 in corresponding Indian Patent Application No. 2019 4704 9138. |
International Search Report, issued in PCT/JP2017/036977, PCT/ISA/210, dated Jan. 9, 2018. |
Japanese Office Action, dated Jul. 21, 2020, for Japanese Application No. 2017-111276, with an English machine translation. |
Also Published As
Publication number | Publication date |
---|---|
CN110691716B (en) | 2023-03-14 |
DE112017007499T5 (en) | 2020-04-02 |
JP6877247B2 (en) | 2021-05-26 |
DE112017007499B4 (en) | 2021-03-11 |
CN110691716A (en) | 2020-01-14 |
US20200164794A1 (en) | 2020-05-28 |
WO2018225281A1 (en) | 2018-12-13 |
JP2018203081A (en) | 2018-12-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11518300B2 (en) | Presentation device | |
CN108136959B (en) | Vehicle state display system | |
US9902399B2 (en) | Vehicle travelling control device for controlling a vehicle in traffic | |
CN107807634B (en) | Driving assistance device for vehicle | |
US9649936B2 (en) | In-vehicle device, control method of in-vehicle device, and computer-readable storage medium | |
KR101760995B1 (en) | Driving support apparatus for vehicle | |
US9159235B2 (en) | Method of operating a driver assistance system of a motor vehicle | |
EP3533665B1 (en) | Apparatus that presents result of recognition of recognition target | |
CN108698551B (en) | Road information recognition system and road information recognition method | |
KR20090125075A (en) | Semiautomatic parking machine | |
EP3040681A1 (en) | Vehicular driving guidance device and method | |
US10600323B2 (en) | Vehicle external notification device | |
US11498482B2 (en) | Notification device | |
EP3825209B1 (en) | Apparatus for displaying steering information of preceding vehicle and method thereof | |
US11299162B2 (en) | Vehicle control device | |
US10946791B2 (en) | Out-of-vehicle notification device | |
US10759329B2 (en) | Out-of-vehicle notification device | |
TW201930116A (en) | Warning system adapted to a vehicle | |
CN107004358A (en) | The method for aiding in the vehicle particularly driver of passenger vehicle | |
US10118642B2 (en) | Method and system of assisting a driver of a vehicle | |
JP7160706B2 (en) | road recognition device | |
US11491986B2 (en) | Moving body control apparatus, moving body, and moving body control method | |
KR20220073898A (en) | Apparatus for controlling a vehicle and method thereof | |
US20200062169A1 (en) | Controlling a controllable headlight of a motor vehicle | |
US20230158889A1 (en) | Vehicle |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: MITSUBISHI ELECTRIC CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KASUGA, KEI;MATSUBARA, TSUTOMU;SAKATA, REIKO;AND OTHERS;SIGNING DATES FROM 20190904 TO 20190924;REEL/FRAME:050997/0302 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |