WO2015166811A1 - 自動運転車及び自動運転車用プログラム - Google Patents
自動運転車及び自動運転車用プログラム Download PDFInfo
- Publication number
- WO2015166811A1 WO2015166811A1 PCT/JP2015/061697 JP2015061697W WO2015166811A1 WO 2015166811 A1 WO2015166811 A1 WO 2015166811A1 JP 2015061697 W JP2015061697 W JP 2015061697W WO 2015166811 A1 WO2015166811 A1 WO 2015166811A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- driving vehicle
- behavior
- user
- getting
- driver
- Prior art date
Links
- 230000006399 behavior Effects 0.000 claims description 218
- 238000004891 communication Methods 0.000 claims description 44
- 210000000554 iris Anatomy 0.000 claims description 7
- 210000003462 vein Anatomy 0.000 claims description 5
- 230000002452 interceptive effect Effects 0.000 claims description 4
- 230000001815 facial effect Effects 0.000 claims description 2
- 230000006870 function Effects 0.000 description 69
- 238000012545 processing Methods 0.000 description 66
- 238000000034 method Methods 0.000 description 36
- 230000008569 process Effects 0.000 description 29
- 230000033001 locomotion Effects 0.000 description 24
- 238000001514 detection method Methods 0.000 description 19
- 230000004044 response Effects 0.000 description 17
- 230000004308 accommodation Effects 0.000 description 6
- 238000012423 maintenance Methods 0.000 description 5
- 230000002093 peripheral effect Effects 0.000 description 4
- 125000002066 L-histidyl group Chemical group [H]N1C([H])=NC(C([H])([H])[C@](C(=O)[*])([H])N([H])[H])=C1[H] 0.000 description 3
- 230000008859 change Effects 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 238000012790 confirmation Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000029305 taxis Effects 0.000 description 2
- 206010010774 Constipation Diseases 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000004397 blinking Effects 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 239000000446 fuel Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 235000012054 meals Nutrition 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 230000002485 urinary effect Effects 0.000 description 1
- 210000002700 urine Anatomy 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/0088—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots characterized by the autonomous decision making process, e.g. artificial intelligence, predefined behaviours
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60N—SEATS SPECIALLY ADAPTED FOR VEHICLES; VEHICLE PASSENGER ACCOMMODATION NOT OTHERWISE PROVIDED FOR
- B60N2/00—Seats specially adapted for vehicles; Arrangement or mounting of seats in vehicles
- B60N2/002—Seats provided with an occupancy detection means mounted therein or thereon
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60N—SEATS SPECIALLY ADAPTED FOR VEHICLES; VEHICLE PASSENGER ACCOMMODATION NOT OTHERWISE PROVIDED FOR
- B60N2/00—Seats specially adapted for vehicles; Arrangement or mounting of seats in vehicles
- B60N2/002—Seats provided with an occupancy detection means mounted therein or thereon
- B60N2/0021—Seats provided with an occupancy detection means mounted therein or thereon characterised by the type of sensor or measurement
- B60N2/0024—Seats provided with an occupancy detection means mounted therein or thereon characterised by the type of sensor or measurement for identifying, categorising or investigation of the occupant or object on the seat
- B60N2/0025—Seats provided with an occupancy detection means mounted therein or thereon characterised by the type of sensor or measurement for identifying, categorising or investigation of the occupant or object on the seat by using weight measurement
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60N—SEATS SPECIALLY ADAPTED FOR VEHICLES; VEHICLE PASSENGER ACCOMMODATION NOT OTHERWISE PROVIDED FOR
- B60N2/00—Seats specially adapted for vehicles; Arrangement or mounting of seats in vehicles
- B60N2/002—Seats provided with an occupancy detection means mounted therein or thereon
- B60N2/0021—Seats provided with an occupancy detection means mounted therein or thereon characterised by the type of sensor or measurement
- B60N2/003—Seats provided with an occupancy detection means mounted therein or thereon characterised by the type of sensor or measurement characterised by the sensor mounting location in or on the seat
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R25/00—Fittings or systems for preventing or indicating unauthorised use or theft of vehicles
- B60R25/20—Means to switch the anti-theft system on or off
- B60R25/23—Means to switch the anti-theft system on or off using manual input of alphanumerical codes
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R25/00—Fittings or systems for preventing or indicating unauthorised use or theft of vehicles
- B60R25/20—Means to switch the anti-theft system on or off
- B60R25/25—Means to switch the anti-theft system on or off using biometry
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/3407—Route searching; Route guidance specially adapted for specific applications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/3407—Route searching; Route guidance specially adapted for specific applications
- G01C21/3438—Rendez-vous, i.e. searching a destination where several users can meet, and the routes to this destination for these users; Ride sharing, i.e. searching a route such that at least two users can share a vehicle for at least part of the route
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0283—Price estimation or determination
- G06Q30/0284—Time or distance, e.g. usage of parking meters or taximeters
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0645—Rental transactions; Leasing transactions
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/123—Traffic control systems for road vehicles indicating the position of vehicles, e.g. scheduled vehicles; Managing passenger vehicles circulating according to a fixed timetable, e.g. buses, trains, trams
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60N—SEATS SPECIALLY ADAPTED FOR VEHICLES; VEHICLE PASSENGER ACCOMMODATION NOT OTHERWISE PROVIDED FOR
- B60N2210/00—Sensor types, e.g. for passenger detection systems or for controlling seats
- B60N2210/10—Field detection presence sensors
- B60N2210/16—Electromagnetic waves
- B60N2210/22—Optical; Photoelectric; Lidar [Light Detection and Ranging]
- B60N2210/24—Cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60N—SEATS SPECIALLY ADAPTED FOR VEHICLES; VEHICLE PASSENGER ACCOMMODATION NOT OTHERWISE PROVIDED FOR
- B60N2210/00—Sensor types, e.g. for passenger detection systems or for controlling seats
- B60N2210/40—Force or pressure sensors
Definitions
- This invention relates to an autonomous vehicle and a program for an autonomous vehicle.
- the next location where the user A starts using the autonomous driving vehicle is estimated, and if the estimated location is the current position, If the estimated location is different from the current location, the user moves to the estimated location and waits.
- Patent Document 1 it is necessary for the user to register and store schedule information including at least date and time and location designation in the autonomous driving vehicle in advance, which is troublesome and actually gets on the vehicle. There is a problem that the inside user cannot give instructions to the autonomous driving vehicle in a flexible manner according to the behavior mode and the purpose of the behavior after getting off.
- An object of the present invention is to provide an autonomous vehicle that can solve the above-described problems.
- the present invention provides: It has an automatic driving mode for autonomous driving, Behavior setting accepting means for accepting the setting of behavior after getting out of the autonomous driving vehicle input when the driver or passenger gets off; Control means for controlling execution of the behavior based on the setting;
- the driver or the rider can set the behavior of the autonomous driving vehicle after getting off.
- a user who is in the self-driving vehicle can give instructions to the self-driving vehicle flexibly according to his / her behavior and purpose after getting off. . *
- FIG. 1 is a block diagram illustrating a hardware configuration example of an electronic control circuit unit 10 of an autonomous driving vehicle 1 according to an embodiment of the present invention.
- the automatic driving vehicle 1 of this embodiment is an example in the case of an electric vehicle.
- the battery is not shown in FIG.
- the automatic driving vehicle 1 of this embodiment has an automatic driving mode and a manual driving mode.
- the manual operation mode is a mode in which the vehicle can travel according to the driver's accelerator pedal operation, brake pedal operation, shift lever operation, and steering operation (handle operation), as in a normal automobile that is not an autonomous driving vehicle.
- the autonomous driving vehicle 1 itself automatically (autonomously) avoids obstacles without having to operate the accelerator pedal, the brake pedal, the shift lever, and the steering operation. This is a mode for driving with changes.
- the driver of the automatic driving vehicle 1 can switch the automatic driving vehicle 1 that is traveling in the manual driving mode to the automatic driving mode by, for example, a predetermined operation through the touch panel 112 described later, and travels in the automatic driving mode.
- a predetermined operation through the touch panel 112 described later
- the driver performs an accelerator pedal operation, a brake pedal operation, a shift lever operation, or a steering operation, it is configured to automatically return to the manual operation mode.
- the electronic control circuit unit 10 has a wireless communication unit 102, a motor drive control unit 103, and a steering drive control unit through a system bus 100 with respect to a control unit 101 configured with a computer.
- 104 manual / automatic operation mode switching control unit 105, radar 106, camera group 107, sensor group 108, surrounding moving body grasping unit 109, current position detection unit 110, display unit 111, touch panel 112, car navigation (hereinafter referred to as car navigation) (Abbreviated) function unit 113, user authentication unit 114, user authentication information storage unit 115, caller authentication unit 116, caller authentication information storage unit 117, behavior setting accepting unit 118 after getting off, behavior storage unit after getting off.
- 119 the behavior control processing unit 120, the timer circuit 121, and the voice input / output unit 122 are connected to each other. It is.
- a motor drive unit 131 is connected to the motor drive control unit 103.
- a steering drive unit 132 is connected to the steering drive control unit 104.
- a manual driving operation detecting unit 133 is connected to the manual / automatic driving mode switching control unit 105.
- a car navigation database 134 is connected to the car navigation function unit 113.
- a user authentication information storage unit 115 is connected to the user authentication unit 114.
- a caller authentication information storage unit 117 is connected to the caller authentication unit 116.
- a behavior storage unit 119 after getting off is connected to the behavior setting receiving unit 118 after getting off.
- the sound input / output unit 122 is connected to a microphone 135 for collecting sound and a speaker 136 for outputting sound.
- the wireless communication unit 102 has a function of performing communication such as telephone communication and e-mail communication through a mobile phone network. That is, the wireless communication unit 102 of the autonomous driving vehicle 1 of this embodiment has a function as a subscriber terminal in the mobile phone network and has a subscriber number as a mobile phone terminal.
- the control unit 101 has a function of performing processing when a call request is received from a caller through the wireless communication unit 102, and a user's mobile phone terminal in which a mobile phone subscriber phone number is registered.
- a function for sending and notifying a necessary message is provided as a software processing function. For example, a smartphone is used as the mobile phone terminal of the user.
- the motor drive control unit 103 controls the supply of a drive signal to the motor drive unit 131 of the automatic driving vehicle 1 constituted by the electric vehicle of this embodiment.
- the start of travel, the travel speed control (including brake control and accelerator control), the travel stop, etc. are controlled.
- the steering drive control unit 104 controls the supply of a drive control signal to the steering drive unit 132 of the autonomous driving vehicle 1 of this embodiment to control the course change of the autonomous driving vehicle 1. To do.
- the manual / automatic operation mode switching control unit 105 performs control to switch the operation mode of the automatic driving vehicle 1 between the manual operation mode and the automatic operation mode in accordance with a selection operation input through the touch panel 112.
- the manual driving operation detection unit 133 receives the operation information of the accelerator pedal operation, the brake pedal operation, the shift lever operation, and the steering operation by the driver, and sends the manual driving operation information to the manual / automatic driving mode switching control unit 105. Supply.
- the manual / automatic driving mode switching control unit 105 supplies the manual driving operation information from the manual driving operation detecting unit 133 to the motor drive control unit 103 and the steering drive control unit 104 when the automatic driving vehicle 1 is in the manual driving mode. Then, the motor drive unit 131 and the steering drive unit 132 are controlled in accordance with the driver's various pedal operations, shift lever operations, and steering operations (handle operations).
- the manual / automatic driving mode switching control unit 105 outputs the output of the radar 106, the camera group 107, the sensor group 108, and the surrounding moving body grasping unit 109 as described later when the automatic driving vehicle 1 is in the automatic driving mode.
- the automatic driving operation information generated by the control unit 101 is supplied to the motor driving control unit 103 and the steering driving control unit 104, and the motor driving unit 131 and the steering driving unit 132 are driven and controlled by the automatic driving operation information.
- the car navigation function unit 113 searches for a route from the current position to the destination (destination) set by the driver or the like, and is controlled so as to travel along the searched route.
- the radar 106 is used to measure the distance to people or objects existing around the vehicle of the autonomous driving vehicle 1 and includes a laser radar, a millimeter wave radar, or the like.
- the laser radar is embedded, for example, near the ceiling or the bumper, and the millimeter wave radar is provided, for example, at the front and rear of the vehicle. Both laser radar and millimeter wave radar may be provided, or only one of them may be provided. Other radars such as a quasi-millimeter wave radar and a microwave radar may be used. Further, a sonar (not shown) can be used for the same purpose as the radar.
- the camera group 107 includes one to a plurality of cameras for photographing the inside of the autonomous driving vehicle 1 and one to a plurality of cameras for photographing the outside of the vehicle such as the front, side, and rear of the autonomous driving vehicle 1.
- a camera that captures the interior of a vehicle is attached to the rear-view mirror (rear-view mirror, rear-view mirror) installed between the driver's seat and front passenger seat, the upper part of the front window, etc.
- a camera for photographing the work of a passenger (passenger) sitting in the passenger seat or the rear seat is included.
- Cameras that photograph the periphery of the autonomous vehicle 1 are attached to the left and right sides of the rearview mirror, for example, and two cameras (stereo cameras) that primarily photograph the left front and right front of the autonomous vehicle 1
- the sensor group 108 includes an open / close detection sensor for detecting opening / closing of a door and opening / closing of a window, a sensor for detecting wearing of a seat belt, and a seating sensor for detecting that a passenger is seated in a seat such as a driver's seat or a passenger seat (for example, a weight sensor or a pressure sensor), a touch sensor (for example, a capacitance sensor) that detects that a person touches a driver's seat handle or a touch panel 112 described later, and a human sensor that detects a person in the vicinity outside the vehicle. (For example, an infrared sensor) and various sensors for acquiring information to assist automatic driving.
- a vibration sensor for detecting vibration of a vehicle or a tire
- a rotation speed sensor for detecting the rotation speed of a tire
- a geomagnetic sensor for detecting acceleration
- a gyro sensor for detecting angle and angular velocity
- the sensor group 108 also includes sensors that detect lighting of a right turn signal, a left turn signal (direction indicator), and a hazard lamp (emergency blinking light).
- the surrounding moving body grasping unit 109 grasps the moving body (including a person) around the own vehicle using the radar 106, the sensor group 108, and the captured image of the camera group 107.
- the surrounding moving body grasping unit 109 grasps surrounding obstacles and moving bodies by performing processing based on, for example, Bayesian theory.
- the current position detector 110 receives radio waves from GPS satellites and detects the current position of the vehicle. Since the current position detection unit 110 has poor accuracy of the position detected by the radio wave from the GPS satellite, the current position detection unit 110 includes not only the information on the current position detected by the reception of the radio wave from the GPS satellite, but also 1 to In addition to using a plurality of sensors, radar 106, captured images of the camera group 107 (with a navigation function), and the like, for example, processing based on Bayesian theory is performed to detect and confirm the current position with higher accuracy. ing.
- the autonomous driving vehicle 1 has position information acquired by receiving radio waves from the radar 106, the camera group 107, the sensor group 108, and GPS satellites in the current position detection unit 110 and the surrounding moving body grasping unit 109.
- Information that is, information corresponding to information obtained from human eyes and ears, is processed by machine learning such as Bayesian theory, and based on this, the control unit 101 knows how to change the course of the vehicle and avoid obstacles.
- Information processing (artificial intelligence) and control (artificial intelligence) are performed to generate automatic driving operation information.
- the display unit 111 includes, for example, an LCD (Liquid Crystal Display).
- the touch panel 112 is arranged by superimposing a touch sensor capable of touch input with a finger on the display screen of the display unit 111 formed of an LCD.
- a display image including software buttons (including keyboard character input buttons) is displayed on the display screen of the display unit 111 based on the control of the control unit 101.
- the touch panel 112 detects a touch by a finger on the software button displayed on the display screen, the touch panel 112 transmits the touch to the control unit 101.
- the control unit 101 is configured to execute control processing corresponding to the software button.
- the car navigation function unit 113 is a function unit for guiding the automatic driving car 1 to assist in moving to a designated destination based on a map stored in the car navigation database 134 and route guidance data. It is. In this embodiment, the car navigation function unit 113 is configured to perform slightly different processing in the manual operation mode and the automatic operation mode.
- the car navigation function unit 113 is detected and confirmed by the current position detection unit 110 on the map that explicitly displays the route (route) to the destination on the display screen of the display unit 111.
- the vehicle position (current position) on the map is moved as the vehicle moves, and route guidance such as intersections and branch points on the route is provided. Provide voice guidance where necessary. This is the same as the normal car navigation function.
- the car navigation function unit 113 notifies the control unit 101 of information on the direction and distance of the vehicle when the current position of the vehicle is away from the route to the destination.
- the control unit 101 is notified of the change direction information along the route before the intersection or branching point on the route as the vehicle moves. To do.
- the control unit 101 instructs the own vehicle on the route.
- the motor drive unit 131 is controlled through the motor drive control unit 103 and the automatic drive operation information for controlling the steering drive unit 132 is generated through the steering drive control unit 104 so as to move along the route as described. . Therefore, by the route guidance to the destination by the car navigation function unit 113 and the control unit 101 in the automatic driving mode, the automatic driving vehicle 1 can move to the destination even when the passenger is unattended.
- the car navigation function unit 113 also has a peripheral search function.
- the car navigation database 134 stores base information (POI (Point Of Interest)) such as parking lots, convenience stores, and the like for peripheral search.
- POI Point Of Interest
- the surrounding search function is not provided in the car navigation function unit 113, but the control unit 101 sends a surrounding search request including position information such as current position information acquired by the current position detection unit 110, for example, to the wireless communication unit 102. It may be realized as a function of sending the search result to a peripheral search server on the Internet through the Internet.
- the user authentication unit 114 performs user authentication using the user authentication information stored in the user authentication information storage unit 115 and the authentication information acquired from the user at that time.
- the user is mainly a driver, but may be a passenger other than the driver.
- the autonomous driving vehicle 1 of this embodiment can autonomously travel in the automatic driving mode even when there is no driver.
- the user authentication unit 114 is configured as an image recognition unit.
- the user authentication information storage unit 115 stores a face image of a passenger who is photographed by a predetermined camera in the camera group 107.
- the user authentication information storage unit 115 is updated and stored by overwriting the previous face image of the person getting off, which was captured by the camera, over the previous face image of the person getting off. The Of course, it is possible to leave the previous image of the person getting off without overwriting.
- user authentication can also be performed by the user's voice.
- the autonomous driving vehicle 1 picks up the passenger's voice with the microphone 135 and stores it in the user authentication information storage unit 115.
- the voice of the person getting off is memorized.
- the user authentication unit 114 is configured to have a speaker voice recognition function, and determines whether the stored voice and the voice of the user collected by the microphone 135 match / do not match. , Perform user authentication.
- user authentication can be performed using the user's fingerprint.
- the automatic driving vehicle 1 is provided with a fingerprint reading device, and the user authentication information storage unit 115 is provided with the information of the person getting off the vehicle.
- the fingerprint is stored, and the user authentication unit 114 is configured to have a fingerprint recognition function, and determines whether the stored fingerprint matches or does not match the new user's fingerprint acquired by the fingerprint reader.
- the first user authentication is performed.
- veins, irises, and other biometric information can be used, and in these cases, it is possible by changing the same configuration.
- the caller authentication unit 116 includes caller authentication information stored in the caller authentication information storage unit 117 and authentication information acquired from the caller acquired when the wireless communication unit 102 receives a call (call). Are used to perform the above-described first caller authentication. In this embodiment, caller authentication is performed by matching / mismatching between the telephone number of the mobile phone terminal of the latest passenger and the telephone number of the mobile phone terminal of the caller.
- the caller authentication information storage unit 117 stores the telephone number of the mobile phone terminal of the latest disembarkation person input through the touch panel 112.
- the telephone number stored in the caller authentication information storage unit 117 is overwritten on the previously stored telephone number, so that it is the telephone number of only the latest passenger.
- the telephone number previously stored may be left without being overwritten.
- the caller authentication unit 116 acquires the incoming telephone number and matches the telephone number stored in the caller authentication information storage unit 117. / It is configured to perform caller authentication by determining a mismatch.
- the difference in information used as authentication information is the same as the configuration of the user authentication unit 114 and the user authentication information storage unit 115.
- the configuration is changed according to the above.
- the behavior setting reception unit 118 after getting off receives the setting of the behavior after getting off by the user when the user such as a driver or a passenger other than the driver gets off.
- the received setting information is stored in a storage unit (not shown).
- the behavior storage unit 119 after getting off is registered and stored in advance by the user.
- the behavior setting accepting unit 118 after getting off presents a list of behaviors after getting off stored in the behavior storing unit 119 after getting off to the user, and information on behaviors after getting off that is selected and set by the user from the list. Accept the settings.
- each after-get-off behavior item in the after-get-off behavior list (see FIG. 2 described later) is configured as an icon button, and the user operates each desired icon button on the touch panel 112. By instructing, it is possible to perform selection setting input of behavior information after getting off.
- An example of the behavior after getting off stored in the behavior storing unit 119 after getting off and the setting receiving process of the behavior setting receiving unit 118 after getting off will be described later.
- the control unit 101 has a voice recognition function for that purpose.
- the behavior control processing unit 120 executes the behavior of the user after getting off the vehicle based on the behavior after getting off received by the behavior setting receiving unit 118 after getting off.
- the control unit 101 in this embodiment, after the user gets off, Only when there is no driver in the vehicle, the control process based on the accepted behavior after getting off the vehicle is executed.
- An example of the behavior control processing by the behavior control processing unit 120 will be described in detail later.
- the timer circuit 121 has a function of providing the current time and measuring the time from a predetermined time point.
- the voice input / output unit 122 takes in the voice collected by the microphone 135 and sends it out to the system bus 100 for voice recognition processing, for example.
- the voice input / output unit 122 has a built-in memory for storing voice message data to be emitted to the outside, and converts voice message data read from the memory into an analog voice signal. Built-in speech synthesizer and DA converter. Then, the voice input / output unit 122 supplies the voice message selected by the control of the control unit 101 to the speaker 136 so as to be emitted to the outside as a voice.
- an inquiry message such as “Do you want to set the behavior after getting off?”
- a notification message such as “Authenticated”, “Could not be authenticated”, etc.
- the electronic control circuit unit 10 of the automatic driving vehicle 1 is configured.
- the motor drive control unit 103, the steering drive control unit 104, and manual / automatic operation mode switching are performed.
- Control unit 105, surrounding moving body grasping unit 109, current position detection unit 110, car navigation function unit 113, user authentication unit 114, caller authentication unit 116, behavior setting accepting unit 118 after getting off, behavior control processing unit 120, voice input Each processing function of the output unit 122 can be realized as software processing performed by the control unit 101 executing a program.
- the after-get-off behavior storage unit 119 stores and holds in advance the after-get-off behavior that the user wants to specify.
- the behavior after getting off it can be selected from those registered in advance by the automobile company or the like in advance in the autonomous driving vehicle 1 or can be set and stored by the user. Further, what is stored on the Internet cloud can be used via the wireless communication unit 102.
- FIG. 2 shows an example of the behavior after getting off stored in the behavior storing unit 119 after getting off in this embodiment. An example of behavior after each getting off will be described.
- “Move to the default parking lot” is to make the automatic driving car 1 move to a pre-registered default parking lot after the user gets off. End.
- the default parking lot a plurality of parking lots can be registered, and “home parking lot”, “company parking lot”, “contracted parking lot”, and the like can be registered.
- the registration of the parking lot means storing the position information and the name (type) of the parking lot such as “home parking lot”, “company parking lot”, “contract parking lot”.
- “Wait until you call at a nearby parking lot” means that the self-driving car 1 searches for a parking lot in the vicinity of the user's disembarking position, waits at that parking lot, and then the user automatically uses a mobile phone terminal.
- the automatic driving car 1 returns to the place where the user got off in response to the calling.
- the user registers the authentication information for calling and the authentication information for re-boarding when getting off the vehicle.
- the behavior control processing unit 120 of the autonomous driving vehicle 1 authenticates the caller using the registered authentication information when the call is made by the user through the mobile phone terminal, and the authentication is OK. In response to a call only when there is a movement control is performed to return to the place where the user got off.
- the authentication information at the time of calling for example, the telephone number (subscriber number) of the caller's mobile phone terminal is registered at the time of getting off, and when receiving an incoming call from the caller, the telephone number of the caller Authentication is performed based on whether or not is a registered telephone number.
- the behavior control processing unit 120 of the autonomous driving vehicle 1 detects the user's re-boarding, the behavior control processing unit 120 authenticates the re-passenger using the registered information for re-boarding, and the authentication is OK. It is controlled so that the use of the own vehicle is permitted by the re-passenger only when there is.
- the authentication information at the time of re-entry for example, the user's face image is registered at the time of getting off, and when the user re-rides, the person who gets off the vehicle is registered by face recognition based on the face image. Authentication is performed depending on whether or not.
- the authentication information at the time of calling is not limited to the phone number of the mobile phone terminal, but may be an e-mail address.
- passwords and IDs may be registered, and in communication based on incoming calls from callers, authentication may be performed by sending the passwords and IDs and confirming their match.
- a combination of a telephone number or e-mail address and a password or ID may be used as authentication information at the time of calling.
- the information for authenticating the re-passenger may be not only a face image but also biological information such as voice (voice), fingerprint, vein, iris, etc. Furthermore, a password or ID is used. Also good. Further, a combination of these may be used as authentication information.
- “Wait here” means that the self-driving vehicle 1 waits at the place where the user got off.
- the user registers authentication information for re-boarding.
- the information for authentication for this re-boarding can be the same as that described in the above-mentioned “standby until calling at a nearby parking lot”, and in this embodiment, a facial image of a user (for example, a driver) To do.
- “Toward point A” means that the autonomous vehicle 1 moves toward the place designated by the user.
- Point A is designated by the user when getting off.
- This point A may be set from points registered in advance, for example, specified on a map, specified by inputting an address, or specifying a identifiable building name. It may be. Further, it may be specified by two-dimensional coordinates of latitude and longitude (or three-dimensional coordinates by adding elevation).
- point A can be specified by phone number, it may be specified by entering the phone number.
- point A can be specified by URL (Uniform Resource Locator) or email address, enter the URL or email address. May be specified.
- URL Uniform Resource Locator
- a password or ID is set as information for authenticating the re-passenger.
- “Going to pick up in response to a call” means that after the user gets off, the self-driving car 1 can move freely until a call is made.
- the automatic driving car 1 is picked up at the place designated by the user in response to the call.
- the information on the place to be picked up is sent to the automatic driving vehicle 1 in the telephone communication when the user makes a telephone call to the automatic driving vehicle 1.
- information on the current position measured by the GPS included in the user's mobile phone terminal is sent from the user's mobile phone terminal to the automatic driving vehicle 1 as information on a place to pick up.
- the user registers authentication information for calling and authentication information for re-boarding when getting off the vehicle.
- the same processing as “waiting until calling at a nearby parking lot” is performed.
- “Waiting at point B” means that the autonomous driving vehicle 1 waits at a place designated by the user and waits for the user to get on again.
- Point B is specified by the user when getting off. This point B may be set from pre-registered points, for example, specified on a map, specified by inputting an address, or specifying a identifiable building name. It may be. Further, it may be specified by two-dimensional coordinates of latitude and longitude (or three-dimensional coordinates by adding elevation). Furthermore, when the point B can be specified by a telephone number, it may be specified by inputting a telephone number. When the point B can be specified by a URL or an e-mail address, it is specified by inputting the URL or the e-mail address. Also good.
- the authentication information at the time of the user's re-entry can be the same as that described in the above-mentioned “standby until calling at a nearby parking lot”.
- the user for example, Driver's face image.
- “Return to the getting-off position after a predetermined time” is based on the assumption that the user returns to the getting-off position (current position) after the predetermined time when the user has performed a predetermined task after getting off (including not moving). This is later behavior.
- the user directly sets the “predetermined time” or inputs information for grasping the “predetermined time” when getting off the vehicle.
- Second half 45 minutes Determine the predetermined time with a little loss time as a guide. Time is added when it becomes an overtime. If broadcast broadcasts are obtained from a predetermined site through the Internet, the game end time can be grasped. ⁇ "Concert” The final time of the announcement by the organizer is a guide. When the user gets off, the end time is input by the user, and the predetermined time is grasped based on the time. ⁇ "Movie watching" The closing time is fixed. When the user gets off, the end time is input by the user, and the predetermined time is grasped based on the time.
- -“Shopping Center” Based on experience values and statistical values, grasp the predetermined time as 2 hours, for example.
- -“Department Store” A predetermined time is grasped as 2 hours, for example, based on experience values and statistical values.
- Mass retailer (home appliances / computers)” Based on experience values / statistical values, grasp a predetermined time, for example, 1 hour.
- ⁇ "Bookstore” Based on experience values and statistical values, grasp the predetermined time as 30 minutes, for example.
- ⁇ "Hanaya” Based on experience values and statistical values, grasp the predetermined time as 15 minutes, for example.
- -“Small store” Based on experience values and statistical values, grasp the predetermined time as 15 minutes, for example.
- -“Convenience store” Based on experience values and statistical values, grasp the predetermined time as 15 minutes, for example. ⁇ "Post office” Depending on the waiting time, for example, grasp as 10 minutes. ⁇ “Bank” Depending on the waiting time, it is assumed that the ATM is 5 minutes and the counter is 20 minutes. ⁇ End time of “school” is fixed. When the user gets off, the end time is input by the user, and a predetermined time, for example, 2 hours, is grasped based on the time. “Restaurant” For example, the predetermined time is grasped as 30 minutes for lunch and 2 hours for dinner. -"Café” For example, the predetermined time is grasped as one hour.
- the predetermined time is grasped as 2 hours. ⁇ "Shrines / Buddhist temples", etc. For example, the predetermined time is grasped as one hour. ⁇ “Zoo / aquarium” For example, 3 hours are used to grasp the predetermined time. ⁇ "Vehicle entry prohibition area""Please refrain from the car from here.” If you want to see it, get off the car and see it. For example, the predetermined time is grasped as 15 minutes.
- the autonomous driving vehicle 1 may determine the time by judging the state of the user (such as hurrying or slowing down) and TPO (situation, degree of congestion, etc.).
- TPO situation, degree of congestion, etc.
- the presence / absence of a parking space at the place where the user is lowered may be a condition for determining the predetermined time.
- the behavior after getting off is “return to the getting-off position after a predetermined time”, but may be “return to the getting-off position at a predetermined time”. Moreover, you may prepare the behavior after both getting off.
- Being at point C after a predetermined time is a behavior after getting off assuming that the user performs a predetermined task after getting off, and is a place where the user gets off again after a predetermined time. It moves to another point C that is not. Similar to “return to the getting-off position after a predetermined time”, the user can directly set “predetermined time” or input information for grasping the “predetermined time” when getting off, Point C is set.
- the point C may be set from points registered in advance. For example, the point C may be specified on a map, specified by inputting an address, or specified by a identifiable building name. May be. Further, it may be specified by two-dimensional coordinates of latitude and longitude (or three-dimensional coordinates by adding elevation).
- the point C when the point C can be specified by a telephone number, it may be specified by inputting a telephone number.
- the point C can be specified by a URL or an e-mail address, it is specified by inputting a URL or an e-mail address. Also good.
- the method of setting the “predetermined time” and the method of grasping the predetermined time by the control unit 101 of the autonomous driving vehicle 1 are as follows. Is the same.
- the point C is far from the getting-off position (current position) (for example, 1 km or more)
- the travel time from the getting-off position (current position) to the point C greatly affects the predetermined time.
- the predetermined time is insufficient only for the time required for the predetermined business, and it is necessary to grasp the time by adding the time required for the movement.
- the behavior after getting off can be set to “being at point C at a predetermined time” instead of “being at point C after a predetermined time” instead of “returning to the getting-off position after a predetermined time”. Moreover, you may prepare the behavior after both getting off.
- control part 101 may authenticate a user's re-boarding by transmitting and confirming with respect to a user's mobile telephone terminal through this radio
- wireless communication part 102 may authenticate a user's re-boarding by transmitting and confirming with respect to a user's mobile telephone terminal through this radio
- wireless communication part 102 may authenticate a user's re-boarding by transmitting and confirming with respect to a user's mobile telephone terminal through this radio
- the user who has been in the autonomous driving vehicle 1, in this example, especially when the driver tries to get off the autonomous driving vehicle 1 tries to get off whether or not to set the behavior after getting off.
- the autonomous driving vehicle 1 tries to get off whether or not to set the behavior after getting off.
- the driver inputs behavior setting after getting off
- the automatic driving vehicle 1 accepts the behavior setting input after getting off by the driver, and processing according to the behavior after getting off after the driver gets off. I do.
- FIG. 3 is a flowchart for explaining an example of the flow of processing operations executed by the control unit 101 of the electronic control circuit unit 10 of the automatic driving vehicle 1 when the driver gets off. Note that the processing of each step in the flowchart of FIG. 3 is realized as software processing performed by the control unit 101 executing a program, with each processing function of the behavior setting accepting unit 118 and behavior control processing unit 120 after getting off the vehicle. The case will be described as follows.
- the control unit 101 determines whether or not the driving of the motor driving unit 131 of the own vehicle has been stopped by the driver (step S1). When it is determined in step S1 that the driving of the motor driving unit 131 is not stopped, the control unit 101 continues the control necessary for traveling (step S2), and then returns to step S1.
- step S1 When it is determined in step S1 that the driving of the motor drive unit 131 has been stopped, the driver is generally expected to get off, so the control unit 101 makes an inquiry about whether to set the behavior after getting off the vehicle. Is displayed on the display unit 111 and is emitted as sound through the speaker 136 (step S3).
- the control unit 101 monitors and discriminates the answer from the driver in response to the inquiry in step S3 (step S4), and when it is judged that the driver does not set the behavior after getting off, the processing routine of FIG. 3 is executed. finish.
- the automatic driving vehicle 1 turns off the power supply while maintaining the power supply to the necessary part as the processing during the stop, with the motor driving unit 131 stopped at the position where the driver got off.
- a predetermined behavior when the driver does not set the behavior after getting off the vehicle may be set in advance and executed.
- the predetermined behavior set in advance may be schedule information.
- step S4 When it is determined in step S4 that the driver has answered that the behavior setting after getting off the vehicle is made, the control unit 101 displays the behavior after getting off the vehicle stored in the behavior storing unit 119 after getting off as a list as shown in FIG. The data is displayed on the display screen of the unit 111 (step S5).
- control unit 101 monitors the input operation of the user through the touch panel 112 and waits for acceptance of the selection operation of the behavior after getting off from the list displayed on the display screen (step S6).
- the control unit 101 performs processing for accepting the selected behavior after getting off (step S7).
- the acceptance process of the behavior after getting off at step S7 will be described in detail later.
- the control unit 101 determines whether or not the process for accepting the selected behavior after getting off is completed (step S8), and when determining that the process for accepting the selected behavior after getting off is completed.
- the selection information of the behavior after getting off selected by the user and the information accompanying it are stored (step S9). If the processing for accepting the selected behavior after getting off is not completed within a predetermined time (for example, 10 minutes), it is determined that the user does not set the behavior after getting off, and the processing routine of FIG. 3 is terminated. May be.
- the automatic driving vehicle 1 turns off the power supply while maintaining the power supply to a necessary part as a process during the stop while the motor driving unit 131 is stopped at the position where the driver gets off.
- a predetermined behavior when the driver does not set the behavior after getting off the vehicle may be set in advance and executed.
- the predetermined behavior set in advance may be schedule information.
- the control unit 101 confirms the user's getting off by using a door sensor or the like, and determines the presence or absence of the driver (step S10).
- the presence / absence of the driver includes a seating sensor including a weight sensor and a pressure sensor provided in the driver's seat, a touch sensor for determining whether or not a person touches the handle or the touch panel 112, It discriminate
- the control unit 101 waits for the driver to get off and determines that the driver has not got off at step S10. Is executed (step S11). In this step S11, when the autonomous driving vehicle 1 travels and moves, the autonomous driving vehicle 1 autonomously travels in the automatic driving mode.
- Example flow of behavior setting acceptance after getting off Next, an example of the flow of behavior setting reception processing after getting off in step S7 will be described.
- 4 to 6 are flowcharts for explaining an example of the flow of behavior setting acceptance processing after getting off, and in this example, the setting when each behavior after getting off of the list shown in FIG. 2 is selected. A description will be given assuming reception processing.
- the processing in the steps of FIGS. 4 to 6 will also be described as a case where the processing function of the behavior setting reception unit 118 after getting off is realized as software processing performed by the control unit 101 executing a program.
- the control unit 101 exchanges with the user in an interactive manner and accepts the setting of the behavior after getting off.
- the control unit 101 determines whether or not “move to default parking” has been selected by the user (step S21), and determines that “move to default parking” has been selected, For the parking lot, a message prompting the user to select one of “home parking lot”, “company parking lot”, and “contracted parking lot” is notified through the display unit 111 (in addition, voice notification through the speaker 136) Then, the selection designation of the default parking lot from the user is accepted (step S22). At this time, after step S22, the control unit 101 determines that the reception is completed in step S8 of FIG. 3, and then in step S9, the selected behavior after getting off the vehicle “move to the default parking lot”. And a selection designation of “home parking lot”, “company parking lot”, or “contract parking lot”. As will be described later, for a predetermined parking lot, “dedicated parking lot for a specific facility” may be added to the selection designation.
- step S21 determines whether “wait until calling in a nearby parking lot” has been selected by the user (step S21). S23).
- step S23 determines whether “wait until calling at nearby parking lot” has been selected by the user.
- the control unit 101 displays a message prompting the user to input authentication information of the caller. Through the speaker (and may be notified through the speaker 136) and accepts the input of the caller authentication information from the user (step S24). As described above, in this example, the telephone number of the caller's mobile phone terminal is input as the caller authentication information.
- the control unit 101 acquires authentication information when the user re-rides (step S25).
- a face image of a driver as a user is captured by a camera that captures a driver's seat in the camera group 107, and the face image is acquired as authentication information for the re-passenger.
- the control unit 101 determines that the reception is completed in step S8 of FIG. 3, and in the next step S9, the behavior after the getting-off is “wait until calling at a nearby parking lot”.
- the selection designation is stored in the storage unit of the behavior setting reception unit 118 after getting off, and the telephone number of the mobile phone terminal as caller authentication information is stored in the caller authentication information storage unit 117.
- the “driver's face image” as the authentication information is stored in the user authentication information storage unit 115.
- control unit 101 determines whether or not "wait here” is selected by the user (step S26). .
- the control unit 101 determines whether or not parking is prohibited at the current position (alighting position) (step S27). Whether or not parking is prohibited is determined by recognizing whether or not there is a parking prohibition sign or a sign from a captured image of a camera that captures the periphery of the camera group 107.
- the parking prohibition is the same as the parking prohibition, but in this embodiment, it is included in “parking prohibition”.
- step S27 If it is determined in step S27 that parking is prohibited, the control unit 101 notifies the user that the current position is prohibited to park, and displays a message prompting another user to set a behavior after getting off the display unit 111. Through (there may be voice notification through the speaker 136) (step S28), and then the process returns to step S21 to wait for the user to reset the behavior after getting off.
- step S29 When it is determined in step S27 that the current position is not parking prohibited, the control unit 101 acquires authentication information when the user re-rides (step S29).
- the process in step S29 is the same as that in step S25 described above.
- the face image of the driver as the user is acquired as authentication information for the re-occupant.
- the control unit 101 determines that the reception is completed in step S8 of FIG. 3, and in the next step S9, selects and designates the selected behavior after getting off the vehicle “wait here”.
- the information is stored in the storage unit of the behavior setting reception unit 118 after getting off, and the “driver's face image” as the authentication information for the re-passenger is stored in the user authentication information storage unit 115.
- step S31 determines whether or not “toward point A” has been selected by the user (step S31 in FIG. 5). ).
- the control unit 101 notifies a message for prompting the user to input the setting of “point A” through the display unit 111 ( At the same time, voice notification may be made through the speaker 136), and the setting of “point A” from the user is accepted (step S32).
- step S32 the control unit 101 displays a map centered on the current position on the display screen of the display unit 111, and accepts the setting of “point A” on the map.
- an input of the address of “point A” by the user may be accepted, or a facility around the current location may be presented and the selection input may be accepted. Further, input of two-dimensional coordinates of latitude and longitude (or three-dimensional coordinates by adding elevation) may be accepted. Furthermore, when “point A” can be specified by a telephone number, input of the telephone number may be accepted, and when “point A” can be specified by URL or mail address, input of URL or mail address is accepted. Also good.
- step S33 the control unit 101 acquires authentication information when the user re-rides at the point A (step S33).
- the processing in step S33 is the same as in steps S25 and S29.
- the driver's face image as a user is acquired as authentication information for the re-passenger.
- the control unit 101 determines that the reception is completed in step S8 of FIG. 3, and in the next step S9, selects and designates the behavior after getting off the vehicle “toward point A”.
- the setting information of “point A” is stored in the storage unit of the behavior setting reception unit 118 after getting off, and the “driver's face image” as information for re-passenger authentication is stored in the user authentication information storage unit 115.
- step S34 determines whether or not “go to pick up in response to the call” has been selected by the user.
- the control unit 101 sends a message prompting the user to input authentication information for the caller through the display unit 111. Notification (along with sound notification through the speaker 136) may be performed, and input of caller authentication information from the user is accepted (step S35). In this example, the telephone number of the caller's mobile phone terminal is input as the caller authentication information.
- step S36 acquires authentication information when the user re-rides.
- the process in this step S36 is the same as that in step S25 and step S29.
- the driver's face image as a user is acquired as authentication information for the re-passenger.
- step S36 the control unit 101 determines that the reception has been completed in step S8 of FIG. 3, and in the next step S9, selects the behavior to be picked up after getting off the vehicle selected in step S9.
- the designation is stored in the storage unit of the behavior setting accepting unit 118 after getting off, the telephone number of the mobile phone terminal as caller authentication information is stored in the caller authentication information storage unit 117, and further, The “driver's face image” as the authentication information is stored in the user authentication information storage unit 115.
- step S34 determines whether or not “Stand by at point B” is selected by the user (step S37). .
- the control unit 101 informs the user of a setting input of “point B” through the display unit 111 ( In addition, voice notification may be made through the speaker 136), and the setting of “point B” from the user is accepted (step S38).
- step S38 the control unit 101 displays a map centered on the current position on the display screen of the display unit 111, and accepts the setting of “point B” on the map.
- an input of the address of “point B” by the user may be accepted, or a facility around the current location may be presented and the selection input may be accepted. Further, input of two-dimensional coordinates of latitude and longitude (or three-dimensional coordinates by adding elevation) may be accepted. Furthermore, when “point B” can be specified by a telephone number, input of the telephone number may be accepted, and when “point B” can be specified by URL or mail address, input of URL or mail address is accepted. Also good.
- step S39 the control unit 101 acquires authentication information when the user re-rides at the point B (step S39).
- the processing in step S39 is the same as that in steps S25 and S29.
- the face image of the driver as the user is acquired as authentication information for the re-occupant.
- the control unit 101 determines that the reception is completed in step S8 of FIG. 3, and in the next step S9, selects and designates the behavior after getting off the vehicle selected “standby at point B”.
- the setting information of “point B” is stored in the storage unit of the behavior setting reception unit 118 after getting off, and the “driver's face image” as the authentication information for the re-passenger is stored in the user authentication information storage unit 115.
- step S37 determines whether “standby at point B” is not selected. If it is determined in step S37 that “standby at point B” is not selected, the control unit 101 determines whether “return to the getting-off position after a predetermined time” is selected by the user (FIG. 6). Step S41). In step S41, when it is determined that “return to the alighting position after a predetermined time” is selected by the user, the control unit 101 calculates information for calculating (estimating) the “predetermined time” for the user.
- Message (including direct input of “predetermined time”) is notified through the display unit 111 (and may be notified by voice through the speaker 136), and the voice input information from the user as described above And input of information for calculating “predetermined time” through touch panel 112 (including direct input of “predetermined time”) is accepted (step S42).
- step S43 acquires information for authentication when the user re-rides.
- the processing in step S43 is the same as that in steps S25 and S29.
- the face image of the driver as the user is acquired as authentication information for the re-occupant.
- step S43 the control unit 101 determines that the reception has been completed in step S8 of FIG. 3, and in the next step S9, selects the behavior after getting off the vehicle “returns to the getting off position after a predetermined time”.
- the designation is stored in the storage unit of the behavior setting reception unit 118 after getting off, and the “driver's face image” as the authentication information for the re-passenger is stored in the user authentication information storage unit 115.
- step S44 determines whether “there is a point C after a predetermined time” has been selected by the user. In this step S44, when it is determined that “I will be at point C after a predetermined time” is selected by the user, the control unit 101 inputs information for calculating (estimating) the “predetermined time” for the user.
- a message prompting (including direct input of “predetermined time”) is notified through the display unit 111 (and may be notified by voice through the speaker 136), and voice input information from the user and the touch panel as described above
- Input of information for calculating “predetermined time” through 112 (including direct input of “predetermined time”) is accepted (step S45).
- step S45 when “point C” is far from the getting-off position (current position), it is preferable to directly input “predetermined time”.
- step S46 the control unit 101 displays a map centered on the getting-off position (current position) on the display screen of the display unit 111, and accepts the setting of “point C” on the map. To do.
- an input of the address of “point C” by the user may be accepted, or a facility around the current location may be presented and the selection input may be accepted.
- input of two-dimensional coordinates of latitude and longitude (or three-dimensional coordinates by adding elevation) may be accepted.
- point C can be specified by a telephone number
- an input of a telephone number may be accepted
- an input of a telephone number may be accepted
- an input of a telephone number may be accepted
- an input of a URL or an email address an input of a URL or an email address is accepted. Also good.
- the control unit 101 takes the time required for the automatic driving vehicle 1 to move from the disembarking position (current position) to “point C” when accepting the setting of “point C”. It is possible to determine whether “being at point C after a predetermined time” is temporally possible as the behavior after getting off. In this case, when accepting the setting of “point C”, the control unit 101 causes the car navigation function unit 113 to search for a route from the getting-off position (current position) to “point C” using the route search function. Determine the route. Next, from the route and the traveling speed of the autonomous driving vehicle 1, the time required for the autonomous driving vehicle 1 to move from the getting-off position (current position) to “point C” is estimated.
- the estimated travel time is compared with the “predetermined time” set as input. If it is determined from the comparison result that the required travel time is longer, it is difficult to realize the behavior after getting off that “it is at point C after a predetermined time”. A message prompting re-input of “predetermined time” or “point C” is notified so as to be shorter than “predetermined time”, and re-input of the information is accepted. Of course, when the point C is close to the getting-off position (current position), there is almost no need for such re-input.
- step S47 acquires authentication information when the user re-rides at the point C (step S47).
- the processing in step S47 is the same as that in steps S25 and S29.
- the driver's face image as a user is acquired as authentication information for the re-passenger.
- step S47 the control unit 101 determines that the acceptance is completed in step S8 of FIG. 3, and selects the “being at point C after a predetermined time” behavior after the getting off in the next step S9.
- the designation and the setting information of “point C” are stored in the storage unit of the behavior setting reception unit 118 after getting off, and the “driver's face image” as the authentication information for the re-passenger is stored as the user authentication information storage unit 115.
- step S48 determines whether or not the user has performed an operation to cancel the behavior setting after getting off.
- step S48 determines whether or not the user has performed an operation for canceling the behavior setting after getting off.
- the control unit 101 returns the process to step S21 in FIG. 4 and repeats the processes after step S21.
- it is determined that the user's desired behavior after getting off the vehicle is not included in the stored contents of the behavior after getting off as shown in FIG. 2, and the user is prompted to input characters from the touch panel 112 or through the microphone 135. Or may be prompted.
- a predetermined behavior when the driver does not set the behavior after getting off the vehicle may be set in advance and executed.
- the predetermined behavior set in advance may be schedule information.
- step S48 If it is determined in step S48 that the user has performed an operation to cancel the behavior setting after getting off the vehicle, the control unit 101 ends this processing routine.
- the automatic driving vehicle 1 supplies power to a necessary part as a process during stopping while the motor driving unit 131 is stopped at the position where the driver gets off.
- the power supply may be turned off while maintaining.
- a predetermined behavior when the driver does not set the behavior after getting off the vehicle may be set in advance and executed.
- step S11 of FIG. 3 shows the flow of the control processing operation of the behavior after getting off executed by the control unit 101 of the electronic control circuit unit 10 of the automatic driving vehicle 1 when the driver gets off and the driver is absent. It is a flowchart for demonstrating an example. This example will also be described assuming processing corresponding to each of the behaviors after getting off in the list shown in FIG.
- the processing of each step in the flowcharts of FIGS. 7 to 12 includes processing functions of the behavior control processing unit 120, the user authentication unit 114, and the caller authentication unit 116, and the control unit 101 executes a program. This will be described as a case where it is realized as software processing performed in (1).
- the control unit 101 first confirms what the set behavior after getting off the vehicle is (step S51). Then, as a result of the confirmation, it is determined whether or not the set behavior after getting off is “movement to a predetermined parking lot” (step S52). In this step S52, when it is determined that the set behavior after getting off is “movement to a default parking lot”, the control unit 101 acquires information on the current position measured by the current position detection unit 110, The car navigation function unit 113 is caused to search for a route from the current position to a preset parking lot that is selected and stored (step S53). The control unit 101 uses the route search result to control the car navigation function unit 113 to perform route guidance, and the automatic driving vehicle 1 is selected and stored in the automatic driving mode. Control is performed to move to a predetermined parking lot (step S54). And the control part 101 will complete
- FIG. 13 shows an example of the situation when this “movement to the default parking lot” is set as the behavior after getting off the vehicle.
- the designated parking lot of the apartment is set as the default parking lot
- “Move” is designated as the behavior after getting off
- the vehicle gets off from the self-driving car 1 and returns to the home.
- the autonomous driving vehicle 1 accepts the setting of the behavior after getting off by the user 2.
- the automatic driving vehicle 1 performs an operation of moving to and parking in the designated parking lot 4 of the designated apartment in the automatic driving mode as the set behavior after getting off.
- the motor driving unit 131 is stopped, and the power supply is turned off while maintaining power supply to a necessary part as a process during the stop.
- the company parking lot is set as the default parking lot, and “moving to the default parking lot” Is set as the behavior after getting off, and it is assumed that the user gets off the automatic driving car 1 and enters the inside of the company building.
- the autonomous driving vehicle 1 accepts the setting of the behavior after getting off by the user 2, and after confirming the getting off of the user 2, moves to the designated company parking lot in the automatic driving mode.
- the motor driving unit 131 is stopped, and the power supply is turned off while maintaining power supply to a necessary part as a process during the stop. The same operation is performed when the contract parking lot is designated as the default parking lot.
- the accommodation facility such as a hotel for accommodation
- the accommodation facility It is assumed that the private parking lot is set as the default parking lot, “movement to the default parking lot” is set as the behavior after getting off, getting off from the automatic driving car 1 and entering the accommodation facility.
- the self-driving car 1 accepts the setting of the behavior after getting off by the user 2, and after confirming that the user 2 gets off, moves to the designated parking lot of the designated accommodation facility in the automatic driving mode.
- the parking operation is executed, and when the parking is completed, the motor driving unit 131 is stopped, and the power supply is also turned off while maintaining power supply to a necessary part as a process during the stop.
- the autonomous driving vehicle 1 is basically not used until the next day. Therefore, if the user re-rides on the same day despite staying, there is a risk of impersonation and trouble for the user, so authentication is performed more strictly. For example, a plurality of pieces of authentication information are always combined, an alarm sound is given at the time of authentication, a re-passenger's face image is taken, a caller's voice is recorded, and the like.
- control part 101 may authenticate a user's re-boarding by transmitting and confirming with respect to a user's mobile telephone terminal through this radio
- FIG. Furthermore, a telephone contact may be made at the accommodation facility to confirm the user's re-boarding.
- the user 2 of the self-driving car 1 designates the dedicated parking lot of the facility as the default parking lot. The same operation may be performed. If there is a specific facility such as a hotel that is frequently used, the user 2 may register the dedicated parking lot as a default parking lot.
- step S52 when it is determined in step S52 that the set behavior after getting off is not “moving to the default parking lot”, the control unit 101 sets the set behavior after getting off the vehicle “until called in a nearby parking lot”. It is determined whether or not it is “waiting” (step S55).
- step S55 when it is determined that the set behavior after getting off is “standby until called at a nearby parking lot”, the control unit 101 controls the car navigation function unit 113 to set the current position as the central position.
- a neighborhood search is performed to detect a nearby parking lot (step S56).
- control unit 101 acquires information on the current position measured by the current position detection unit 110, and sends a route from the current position to the nearby parking lot detected in step S56 to the car navigation function unit 113.
- the route search function is used for searching (step S57).
- the control unit 101 uses the route search result to control the car navigation function unit 113 to perform route guidance, and in the automatic driving mode, the nearby parking location detected in step S56 is the automatic driving vehicle 1. Control is performed to move the vehicle to the parking lot (step S58).
- Step S59 When the control unit 101 parks at the target parking lot, the control unit 101 stops the motor drive unit 131 to enter a standby state, and monitors the wireless communication unit 102 to determine whether a call (incoming call) has been received.
- the target parking lot is full, it is a matter of course that another parking lot is searched again and moved to the detected parking lot. Needless to say, the waiting place is not a parking lot, but can play an equivalent role as long as it can be parked.
- step S59 If it is determined in step S59 that the call has been received, the control unit 101 obtains the mobile phone terminal phone number stored in the caller authentication information storage unit 117 and the incoming mobile phone terminal phone number. It is checked to determine whether the caller authentication is OK (step S61 in FIG. 8).
- step S61 When it is determined in step S61 that the authentication of the caller is not OK, the control unit 101 sends a call NG voice message to the destination of the incoming call through the wireless communication unit 102 (step S62). And the control part 101 returns a process to FIG.7 S59. If it is determined in step S61 that the caller authentication is OK, the control unit 101 controls the car navigation function unit 113 to perform route guidance by tracing back the route searched in step S57. Thus, the automatic driving vehicle 1 is moved from the parking lot at the current position to the original getting-off position in the automatic driving mode (step S63).
- the control unit 101 determines whether or not a user has got on the driver's seat, for example, whether or not a person touches a door sensor, a seating sensor including a weight sensor or a pressure sensor, a handle, or a touch panel 112. It is determined by monitoring a sensor, a captured image of a camera for photographing a driver in the driver's seat in the camera group 107, the presence / absence of a voice of the driver in the driver's seat picked up by the microphone 135, etc. (step S64). .
- the control unit 101 uses the face image as the re-passenger authentication information stored in the user authentication information storage unit 115 and the boarding. By comparing and referring to the face image obtained by photographing the driver who has taken the camera with the camera, it is determined whether or not the re-passenger authentication is OK (step). S65).
- step S65 When it is determined in this step S65 that the re-passenger authentication is not OK, the control unit 101 displays a message prompting to get off because the re-passenger authentication cannot be performed on the display screen of the display unit 111, and Sound is emitted through the speaker 136 to prompt the re-occupant to get off (step S67). Then, the control unit 101 monitors the door sensor, the seating sensor, the touch sensor, the captured image of the camera, the sound collected by the microphone 135, etc., and determines whether or not the passenger has got off (step S68). If it is determined that there is not, the process returns to step S67. Further, when it is determined in step S68 that the vehicle has got off, the control unit 101 returns the process to step S64 and repeats the processes after step S64.
- step S65 If it is determined in step S65 that the re-passenger authentication is not OK, there is a possibility of a suspicious person. Therefore, the control unit 101 does not simply prompt the re-passenger to get off as in step S67. Instead, the possibility of a suspicious person may be communicated to the mobile phone terminal of the latest dismounter or the owner of the autonomous driving vehicle 1 by telephone or e-mail. In the case of contact by e-mail, a face image for authenticating the re-passenger obtained by photographing with a camera may be attached. Of course, unless the latest disembarkation person or the owner of the autonomous driving vehicle 1 permits, the driving operation of the autonomous driving vehicle is impossible. Further, when it is determined that the possibility of a suspicious person is extremely high, the 110th notification may be made.
- step S65 If it is determined in step S65 that the re-passenger authentication is OK, the control unit 101 permits the re-boarding of the driver who has boarded (step S66), and then the driver (re-passenger). The process shifts to the operation control mode processing routine according to the instruction.
- step S55 in FIG. 7 When it is determined in step S55 in FIG. 7 that the set behavior after getting off is not “standby until called at a nearby parking lot”, the control unit 101 sets the set behavior after getting off is “standby here”. (Step S71 in FIG. 9). When it is determined in this step S71 that the set behavior after getting off is “standby here”, the control unit 101 stops the motor driving unit 131 to put the automatic driving vehicle 1 in the standby state and performs processing. Proceeding to step S64 in FIG. 8, the user's re-boarding is monitored, and the processing after step S64 is repeated.
- step S72 it is determined whether or not the set behavior after getting off is “toward point A” (step S72).
- the control unit 101 acquires information on the current position measured by the current position detection unit 110, and the current position The route to the point A that is set and stored is caused to be searched by the car navigation function unit 113 using the route search function (step S73). Then, using the route search result, the control unit 101 controls the car navigation function unit 113 to perform route guidance, and controls the automatic driving vehicle 1 to move to the point A in the automatic driving mode. (Step S74).
- step S74 the control unit 101 advances the process to step S64 in FIG. 8, monitors the user's re-boarding at the point A, and repeats the processes after step S64.
- step S72 when it is determined in step S72 that the set behavior after getting off is not “toward the point A”, the control unit 101 sets “being picked up in response to the call”. Is determined (step S75).
- step S75 When it is determined in this step S75 that the set behavior after getting off is “to pick up in response to the call”, the control unit 101 determines whether or not the current position is prohibited from parking (step S76). Whether or not parking is prohibited is determined by recognizing whether or not there is a parking prohibition sign or a sign from a captured image of a camera that captures the periphery of the camera group 107.
- step S76 When it is determined in this step S76 that parking is not prohibited, the control unit 101 stops the motor driving unit 131 and waits while setting a state in which a call from the user is accepted at the current position (step S77). If it is determined in step S76 that parking is prohibited, the control unit 101 controls the car navigation function unit 113 to perform a peripheral search with the current position as the central position to search for a nearby parking available place. (Step S78).
- control unit 101 determines whether or not there is a nearby parking available place as a result of the search in step S78 (step S79). If it is determined that there is a nearby parking available place, the route to the parking available place is given to the car navigation function unit 113. The automatic driving vehicle 1 is moved in the automatic driving mode to the parking available place by performing the search along the route, and the motor driving unit 131 is set in a state of accepting a call from the user. Stop and wait (step S80).
- step S79 When it is determined in step S79 that a nearby parking available place is not found, the control unit 101 continues to move in the range of the surrounding area centering on the user's getting-off place and waits for a call from the user (step S79). S81).
- step S77 step S80 or step S81
- the control unit 101 monitors whether or not a call from the user has been received and waits for the call (step S91 in FIG. 10).
- step S91 the control unit 101 determines whether or not the caller authentication is OK in the same manner as in step S61 (step S92).
- step S92 When it is determined in step S92 that the authentication of the caller is not OK, the control unit 101 sends a call NG voice message to the other party that has received the call through the wireless communication unit 102 (step S93). Then, the control unit 101 returns the process to step S91. If it is determined in step S92 that the authentication of the caller is OK, the control unit 101 generates a voice message by voice synthesis requesting the other party that has received the call to transmit the location information of the place to pick up. In response to this, position information of a place to be picked up sent from the other party is acquired (step S94).
- control unit 101 acquires information on the current position measured by the current position detection unit 110, and sends the route from the current position to the location to be picked up acquired in step S94 to the car navigation function unit 113.
- Search is performed using the search function (step S95).
- control unit 101 controls the car navigation function unit 113 to perform route guidance using the route search result, and picks up the automatic driving vehicle 1 in step S94 in the automatic driving mode.
- Control is made to move to a place (step S96). Thereafter, the control unit 101 advances the process to step S64 in FIG. 8, monitors the user's re-boarding, and repeats the processes after step S64.
- step S75 of FIG. 9 when it is determined in step S75 of FIG. 9 that the set behavior after getting off the vehicle is not “going to pick up in response to the call”, the control unit 101 sets the set behavior after getting off the vehicle to “wait at point B”. Is determined (step S101 in FIG. 11).
- step S101 the control unit 101 acquires information on the current position measured by the current position detection unit 110, and the current position The route to the point B that has been set and stored is caused to be searched by the car navigation function unit 113 using the route search function (step S102).
- Step S103 the control unit 101 controls the car navigation function unit 113 to perform route guidance, and controls the automatic driving vehicle 1 to move to the point B in the automatic driving mode.
- Step S103 the control unit 101 advances the process to step S64 in FIG. 8, monitors the user's re-boarding, and repeats the processes after step S64.
- step S104 when it is determined in step S101 that the set behavior after getting off is not “standby at point B”, it is determined whether or not the set behavior after getting off is “return to the getting-off position after a predetermined time”.
- step S104 when it is determined that the set behavior after getting off is “return to the getting-off position after a predetermined time”, the control unit 101 performs a predetermined operation based on the stored information for measuring the predetermined time. Time is estimated (step S105).
- the control unit 101 determines whether or not the getting-off position (current position) is prohibited from parking (step S106). Whether or not parking is prohibited is determined by recognizing whether or not there is a parking prohibition sign or a sign from a captured image of a camera that captures the periphery of the camera group 107. If it is determined in step S106 that parking is not prohibited, the control unit 101 stops the motor driving unit 131 at the getting-off position (current position) and waits (step S107).
- step S106 when it is determined in step S106 that parking is prohibited, the control unit 101 continues to move, for example, around the area around the getting-off position (current position) for the predetermined time estimated in step S105.
- the car navigation function unit 113 is controlled to search the vicinity with the getting-off position (current position) as the central position, to search for a nearby parking available place, and to the parking lot detected as a result of the search
- step S108 To wait for the predetermined time estimated in step S105 (step S108).
- the predetermined time is a predetermined threshold, for example, 20 minutes or less, the movement is continued, and if the predetermined time is equal to or greater than the threshold, a parking area such as a parking lot is searched for and parking is possible. You may make it move to a different place and wait.
- step S108 the control unit 101 monitors whether or not the predetermined time estimated in step S105 has been reached (step S109). 113 is controlled to guide the route and return to the getting-off position (step S110). Then, after step S107 or step S110, the control unit 101 advances the processing to step S64, monitors the user's re-boarding, and repeats the processing after step S64.
- FIG. 14 shows an example of the situation when this “return to the getting-off position after a predetermined time” is set as the behavior after getting off.
- the example of FIG. 14 is a situation in which the driver 2 of the autonomous driving vehicle 1 makes a urine or constipation and stops at a toilet such as a park.
- the driver 2 sets the behavior after getting off by inputting voice, for example, “toilet” to the microphone 135 when getting off, and gets off the automatic driving car 1.
- the control unit 101 of the autonomous driving vehicle 1 recognizes the voice input “toilet” of the driver 2 by voice, estimates a predetermined time as 5 minutes in this example, and stands by at the place.
- the driver is authenticated to re-ride within a predetermined time or within a slightly longer time than the predetermined time, and the re-ride is permitted.
- a voice input such as “Toilet is large” may be input, and the predetermined time may be estimated to be longer, for example, 10 minutes.
- the predetermined time may be estimated longer than that of a man.
- FIG. 15 is a case where the driver 2 of the autonomous driving vehicle 1 goes to the national sports hall of Ryogoku, for example, with the autonomous driving vehicle 1 for watching the sumo wrestling.
- the control unit 101 of the autonomous driving vehicle 1 determines that the sumo wrestling place will end at 6:00 pm, estimates a predetermined time, and moves to, for example, a parking lot of the Kokugikan or a nearby empty parking lot. It waits, and when it is 6:00 pm, the driver 2 is picked up and controlled to move to the entrance / exit of the Kokugikan.
- step S104 when it is determined in step S104 that the set behavior after getting off is not “return to the getting-off position after a predetermined time”, the control unit 101 sets the behavior after getting off “is at point C after a predetermined time”. Is determined (step S111 in FIG. 12). When it is determined in this step S111 that the set behavior after getting off is not “at point C after a predetermined time”, the control unit 101 shifts to a processing routine for other behavior after getting off (step S112).
- step S111 when it is determined that the set behavior after getting off is “at point C after a predetermined time”, the control unit 101 determines that the predetermined time is based on the stored information for measuring the predetermined time. Is estimated (step S113).
- the control unit 101 acquires information on the current position measured by the current position detection unit 110, and sends a route from the current position to the set and stored point C to the car navigation function unit 113.
- a search is performed using the route search function (step S114).
- the control unit 101 controls the car navigation function unit 113 to perform route guidance, and controls the automatic driving vehicle 1 to move to the point C in the automatic driving mode. (Step S115).
- the control unit 101 determines whether or not the point C is prohibited from parking (step S116). Whether or not parking is prohibited is determined by recognizing whether or not there is a parking prohibition sign or a sign from a captured image of a camera that captures the periphery of the camera group 107. When it is determined in step S116 that parking is not prohibited, the control unit 101 stops the motor driving unit 131 and waits while accepting a call from the user at the point C (step S117).
- control unit 101 performs movement such as circling around the area around the point C for the predetermined time estimated in step S113 in the same manner as in step S108.
- movement such as circling around the area around the point C for the predetermined time estimated in step S113 in the same manner as in step S108.
- step S118 the control unit 101 monitors whether or not it has become close to the predetermined time estimated in step S113 (step S119), and if it is determined that it has become close to the predetermined time, the car navigation function unit 113 is controlled to guide the route and move to point C (step S120). Then, after step S117 or step S120, the control unit 101 advances the processing to step S64 in FIG. 8, monitors the user's re-boarding, and repeats the processing after step S64.
- the automatic driving vehicle 1 has an automatic driving mode and a manual driving mode.
- autonomous driving is performed only in the automatic driving mode.
- the manual / automatic operation mode switching control unit 105 and the manual driving operation detection unit 123 are turned off (invalid).
- FIG. 16 shows an example of a situation that is assumed to be used for business and is supposed to behave after getting off the vehicle.
- the automatic driving vehicle 1A in the example of FIG. 16 is not an individual passenger car but an example of a delivery truck of a company that owns the factory A.
- the products manufactured at the factory A are loaded on the automatic driving vehicle 1A and carried to the B center, and a part of the products loaded on the automatic driving vehicle 1 is lowered at the B center.
- the driving car 1 is moved from the B center to the C center, and all the remaining products loaded on the automatic driving car 1 are lowered at the C center.
- the automatic driving vehicle 1 is returned to the A factory. It is assumed that the position information of the factory A, the center B, and the center C is registered and stored in the automatic driving vehicle 1A.
- the unloading workers 2A and 2B get on the automatic driving vehicle 1A composed of a delivery truck and move.
- one of the workers 2A and 2B gets on the driver's seat, and the others get on the passenger seat.
- One of the workers 2A and 2B inputs the setting of the behavior after getting off when getting off, but it is not always necessary that one of the workers 2A and 2B gets on the driver's seat.
- the automatic driving vehicle 1A always moves in the automatic driving mode.
- the autonomous driving vehicle 1A first takes the workers 2A and 2B and moves from the garage to the factory A.
- one of the workers 2A and 2B selects and sets “stand by at the current position until there is a movement instruction and moves to the B center when there is a movement instruction” as the behavior after getting off.
- the workers 2A and 2B load products on the automatic driving vehicle 1A and re-ride them.
- the autonomous driving vehicle 1A authenticates both or one of the workers 2A and 2B as a user, and if it is determined that the authentication is OK, one of the workers 2A and 2B is determined based on the set behavior after getting off the vehicle. Waiting for a movement instruction from the factory, if it is determined that a movement instruction has been made, the factory moves from the A factory to the B center in the automatic operation mode.
- the automatic driving vehicle 1A When the automatic driving vehicle 1A arrives at the B center, one of the workers 2A and 2B will behave as after getting off, “waiting at the current position until a movement instruction is given, and moving to the C center when there is a movement instruction”. Select and set. Then, the workers 2A and 2B drop a part of the products to be delivered to the B center from the automatic driving vehicle 1A among the products loaded on the automatic driving vehicle 1A. Thereafter, the workers 2A and 2B get on the autonomous driving vehicle 1A again. Then, the autonomous driving vehicle 1A authenticates both or one of the workers 2A and 2B as a user, and if it is determined that the authentication is OK, one of the workers 2A and 2B is determined based on the set behavior after getting off the vehicle. In response to the movement instruction, the center B moves from the center B to the center C in the automatic operation mode.
- the self-driving car 1A arrives at the C center, one of the workers 2A and 2B, as the behavior after getting off, “waits at the current position until there is a movement instruction and returns to the factory A when there is a movement instruction”. Select and set. Then, the workers 2A and 2B drop all the remaining commodities loaded on the automatic driving vehicle 1A from the automatic driving vehicle 1A. Thereafter, the workers 2A and 2B get on the autonomous driving vehicle 1A again. Then, the autonomous driving vehicle 1A authenticates both or one of the workers 2A and 2B as a user, and if it is determined that the authentication is OK, one of the workers 2A and 2B is determined based on the set behavior after getting off the vehicle. Move to factory A according to the movement instruction.
- the number of factories and centers is not limited to this example, and there may be a plurality of delivery trucks.
- the delivery route and work procedure are determined in advance, they are set as defaults, and the behavior after getting off the vehicle is set as the default depending on the status of the baggage on the day, the number and situation of workers, and the weather. Can be changed flexibly when the worker gets off.
- the delivery route can be changed or the work procedure can be changed.
- the autonomous driving vehicle of the above-described embodiment when the user who gets on the autonomous driving vehicle gets off the vehicle, the user automatically changes in accordance with his / her behavior and purpose after getting off the vehicle. It can give instructions to the driving car and is very convenient. For example, when a user gets on an autonomous driving vehicle and returns home, the user can move to a parking lot away from home in the automatic driving mode by setting the behavior after getting off the autonomous driving vehicle. Easy to do.
- the user moves the autonomous driving vehicle to a desired place other than the getting-off place by setting the behavior after getting off.
- the user can re-ride at an appropriate position according to the user's behavior.
- the re-passenger can be authenticated using the pre-stored information for authenticating the re-passenger. It is possible to prevent a situation in which a person rides on an autonomous driving vehicle and uses it.
- the user can call using the communication means. Therefore, the user can re-ride on the autonomous driving vehicle from the arbitrary location by calling an autonomous driving vehicle by designating an arbitrary location other than the alighting place, which is very convenient.
- the caller since the caller is authenticated, it is possible to prevent a situation in which an autonomous vehicle is called and used by an unauthorized person.
- the self-driving vehicle of the present invention has an advantage that it can be used not only for personal passenger cars but also for commercial trucks.
- the control for switching from the manual operation mode to the automatic operation mode is performed by a selection operation input through the touch panel 112 by the driver, and the control for switching from the automatic operation mode to the manual operation mode is performed through the selection operation through the touch panel 112.
- these switching controls can also be performed automatically.
- when entering an expressway it can be detected by the car navigation system or GPS of an autonomous driving vehicle and automatically switched to the automatic driving mode, and conversely, when leaving the highway to a general road, it can be switched to the manual driving mode.
- an area where automatic driving is possible is specified, whether or not the vehicle is traveling in that area is determined by the car navigation system or GPS of the autonomous driving vehicle.
- the manual operation mode may be switched.
- an emergency response mode may be provided for switching between the manual operation mode and the automatic operation mode.
- the automatic operation mode for example, by pressing an emergency response button (not shown) or preparing an emergency response menu on the touch panel 112 and touching it.
- the automatic operation mode is canceled, the vehicle is decelerated promptly, and the vehicle stops on the roadside belt.
- an autonomous driving vehicle recognizes from a camera image that it is likely to cause an accident due to falling asleep or being distracted during manual driving, the accident is urgently avoided by, for example, an automatic braking function or an automatic steering operation.
- an example of an automatic driving vehicle having both the automatic driving mode and the manual driving mode has been described.
- the present invention is configured even for an automatic driving vehicle having no manual driving mode. it can.
- the automatic operation mode is always set, there is no need to perform control for switching from the manual operation mode to the automatic operation mode, or control for switching from the automatic operation mode to the manual operation mode, and the manual / automatic operation mode switching control unit 105 and the manual operation operation are not required.
- the detection unit 133 is not necessary. Instead, an automatic operation control unit (not shown) that performs control processing for automatic operation is provided. Further, there may be no accelerator pedal, brake pedal, shift lever, steering (handle), etc. necessary for manual driving.
- a brake pedal (a brake button, a touch panel input corresponding to a brake, a voice input, or the like may be provided), and only a brake operation may be manually handled. Further, even in the case of automatic driving, it may be possible to manually handle travel speed designation and travel speed adjustment (for example, 5 km faster / 5 km slower) by button input, touch panel input, voice input, or the like.
- the wireless communication unit has been described as using a mobile phone network.
- the wireless communication unit is not limited to this, and other devices such as Wi-Fi (Wireless Fidelity (registered trademark)) and WiMAX (Worldwide Interoperability for Microwave Access) are available. May be used.
- Wi-Fi Wireless Fidelity (registered trademark)
- WiMAX Worldwide Interoperability for Microwave Access
- the wireless communication unit is configured by these communication methods, the above-described embodiment can be realized without a mobile phone function. Therefore, the mobile phone terminal may be a mobile terminal such as a tablet without a mobile phone function. Good.
- the mobile phone terminal and mobile terminal are not limited to smartphones and tablets, but may be wristwatch-type terminals, eyeglass-type terminals, clothes-type terminals, ring-type terminals, notebook PCs, and the like.
- the re-passenger authentication information is a predetermined disembarkation face image, password, or ID, but the user uses the face image as re-passenger authentication information.
- the caller authentication information what the user uses when getting off the vehicle may be selectively set.
- the driver gets off from the automatic driving vehicle 1 .
- the driver is present from the state in which the autonomous driving vehicle 1 autonomously travels with the presence of a passenger but no driver.
- the same processing operation can be performed.
- the behavior after getting off is set by a passenger other than the driver, and the re-passenger is also the passenger.
- caller authentication information stored in the storage unit for caller authentication are not limited to the telephone numbers described above, but include biometric information such as a caller's face image, name, address, and date of birth. ID, password, e-mail address, etc. can be used alone or in combination.
- the caller sends this authentication information from the mobile phone terminal to the automatic driving vehicle 1 through the mobile phone network.
- the face image of the caller can be taken by a camera provided in the mobile phone terminal, and the other biometric information of the caller can be prestored in the memory of the mobile phone terminal.
- the caller authentication information such as the biometric information such as the caller's face image, name, and address can be input by the caller, and can be stored on the Internet cloud via the wireless communication unit 102. You can also use what is stored in
- examples of user authentication information stored in the storage unit for user authentication are not limited to the user's face image as in the above-described example.
- the user's fingerprint, voice ( Other biological information of the user such as voice), veins, and irises can also be used.
- user authentication information in addition to the user's biometric information, a name, address, date of birth, ID, password, telephone number and / or email address, etc., alone or in combination You may make it use.
- the user authentication information such as the biometric information such as the user's face image, the name, and the address can be stored in advance in the user authentication information storage unit 115 of the automatic driving vehicle 1. , What the user has entered can be used. Further, what is stored on the Internet cloud can be used via the wireless communication unit 102.
- the car navigation database 134 is installed in the autonomous driving vehicle 1, a part or all of those databases are installed not on the autonomous driving vehicle 1 but on the cloud of the Internet, and the wireless communication unit 102. Can also be used.
- some or all of the processing blocks of the car navigation function unit 113, the user authentication unit 114, the caller authentication unit 116, and the behavior control processing unit 120 are not via the automatic driving vehicle 1 but via the wireless communication unit 102. It can also be processed on the cloud.
- the cancellation of the behavior after getting off the vehicle is not mentioned, but a canceling means for receiving the cancellation of the behavior after getting off the vehicle received by the behavior setting receiving means after getting off the vehicle is provided. May be.
- the self-driving vehicle is provided with a registrant storage unit that accepts and stores in advance the registration of the person who has the right to cancel, and the canceling means is registered by the person who has requested cancellation. The cancellation of the behavior after getting off is accepted only when the registered person is stored in the person storage unit.
- car of the above-described embodiment has been described as an electric car, it may be a gasoline car, a hybrid car, a fuel cell car, or other driving type car.
- the motor vehicle of this invention includes a motorcycle, a three-wheeled motor vehicle, etc.
- the automobile of the present invention is not limited to ordinary cars, light cars, and trucks, but may be buses, tractors, dump trucks, excavators, forklifts, single-seat cars, motorbikes, and electric wheelchairs.
- the self-driving car of the present invention can also be used as a car that can be used temporarily, such as a rental car.
- a user rents a self-driving car at a rental car company that provides a rental car service
- the user is registered as a temporary owner (driver).
- One or more users can be registered.
- biometric information such as a user's face image, voice, fingerprint, and iris, and an ID and password issued by a rental car company are stored in the user authentication information storage unit and used as user authentication information.
- the telephone number and mail address of the user's mobile phone terminal, the ID and password issued by the rental car company are stored in the caller authentication information storage unit and used as caller authentication information.
- the user can use the information for user authentication at the time of registration as the information for authentication at the time of getting on again, and does not need to register biometric information of the user again when getting off the vehicle.
- the user is face-authenticated with a face image taken by the camera at all times or during a predetermined time (for example, 3 minutes) while using the rental car so that the user can be recognized as a legitimate user. Also good.
- face authentication is not performed, so troubles such as theft are detected and the temporary owner can use the communication function such as the mobile phone terminal function installed in the rental car. Make contact with a rental car company with a user.
- 110 may be reported to the police.
- an automatic driving car as a rental car is provided with a dedicated return button or an easy-to-understand display of the return on the touch panel so that the user can easily input “return” to improve convenience be able to.
- the self-driving car as a rental car receives "return”, it searches for the nearest parking lot of its own rental car company as a place for return, and moves autonomously (automatic driving) unattended.
- an autonomous driving car as a rental car moves unmanned and autonomously (automatic driving) to a return place designated in advance.
- the rental car fee is calculated including the electricity fee used at the end of use or after the end of the move (gasoline fee in the case of a gasoline car), and is settled with a credit card or debit card registered by the user at the time of rental. If the rental car has an electronic money settlement function, it can be settled (paid) with electronic money, and if it has a prepaid card settlement function, it can be settled (paid) with a prepaid card. In addition, it may be prepaid at the time of rental such as time rental, one-day rental, and multi-day rental. Also, an ETC (Electronic Toll Collection System) card can be used. After the return, the user authentication information and the caller authentication information of the user as a temporary owner are reset (erased), and the user is not allowed to board again.
- ETC Electronic Toll Collection System
- rental cars cannot be returned unless you go to a rental car company in the vicinity of the airport or station.
- the self-driving car rental car of this invention it is possible to drop off at any time convenient for the user. Convenient to. For example, if the departure time of a train or airplane is approaching, you can get off at the station or airport entrance (rotary) and jump on the train or check in to the airplane without going to the rental car company. If the rental car behaves as "return” after getting off, it will unmanned and autonomously run (automatic driving) and return to a predetermined return place such as the parking lot of the nearest rental car company.
- the self-driving car rental car of the present invention has a dedicated autonomous driving (automatic driving) road and a dedicated lane in the vicinity of the station and airport, and there is a place for return within the maintenance area. After getting off the rental car for return, it can move to the return place by unmanned autonomous driving (automatic driving) very smoothly and can be completed easily.
- the self-driving car rental car of the present invention may be an automatic driving car having both an automatic driving mode and a manual driving mode, or an automatic driving car without a manual driving mode.
- the invention can also be applied to car sharing.
- car sharing a member who shares the use of an autonomous vehicle is registered as a user of the autonomous vehicle (share car).
- biometric information such as a member's face image, voice, fingerprint, and iris, and an ID and password issued by a car sharing company are stored in the user authentication information storage unit and used as user authentication information.
- the telephone number and mail address of the member's mobile phone terminal, the ID and password issued by the car sharing company are stored in the caller authentication information storage unit and used as caller authentication information.
- the registered user authentication information and caller authentication information can be used exclusively by the share car as a temporary exclusive user, and the share car can be freely used. Can be used. Even if it is not your own car, the behavior after getting off can be set flexibly according to your purpose of use, so you can treat it like your own car and it is very useful.
- a self-driving car as a share car is provided with a button dedicated to (use) termination, or a (use) end indication is provided in an easy-to-understand manner on the touch panel, so that the user can easily input "(use) end” It is possible to improve convenience.
- the self-driving car as a share car moves unattended and autonomously (automatic driving) to a predetermined parking lot of the share car or a place where the next user calls.
- the user authentication information and the caller authentication information as a temporary exclusive user of the user are reset (erased), and the user is not allowed to re-ride. If you wish to re-ride, apply for a shared car again.
- the present invention can also be applied to taxis and hires that are provided by taxi companies or higher service companies and that run unattended and autonomously (automatic driving).
- the taxi or hire has no manual operation mode, or even if there is a manual operation mode, the user cannot switch from the automatic operation mode.
- the taxi or hire can be used as user authentication information and caller authentication information by entering personal information such as the user's biometric information or issuing a one-time password when the user gets on. Is stored in the storage unit.
- the user can input the behavior after getting off to the behavior setting accepting means after getting off.
- the user authentication information and the caller authentication information stored when the user gets on the board are used for authentication.
- the user can get off at any time according to his / her schedule.
- the use of the taxi or hire ends, the user gets off the vehicle and inputs “end of use” or simply “end” to the behavior setting receiving means after getting off.
- the taxi or hire will be provided with a (use) end-only button, or a (use) end display will be provided on the touch panel in an easy-to-understand manner so that the user can easily input "(use) end”.
- convenience can be improved.
- the taxi or hire accepts "(use) end it searches the nearest taxi company or hire service company's parking lot as the end of use and moves autonomously (automatic driving) unattended. To do.
- the taxi or hire moves unmanned and autonomously (automatic driving) to a place where the next passenger (user) waiting in advance is waiting.
- the taxi searches for the nearest taxi boarding location, and the hire searches for the nearest hire boarding location as the location at the end of use, and moves autonomously (automatic driving) unattended. You may wait for passengers (users) to board.
- a taxi or hire has an electronic money payment function or a prepaid card payment function
- the user can pay a usage fee (boarding fee) with electronic money or a prepaid card when getting off.
- a usage fee boarding fee
- the credit card number and debit card number of the user can be registered at the time of boarding, so that the credit card can be automatically registered after getting off. You can settle (pay) the usage fee (ride fee) from the debit card. Further, it may be prepaid at the time of boarding with a fixed fee for every predetermined time use (for example, 1000 yen within 10 minutes) or a fixed fee for every predetermined distance use (for example, 1000 yen within 1 km).
- Prepayment may be performed with electronic money or a prepaid card, or a cash storage machine is installed to enable cash payment.
- a cash storage machine is installed to enable cash payment.
- fishing may be impossible for the sake of simplicity, or only coins (for example, only 500 Yen coins) may be accepted.
- An ETC card can also be used.
- taxis and hires for autonomous driving vehicles of the present invention are autonomously driven in the area where there are facilities such as Olympic and soccer World Cup competition venues and athlete villages, and in large facilities such as an amusement park ( (Automatic operation) If dedicated roads and lanes are maintained, passengers (users) can be dropped freely in the maintenance area, and after the use is completed, unmanned It moves very smoothly by autonomous driving (automatic driving) and can efficiently wait for the next passenger.
- facilities such as Olympic and soccer World Cup competition venues and athlete villages
- amusement park amusement park
- a new service is available in the above-mentioned maintenance area, where a number of self-driving cars dedicated to movement within the maintenance area are deployed so that users can use the service anytime, anywhere, for a fee or free of charge. .
- This service is very convenient because the user does not have to worry about the means of transportation within the maintenance area.
- an ultra-small vehicle such as a single-seat vehicle or a two-seater vehicle is a very useful moving means and transport means from the viewpoint of space efficiency.
Landscapes
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Accounting & Taxation (AREA)
- Development Economics (AREA)
- Finance (AREA)
- Mechanical Engineering (AREA)
- Strategic Management (AREA)
- Automation & Control Theory (AREA)
- Aviation & Aerospace Engineering (AREA)
- Transportation (AREA)
- Economics (AREA)
- Theoretical Computer Science (AREA)
- General Business, Economics & Management (AREA)
- Marketing (AREA)
- Game Theory and Decision Science (AREA)
- Entrepreneurship & Innovation (AREA)
- Evolutionary Computation (AREA)
- Combustion & Propulsion (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Chemical & Material Sciences (AREA)
- Medical Informatics (AREA)
- Traffic Control Systems (AREA)
- Navigation (AREA)
- Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
- Geophysics And Detection Of Objects (AREA)
- Chair Legs, Seat Parts, And Backrests (AREA)
- Seats For Vehicles (AREA)
Abstract
Description
自律走行を行う自動運転モードを有し、
運転者または乗車者の降車時に入力される前記自動運転車の降車後の振る舞いの設定を受け付けるための振る舞い設定受付手段と、
前記設定に基づいて前記振る舞いの実行を制御するための制御手段と、
を備える自動運転車を提供する。
図1は、この発明の実施形態の自動運転車1の電子制御回路部10のハードウエア構成例を示すブロック図である。なお、この実施形態の自動運転車1は、電気自動車の場合の例である。ただし、バッテリーは、図1では図示を省略した。
前述したように降車後振る舞い記憶部119には、利用者が指定したい降車後振る舞いが予め記憶保持される。この降車後振る舞いとしては、自動運転車1に予め自動車会社などによりデフォルトとして登録されているものの中から選択することもできるし、利用者が設定して記憶させるようにすることもできる。また、無線通信部102を介して、インターネットのクラウド上に記憶したものを利用することもできる。
(a)タイマー回路121に対するタイマー時間の設定、
(b)音声入力による設定、例えば「ちょっとだけ用事」、「トイレ」、「食事」、「コンサート」、「野球観戦」、「サッカー観戦」、「大相撲観戦」、「2,3分」、「1時間ほど」など、
(c)表示部111に表示された所定の用事のリストの中からの選択
などが挙げられる。
・「大相撲観戦」 本場所の場合は、ほぼ18時に終了するので、その終了時刻を目安に所定時間を把握する。
・「野球観戦」 プロ野球の場合、開始から約3時間、高校野球の場合、開始から約2時間が目安である。延長戦になると時間が加算される。インターネットを通じて所定のサイトから放送中継などを入手すれば、試合終了時刻が把握できる。
・「サッカー観戦」 前半45分 ハーフタイム15分 後半45分 ロスタイム少々を目安に所定時間を把握する。延長戦になると時間が加算される。インターネットを通じて所定のサイトから放送中継などを入手すれば、試合終了時刻が把握できる。
・「コンサート」 主催者側発表の終演時間が目安である。降車時に、利用者により、その終演時間が入力され、その時刻を目安に所定時間を把握する。
・「映画鑑賞」 終演時間は確定している。降車時に、利用者により、その終演時間が入力され、その時刻を目安に所定時間を把握する。
・「ショッピングセンター」 経験値・統計値を基に所定時間を、例えば2時間として把握する。
・「デパート」 経験値・統計値を基に所定時間を、例えば2時間として把握する。
・「量販店(家電・コンピュータ)」 経験値・統計値を基に所定時間を、例えば1時間として把握する。
・「本屋」 経験値・統計値を基に所定時間を、例えば30分として把握する。
・「花屋」 経験値・統計値を基に所定時間を、例えば15分として把握する。
・「小さい商店」 経験値・統計値を基に所定時間を、例えば15分として把握する。
・「コンビニ」 経験値・統計値を基に所定時間を、例えば15分として把握する。
・「郵便局」 待ち時間によるが、例えば10分として把握する。
・「銀行」 待ち時間によるが、ATMなら5分、窓口なら20分として把握する。
・「塾」 終了時間は確定している。降車時に、利用者により、その終了時間が入力され、その時刻を目安に所定時間、例えば2時間を把握する。
・「レストラン」 例えば昼食30分、夕食2時間として、所定時間を把握する。
・「喫茶」 例えば1時間として、所定時間を把握する。
・「釣堀」 例えば2時間として、所定時間を把握する。
・「神社・仏閣」 初詣など。例えば1時間として、所定時間を把握する。
・「動物園・水族館」 例えば3時間として、所定時間を把握する。
・「車両進入禁止エリア」 「ここから先、車はご遠慮ください」のような表示があるが、その先が見てみたい場合、車を降りて見てくる。例えば15分として、所定時間を把握する。
次に、自動運転車1に乗車していた利用者、この例では、特に運転者が降車をしようとする際における自動運転車1の制御部101の処理動作の概要について説明する。
次に、ステップS7の降車後振る舞い設定受付処理の流れの例について説明する。図4~図6は、降車後振る舞い設定受付処理の流れの例を説明するためのフローチャートであり、この例では、図2に示したリストの各降車後振る舞いのそれぞれが選択されたときの設定受付処理を想定して説明する。
次に、図3のステップS11における利用者の降車後の処理動作例について説明する。図7~図12は、運転者が降車して、運転者が不在となったときの自動運転車1の電子制御回路部10の制御部101が実行する降車後振る舞いの制御処理動作の流れの例を説明するためのフローチャートである。この例も、図2に示したリストの各降車後振る舞いのそれぞれが設定受付されたときのそれぞれに対応する処理を想定して説明する。
以上説明したようにして、上述の実施形態の自動運転車によれば、自動運転車に乗車中の利用者が、降車時に、降車後の自分の行動態様や行動目的に応じて臨機応変に自動運転車に指示を与えることができ、非常に便利である。例えば利用者が自動運転車に乗車して自宅に帰宅したときに、自動運転車に対して降車後振る舞いの設定をすることにより、自宅から離れている駐車場に自動運転モードにより移動させることが簡単にできる。
上述の実施形態では、手動運転モードから自動運転モードに切り替える制御は、運転者によるタッチパネル112を通じた選択操作入力で行われ、自動運転モードから手動運転モードに切り替える制御は、タッチパネル112を通じた選択操作入力や、アクセルペダル操作、ブレーキペダル操作、シフトレバー操作またはステアリング操作で行われたが、これらの切り替える制御は自動で行うこともできる。例えば、高速道路に進入したら、それを自動運転車のカーナビやGPSで検出し、自動的に自動運転モードにし、逆に、高速道路から一般道路に出るときに、手動運転モードに切り替えることができる。自動運転可能なエリアが規定される場合は、そのエリアを走行しているか否かを自動運転車のカーナビやGPSで判別し、当該エリアを走行している場合は自動運転モード、当該エリア外を走行している場合は手動運転モードとなるよう切り替えてもよい。
2 利用者
10 電子制御回路部
107 カメラ群
108 センサ群
110 現在位置検出部
111 表示部
112 タッチパネル
113 カーナビ機能部
114 利用者認証部
115 利用者認証用情報記憶部
116 呼出者認証部
117 呼出者認証用情報記憶部
118 降車後振る舞い設定受付部
119 降車後振る舞い記憶部
120 振る舞い制御処理部
121 タイマー回路
122 音声入出力部
135 マイクロフォン
136 スピーカ
Claims (20)
- 自律走行を行う自動運転モードを有し、
運転者または乗車者の降車時に入力される前記自動運転車の降車後の振る舞いの設定を受け付けるための振る舞い設定受付手段と、
前記設定に基づいて前記振る舞いの実行を制御するための制御手段と、
を備える自動運転車。 - 前記運転者の乗車の有無を判別する運転者有無判別手段及び/または前記乗車者の乗車の有無を判別する乗車者有無判別手段を備え、
前記制御手段は、前記運転者有無判別手段または乗車者有無判別手段で前記運転者または乗車者が乗車していないと判別されたときに、前記設定に基づいて前記振る舞いの実行を制御する
ことを特徴とする請求項1に記載の自動運転車。 - 前記運転者有無判別手段は、画像認識装置、音声認識装置、運転席に設けられた着座センサ、及びハンドルまたはタッチパネルに設けられたタッチセンサのうちのいずれか1つ以上を有する
ことを特徴とする請求項1または請求項2に記載の自動運転車。 - 前記乗車者有無判別手段は、画像認識装置、音声認識装置、運転席以外の乗車席に設けられた着座センサ、及びハンドルまたはタッチパネルに設けられたタッチセンサのうちのいずれか1つ以上を有する
ことを特徴とする請求項1~請求項3いずれかに記載の自動運転車。 - 前記運転者または乗車者の顔画像、音声、指紋、静脈、虹彩などの生体情報を記憶する生体情報記憶部と、
前記運転者または乗車者の降車後の再乗車時に、前記運転者または乗車者の生体情報を取得する生体情報取得部と、
前記取得された生体情報と前記記憶された生体情報とを比較することで認証を行う生体情報認証手段と、
を備える
ことを特徴とする請求項1~請求項4のいずれかに記載の自動運転車。 - 1種または複数種の振る舞いを記憶する振る舞い記憶部を備え、
前記振る舞い設定受付手段は、前記振る舞い記憶部に記憶される前記振る舞いのうちから選択された振る舞いの設定を受け付ける
ことを特徴とする請求項1~請求項5のいずれかに記載の自動運転車。 - メッセージを表示する表示画面及び/またはメッセージを音響再生するスピーカと、 前記設定を受け付ける際に用いられるメッセージを記憶するメッセージ記憶部とを備え、
前記振る舞い設定受付手段は、前記メッセージ記憶部に記憶されているメッセージを用いた対話形式で前記設定を受け付ける
ことを特徴とする請求項1~請求項6のいずれかに記載の自動運転車。 - 前記設定を入力した運転者または乗車者に対する通信のための連絡先を記憶する通信先記憶部と、
前記制御手段が前記振る舞いの実行を制御する前に前記通信を行う通信手段と、
を備える
ことを特徴とする請求項1~請求項7のいずれかに記載の自動運転車。 - 前記制御手段は、前記設定を入力した運転者または乗車者の承認を得た後に、前記振る舞いの実行を制御する
ことを特徴とする請求項8に記載の自動運転車。 - 前記設定を入力した運転者または乗車者に対する通信のための連絡先を記憶する通信先記憶部と、
前記運転者または乗車者から前記自動運転車が呼出を受けた時に、前記運転者または乗車者の連絡先が前記通信先記憶部に記憶された連絡先と一致するか否かで認証を行う通信先認証手段と、
を備える
ことを特徴とする請求項1~請求項9のいずれかに記載の自動運転車。 - 手動運転を行う手動運転モードを有し、
前記運転者が降車時に、前記自動運転モードにするとともに、前記手動運転モードへの切替えは禁止するよう制御する
ことを特徴とする請求項1~請求項10のいずれかに記載の自動運転車。 - 一時的利用が可能な自動運転車であって、
前記振る舞いが前記自動運転車の返却または利用の終了を含む
ことを特徴とする請求項1~請求項11のいずれかに記載の自動運転車。 - 前記自動運転車を一時的に利用する利用者の顔画像、音声、指紋、静脈、虹彩などの生体情報を登録する一時的生体情報登録手段と、
前記自動運転車の返却時または利用の終了時に、前記登録されていた前記利用者の生体情報を抹消する生体情報抹消手段と、
を備える
ことを特徴とする請求項12に記載の自動運転車。 - 前記自動運転車を一時的に利用する利用者に対する通信のための連絡先を登録する一時的通信先登録手段と、
前記自動運転車の返却時または利用の終了時に、前記登録されていた前記利用者の通信先を抹消する通信先抹消手段と、
を備える
ことを特徴とする請求項12または請求項13に記載の自動運転車。 - 前記一時的利用が可能な自動運転車を提供する提供会社が発行するID及び/またはパスワードを前記一時的に利用する利用者に発行する発行手段と、
前記発行されたID及び/またはパスワードで認証を行うID認証手段と、
を備える
ことを特徴とする請求項12~請求項14のいずれかに記載の自動運転車。 - 利用実績に応じた利用料金の決済手段を備える
ことを特徴とする請求項12~請求項15のいずれかに記載の自動運転車。 - 利用可能時間または利用可能距離に応じた利用料金、または固定利用料金であって、
前記利用料金の利用開始前の決済手段を備える
ことを特徴とする請求項12~請求項16のいずれかに記載の自動運転車。 - 前記一時的利用が可能な自動運転車は、レンタカー会社から提供されるレンタカー、タクシー会社から提供されるタクシー、ハイヤー会社から提供されるハイヤー、カーシェアリング会社から提供される自動運転車である
ことを特徴とする請求項12~請求項17のいずれかに記載の自動運転車。 - 自律走行を行う自動運転モードを有する自動運転車が備えるコンピュータを、
降車する運転者または乗車者によって入力される前記自動運転車の降車後の振る舞いの設定を受け付けるための振る舞い設定受付手段、及び、前記設定に基づいて前記振る舞いの実行を制御するための制御手段
として機能させるための自動運転車用プログラム。 - 一時的利用が可能な自動運転車であって、
前記振る舞いが前記自動運転車の返却または利用の終了を含む
ことを特徴とする請求項19に記載の自動運転車用プログラム。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/305,254 US10139824B2 (en) | 2014-04-30 | 2015-04-16 | Automatic driving vehicle and program for automatic driving vehicle |
JP2015540769A JPWO2015166811A1 (ja) | 2014-04-30 | 2015-04-16 | 自動運転車及び自動運転車用プログラム |
US16/157,797 US10656648B2 (en) | 2014-04-30 | 2018-10-11 | Automatic driving vehicle and program for automatic driving vehicle |
US16/848,387 US11397435B2 (en) | 2014-04-30 | 2020-04-14 | Automatic driving vehicle and program for automatic driving vehicle |
US17/841,920 US20220308584A1 (en) | 2014-04-30 | 2022-06-16 | Automatic driving vehicle and program for automatic driving vehicle |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014-093515 | 2014-04-30 | ||
JP2014093515 | 2014-04-30 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/305,254 A-371-Of-International US10139824B2 (en) | 2014-04-30 | 2015-04-16 | Automatic driving vehicle and program for automatic driving vehicle |
US16/157,797 Continuation US10656648B2 (en) | 2014-04-30 | 2018-10-11 | Automatic driving vehicle and program for automatic driving vehicle |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015166811A1 true WO2015166811A1 (ja) | 2015-11-05 |
Family
ID=54358545
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2015/061697 WO2015166811A1 (ja) | 2014-04-30 | 2015-04-16 | 自動運転車及び自動運転車用プログラム |
Country Status (3)
Country | Link |
---|---|
US (4) | US10139824B2 (ja) |
JP (7) | JPWO2015166811A1 (ja) |
WO (1) | WO2015166811A1 (ja) |
Cited By (42)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106549981A (zh) * | 2017-01-13 | 2017-03-29 | 邹城众达知识产权咨询服务有限公司 | 一种基于大数据云计算服务的无方向盘智能网约车系统及其运营方法 |
JP2017102007A (ja) * | 2015-12-01 | 2017-06-08 | 株式会社デンソー | 報知処理装置 |
JP2017136977A (ja) * | 2016-02-04 | 2017-08-10 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP2017197066A (ja) * | 2016-04-28 | 2017-11-02 | 本田技研工業株式会社 | 車両制御システム、交通制御システム、車両制御方法、および車両制御プログラム |
WO2017213079A1 (ja) * | 2016-06-07 | 2017-12-14 | 株式会社デンソー | 車両用装置、及び車両用システム |
JP2017228266A (ja) * | 2016-06-21 | 2017-12-28 | バイドゥ オンライン ネットワーク テクノロジー (ベイジン) カンパニー リミテッド | スマート車両に用いられる局部軌跡計画方法及び装置 |
JP6274699B1 (ja) * | 2017-03-16 | 2018-02-07 | 三菱ロジスネクスト株式会社 | 配車システム |
JP2018024485A (ja) * | 2016-08-08 | 2018-02-15 | 三菱ロジスネクスト株式会社 | 荷役車両呼出装置、荷役車両システム、荷役車両呼出方法及び荷役車両呼出装置の制御プログラム |
JP2018036754A (ja) * | 2016-08-30 | 2018-03-08 | トヨタ自動車株式会社 | 車両の制御装置 |
JP6292646B1 (ja) * | 2017-03-16 | 2018-03-14 | 三菱ロジスネクスト株式会社 | 配車システム |
JP2018067172A (ja) * | 2016-10-20 | 2018-04-26 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
KR20180053081A (ko) * | 2016-11-11 | 2018-05-21 | 엘지전자 주식회사 | 자율 주행 차량 및 그 제어방법 |
WO2018102425A1 (en) * | 2016-12-02 | 2018-06-07 | Starsky Robotics, Inc. | Vehicle control system and method of use |
JP6341525B1 (ja) * | 2017-06-22 | 2018-06-13 | 三菱ロジスネクスト株式会社 | 配車システムおよび配車方法 |
WO2018116358A1 (ja) * | 2016-12-19 | 2018-06-28 | 株式会社Doog | 移動支援ロボットの運行管理システム |
WO2018123032A1 (ja) * | 2016-12-28 | 2018-07-05 | 株式会社Doog | 移動支援ロボットの運行管理システム |
JP2018103858A (ja) * | 2016-12-27 | 2018-07-05 | トヨタ自動車株式会社 | 自動運転装置 |
CN108372858A (zh) * | 2017-01-30 | 2018-08-07 | 松下电器(美国)知识产权公司 | 自动驾驶车辆的控制装置、控制方法以及记录有程序的记录介质 |
EP3358306A1 (en) | 2017-01-27 | 2018-08-08 | Panasonic Intellectual Property Corporation of America | Vehicle control apparatus and vehicle control method |
GB2561054A (en) * | 2017-01-27 | 2018-10-03 | Ford Global Tech Llc | Self parking vehicle |
US10209718B2 (en) | 2017-03-14 | 2019-02-19 | Starsky Robotics, Inc. | Vehicle sensor system and method of use |
JP2019067439A (ja) * | 2015-05-26 | 2019-04-25 | ウェイモ エルエルシー | 自律車両のためのフォールバック要求 |
JP2019073279A (ja) * | 2018-11-05 | 2019-05-16 | 本田技研工業株式会社 | 車両制御システム、車両制御方法、および車両制御プログラム |
KR20190067430A (ko) * | 2017-12-07 | 2019-06-17 | 엘케이시스(주) | 자율 주행 차량을 이용한 자동 선적 방법 및 장치 |
JP2019164444A (ja) * | 2018-03-19 | 2019-09-26 | トヨタ自動車株式会社 | 駐車支援装置 |
JP2019218057A (ja) * | 2019-09-05 | 2019-12-26 | みこらった株式会社 | 自動車及び自動車用プログラム |
EP3551510A4 (en) * | 2016-12-09 | 2020-01-29 | Gentex Corporation | USER AUTHENTICATION ACTIVATION SYSTEMS AND METHODS |
JP2020038585A (ja) * | 2018-09-06 | 2020-03-12 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
JP2020042850A (ja) * | 2019-11-28 | 2020-03-19 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
JP2020077385A (ja) * | 2019-09-25 | 2020-05-21 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
CN111445722A (zh) * | 2019-01-16 | 2020-07-24 | 丰田自动车株式会社 | 自动驻车管理装置 |
US10726360B2 (en) | 2016-11-04 | 2020-07-28 | Honda Motor Co., Ltd. | Vehicle control system, vehicle control method, and vehicle control program |
JP2020119327A (ja) * | 2019-01-24 | 2020-08-06 | トヨタ自動車株式会社 | 配車装置および配車方法 |
WO2020209402A1 (ko) * | 2019-04-10 | 2020-10-15 | 엘케이시스(주) | 자율 주행 차량을 이용한 자동 선적 방법 및 장치 |
US10884413B2 (en) | 2016-08-18 | 2021-01-05 | Motional Ad Llc | Hailing a vehicle |
JP2021004029A (ja) * | 2020-08-13 | 2021-01-14 | みこらった株式会社 | 自動車及び自動車用プログラム |
WO2021085446A1 (ja) * | 2019-10-29 | 2021-05-06 | Whill株式会社 | 施設内におけるシステムおよび電動モビリティ |
JP2021093211A (ja) * | 2021-03-08 | 2021-06-17 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
JP2022046490A (ja) * | 2020-08-13 | 2022-03-23 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP2022171722A (ja) * | 2017-01-30 | 2022-11-11 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 自動運転車両の制御装置、制御方法及びプログラム |
US11827503B2 (en) | 2020-03-18 | 2023-11-28 | Crown Equipment Corporation | Adaptive acceleration for materials handling vehicle |
US11947353B1 (en) | 2015-05-26 | 2024-04-02 | Waymo Llc | Non-passenger requests for autonomous vehicles |
Families Citing this family (122)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10139824B2 (en) * | 2014-04-30 | 2018-11-27 | Mico Latta Inc. | Automatic driving vehicle and program for automatic driving vehicle |
CN105094767B (zh) * | 2014-05-06 | 2019-02-12 | 华为技术有限公司 | 自动驾驶车辆调度方法、车辆调度服务器及自动驾驶车辆 |
CN107249954B (zh) * | 2014-12-29 | 2020-07-10 | 罗伯特·博世有限公司 | 用于使用个性化驾驶简档操作自主车辆的系统和方法 |
JP6375237B2 (ja) * | 2015-01-28 | 2018-08-15 | 日立オートモティブシステムズ株式会社 | 自動運転制御装置 |
KR101709521B1 (ko) * | 2015-07-30 | 2017-03-09 | 주식회사 한글과컴퓨터 | 무인 스마트카를 이용한 공익서비스 시스템 및 방법 |
EP3345138A1 (en) * | 2015-09-04 | 2018-07-11 | Robert Bosch GmbH | Access and control for driving of autonomous vehicle |
DE102015225157A1 (de) * | 2015-12-14 | 2017-06-14 | Robert Bosch Gmbh | Verfahren zum Übertragen, Empfangen und Verarbeiten von Datenwerten, sowie eine Übertragungs- und Empfangsvorrichtung |
US9637117B1 (en) * | 2016-01-12 | 2017-05-02 | Ford Global Technologies, Llc | System and method for automatic activation of autonomous parking |
US20170253237A1 (en) * | 2016-03-02 | 2017-09-07 | Magna Electronics Inc. | Vehicle vision system with automatic parking function |
US10035519B2 (en) * | 2016-03-15 | 2018-07-31 | GM Global Technology Operations LLC | System and method for autonomous vehicle driving behavior modification |
US11400997B2 (en) | 2016-05-23 | 2022-08-02 | Indian Motorcycle International, LLC | Display systems and methods for a recreational vehicle |
CN106440192B (zh) * | 2016-09-19 | 2019-04-09 | 珠海格力电器股份有限公司 | 一种家电控制方法、装置、系统及智能空调 |
CN107979410B (zh) * | 2016-10-21 | 2020-09-04 | 菜鸟智能物流控股有限公司 | 一种自动进入、离开乘坐设备的方法及相关装置 |
DE112017006295T5 (de) * | 2017-01-10 | 2019-08-29 | Ford Global Technologies, Llc | Abholen und absetzen von fluggästen an einem flughafen unter verwendung eines autonomen fahrzeugs |
US10369988B2 (en) | 2017-01-13 | 2019-08-06 | Ford Global Technologies, Llc | Autonomous parking of vehicles inperpendicular parking spots |
CN107187420B (zh) * | 2017-04-21 | 2019-07-02 | 百度在线网络技术(北京)有限公司 | 无人驾驶汽车及其启动方法、计算机可读介质 |
CN107293058B (zh) * | 2017-05-05 | 2018-03-06 | 北京摩拜科技有限公司 | 车辆管理方法、系统、服务器及车辆 |
JP6971490B2 (ja) * | 2017-05-16 | 2021-11-24 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP6628441B2 (ja) * | 2017-05-16 | 2020-01-08 | みこらった株式会社 | 自動車及び自動車用プログラム |
US10683034B2 (en) | 2017-06-06 | 2020-06-16 | Ford Global Technologies, Llc | Vehicle remote parking systems and methods |
US10926737B2 (en) * | 2017-06-14 | 2021-02-23 | Sumitomo Electric Industries, Ltd. | Extra-vehicular communication device, communication control method, and communication control program |
US10234868B2 (en) | 2017-06-16 | 2019-03-19 | Ford Global Technologies, Llc | Mobile device initiation of vehicle remote-parking |
US10775781B2 (en) | 2017-06-16 | 2020-09-15 | Ford Global Technologies, Llc | Interface verification for vehicle remote park-assist |
US10585430B2 (en) * | 2017-06-16 | 2020-03-10 | Ford Global Technologies, Llc | Remote park-assist authentication for vehicles |
JP6341524B1 (ja) * | 2017-06-22 | 2018-06-13 | 三菱ロジスネクスト株式会社 | 配車システムおよび配車方法 |
JP6982422B2 (ja) * | 2017-07-10 | 2021-12-17 | 株式会社Subaru | 車両の走行支援装置 |
US20190018418A1 (en) * | 2017-07-13 | 2019-01-17 | Horng-Huei Tseng | Vehicle summon system |
KR102160968B1 (ko) * | 2017-07-17 | 2020-09-29 | 한국전자통신연구원 | 자율주행 로봇 장치 및 자율 주행 방법 |
US11126191B2 (en) | 2017-08-07 | 2021-09-21 | Panasonic Intellectual Property Corporation Of America | Control device and control method |
US10816975B2 (en) * | 2017-08-09 | 2020-10-27 | Toyota Motor Engineering & Manufacturing North America, Inc. | Autonomous acceleration profile feedback system |
US10627815B2 (en) * | 2017-08-22 | 2020-04-21 | Waymo Llc | Context aware stopping for autonomous vehicles |
DE102017008084A1 (de) | 2017-08-25 | 2019-02-28 | Daimler Ag | Verfahren zur Vergabe von Zugangs- und Fahrberechtigungen |
US11507894B2 (en) | 2017-09-08 | 2022-11-22 | Beijing Didi Infinity Technology And Development Co., Ltd. | System and method for ride order dispatching |
JP6866811B2 (ja) | 2017-09-14 | 2021-04-28 | トヨタ自動車株式会社 | 自動運転車両 |
US10580304B2 (en) | 2017-10-02 | 2020-03-03 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for voice controlled autonomous parking |
US10281921B2 (en) | 2017-10-02 | 2019-05-07 | Ford Global Technologies, Llc | Autonomous parking of vehicles in perpendicular parking spots |
US10787174B2 (en) | 2017-10-13 | 2020-09-29 | Toyota Motor Engineering & Manufacutring North America, Inc. | Automatic vehicle driving mode system |
US20190129413A1 (en) * | 2017-10-26 | 2019-05-02 | GM Global Technology Operations LLC | Flexible remote vehicle control |
JP6885298B2 (ja) * | 2017-10-27 | 2021-06-09 | トヨタ自動車株式会社 | 自動運転車両 |
US10627811B2 (en) | 2017-11-07 | 2020-04-21 | Ford Global Technologies, Llc | Audio alerts for remote park-assist tethering |
JP6958273B2 (ja) * | 2017-11-16 | 2021-11-02 | トヨタ自動車株式会社 | ドライバレス輸送システム |
US10336320B2 (en) | 2017-11-22 | 2019-07-02 | Ford Global Technologies, Llc | Monitoring of communication for vehicle remote park-assist |
US10578676B2 (en) | 2017-11-28 | 2020-03-03 | Ford Global Technologies, Llc | Vehicle monitoring of mobile device state-of-charge |
KR102058049B1 (ko) * | 2017-12-01 | 2020-01-22 | 엘지전자 주식회사 | 차량의 운행 시스템 |
JP2019099086A (ja) * | 2017-12-07 | 2019-06-24 | Joyson Safety Systems Japan株式会社 | 乗員検知装置及び警告装置 |
US20190185012A1 (en) | 2017-12-18 | 2019-06-20 | PlusAI Corp | Method and system for personalized motion planning in autonomous driving vehicles |
US11273836B2 (en) * | 2017-12-18 | 2022-03-15 | Plusai, Inc. | Method and system for human-like driving lane planning in autonomous driving vehicles |
US11130497B2 (en) | 2017-12-18 | 2021-09-28 | Plusai Limited | Method and system for ensemble vehicle control prediction in autonomous driving vehicles |
JP6925951B2 (ja) * | 2017-12-19 | 2021-08-25 | アルパイン株式会社 | 車載システム及びエネルギー源供給システム |
US10974717B2 (en) | 2018-01-02 | 2021-04-13 | Ford Global Technologies, I.LC | Mobile device tethering for a remote parking assist system of a vehicle |
US10585431B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10688918B2 (en) | 2018-01-02 | 2020-06-23 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10737690B2 (en) | 2018-01-02 | 2020-08-11 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10583830B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10814864B2 (en) | 2018-01-02 | 2020-10-27 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US11148661B2 (en) | 2018-01-02 | 2021-10-19 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10684773B2 (en) | 2018-01-03 | 2020-06-16 | Ford Global Technologies, Llc | Mobile device interface for trailer backup-assist |
KR20190093763A (ko) | 2018-01-12 | 2019-08-12 | 현대자동차주식회사 | 자율주차시스템의 안전 제어 장치 및 방법 |
US10747218B2 (en) | 2018-01-12 | 2020-08-18 | Ford Global Technologies, Llc | Mobile device tethering for remote parking assist |
US10917748B2 (en) | 2018-01-25 | 2021-02-09 | Ford Global Technologies, Llc | Mobile device tethering for vehicle systems based on variable time-of-flight and dead reckoning |
US10684627B2 (en) | 2018-02-06 | 2020-06-16 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for position aware autonomous parking |
US11188070B2 (en) | 2018-02-19 | 2021-11-30 | Ford Global Technologies, Llc | Mitigating key fob unavailability for remote parking assist systems |
US10507868B2 (en) | 2018-02-22 | 2019-12-17 | Ford Global Technologies, Llc | Tire pressure monitoring for vehicle park-assist |
US20210114626A1 (en) * | 2018-02-22 | 2021-04-22 | Honda Motor Co., Ltd. | Vehicle control system, vehicle control device, and vehicle control method |
CN111602186B (zh) * | 2018-02-22 | 2022-06-17 | 本田技研工业株式会社 | 车辆控制系统、车辆控制装置、及车辆控制方法 |
JP7080684B2 (ja) * | 2018-03-16 | 2022-06-06 | 本田技研工業株式会社 | 車両利用システム、および車両利用方法 |
WO2019190481A1 (en) * | 2018-03-27 | 2019-10-03 | Ford Global Technologies, Llc | Display for an autonomous taxi |
US10732622B2 (en) | 2018-04-05 | 2020-08-04 | Ford Global Technologies, Llc | Advanced user interaction features for remote park assist |
US10493981B2 (en) | 2018-04-09 | 2019-12-03 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10793144B2 (en) | 2018-04-09 | 2020-10-06 | Ford Global Technologies, Llc | Vehicle remote park-assist communication counters |
US10759417B2 (en) | 2018-04-09 | 2020-09-01 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10683004B2 (en) | 2018-04-09 | 2020-06-16 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
JP7259211B2 (ja) | 2018-05-07 | 2023-04-18 | トヨタ自動車株式会社 | 情報処理装置、情報処理方法、及びプログラム |
KR102175298B1 (ko) * | 2018-06-14 | 2020-11-06 | 엘지전자 주식회사 | 이동 로봇의 동작 방법 |
JP6962280B2 (ja) * | 2018-06-26 | 2021-11-05 | 株式会社デンソー | 車両制御方法、車両制御システム、及び車両制御装置 |
US10909866B2 (en) | 2018-07-20 | 2021-02-02 | Cybernet Systems Corp. | Autonomous transportation system and methods |
JP2020021221A (ja) * | 2018-07-31 | 2020-02-06 | クラリオン株式会社 | 駐車場管理装置および駐車場管理方法 |
KR102625398B1 (ko) * | 2018-08-17 | 2024-01-17 | 현대자동차주식회사 | 차량 및 그 제어방법 |
US10384605B1 (en) | 2018-09-04 | 2019-08-20 | Ford Global Technologies, Llc | Methods and apparatus to facilitate pedestrian detection during remote-controlled maneuvers |
US10821972B2 (en) | 2018-09-13 | 2020-11-03 | Ford Global Technologies, Llc | Vehicle remote parking assist systems and methods |
US10717432B2 (en) | 2018-09-13 | 2020-07-21 | Ford Global Technologies, Llc | Park-assist based on vehicle door open positions |
US10752258B2 (en) * | 2018-09-17 | 2020-08-25 | GM Global Technology Operations LLC | Apparatus and method for audible driver confirmation for maneuvers in an autonomous vehicle |
JP7070275B2 (ja) * | 2018-09-20 | 2022-05-18 | トヨタ自動車株式会社 | 乗員支援装置 |
JP7084835B2 (ja) * | 2018-09-21 | 2022-06-15 | 日産自動車株式会社 | 自動運転方法及び車両 |
US11200807B2 (en) * | 2018-09-24 | 2021-12-14 | Here Global B.V. | Method and apparatus for detecting an availability of a vehicle based on parking search behaviors |
US10529233B1 (en) | 2018-09-24 | 2020-01-07 | Ford Global Technologies Llc | Vehicle and method for detecting a parking space via a drone |
US10967851B2 (en) | 2018-09-24 | 2021-04-06 | Ford Global Technologies, Llc | Vehicle system and method for setting variable virtual boundary |
US10908603B2 (en) | 2018-10-08 | 2021-02-02 | Ford Global Technologies, Llc | Methods and apparatus to facilitate remote-controlled maneuvers |
US10628687B1 (en) | 2018-10-12 | 2020-04-21 | Ford Global Technologies, Llc | Parking spot identification for vehicle park-assist |
US11097723B2 (en) | 2018-10-17 | 2021-08-24 | Ford Global Technologies, Llc | User interfaces for vehicle remote park assist |
US11137754B2 (en) | 2018-10-24 | 2021-10-05 | Ford Global Technologies, Llc | Intermittent delay mitigation for remote vehicle operation |
US11312331B2 (en) * | 2018-11-13 | 2022-04-26 | Carrier Corporation | System and method for providing temporary access to a vehicle |
US10877479B2 (en) * | 2018-12-12 | 2020-12-29 | Waymo Llc | Multiple destination trips for autonomous vehicles |
CN109788030B (zh) * | 2018-12-17 | 2021-08-03 | 北京百度网讯科技有限公司 | 无人车数据处理方法、装置、系统及存储介质 |
US11789442B2 (en) | 2019-02-07 | 2023-10-17 | Ford Global Technologies, Llc | Anomalous input detection |
US11899448B2 (en) * | 2019-02-21 | 2024-02-13 | GM Global Technology Operations LLC | Autonomous vehicle that is configured to identify a travel characteristic based upon a gesture |
JP7101630B2 (ja) * | 2019-02-26 | 2022-07-15 | 本田技研工業株式会社 | 車両制御装置、車両制御方法、およびプログラム |
JP7172760B2 (ja) * | 2019-03-13 | 2022-11-16 | トヨタ自動車株式会社 | 情報処理装置および情報処理方法 |
US11195344B2 (en) | 2019-03-15 | 2021-12-07 | Ford Global Technologies, Llc | High phone BLE or CPU burden detection and notification |
JP7217655B2 (ja) * | 2019-03-26 | 2023-02-03 | 日立建機株式会社 | 作業車両 |
US11275368B2 (en) | 2019-04-01 | 2022-03-15 | Ford Global Technologies, Llc | Key fobs for vehicle remote park-assist |
US11169517B2 (en) | 2019-04-01 | 2021-11-09 | Ford Global Technologies, Llc | Initiation of vehicle remote park-assist with key fob |
JP7088118B2 (ja) * | 2019-04-09 | 2022-06-21 | トヨタ自動車株式会社 | 自動運転車両管理システム |
WO2020217739A1 (ja) * | 2019-04-20 | 2020-10-29 | 株式会社Zmp | 自動運転車両による利用者搬送システム |
KR20200130886A (ko) * | 2019-05-02 | 2020-11-23 | 현대자동차주식회사 | 자율주행 차량용 엔터테인먼트 방법 |
KR20200138541A (ko) * | 2019-05-31 | 2020-12-10 | 현대자동차주식회사 | 자율주행 차량용 엔터테인먼트 장치 |
US11281217B2 (en) * | 2019-06-25 | 2022-03-22 | Ford Global Technologies, Llc | Enhanced vehicle operation |
WO2020261543A1 (ja) * | 2019-06-28 | 2020-12-30 | ヤマハ発動機株式会社 | 自動運転専用道路管理装置 |
US11377104B2 (en) * | 2019-07-24 | 2022-07-05 | Robotic Research Opco, Llc | Autonomous waste collection truck |
US11587014B1 (en) | 2019-10-22 | 2023-02-21 | Wells Fargo Bank, N.A. | Ride sharing with cash delivery |
JP7433034B2 (ja) | 2019-12-12 | 2024-02-19 | 株式会社デンソーテン | 管理装置及び管理方法 |
JP7407631B2 (ja) * | 2020-03-19 | 2024-01-04 | 本田技研工業株式会社 | 車両制御装置 |
WO2021203385A1 (zh) * | 2020-04-09 | 2021-10-14 | 华为技术有限公司 | 一种生成指令单元组的方法及装置 |
JP2021175149A (ja) * | 2020-04-30 | 2021-11-01 | キヤノン株式会社 | 電子機器、電子機器の制御方法、プログラム |
KR102518175B1 (ko) * | 2020-07-22 | 2023-04-07 | 현대자동차주식회사 | 이동식 교육 서비스 제공방법 및 시스템 |
US11807278B2 (en) * | 2020-10-21 | 2023-11-07 | Gm Cruise Holdings Llc | Autonomous vehicle passenger safety monitoring |
TWI748897B (zh) * | 2021-03-11 | 2021-12-01 | 宜成投資股份有限公司 | 四輪以上免牌照自動駕駛電動車之中央控制調派車輛系統 |
KR102458673B1 (ko) * | 2021-05-21 | 2022-10-25 | 유니텔전자 (주) | 인공 지능 기반의 스마트 시트 제어 방법 및 그를 위한 장치 |
WO2023276399A1 (ja) | 2021-07-02 | 2023-01-05 | Whill株式会社 | 電動モビリティ |
US11761781B2 (en) | 2021-09-30 | 2023-09-19 | Gm Cruise Holdings Llc | User preview of rideshare service vehicle surroundings |
US11869102B2 (en) * | 2021-10-26 | 2024-01-09 | Honda Motor Co., Ltd. | Systems and methods for providing distance based notifications for electric vehicles |
CN114475631B (zh) * | 2022-02-22 | 2023-03-24 | 阿波罗智能技术(北京)有限公司 | 驾驶数据处理方法装置、设备、自动驾驶车辆介质及产品 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001288940A (ja) * | 2000-04-04 | 2001-10-19 | Honda Motor Co Ltd | 共同利用車両の予約方法 |
JP2002505219A (ja) * | 1997-08-01 | 2002-02-19 | カルカー インコーポレーテッド | 自動車用集中制御・管理システム |
JP2003259420A (ja) * | 2002-02-27 | 2003-09-12 | Yamaha Corp | 自動車位置通信システム、カーナビゲーション装置及び携帯通信装置 |
JP2006302187A (ja) * | 2005-04-25 | 2006-11-02 | Denso Corp | 自動駐車支援システム及び自動駐車支援装置 |
JP2009271631A (ja) * | 2008-05-01 | 2009-11-19 | Pioneer Electronic Corp | 車両制御装置、車両制御方法、車両制御プログラム、および記録媒体 |
JP2012048563A (ja) * | 2010-08-27 | 2012-03-08 | Toyota Motor Corp | 自動運転制御装置 |
WO2014024254A1 (ja) * | 2012-08-07 | 2014-02-13 | 株式会社日立製作所 | 自律走行装置の利用支援ツール、運用管理センタ、運用システム及び自律走行装置 |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000268290A (ja) * | 1999-03-12 | 2000-09-29 | Honda Motor Co Ltd | 自動運転領域における侵入検知制御装置 |
JP2000348298A (ja) * | 1999-06-03 | 2000-12-15 | Honda Motor Co Ltd | 共有電動車両の貸出ポート |
US6778068B2 (en) * | 2001-03-02 | 2004-08-17 | Qualcomm, Incorporated | Electronic locking device and method of operating same |
JP2003038580A (ja) * | 2001-07-27 | 2003-02-12 | Suzuki Motor Corp | 電動車椅子レンタルシステム |
JP2003214866A (ja) * | 2002-01-29 | 2003-07-30 | Clarion Co Ltd | ナビゲーション方法及び車載ナビゲーション装置並びにナビゲーション用プログラム |
JP2005228020A (ja) * | 2004-02-13 | 2005-08-25 | Zenrin Co Ltd | スケジュール管理装置 |
JP4059224B2 (ja) * | 2004-04-13 | 2008-03-12 | 株式会社デンソー | 運転者の外観認識システム |
JP2007108918A (ja) * | 2005-10-12 | 2007-04-26 | Matsushita Electric Ind Co Ltd | 情報管理システムおよび情報送信端末 |
US9459622B2 (en) * | 2007-01-12 | 2016-10-04 | Legalforce, Inc. | Driverless vehicle commerce network and community |
US20080027599A1 (en) * | 2006-07-28 | 2008-01-31 | James Logan | Autonomous vehicle and systems and methods for the operation thereof |
JP2009031196A (ja) * | 2007-07-30 | 2009-02-12 | Aisin Aw Co Ltd | 情報通知システム及びプログラム |
JP2009103623A (ja) * | 2007-10-24 | 2009-05-14 | Clarion Co Ltd | 規制情報提供装置、その制御方法及びその制御プログラム |
JP2009244032A (ja) * | 2008-03-31 | 2009-10-22 | Aisin Aw Co Ltd | 車載ナビゲーション装置 |
JP5361619B2 (ja) | 2009-09-04 | 2013-12-04 | 株式会社東芝 | 駐車場管理装置及び駐車場管理方法 |
JP5608592B2 (ja) * | 2011-03-18 | 2014-10-15 | 東芝テック株式会社 | 自律移動装置及び自律移動制御方法 |
FR2984254B1 (fr) * | 2011-12-16 | 2016-07-01 | Renault Sa | Controle de vehicules autonomes |
US9429943B2 (en) | 2012-03-05 | 2016-08-30 | Florida A&M University | Artificial intelligence valet systems and methods |
US20140309863A1 (en) * | 2013-04-15 | 2014-10-16 | Flextronics Ap, Llc | Parental control over vehicle features and child alert system |
US9082239B2 (en) * | 2012-03-14 | 2015-07-14 | Flextronics Ap, Llc | Intelligent vehicle for assisting vehicle occupants |
US20150166009A1 (en) * | 2013-11-11 | 2015-06-18 | Chris Outwater | System and Method for Wirelessly Rostering a Vehicle |
US20150149263A1 (en) * | 2013-11-22 | 2015-05-28 | Here Global B.V. | Automated Parking Payment |
US9304009B2 (en) * | 2014-02-04 | 2016-04-05 | Here Global B.V. | Method and apparatus for providing passenger embarkation points for points of interests |
US9205805B2 (en) * | 2014-02-14 | 2015-12-08 | International Business Machines Corporation | Limitations on the use of an autonomous vehicle |
US9146118B2 (en) * | 2014-02-27 | 2015-09-29 | Telenav Inc. | Navigation system with point of interest detour mechanism and method of operation thereof |
US9567007B2 (en) * | 2014-02-27 | 2017-02-14 | International Business Machines Corporation | Identifying cost-effective parking for an autonomous vehicle |
US10139824B2 (en) * | 2014-04-30 | 2018-11-27 | Mico Latta Inc. | Automatic driving vehicle and program for automatic driving vehicle |
-
2015
- 2015-04-16 US US15/305,254 patent/US10139824B2/en active Active
- 2015-04-16 WO PCT/JP2015/061697 patent/WO2015166811A1/ja active Application Filing
- 2015-04-16 JP JP2015540769A patent/JPWO2015166811A1/ja active Pending
-
2016
- 2016-12-02 JP JP2016234653A patent/JP6332817B2/ja active Active
- 2016-12-02 JP JP2016234654A patent/JP6320496B2/ja active Active
-
2018
- 2018-04-19 JP JP2018080294A patent/JP6590348B2/ja active Active
- 2018-06-20 JP JP2018116653A patent/JP2018160270A/ja active Pending
- 2018-10-11 US US16/157,797 patent/US10656648B2/en active Active
-
2020
- 2020-04-14 US US16/848,387 patent/US11397435B2/en active Active
- 2020-05-21 JP JP2020088514A patent/JP7034502B2/ja active Active
-
2022
- 2022-02-22 JP JP2022025187A patent/JP2022066246A/ja active Pending
- 2022-06-16 US US17/841,920 patent/US20220308584A1/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002505219A (ja) * | 1997-08-01 | 2002-02-19 | カルカー インコーポレーテッド | 自動車用集中制御・管理システム |
JP2001288940A (ja) * | 2000-04-04 | 2001-10-19 | Honda Motor Co Ltd | 共同利用車両の予約方法 |
JP2003259420A (ja) * | 2002-02-27 | 2003-09-12 | Yamaha Corp | 自動車位置通信システム、カーナビゲーション装置及び携帯通信装置 |
JP2006302187A (ja) * | 2005-04-25 | 2006-11-02 | Denso Corp | 自動駐車支援システム及び自動駐車支援装置 |
JP2009271631A (ja) * | 2008-05-01 | 2009-11-19 | Pioneer Electronic Corp | 車両制御装置、車両制御方法、車両制御プログラム、および記録媒体 |
JP2012048563A (ja) * | 2010-08-27 | 2012-03-08 | Toyota Motor Corp | 自動運転制御装置 |
WO2014024254A1 (ja) * | 2012-08-07 | 2014-02-13 | 株式会社日立製作所 | 自律走行装置の利用支援ツール、運用管理センタ、運用システム及び自律走行装置 |
Non-Patent Citations (2)
Title |
---|
"ISIT Dai 12 Kai Car Electronics Kenkyukai Koen Gaiyo", INSTITUTE OF SYSTEMS, INFORMATION TECHNOLOGIES AND NANOTECHNOLOGIES, 27 September 2012 (2012-09-27), XP055234076, [retrieved on 20150710] * |
HARUTOSHI OGAI, KITAKYUSHU-SHI NI OKERU JIDO UNTEN NO JISSHO JIKKEN -KOREISHA-YO KOGATA DENKI JIDOSHA NO JIRITSU SOKO JIKKEN, 27 September 2012 (2012-09-27), pages 1, 31, 53, XP055234065, Retrieved from the Internet <URL:http://web.archive.org/web/20130123173244/http://www.isit.or.jp/wg2/files/2012/08/20120927_ISIT_Ogai.pdf> [retrieved on 20150710] * |
Cited By (79)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11947353B1 (en) | 2015-05-26 | 2024-04-02 | Waymo Llc | Non-passenger requests for autonomous vehicles |
JP2022109937A (ja) * | 2015-05-26 | 2022-07-28 | ウェイモ エルエルシー | 自律車両のためのフォールバック要求 |
US10796581B2 (en) | 2015-05-26 | 2020-10-06 | Waymo Llc | Fallback requests for autonomous vehicles |
JP7109352B2 (ja) | 2015-05-26 | 2022-07-29 | ウェイモ エルエルシー | 自律車両のためのフォールバック要求 |
US11276314B2 (en) | 2015-05-26 | 2022-03-15 | Waymo Llc | Fallback requests for autonomous vehicles |
JP2019067439A (ja) * | 2015-05-26 | 2019-04-25 | ウェイモ エルエルシー | 自律車両のためのフォールバック要求 |
JP2017102007A (ja) * | 2015-12-01 | 2017-06-08 | 株式会社デンソー | 報知処理装置 |
WO2017094848A1 (ja) * | 2015-12-01 | 2017-06-08 | 株式会社デンソー | 報知処理装置 |
JP2017136977A (ja) * | 2016-02-04 | 2017-08-10 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP2017197066A (ja) * | 2016-04-28 | 2017-11-02 | 本田技研工業株式会社 | 車両制御システム、交通制御システム、車両制御方法、および車両制御プログラム |
US10336341B2 (en) | 2016-04-28 | 2019-07-02 | Honda Motor Co., Ltd. | Vehicle control system, traffic control system, vehicle control method, and vehicle control program |
JP2017218019A (ja) * | 2016-06-07 | 2017-12-14 | 株式会社デンソー | 車両用装置、及び車両用システム |
CN109478292B (zh) * | 2016-06-07 | 2023-04-28 | 株式会社电装 | 车辆用装置以及车辆用系统 |
WO2017213079A1 (ja) * | 2016-06-07 | 2017-12-14 | 株式会社デンソー | 車両用装置、及び車両用システム |
CN109478292A (zh) * | 2016-06-07 | 2019-03-15 | 株式会社电装 | 车辆用装置以及车辆用系统 |
JP2017228266A (ja) * | 2016-06-21 | 2017-12-28 | バイドゥ オンライン ネットワーク テクノロジー (ベイジン) カンパニー リミテッド | スマート車両に用いられる局部軌跡計画方法及び装置 |
JP2018024485A (ja) * | 2016-08-08 | 2018-02-15 | 三菱ロジスネクスト株式会社 | 荷役車両呼出装置、荷役車両システム、荷役車両呼出方法及び荷役車両呼出装置の制御プログラム |
US11449056B2 (en) | 2016-08-18 | 2022-09-20 | Motional Ad Llc | Hailing a vehicle |
US11892844B2 (en) | 2016-08-18 | 2024-02-06 | Motional Ad Llc | Hailing a vehicle |
US10884413B2 (en) | 2016-08-18 | 2021-01-05 | Motional Ad Llc | Hailing a vehicle |
US10392017B2 (en) | 2016-08-30 | 2019-08-27 | Toyota Jidosha Kabushiki Kaisha | Vehicle control system |
JP2018036754A (ja) * | 2016-08-30 | 2018-03-08 | トヨタ自動車株式会社 | 車両の制御装置 |
JP2018067172A (ja) * | 2016-10-20 | 2018-04-26 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
US10726360B2 (en) | 2016-11-04 | 2020-07-28 | Honda Motor Co., Ltd. | Vehicle control system, vehicle control method, and vehicle control program |
KR102003940B1 (ko) * | 2016-11-11 | 2019-10-01 | 엘지전자 주식회사 | 자율 주행 차량 및 그 제어방법 |
KR20180053081A (ko) * | 2016-11-11 | 2018-05-21 | 엘지전자 주식회사 | 자율 주행 차량 및 그 제어방법 |
US10775788B2 (en) | 2016-11-11 | 2020-09-15 | Lg Electronics Inc. | Autonomous vehicle and control method thereof |
US11422555B2 (en) | 2016-11-11 | 2022-08-23 | Lg Electronics Inc. | Autonomous vehicle and control method thereof |
WO2018102425A1 (en) * | 2016-12-02 | 2018-06-07 | Starsky Robotics, Inc. | Vehicle control system and method of use |
EP3551510A4 (en) * | 2016-12-09 | 2020-01-29 | Gentex Corporation | USER AUTHENTICATION ACTIVATION SYSTEMS AND METHODS |
WO2018116358A1 (ja) * | 2016-12-19 | 2018-06-28 | 株式会社Doog | 移動支援ロボットの運行管理システム |
US11048263B2 (en) | 2016-12-27 | 2021-06-29 | Toyota Jidosha Kabushiki Kaisha | Autonomous driving device for vehicle |
JP2018103858A (ja) * | 2016-12-27 | 2018-07-05 | トヨタ自動車株式会社 | 自動運転装置 |
WO2018123032A1 (ja) * | 2016-12-28 | 2018-07-05 | 株式会社Doog | 移動支援ロボットの運行管理システム |
CN106549981A (zh) * | 2017-01-13 | 2017-03-29 | 邹城众达知识产权咨询服务有限公司 | 一种基于大数据云计算服务的无方向盘智能网约车系统及其运营方法 |
US11441910B2 (en) | 2017-01-27 | 2022-09-13 | Panasonic Intellectual Property Corporation Of America | Vehicle control apparatus and vehicle control method |
EP3358306A1 (en) | 2017-01-27 | 2018-08-08 | Panasonic Intellectual Property Corporation of America | Vehicle control apparatus and vehicle control method |
US10845201B2 (en) | 2017-01-27 | 2020-11-24 | Panasonic Intellectual Property Corporation Of America | Vehicle control apparatus and vehicle control method |
GB2561054A (en) * | 2017-01-27 | 2018-10-03 | Ford Global Tech Llc | Self parking vehicle |
US10870425B2 (en) | 2017-01-30 | 2020-12-22 | Panasonic Intellectual Property Corporation Of America | Control device, control method, and recording medium having program recorded thereon for automatic driving vehicle |
JP2018122650A (ja) * | 2017-01-30 | 2018-08-09 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 自動運転車両の制御装置、制御方法及びプログラム |
JP7308340B2 (ja) | 2017-01-30 | 2023-07-13 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 自動運転車両の制御装置、制御方法及びプログラム |
CN108372858B (zh) * | 2017-01-30 | 2022-07-08 | 松下电器(美国)知识产权公司 | 自动驾驶车辆的控制装置、控制方法以及记录有程序的记录介质 |
JP2022171722A (ja) * | 2017-01-30 | 2022-11-11 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 自動運転車両の制御装置、制御方法及びプログラム |
JP2021178632A (ja) * | 2017-01-30 | 2021-11-18 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 自動運転車両の制御装置、制御方法及びプログラム |
CN108372858A (zh) * | 2017-01-30 | 2018-08-07 | 松下电器(美国)知识产权公司 | 自动驾驶车辆的控制装置、控制方法以及记录有程序的记录介质 |
JP7136974B2 (ja) | 2017-01-30 | 2022-09-13 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 自動運転車両の制御装置、制御方法及びプログラム |
US11073836B2 (en) | 2017-03-14 | 2021-07-27 | Gatik Ai Inc. | Vehicle sensor system and method of use |
US10209718B2 (en) | 2017-03-14 | 2019-02-19 | Starsky Robotics, Inc. | Vehicle sensor system and method of use |
US11681299B2 (en) | 2017-03-14 | 2023-06-20 | Gatik Ai Inc. | Vehicle sensor system and method of use |
JP2018156257A (ja) * | 2017-03-16 | 2018-10-04 | 三菱ロジスネクスト株式会社 | 配車システム |
JP2018156258A (ja) * | 2017-03-16 | 2018-10-04 | 三菱ロジスネクスト株式会社 | 配車システム |
JP6292646B1 (ja) * | 2017-03-16 | 2018-03-14 | 三菱ロジスネクスト株式会社 | 配車システム |
JP6274699B1 (ja) * | 2017-03-16 | 2018-02-07 | 三菱ロジスネクスト株式会社 | 配車システム |
JP6341525B1 (ja) * | 2017-06-22 | 2018-06-13 | 三菱ロジスネクスト株式会社 | 配車システムおよび配車方法 |
JP2019006202A (ja) * | 2017-06-22 | 2019-01-17 | 三菱ロジスネクスト株式会社 | 配車システムおよび配車方法 |
KR20190067430A (ko) * | 2017-12-07 | 2019-06-17 | 엘케이시스(주) | 자율 주행 차량을 이용한 자동 선적 방법 및 장치 |
KR102045126B1 (ko) * | 2017-12-07 | 2019-11-14 | 엘케이시스(주) | 자율 주행 차량을 이용한 자동 선적 방법 및 장치 |
JP2019164444A (ja) * | 2018-03-19 | 2019-09-26 | トヨタ自動車株式会社 | 駐車支援装置 |
JP2020038585A (ja) * | 2018-09-06 | 2020-03-12 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
JP2019073279A (ja) * | 2018-11-05 | 2019-05-16 | 本田技研工業株式会社 | 車両制御システム、車両制御方法、および車両制御プログラム |
CN111445722B (zh) * | 2019-01-16 | 2022-12-30 | 丰田自动车株式会社 | 自动驻车管理装置 |
CN111445722A (zh) * | 2019-01-16 | 2020-07-24 | 丰田自动车株式会社 | 自动驻车管理装置 |
JP7172647B2 (ja) | 2019-01-24 | 2022-11-16 | トヨタ自動車株式会社 | 配車装置および配車方法 |
JP2020119327A (ja) * | 2019-01-24 | 2020-08-06 | トヨタ自動車株式会社 | 配車装置および配車方法 |
WO2020209402A1 (ko) * | 2019-04-10 | 2020-10-15 | 엘케이시스(주) | 자율 주행 차량을 이용한 자동 선적 방법 및 장치 |
JP2019218057A (ja) * | 2019-09-05 | 2019-12-26 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP2020077385A (ja) * | 2019-09-25 | 2020-05-21 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
WO2021085446A1 (ja) * | 2019-10-29 | 2021-05-06 | Whill株式会社 | 施設内におけるシステムおよび電動モビリティ |
JP7248297B2 (ja) | 2019-11-28 | 2023-03-30 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
JP2020042850A (ja) * | 2019-11-28 | 2020-03-19 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
US11827503B2 (en) | 2020-03-18 | 2023-11-28 | Crown Equipment Corporation | Adaptive acceleration for materials handling vehicle |
US11919761B2 (en) | 2020-03-18 | 2024-03-05 | Crown Equipment Corporation | Based on detected start of picking operation, resetting stored data related to monitored drive parameter |
JP2022046490A (ja) * | 2020-08-13 | 2022-03-23 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP7257066B2 (ja) | 2020-08-13 | 2023-04-13 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP2021004029A (ja) * | 2020-08-13 | 2021-01-14 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP6994265B2 (ja) | 2020-08-13 | 2022-01-14 | みこらった株式会社 | 自動車及び自動車用プログラム |
JP2021093211A (ja) * | 2021-03-08 | 2021-06-17 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
JP7142379B2 (ja) | 2021-03-08 | 2022-09-27 | みこらった株式会社 | 自動運転車及び自動運転車用プログラム |
Also Published As
Publication number | Publication date |
---|---|
JP6320496B2 (ja) | 2018-05-09 |
JP2017091557A (ja) | 2017-05-25 |
US10656648B2 (en) | 2020-05-19 |
US20170123423A1 (en) | 2017-05-04 |
US20220308584A1 (en) | 2022-09-29 |
JP2017059254A (ja) | 2017-03-23 |
JP7034502B2 (ja) | 2022-03-14 |
US20200241532A1 (en) | 2020-07-30 |
US10139824B2 (en) | 2018-11-27 |
JP2022066246A (ja) | 2022-04-28 |
US20190041855A1 (en) | 2019-02-07 |
JP2018113074A (ja) | 2018-07-19 |
JP6590348B2 (ja) | 2019-10-16 |
JPWO2015166811A1 (ja) | 2017-04-20 |
JP2018160270A (ja) | 2018-10-11 |
US11397435B2 (en) | 2022-07-26 |
JP2020149708A (ja) | 2020-09-17 |
JP6332817B2 (ja) | 2018-05-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6590348B2 (ja) | 自動運転車及び自動運転車用プログラム | |
JP6991609B2 (ja) | 移動体呼び寄せシステム及び呼び寄せ装置並びに呼び寄せ装置用プログラム | |
US11180116B2 (en) | Vehicle dispatch management system and vehicle dispatch management server | |
JP6590281B2 (ja) | 自動車及び自動車用プログラム | |
JP6598254B2 (ja) | 自動運転車及び自動運転車用プログラム | |
JP6854018B2 (ja) | 自動運転車及び自動運転車用プログラム | |
JP6762499B2 (ja) | 自動車及び自動車用プログラム | |
JP7142379B2 (ja) | 自動運転車及び自動運転車用プログラム | |
JP6994265B2 (ja) | 自動車及び自動車用プログラム | |
JP2022046490A (ja) | 自動車及び自動車用プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2015540769 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15786158 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15305254 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15786158 Country of ref document: EP Kind code of ref document: A1 |