US10935385B2 - Information processing apparatus, vehicle, information processing method, running control method, and map updating method - Google Patents
Information processing apparatus, vehicle, information processing method, running control method, and map updating method Download PDFInfo
- Publication number
- US10935385B2 US10935385B2 US16/045,858 US201816045858A US10935385B2 US 10935385 B2 US10935385 B2 US 10935385B2 US 201816045858 A US201816045858 A US 201816045858A US 10935385 B2 US10935385 B2 US 10935385B2
- Authority
- US
- United States
- Prior art keywords
- vehicle
- map
- mode
- information processing
- road
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 126
- 238000000034 method Methods 0.000 title claims description 40
- 238000003672 processing method Methods 0.000 title claims description 9
- 230000008859 change Effects 0.000 claims abstract description 50
- 230000009471 action Effects 0.000 claims abstract description 30
- 238000001514 detection method Methods 0.000 claims description 240
- 238000004891 communication Methods 0.000 claims description 69
- 230000001133 acceleration Effects 0.000 claims description 22
- 238000004590 computer program Methods 0.000 claims description 8
- 230000003247 decreasing effect Effects 0.000 claims description 2
- 238000003860 storage Methods 0.000 description 26
- 238000010586 diagram Methods 0.000 description 21
- 238000003384 imaging method Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 10
- 230000000694 effects Effects 0.000 description 9
- 238000004422 calculation algorithm Methods 0.000 description 7
- 239000000470 constituent Substances 0.000 description 7
- 230000006866 deterioration Effects 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 230000008901 benefit Effects 0.000 description 6
- 230000006735 deficit Effects 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 5
- 238000010295 mobile communication Methods 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 3
- 230000002401 inhibitory effect Effects 0.000 description 3
- 230000007423 decrease Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000014759 maintenance of location Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000000717 retained effect Effects 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 230000009931 harmful effect Effects 0.000 description 1
- 230000000670 limiting effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000036961 partial effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/28—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network with correlation of data from several navigational instruments
- G01C21/30—Map- or contour-matching
- G01C21/32—Structuring or formatting of map data
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/38—Electronic maps specially adapted for navigation; Updating thereof
- G01C21/3804—Creation or updating of map data
- G01C21/3807—Creation or updating of map data characterised by the type of data
- G01C21/3815—Road data
- G01C21/3819—Road shape data, e.g. outline of a route
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/38—Electronic maps specially adapted for navigation; Updating thereof
- G01C21/3804—Creation or updating of map data
- G01C21/3833—Creation or updating of map data characterised by the source of data
- G01C21/3848—Data obtained from both position sensors and additional sensors
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/38—Electronic maps specially adapted for navigation; Updating thereof
- G01C21/3885—Transmission of map data to client devices; Reception of map data by client devices
- G01C21/3896—Transmission of map data from central databases
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/42—Simultaneous measurement of distance and other co-ordinates
-
- G05D2201/0213—
Definitions
- the present disclosure relates to an information processing apparatus, a vehicle, an information processing method, a running control method, and a map updating method for updating map data.
- One non-limiting and exemplary embodiment provides an information processing apparatus, a vehicle, and the like that make it possible to effectively update map data so that it becomes more accurate.
- the techniques disclosed here feature an information processing apparatus for distributing, to a vehicle, a map showing a road on which the vehicle runs and a road shape that is a shape of an area around the road, including: a processor; and a non-transitory recording medium storing thereon (i) the map divided into a plurality of areas and (ii) a computer program, which when executed by the processor, causes the processor to perform operations including: determining, for each of the plurality of areas, whether it is necessary to update the map of that area, the determining being based on at least either an amount of change from the road shape shown by the map to the road shape of a present time or the number of occurrences of a predetermined action of a vehicle having run through that area.
- An information processing apparatus, a vehicle, and the like according to the present disclosure make it possible to effectively update map data so that it becomes more accurate.
- FIG. 1 is a schematic view of a map updating system according to an embodiment
- FIG. 2 is a block diagram showing an example of a hardware configuration of an information processing apparatus according to the embodiment
- FIG. 3 is a block diagram showing an example of a hardware configuration of a vehicle according to the embodiment.
- FIG. 4 is a block diagram showing an example of a functional configuration of the map updating system according to the embodiment.
- FIG. 5 is a graph for explaining the computation of a score by a first determiner
- FIG. 6 is a diagram showing a part of a three-dimensional map divided into a plurality of areas
- FIG. 7A is a diagram showing an example of an event history database
- FIG. 7B is a diagram showing an example of a vehicle database
- FIG. 8A is a diagram showing an example of a UI that notifies the reception of a detection request
- FIG. 8B is a diagram showing an example of a UI that asks a user whether to switch a running mode to a detection mode
- FIG. 9 is a sequence chart showing an example of a map updating method by the map updating system according to the embodiment.
- FIG. 10 is a flow chart showing a detailed example of an update determination process according to the embodiment.
- FIG. 11 is a flow chart showing a detailed example of running control according to the embodiment.
- FIG. 12 is a diagram for explaining an effect that is brought about by running at a larger intervehicular distance in the detection mode
- FIG. 13A is a diagram for explaining an effect that is brought about by running at a lower maximum speed or maximum acceleration in the detection mode
- FIG. 13B is a diagram for explaining an effect that is brought about by running at a lower maximum speed or maximum acceleration in the detection mode
- FIG. 14A is a diagram for explaining an effect of inhibiting the occurrence of occlusion by a plurality of objects.
- FIG. 14B is a diagram for explaining an effect of inhibiting the occurrence of occlusion by a plurality of objects.
- Japanese Unexamined Patent Application Publication No. 2016-156973 discloses a map system including a server that manages map data and a vehicle mounted with a sensor.
- the vehicle detects a point of change in environment from a comparison between data acquired from the sensor and the map data
- the server acquires data on the point of change and sensor accuracy information at the time of detection of the point of change from the vehicle, whereby the map data is updated by assigning a weight to the point-of-change data according to the sensor accuracy information.
- Japanese Unexamined Patent Application Publication No. 2016-153832 discloses a system that updates map data by assigning weights to points of change on the basis of the amount of traffic in an area through which the vehicle has run.
- the technologies disclosed in these documents are undesirably excessive in data communication in an area with a large amount of traffic and insufficient in update data in an area with a small amount of traffic, as the amount of data that is collected by the server depends on the amount of traffic in each map area.
- the volume of detection data that is obtained by a LiDAR or a camera is large, unnecessary communication is often performed in an area with a large amount of traffic.
- detection data that is obtained by a LiDAR or a camera tends to be insufficient; therefore, even if there is a point of change on an actual road, the frequency of updates cannot be maintained and, accordingly, information freshness cannot be maintained.
- Japanese Unexamined Patent Application Publication No. 2016-153832 discloses a method for maintaining the frequency of updates by increasing the weight of detection data obtained in an area with a small amount of traffic. However, this method undesirably sacrifices update reliability in a case where the accuracy of data detected is low.
- the technologies disclosed in the foregoing documents entail a risk of collecting low-accuracy detection data in collecting detection data from the vehicle's sensor. For example, since the vehicle detects its surroundings with the sensor while running, the accuracy with which the vehicle estimates its own position deteriorates due to vibration or sudden acceleration of the vehicle during sensor detection so that there is a possibility of deterioration in matching accuracy between detection data and map data. Further, in a case where the vehicle is approaching a large-sized vehicle while detecting data, occlusion occurs in LiDAR detection data to make it impossible to appropriately detect a surrounding situation.
- Japanese Unexamined Patent Application Publication No. 2016-156973 discloses a method for, according to an element responsible for sensor accuracy at the time of data detection, changing the weight of data detected. However, this method undesirably slow in updating, as only low weights are assigned in a situation where sensor accuracy is not stable.
- the techniques disclosed here feature an information processing apparatus for distributing, to a vehicle, a map showing a road on which the vehicle runs and a road shape that is a shape of an area around the road, including: a processor; and a non-transitory recording medium storing thereon (i) the map divided into a plurality of areas and (ii) a computer program, which when executed by the processor, causes the processor to perform operations including: determining, for each of the plurality of areas, whether it is necessary to update the map of that area, the determining being based on at least either an amount of change from the road shape shown by the map to the road shape of a present time or the number of occurrences of a predetermined action of a vehicle having run through that area.
- the map is divided into the plurality of areas, and it is determined, for each of the plurality of areas, whether it is necessary to update the map.
- the information processing apparatus can efficiently determine an area where the map needs to be updated.
- the operations may further include: selecting, from among one or more vehicles that are running or planning to run in an area where it has been determined that the map needs to be updated, a detecting vehicle that detects the road shape; and transmitting, via a communication network to the detecting vehicle thus selected, a detection request that causes the detecting vehicle to detect the road shape of the area.
- one or more vehicles are requested to detect detection data for an area where the map needs to be updated.
- the information processing apparatus since the information processing apparatus selectively transmits a detection request to a vehicle that is running or planning to run in an area where updating is needed, the information processing apparatus can effectively acquire accurate detection data. For this reason, the information processing apparatus can effectively update map data so that it becomes more accurate.
- the vehicle may be a vehicle that performs automatic driving
- the predetermined action may be at least one of sudden deceleration, sudden acceleration, manual driving interruption during the automatic driving, and quick avoidance that have occurred during a predetermined period of time up to the present time.
- the information processing apparatus determines that a larger number of occurrences of predetermined actions in a vehicle that performs automatic driving lead to a greater difference between a road shape shown by a map and an actual road shape. This is because a difference between a road shape shown by a map and an actual road shape leads to deterioration in accuracy of estimation of the position of the vehicle, thus making stable running control difficult.
- the information processing apparatus determines that a larger number of occurrences of predetermined actions in a vehicle that performs automatic driving lead to a greater difference between a road shape shown by a map and an actual road shape, the information processing apparatus can effectively determine whether it is necessary to update the map.
- the selecting may include selecting, as the detecting vehicle, a vehicle whose vehicle height is not lower than a predetermined height.
- the selecting may include selecting, as the detecting vehicle, a vehicle having a sensor that detects the road shape and whose detection accuracy is not lower than a predetermined degree of accuracy.
- the selecting may include selecting, as the detecting vehicle, a vehicle that has ever run in a detection mode in which to detect the road shape in order to update the map.
- the determining may include: computing a score for each of the plurality of areas on the basis of at least either the amount of change or the number of occurrences; and determining that it is necessary to update the map of that one of the plurality of areas whose computed score exceeds a predetermined score.
- the determining may include increasing the predetermined score of that one of the plurality of areas where the number of times it has been determined per unit period that updating is needed is not smaller than a first number of times and decreasing the predetermined score of that one of the plurality of areas where the number of times it has been determined per unit period that updating is needed is smaller than a second number of times that is not larger than the first number of times.
- an area with a greater time series variation in the road shape can be updated at a higher frequency.
- the techniques disclosed here feature a vehicle including: a processor; and a memory storing thereon a computer program, which when executed by the processor, causes the processor to perform operations including: detecting a road on which the vehicle runs and a road shape that is a shape of an area around the road; receiving, via a communication network from an external information processing apparatus, a detection request to detect the road shape; upon receiving the detection request, determining whether a running mode of the vehicle is switched from a normal mode in which to perform normal running to a detection mode in which to, in order to update the map, perform running that detects the road shape of a road on which the vehicle is running; in a case where it has been determined that switching to the detection mode is made, switching the running mode to the detection mode and controlling the vehicle to run in the detection mode; and in a case where it has been determined that switching to the detection mode is not made, keeping the running mode in the detection mode and controlling the vehicle to run in the normal mode.
- the vehicle since the vehicle runs in a detection mode, different from a normal mode, in which to detect a road shape for use in updating of a map, the vehicle can detect the road shape with a high degree of accuracy. Further, even upon receiving a detection request from an information processing apparatus, the vehicle keeps running in the normal mode without switching to the detection mode, depending on the situation of the vehicle; therefore, the vehicle can inhibit impairment of convenience for a user of the vehicle.
- the operation may further include: transmitting, via the communication network to the information processing apparatus, detection data indicating the road shape thus detected.
- the vehicle can transmit, to the information processing apparatus, detection data detected with a high degree of accuracy in an area where the map needs to be updated.
- the detection mode may be a running mode in which to run in a state of change in at least one of settings of making a maximum running speed lower, making a maximum running acceleration lower, making an intervehicular distance larger, or making a detection frequency higher than in the normal mode.
- the vehicle can detect the road shape with a high degree of accuracy in the detection mode.
- the operations may further include: detecting a following vehicle falling within a certain distance from the vehicle; and in a case where the running mode has been switched to the detection mode and the following vehicle has been detected, controlling the vehicle to run in the normal mode.
- the vehicle can run out of the way of a following vehicle in a case where the following vehicle has been detected.
- the operations may further include: detecting an occurrence of congestion in an area around the vehicle; and in a case where the running mode has been switched to the detection mode and the occurrence of congestion has been detected, controlling the vehicle to run in the normal mode.
- the vehicle can run out of the way of a stuck vehicle when the density of other vehicles is not lower than a predetermined density and congestion is taking place.
- the operations may further include receiving an input indicating whether to give approval to switch the running mode from the normal mode to the detection mode, and the determining may include, in a case where the input indicates approval, determining that the running mode is switched and, in a case where the input indicates disapproval, determining that the running mode is not switched.
- the vehicle since, in the absence of approval from a person who is in the vehicle interior, the vehicle runs in the normal mode without switching the running mode, the vehicle can inhibit impairment of convenience for the person.
- the vehicle may be a vehicle that carries a passenger
- the determining may include, in a case where the vehicle is not on its way to pick up the passenger and the vehicle is not carrying the passenger, determining that the running mode is switched and, in a case where the vehicle is on its way to pick up the passenger or the vehicle is carrying the passenger, determining that the running mode is not switched.
- the vehicle since, in a case where the vehicle is on its way to pick up the passenger or is carrying the passenger, the vehicle runs while keeping the running mode in the detection mode, the vehicle can inhibit impairment of convenience for the passenger.
- the techniques disclosed here feature a map updating method including updating a map in a map updating system including an information processing apparatus for distributing, to one or more vehicles, the map showing a road on which the one or more vehicles run and a road shape that is a shape of an area around the road and the one or more vehicles, wherein the information processing apparatus performs operations including: determining, for each of the plurality of areas, whether it is necessary to update the map of that area, the determining being based on at least either an amount of change from the road shape shown by the map to the road shape of a present time or the number of occurrences of a predetermined action of a vehicle having run through that area, the map being stored in a memory of the information processing apparatus; selecting, from among one or more vehicles that are running or planning to run in an area where it has been determined that the map needs to be updated, a detection vehicle that detects the road shape with a high degree of accuracy; and transmitting, via a communication network to the detecting vehicle thus selected, a detection request that causes the
- FIG. 1 is a schematic view of a map updating system 1 according to the embodiment.
- FIG. 1 shows an information processing apparatus 100 , a plurality of vehicles 200 , a communication network 300 , and a base station 310 of a mobile communication system.
- the map updating system 1 includes the information processing apparatus 100 and the plurality of vehicles 200 . It should be noted that although FIG. 1 shows two vehicles 200 , the number of vehicles 200 is not limited, provided it is one or more.
- the information processing apparatus 100 is an apparatus that distributes, to the plurality of vehicles 200 , a three-dimensional map showing a road on which the plurality of vehicles 200 run and a three-dimensional road shape of an area around the road.
- the information processing apparatus 100 is for example a server.
- Each of the plurality of vehicles 200 includes a three-dimensional road shape shown by the three-dimensional map and a detector (sensor) of that vehicle 200 .
- Each of the plurality of vehicles 200 is for example a vehicle that estimates its own position by performing matching between a road, detected by the detector, on which that vehicle 200 is running and a three-dimensional road shape, detected by the detector, of an area around the road and performs automatic driving or drive assist with use of its own position thus estimated.
- Each of the plurality of vehicles 200 performs the aforementioned matching, for example, by using an ICP (iterative closest point) algorithm or NDT (normal distributions transform) algorithm.
- the road shape shown by the three-dimensional map no longer agrees with the actual road shape.
- the shapes of a road and an area around the road change from day to day.
- the shapes of trees greatly vary according to seasons.
- the shape of an area around the road changes from day to day. This causes deterioration in matching accuracy.
- the map updating system 1 causes the plurality of vehicles 200 , which are general-purpose vehicles mounted with detectors (sensors), to detect detection data for use in map updating and causes the information processing apparatus 100 to aggregate the detection data.
- the three-dimensional map retained by the information processing apparatus 100 is updated with use of detection data obtained by the sensors of the plurality of vehicles 200 .
- the information processing apparatus 100 requires plural pieces of detection data even on one change in environment.
- uploading of detection data by all of the vehicles that pass through that point is greatly wasteful both in terms of communications traffic and in terms of computational resources of the information processing apparatus 100 .
- the volume of detection data that is obtained by a LIDAR or a camera is large, communications traffic is enormous.
- Japanese Unexamined Patent Application Publication No. 2016-153832 achieves updating by correcting reliability per piece of data on a road with a small amount of traffic so that it becomes higher. That is, in an area with a small amount of traffic, updating is finalized with a small quantity of detection data, and that way, reliability is sacrificed. To address this problem, it is necessary to detect a larger quantity of detection data in an area with a small amount of traffic and to improve the accuracy of every single piece of detection data.
- the information processing apparatus 100 divides the three-dimensional map into a plurality of areas, determines, for each of the plurality of areas, whether it is necessary to update the three-dimensional map, and requests one or more vehicles 200 to detect detection data only in an area where the three-dimensional map needs to be updated. Further, the information processing apparatus 100 selects a vehicle 200 that is capable of detecting a road shape with a high degree of accuracy as a vehicle 200 that the information processing apparatus 100 requests to detect detection data from among one or more vehicles 200 that are running or planning to run in the area. In this way, the information processing apparatus 100 selects, for an area where updating is needed, a vehicle 200 that is capable of detecting a road shape with a high degree of accuracy and transmits a detection request to the vehicle 200 . This makes it possible to keep the information freshness of the three-dimensional map as high as possible and reduce communications traffic that is generated for aggregation of detection data and the processing load of the detection data.
- a hardware configuration of the information processing apparatus 100 is described with reference to FIG. 2 .
- FIG. 2 is a block diagram showing an example of the hardware configuration of the information processing apparatus 100 according to the embodiment.
- the information processing apparatus 100 includes a CPU (central processing unit) 11 , a main memory 12 , a storage 13 , and a communication IF (interface) 14 as hardware components.
- a CPU central processing unit
- main memory 12 main memory
- storage 13 main memory
- communication IF interface
- the CPU 11 is a processor that executes a control program stored in the storage 13 or the like.
- the main memory 12 is a volatile storage area that is used as a work area that is used when the CPU 11 executes the control program.
- the storage 13 is a nonvolatile storage area that retains the control program, contents, and the like.
- the communication IF 14 is a communication interface that communicates with the plurality of vehicles 200 via a communication network.
- the communication IF 14 is for example a cable LAN interface.
- the communication IF 14 may be a wireless LAN interface.
- the communication IF 14 may be any communication interface that can establish a communication connection with the communication network.
- FIG. 3 is a block diagram showing an example of the hardware configuration of the vehicle 200 according to the embodiment.
- the vehicle 200 includes a CPU (central processing unit) 21 , a main memory 22 , a storage 23 , a communication IF (interface) 24 , an imaging device 25 , a LIDAR 26 , an IMU 27 , a GNSS 28 , a display 29 , and an input IF (interface) 30 as hardware components.
- FIG. 3 shows a configuration of each of the plurality of vehicles 200 . It should be noted that it is not necessary for each of the plurality of vehicles 200 to include all of the hardware components shown in FIG. 3 .
- the vehicle 200 needs only include at least either the imaging device 25 or the LIDAR 26 .
- the vehicle 200 does not need to include the display 29 .
- the vehicle 200 does not need to include the input IF 30 .
- the vehicle 200 may include a dedicated circuit instead of the CPU (central processing unit) 21 and the main memory 22 .
- the CPU 21 is a processor that executes a control program stored in the storage 23 or the like.
- the main memory 22 is a volatile storage area that is used as a work area that is used when the CPU 21 executes the control program.
- the storage 23 is a nonvolatile storage area that retains the control program, contents, and the like.
- the communication IF 24 is a communication interface that communicates with the information processing apparatus 10 via the communication network 300 . That is, the communication IF 24 needs only be a communication interface that is capable of making a communication connection to the communication network 300 . Specifically, the communication IF 24 is a communication interface that makes a communication connection to the communication network 300 through a communication connection to the base station 310 of the mobile communication system.
- the communication IF 24 may for example be a radio communication interface that complies with a communication standard that is used in a mobile communication system such as a third-generation (3G) mobile communication system, a fourth-generation (4G) mobile communication system, or LTE (registered trademark).
- Another example of the communication IF 24 may be a wireless LAN (local area network) interface that complies with the IEEE802.11a, b, g, or n standard or a communication interface that makes a communication connection to the communication network 300 through a communication connection to a router (e.g. a mobile wireless LAN router; not illustrated).
- a wireless LAN local area network
- the imaging device 25 is a camera including an optical system such as a lens and an image sensor.
- the LIDAR 26 is a laser sensor that detects a distance to an object located within a range of detection in an angular range of 360 degrees all around in a horizontal direction of the vehicle 200 and a predetermined angle (e.g. 30 degrees) in a vertical direction of the vehicle 200 .
- the LIDAR 26 measures the distance from the LIDAR 26 to an object by emitting a laser toward its surroundings and sensing a laser reflected by a surrounding object. For example, the LIDAR 26 measures the distance in the order of centimeters. In this way, the LIDAR 26 detects the three-dimensional coordinates of each of a plurality of points on a terrain surface of an area around the vehicle 200 . That is, the LIDAR 26 can detect the three-dimensional shape of a terrain including an object around the vehicle 200 by detecting a plurality of three-dimensional coordinates of the surrounding terrain surface.
- the IMU 27 is a sensor device that includes an acceleration sensor and a gyro sensor.
- the acceleration sensor is a sensor that detects accelerations applied in three different directions of the vehicle 200 , respectively.
- the gyro sensor is a sensor that detects the angular velocity of each of three axial rotations about the three different directions of the vehicle 200 , respectively.
- the GNSS 28 receives information indicating the position of the GNSS 28 from an artificial satellite including a GPS satellite. That is, the GNSS 28 detects the present position of the vehicle 200 .
- the display 29 is a display device that displays a result of processing in the CPU 21 .
- the display 29 is for example a liquid crystal display or an organic EL display.
- the display 29 is disposed in a vehicle interior of the vehicle 200 .
- the input IF 30 is for example a touch panel, disposed on a surface of the display 29 , that receives an input from a user to a UI (user interface) that is displayed on the display 29 .
- the input IF 30 may for example by an input device such as a numeric keypad, a keyboard, or a mouse.
- the input IF 30 is disposed in the vehicle interior of the vehicle 200 .
- the communication network 300 is a communication network through which the information processing apparatus 100 and the plurality of vehicles 200 communicate with each other.
- the communication network 300 may be a general-purpose network such as the Internet or may be a network dedicated to the map updating system 1 .
- FIG. 4 is a block diagram showing an example of the functional configuration of the map updating system 1 according to the embodiment.
- the information processing apparatus 100 includes a first determiner 110 , a selector 120 , a first communicator 130 , and a database 150 as functional components.
- the information processing apparatus 100 may further include a position acquirer 140 .
- the first determiner 110 determines, on the basis of at least either an amount of change from the road shape shown by the three-dimensional map to the present road shape of the road or the number of occurrences of a predetermined action of a vehicle 200 having run through that area, whether it is necessary to update the three-dimensional map of that area. Specifically, the first determiner 110 computes a score for each of the plurality of areas on the basis of at least either the amount of change or the number of occurrences. Moreover, the first determiner 110 may determine that it is necessary to update the three-dimensional map of that one of the plurality of areas where the score thus computed exceeds a predetermined score.
- the first determiner 110 may use Formula 1 below to compute a score for determining whether it is necessary to update the map.
- FIG. 5 is a graph for explaining the computation of a score by the first determiner 110 .
- S denotes the score
- c i denotes the amount of change in past update score
- P denotes a predetermined period of time preceding the last update time
- T denotes the time elapsed since the last update time up to the current time
- p j denotes the number of occurrences of various events
- ⁇ and ⁇ j denote coefficients.
- the number of occurrences of various events is the number of occurrences of predetermined actions of a vehicle 200 .
- the predetermined actions are sudden deceleration, sudden acceleration, manual driving interruption during automatic driving, and quick avoidance that have occurred in the vehicle 200 during the predetermined period of time T up to the present time.
- the predetermined actions do not need to be all of sudden deceleration, sudden acceleration, manual driving interruption, and quick avoidance, but needs only be at least one of them.
- the score that the first determiner 110 computes is a weighted sum of a score indicated by the first term of Formula 1 and based on an amount of change from the road shape shown by the three-dimensional map to the present road shape of the road and a score indicated by the second term of Formula 1 and based on the number of occurrences of a predetermined action of a vehicle 20 having run through an area. That is, an area where updating frequently takes place in the past update history, an area where the time elapsed since the last update time is long, or an area where an event such as sudden deceleration or quick avoidance frequency takes place tends to be high in score.
- the score may be either the score indicated by the first term of Formula 1 or the score indicated by the second term of Formula 1.
- FIG. 6 is a diagram showing a part of a three-dimensional map divided into a plurality of areas.
- the first determiner 110 computes a score by performing a calculation according to Formula 1 for each of the plurality of areas A 1 to A 4 , B 1 to B 4 , and C 1 to C 4 . Since the plurality of areas that constitute the three-dimensional map shown in FIG. 6 are partial, the first determiner 110 also performs a calculation according to Formula 1 for each of a plurality of areas that are not shown in FIG. 6 .
- the first determiner 110 may increase a predetermined score of that one of the plurality of areas where the number of times it has been determined per unit period that updating is needed is not smaller than a first number of times and decrease a predetermined score of that one of the plurality of areas where the number of times it has been determined per unit period that updating is needed is smaller than a second number of times that is not larger than the first number of times.
- the first determiner 110 is realized, for example, by the CPU 11 , the main memory 12 , the storage 13 , and the like.
- the selector 120 selects, from among one or more vehicles 200 that are running or planning to run in an area where it has been determined by the first determiner 110 that the three-dimensional map needs to be updated, a detection vehicle that detects the road shape. With use of the present positions or running routes of a plurality of vehicles 200 as acquired by the after-mentioned position acquirer 140 , the selector 120 may identify the one or more vehicles 200 that are running or planning to run in the area.
- the selector 120 may select, from among the one or more vehicles 200 thus identified that are running or planning to run in the area, a detecting vehicle that detects the road shape. For example, with use of vehicle height information or model information included in the vehicle information, the selector 120 may select, as the detecting vehicle, a vehicle whose vehicle height is not lower than a predetermined height. Further, for example, with use of sensor information or model information included in the vehicle information, the selector 120 may select, as the detecting vehicle, a vehicle having a sensor that detects the road shape and whose detection accuracy is not lower than a predetermined degree of accuracy. Further, for example, with use of history information included in the vehicle information, the selector 120 may select, as the detecting vehicle, a vehicle that has ever run in a detection mode in which to detect the road shape in order to update the three-dimensional map.
- the selector 120 is realized, for example, by the CPU 11 , the main memory 12 , the storage 13 , and the like.
- the first communicator 130 transmits, via the communication network 300 to the detecting vehicle selected by the selector 120 , a detection request that causes the detecting vehicle to detect the road shape of the area.
- the first communicator 130 is an example of a first communicator.
- the first communicator 130 may receive, via the communication network 300 from a vehicle 200 to which it has transmitted the detection request as the detecting vehicle, detection data detected by a sensor of the vehicle 200 . Further, the first communicator 130 may receive position information indicating the present position of each of the plurality of vehicles 200 or route information indicating the running route of each of the plurality of vehicles 200 . Further, the first communicator 130 may receive a history of occurrences of a predetermined action of each of the plurality of vehicles 200 .
- a history of occurrences of a predetermined action received by the first communicator 130 is stored in an event history database 153 for each of the plurality of vehicle 200 .
- the first communicator 130 is realized, for example, by the CPU 11 , the main memory 12 , the storage 13 , the communication IF 14 , and the like.
- the position acquirer 140 acquires the present positions or running routes of the plurality of vehicles 200 via the first communicator 130 .
- the position acquirer 140 acquires the present positions or running routes of the plurality of vehicles 200 every few seconds to several tens of seconds.
- the position determiner 140 is realized, for example, by the CPU 11 , the main memory 12 , the storage 13 , and the like.
- the database 150 includes a map database 151 , a map history database 152 , the event history database 153 , and the vehicle database 154 .
- the map database 151 stores a three-dimensional map divided into a plurality of areas.
- the plurality of areas are a plurality of areas into which an area of a two-dimensional map has been divided.
- the three-dimensional map is a point cloud map, and is data in which a terrain (shape of land) including a road is constituted by a point cloud on a three-dimensional space.
- the point cloud is constituted by a plurality of points (e.g. millions of points or more) each represented by three-dimensional coordinates (x, y, z).
- the three-dimensional map of the map database 151 is updated to be a three-dimensional map based on the detection data thus received. For this reason, the map database 151 stores the most recent three-dimensional map.
- the three-dimensional map stored in the map database 151 that is yet to be updated is based on data on a road shape detected in advance with use of an apparatus including a predetermined sensor.
- the map history database 152 stores a three-dimensional map that was stored in the map database 151 in the past. That is, the map history database 152 stores a three-dimensional map that was stored in the map database 151 and that has been brought out of date by being updated.
- the map history database 152 may be included in the map database 151 .
- the event history database 153 stores, for each of the plurality of vehicles 200 , a history of occurrences of a predetermined action of that vehicle 200 as acquired via the first communicator 130 .
- the event history database 153 holds data of association between vehicle IDs that identify the plurality of vehicles 200 , predetermined actions having occurred, times of occurrence of the predetermined actions, and areas of occurrence of the predetermined actions.
- the vehicle database 154 stores vehicle information.
- the vehicle information is information regarding the plurality of vehicles 200 .
- the vehicle information may also include vehicle height information indicating the vehicle height of each of the plurality of vehicles 200 .
- the vehicle information may also include sensor information indicating the detection accuracy of a sensor of each of the plurality of vehicles 200 .
- the vehicle information may also include model information indicating models.
- the model information may for example include the vehicle height information.
- the model information may also be used to identify the detection accuracy of a sensor. That is, since a model can be identified by using the model information, the type of a sensor of a vehicle of the model can be identified. Accordingly, the detection accuracy of the sensor can be identified by using the model information.
- the vehicle information may also include history information indicating, for each of the plurality of vehicles 200 , whether that vehicle 200 ran in the detection mode in which to detect the road shape.
- the vehicle database 154 holds data of association between vehicle IDs that identify the plurality of vehicles 200 , the vehicle heights of the vehicles 200 , the sensor accuracy of sensors of the vehicles 200 , the numbers of times the vehicles 200 ran in the detection mode, which serves as history information.
- FIG. 7B has, as vehicle heights, information categorized into three levels “high”, “medium”, and “low” according to vehicle heights, it may alternatively have the numerical values of vehicle heights per se.
- the sensor accuracy in FIG. 7B may be defined by being categorized into “high”, “medium”, “low”, “unknown”, and the like by performing a predetermined detection test in advance for each of the plural types of sensor.
- the type of a sensor and a vehicle may be associated according to the model, or the type of a sensor of a vehicle may be identified by acquiring information indicating the type of the sensor from the vehicle. It should be noted that “unknown” represents a sensor that has not been subjected to the predetermined test.
- the database 150 is realized, for example, by the storage 13 and the like. It should be noted that the map database 151 , map history database 152 , event history database 153 , and vehicle database 154 of the database 150 may be realized by one storage, may be realized by a plurality of storages, or may be realized by a plurality of respectively corresponding storages. Further, the information processing apparatus 100 does not need to include the database 150 , and a second apparatus that is separate from the information processing apparatus 100 may include the database 150 . In this case, the information processing apparatus 100 and the second apparatus are communicably connected to each other.
- the vehicle 200 includes a second communicator 210 , a second determiner 202 , a switcher 230 , a controller 240 , a first detector 250 , and a map database 260 as functional components.
- the vehicle 200 may further include an input receiver 270 .
- the vehicle 200 may further include a second detector 280 .
- the vehicle 200 may further include a congestion detector 290 .
- the second communicator 210 receives, via the communicator network 300 from the external information processing apparatus 100 , a detection request to detect the current road shape.
- the second communicator 210 is an example of a second communicator.
- the second communicator 210 may also transmit, via the communication network 300 to the information processing apparatus 100 , detection data indicating a road shape as detected by the first detector 250 . Further, the second communicator 210 may also transmit, via the communication network 300 to the information processing apparatus 100 , position information indicating the present position or route information indicating the running route. Further, the second communicator may also transmit a history of occurrences of a predetermined action of the vehicle 200 to the information processing apparatus 100 via the communication network 300 .
- the second communicator 210 is realized, for example, by the CPU 21 , the main memory 22 , the storage 23 , the communication IF 24 , and the like.
- the second determiner 220 determines whether a running mode of the vehicle 200 is switched from a normal mode to the detection mode.
- the normal mode is a running mode in which to perform normal running.
- the detection mode is a running mode in which to perform running that detects a three-dimensional road shape of a road on which the vehicle 200 is running. More specifically, the detection mode is a running mode in which to run in a state of change in at least one of the settings of making the maximum speed lower, making the maximum acceleration lower, making the intervehicular distance between the vehicle 200 and another vehicle larger, or making the first detector 250 perform detection at a higher detection frequency than in the normal mode.
- the maximum speed, maximum acceleration, intervehicular distance, and detection frequency to be set may take on predetermined fixed values or may be changed according to scores to values that are computed according to Formula 2 below. Control after change in setting may involve the use of a technology similar to a cruise control function or the like that is already in practical use.
- the detection frequency of the first detector 250 is a sampling frequency at which the first detector 250 performs detection.
- the detection frequency of the first detector 250 is a sampling frequency at which the first detector 250 detects positions in one direction.
- the detection frequency is a rotational frequency.
- v new max ⁇ ( v old ⁇ S th S , v min ) ( 2 )
- v new Maximum speed of detection mode
- v old Maximum speed of normal mode (such as speed limit)
- S th Threshold score
- Current score v min Minimum speed
- the second determiner 220 may determine that the running mode is switched.
- the second determiner 220 may determine that the running mode is not switched.
- the second determiner 220 may determine that the running mode is switched.
- the second determiner 220 may determine that the running mode is not switched.
- the clause “the vehicle 200 is on its way to pick up the passenger” indicates a state where the vehicle 200 is running to a place where the passenger is in order to carry the passenger.
- the clause “the vehicle 200 is carrying the passenger” indicates a state where the vehicle 200 is running with the passenger on board the vehicle 200 .
- the second determiner 220 is realized, for example, by the CPU 21 , the main memory 22 , the storage 23 , and the like.
- the switcher 230 switches the running mode to the detection mode, and in a case where the second determiner 220 has determined that switching to the detection is not made, the switcher 230 keeps the running mode in the normal mode.
- the controller 240 causes the vehicle 200 to run in a running mode to which the switcher 230 has made switching.
- the controller 240 may cause the vehicle 200 to run in the normal mode when the after-mentioned second detector 280 is detecting a following vehicle.
- the controller 240 may cause the vehicle 200 to run in the normal mode when the after-mentioned congestion detector 290 is detecting an occurrence of congestion in an area around the vehicle 200 .
- the controller 240 may cause the vehicle 200 to stop running in the detection mode and start to run in the normal mode in a case where a following vehicle has been detected or in a case where congestion has been detected.
- the first detector 250 detects a three-dimensional road shape of a road. In a case where the switcher 230 has switched the running mode to the detection mode, the first detector 250 detects a three-dimensional road shape of a road on which the vehicle 200 is running, with the vehicle 200 running in the detection mode.
- the first detector 250 is realized, for example, by the CPU 21 , the main memory 22 , the storage 23 , the imaging device 25 , the LIDAR 26 , and the like.
- the map database 260 stores a three-dimensional map.
- the most recent three-dimensional map stored in the map database 151 of the information processing apparatus 100 is acquired through the second communicator 210 every predetermined period of time, and the most recent three-dimensional map acquired through the second communicator 210 is stored in the map database 260 .
- the map database 260 is realized, for example, by the CPU 21 , the main memory 22 , the storage 23 , and the like.
- the input receiver 270 receives an input indicating whether to give approval to switch the running mode from the normal mode to the detection mode.
- the input receiver 270 receives the input from a person who is in the vehicle interior of the vehicle 200 .
- the input receiver 270 is realized, for example, by the display 29 and the input IF 30 and, as shown in FIG. 8A or 8B , displays, on the display 29 , a UI 291 that notifies reception of a detection request or a UI 292 that asks the user whether to switch the running mode to the detection mode.
- the user can operate the input IF 30 to select “YES” or “NO” being displayed on the UI 291 or the UI 292 .
- the input receiver 270 receives, as the input indicating whether to give approval to switch the running mode from the normal mode to the detection mode, the input indicating “YES” or “NO” received through the UI 291 or the UI 292 .
- the input receiver 270 receives, as an input indicating approval to switch to the detection mode, an input indicating “YES” received through the UI 291 or the UI 292 , and receives, as input indicating disapproval to switch to the detection mode, an input indicating “NO” received through the UI 291 or the UI 292 .
- the second detector 280 detects a following vehicle falling within a certain distance from the vehicle 200 .
- the second detector 280 is realized, for example, by the CPU 21 , the main memory 22 , the storage 23 , the imaging device 25 , the LIDAR 26 , and the like. It should be noted that in a case where the imaging device 25 fulfills a part of the function of the second detector 280 , the imaging device 25 takes an image of an area behind the vehicle 200 .
- the congestion detector 290 detects an occurrence of congestion in an area around the vehicle 200 .
- the congestion detector 290 may detect an occurrence of congestion by estimating whether the density of a plurality of other vehicles 200 in an area around the vehicle 200 is not lower than a predetermined density. Specifically, the congestion detector 290 may estimate the density by estimating how many other vehicles 200 there are in a predetermined range around the vehicle 200 .
- the congestion detector 290 is realized, for example, by the CPU 21 , the main memory 22 , the storage 23 , the imaging device 25 , the LIDAR 26 , and the like.
- the congestion detector 290 may detect an occurrence of congestion in an area around the vehicle 200 by, on the basis of the present position of the vehicle 200 , acquiring information on congestion in an area around the present position from the external information processing apparatus 100 .
- FIG. 9 is a sequence chart showing an example of a map updating method by the map updating system 1 according to the embodiment.
- the map updating method includes an information processing method that is executed by the information processing apparatus 100 and a running control method that is executed by a vehicle 200 .
- the first determiner 110 determines, for each of the plurality of areas, whether it is necessary to update the map of that area (S 11 ). An update determination process of determining whether it is necessary to update the map will be described in detail later.
- the selector 120 selects, from among one or more vehicles 200 that are running or planning to run in an area where it has been determined by the first determiner 110 that the map is needs to be updated, a detection vehicle that detects the road shape (S 12 ).
- the first communicator 130 transmits, via the communication network 300 to the detecting vehicle, a detection request that causes the detecting vehicle to detect the road shape of the area (S 13 ).
- the vehicle 200 selected as the detecting vehicle starts running control by receiving the detection request (S 21 ).
- the vehicle 200 runs in the detection mode, when permitted to run in the detection mode, and detects the road shape.
- the running control including the reception of the detection request will be described in detail later.
- the vehicle 200 transmits detection data obtained to the information processing apparatus 100 via the communication network 300 (S 22 ).
- the first communicator 130 receives the detection data transmitted from the vehicle 200 (S 14 ), and updating of the three-dimensional map of the map database 151 is performed (S 15 ).
- the information processing apparatus 100 does not need to include the map database 151 but may have its configuration separated into a first information processing apparatus including the first determiner 110 , the selector 120 , and the first communicator 130 and a second information processing apparatus including the database 150 .
- the first information processing apparatus acquires, from the second information processing apparatus, information for determining whether it is necessary to update the map and executes steps S 11 to S 13 .
- the vehicle 200 executes step S 21 and executes, instead of step S 22 , a step of transmitting the detection data to the second information processing apparatus. Then, the second information processing apparatus executes steps S 14 and S 15 .
- FIG. 10 is a flow chart showing a detailed example of an update determination process according to the embodiment.
- the information processing apparatus 100 repeatedly performs the update determination process every predetermined period of time.
- the first determiner 110 performs a loop of executing the following steps S 32 to S 35 for each of the plurality of areas A 1 to A 4 , B 1 to B 4 , and C 1 to C 4 (S 31 ).
- the first determiner 110 computes a score of one of the plurality of areas, e.g. the area A 1 (S 32 ).
- the first determiner 110 computes the score, for example, according to Formula 1.
- the first determiner 110 determines whether the score thus computed exceeds a predetermined score (S 33 ).
- the first determiner 110 determines that the area A 1 is needs updating (S 34 ).
- the first determiner 110 determines that the area A 1 does not need updating (S 35 ).
- step S 34 or S 35 is executed for the area A 1 , the first determiner 110 repeats the loop in sequence for each of the remaining areas A 2 to A 4 , B 1 to B 4 , and C 1 to C 4 , for which the loop has not been performed yet. In a case where the loop has been performed for all of the areas A 1 to A 4 , B 1 to B 4 , and C 1 to C 4 , the first determiner 110 ends the loop and end the update determination process.
- the first determiner 110 may alternatively perform the update determination process by executing steps S 32 to S 35 in parallel for each of the plurality of areas A 1 to A 4 , B 1 to B 4 , and C 1 to C 4 .
- FIG. 11 is a flow chart showing a detailed example of running control according to the embodiment.
- a vehicle 200 to which a detection request has been transmitted by the information processing apparatus 100 i.e. a vehicle 200 selected as a detecting vehicle by the information processing apparatus 100 , performs running control.
- the second communicator 210 receives the detection request (S 41 ).
- the second determiner 220 determines whether the running mode of the vehicle 200 is switched from the normal mode to the detection mode (S 42 ).
- the switcher 230 makes switching to the detection mode (S 43 ). In a case where the second determiner 220 has determined that the running mode is not switched to the detection mode (No in S 42 ), the switcher 230 keeps the running mode in the normal mode and ends the running control process.
- the controller 240 causes the vehicle 200 to run in the detection mode, to which switching has been made by the switcher 230 , in an area where it has been determined that updating is needed (S 44 ), and causes the first detector 250 to detect a road shape in the area.
- the controller 240 causes the vehicle 200 to run in a state of change in at least one of the settings of making the maximum speed lower, making the maximum acceleration lower, making the intervehicular distance between the vehicle 200 and another vehicle larger, or making the first detector 250 perform detection at a higher detection frequency than in the normal mode, and causes the first detector 250 to detect a road shape in this state.
- the switcher 230 determines whether detection of a road shape in the area has ended (S 45 ).
- the switcher 230 switches the running mode of the vehicle 200 from the detection mode to the normal mode (S 46 ).
- the switcher 230 repeats step S 45 . That is, the switcher 230 continues the detection mode until the end of detection of a road shape in the area and, once the detection ends, switches from the detection mode to the normal mode.
- the controller 240 causes the vehicle 200 to run in the normal mode, to which switching has been made by the switcher 230 (S 47 ), and ends the running control.
- FIG. 12 is a diagram for explaining an effect that is brought about by running at a larger intervehicular distance in the detection mode.
- the LIDAR 26 is disposed as a sensor on top of the vehicle 200 .
- the plurality of dashed lines indicate a laser that is emitted from the LIDAR 26 , and the LIDAR 26 emits a laser in a range from the dashed line 40 to the dashed line 41 .
- detection data detected by the LIDAR 26 is preferably data obtained by detecting an object described on a map, such as a surrounding building, tree, or road, but not data obtained by detecting the truck 400 temporarily running ahead.
- the LIDAR 26 cannot see through, it is difficult to detect an object that is on an opposite side of the truck 400 .
- the LIDAR 26 detects the truck 400 , which is an object that should not be detected, and cannot detect an object that should be detected. This raises the possibility of misrecognizing the truck 400 as a change in environment on the basis of the detection data thus detected. It is of course possible to detect the truck 400 from an image or the like obtained by the imaging device 25 in parallel to the LIDAR 26 and exclude updating the map with use of the detection data obtained by detecting the truck 400 . However, it is impossible to avoid being incapable of detecting an object that should be detected, as the process for detecting the truck 400 imposes a load. Therefore, it is effective to keep a large intervehicular distance from a vehicle ahead in order to effectively detect an object that should be detected, as in such a case as that shown in FIG. 12 where the truck 400 is running ahead.
- congestion information can be acquired by stock cars and lane information is usually included in map information, a following vehicle can be easily sensed by an all-round LiDAR or a rearview camera.
- FIGS. 13A and 13B are diagrams for explaining an effect that is brought about by running at a lower maximum speed or maximum acceleration in the detection mode.
- FIGS. 13A and 13B are diagrams of a vehicle 200 and an object 50 as viewed from above.
- the vehicle 200 is simply illustrated for the purpose of orientation or the like.
- the LIDAR 26 of the vehicle 200 is located at the origin point of an x-y plane in FIG. 13A and emits a laser 360-degree all around.
- the laser is indicated by dashed lines.
- the LIDAR 26 detects, as a point cloud, the coordinates of each of the points 51 to 57 , where the laser is reflected by the object 50 .
- the coordinates thus detected are values based on the center coordinates of the LIDAR 26 of the vehicle 200 , i.e. the origin point of the x-y plane, and are not coordinates on a map.
- the coordinates of the points 51 to 57 on the map can be found by performing coordinate transformation from a first coordinate system centered at the LIDAR 26 of the vehicle 200 into a second coordinate system on the map in consideration of the position and/or posture of the vehicle 200 on the map.
- the coordinate transformation involves the use of Formula 3.
- the position and/or posture of the vehicle 200 on the second coordinate system on the map is/are estimated by the aforementioned ICP or NDT algorithm.
- These algorithms have initial value dependence, as they are generally repetition algorithms. For this reason, in the estimation of the position and/or posture of the vehicle 200 with use of these algorithms, giving an inappropriate initial value leads to deterioration in estimation accuracy.
- the position and/or posture at the current time or the position and/or posture at the next time as estimated from speed, acceleration, angular acceleration, and the like is/are used as an initial value(s), the accuracy deteriorates during sudden acceleration or quick turning.
- the detection frequency of a sensor may be changed according to scores.
- the LIDAR 26 has its laser light receiver and emitter rotating therein and can change the rotational frequency of the laser light receiver and emitter.
- the LIDAR 26 causes the light receiver and emitter to rotate at a higher rotational frequency and thereby allows detection data to detected at a higher temporal density.
- a point cloud included in detection data hardly has a harmful effect, as it is sampled down for use.
- increasing the rotational frequency of the light receiver and emitter of the LIDAR 26 makes it possible to improve the accuracy of an initial position that is given in matching process in estimating the position of the vehicle 200 by increasing the temporal density of detection data and makes it possible to improve the accuracy of estimation of the position of the vehicle 200 .
- FIGS. 14A and 14B are diagrams for explaining an effect of inhibiting the occurrence of occlusion by a plurality of objects.
- the LIDAR 26 is capable of changing its detection frequency.
- an imaging device such as a stereo camera.
- the information processing apparatus 100 divides a three-dimensional map into a plurality of areas and determines, for each of the plurality of areas, whether it is necessary to update the map. Thus, the information processing apparatus 100 can efficiently determine an area where the map needs to be updated.
- the information processing apparatus 100 requests one or more vehicles 200 to detect detection data for an area where the map needs to be updated.
- the information processing apparatus 100 selectively transmits a detection request to a vehicle 200 that is running or planning to run in an area where updating is needed, the information processing apparatus 100 can effectively acquire accurate detection data. For this reason, the information processing apparatus 100 can effectively update map data so that it becomes more accurate.
- the vehicle 200 is a vehicle that performs automatic driving.
- the predetermined action is at least one of sudden deceleration, sudden acceleration, manual driving interruption during the automatic driving, and quick avoidance that have occurred during a predetermined period of time up to the present time. That is, the information processing apparatus 100 determines that a larger number of occurrences of predetermined actions in a vehicle that performs automatic driving lead to a greater difference between a road shape shown by a map and an actual road shape. This is because a difference between a road shape shown by a map and an actual road shape leads to deterioration in accuracy of estimation of the position of the vehicle, thus making stable running control difficult.
- the information processing apparatus 100 determines that a larger number of occurrences of predetermined actions in a vehicle that performs automatic driving lead to a greater difference between a road shape shown by a map and an actual road shape, the information processing apparatus 100 can effectively determine whether it is necessary to update the map.
- the selector 120 selects, as the detecting vehicle, a vehicle 200 having a sensor that detects the road shape and whose detection accuracy is not lower than a predetermined degree of accuracy. This makes it possible to select a vehicle 200 that can detect the road shape with a high degree of accuracy.
- the selecting 120 selects, as the detecting vehicle, a vehicle 200 that has ever run in a detection mode in which to detect the road shape in order to update the map. This makes it possible to select a vehicle that is highly likely to be able to run in the detection mode in which to detect the road shape with a high degree of accuracy. This makes it possible to reduce the time it takes to select a vehicle that run in the detection mode and the processing load on the selection.
- the first determiner 110 computes a score for each of the plurality of areas on the basis of at least either the amount of change in the road shape or the number of occurrences of predetermined actions. The first determiner 110 determines that it is necessary to update the three-dimensional map of that one of the plurality of areas where the score thus computed exceeds a predetermined score. This makes it possible to effectively and easily determine, for each of the plurality of areas, whether it is necessary to update the map of that area.
- the first determiner 110 increases a predetermined score of that one of the plurality of areas where the number of times it has been determined per unit period that updating is needed is not smaller than a first number of times and decreases a predetermined score of that one of the plurality of areas where the number of times it has been determined per unit period that updating is needed is smaller than a second number of times that is not larger than the first number of times.
- a threshold score for determining whether it is necessary to update the map an area with a greater time series variation in the road shape can be updated at a higher frequency. This makes it possible to update the map of the area at a frequency according to the time series variation in the road shape. This makes it possible to update the map of each of the plurality of areas at a frequency appropriate for that area.
- a vehicle 200 Upon receiving a detection request, a vehicle 200 according to the present embodiment runs in a detection mode, different from a normal mode, in which to detect a road shape for use in updating of a map. For this reason, the vehicle 200 can detect the road shape with a high degree of accuracy. Further, even upon receiving a detection request from an information processing apparatus, the vehicle 200 keeps running in the normal mode without switching to the detection mode, depending on the situation of the vehicle 200 ; therefore, the vehicle 200 can inhibit impairment of convenience for a user of the vehicle 200 .
- the second communicator 210 transmits, via the communication network 300 to the information processing apparatus 100 , detection data indicating the road shape detected by the first detector 250 . For this reason, the vehicle 200 can transmit, to the information processing apparatus 100 , detection data detected with a high degree of accuracy in an area where the map needs to be updated.
- the detection mode is a running mode in which to run in a state of change in at least one of the settings of making the maximum running speed lower, making the maximum running acceleration lower, making the intervehicular distance larger, or making the first detector 250 perform detection at a higher detection frequency than in the normal mode. For this reason, the vehicle 200 can detect the road shape with a high degree of accuracy in the detection mode.
- the vehicle 200 further includes a second detector 280 that detects a following vehicle falling within a certain distance from the vehicle 200 .
- the controller 240 causes the vehicle 200 to run in the normal mode when the second detector 280 is detecting a following vehicle. For this reason, even in a case where the running mode has been switched to the detection mode, the vehicle 200 can run out of the way of a following vehicle in a case where the second detector 280 has detected the following vehicle.
- the vehicle 200 further includes a congestion detector 290 that detects an occurrence of congestion in an area around the vehicle 200 .
- the controller 240 causes the vehicle 200 to run in the normal mode when the congestion detector 290 is detecting an occurrence of congestion. For this reason, even in a case where the running mode has been switched to the detection mode, the vehicle 200 can run out of the way of a stuck vehicle in a case where an area around the vehicle 200 is congested.
- the vehicle 200 further includes an input receiver 270 that receives an input indicating whether to give approval to switch the running mode from the normal mode to the detection mode.
- the second determiner 220 determines that the running mode is switched, and in a case where an input received by the input receiver 270 indicates disapproval, the second determiner 220 determines that the running mode is not switched. For this reason, since, in the absence of approval from a person who is in the vehicle interior, the vehicle 200 runs in the normal mode without switching the running mode, the vehicle 200 can inhibit impairment of convenience for the person.
- the vehicle 200 is a vehicle that carries a passenger. Further, in a case where the vehicle 200 is not on its way to pick up the passenger and the vehicle 200 is not carrying the passenger, the second determiner 220 determines that the running mode is switched. In a case where the vehicle 200 is on its way to pick up the passenger or the vehicle 200 is carrying the passenger, the second determiner 220 determines that the running mode is not switched. For this reason, since, in a case where the vehicle 200 is on its way to pick up the passenger or is carrying the passenger, the vehicle 200 runs while keeping the running mode in the detection mode, the vehicle 200 can inhibit impairment of convenience for the passenger.
- the LIDAR 26 is used to detect a three-dimensional shape of a terrain surface of an area around the vehicle 200 , this does not imply any limitation.
- a stereo camera may be used to detect a three-dimensional shape of a terrain surface of an area around the vehicle 200 from an image taken by the stereo camera, or a combination of a distance sensor and a camera may be used to detect a three-dimensional shape of a terrain surface of an area around the vehicle 200 from a result of detection yielded by the distance sensor and an image taken by the camera.
- the first detector 250 which is realized by the LIDAR 26 and the like, may detect a road around the vehicle 200 and a road shape, i.e. a shape of an area around the road, in two-dimensional shapes, although the foregoing embodiment assumes that the first detector 250 detects them in three-dimensional shapes.
- the first detector 250 may also perform two-dimensional shape detection by transforming a road shape detected in a three-dimensional shape into a two-dimensional shape.
- the vehicle 200 is configured such that in a case where the switcher 230 has switched the running mode to the detection mode, the controller 240 causes the vehicle 200 to run in the normal mode when the second detector 280 is detecting a following vehicle.
- the second determiner 220 may determine that the running mode is not switched to the detection mode.
- the vehicle 200 is configured such that in a case where the switcher 230 has switched the running mode to the detection mode, the controller 240 causes the vehicle 200 to run in the normal mode when the congestion detector 290 is detecting an occurrence of congestion in an area around the vehicle 200 .
- the second determiner 220 may determine that the running mode is not switched to the detection mode.
- the information processing apparatus 100 may be configured such that the information processing apparatus 100 , a service server, or the like includes such a service that in a case where a vehicle 200 having received a detection request that causes a detection vehicle selected by the information processing apparatus 100 to detect the road shape of an area changes its running mode from the normal mode to the detection mode, remuneration is given to an owner of the vehicle 200 or a person who is on board the vehicle 200 .
- the remuneration here may be giving money or granting bonus point service.
- the information processing apparatus 100 can uniquely identify a vehicle 200 . Therefore, for example, in a case where the vehicle 200 is a self-guided taxi, an example of a service in a case where a passenger who is on board the vehicle 200 gives approval to change the running mode may be a mechanism by which a terminal attached to the vehicle 200 displays a particular QR code (registered trademark), the passenger then scans the QR code (registered trademark) with his/her own smartphone or the like, the smartphone or the like transmits, to the information processing apparatus 100 or the service server, information indicating the approval to change the running mode, and the passenger thereby obtains the remuneration; that is, the information processing apparatus 100 gives the remuneration to the passenger.
- an account of an owner may be associated with each vehicle 200 , and information indicating approval to change the running mode may be stored in the account.
- each constituted element may be configured by dedicated hardware or realized by executing a software program suited to that constituted element.
- Each constituent element may be realized by a program executor such as a CPU or a processor reading out a software program stored on a recording medium such as a hard disk or a semiconductor memory.
- a software that realizes the information processing method, the running control method, the map updating method, and the like according to each of the embodiments is the following program.
- this program causes a computer to execute an information processing method that is executed by an information processing apparatus for distributing, to a vehicle, a map showing a road on which the vehicle runs and a road shape that is a shape of an area around the road, including: determining, for each of the plurality of areas, whether it is necessary to update the map of that area, the determining being based on at least either an amount of change from the road shape shown by the map to the road shape of a present time or the number of occurrences of a predetermined action of a vehicle having run through that area, the map being stored in a memory of the information processing apparatus.
- this program causes a computer to execute a running control method that is executed by a vehicle, including: detecting a road on which the vehicle runs and a road shape that is a shape of an area around the road; receiving, via a communication network from an external information apparatus, a detection request to detect the road shape; determining whether a running mode of the vehicle is switched from a normal mode in which to perform normal running to a detection mode in which to, in order to update the map, perform running that detects the road shape of a road on which the vehicle is running; in a case where it has been determined that switching to the detection mode is made, switching the running mode to the detection mode and controlling the vehicle to run in the detection mode; and in a case where it has been determined that switching to the detection mode is not made, keeping the running mode in the detection mode and controlling the vehicle to run in the normal mode.
- this program causes a computer to execute a map updating method including updating a map in a map updating system including an information processing apparatus for distributing, to one or more vehicles, the map showing a road on which the one or more vehicles run and a road shape that is a shape of an area around the road and the one or more vehicles, wherein the information processing apparatus performs operations including: determining, for each of the plurality of areas, whether it is necessary to update the map of that area, the determining being based on at least either an amount of change from the road shape shown by the map to the road shape of a present time or the number of occurrences of a predetermined action of a vehicle having run through that area, the map being stored in a memory of the information processing apparatus; selecting, from among one or more vehicles that are running or planning to run in an area where it has been determined that the map needs to be updated, a detection vehicle that detects the road shape with a high degree of accuracy; and transmitting, via a communication network to the detecting vehicle thus selected, a detection request that causes the a
- the present disclosure is useful as an information processing apparatus, a vehicle, an information processing method, a running control method, a map updating method, and the like that make it possible to update map data so that it becomes more accurate.
Landscapes
- Engineering & Computer Science (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Automation & Control Theory (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- Aviation & Aerospace Engineering (AREA)
- Traffic Control Systems (AREA)
- Navigation (AREA)
Abstract
Description
vnew: Maximum speed of detection mode
vold: Maximum speed of normal mode (such as speed limit)
Sth: Threshold score
S: Current score
vmin: Minimum speed
xb: x coordinate of point on second coordinate system
yb: y coordinate of point on second coordinate system
zb: z coordinate of point on second coordinate system
R: 3×3 rotation matrix of posture of
t: position (xc, yc, zc) T of
x: x coordinate of point on first coordinate system
y: y coordinate of point on first coordinate system
z: z coordinate of point on first coordinate system
Claims (19)
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017-163739 | 2017-08-28 | ||
JPJP2017-163739 | 2017-08-28 | ||
JP2017163739 | 2017-08-28 | ||
JP2018-075712 | 2018-04-10 | ||
JP2018075712A JP7141845B2 (en) | 2017-08-28 | 2018-04-10 | Information processing device, information processing method, and map update method |
JPJP2018-075712 | 2018-04-10 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20190063929A1 US20190063929A1 (en) | 2019-02-28 |
US10935385B2 true US10935385B2 (en) | 2021-03-02 |
Family
ID=62951980
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/045,858 Active US10935385B2 (en) | 2017-08-28 | 2018-07-26 | Information processing apparatus, vehicle, information processing method, running control method, and map updating method |
Country Status (3)
Country | Link |
---|---|
US (1) | US10935385B2 (en) |
EP (1) | EP3460406B1 (en) |
CN (1) | CN109425358B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11226207B2 (en) * | 2019-12-18 | 2022-01-18 | GM Cruise Holdings, LLC | Semantic map with live updates |
US20220048524A1 (en) * | 2018-11-05 | 2022-02-17 | Toyota Jidosha Kabushiki Kaisha | Map information system |
US11409728B2 (en) * | 2018-10-02 | 2022-08-09 | Toyota Jidosha Kabushiki Kaisha | Map information system |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPWO2016147623A1 (en) * | 2015-03-18 | 2018-01-25 | 日本電気株式会社 | Operation control device, operation control method, and inter-vehicle communication system |
TWI674393B (en) * | 2018-11-09 | 2019-10-11 | 財團法人車輛研究測試中心 | Multi-positioning system switching and fusion correction method and device thereof |
US11402220B2 (en) * | 2019-03-13 | 2022-08-02 | Here Global B.V. | Maplets for maintaining and updating a self-healing high definition map |
WO2020226085A1 (en) * | 2019-05-09 | 2020-11-12 | ソニー株式会社 | Information processing device, information processing method, and program |
DE112020002824T5 (en) * | 2019-06-13 | 2022-03-10 | Denso Corporation | Map data generation system, data center and on-vehicle device |
CN112347206A (en) * | 2019-08-06 | 2021-02-09 | 华为技术有限公司 | Map updating method, device and storage medium |
JP7111118B2 (en) * | 2020-01-29 | 2022-08-02 | トヨタ自動車株式会社 | Map generation data collection device and map generation data collection method |
JP7136138B2 (en) * | 2020-01-29 | 2022-09-13 | トヨタ自動車株式会社 | Map generation data collection device, map generation data collection method, and vehicle |
CN115280393B (en) * | 2020-03-18 | 2024-06-04 | 三菱电机株式会社 | Travel road determination device and travel road determination method |
JP7013512B2 (en) * | 2020-03-30 | 2022-01-31 | 本田技研工業株式会社 | Information processing device, route guidance device, program, and information processing method |
JP7302525B2 (en) * | 2020-04-27 | 2023-07-04 | トヨタ自動車株式会社 | Map management device and map management method |
JP7287353B2 (en) * | 2020-06-11 | 2023-06-06 | トヨタ自動車株式会社 | Position estimation device and position estimation computer program |
JP7238853B2 (en) * | 2020-06-11 | 2023-03-14 | トヨタ自動車株式会社 | Map generation data collection device and map generation data collection method |
CN111880527B (en) * | 2020-06-19 | 2022-12-27 | 中国煤炭科工集团太原研究院有限公司 | Robot control method for underground unmanned transport vehicle |
CN112163063B (en) * | 2020-10-22 | 2023-07-25 | 腾讯科技(深圳)有限公司 | Method, apparatus and computer readable storage medium for generating high precision map |
JP7327355B2 (en) * | 2020-11-05 | 2023-08-16 | トヨタ自動車株式会社 | Map update device and map update method |
CN112466005B (en) * | 2020-11-26 | 2022-08-09 | 重庆长安汽车股份有限公司 | User use habit-based automatic driving fence updating system and method |
GB2601780A (en) * | 2020-12-10 | 2022-06-15 | Daimler Ag | Method for verifying map data of an environment of a vehicle |
KR20220105777A (en) * | 2021-01-21 | 2022-07-28 | 현대자동차주식회사 | Apparatus and method for generating a road map |
DE102021107716A1 (en) * | 2021-03-26 | 2022-09-29 | Bayerische Motoren Werke Aktiengesellschaft | Method for providing a navigation map to a vehicle, computer-readable medium, system, and vehicle |
CN113722342A (en) * | 2021-09-03 | 2021-11-30 | 北京百度网讯科技有限公司 | High-precision map element change detection method, device and equipment and automatic driving vehicle |
CN117367456B (en) * | 2023-12-05 | 2024-05-28 | 文诚恒远(天津)供应链管理服务有限公司 | Vehicle route deviation detection method and device, storage medium and electronic equipment |
Citations (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5994570A (en) * | 1997-10-31 | 1999-11-30 | Shin-Etsu Chemical Co., Ltd. | Platinum complex catalyst |
US20050278098A1 (en) * | 1994-05-23 | 2005-12-15 | Automotive Technologies International, Inc. | Vehicular impact reactive system and method |
US20070109111A1 (en) * | 1997-10-22 | 2007-05-17 | Intelligent Technologies International, Inc. | Accident Avoidance Systems and Methods |
US20070213922A1 (en) * | 2006-03-10 | 2007-09-13 | Van Buer Darrel J | Traffic notification system for reporting traffic anomalies based on historical probe vehicle data |
US20080015771A1 (en) * | 1997-10-22 | 2008-01-17 | Intelligent Technologies International, Inc. | Information Transfer Arrangement and Method for Vehicles |
US20080040029A1 (en) * | 1997-10-22 | 2008-02-14 | Intelligent Technologies International, Inc. | Vehicle Position Determining System and Method |
US20080040004A1 (en) * | 1994-05-23 | 2008-02-14 | Automotive Technologies International, Inc. | System and Method for Preventing Vehicular Accidents |
US20080046150A1 (en) * | 1994-05-23 | 2008-02-21 | Automotive Technologies International, Inc. | System and Method for Detecting and Protecting Pedestrians |
US20080042815A1 (en) * | 1997-10-22 | 2008-02-21 | Intelligent Technologies International, Inc. | Vehicle to Infrastructure Information Conveyance System and Method |
US20080106436A1 (en) * | 1997-10-22 | 2008-05-08 | Intelligent Technologies International, Inc. | In-Vehicle Signage Techniques |
US20080133136A1 (en) * | 1997-10-22 | 2008-06-05 | Intelligent Technologies International, Inc. | Intersection Collision Avoidance Techniques |
US20080140318A1 (en) * | 1997-10-22 | 2008-06-12 | Intelligent Technologies International, Inc. | Weather Monitoring Techniques |
US20080147253A1 (en) * | 1997-10-22 | 2008-06-19 | Intelligent Technologies International, Inc. | Vehicular Anticipatory Sensor System |
US20080154495A1 (en) * | 1997-10-22 | 2008-06-26 | Intelligent Technologies International, Inc. | Inertial Measurement Unit for Aircraft |
US20080150786A1 (en) * | 1997-10-22 | 2008-06-26 | Intelligent Technologies International, Inc. | Combined Imaging and Distance Monitoring for Vehicular Applications |
US20080154629A1 (en) * | 1997-10-22 | 2008-06-26 | Intelligent Technologies International, Inc. | Vehicle Speed Control Method and Arrangement |
US20080161986A1 (en) * | 1997-10-22 | 2008-07-03 | Intelligent Technologies International, Inc. | Autonomous Vehicle Travel Control Systems and Methods |
US7769544B2 (en) * | 2001-05-07 | 2010-08-03 | Ansaldo Sts Usa, Inc. | Autonomous vehicle railroad crossing warning system |
US7979172B2 (en) * | 1997-10-22 | 2011-07-12 | Intelligent Technologies International, Inc. | Autonomous vehicle travel control systems and methods |
US20130024548A1 (en) * | 2011-07-21 | 2013-01-24 | Hitachi Automotive Systems, Ltd. | Map data distribution server, map data distribution system, and map data distribution method |
US8571722B2 (en) * | 2010-10-22 | 2013-10-29 | Toyota Motor Engineering & Manufacturing North America, Inc. | Method for safely parking vehicle near obstacles |
US20130304513A1 (en) * | 2012-05-08 | 2013-11-14 | Elwha Llc | Systems and methods for insurance based on monitored characteristics of an autonomous drive mode selection system |
US8983738B2 (en) * | 2010-02-23 | 2015-03-17 | Israel Aerospace Industries Ltd. | System and method of autonomous operation of multi-tasking earth moving machinery |
US9103671B1 (en) * | 2007-11-29 | 2015-08-11 | American Vehicular Sciences, LLC | Mapping techniques using probe vehicles |
US9141107B2 (en) * | 2013-04-10 | 2015-09-22 | Google Inc. | Mapping active and inactive construction zones for autonomous driving |
JP2016153832A (en) | 2015-02-20 | 2016-08-25 | パイオニア株式会社 | Map data storage device, control method, program, and storage medium |
JP2016156973A (en) | 2015-02-25 | 2016-09-01 | パイオニア株式会社 | Map data storage device, control method, program and recording medium |
US20170122749A1 (en) * | 2015-11-04 | 2017-05-04 | Toyota Jidosha Kabushiki Kaisha | Map update determination system |
US20180149487A1 (en) * | 2016-11-26 | 2018-05-31 | Thinkware Corporation | Image processing apparatus, image processing method, computer program and computer readable recording medium |
US20180188045A1 (en) * | 2016-12-30 | 2018-07-05 | DeepMap Inc. | High definition map updates based on sensor data collected by autonomous vehicles |
US20180217612A1 (en) * | 2017-01-30 | 2018-08-02 | Vladimeros Vladimerou | Determination of roadway features |
US20180345962A1 (en) * | 2017-06-02 | 2018-12-06 | Honda Motor Co., Ltd. | Automatic traveling control system and server device |
US10232856B2 (en) * | 2016-07-28 | 2019-03-19 | Ford Global Technologies, Llc | Vehicle user-communication system and method |
US20190347249A1 (en) * | 2017-03-28 | 2019-11-14 | Clarion Co., Ltd. | In-vehicle device and map updating system |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005165639A (en) * | 2003-12-02 | 2005-06-23 | Matsushita Electric Ind Co Ltd | Difference map data generation device, map data update device, difference map data generation program and map difference update system |
JP2007052341A (en) * | 2005-08-19 | 2007-03-01 | Fujitsu Ten Ltd | Road data updating device |
WO2007080829A1 (en) * | 2006-01-10 | 2007-07-19 | Pioneer Corporation | Data update device, data update method, data update program, and recording medium |
JP4702108B2 (en) * | 2006-03-03 | 2011-06-15 | 株式会社デンソー | Method for updating map data of map display device and map display device |
JP5315363B2 (en) * | 2011-01-19 | 2013-10-16 | 株式会社ゼンリン | Road network analysis system |
JP6322812B2 (en) * | 2014-08-21 | 2018-05-16 | パナソニックIpマネジメント株式会社 | Information management apparatus, vehicle, and information management method |
WO2016077690A1 (en) * | 2014-11-14 | 2016-05-19 | Henderson Ricky Jay | Power docking port system |
JP5997797B2 (en) * | 2015-03-03 | 2016-09-28 | 富士重工業株式会社 | Vehicle map data processing device |
JP6849330B2 (en) * | 2015-08-28 | 2021-03-24 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | Map generation method, self-position estimation method, robot system, and robot |
-
2018
- 2018-07-13 EP EP18183505.9A patent/EP3460406B1/en active Active
- 2018-07-26 US US16/045,858 patent/US10935385B2/en active Active
- 2018-08-17 CN CN201810960398.1A patent/CN109425358B/en active Active
Patent Citations (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080040004A1 (en) * | 1994-05-23 | 2008-02-14 | Automotive Technologies International, Inc. | System and Method for Preventing Vehicular Accidents |
US20050278098A1 (en) * | 1994-05-23 | 2005-12-15 | Automotive Technologies International, Inc. | Vehicular impact reactive system and method |
US20080046150A1 (en) * | 1994-05-23 | 2008-02-21 | Automotive Technologies International, Inc. | System and Method for Detecting and Protecting Pedestrians |
US20080161986A1 (en) * | 1997-10-22 | 2008-07-03 | Intelligent Technologies International, Inc. | Autonomous Vehicle Travel Control Systems and Methods |
US20080150786A1 (en) * | 1997-10-22 | 2008-06-26 | Intelligent Technologies International, Inc. | Combined Imaging and Distance Monitoring for Vehicular Applications |
US20080015771A1 (en) * | 1997-10-22 | 2008-01-17 | Intelligent Technologies International, Inc. | Information Transfer Arrangement and Method for Vehicles |
US7979172B2 (en) * | 1997-10-22 | 2011-07-12 | Intelligent Technologies International, Inc. | Autonomous vehicle travel control systems and methods |
US20070109111A1 (en) * | 1997-10-22 | 2007-05-17 | Intelligent Technologies International, Inc. | Accident Avoidance Systems and Methods |
US20080042815A1 (en) * | 1997-10-22 | 2008-02-21 | Intelligent Technologies International, Inc. | Vehicle to Infrastructure Information Conveyance System and Method |
US20080106436A1 (en) * | 1997-10-22 | 2008-05-08 | Intelligent Technologies International, Inc. | In-Vehicle Signage Techniques |
US20080133136A1 (en) * | 1997-10-22 | 2008-06-05 | Intelligent Technologies International, Inc. | Intersection Collision Avoidance Techniques |
US20080140318A1 (en) * | 1997-10-22 | 2008-06-12 | Intelligent Technologies International, Inc. | Weather Monitoring Techniques |
US20080147253A1 (en) * | 1997-10-22 | 2008-06-19 | Intelligent Technologies International, Inc. | Vehicular Anticipatory Sensor System |
US20080154495A1 (en) * | 1997-10-22 | 2008-06-26 | Intelligent Technologies International, Inc. | Inertial Measurement Unit for Aircraft |
US20080040029A1 (en) * | 1997-10-22 | 2008-02-14 | Intelligent Technologies International, Inc. | Vehicle Position Determining System and Method |
US20080154629A1 (en) * | 1997-10-22 | 2008-06-26 | Intelligent Technologies International, Inc. | Vehicle Speed Control Method and Arrangement |
US5994570A (en) * | 1997-10-31 | 1999-11-30 | Shin-Etsu Chemical Co., Ltd. | Platinum complex catalyst |
US7769544B2 (en) * | 2001-05-07 | 2010-08-03 | Ansaldo Sts Usa, Inc. | Autonomous vehicle railroad crossing warning system |
US20070213922A1 (en) * | 2006-03-10 | 2007-09-13 | Van Buer Darrel J | Traffic notification system for reporting traffic anomalies based on historical probe vehicle data |
US9103671B1 (en) * | 2007-11-29 | 2015-08-11 | American Vehicular Sciences, LLC | Mapping techniques using probe vehicles |
US8983738B2 (en) * | 2010-02-23 | 2015-03-17 | Israel Aerospace Industries Ltd. | System and method of autonomous operation of multi-tasking earth moving machinery |
US8571722B2 (en) * | 2010-10-22 | 2013-10-29 | Toyota Motor Engineering & Manufacturing North America, Inc. | Method for safely parking vehicle near obstacles |
US20130024548A1 (en) * | 2011-07-21 | 2013-01-24 | Hitachi Automotive Systems, Ltd. | Map data distribution server, map data distribution system, and map data distribution method |
US20130304513A1 (en) * | 2012-05-08 | 2013-11-14 | Elwha Llc | Systems and methods for insurance based on monitored characteristics of an autonomous drive mode selection system |
US9141107B2 (en) * | 2013-04-10 | 2015-09-22 | Google Inc. | Mapping active and inactive construction zones for autonomous driving |
JP2016153832A (en) | 2015-02-20 | 2016-08-25 | パイオニア株式会社 | Map data storage device, control method, program, and storage medium |
JP2016156973A (en) | 2015-02-25 | 2016-09-01 | パイオニア株式会社 | Map data storage device, control method, program and recording medium |
US20170122749A1 (en) * | 2015-11-04 | 2017-05-04 | Toyota Jidosha Kabushiki Kaisha | Map update determination system |
US10232856B2 (en) * | 2016-07-28 | 2019-03-19 | Ford Global Technologies, Llc | Vehicle user-communication system and method |
US20180149487A1 (en) * | 2016-11-26 | 2018-05-31 | Thinkware Corporation | Image processing apparatus, image processing method, computer program and computer readable recording medium |
US20180188045A1 (en) * | 2016-12-30 | 2018-07-05 | DeepMap Inc. | High definition map updates based on sensor data collected by autonomous vehicles |
US20180217612A1 (en) * | 2017-01-30 | 2018-08-02 | Vladimeros Vladimerou | Determination of roadway features |
US20190347249A1 (en) * | 2017-03-28 | 2019-11-14 | Clarion Co., Ltd. | In-vehicle device and map updating system |
US20180345962A1 (en) * | 2017-06-02 | 2018-12-06 | Honda Motor Co., Ltd. | Automatic traveling control system and server device |
Non-Patent Citations (2)
Title |
---|
A. K. Aijazi et al., "Detecting and Updating Changes in Lidar Point Clouds for Automatic 3D Urban Cartography", ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci., II-5/W2, Nov. 11, 2013, pp. 7-12, http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/II-5-W2/7/2013/isprsannals-II-5-W2-7-2013.pdf. |
The Extended European Search Report dated Feb. 27, 2019 for European Patent Application No. 18183505.9. |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11409728B2 (en) * | 2018-10-02 | 2022-08-09 | Toyota Jidosha Kabushiki Kaisha | Map information system |
US20220048524A1 (en) * | 2018-11-05 | 2022-02-17 | Toyota Jidosha Kabushiki Kaisha | Map information system |
US11680820B2 (en) | 2018-11-05 | 2023-06-20 | Toyota Jidosha Kabushiki Kaisha | Map information system |
US11692848B2 (en) * | 2018-11-05 | 2023-07-04 | Toyota Jidosha Kabushiki Kaisha | Map information system |
US11226207B2 (en) * | 2019-12-18 | 2022-01-18 | GM Cruise Holdings, LLC | Semantic map with live updates |
US11808600B2 (en) | 2019-12-18 | 2023-11-07 | Gm Cruise Holdings Llc | Semantic map with live updates |
US12104920B2 (en) | 2019-12-18 | 2024-10-01 | Gm Cruise Holdings Llc | Semantic map with live updates |
Also Published As
Publication number | Publication date |
---|---|
EP3460406B1 (en) | 2024-04-03 |
CN109425358B (en) | 2023-12-22 |
US20190063929A1 (en) | 2019-02-28 |
CN109425358A (en) | 2019-03-05 |
EP3460406A1 (en) | 2019-03-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10935385B2 (en) | Information processing apparatus, vehicle, information processing method, running control method, and map updating method | |
JP7141845B2 (en) | Information processing device, information processing method, and map update method | |
EP3327464B1 (en) | Algorithm and infrastructure for robust and efficient vehicle localization | |
US9977431B2 (en) | Automotive drone deployment system | |
US10471955B2 (en) | Stop sign and traffic light alert | |
US11155268B2 (en) | Utilizing passenger attention data captured in vehicles for localization and location-based services | |
US20220188695A1 (en) | Autonomous vehicle system for intelligent on-board selection of data for training a remote machine learning model | |
US11613253B2 (en) | Method of monitoring localization functions in an autonomous driving vehicle | |
US11073403B2 (en) | Map selection for vehicle pose system | |
US20200133303A1 (en) | Map information system | |
WO2022072412A1 (en) | Methods and systems for performing outlet inference by an autonomous vehicle to determine feasible paths through an intersection | |
US11851083B2 (en) | Methods and system for constructing data representation for use in assisting autonomous vehicles navigate intersections | |
US20210325898A1 (en) | Using drone data to generate high-definition map for autonomous vehicle navigation | |
US11970185B2 (en) | Data structure for storing information relating to an environment of an autonomous vehicle and methods of use thereof | |
US11880203B2 (en) | Methods and system for predicting trajectories of uncertain road users by semantic segmentation of drivable area boundaries | |
CN115220439A (en) | System and method for a vehicle and storage medium | |
US11358598B2 (en) | Methods and systems for performing outlet inference by an autonomous vehicle to determine feasible paths through an intersection | |
US11588957B2 (en) | Dirty monocular camera detection methods for autonomous driving vehicle | |
KR20220059388A (en) | LIGHT DETECTION AND RANGING (LiDAR) SCAN SMOOTHING | |
US20220212694A1 (en) | Methods and systems for generating a longitudinal plan for an autonomous vehicle based on behavior of uncertain road users | |
WO2021117464A1 (en) | Driving assistance device and driving assistance system | |
JP2010107410A (en) | Navigation apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: PANASONIC INTELLECTUAL PROPERTY CORPORATION OF AME Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OHYAMA, SHINJI;NAKATA, YOHEI;YAMAGUCHI, TAKUYA;SIGNING DATES FROM 20180622 TO 20180626;REEL/FRAME:047077/0985 Owner name: PANASONIC INTELLECTUAL PROPERTY CORPORATION OF AMERICA, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OHYAMA, SHINJI;NAKATA, YOHEI;YAMAGUCHI, TAKUYA;SIGNING DATES FROM 20180622 TO 20180626;REEL/FRAME:047077/0985 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |