US11252380B2 - Road management system - Google Patents

Road management system Download PDF

Info

Publication number
US11252380B2
US11252380B2 US16/800,685 US202016800685A US11252380B2 US 11252380 B2 US11252380 B2 US 11252380B2 US 202016800685 A US202016800685 A US 202016800685A US 11252380 B2 US11252380 B2 US 11252380B2
Authority
US
United States
Prior art keywords
moving body
vehicle
avoidance position
section
avoidance
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US16/800,685
Other versions
US20200275061A1 (en
Inventor
Yasuo Oishi
Kazuya Matsuura
Atsuki Kakinuma
Takeo Tokunaga
Akira Iihoshi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honda Motor Co Ltd
Original Assignee
Honda Motor Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honda Motor Co Ltd filed Critical Honda Motor Co Ltd
Assigned to HONDA MOTOR CO., LTD. reassignment HONDA MOTOR CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: IIHOSHI, AKIRA, KAKINUMA, ATSUKI, MATSUURA, KAZUYA, OISHI, YASUO, TOKUNAGA, TAKEO
Publication of US20200275061A1 publication Critical patent/US20200275061A1/en
Application granted granted Critical
Publication of US11252380B2 publication Critical patent/US11252380B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/188Capturing isolated or intermittent images triggered by the occurrence of a predetermined event, e.g. an object reaching a predetermined position
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/09Arrangements for giving variable traffic instructions
    • G08G1/0962Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
    • G08G1/0967Systems involving transmission of highway information, e.g. weather, speed limits
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R11/00Arrangements for holding or mounting articles, not otherwise provided for
    • B60R11/04Mounting of cameras operative during drive; Arrangement of controls thereof relative to the vehicle
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R21/00Arrangements or fittings on vehicles for protecting or preventing injuries to occupants or pedestrians in case of accidents or other traffic risks
    • B60R21/01Electrical circuits for triggering passive safety arrangements, e.g. airbags, safety belt tighteners, in case of vehicle accidents or impending vehicle accidents
    • B60R21/013Electrical circuits for triggering passive safety arrangements, e.g. airbags, safety belt tighteners, in case of vehicle accidents or impending vehicle accidents including means for detecting collisions, impending collisions or roll-over
    • G06K9/00791
    • G06K9/00805
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/166Anti-collision systems for active traffic, e.g. moving vehicles, pedestrians, bikes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/183Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/165Anti-collision systems for passive traffic, e.g. including static obstacles, trees

Definitions

  • the present invention relates to a road management system that acquires an image of a road surface.
  • Japanese Patent No. 5853719 discloses: an image processing system that detects a state (a dry state or a frozen state) of a road surface and utilizes the detected state for such objects as calling attention of a driver to the detected state; and a vehicle comprising the image processing system.
  • the present invention which was made considering such a problem, has an object of providing a road management system that can suppress a load of arithmetic processing when performing analysis of a road surface.
  • An aspect of the present invention comprises:
  • an avoiding action detecting section configured to detect an avoiding action of a moving body moving along a road
  • a position detecting section configured to detect a position of the moving body
  • an avoidance position determining section configured to determine an avoidance position where the avoiding action has occurred, based on a detection result of the avoiding action detecting section and a detection result of the position detecting section;
  • an imaging section configured to image a road surface of the avoidance position.
  • the load of arithmetic processing when performing analysis of the road surface can be suppressed.
  • FIG. 1 is a schematic view of a road management system according to a first embodiment
  • FIG. 2 is a configuration diagram of a vehicle in the first embodiment
  • FIG. 3 is a sequence chart of the first embodiment
  • FIG. 4 is a schematic view of a road management system according to a second embodiment
  • FIG. 5 is configuration diagram of a first vehicle and a second vehicle in the second and a third embodiment
  • FIG. 6 is a sequence chart of the second and third embodiments
  • FIG. 7 is a schematic view of a road management system according to the third embodiment.
  • FIG. 8 is a schematic view of a road management system according to a fourth embodiment.
  • FIG. 9 is a configuration diagram of a second vehicle in the fourth embodiment.
  • FIG. 10 is a sequence chart of the fourth embodiment.
  • FIG. 11 is a flowchart of processing performed by a server in the fourth embodiment.
  • a road management system 10 when an avoiding action has been performed by a vehicle 20 (a moving body) moving along a road 70 , an image of an avoidance position 74 is transmitted to a server 50 from that vehicle 20 .
  • a two-wheeled vehicle is assumed as a kind of the vehicle 20 .
  • the kind of the vehicle 20 may be a four-wheeled vehicle, not a two-wheeled vehicle.
  • the server 50 is a computer and includes a server arithmetic section 52 and a server recording section 54 .
  • the server arithmetic section 52 is configured by a processor comprising the likes of a CPU, for example.
  • the server arithmetic section 52 realizes various kinds of functions by executing a program recorded in the server recording section 54 .
  • the server recording section 54 is configured by a RAM, a ROM, or the like.
  • the vehicle 20 includes an imaging section 22 , an avoiding action detecting section 24 , a position detecting section 26 , a network communication section 28 , and an ECU 30 .
  • the imaging section 22 includes at least one camera, and images a road surface in front of the vehicle 20 .
  • the imaging section 22 outputs an image of the road surface to the ECU 30 .
  • the avoiding action detecting section 24 includes the likes of a gyro sensor or steering angle sensor, for example, and detects a turning action and an avoiding action performed by the vehicle 20 on the road 70 .
  • an avoiding action refers to consecutive turning actions to right and left performed by the vehicle 20 within a certain time or within a certain distance.
  • the avoiding action detecting section 24 outputs detected information to the ECU 30 .
  • the position detecting section 26 is a navigation device that includes, for example, map information, a GNSS module, an acceleration sensor, a gyro sensor, and so on, and detects a position of the vehicle 20 using at least one of satellite navigation and inertial navigation.
  • the position detecting section 26 outputs detected information to the ECU 30 .
  • the network communication section 28 transmits to the server 50 via a public line 60 information generated by a vehicle arithmetic section 34 or information recorded in a vehicle recording section 36 , and, moreover, receives information transmitted via the public line 60 from the server 50 , and outputs the received information to the vehicle arithmetic section 34 or the vehicle recording section 36 .
  • the ECU 30 includes an input/output section 32 , the vehicle arithmetic section 34 , and the vehicle recording section 36 .
  • the input/output section 32 is configured by an A/D converter circuit, a communication interface, a driver, and so on.
  • the vehicle arithmetic section 34 is configured by a processor comprising the likes of a CPU, for example.
  • the vehicle arithmetic section 34 realizes various kinds of functions by executing a program recorded in the vehicle recording section 36 .
  • the vehicle arithmetic section 34 functions as an avoidance position determining section 38 , an information processing section 40 , and a control section 42 .
  • the avoidance position determining section 38 determines a position where the avoiding action has occurred (hereafter called the avoidance position 74 ), based on a detection result of the avoiding action detecting section 24 and a detection result of the position detecting section 26 .
  • the information processing section 40 associates image information and position information, and records the associated image information and position information in the vehicle recording section 36 .
  • the control section 42 controls various kinds of instruments.
  • the vehicle recording section 36 is configured by a RAM, a ROM, or the like.
  • the vehicle recording section 36 records various kinds of programs, and various kinds of information used in processing performed by the vehicle arithmetic section 34 .
  • the vehicle recording section 36 records associated with each other the position information and image information from a latest time to a predetermined time before.
  • the position information and the image information may be associated with time information, for example, an imaging time of the image, or a road passing time.
  • FIG. 3 Processing performed by the road management system 10 according to the first embodiment will be described using FIG. 3 .
  • the processing below is performed during running of the vehicle 20 .
  • step S 1 the control section 42 controls the imaging section 22 to image the road surface. Thereafter, the imaging section 22 continues imaging. Moreover, the avoidance position determining section 38 monitors turning action of the vehicle 20 based on the detection result of the avoiding action detecting section 24 . Now, the information processing section 40 associates image information of the road surface imaged by the imaging section 22 and position information detected by the position detecting section 26 , and records the associated image information and position information in the vehicle recording section 36 . Moreover, the information processing section 40 records, in the vehicle recording section 36 , turning position information indicating a position where a turning action has been performed, based on information detected by the avoiding action detecting section 24 and position information detected by the position detecting section 26 .
  • the vehicle 20 when there is an abnormality, for example, a hole 72 (or an obstacle) in the road 70 , the vehicle 20 performs an avoiding action avoiding the hole 72 before reaching the hole 72 .
  • the avoiding action detecting section 24 detects consecutive turning actions to right and left within a certain time or within a certain distance.
  • step S 2 the avoidance position determining section 38 recognizes that an avoiding action has been performed by the vehicle 20 , based on the detection result of the avoiding action detecting section 24 .
  • the avoidance position determining section 38 determines the avoidance position 74 based on the turning position information recorded in the vehicle recording section 36 .
  • the avoidance position determining section 38 may determine a start position 76 of the avoiding action, that is, an initial turning position in a section where the avoiding action has been performed, to be the avoidance position 74 .
  • the avoidance position determining section 38 may determine an end position 78 of the avoiding action, that is, a final turning position in the section where the avoiding action has been performed, to be the avoidance position 74 .
  • the avoidance position determining section 38 may determine an intermediate position 80 of the avoiding action, that is, an intermediate position between the start position 76 and the end position 78 of the avoiding action, to be the avoidance position 74 . Moreover, the avoidance position determining section 38 may determine the whole or part of the section from the start position 76 to the end position 78 of the avoiding action, to be the avoidance position 74 .
  • step S 4 the information processing section 40 extracts an image in the avoidance position 74 .
  • the information processing section 40 extracts image information associated with position information of the avoidance position 74 from the image information recorded in the vehicle recording section 36 up to a predetermined time before.
  • step S 5 the control section 42 controls the network communication section 28 to transmit the image information of the road surface of the avoidance position 74 and the position information associated with that image information.
  • step S 6 the server 50 performs image analysis, based on the image information of the road surface of the avoidance position 74 transmitted from the vehicle 20 , and specifies a kind of the abnormality. Due to this processing, the kind and position of the abnormality are specified.
  • a two-wheeled vehicle is assumed as a kind of the first vehicle 20 a and a kind of the second vehicle 20 b .
  • the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may be a four-wheeled vehicle, not a two-wheeled vehicle.
  • the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may differ.
  • a configuration of the server 50 is the same as the configuration of the server 50 in the first embodiment.
  • Configurations of the first vehicle 20 a and the second vehicle 20 b will be described using FIG. 5 .
  • the configurations of the first vehicle 20 a and the second vehicle 20 b are roughly the same as the configuration of the vehicle 20 in the first embodiment.
  • those configurations of the first vehicle 20 a that are the same as the configurations of the vehicle 20 in the first embodiment will be given the same names as in the first embodiment, and will be assigned with reference symbols adding ‘a’ to the reference symbols of the first embodiment.
  • those configurations of the second vehicle 20 b that are the same as the configurations of the vehicle 20 in the first embodiment will be given the same names as in the first embodiment, and will be assigned with reference symbols adding ‘b’ to the reference symbols of the first embodiment.
  • the vehicle 20 (the first vehicle 20 a and the second vehicle 20 b ) of the second embodiment includes (include) inter-vehicle communication sections 44 a , 44 b .
  • the inter-vehicle communication sections 44 a , 44 b each include a communication device and an antenna for performing inter-vehicle communication.
  • FIG. 6 Processing performed by the road management system 10 according to the second embodiment will be described using FIG. 6 .
  • the processing below is performed during running of the first vehicle 20 a and the second vehicle 20 b.
  • step S 11 an avoidance position determining section 38 a of the first vehicle 20 a monitors turning action of the first vehicle 20 a based on the detection result of an avoiding action detecting section 24 a .
  • an information processing section 40 a records, in a vehicle recording section 36 a , turning position information indicating the position where a turning action has been performed, based on the information detected by the avoiding action detecting section 24 a and the position information detected by a position detecting section 26 a.
  • step S 12 the avoidance position determining section 38 a recognizes that an avoiding action has been performed by the first vehicle 20 a , based on the detection result of the avoiding action detecting section 24 a.
  • step S 13 the avoidance position determining section 38 a determines the avoidance position 74 based on the turning position information recorded in the vehicle recording section 36 a .
  • a method of determining the avoidance position 74 is the same as the determining method in step S 3 of the first embodiment.
  • step S 14 a control section 42 a controls the inter-vehicle communication section 44 a to transmit the position information indicating the avoidance position 74 .
  • a control section 42 b of the second vehicle 20 b compares the position information received by the inter-vehicle communication section 44 b and position information detected by a position detecting section 26 b , and determines whether or not imaging is to be performed by the second vehicle 20 b .
  • the control section 42 b determines imaging is to be performed when the avoidance position 74 is ahead on the road 70 where the second vehicle 20 b is running, in a range within a first predetermined distance from the second vehicle 20 b .
  • the control section 42 b controls an imaging section 22 b to image the road surface. Note that in the case where, due to imaging being performed constantly or every certain period, imaging is already performed by the imaging section 22 b , the control section 42 b extracts an image of the avoidance position 74 .
  • an information processing section 40 b associates the image information of the road surface of the avoidance position 74 imaged by the imaging section 22 b and the position information detected by the position detecting section 26 b.
  • step S 17 the control section 42 b controls a network communication section 28 b to transmit the image information of the road surface of the avoidance position 74 and the position information associated with that image information.
  • step S 18 the server 50 performs image analysis, based on the image information of the road surface of the avoidance position 74 transmitted from the second vehicle 20 b , and specifies a kind of the abnormality. Due to this processing, the kind and position of the abnormality are specified.
  • the road management system 10 when an avoiding action has been performed by the first vehicle 20 a (the first moving body) moving along the road 70 , an image of the avoidance position 74 is transmitted to the server 50 from the second vehicle 20 b (the second moving body) moving in an opposite direction to the first vehicle 20 a and approaching the avoidance position 74 .
  • a two-wheeled vehicle is assumed as the first vehicle 20 a and the second vehicle 20 b .
  • the first vehicle 20 a and the second vehicle 20 b may each be a four-wheeled vehicle, not a two-wheeled vehicle.
  • the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may differ.
  • a configuration of the server 50 is the same as the configuration of the server 50 in the first and second embodiments. Moreover, configurations of the first vehicle 20 a and the second vehicle 20 b are the same as the configurations of the first vehicle 20 a and the second vehicle 20 b in the second embodiment.
  • step S 17 the control section 42 b preferably controls the network communication section 28 b to add, to the image information and the position information, information indicating that the second vehicle 20 b is moving in an opposite direction to a moving direction of the first vehicle 20 a .
  • This information means that the avoidance position 74 is present in a roadway on an opposite side to a roadway where the second vehicle 20 b is running.
  • the road management system 10 when an avoiding action has been performed by at least one first vehicle 20 a (first moving body) moving along the road 70 , an image of the avoidance position 74 is transmitted to the server 50 from the second vehicle 20 b (the second moving body) approaching the avoidance position 74 .
  • a two-wheeled vehicle is assumed as the first vehicle 20 a and the second vehicle 20 b .
  • the first vehicle 20 a and the second vehicle 20 b may each be a four-wheeled vehicle, not a two-wheeled vehicle.
  • the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may differ.
  • the server 50 includes a timer 56 and a server transmitting section 58 .
  • the server transmitting section 58 transmits information over a broad region.
  • the second vehicle 20 b includes a broadcast receiving section 46 b that receives the information transmitted from the server transmitting section 58 .
  • FIG. 10 Processing performed by the road management system 10 according to the fourth embodiment will be described using FIG. 10 .
  • the processing below is performed during running of the first vehicle 20 a and the second vehicle 20 b.
  • Processing of steps S 21 -S 23 is the same as processing of steps S 11 -S 13 of the second embodiment.
  • step S 24 the control section 42 a of the first vehicle 20 a controls a network communication section 28 a to transmit position information indicating the avoidance position 74 .
  • step S 25 the server arithmetic section 52 determines whether imaging is required or not. A determining method will be described below in [4.2].
  • step S 26 the server transmitting section 58 transmits the position information indicating the avoidance position 74 .
  • step S 27 the control section 42 b of the second vehicle 20 b compares the position information received by the broadcast receiving section 46 b and the position information detected by the position detecting section 26 b , and determines whether or not imaging is to be performed by the second vehicle 20 b .
  • the control section 42 b determines imaging is to be performed when the avoidance position 74 is ahead on the road 70 where the second vehicle 20 b is running, in a range within a first predetermined distance from the second vehicle 20 b .
  • the control section 42 b controls the imaging section 22 b to image the road surface. Note that in the case where imaging is performed constantly or every certain period by the imaging section 22 b , the control section 42 b extracts an image of the avoidance position 74 .
  • Processing of steps S 28 -S 30 is the same as processing of steps S 16 -S 18 of the second embodiment.
  • step S 25 of FIG. 10 The method of determining whether imaging is required or not, performed in step S 25 of FIG. 10 , will be described using FIG. 11 . Processing described below is performed each time the server 50 receives position information from the first vehicle 20 a .
  • the avoidance position 74 and the number of times of occurrence of the avoiding action are recorded by the server recording section 54 , and elapsed time with an initial avoiding action as a starting point is clocked by the timer 56 .
  • step S 31 the server arithmetic section 52 determines whether or not position information of the avoidance position 74 is recorded in the server recording section 54 . If the position information is present (step S 31 : YES), then processing shifts to step S 32 . On the other hand, if there is no position information (step S 31 : NO), then processing shifts to step S 36 .
  • step S 32 the server arithmetic section 52 determines whether or not the elapsed time clocked by the timer 56 is within a predetermined time. If the elapsed time is within the predetermined time (step S 32 : YES), then processing shifts to step S 33 . On the other hand, if the elapsed time exceeds the predetermined time (step S 32 : NO), then processing shifts to step S 38 .
  • step S 33 the server arithmetic section 52 adds 1 to the number of times of occurrence recorded in the server recording section 54 . Processing shifts to step S 34 .
  • step S 34 the server arithmetic section 52 determines whether or not the number of times of occurrence exceeds a predetermined number of times. If the number of times of occurrence exceeds the predetermined number of times (step S 34 : YES), then processing shifts to step S 35 . Then, in step S 35 , the server arithmetic section 52 determines that imaging is required. On the other hand, if the number of times of occurrence is within the predetermined number of times (step S 34 : NO), then processing once ends.
  • the server arithmetic section 52 associates the position information of the avoidance position 74 and an initial value of the number of times of occurrence, and records the associated position information and initial value in the server recording section 54 . Then, in step S 37 , the server arithmetic section 52 activates the timer 56 to start clocking of elapsed time.
  • the server arithmetic section 52 resets the timer 56 , and sets, to an initial value, the number of times of occurrence recorded in the server recording section 54 .
  • a configuration is adopted whereby the server 50 transmits position information over a broad region using the server transmitting section 58 .
  • a configuration may be adopted whereby the server 50 transmits position information to a roadside device in a periphery of the avoidance position 74 , and the position information is transmitted to the second vehicle 20 b from the roadside device.
  • An aspect of the present invention comprises:
  • an avoiding action detecting section 24 , 24 a that detects an avoiding action of a moving body (a vehicle 20 , a first vehicle 20 a ) moving along a road 70 ;
  • a position detecting section 26 , 26 a that detects a position of the moving body (the vehicle 20 , the first vehicle 20 a );
  • an avoidance position determining section 38 , 38 a that determines an avoidance position 74 where the avoiding action has occurred, based on a detection result of the avoiding action detecting section 24 , 24 a and a detection result of the position detecting section 26 , 26 a ;
  • the road surface of the avoidance position 74 where a possibility of there being an abnormality is high is selectively imaged, hence a load of arithmetic processing to perform analysis of the road surface can be suppressed.
  • An aspect of the present invention may further comprise
  • a transmitting section (an inter-vehicle communication section 44 a ) that transmits position information of the avoidance position 74 to a device (a second vehicle 20 b ) on an outside of the moving body (the first vehicle 20 a ), wherein
  • the imaging section 22 b may be provided in the device (the second vehicle 20 b ).
  • the above-described configuration results in that, after the avoiding action has been performed by the moving body (the first vehicle 20 a ), the road surface of the avoidance position 74 can be imaged by the device (the second vehicle 20 b ) other than the moving body (the first vehicle 20 a ).
  • the moving body may be a first moving body (the first vehicle 20 a ), and
  • the device may be a second moving body (the second vehicle 20 b ) approaching the avoidance position 74 .
  • the second vehicle 20 b the second moving body approaching the avoidance position 74 .
  • the above-described configuration results in that, after the avoiding action has been performed by the first moving body (the first vehicle 20 a ), the road surface of the avoidance position 74 can be imaged by the second moving body (the second vehicle 20 b ) approaching the avoidance position 74 .
  • the device may be the second moving body (the second vehicle 20 b ) moving in the same direction as the first moving body (the first vehicle 20 a ) and approaching the avoidance position 74 .
  • the above-described configuration results in that, after the avoiding action has been performed by the first moving body (the first vehicle 20 a ), the road surface of the avoidance position 74 can be imaged by the second moving body (the second vehicle 20 b ) following the first moving body (the first vehicle 20 a ).
  • the device may be the second moving body (the second vehicle 20 b ) moving in an opposite direction to the first moving body (the first vehicle 20 a ) and approaching the avoidance position 74 .
  • the above-described configuration results in that, after the avoiding action has been performed by the first moving body (the first vehicle 20 a ), the road surface of the avoidance position 74 can be imaged by the second moving body (the second vehicle 20 b ) approaching from in front of the first moving body (the first vehicle 20 a ).
  • the moving body (the vehicle 20 , the first vehicle 20 a , the second vehicle 20 b ) may be a two-wheeled vehicle.
  • the moving body (the vehicle 20 , the first vehicle 20 a , the second vehicle 20 b ) is a two-wheeled vehicle as in the above-described configuration, it is easy for the avoiding action to be detected. (All Embodiments)
  • a kind of the first moving body (the first vehicle 20 a ) may be identical to a kind of the second moving body (the second moving vehicle 20 b ).
  • An aspect of the present invention may further comprise:
  • an information processing section 40 that associates image information of the road surface of the avoidance position 74 imaged by the imaging section 22 and position information of the avoidance position 74 ;
  • a transmitting section (a network communication section 28 ) that transmits, to a server 50 on an outside, the image information and the position information that have been associated, wherein
  • the avoiding action detecting section 24 the position detecting section 26 , the avoidance position determining section 38 , the imaging section 22 , the information processing section 40 , and the transmitting section (the network communication section 28 ) may be provided in the moving body (the vehicle 20 ).
  • the transmitting section the network communication section 28
  • the imaging section 22 b images the road surface of the avoidance position 74 .
  • the above-described configuration results in that, when the avoiding action by the plurality of moving bodies (first vehicles 20 a ) has occurred, the road surface of the avoidance position 74 is imaged, hence a road surface where there is an even higher possibility of there being an abnormality can be imaged, and wasteful imaging and image analysis can be suppressed.

Abstract

A road management system includes: an avoiding action detecting section that detects an avoiding action of a moving body (a vehicle, a first vehicle) moving along a road; a position detecting section that detects a position of the moving body (the vehicle, the first vehicle); an avoidance position determining section that determines an avoidance position where the avoiding action has occurred, based on a detection result of the avoiding action detecting section and a detection result of the position detecting section; and an imaging section that images a road surface of the avoidance position.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application is based upon and claims the benefit of priority from Japanese Patent Application No. 2019-032474 filed on Feb. 26, 2019, the contents of which are incorporated herein by reference.
BACKGROUND OF THE INVENTION Field of the Invention
The present invention relates to a road management system that acquires an image of a road surface.
Description of the Related Art
Japanese Patent No. 5853719 discloses: an image processing system that detects a state (a dry state or a frozen state) of a road surface and utilizes the detected state for such objects as calling attention of a driver to the detected state; and a vehicle comprising the image processing system.
SUMMARY OF THE INVENTION
In the case of using the system of Japanese Patent No. 5853719 with an object of analyzing an abnormality unexpectedly present in the road surface, it is required that analysis of the road surface is performed continuously, regardless of presence/absence of an abnormality. Therefore, a load of arithmetic processing increases.
The present invention, which was made considering such a problem, has an object of providing a road management system that can suppress a load of arithmetic processing when performing analysis of a road surface.
An aspect of the present invention comprises:
an avoiding action detecting section configured to detect an avoiding action of a moving body moving along a road;
a position detecting section configured to detect a position of the moving body;
an avoidance position determining section configured to determine an avoidance position where the avoiding action has occurred, based on a detection result of the avoiding action detecting section and a detection result of the position detecting section; and
an imaging section configured to image a road surface of the avoidance position.
According to the present invention, the load of arithmetic processing when performing analysis of the road surface can be suppressed.
The above and other objects, features, and advantages of the present invention will become more apparent from the following description when taken in conjunction with the accompanying drawings, in which preferred embodiments of the present invention are shown by way of illustrative example.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a schematic view of a road management system according to a first embodiment;
FIG. 2 is a configuration diagram of a vehicle in the first embodiment;
FIG. 3 is a sequence chart of the first embodiment;
FIG. 4 is a schematic view of a road management system according to a second embodiment;
FIG. 5 is configuration diagram of a first vehicle and a second vehicle in the second and a third embodiment;
FIG. 6 is a sequence chart of the second and third embodiments;
FIG. 7 is a schematic view of a road management system according to the third embodiment;
FIG. 8 is a schematic view of a road management system according to a fourth embodiment;
FIG. 9 is a configuration diagram of a second vehicle in the fourth embodiment;
FIG. 10 is a sequence chart of the fourth embodiment; and
FIG. 11 is a flowchart of processing performed by a server in the fourth embodiment.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
Preferred embodiments of a road management system according to the present invention will be presented and described in detail below with reference to the accompanying drawings.
1. First Embodiment
As shown in FIG. 1, in a road management system 10 according to a first embodiment, when an avoiding action has been performed by a vehicle 20 (a moving body) moving along a road 70, an image of an avoidance position 74 is transmitted to a server 50 from that vehicle 20. Here, a two-wheeled vehicle is assumed as a kind of the vehicle 20. However, the kind of the vehicle 20 may be a four-wheeled vehicle, not a two-wheeled vehicle.
The server 50 is a computer and includes a server arithmetic section 52 and a server recording section 54. The server arithmetic section 52 is configured by a processor comprising the likes of a CPU, for example. The server arithmetic section 52 realizes various kinds of functions by executing a program recorded in the server recording section 54. The server recording section 54 is configured by a RAM, a ROM, or the like.
[1.1 Configuration of Vehicle 20]
A configuration of the vehicle 20 will be described using FIG. 2. The vehicle 20 includes an imaging section 22, an avoiding action detecting section 24, a position detecting section 26, a network communication section 28, and an ECU 30.
The imaging section 22 includes at least one camera, and images a road surface in front of the vehicle 20. The imaging section 22 outputs an image of the road surface to the ECU 30. The avoiding action detecting section 24 includes the likes of a gyro sensor or steering angle sensor, for example, and detects a turning action and an avoiding action performed by the vehicle 20 on the road 70. In the present specification, an avoiding action refers to consecutive turning actions to right and left performed by the vehicle 20 within a certain time or within a certain distance. The avoiding action detecting section 24 outputs detected information to the ECU 30. The position detecting section 26 is a navigation device that includes, for example, map information, a GNSS module, an acceleration sensor, a gyro sensor, and so on, and detects a position of the vehicle 20 using at least one of satellite navigation and inertial navigation. The position detecting section 26 outputs detected information to the ECU 30.
The network communication section 28 transmits to the server 50 via a public line 60 information generated by a vehicle arithmetic section 34 or information recorded in a vehicle recording section 36, and, moreover, receives information transmitted via the public line 60 from the server 50, and outputs the received information to the vehicle arithmetic section 34 or the vehicle recording section 36.
The ECU 30 includes an input/output section 32, the vehicle arithmetic section 34, and the vehicle recording section 36. The input/output section 32 is configured by an A/D converter circuit, a communication interface, a driver, and so on. The vehicle arithmetic section 34 is configured by a processor comprising the likes of a CPU, for example. The vehicle arithmetic section 34 realizes various kinds of functions by executing a program recorded in the vehicle recording section 36. Here, the vehicle arithmetic section 34 functions as an avoidance position determining section 38, an information processing section 40, and a control section 42. The avoidance position determining section 38 determines a position where the avoiding action has occurred (hereafter called the avoidance position 74), based on a detection result of the avoiding action detecting section 24 and a detection result of the position detecting section 26. The information processing section 40 associates image information and position information, and records the associated image information and position information in the vehicle recording section 36. The control section 42 controls various kinds of instruments. The vehicle recording section 36 is configured by a RAM, a ROM, or the like. The vehicle recording section 36 records various kinds of programs, and various kinds of information used in processing performed by the vehicle arithmetic section 34. Moreover, the vehicle recording section 36 records associated with each other the position information and image information from a latest time to a predetermined time before. The position information and the image information may be associated with time information, for example, an imaging time of the image, or a road passing time.
[1.2 Processing]
Processing performed by the road management system 10 according to the first embodiment will be described using FIG. 3. The processing below is performed during running of the vehicle 20.
In step S1, the control section 42 controls the imaging section 22 to image the road surface. Thereafter, the imaging section 22 continues imaging. Moreover, the avoidance position determining section 38 monitors turning action of the vehicle 20 based on the detection result of the avoiding action detecting section 24. Now, the information processing section 40 associates image information of the road surface imaged by the imaging section 22 and position information detected by the position detecting section 26, and records the associated image information and position information in the vehicle recording section 36. Moreover, the information processing section 40 records, in the vehicle recording section 36, turning position information indicating a position where a turning action has been performed, based on information detected by the avoiding action detecting section 24 and position information detected by the position detecting section 26.
As shown in FIG. 1, when there is an abnormality, for example, a hole 72 (or an obstacle) in the road 70, the vehicle 20 performs an avoiding action avoiding the hole 72 before reaching the hole 72. At this time, the avoiding action detecting section 24 detects consecutive turning actions to right and left within a certain time or within a certain distance.
In step S2, the avoidance position determining section 38 recognizes that an avoiding action has been performed by the vehicle 20, based on the detection result of the avoiding action detecting section 24.
In step S3, the avoidance position determining section 38 determines the avoidance position 74 based on the turning position information recorded in the vehicle recording section 36. For example, the avoidance position determining section 38 may determine a start position 76 of the avoiding action, that is, an initial turning position in a section where the avoiding action has been performed, to be the avoidance position 74. Moreover, the avoidance position determining section 38 may determine an end position 78 of the avoiding action, that is, a final turning position in the section where the avoiding action has been performed, to be the avoidance position 74. Moreover, the avoidance position determining section 38 may determine an intermediate position 80 of the avoiding action, that is, an intermediate position between the start position 76 and the end position 78 of the avoiding action, to be the avoidance position 74. Moreover, the avoidance position determining section 38 may determine the whole or part of the section from the start position 76 to the end position 78 of the avoiding action, to be the avoidance position 74.
In step S4, the information processing section 40 extracts an image in the avoidance position 74. Now, the information processing section 40 extracts image information associated with position information of the avoidance position 74 from the image information recorded in the vehicle recording section 36 up to a predetermined time before.
In step S5, the control section 42 controls the network communication section 28 to transmit the image information of the road surface of the avoidance position 74 and the position information associated with that image information.
In step S6, the server 50 performs image analysis, based on the image information of the road surface of the avoidance position 74 transmitted from the vehicle 20, and specifies a kind of the abnormality. Due to this processing, the kind and position of the abnormality are specified.
2. Second Embodiment
As shown in FIG. 4, in the road management system 10 according to a second embodiment, when an avoiding action has been performed by a first vehicle 20 a (a first moving body) moving along the road 70, an image of the avoidance position 74 is transmitted to the server 50 from a second vehicle 20 b (a second moving body) moving in the same direction as the first vehicle 20 a and approaching the avoidance position 74. Here, a two-wheeled vehicle is assumed as a kind of the first vehicle 20 a and a kind of the second vehicle 20 b. However, the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may be a four-wheeled vehicle, not a two-wheeled vehicle. Moreover, the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may differ. A configuration of the server 50 is the same as the configuration of the server 50 in the first embodiment.
[2.1 Configurations of First Vehicle 20 a and Second Vehicle 20 b]
Configurations of the first vehicle 20 a and the second vehicle 20 b will be described using FIG. 5. The configurations of the first vehicle 20 a and the second vehicle 20 b are roughly the same as the configuration of the vehicle 20 in the first embodiment. In the description below, those configurations of the first vehicle 20 a that are the same as the configurations of the vehicle 20 in the first embodiment, will be given the same names as in the first embodiment, and will be assigned with reference symbols adding ‘a’ to the reference symbols of the first embodiment. Similarly, those configurations of the second vehicle 20 b that are the same as the configurations of the vehicle 20 in the first embodiment, will be given the same names as in the first embodiment, and will be assigned with reference symbols adding ‘b’ to the reference symbols of the first embodiment.
Unlike in the first embodiment, the vehicle 20 (the first vehicle 20 a and the second vehicle 20 b) of the second embodiment includes (include) inter-vehicle communication sections 44 a, 44 b. The inter-vehicle communication sections 44 a, 44 b each include a communication device and an antenna for performing inter-vehicle communication.
[2.2 Processing]
Processing performed by the road management system 10 according to the second embodiment will be described using FIG. 6. The processing below is performed during running of the first vehicle 20 a and the second vehicle 20 b.
In step S11, an avoidance position determining section 38 a of the first vehicle 20 a monitors turning action of the first vehicle 20 a based on the detection result of an avoiding action detecting section 24 a. Now, an information processing section 40 a records, in a vehicle recording section 36 a, turning position information indicating the position where a turning action has been performed, based on the information detected by the avoiding action detecting section 24 a and the position information detected by a position detecting section 26 a.
In step S12, the avoidance position determining section 38 a recognizes that an avoiding action has been performed by the first vehicle 20 a, based on the detection result of the avoiding action detecting section 24 a.
In step S13, the avoidance position determining section 38 a determines the avoidance position 74 based on the turning position information recorded in the vehicle recording section 36 a. A method of determining the avoidance position 74 is the same as the determining method in step S3 of the first embodiment.
In step S14, a control section 42 a controls the inter-vehicle communication section 44 a to transmit the position information indicating the avoidance position 74.
In step S15, a control section 42 b of the second vehicle 20 b compares the position information received by the inter-vehicle communication section 44 b and position information detected by a position detecting section 26 b, and determines whether or not imaging is to be performed by the second vehicle 20 b. For example, the control section 42 b determines imaging is to be performed when the avoidance position 74 is ahead on the road 70 where the second vehicle 20 b is running, in a range within a first predetermined distance from the second vehicle 20 b. Then, at a time point when the second vehicle 20 b has reached the avoidance position 74 or a position a second predetermined distance before the avoidance position 74, the control section 42 b controls an imaging section 22 b to image the road surface. Note that in the case where, due to imaging being performed constantly or every certain period, imaging is already performed by the imaging section 22 b, the control section 42 b extracts an image of the avoidance position 74.
In step S16, an information processing section 40 b associates the image information of the road surface of the avoidance position 74 imaged by the imaging section 22 b and the position information detected by the position detecting section 26 b.
In step S17, the control section 42 b controls a network communication section 28 b to transmit the image information of the road surface of the avoidance position 74 and the position information associated with that image information.
In step S18, the server 50 performs image analysis, based on the image information of the road surface of the avoidance position 74 transmitted from the second vehicle 20 b, and specifies a kind of the abnormality. Due to this processing, the kind and position of the abnormality are specified.
3. Third Embodiment
As shown in FIG. 7, in the road management system 10 according to a third embodiment, when an avoiding action has been performed by the first vehicle 20 a (the first moving body) moving along the road 70, an image of the avoidance position 74 is transmitted to the server 50 from the second vehicle 20 b (the second moving body) moving in an opposite direction to the first vehicle 20 a and approaching the avoidance position 74. Here, a two-wheeled vehicle is assumed as the first vehicle 20 a and the second vehicle 20 b. However, the first vehicle 20 a and the second vehicle 20 b may each be a four-wheeled vehicle, not a two-wheeled vehicle. Moreover, the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may differ. A configuration of the server 50 is the same as the configuration of the server 50 in the first and second embodiments. Moreover, configurations of the first vehicle 20 a and the second vehicle 20 b are the same as the configurations of the first vehicle 20 a and the second vehicle 20 b in the second embodiment.
Processing performed by the road management system 10 according to the third embodiment is the same as in the second embodiment. However, in step S17, the control section 42 b preferably controls the network communication section 28 b to add, to the image information and the position information, information indicating that the second vehicle 20 b is moving in an opposite direction to a moving direction of the first vehicle 20 a. This information means that the avoidance position 74 is present in a roadway on an opposite side to a roadway where the second vehicle 20 b is running.
4. Fourth Embodiment
As shown in FIG. 8, in the road management system 10 according to a fourth embodiment, when an avoiding action has been performed by at least one first vehicle 20 a (first moving body) moving along the road 70, an image of the avoidance position 74 is transmitted to the server 50 from the second vehicle 20 b (the second moving body) approaching the avoidance position 74. Here, a two-wheeled vehicle is assumed as the first vehicle 20 a and the second vehicle 20 b. However, the first vehicle 20 a and the second vehicle 20 b may each be a four-wheeled vehicle, not a two-wheeled vehicle. Moreover, the kind of the first vehicle 20 a and the kind of the second vehicle 20 b may differ.
In the fourth embodiment, the server 50 includes a timer 56 and a server transmitting section 58. The server transmitting section 58 transmits information over a broad region. Moreover, as shown in FIG. 9, the second vehicle 20 b includes a broadcast receiving section 46 b that receives the information transmitted from the server transmitting section 58.
[4.1 Processing]
Processing performed by the road management system 10 according to the fourth embodiment will be described using FIG. 10. The processing below is performed during running of the first vehicle 20 a and the second vehicle 20 b.
Processing of steps S21-S23 is the same as processing of steps S11-S13 of the second embodiment.
In step S24, the control section 42 a of the first vehicle 20 a controls a network communication section 28 a to transmit position information indicating the avoidance position 74.
In step S25, the server arithmetic section 52 determines whether imaging is required or not. A determining method will be described below in [4.2].
If it is determined in step S25 that imaging is required, then in step S26, the server transmitting section 58 transmits the position information indicating the avoidance position 74.
In step S27, the control section 42 b of the second vehicle 20 b compares the position information received by the broadcast receiving section 46 b and the position information detected by the position detecting section 26 b, and determines whether or not imaging is to be performed by the second vehicle 20 b. For example, the control section 42 b determines imaging is to be performed when the avoidance position 74 is ahead on the road 70 where the second vehicle 20 b is running, in a range within a first predetermined distance from the second vehicle 20 b. Then, at a time point when the second vehicle 20 b has reached the avoidance position 74 or a position a second predetermined distance before the avoidance position 74, the control section 42 b controls the imaging section 22 b to image the road surface. Note that in the case where imaging is performed constantly or every certain period by the imaging section 22 b, the control section 42 b extracts an image of the avoidance position 74.
Processing of steps S28-S30 is the same as processing of steps S16-S18 of the second embodiment.
[4.2 Method of Determining Whether Imaging is Required or Not]
The method of determining whether imaging is required or not, performed in step S25 of FIG. 10, will be described using FIG. 11. Processing described below is performed each time the server 50 receives position information from the first vehicle 20 a. Here, the avoidance position 74 and the number of times of occurrence of the avoiding action are recorded by the server recording section 54, and elapsed time with an initial avoiding action as a starting point is clocked by the timer 56.
In step S31, the server arithmetic section 52 determines whether or not position information of the avoidance position 74 is recorded in the server recording section 54. If the position information is present (step S31: YES), then processing shifts to step S32. On the other hand, if there is no position information (step S31: NO), then processing shifts to step S36.
In step S32, the server arithmetic section 52 determines whether or not the elapsed time clocked by the timer 56 is within a predetermined time. If the elapsed time is within the predetermined time (step S32: YES), then processing shifts to step S33. On the other hand, if the elapsed time exceeds the predetermined time (step S32: NO), then processing shifts to step S38.
In step S33, the server arithmetic section 52 adds 1 to the number of times of occurrence recorded in the server recording section 54. Processing shifts to step S34.
In step S34, the server arithmetic section 52 determines whether or not the number of times of occurrence exceeds a predetermined number of times. If the number of times of occurrence exceeds the predetermined number of times (step S34: YES), then processing shifts to step S35. Then, in step S35, the server arithmetic section 52 determines that imaging is required. On the other hand, if the number of times of occurrence is within the predetermined number of times (step S34: NO), then processing once ends.
In the case of processing shifting from step S31 to step S36, the server arithmetic section 52 associates the position information of the avoidance position 74 and an initial value of the number of times of occurrence, and records the associated position information and initial value in the server recording section 54. Then, in step S37, the server arithmetic section 52 activates the timer 56 to start clocking of elapsed time.
In the case of processing shifting from step S32 to step S38, the server arithmetic section 52 resets the timer 56, and sets, to an initial value, the number of times of occurrence recorded in the server recording section 54.
[4.3 Modified Example]
In the fourth embodiment, a configuration is adopted whereby the server 50 transmits position information over a broad region using the server transmitting section 58. Instead, a configuration may be adopted whereby the server 50 transmits position information to a roadside device in a periphery of the avoidance position 74, and the position information is transmitted to the second vehicle 20 b from the roadside device.
5. Technical Ideas Obtained from Embodiments
Technical ideas understandable from the above-described embodiments and modified example will be described below.
An aspect of the present invention comprises:
an avoiding action detecting section 24, 24 a that detects an avoiding action of a moving body (a vehicle 20, a first vehicle 20 a) moving along a road 70;
a position detecting section 26, 26 a that detects a position of the moving body (the vehicle 20, the first vehicle 20 a);
an avoidance position determining section 38, 38 a that determines an avoidance position 74 where the avoiding action has occurred, based on a detection result of the avoiding action detecting section 24, 24 a and a detection result of the position detecting section 26, 26 a; and
an imaging section 22, 22 b that images a road surface of the avoidance position 74. (All Embodiments)
Due to the above-described configuration, the road surface of the avoidance position 74 where a possibility of there being an abnormality is high is selectively imaged, hence a load of arithmetic processing to perform analysis of the road surface can be suppressed.
An aspect of the present invention may further comprise
a transmitting section (an inter-vehicle communication section 44 a) that transmits position information of the avoidance position 74 to a device (a second vehicle 20 b) on an outside of the moving body (the first vehicle 20 a), wherein
the imaging section 22 b may be provided in the device (the second vehicle 20 b). (Second Embodiment, Third Embodiment)
The above-described configuration results in that, after the avoiding action has been performed by the moving body (the first vehicle 20 a), the road surface of the avoidance position 74 can be imaged by the device (the second vehicle 20 b) other than the moving body (the first vehicle 20 a).
In an aspect of the present invention,
the moving body may be a first moving body (the first vehicle 20 a), and
the device may be a second moving body (the second vehicle 20 b) approaching the avoidance position 74. (Second Embodiment, Third Embodiment)
The above-described configuration results in that, after the avoiding action has been performed by the first moving body (the first vehicle 20 a), the road surface of the avoidance position 74 can be imaged by the second moving body (the second vehicle 20 b) approaching the avoidance position 74.
In an aspect of the present invention,
the device may be the second moving body (the second vehicle 20 b) moving in the same direction as the first moving body (the first vehicle 20 a) and approaching the avoidance position 74. (Second Embodiment)
The above-described configuration results in that, after the avoiding action has been performed by the first moving body (the first vehicle 20 a), the road surface of the avoidance position 74 can be imaged by the second moving body (the second vehicle 20 b) following the first moving body (the first vehicle 20 a).
In an aspect of the present invention,
the device may be the second moving body (the second vehicle 20 b) moving in an opposite direction to the first moving body (the first vehicle 20 a) and approaching the avoidance position 74. (Third Embodiment)
The above-described configuration results in that, after the avoiding action has been performed by the first moving body (the first vehicle 20 a), the road surface of the avoidance position 74 can be imaged by the second moving body (the second vehicle 20 b) approaching from in front of the first moving body (the first vehicle 20 a).
In an aspect of the present invention,
the moving body (the vehicle 20, the first vehicle 20 a, the second vehicle 20 b) may be a two-wheeled vehicle.
When the moving body (the vehicle 20, the first vehicle 20 a, the second vehicle 20 b) is a two-wheeled vehicle as in the above-described configuration, it is easy for the avoiding action to be detected. (All Embodiments)
In an aspect of the present invention,
a kind of the first moving body (the first vehicle 20 a) may be identical to a kind of the second moving body (the second moving vehicle 20 b). (Second Embodiment, Third Embodiment)
A possibility of the second moving body (the second vehicle 20 b) performing the same avoiding action as the first moving body (the first vehicle 20 a) is high. Therefore, the above-described configuration enables the image of the road surface of the avoidance position 74 to be utilized by the second moving body (the second vehicle 20 b) too.
An aspect of the present invention may further comprise:
an information processing section 40 that associates image information of the road surface of the avoidance position 74 imaged by the imaging section 22 and position information of the avoidance position 74; and
a transmitting section (a network communication section 28) that transmits, to a server 50 on an outside, the image information and the position information that have been associated, wherein
the avoiding action detecting section 24, the position detecting section 26, the avoidance position determining section 38, the imaging section 22, the information processing section 40, and the transmitting section (the network communication section 28) may be provided in the moving body (the vehicle 20). (First Embodiment)
In an aspect of the present invention,
a configuration may be adopted whereby
in a case that, within a predetermined time, the avoiding action by a plurality of the moving bodies (the first vehicles 20 a) has occurred in the same avoidance position 74,
the imaging section 22 b images the road surface of the avoidance position 74. (Fourth Embodiment)
The above-described configuration results in that, when the avoiding action by the plurality of moving bodies (first vehicles 20 a) has occurred, the road surface of the avoidance position 74 is imaged, hence a road surface where there is an even higher possibility of there being an abnormality can be imaged, and wasteful imaging and image analysis can be suppressed.
Note that the road management system according to the present invention is not limited to the above-mentioned embodiments, and that, of course, a variety of configurations may be adopted without departing from the spirit of the present invention.

Claims (8)

What is claimed is:
1. A road management system comprising a position determining device mounted on a first moving body moving along a road and an imaging device mounted on a second moving body moving along the road, the first moving body and the second moving body being provided in plural, wherein
the position determining device comprises a first processor that executes first computer-readable instructions stored in a first memory,
the first processor executes the first computer-readable instructions to cause the position determining device to:
determine an avoidance position where an avoiding action has occurred, based on avoidance information indicating that the first moving body has avoided an abnormality on the road and position information indicating an avoidance position of the first moving body; and
externally transmit avoidance position information indicating the avoidance position through a transmitting circuit, and
the imaging device comprises a second processor that executes second computer-readable instructions stored in a second memory, and
the second processor executes the second computer-readable instructions to cause the imaging device to image with a camera a road surface at the avoidance position indicated by the avoidance position information received through a receiving circuit.
2. The road management system according to claim 1, wherein
the second moving body moves in a same direction as the first moving body and approaches the avoidance position.
3. The road management system according to claim 1, wherein
the second moving body moves in an opposite direction to the first moving body and approaches the avoidance position.
4. The road management system according to claim 1, wherein at least one of the first moving body and the second moving body is a two-wheeled vehicle.
5. The road management system according to claim 1, wherein the second processor causes the imaging device to:
associate image information of the road surface of the avoidance position imaged by the camera and the position information of the avoidance position; and
transmit, through a transmitting circuit to a server on an outside, the image information and the position information that have been associated.
6. The road management system according to claim 1, wherein the transmitting circuit of the first moving body and the receiving circuit of the second moving body are configured to perform inter-vehicle communication, and the transmitting circuit of the first moving body and the receiving circuit of the second moving body perform the inter-vehicle communication to transmit and receive the avoidance position information.
7. The road management system according to claim 1, wherein the transmitting circuit of the first moving body and the receiving circuit of the second moving body are configured to perform communication with a server, the transmitting circuit of the first moving body transmits the avoidance position information to the server, and the receiving circuit of the second moving body receives the avoidance position information from the server.
8. The road management system according to claim 7, wherein in a case that, within a predetermined time, the avoiding action has been performed by more than one said first moving body in a same avoidance position, the server transmits the avoidance position information to the receiving circuit of the second moving body, and the second processor executes the second computer-readable instructions to cause the imaging device to image with the camera the road surface at the avoidance position indicated by the avoidance position information received through the receiving circuit.
US16/800,685 2019-02-26 2020-02-25 Road management system Active US11252380B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2019-032474 2019-02-26
JPJP2019-032474 2019-02-26
JP2019032474A JP7359551B2 (en) 2019-02-26 2019-02-26 road management system

Publications (2)

Publication Number Publication Date
US20200275061A1 US20200275061A1 (en) 2020-08-27
US11252380B2 true US11252380B2 (en) 2022-02-15

Family

ID=72141912

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/800,685 Active US11252380B2 (en) 2019-02-26 2020-02-25 Road management system

Country Status (3)

Country Link
US (1) US11252380B2 (en)
JP (1) JP7359551B2 (en)
CN (1) CN111613077A (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022201860A1 (en) * 2021-03-24 2022-09-29 本田技研工業株式会社 Information processing device and system

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100235035A1 (en) * 2006-03-22 2010-09-16 Nissan Motor Co., Ltd. Avoidance maneuver calculation device, avoidance control device, vehicle having each device, avoidance maneuver calculating method, and avoidance controlling method
US20130151058A1 (en) * 2011-12-09 2013-06-13 GM Global Technology Operations LLC Method and system for controlling a host vehicle
US20140229069A1 (en) * 2011-09-26 2014-08-14 Toyota Jidosha Kabushiki Kaisha Driving support system for a vehicle
JP5853719B2 (en) 2012-01-20 2016-02-09 株式会社リコー Image processing system and vehicle equipped with the same
US20160217333A1 (en) * 2015-01-26 2016-07-28 Ricoh Company, Ltd. Information processing apparatus and information processing system
US20170168503A1 (en) * 2015-12-11 2017-06-15 Avishtech, Llc Autonomous vehicle towing system and method
US20170291608A1 (en) * 2016-04-11 2017-10-12 Volkswagen Ag Method for the autonomous or partly autonomous execution of a cooperative driving maneuver
US20180032082A1 (en) * 2016-01-05 2018-02-01 Mobileye Vision Technologies Ltd. Machine learning navigational engine with imposed constraints
US20180107946A1 (en) * 2015-06-22 2018-04-19 Denso Corporation Alert output apparatus
US20180165966A1 (en) * 2016-12-13 2018-06-14 Toyota Jidosha Kabushiki Kaisha Obstacle determination system and non-transitory computer-readable recording medium storing obstacle determination program
JP2018171959A (en) * 2017-03-31 2018-11-08 株式会社Subaru Travel support device for vehicle
US20190256088A1 (en) * 2018-11-02 2019-08-22 Intel Corporation Methods and apparatus to generate vehicle warnings
US20200166360A1 (en) * 2017-06-09 2020-05-28 Bosch Engineering K.K. Travel obstruction detection device and vehicle navigation system

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4367174B2 (en) * 2004-02-25 2009-11-18 株式会社デンソー On-vehicle transmitter and obstacle detection system
JP4845783B2 (en) * 2007-03-16 2011-12-28 パイオニア株式会社 Information processing method, in-vehicle device, and information distribution device
WO2011013238A1 (en) * 2009-07-31 2011-02-03 パイオニア株式会社 Driving support device, driving support system, driving support method, and program
JP5585405B2 (en) * 2010-11-15 2014-09-10 株式会社デンソー Traffic information system, traffic information acquisition device, and traffic information provision device
CN103176185B (en) * 2011-12-26 2015-01-21 上海汽车集团股份有限公司 Method and system for detecting road barrier
CN107290797B (en) * 2017-06-23 2019-01-18 西北工业大学 A kind of obstacle detection system and method based on quorum-sensing system
CN109035796A (en) * 2018-08-27 2018-12-18 山东飞天激光光电科技有限公司 A kind of laser radar Overheight Vehicles linear detector and detection method

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100235035A1 (en) * 2006-03-22 2010-09-16 Nissan Motor Co., Ltd. Avoidance maneuver calculation device, avoidance control device, vehicle having each device, avoidance maneuver calculating method, and avoidance controlling method
US20140229069A1 (en) * 2011-09-26 2014-08-14 Toyota Jidosha Kabushiki Kaisha Driving support system for a vehicle
US20130151058A1 (en) * 2011-12-09 2013-06-13 GM Global Technology Operations LLC Method and system for controlling a host vehicle
JP5853719B2 (en) 2012-01-20 2016-02-09 株式会社リコー Image processing system and vehicle equipped with the same
US20160217333A1 (en) * 2015-01-26 2016-07-28 Ricoh Company, Ltd. Information processing apparatus and information processing system
US20180107946A1 (en) * 2015-06-22 2018-04-19 Denso Corporation Alert output apparatus
US20170168503A1 (en) * 2015-12-11 2017-06-15 Avishtech, Llc Autonomous vehicle towing system and method
US20180032082A1 (en) * 2016-01-05 2018-02-01 Mobileye Vision Technologies Ltd. Machine learning navigational engine with imposed constraints
US20170291608A1 (en) * 2016-04-11 2017-10-12 Volkswagen Ag Method for the autonomous or partly autonomous execution of a cooperative driving maneuver
US20180165966A1 (en) * 2016-12-13 2018-06-14 Toyota Jidosha Kabushiki Kaisha Obstacle determination system and non-transitory computer-readable recording medium storing obstacle determination program
JP2018171959A (en) * 2017-03-31 2018-11-08 株式会社Subaru Travel support device for vehicle
US20200166360A1 (en) * 2017-06-09 2020-05-28 Bosch Engineering K.K. Travel obstruction detection device and vehicle navigation system
US20190256088A1 (en) * 2018-11-02 2019-08-22 Intel Corporation Methods and apparatus to generate vehicle warnings

Also Published As

Publication number Publication date
US20200275061A1 (en) 2020-08-27
JP2020135797A (en) 2020-08-31
JP7359551B2 (en) 2023-10-11
CN111613077A (en) 2020-09-01

Similar Documents

Publication Publication Date Title
US11941985B2 (en) Vehicular environment estimation device
US10688997B2 (en) Lane merging determination apparatus
US9620008B2 (en) Method and system for using global scene context for adaptive prediction and corresponding program, and vehicle equipped with such system
US11136037B2 (en) Vehicle control device and vehicle control method
US20210070317A1 (en) Travel plan generation device, travel plan generation method, and non-transitory tangible computer readable storage medium
US20080114531A1 (en) Communication apparatus for use in intervehicle communication system and program for same
JP7247042B2 (en) Vehicle control system, vehicle control method, and program
US20200013242A1 (en) Sensor failure detection device and control method for same
JP7135796B2 (en) Server device, server control method, server control program, vehicle, vehicle control method, and vehicle control program
JP4225190B2 (en) Vehicle driving support device
US11383734B2 (en) Vehicle control apparatus, vehicle control method, vehicle, and non-transitory computer-readable storage medium
US20190071073A1 (en) Vehicle control apparatus, vehicle, processing method of vehicle control apparatus, and storage medium
US11252380B2 (en) Road management system
CN115123213A (en) Driving support device, vehicle, and driving support method
JP2003312415A (en) Vehicular pre-presenting device
US20200312125A1 (en) Communication system
JP2002148336A (en) Interruption prediction device
JP7069726B2 (en) Notification device and in-vehicle device
JP2009137494A (en) Obstacle determination device
WO2022039040A1 (en) Vehicular system and target identification program
JP2011134087A (en) Driving assist system
JP7363062B2 (en) Driving environment abnormality determination system
CN117916786A (en) In-vehicle apparatus, in-vehicle system, control method, and computer program
JP2018180882A (en) Driving support device

Legal Events

Date Code Title Description
AS Assignment

Owner name: HONDA MOTOR CO., LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OISHI, YASUO;MATSUURA, KAZUYA;KAKINUMA, ATSUKI;AND OTHERS;REEL/FRAME:051923/0625

Effective date: 20200212

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE