US20190213884A1 - Detection system and method - Google Patents

Detection system and method Download PDF

Info

Publication number
US20190213884A1
US20190213884A1 US15/866,188 US201815866188A US2019213884A1 US 20190213884 A1 US20190213884 A1 US 20190213884A1 US 201815866188 A US201815866188 A US 201815866188A US 2019213884 A1 US2019213884 A1 US 2019213884A1
Authority
US
United States
Prior art keywords
sensor
vehicle
interest
accordance
host
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/866,188
Other versions
US11488476B2 (en
Inventor
Junsung KIM
Junqing Wei
Wenda Xu
Gaurav Bhatia
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Delphi Technologies LLC
Motional AD LLC
Original Assignee
Aptiv Technologies Ltd
Delphi Technologies LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aptiv Technologies Ltd, Delphi Technologies LLC filed Critical Aptiv Technologies Ltd
Priority to US15/866,188 priority Critical patent/US11488476B2/en
Assigned to DELPHI TECHNOLOGIES, LLC reassignment DELPHI TECHNOLOGIES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BHATIA, GAURAV, Kim, Junsung, WEI, JUNQING, XU, Wenda
Publication of US20190213884A1 publication Critical patent/US20190213884A1/en
Assigned to APTIV TECHNOLOGIES LIMITED reassignment APTIV TECHNOLOGIES LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DELPHI TECHNOLOGIES LLC
Assigned to MOTIONAL AD LLC reassignment MOTIONAL AD LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: APTIV TECHNOLOGIES LIMITED
Application granted granted Critical
Publication of US11488476B2 publication Critical patent/US11488476B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/87Combinations of radar systems, e.g. primary radar and secondary radar
    • G01S13/878Combination of several spaced transmitters or receivers of known location for determining the position of a transponder or a reflector
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/87Combinations of radar systems, e.g. primary radar and secondary radar
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/93Radar or analogous systems specially adapted for specific applications for anti-collision purposes
    • G01S13/931Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/93Lidar systems specially adapted for specific applications for anti-collision purposes
    • G01S17/931Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/003Transmission of data between radar, sonar or lidar systems and remote stations
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/02Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
    • G01S7/40Means for monitoring or calibrating
    • G01S7/4004Means for monitoring or calibrating of parts of a radar system
    • G01S7/4026Antenna boresight
    • G06K9/00805
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/161Decentralised systems, e.g. inter-vehicle communication
    • G08G1/163Decentralised systems, e.g. inter-vehicle communication involving continuous checking
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/165Anti-collision systems for passive traffic, e.g. including static obstacles, trees
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/166Anti-collision systems for active traffic, e.g. moving vehicles, pedestrians, bikes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/12Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W84/00Network topologies
    • H04W84/18Self-organising networks, e.g. ad-hoc networks or sensor networks
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2420/00Indexing codes relating to the type of sensors based on the principle of their operation
    • B60W2420/40Photo, light or radio wave sensitive means, e.g. infrared sensors
    • B60W2420/403Image sensing, e.g. optical camera
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2420/00Indexing codes relating to the type of sensors based on the principle of their operation
    • B60W2420/40Photo, light or radio wave sensitive means, e.g. infrared sensors
    • B60W2420/408Radar; Laser, e.g. lidar
    • B60W2420/42
    • B60W2420/52
    • B60W2550/30
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2554/00Input parameters relating to objects
    • B60W2554/80Spatial relation or speed relative to objects
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/66Radar-tracking systems; Analogous systems
    • G01S13/72Radar-tracking systems; Analogous systems for two-dimensional tracking, e.g. combination of angle and range tracking, track-while-scan radar
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/86Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
    • G01S13/867Combination of radar systems with cameras
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/91Radar or analogous systems specially adapted for specific applications for traffic control
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/93Radar or analogous systems specially adapted for specific applications for anti-collision purposes
    • G01S13/931Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
    • G01S2013/9316Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles combined with communication equipment with other vehicles or with base stations
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/93Radar or analogous systems specially adapted for specific applications for anti-collision purposes
    • G01S13/931Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
    • G01S2013/9327Sensor installation details
    • G01S2013/93271Sensor installation details in the front of the vehicles
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/02Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
    • G01S7/40Means for monitoring or calibrating
    • G01S7/4004Means for monitoring or calibrating of parts of a radar system
    • G01S7/4039Means for monitoring or calibrating of parts of a radar system of sensor or antenna obstruction, e.g. dirt- or ice-coating

Definitions

  • This disclosure generally relates to a detection system, and more particularly relates to a detection system with a selectable perspective view.
  • FIG. 1 is an illustration of a detection system in accordance with one embodiment
  • FIG. 2 is an illustration of a traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment
  • FIG. 3 is an illustration of another traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment
  • FIG. 4 is an illustration of yet another traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment
  • FIG. 5 is an illustration of yet another traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment.
  • FIG. 6 is a flowchart of a method of operating a detection system in accordance with another embodiment.
  • FIG. 1 illustrates a non-limiting example of a detection system 10 , hereafter referred to as the system 10 , installed in a host-vehicle 12 .
  • the system 10 is an improvement over prior detection systems because the system 10 uses data from a second-sensor 14 , remote from a first-sensor 16 , when a first-field-of-view 18 of the first-sensor 16 is occluded, blocked, or otherwise obstructed.
  • the host-vehicle 12 may be characterized as an automated vehicle.
  • the term automated vehicle may apply to instances when the host-vehicle 12 is being operated in an automated-mode, i.e. a fully autonomous mode, where a human-operator (not shown) of the host-vehicle 12 may do little more than designate a destination to operate the host-vehicle 12 .
  • full automation is not a requirement.
  • the teachings presented herein are useful when the host-vehicle 12 is operated in a manual-mode where the degree or level of automation may be little more than providing an audible or visual warning to the human-operator who is generally in control of the steering, accelerator, and brakes of the host-vehicle 12 .
  • the system 10 may merely assist the human-operator as needed to change lanes and/or avoid interference with and/or a collision with, for example, an object such as an other-vehicle, a pedestrian, or a road sign.
  • the system 10 includes the first-sensor 16 positioned on the host-vehicle 12 at a first-location 20 , such as on the front of the host-vehicle 12 , or in the passenger-compartment of the host-vehicle 12 .
  • the first-sensor 16 detects objects 22 in the first- field-of-view 18 , and is illustrated as a camera 24 in FIG. 1 .
  • the first-sensor 16 may be a ranging-sensor 26 , such as a radar-sensor or a lidar-sensor.
  • Examples of the camera 24 suitable for use on the host-vehicle 12 are commercially available as will be recognized by those in the art, one such being the APTINA MT9V023 from Micron Technology, Inc.
  • the camera 24 may be mounted in the first-location 20 that is on the front of the host-vehicle 12 , or mounted in the interior of the host-vehicle 12 at a location suitable for the camera 24 to view the area around the host-vehicle 12 through the windshield of the host-vehicle 12 .
  • the camera 24 is preferably a video-type camera 24 or camera 24 that can capture images of the surrounding area at a sufficient frame-rate, of ten frames per second, for example.
  • the system 10 also includes the second-sensor 14 positioned at a second-location 28 that is different than the first-location 20 .
  • the second-location 28 may be on the host-vehicle 12 (shown as 28 A), on a second-vehicle 30 (shown as 28 B), or part of a transportation-infrastructure 32 (shown as 28 C), as will be described in more detail below.
  • the second-vehicle 30 may be referred to as the alternative-vehicle 30 .
  • the second-sensor 14 detects objects 22 in a second-field-of-view 34 that at least partially overlaps the first-field-of-view 18 .
  • the objects 22 in the second-field-of-view 34 may include the same or different objects 22 than the objects 22 in the first- field-of-view 18 , as will become apparent with a reading of the descriptions of the traffic scenarios of FIGS. 2-5 below.
  • the object 22 that is detected in both the first-field-of-view 18 and the second-field-of-view 34 may be determined to be the same object 22 based on the position of the object 22 . That is, the position of the object 22 may have the same coordinates of latitude, longitude, and elevation in the first-field-of-view 18 as the coordinates in the second-field-of-view 34 .
  • the system 10 also includes a controller 36 in communication with the first-sensor 16 and the second-sensor 14 .
  • the first-sensor 16 and the second-sensor 14 may be hard-wired to the controller 36 through the host-vehicle's 12 electrical-system (not shown), or may be in wireless communication with the controller 36 .
  • the controller 36 may include a processor (not shown) such as a microprocessor or other control circuitry such as analog and/or digital control circuitry including an application specific integrated circuit (ASIC) for processing data as should be evident to those in the art.
  • the controller 36 may include a memory (not specifically shown), including non-volatile memory, such as electrically erasable programmable read-only memory (EEPROM) for storing one or more routines, thresholds, and captured data.
  • the one or more routines may be executed by the processor to perform steps for detecting the objects 22 based on signals received by the controller 36 from the first-sensor 16 and the second-sensor 14 as described herein.
  • FIG. 2 illustrates a traffic scenario where the host-vehicle 12 equipped with the system 10 is approaching a crosswalk on a roadway 40 and a pedestrian 38 is also attempting to cross the roadway 40 in the crosswalk.
  • a lead-vehicle 42 is disposed between the host-vehicle 12 and the pedestrian 38 such that the lead-vehicle 42 blocks a portion 43 of the first-field-of-view 18 of the first-sensor 16 (i.e. the camera 24 ).
  • the controller 36 selects the second-sensor 14 A to detect an object-of-interest 44 (i.e. the pedestrian 38 ) in accordance with a determination that an obstruction 46 (i.e. the lead-vehicle 42 ) blocks a first-line-of-sight 48 between the first-sensor 16 and the object-of-interest 44 .
  • the second-sensor 14 A is mounted on a front right-corner of the host-vehicle 12 .
  • the second-sensor 14 A may be one of either the camera 24 or the ranging-sensor 26 that is in a different location (i.e. the second-location 28 A) relative to the first-sensor 16 .
  • the obstruction 46 may be any obstruction 46 including a building, signage, or construction-barriers, for example.
  • the object-of-interest 44 is the pedestrian 38 in the roadway 40 , but may be any object-of-interest 44 , including a lane-marking 50 , a traffic-intersection 52 , an edge 54 of the roadway 40 , or an other-vehicle 56 .
  • the system 10 provides the technical benefits of improving a road-model and/or improving a lane-keeping-algorithm that may use these features of the roadway 40 as inputs.
  • the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is passing through the traffic-intersection 52 that is obscured by buildings, trucks, etc.
  • the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is making a left-hand turn at the traffic-intersection 52 behind a large vehicle (e.g. a Class 8 truck) that may block oncoming traffic or may obscure a traffic-light.
  • a large vehicle e.g. a Class 8 truck
  • the controller 36 is able to determine that the obstruction 46 blocks the first-line-of-sight 48 based on a tracking 58 of the object-of-interest 44 prior to the obstruction 46 blocking the first-line-of-sight 48 .
  • the specific process of tracking 58 used by the controller 36 may be dependent upon the sensor-type (i.e., camera 24 , ranging-sensor 26 ) and will be understood by those in the art.
  • the controller 36 performs the tracking 58 based initially on signals received by the first-sensor 16 while the object-of-interest 44 is exposed (i.e. visible, unobstructed first-line-of-sight 48 , etc.) in the first-field-of-view 18 .
  • the obstruction 46 may cause an abrupt loss of tracking-data from the object-of-interest 44 to the first-sensor 16 , whereby the controller 36 determines that the obstruction 46 is present.
  • the controller 36 selects the second-sensor 14 with an unobstructed second-line-of-sight 60 to the object-of-interest 44 .
  • the second-sensor 14 tracks 58 the object-of-interest 44 in the second-field-of-view 34 , wherein the second-sensor 14 detects at least a part of the object-of-interest 44 .
  • the controller 36 selects the second-sensor 14 A mounted on the host-vehicle 12 with the second-line-of-sight 60 A.
  • the controller 36 tracks 58 the pedestrian 38 while the pedestrian 38 is in plain view of the first-sensor 16 .
  • the controller 36 determines that the associated loss of tracking-data from the pedestrian 38 is caused by the lead-vehicle 42 .
  • the controller 36 selects the second-sensor 14 A, that has the unobstructed view of the pedestrian 38 , to resume the tracking 58 of the pedestrian 38 . If the view of the second-sensor 14 A to the pedestrian 38 is blocked, then the controller 36 may select another second-sensor ( 14 B or 14 C) that has the unobstructed view of the pedestrian 38 .
  • FIG. 3 illustrates another traffic scenario where the second-vehicle 30 is at the traffic-intersection 52 and the second-sensor 14 B is mounted on the second-vehicle 30 that has the unobstructed view of the object-of-interest 44 (i.e. the pedestrian 38 ).
  • the second-sensor 14 B may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16 (i.e. the second-location 28 B).
  • Data from the second-sensor 14 B mounted on the second-vehicle 30 is both requested 64 and transferred 66 from the second-vehicle 30 to the host-vehicle 12 using a communications protocol, such as dedicated short range communications (DSRC), Bluetooth ®, cellular or other ad hoc protocols.
  • the data is both requested 64 and transferred 66 from the second-vehicle 30 by way of a transceiver 68 (see FIG. 1 ) mounted on the host-vehicle 12 that is also in communication with the controller 36
  • FIG. 4 illustrates yet another traffic scenario where the second-sensor 14 C is part of the transportation-infrastructure 32 (e.g. mounted to a traffic light at the second-location 28 C) located proximate the traffic-intersection 52 , and has the unobstructed view of the object-of-interest 44 .
  • the second-sensor 14 C may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16 .
  • Data from the second-sensor 14 C that is part of the transportation-infrastructure 32 is both requested 64 and transferred 66 from the transportation-infrastructure 32 to the host-vehicle 12 using a communications protocol, such as DSRC, UHF, VHF, Bluetooth ®, cellular WiMAX, GSM, 3G or other ad hoc protocols.
  • the data is both requested 64 and transferred 66 from the transportation-infrastructure 32 by way of the transceiver 68 mounted on the host-vehicle 12 that is also in communication with the controller 36 , as illustrated in FIG. 1 .
  • the second-sensor 14 may be a directional-sensor 72 where the second-field-of-view 34 (i.e. 34 A- 34 C) may be adjusted to detect the object-of-interest 44 . That is, the controller 36 may rotate the second-sensor 14 that is the directional-sensor 72 as necessary, and/or may change an angle-of-elevation of a bore-site of the second-sensor 14 as necessary to capture the object-of-interest 44 . The controller 36 may adjust the second-field-of-view 34 as required to resume tracking 58 the object-of-interest 44 .
  • the controller 36 may rotate the second-sensor 14 that is the directional-sensor 72 as necessary, and/or may change an angle-of-elevation of a bore-site of the second-sensor 14 as necessary to capture the object-of-interest 44 .
  • the controller 36 may adjust the second-field-of-view 34 as required to resume tracking 58 the object-of-interest 44 .
  • FIG. 5 illustrates yet another traffic scenario with all three second-sensors 14 A- 14 C 14 C present.
  • the controller 36 may further select the second-sensor 14 based on a proximity of the second-sensor 14 to both the host-vehicle 12 and the object-of-interest 44 .
  • the controller 36 polls (i.e. samples, surveys) the available second-sensors 14 within a communications-range via the second-sensor's 14 respective communications protocol and prioritizes the data based on the second-sensor's 14 point-of-view 74 , sensor-distance 76 from the host-vehicle 12 , and object-distance 78 to the object-of-interest 44 .
  • the point-of-view 74 is the position of the second-sensor 14 from which the object-of-interest 44 is detected.
  • the controller 36 may select the second-sensor 14 with the point-of-view 74 that provides a least obstructed view of the object-of-interest 44 .
  • the second-sensor 14 B mounted on the second-vehicle 30 provides the least obstructed point-of-view 74 B of the pedestrian 38 compared to the point-of-view 74 A of the second-sensor 14 A mounted on the host-vehicle 12 . Therefore, the second-sensor 14 B may be assigned a higher priority than the second-sensor 14 A.
  • the second-sensor 14 C mounted to the traffic light may have the least obstructed view of the pedestrian 38 compared to the second-sensor 14 B mounted on the second-vehicle 30 (possibly due to its greater elevation above the traffic-intersection 52 ), and may be assigned a higher priority than the second-sensor 14 B.
  • the sensor-distance 76 from the host-vehicle 12 (i.e. the distance from the second-sensor 14 B, 14 C to the host-vehicle 12 ) is prioritized because a signal-strength and a signal-quality of the data transferred 66 by the second-sensor 14 B- 14 C may be affected by the sensor-distance 76 B, 76 C. Therefore, the second-sensor 14 B, 14 C with the shortest sensor-distance 76 B, 76 C may be assigned a higher priority by the controller 36 .
  • the object-distance 78 (i.e. the distance from the second-sensor 14 to the object-of-interest 44 ) is prioritized because a resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14 may be affected by the object-distance 78 A, 78 B, 78 C. Therefore, the second-sensor 14 with the shortest object-distance 78 A, 78 B, 78 C may be assigned a higher priority by the controller 36 .
  • the controller 36 may also prioritize the second-sensor 14 by a sensor-type (i.e. the camera 24 or the ranging-sensor 26 ).
  • the controller 36 may assign the higher priority to the second-sensor 14 that provides the higher resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14 .
  • the inherent resolutions of the various sensor-types will be understood by one skilled in the art.
  • FIG. 6 illustrates a method 200 of operating a detection system 10 , hereafter referred to as the system 10 , installed in a host-vehicle 12 .
  • Step 202 DETECT OBJECTS IN FIRST-FIELD-OF-VIEW, includes detecting objects 22 in a first-field-of-view 18 with a first-sensor 16 mounted on the host-vehicle 12 .
  • FIG. 1 illustrates a non-limiting example of the detection system 10 installed in the host-vehicle 12 .
  • the system 10 is an improvement over prior detection systems because the system 10 uses data from a second-sensor 14 , remote from the first-sensor 16 , when the first-field-of-view 18 of the first-sensor 16 is occluded, blocked, or otherwise obstructed.
  • the system 10 includes the first-sensor 16 positioned on the host-vehicle 12 at a first-location 20 , such as on the front of the host-vehicle 12 , or in the passenger-compartment of the host-vehicle 12 .
  • the first-sensor 16 detects objects 22 in the first- field-of-view 18 , and is illustrated as a camera 24 in FIG. 1 .
  • the first-sensor 16 may be a ranging-sensor 26 , such as a radar-sensor or a lidar-sensor.
  • Examples of the camera 24 suitable for use on the host-vehicle 12 are commercially available as will be recognized by those in the art, one such being the APTINA MT9V023 from Micron Technology, Inc.
  • the camera 24 may be mounted in the first-location 20 that is on the front of the host-vehicle 12 , or mounted in the interior of the host-vehicle 12 at a location suitable for the camera 24 to view the area around the host-vehicle 12 through the windshield of the host-vehicle 12 .
  • the camera 24 is preferably a video-type camera 24 or camera 24 that can capture images of the surrounding area at a sufficient frame-rate, of ten frames per second, for example.
  • Step 204 DETERMINE OBSTRUCTION PRESENT, includes determining whether an obstruction 46 blocks a first-line-of-sight 48 to an object-of-interest 44 .
  • the system 10 also includes a controller 36 in communication with the first-sensor 16 and the second-sensor 14 .
  • the first-sensor 16 and the second-sensor 14 may be hard-wired to the controller 36 through the host-vehicle's 12 electrical-system (not shown), or may be in wireless communication with the controller 36 .
  • the controller 36 may include a processor (not shown) such as a microprocessor or other control circuitry such as analog and/or digital control circuitry including an application specific integrated circuit (ASIC) for processing data as should be evident to those in the art.
  • the controller 36 may include a memory (not specifically shown), including non-volatile memory, such as electrically erasable programmable read-only memory (EEPROM) for storing one or more routines, thresholds, and captured data.
  • the one or more routines may be executed by the processor to perform steps for detecting the objects 22 based on signals received by the controller 36 from the first-sensor 16 and the second-sensor 14 as described herein.
  • the controller 36 is able to determine that the obstruction 46 blocks the first-line-of-sight 48 based on a tracking 58 of the object-of-interest 44 prior to the obstruction 46 blocking the first-line-of-sight 48 .
  • the specific process of tracking 58 used by the controller 36 may be dependent upon the sensor-type (i.e., camera 24 , ranging-sensor 26 ) and will be understood by those in the art.
  • the controller 36 performs the tracking 58 based initially on signals received by the first-sensor 16 while the object-of-interest 44 is exposed (i.e. visible, unobstructed first-line-of-sight 48 , etc.) in the first-field-of-view 18 .
  • the obstruction 46 may cause an abrupt loss of tracking-data from the object-of-interest 44 to the first-sensor 16 , whereby the controller 36 determines that the obstruction 46 is present.
  • Step 206 DETECT OBJECTS IN SECOND-FIELD-OF-VIEW, includes detecting objects 22 in a second-field-of-view 34 with the second-sensor 14 .
  • the system 10 also includes the second-sensor 14 positioned at a second-location 28 that is different than the first-location 20 .
  • the second-location 28 may be on the host-vehicle 12 (shown as 28 A), on a second-vehicle 30 (shown as 28 B), or part of a transportation-infrastructure 32 (shown as 28 C), as will be described in more detail below.
  • the second-vehicle 30 may be referred to as the alternative-vehicle 30 .
  • the second-sensor 14 detects objects 22 in a second-field-of-view 34 that at least partially overlaps the first-field-of-view 18 .
  • the objects 22 in the second-field-of-view 34 may include the same or different objects 22 than the objects 22 in the first-field-of-view 18 , as will become apparent with a reading of the descriptions of the traffic scenarios of FIGS. 2-5 below.
  • the object 22 that is detected in both the first-field-of-view 18 and the second-field-of-view 34 may be determined to be the same object 22 based on the position of the object 22 . That is, the position of the object 22 may have the same coordinates of latitude, longitude, and elevation in the first-field-of-view 18 as the coordinates in the second-field-of-view 34 .
  • Step 208 includes selecting the second-sensor 14 with the controller 36 to detect the object-of-interest 44 in accordance with a determination that the obstruction 46 blocks the first-line-of-sight 48 between the first-sensor 16 and the object-of-interest 44 .
  • FIG. 2 illustrates a traffic scenario where the host-vehicle 12 equipped with the system 10 is approaching a crosswalk on a roadway 40 and a pedestrian 38 is also attempting to cross the roadway 40 in the crosswalk.
  • a lead-vehicle 42 is disposed between the host-vehicle 12 and the pedestrian 38 such that the lead-vehicle 42 blocks a portion 43 of the first-field-of-view 18 of the first-sensor 16 (i.e. the camera 24 ).
  • the controller 36 selects the second-sensor 14 A to detect the object-of-interest 44 (i.e. the pedestrian 38 ) in accordance with the determination that the obstruction 46 (i.e. the lead-vehicle 42 ) blocks the first-line-of-sight 48 between the first-sensor 16 and the object-of-interest 44 .
  • the second-sensor 14 A is mounted on a front right-corner of the host-vehicle 12 .
  • the second-sensor 14 A may be one of either the camera 24 or the ranging-sensor 26 that is in a different location (i.e. the second-location 28 A) relative to the first-sensor 16 .
  • the obstruction 46 may be any obstruction 46 including a building, signage, or construction-barriers, for example.
  • the object-of-interest 44 is the pedestrian 38 in the roadway 40 , but may be any object-of-interest 44 , including a lane-marking 50 , a traffic-intersection 52 , an edge 54 of the roadway 40 , or an other-vehicle 56 .
  • the system 10 provides the technical benefits of improving a road-model and/or improving a lane-keeping-algorithm that may use these features of the roadway 40 as inputs.
  • the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is passing through the traffic-intersection 52 that is obscured by buildings, trucks, etc.
  • the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is making a left-hand turn at the traffic-intersection 52 behind a large vehicle (e.g. a Class 8 truck) that may block oncoming traffic or may obscure a traffic-light.
  • a large vehicle e.g. a Class 8 truck
  • the controller 36 selects the second-sensor 14 with an unobstructed second-line-of-sight 60 to the object-of-interest 44 .
  • the second-sensor 14 then tracks 58 the object-of-interest 44 in the second-field-of-view 34 , wherein the second-sensor 14 detects at least a part of the object-of-interest 44 .
  • the controller 36 selects the second-sensor 14 A mounted on the host-vehicle 12 with the second-line-of-sight 60 A.
  • the controller 36 tracks 58 the pedestrian 38 while the pedestrian 38 is in plain view of the first-sensor 16 .
  • the controller 36 determines that the associated loss of tracking-data from the pedestrian 38 is caused by the lead-vehicle 42 .
  • the controller 36 selects the second-sensor 14 A, that has the unobstructed view of the pedestrian 38 , to resume the tracking 58 of the pedestrian 38 . If the view of the second-sensor 14 A to the pedestrian 38 is blocked, then the controller 36 may select another second-sensor ( 14 B or 14 C) that has the unobstructed view of the pedestrian 38 .
  • FIG. 3 illustrates another traffic scenario where the second-vehicle 30 is at the traffic-intersection 52 and the second-sensor 14 B is mounted on the second-vehicle 30 that has the unobstructed view of the object-of-interest 44 (i.e. the pedestrian 38 ).
  • the second-sensor 14 B may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16 (i.e. the second-location 28 B).
  • Data from the second-sensor 14 B mounted on the second-vehicle 30 is both requested 64 and transferred 66 from the second-vehicle 30 to the host-vehicle 12 using a communications protocol, such as dedicated short range communications (DSRC), Bluetooth ®, cellular or other ad hoc protocols.
  • the data is both requested 64 and transferred 66 from the second-vehicle 30 by way of a transceiver 68 (see FIG. 1 ) mounted on the host-vehicle 12 that is also in communication with the controller 36
  • FIG. 4 illustrates yet another traffic scenario where the second-sensor 14 C is part of a transportation-infrastructure 32 (e.g. mounted to a traffic light at the second-location 28 C) located proximate the traffic-intersection 52 , and has the unobstructed view of the object-of-interest 44 .
  • the second-sensor 14 C may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16 .
  • Data from the second-sensor 14 C that is part of the transportation-infrastructure 32 is both requested 64 and transferred 66 from the transportation-infrastructure 32 to the host-vehicle 12 using a communications protocol, such as DSRC, UHF, VHF, Bluetooth ®, cellular WiMAX, GSM, 3G or other ad hoc protocols.
  • the data is both requested 64 and transferred 66 from the transportation-infrastructure 32 by way of the transceiver 68 mounted on the host-vehicle 12 that is also in communication with the controller 36 , as illustrated in FIG. 1 .
  • the second-sensor 14 may be a directional-sensor 72 where the second-field-of-view 34 (i.e. 34 A- 34 C) may be adjusted to detect the object-of-interest 44 . That is, the controller 36 may rotate the second-sensor 14 that is the directional-sensor 72 as necessary, and/or may change an angle-of-elevation of a bore-site of the second-sensor 14 as necessary to capture the object-of-interest 44 . The controller 36 may adjust the second-field-of-view 34 as required to resume tracking 58 the object-of-interest 44 .
  • the controller 36 may rotate the second-sensor 14 that is the directional-sensor 72 as necessary, and/or may change an angle-of-elevation of a bore-site of the second-sensor 14 as necessary to capture the object-of-interest 44 .
  • the controller 36 may adjust the second-field-of-view 34 as required to resume tracking 58 the object-of-interest 44 .
  • FIG. 5 illustrates yet another traffic scenario with all three second-sensors 14 A- 14 C present.
  • the controller 36 may further select the second-sensor 14 based on a proximity of the second-sensor 14 to both the host-vehicle 12 and the object-of-interest 44 .
  • the controller 36 polls (i.e. samples, surveys) the available second-sensors 14 within a communications-range via the second-sensor's 14 respective communications protocol and prioritizes the data based on the second-sensor's 14 point-of-view 74 , sensor-distance 76 from the host-vehicle 12 , and object-distance 78 to the object-of-interest 44 .
  • the point-of-view 74 is the position of the second-sensor 14 from which the object-of-interest 44 is detected.
  • the controller 36 may select the second-sensor 14 with the point-of-view 74 that provides a least obstructed view of the object-of-interest 44 .
  • the second-sensor 14 B mounted on the second-vehicle 30 provides the least obstructed point-of-view 74 B of the pedestrian 38 compared to the point-of-view 74 A of the second-sensor 14 A mounted on the host-vehicle 12 . Therefore, the second-sensor 14 B may be assigned a higher priority than the second-sensor 14 A.
  • the second-sensor 14 C mounted to the traffic light may have the least obstructed view of the pedestrian 38 compared to the second-sensor 14 B mounted on the second-vehicle 30 (possibly due to its greater elevation above the traffic-intersection 52 ), and may be assigned a higher priority than the second-sensor 14 B.
  • the sensor-distance 76 from the host-vehicle 12 (i.e. the distance from the second-sensor 14 B, 14 C to the host-vehicle 12 ) is prioritized because a signal-strength and a signal-quality of the data transferred 66 by the second-sensor 14 B- 14 C may be affected by the sensor-distance 76 B, 76 C. Therefore, the second-sensor 14 B, 14 C with the shortest sensor-distance 76 B, 76 C may be assigned a higher priority by the controller 36 .
  • the object-distance 78 (i.e. the distance from the second-sensor 14 to the object-of-interest 44 ) is prioritized because a resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14 may be affected by the object-distance 78 A, 78 B, 78 C. Therefore, the second-sensor 14 with the shortest object-distance 78 A, 78 B, 78 C may be assigned a higher priority by the controller 36 .
  • the controller 36 may also prioritize the second-sensor 14 by a sensor-type (i.e. the camera 24 or the ranging-sensor 26 ).
  • the controller 36 may assign the higher priority to the second-sensor 14 that provides the higher resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14 .
  • the inherent resolutions of the various sensor-types will be understood by one skilled in the art.
  • a detection system 10 (the system 10 ), a controller 36 for the system 10 , and a method 200 of operating the system 10 are provided.
  • the system 10 is beneficial because the system 10 uses the second-sensor 14 to track 58 the object-of-interest 44 when the obstruction 46 blocks the first-line-of-sight 48 of the first-sensor 16 .

Landscapes

  • Engineering & Computer Science (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Electromagnetism (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Automation & Control Theory (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Traffic Control Systems (AREA)

Abstract

A detection system includes a first-sensor, a second-sensor, and a controller. The first-sensor is mounted on a host-vehicle. The first-sensor detects objects in a first-field-of-view. The second-sensor is positioned at a second-location different than the first-location. The second-sensor detects objects in a second-field-of-view that at least partially overlaps the first-field of view. The controller is in communication with the first-sensor and the second-sensor. The controller selects the second-sensor to detect an object-of-interest in accordance with a determination that an obstruction blocks a first-line-of-sight between the first-sensor and the object-of-interest.

Description

    TECHNICAL FIELD OF INVENTION
  • This disclosure generally relates to a detection system, and more particularly relates to a detection system with a selectable perspective view.
  • BRIEF DESCRIPTION OF DRAWINGS
  • The present invention will now be described, by way of example with reference to the accompanying drawings, in which:
  • FIG. 1 is an illustration of a detection system in accordance with one embodiment;
  • FIG. 2 is an illustration of a traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment;
  • FIG. 3 is an illustration of another traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment;
  • FIG. 4 is an illustration of yet another traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment;
  • FIG. 5 is an illustration of yet another traffic scenario of a vehicle equipped with the detection system of FIG. 1 in accordance with one embodiment; and
  • FIG. 6 is a flowchart of a method of operating a detection system in accordance with another embodiment.
  • DETAILED DESCRIPTION
  • FIG. 1 illustrates a non-limiting example of a detection system 10, hereafter referred to as the system 10, installed in a host-vehicle 12. As will be described in more detail below, the system 10 is an improvement over prior detection systems because the system 10 uses data from a second-sensor 14, remote from a first-sensor 16, when a first-field-of-view 18 of the first-sensor 16 is occluded, blocked, or otherwise obstructed.
  • The host-vehicle 12 may be characterized as an automated vehicle. As used herein, the term automated vehicle may apply to instances when the host-vehicle 12 is being operated in an automated-mode, i.e. a fully autonomous mode, where a human-operator (not shown) of the host-vehicle 12 may do little more than designate a destination to operate the host-vehicle 12. However, full automation is not a requirement. It is contemplated that the teachings presented herein are useful when the host-vehicle 12 is operated in a manual-mode where the degree or level of automation may be little more than providing an audible or visual warning to the human-operator who is generally in control of the steering, accelerator, and brakes of the host-vehicle 12. For example, the system 10 may merely assist the human-operator as needed to change lanes and/or avoid interference with and/or a collision with, for example, an object such as an other-vehicle, a pedestrian, or a road sign.
  • The system 10 includes the first-sensor 16 positioned on the host-vehicle 12 at a first-location 20, such as on the front of the host-vehicle 12, or in the passenger-compartment of the host-vehicle 12. The first-sensor 16 detects objects 22 in the first- field-of-view 18, and is illustrated as a camera 24 in FIG. 1. Alternatively, the first-sensor 16 may be a ranging-sensor 26, such as a radar-sensor or a lidar-sensor. Examples of the camera 24 suitable for use on the host-vehicle 12 are commercially available as will be recognized by those in the art, one such being the APTINA MT9V023 from Micron Technology, Inc. of Boise, Id., USA. The camera 24 may be mounted in the first-location 20 that is on the front of the host-vehicle 12, or mounted in the interior of the host-vehicle 12 at a location suitable for the camera 24 to view the area around the host-vehicle 12 through the windshield of the host-vehicle 12. The camera 24 is preferably a video-type camera 24 or camera 24 that can capture images of the surrounding area at a sufficient frame-rate, of ten frames per second, for example.
  • The system 10 also includes the second-sensor 14 positioned at a second-location 28 that is different than the first-location 20. The second-location 28 may be on the host-vehicle 12 (shown as 28A), on a second-vehicle 30 (shown as 28B), or part of a transportation-infrastructure 32 (shown as 28C), as will be described in more detail below. Alternatively, the second-vehicle 30 may be referred to as the alternative-vehicle 30. The second-sensor 14 detects objects 22 in a second-field-of-view 34 that at least partially overlaps the first-field-of-view 18. Depending on the amount of overlap of the second-field-of-view 34 with the first-field-of-view 18, the objects 22 in the second-field-of-view 34 may include the same or different objects 22 than the objects 22 in the first- field-of-view 18, as will become apparent with a reading of the descriptions of the traffic scenarios of FIGS. 2-5 below. The object 22 that is detected in both the first-field-of-view 18 and the second-field-of-view 34 may be determined to be the same object 22 based on the position of the object 22. That is, the position of the object 22 may have the same coordinates of latitude, longitude, and elevation in the first-field-of-view 18 as the coordinates in the second-field-of-view 34.
  • The system 10 also includes a controller 36 in communication with the first-sensor 16 and the second-sensor 14. The first-sensor 16 and the second-sensor 14 may be hard-wired to the controller 36 through the host-vehicle's 12 electrical-system (not shown), or may be in wireless communication with the controller 36. The controller 36 may include a processor (not shown) such as a microprocessor or other control circuitry such as analog and/or digital control circuitry including an application specific integrated circuit (ASIC) for processing data as should be evident to those in the art. The controller 36 may include a memory (not specifically shown), including non-volatile memory, such as electrically erasable programmable read-only memory (EEPROM) for storing one or more routines, thresholds, and captured data. The one or more routines may be executed by the processor to perform steps for detecting the objects 22 based on signals received by the controller 36 from the first-sensor 16 and the second-sensor 14 as described herein.
  • FIG. 2 illustrates a traffic scenario where the host-vehicle 12 equipped with the system 10 is approaching a crosswalk on a roadway 40 and a pedestrian 38 is also attempting to cross the roadway 40 in the crosswalk. A lead-vehicle 42 is disposed between the host-vehicle 12 and the pedestrian 38 such that the lead-vehicle 42 blocks a portion 43 of the first-field-of-view 18 of the first-sensor 16 (i.e. the camera 24).
  • The controller 36 selects the second-sensor 14A to detect an object-of-interest 44 (i.e. the pedestrian 38) in accordance with a determination that an obstruction 46 (i.e. the lead-vehicle 42) blocks a first-line-of-sight 48 between the first-sensor 16 and the object-of-interest 44. In the example illustrated in FIG. 2, the second-sensor 14A is mounted on a front right-corner of the host-vehicle 12. The second-sensor 14A may be one of either the camera 24 or the ranging-sensor 26 that is in a different location (i.e. the second-location 28A) relative to the first-sensor 16. The obstruction 46 may be any obstruction 46 including a building, signage, or construction-barriers, for example. In the example illustrated in FIG. 2, the object-of-interest 44 is the pedestrian 38 in the roadway 40, but may be any object-of-interest 44, including a lane-marking 50, a traffic-intersection 52, an edge 54 of the roadway 40, or an other-vehicle 56.
  • When the object-of-interest 44 is the lane-marking 50 and/or the edge 54 of the roadway 40, the system 10 provides the technical benefits of improving a road-model and/or improving a lane-keeping-algorithm that may use these features of the roadway 40 as inputs. When the object-of-interest 44 is the traffic-intersection 52, the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is passing through the traffic-intersection 52 that is obscured by buildings, trucks, etc. In addition, the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is making a left-hand turn at the traffic-intersection 52 behind a large vehicle (e.g. a Class 8 truck) that may block oncoming traffic or may obscure a traffic-light.
  • The controller 36 is able to determine that the obstruction 46 blocks the first-line-of-sight 48 based on a tracking 58 of the object-of-interest 44 prior to the obstruction 46 blocking the first-line-of-sight 48. The specific process of tracking 58 used by the controller 36 may be dependent upon the sensor-type (i.e., camera 24, ranging-sensor 26) and will be understood by those in the art. The controller 36 performs the tracking 58 based initially on signals received by the first-sensor 16 while the object-of-interest 44 is exposed (i.e. visible, unobstructed first-line-of-sight 48, etc.) in the first-field-of-view 18. The obstruction 46 may cause an abrupt loss of tracking-data from the object-of-interest 44 to the first-sensor 16, whereby the controller 36 determines that the obstruction 46 is present. The controller 36 then selects the second-sensor 14 with an unobstructed second-line-of-sight 60 to the object-of-interest 44. The second-sensor 14 then tracks 58 the object-of-interest 44 in the second-field-of-view 34, wherein the second-sensor 14 detects at least a part of the object-of-interest 44. In the example illustrated in FIG. 2, the controller 36 selects the second-sensor 14A mounted on the host-vehicle 12 with the second-line-of-sight 60A.
  • Described another way, the controller 36 tracks 58 the pedestrian 38 while the pedestrian 38 is in plain view of the first-sensor 16. When the view of the pedestrian 38 to the first-sensor 16 is blocked by the lead-vehicle 42, the controller 36 determines that the associated loss of tracking-data from the pedestrian 38 is caused by the lead-vehicle 42. The controller 36 then selects the second-sensor 14A, that has the unobstructed view of the pedestrian 38, to resume the tracking 58 of the pedestrian 38. If the view of the second-sensor 14A to the pedestrian 38 is blocked, then the controller 36 may select another second-sensor (14B or 14C) that has the unobstructed view of the pedestrian 38.
  • FIG. 3 illustrates another traffic scenario where the second-vehicle 30 is at the traffic-intersection 52 and the second-sensor 14B is mounted on the second-vehicle 30 that has the unobstructed view of the object-of-interest 44 (i.e. the pedestrian 38). The second-sensor 14B may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16 (i.e. the second-location 28B). Data from the second-sensor 14B mounted on the second-vehicle 30 is both requested 64 and transferred 66 from the second-vehicle 30 to the host-vehicle 12 using a communications protocol, such as dedicated short range communications (DSRC), Bluetooth ®, cellular or other ad hoc protocols. The data is both requested 64 and transferred 66 from the second-vehicle 30 by way of a transceiver 68 (see FIG. 1) mounted on the host-vehicle 12 that is also in communication with the controller 36.
  • FIG. 4 illustrates yet another traffic scenario where the second-sensor 14C is part of the transportation-infrastructure 32 (e.g. mounted to a traffic light at the second-location 28C) located proximate the traffic-intersection 52, and has the unobstructed view of the object-of-interest 44. The second-sensor 14C may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16. Data from the second-sensor 14C that is part of the transportation-infrastructure 32 is both requested 64 and transferred 66 from the transportation-infrastructure 32 to the host-vehicle 12 using a communications protocol, such as DSRC, UHF, VHF, Bluetooth ®, cellular WiMAX, GSM, 3G or other ad hoc protocols. The data is both requested 64 and transferred 66 from the transportation-infrastructure 32 by way of the transceiver 68 mounted on the host-vehicle 12 that is also in communication with the controller 36, as illustrated in FIG. 1.
  • In the traffic scenarios illustrated in FIGS. 2-4, the second-sensor 14 may be a directional-sensor 72 where the second-field-of-view 34 (i.e. 34A-34C) may be adjusted to detect the object-of-interest 44. That is, the controller 36 may rotate the second-sensor 14 that is the directional-sensor 72 as necessary, and/or may change an angle-of-elevation of a bore-site of the second-sensor 14 as necessary to capture the object-of-interest 44. The controller 36 may adjust the second-field-of-view 34 as required to resume tracking 58 the object-of-interest 44.
  • FIG. 5 illustrates yet another traffic scenario with all three second-sensors 14A- 14 C 14C present. The controller 36 may further select the second-sensor 14 based on a proximity of the second-sensor 14 to both the host-vehicle 12 and the object-of-interest 44. The controller 36 polls (i.e. samples, surveys) the available second-sensors 14 within a communications-range via the second-sensor's 14 respective communications protocol and prioritizes the data based on the second-sensor's 14 point-of-view 74, sensor-distance 76 from the host-vehicle 12, and object-distance 78 to the object-of-interest 44. As used herein, the point-of-view 74 is the position of the second-sensor 14 from which the object-of-interest 44 is detected. The controller 36 may select the second-sensor 14 with the point-of-view 74 that provides a least obstructed view of the object-of-interest 44. For example, the second-sensor 14B mounted on the second-vehicle 30 provides the least obstructed point-of-view 74B of the pedestrian 38 compared to the point-of-view 74A of the second-sensor 14A mounted on the host-vehicle 12. Therefore, the second-sensor 14B may be assigned a higher priority than the second-sensor 14A. Similarly, the second-sensor 14C mounted to the traffic light may have the least obstructed view of the pedestrian 38 compared to the second-sensor 14B mounted on the second-vehicle 30 (possibly due to its greater elevation above the traffic-intersection 52), and may be assigned a higher priority than the second-sensor 14B.
  • The sensor-distance 76 from the host-vehicle 12 (i.e. the distance from the second- sensor 14B, 14C to the host-vehicle 12) is prioritized because a signal-strength and a signal-quality of the data transferred 66 by the second-sensor 14B-14C may be affected by the sensor-distance 76B, 76C. Therefore, the second- sensor 14B, 14C with the shortest sensor-distance 76B, 76C may be assigned a higher priority by the controller 36.
  • The object-distance 78 (i.e. the distance from the second-sensor 14 to the object-of-interest 44) is prioritized because a resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14 may be affected by the object-distance 78A, 78B, 78C. Therefore, the second-sensor 14 with the shortest object-distance 78A, 78B, 78C may be assigned a higher priority by the controller 36.
  • The controller 36 may also prioritize the second-sensor 14 by a sensor-type (i.e. the camera 24 or the ranging-sensor 26). The controller 36 may assign the higher priority to the second-sensor 14 that provides the higher resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14. The inherent resolutions of the various sensor-types will be understood by one skilled in the art.
  • FIG. 6 illustrates a method 200 of operating a detection system 10, hereafter referred to as the system 10, installed in a host-vehicle 12.
  • Step 202, DETECT OBJECTS IN FIRST-FIELD-OF-VIEW, includes detecting objects 22 in a first-field-of-view 18 with a first-sensor 16 mounted on the host-vehicle 12.
  • FIG. 1 illustrates a non-limiting example of the detection system 10 installed in the host-vehicle 12. As will be described in more detail below, the system 10 is an improvement over prior detection systems because the system 10 uses data from a second-sensor 14, remote from the first-sensor 16, when the first-field-of-view 18 of the first-sensor 16 is occluded, blocked, or otherwise obstructed.
  • The system 10 includes the first-sensor 16 positioned on the host-vehicle 12 at a first-location 20, such as on the front of the host-vehicle 12, or in the passenger-compartment of the host-vehicle 12. The first-sensor 16 detects objects 22 in the first- field-of-view 18, and is illustrated as a camera 24 in FIG. 1. Alternatively, the first-sensor 16 may be a ranging-sensor 26, such as a radar-sensor or a lidar-sensor. Examples of the camera 24 suitable for use on the host-vehicle 12 are commercially available as will be recognized by those in the art, one such being the APTINA MT9V023 from Micron Technology, Inc. of Boise, Id., USA. The camera 24 may be mounted in the first-location 20 that is on the front of the host-vehicle 12, or mounted in the interior of the host-vehicle 12 at a location suitable for the camera 24 to view the area around the host-vehicle 12 through the windshield of the host-vehicle 12. The camera 24 is preferably a video-type camera 24 or camera 24 that can capture images of the surrounding area at a sufficient frame-rate, of ten frames per second, for example.
  • Step 204, DETERMINE OBSTRUCTION PRESENT, includes determining whether an obstruction 46 blocks a first-line-of-sight 48 to an object-of-interest 44.
  • The system 10 also includes a controller 36 in communication with the first-sensor 16 and the second-sensor 14. The first-sensor 16 and the second-sensor 14 may be hard-wired to the controller 36 through the host-vehicle's 12 electrical-system (not shown), or may be in wireless communication with the controller 36. The controller 36 may include a processor (not shown) such as a microprocessor or other control circuitry such as analog and/or digital control circuitry including an application specific integrated circuit (ASIC) for processing data as should be evident to those in the art. The controller 36 may include a memory (not specifically shown), including non-volatile memory, such as electrically erasable programmable read-only memory (EEPROM) for storing one or more routines, thresholds, and captured data. The one or more routines may be executed by the processor to perform steps for detecting the objects 22 based on signals received by the controller 36 from the first-sensor 16 and the second-sensor 14 as described herein.
  • The controller 36 is able to determine that the obstruction 46 blocks the first-line-of-sight 48 based on a tracking 58 of the object-of-interest 44 prior to the obstruction 46 blocking the first-line-of-sight 48. The specific process of tracking 58 used by the controller 36 may be dependent upon the sensor-type (i.e., camera 24, ranging-sensor 26) and will be understood by those in the art. The controller 36 performs the tracking 58 based initially on signals received by the first-sensor 16 while the object-of-interest 44 is exposed (i.e. visible, unobstructed first-line-of-sight 48, etc.) in the first-field-of-view 18. The obstruction 46 may cause an abrupt loss of tracking-data from the object-of-interest 44 to the first-sensor 16, whereby the controller 36 determines that the obstruction 46 is present.
  • Step 206, DETECT OBJECTS IN SECOND-FIELD-OF-VIEW, includes detecting objects 22 in a second-field-of-view 34 with the second-sensor 14.
  • The system 10 also includes the second-sensor 14 positioned at a second-location 28 that is different than the first-location 20. The second-location 28 may be on the host-vehicle 12 (shown as 28A), on a second-vehicle 30 (shown as 28B), or part of a transportation-infrastructure 32 (shown as 28C), as will be described in more detail below. Alternatively, the second-vehicle 30 may be referred to as the alternative-vehicle 30. The second-sensor 14 detects objects 22 in a second-field-of-view 34 that at least partially overlaps the first-field-of-view 18. Depending on the amount of overlap of the second-field-of-view 34 with the first-field-of-view 18, the objects 22 in the second-field-of-view 34 may include the same or different objects 22 than the objects 22 in the first-field-of-view 18, as will become apparent with a reading of the descriptions of the traffic scenarios of FIGS. 2-5 below. The object 22 that is detected in both the first-field-of-view 18 and the second-field-of-view 34 may be determined to be the same object 22 based on the position of the object 22. That is, the position of the object 22 may have the same coordinates of latitude, longitude, and elevation in the first-field-of-view 18 as the coordinates in the second-field-of-view 34.
  • Step 208, SELECT SECOND-SENSOR, includes selecting the second-sensor 14 with the controller 36 to detect the object-of-interest 44 in accordance with a determination that the obstruction 46 blocks the first-line-of-sight 48 between the first-sensor 16 and the object-of-interest 44.
  • FIG. 2 illustrates a traffic scenario where the host-vehicle 12 equipped with the system 10 is approaching a crosswalk on a roadway 40 and a pedestrian 38 is also attempting to cross the roadway 40 in the crosswalk. A lead-vehicle 42 is disposed between the host-vehicle 12 and the pedestrian 38 such that the lead-vehicle 42 blocks a portion 43 of the first-field-of-view 18 of the first-sensor 16 (i.e. the camera 24).
  • The controller 36 selects the second-sensor 14A to detect the object-of-interest 44 (i.e. the pedestrian 38) in accordance with the determination that the obstruction 46 (i.e. the lead-vehicle 42) blocks the first-line-of-sight 48 between the first-sensor 16 and the object-of-interest 44. In the example illustrated in FIG. 2, the second-sensor 14A is mounted on a front right-corner of the host-vehicle 12. The second-sensor 14A may be one of either the camera 24 or the ranging-sensor 26 that is in a different location (i.e. the second-location 28A) relative to the first-sensor 16. The obstruction 46 may be any obstruction 46 including a building, signage, or construction-barriers, for example. In the example illustrated in FIG. 2, the object-of-interest 44 is the pedestrian 38 in the roadway 40, but may be any object-of-interest 44, including a lane-marking 50, a traffic-intersection 52, an edge 54 of the roadway 40, or an other-vehicle 56.
  • When the object-of-interest 44 is the lane-marking 50 and/or the edge 54 of the roadway 40, the system 10 provides the technical benefits of improving a road-model and/or improving a lane-keeping-algorithm that may use these features of the roadway 40 as inputs. When the object-of-interest 44 is the traffic-intersection 52, the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is passing through the traffic-intersection 52 that is obscured by buildings, trucks, etc. In addition, the system 10 provides the technical benefit of improving safety when the host-vehicle 12 is making a left-hand turn at the traffic-intersection 52 behind a large vehicle (e.g. a Class 8 truck) that may block oncoming traffic or may obscure a traffic-light.
  • The controller 36 selects the second-sensor 14 with an unobstructed second-line-of-sight 60 to the object-of-interest 44. The second-sensor 14 then tracks 58 the object-of-interest 44 in the second-field-of-view 34, wherein the second-sensor 14 detects at least a part of the object-of-interest 44. In the example illustrated in FIG. 2, the controller 36 selects the second-sensor 14A mounted on the host-vehicle 12 with the second-line-of-sight 60A.
  • Described another way, the controller 36 tracks 58 the pedestrian 38 while the pedestrian 38 is in plain view of the first-sensor 16. When the view of the pedestrian 38 to the first-sensor 16 is blocked by the lead-vehicle 42, the controller 36 determines that the associated loss of tracking-data from the pedestrian 38 is caused by the lead-vehicle 42. The controller 36 then selects the second-sensor 14A, that has the unobstructed view of the pedestrian 38, to resume the tracking 58 of the pedestrian 38. If the view of the second-sensor 14A to the pedestrian 38 is blocked, then the controller 36 may select another second-sensor (14B or 14C) that has the unobstructed view of the pedestrian 38.
  • FIG. 3 illustrates another traffic scenario where the second-vehicle 30 is at the traffic-intersection 52 and the second-sensor 14B is mounted on the second-vehicle 30 that has the unobstructed view of the object-of-interest 44 (i.e. the pedestrian 38). The second-sensor 14B may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16 (i.e. the second-location 28B). Data from the second-sensor 14B mounted on the second-vehicle 30 is both requested 64 and transferred 66 from the second-vehicle 30 to the host-vehicle 12 using a communications protocol, such as dedicated short range communications (DSRC), Bluetooth ®, cellular or other ad hoc protocols. The data is both requested 64 and transferred 66 from the second-vehicle 30 by way of a transceiver 68 (see FIG. 1) mounted on the host-vehicle 12 that is also in communication with the controller 36.
  • FIG. 4 illustrates yet another traffic scenario where the second-sensor 14C is part of a transportation-infrastructure 32 (e.g. mounted to a traffic light at the second-location 28C) located proximate the traffic-intersection 52, and has the unobstructed view of the object-of-interest 44. The second-sensor 14C may be one of either the camera 24 or the ranging-sensor 26 that is in a different location relative to the first-sensor 16. Data from the second-sensor 14C that is part of the transportation-infrastructure 32 is both requested 64 and transferred 66 from the transportation-infrastructure 32 to the host-vehicle 12 using a communications protocol, such as DSRC, UHF, VHF, Bluetooth ®, cellular WiMAX, GSM, 3G or other ad hoc protocols. The data is both requested 64 and transferred 66 from the transportation-infrastructure 32 by way of the transceiver 68 mounted on the host-vehicle 12 that is also in communication with the controller 36, as illustrated in FIG. 1.
  • In the traffic scenarios illustrated in FIGS. 2-4, the second-sensor 14 may be a directional-sensor 72 where the second-field-of-view 34 (i.e. 34A-34C) may be adjusted to detect the object-of-interest 44. That is, the controller 36 may rotate the second-sensor 14 that is the directional-sensor 72 as necessary, and/or may change an angle-of-elevation of a bore-site of the second-sensor 14 as necessary to capture the object-of-interest 44. The controller 36 may adjust the second-field-of-view 34 as required to resume tracking 58 the object-of-interest 44.
  • FIG. 5 illustrates yet another traffic scenario with all three second-sensors 14A-14C present. The controller 36 may further select the second-sensor 14 based on a proximity of the second-sensor 14 to both the host-vehicle 12 and the object-of-interest 44. The controller 36 polls (i.e. samples, surveys) the available second-sensors 14 within a communications-range via the second-sensor's 14 respective communications protocol and prioritizes the data based on the second-sensor's 14 point-of-view 74, sensor-distance 76 from the host-vehicle 12, and object-distance 78 to the object-of-interest 44. As used herein, the point-of-view 74 is the position of the second-sensor 14 from which the object-of-interest 44 is detected. The controller 36 may select the second-sensor 14 with the point-of-view 74 that provides a least obstructed view of the object-of-interest 44. For example, the second-sensor 14B mounted on the second-vehicle 30 provides the least obstructed point-of-view 74B of the pedestrian 38 compared to the point-of-view 74A of the second-sensor 14A mounted on the host-vehicle 12. Therefore, the second-sensor 14B may be assigned a higher priority than the second-sensor 14A. Similarly, the second-sensor 14C mounted to the traffic light may have the least obstructed view of the pedestrian 38 compared to the second-sensor 14B mounted on the second-vehicle 30 (possibly due to its greater elevation above the traffic-intersection 52), and may be assigned a higher priority than the second-sensor 14B.
  • The sensor-distance 76 from the host-vehicle 12 (i.e. the distance from the second- sensor 14B, 14C to the host-vehicle 12) is prioritized because a signal-strength and a signal-quality of the data transferred 66 by the second-sensor 14B-14C may be affected by the sensor-distance 76B, 76C. Therefore, the second- sensor 14B, 14C with the shortest sensor-distance 76B, 76C may be assigned a higher priority by the controller 36.
  • The object-distance 78 (i.e. the distance from the second-sensor 14 to the object-of-interest 44) is prioritized because a resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14 may be affected by the object-distance 78A, 78B, 78C. Therefore, the second-sensor 14 with the shortest object-distance 78A, 78B, 78C may be assigned a higher priority by the controller 36.
  • The controller 36 may also prioritize the second-sensor 14 by a sensor-type (i.e. the camera 24 or the ranging-sensor 26). The controller 36 may assign the higher priority to the second-sensor 14 that provides the higher resolution of the object-of-interest 44 depicted in the image produced by the second-sensor 14. The inherent resolutions of the various sensor-types will be understood by one skilled in the art.
  • Accordingly, a detection system 10 (the system 10), a controller 36 for the system 10, and a method 200 of operating the system 10 are provided. The system 10 is beneficial because the system 10 uses the second-sensor 14 to track 58 the object-of-interest 44 when the obstruction 46 blocks the first-line-of-sight 48 of the first-sensor 16.
  • While this invention has been described in terms of the preferred embodiments thereof, it is not intended to be so limited, but rather only to the extent set forth in the claims that follow.

Claims (30)

We claim:
1. A detection system, comprising:
a first-sensor positioned at a first-location on a host-vehicle, the first-sensor detecting objects in a first-field-of-view;
a second-sensor positioned at a second-location different than the first-location, said second-sensor detecting objects in a second-field-of-view that at least partially overlaps the first-field-of-view; and
a controller in communication with the first-sensor and the second-sensor, the controller selecting the second-sensor to detect an object-of-interest in accordance with a determination that an obstruction blocks a first-line-of-sight between the first-sensor and the object-of-interest.
2. The detection system in accordance with claim 1, wherein the second-sensor is mounted on the host-vehicle.
3. The detection system in accordance with claim 1, wherein the second-sensor is mounted on a second-vehicle.
4. The detection system in accordance with claim 3, wherein data from the second-sensor is both requested and transferred from the second-vehicle to the host- vehicle using a communications protocol.
5. The detection system in accordance with claim 1, wherein the second-sensor is part of a transportation-infrastructure.
6. The detection system in accordance with claim 5, wherein data from the second-sensor is requested from the transportation-infrastructure by the host-vehicle using a communications protocol.
7. The detection system in accordance with claim 5, wherein data from the second-sensor is transferred from the transportation-infrastructure to the host-vehicle using a communications protocol.
8. The detection system in accordance with claim 1, wherein the second-sensor is directional and the second-field-of-view is adjusted to detect the object-of-interest.
9. The detection system in accordance with claim 1, wherein the controller selects the second-sensor based on a proximity of the second-sensor to both the host-vehicle and the object-of-interest.
10. The detection system in accordance with claim 9, wherein the controller polls available sensors within a communications-range and prioritizes data from the available sensors by a point-of-view of the second-sensor, a sensor-distance of the second-sensor from the host-vehicle, and an object-distance of the second-sensor to the object-of-interest.
11. The detection system in accordance with claim 1, wherein the object-of-interest is a pedestrian.
12. The detection system in accordance with claim 1, wherein the object-of-interest is an other-vehicle.
13. The detection system in accordance with claim 1, wherein the object-of-interest is a lane-marking.
14. The detection system in accordance with claim 1, wherein the object-of-interest is an edge of a roadway.
15. The detection system in accordance with claim 1, wherein the object-of-interest is a traffic-intersection.
16. A detection method, comprising:
detecting objects in a first-field-of-view with a first-sensor, the first-sensor positioned at a first-location on a host-vehicle;
detecting objects in a second-field-of-view with a second-sensor, the second-sensor positioned at a second-location different than the first-location, the second-field-of-view at least partially overlaps the first-field-of-view; and
selecting the second-sensor, with a controller in communication with the first-sensor and the second-sensor, to detect an object-of-interest in accordance with a determination that an obstruction blocks a first-line-of-sight between the first-sensor and the object-of-interest.
17. The detection method in accordance with claim 16, wherein the second-sensor is mounted on the host-vehicle.
18. The detection method in accordance with claim 16, wherein the second-sensor is mounted on a second-vehicle.
19. The detection method in accordance with claim 18, wherein data from the second-sensor is both requested and transferred from the second-vehicle to the host-vehicle using a communications protocol.
20. The detection method in accordance with claim 16, wherein the second-sensor is part of a transportation-infrastructure.
21. The detection method in accordance with claim 20, wherein data from the second-sensor is requested from the transportation-infrastructure by the host-vehicle using a communications protocol.
22. The detection method in accordance with claim 20, wherein data from the second-sensor is transferred from the transportation-infrastructure to the host-vehicle using a communications protocol.
23. The detection method in accordance with claim 16, wherein the second-sensor is directional and the second-field-of-view is adjusted to detect the object-of-interest.
24. The detection method in accordance with claim 16, wherein the controller selects the second-sensor based on a proximity of the second-sensor to both the host-vehicle and the object-of-interest.
25. The detection method in accordance with claim 24, wherein the controller polls available sensors within a communications-range and prioritizes data from the available sensors by a point-of-view of the second-sensor, a sensor-distance of the second-sensor from the host-vehicle, and an object-distance of the second-sensor to the object-of-interest.
26. The detection method in accordance with claim 16, wherein the object-of-interest is a pedestrian.
27. The detection method in accordance with claim 16, wherein the object-of-interest is an other-vehicle.
28. The detection method in accordance with claim 16, wherein the object-of-interest is a lane-marking.
29. The detection method in accordance with claim 16, wherein the object-of-interest is an edge of a roadway.
30. The detection method in accordance with claim 16, wherein the object-of-interest is a traffic-intersection.
US15/866,188 2018-01-09 2018-01-09 Detection system and method Active 2038-11-29 US11488476B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/866,188 US11488476B2 (en) 2018-01-09 2018-01-09 Detection system and method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/866,188 US11488476B2 (en) 2018-01-09 2018-01-09 Detection system and method

Publications (2)

Publication Number Publication Date
US20190213884A1 true US20190213884A1 (en) 2019-07-11
US11488476B2 US11488476B2 (en) 2022-11-01

Family

ID=67140917

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/866,188 Active 2038-11-29 US11488476B2 (en) 2018-01-09 2018-01-09 Detection system and method

Country Status (1)

Country Link
US (1) US11488476B2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200068044A1 (en) * 2018-02-01 2020-02-27 Toyota Jidosha Kabushiki Kaisha Vehicular millimeter wave communication for non-line of sight scenarios
US10882521B2 (en) * 2018-02-21 2021-01-05 Blackberry Limited Method and system for use of sensors in parked vehicles for traffic safety
WO2021069568A1 (en) * 2019-10-10 2021-04-15 Starship Technologies Oü Device, system and method for assisting mobile robots in autonomously crossing roads
US11280897B2 (en) * 2019-03-31 2022-03-22 Waymo Llc Radar field of view extensions
US20220324441A1 (en) * 2019-09-27 2022-10-13 Intel Corporation Potential collision warning system based on road user intent prediction
US20230095194A1 (en) * 2021-09-30 2023-03-30 AyDeeKay LLC dba Indie Semiconductor Dynamic and Selective Pairing Between Proximate Vehicles
WO2024060575A1 (en) * 2022-09-19 2024-03-28 智道网联科技(北京)有限公司 Road side unit data processing method and apparatus, electronic device, and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180047287A1 (en) * 2016-08-10 2018-02-15 Panasonic Intellectual Property Corporation Of America Communication method and server
US20180101736A1 (en) * 2016-10-11 2018-04-12 Samsung Electronics Co., Ltd. Method for providing a sight securing image to vehicle, electronic apparatus and computer readable recording medium therefor
US20180107225A1 (en) * 2016-10-17 2018-04-19 Denso International America, Inc. Sensor System for Vehicle

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111727413A (en) * 2017-12-27 2020-09-29 直流电资本有限责任公司 Method for accessing supplemental perception data from other vehicles

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180047287A1 (en) * 2016-08-10 2018-02-15 Panasonic Intellectual Property Corporation Of America Communication method and server
US20180101736A1 (en) * 2016-10-11 2018-04-12 Samsung Electronics Co., Ltd. Method for providing a sight securing image to vehicle, electronic apparatus and computer readable recording medium therefor
US20180107225A1 (en) * 2016-10-17 2018-04-19 Denso International America, Inc. Sensor System for Vehicle

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200068044A1 (en) * 2018-02-01 2020-02-27 Toyota Jidosha Kabushiki Kaisha Vehicular millimeter wave communication for non-line of sight scenarios
US11057497B2 (en) * 2018-02-01 2021-07-06 Toyota Jidosha Kabushiki Kaisha Vehicular millimeter wave communication for non-line of sight scenarios
US10882521B2 (en) * 2018-02-21 2021-01-05 Blackberry Limited Method and system for use of sensors in parked vehicles for traffic safety
US11280897B2 (en) * 2019-03-31 2022-03-22 Waymo Llc Radar field of view extensions
US20220324441A1 (en) * 2019-09-27 2022-10-13 Intel Corporation Potential collision warning system based on road user intent prediction
WO2021069568A1 (en) * 2019-10-10 2021-04-15 Starship Technologies Oü Device, system and method for assisting mobile robots in autonomously crossing roads
US20230095194A1 (en) * 2021-09-30 2023-03-30 AyDeeKay LLC dba Indie Semiconductor Dynamic and Selective Pairing Between Proximate Vehicles
WO2024060575A1 (en) * 2022-09-19 2024-03-28 智道网联科技(北京)有限公司 Road side unit data processing method and apparatus, electronic device, and storage medium

Also Published As

Publication number Publication date
US11488476B2 (en) 2022-11-01

Similar Documents

Publication Publication Date Title
US11488476B2 (en) Detection system and method
US11772489B2 (en) Visually obstructed object detection for automated vehicle using V2V/V2I communications
CN108454631B (en) Information processing apparatus, information processing method, and recording medium
US10885789B2 (en) Device and method for lateral guidance assistance for a road vehicle
US11639174B2 (en) Automated speed control system
US8854458B2 (en) Object detection device
US20150293534A1 (en) Vehicle control system and method
US10466706B2 (en) Automated guidance system
US11099264B2 (en) Variable range and frame-rate radar operation for automated vehicle
JP2008250503A (en) Operation support device
US20170349181A1 (en) Lane management system for an automated vehicle
EP3524935B1 (en) Vehicle perception-data gathering system and method
CN108974007B (en) Determining an object of interest for active cruise control
US20180362038A1 (en) Device for Lateral Guidance Assistance for a Road Vehicle
US11519735B2 (en) Vehicle navigation system and method
CN110120158B (en) Traffic control system
US20190281260A1 (en) Vehicle sensor configuration based on map data
JP7015665B2 (en) Information processing equipment, information processing methods and programs
US11143760B2 (en) Object-detector configuration based on human-override of automated vehicle control
US20190236949A1 (en) Traffic control system

Legal Events

Date Code Title Description
AS Assignment

Owner name: DELPHI TECHNOLOGIES, LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KIM, JUNSUNG;WEI, JUNQING;XU, WENDA;AND OTHERS;REEL/FRAME:044577/0356

Effective date: 20180103

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

AS Assignment

Owner name: APTIV TECHNOLOGIES LIMITED, BARBADOS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DELPHI TECHNOLOGIES LLC;REEL/FRAME:052044/0428

Effective date: 20180101

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

AS Assignment

Owner name: MOTIONAL AD LLC, MASSACHUSETTS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:APTIV TECHNOLOGIES LIMITED;REEL/FRAME:053863/0399

Effective date: 20200917

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE