WO2020263982A1 - Image based locationing - Google Patents

Image based locationing Download PDF

Info

Publication number
WO2020263982A1
WO2020263982A1 PCT/US2020/039362 US2020039362W WO2020263982A1 WO 2020263982 A1 WO2020263982 A1 WO 2020263982A1 US 2020039362 W US2020039362 W US 2020039362W WO 2020263982 A1 WO2020263982 A1 WO 2020263982A1
Authority
WO
WIPO (PCT)
Prior art keywords
motion
estimates
orientation
image
images
Prior art date
Application number
PCT/US2020/039362
Other languages
French (fr)
Inventor
Erik Volkerink
Ajay Khoche
Original Assignee
Trackonomy Systems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Trackonomy Systems, Inc. filed Critical Trackonomy Systems, Inc.
Publication of WO2020263982A1 publication Critical patent/WO2020263982A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/285Analysis of motion using a sequence of stereo image pairs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/147Details of sensors, e.g. sensor lenses
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/751Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • G06T2207/10021Stereoscopic video; Stereoscopic image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30244Camera pose
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30252Vehicle exterior; Vicinity of vehicle

Definitions

  • Locationing systems can track mobile targets in real time. These systems typically ascertain information relating to their geographic locations based on communications with a variety of different wireless locationing systems (e.g., the Global Positioning System (GPS), cellular network systems (e.g., GSM), and wireless local area networks (e.g., a system of Wi-Fi access points).
  • GPS Global Positioning System
  • GSM cellular network systems
  • Wi-Fi access points wireless local area networks
  • No single approach provides continuous tracking information under all circumstances. For example, GPS tracking requires a tracking device to have an unobstructed view of at least four GPS satellites at the same time, making GPS tracking in urban and indoor environments problematic. Dead reckoning may be used to supplement GPS locationing when GPS signals are unavailable or inaccurate (e.g., as a result of signal multipath error).
  • dead-reckoning navigation is limited by the rapid accumulation of errors and, requires a complex fusion process to integrate dead-reckoning navigation data with GPS navigation data.
  • Map-matching techniques can improve locationing accuracy by identifying the most likely locations of a vehicle on a road network.
  • the accuracy of map-matching techniques depends significantly on the accuracy of the position estimates for the mobile target being tracked and the fidelity of the spatial road map used to locate the mobile target in a geographic region.
  • FIGS. 1 A and 1 B respectively show diagrammatic side and bottom views of an example vehicle that includes a locationing system on the vehicle chassis.
  • FIG. 2 shows a diagrammatic view of an example locationing system.
  • FIG. 3 is a diagrammatic view of one or more light sources illuminating a surface and one or more imaging devices capturing images of the illuminated surface.
  • FIG. 4 shows components of an example of the locationing system shown in FIGS. 1A and 1 B.
  • FIG. 5 is a flow diagram of an example image based locationing method.
  • FIG. 6A is a diagrammatic view of an example of an imaged region divided into a set of parallel tracks.
  • FIG. 6B is a diagrammatic view of an example of an imaged region divided into an array of tracks.
  • FIG. 7 is a flow diagram of an example process of estimating motion during a time step based on the positions determined for multiple tracks.
  • FIG. 8 shows one or more locationing systems located at different locations on an example vehicle.
  • FIG. 9 shows an example vehicle that includes multiple locationing systems located on the vehicle chassis.
  • FIGS. 10A and 10B respectively show examples of different systems for maintaining the performance of the locationing system.
  • This disclosure relates to systems and methods of obtaining accurate motion and orientation estimates for a vehicle traveling at high speed based on images of a road surface.
  • a purpose of these systems and methods is to provide a
  • FIGS. 1A and 1 B show an example of a vehicle 10 that includes an example locationing system 12 on the chassis 14 of the vehicle 10.
  • the locationing system 12 may be integrated with the vehicle 10 during manufacture or it may be an add-on component that is retrofitted to the vehicle 10.
  • the locationing system 12 may be located anywhere on the chassis 14 that provides a vantage point from which the locationing system 12 can capture images of the road surface.
  • the locationing system 12 is configured to simultaneously capture images of both the road surface and at least one wheel (e.g., wheel 16) so that wheel slippage can be optically detected and incorporated into the locationing process.
  • the locationing system 12 includes at least one light source 18, at least one image capture device 20, a data processing system 22, and a data storage system 24.
  • the data processing system 22 is implemented by or integrated with the central control system of the vehicle.
  • the data processing system 22 is implemented, at least in part, by a remote data processing system or service that communicates with the locationing system 12.
  • the light source 18 typically is a high brightness illumination source, such as a light emitting diode (LED) that emits light 19 of a particular wavelength towards a road surface 25.
  • the light source 18 emits light in the infrared wavelength range (e.g., 750-1400 nm). In other examples, different wavelengths of light may be used.
  • the image capture device 20 may include one or more imaging devices 26 that are capable of capturing images of a road surface at a high rate (e.g., 750-1000 frames per second). In some examples, the images are captured as grayscale images. In some examples (see FIG. 4), the image capture device 20 includes two stereoscopic imaging devices 28, 30 that are configured to capture a respective frame that includes at least two simultaneous stereoscopic images 38 for each image capture event. In some examples, a trigger signal or a global shutter 32 synchronizes the capture of the image frames by the imaging devices 28, 30.
  • the image capture device 20 also may include one or more lenses 34 (e.g., fixed lenses or automatic zoom lenses), one or more optical filters 36 at least one of which may be matched to pass a wavelength of light generated by the light source 18, and an automatic exposure control system.
  • the various components of the light source 18 and the image capture device 20 may be integrated and configured to optimize one or more parameters of the captured images (e.g., contrast).
  • the data processing system 22 processes the images captured by the image capture device 20 to determine an image-based trajectory of the vehicle along a road.
  • the image capture device 20 consists of a monocular camera with an optional ranging sensor that enables vehicle motion and depth to be calculated in terms of pixel motion across successive images.
  • the image capture device 20 includes two or more cameras that are fixed in position in a known geometry, which allows vehicle motion and depth to be calculated in real world dimensions (e.g., meters).
  • the data processing system 22 processes a sequence of frames of stereoscopic image pairs that are captured at different times (e.g., at equal time intervals) to determine respective estimates of three- dimensional points on the road surface.
  • the data processing system 22 rectifies each stereo image pair to a common epipolar plane (FIG. 5, block 40).
  • the data processing system 22 may use any of a wide variety of methods to rectify the stereo image pair in each frame.
  • the resulting rectified images have epipolar lines that are parallel to the horizontal axis and corresponding points have identical vertical coordinates.
  • the images of the stereo pair are not grayscale images, the images optionally may be converted to a grayscale format (FIG. 5, block 42).
  • the data processing system 22 detects features in each of the rectified stereo pair of images in each successive frame (FIG. 5, block 44).
  • features are points in an image that can be uniquely identified based on comparisons of colors, highlights, and other features in the pair of images. Examples include points that have high contrast with its neighbors in a local region of an image.
  • the data processing system 22 matches corresponding features in the rectified stereo pair of images (FIG. 5, block 46).
  • the processing system 22 matches points (e.g., pixels) or features in one image of the stereo pair with corresponding interest points or features in the other image of the stereo pair.
  • the result of this process is a feature disparity map, where the disparity is the pixel displacement between the matched features.
  • the depth at each feature point is calculated to obtain a sparse three-dimensional surface profile (FIG. 5, block 48).
  • the data processing system 22 produces a depth map by calculating an elevation value for each feature in the disparity map using triangulation.
  • the data processing system 22 determines the motion of the vehicle (e.g., orientation and translation) between successive frames (FIG. 5, block 50).
  • an optical flow process is used to determine the motion and orientation of the vehicle between successive stereoscopic image frames.
  • a stereoscopic image set in a frame is used to estimate the current trajectory of the vehicle.
  • the imaged region of the road surface under the vehicle is divided into multiple tracks that can be processed to obtain multiple independent estimates of the vehicle’s trajectory. Multiple of these estimates can be combined or otherwise used to determine estimates of the vehicle’s motion and orientation between successive stereoscopic image frames.
  • the multiple independent sets of estimates can be used to identify and reject as outliers motion and orientation estimates that are inconsistent with the majority of other estimates.
  • an imaged road region 60 is divided into three parallel tracks 62, 64, 66, each of which is processed as an
  • Each track 62, 64, 66 is processed
  • the data processing system 22 can determine a respective set of disparity and depth maps for each track 62, 64, 66.
  • FIG. 6B shows another example, in which the imaged road region 68 is divided into a rectangular array of six tracks 68, 70, 72, 74, 76, 78, each of which can be processed to obtain respective independent sets of estimates of motion and orientation along the vehicle’s trajectory.
  • the imaged region 60 is divided into the tracks 62- 66 and 68-78 by capturing a stereoscopic image set and windowing or bucketing the frame into regions that correspond to the sets of tracks 62-66 and 68-78.
  • the imaged region 60 is divided into the sets of tracks 62-66 and 68-78 using lenses or diffractive optical elements that are configured to direct the light reflecting from the road surface to respective ones of the sets of tracks 62-68 and 68-78.
  • the imaged region 60 is divided into the sets of tracks 62-66 and 68-78 using respective pairs of stereoscopic image devices 28, 30 (e.g., cameras) arrayed according to the desired track layout.
  • FIG. 7 shows an example method of determining estimates of the motion and orientation of a vehicle from successive stereoscopic image frames.
  • the data processing system 22 estimates the motion and orientation between corresponding features in a respective stereoscopic pair images (FIG. 7, block 80).
  • the data processing system 22 removes inconsistent motion and orientation estimates as outliers (FIG. 7, block 82).
  • Estimates of the motion and orientation are determined based on an aggregation of the consistent motion and orientation estimates (FIG. 7, block 84). In some examples, the consistent motion and orientation estimates are averaged to determine the motion and orientation estimates.
  • FIG. 8 shows an example of a vehicle 85, in which one or more image- based locationing systems 86, 88 are mounted on the front and/or rear ends of the vehicle 10 and oriented to obtain images of the road surface.
  • FIG. 9 shows an example of a vehicle 96 in which first and second imaging devices 90, 92 and a light source 94 are located on the chassis of the vehicle 96 at different locations along the longitudinal dimension of the vehicle 10.
  • Each of the image capture devices 90, 92 is operable to obtain a series of successive stereoscopic image frames of the light reflecting from the road surface.
  • the data processing system 22 is operable to determine the vehicle’s direction of travel based on the locations of features detected in the images captured by the first imaging device 90 and the locations of matching features detected in the images captured by the second imaging device 92. By increasing the distance between the complementary images, this approach increases the accuracy of measured changes in direction.
  • the first image capture device 90 is located under the vehicle 96 on the chassis and the second image capture device 92 is located on the roof of the vehicle 96.
  • FIG. 10A shows an example of a vehicle 98 that includes a cleaning system 100 for the locationing system 12.
  • the cleaning system 100 includes a funnel that includes a large input end 102 and a small output end 104.
  • Movement of the vehicle 98 over the road causes air 106 to flow into the input end 102 of the funnel 100 at a first rate that depends at least in part on the speed of the vehicle 98.
  • the optical cleaning system 100 is designed so that , under typical driving conditions, the input air is ejected from the output end 104 of the funnel 100 at pressure that is sufficient to remove dust and other debris from the optical and other components of the locationing system 12.
  • FIG. 10B shows an example of a vehicle that includes a cleaning system 108 for the locationing system 12.
  • the cleaning system 108 includes an electrically powered cylindrical brush that is configured to remove dust and other debris from the optical and other components of the locationing system 12.
  • the rotating brush is configured to move into and out of contact with the locationing system components either on a regular basis (e.g., each time the vehicle is turned on) or on demand (e.g., when a sensor detects that the optical and/or other components require cleaning).
  • the data processing system 22 also integrates data received from one or more auxiliary locationing sensors 38 into the locationing process.
  • Example auxiliary locationing sensors 38 include a wheel encoder, an inertial measurement unit (IMU), an inertial navigation system (INS), a global positioning system (GPS), a sound navigation and ranging (SONAR) system, a LIDAR system, and a radar system.
  • Examples of the subject matter described herein can be implemented in data processing apparatus (e.g., computer hardware and digital electronic circuitry) operable to perform functions by operating on input and generating output. Examples of the subject matter described herein also can be tangibly embodied in software or firmware, as one or more sets of computer instructions encoded on one or more tangible non-transitory carrier media (e.g., a machine readable storage device, substrate, or sequential access memory device) for execution by data processing apparatus.
  • data processing apparatus e.g., computer hardware and digital electronic circuitry
  • Examples of the subject matter described herein also can be tangibly embodied in software or firmware, as one or more sets of computer instructions encoded on one or more tangible non-transitory carrier media (e.g., a machine readable storage device, substrate, or sequential access memory device) for execution by data processing apparatus.
  • tangible non-transitory carrier media e.g., a machine readable storage device, substrate, or sequential access memory device

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Vascular Medicine (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

This disclosure relates to systems and methods of obtaining accurate motion and orientation estimates for a vehicle (10) traveling at high speed based on images of a road surface. A purpose of these systems and methods is to provide a supplementary or alternative means of locating a vehicle (10) on a map, particularly in cases where other locationing approaches (e.g., GPS) are unreliable or unavailable.

Description

IMAGE BASED LOCATIONING
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims the benefit of U.S. Provisional Application No. 62/865,192, filed 22 June 2019, which is incorporated herein by reference.
BACKGROUND
[0002] Locationing systems can track mobile targets in real time. These systems typically ascertain information relating to their geographic locations based on communications with a variety of different wireless locationing systems (e.g., the Global Positioning System (GPS), cellular network systems (e.g., GSM), and wireless local area networks (e.g., a system of Wi-Fi access points). No single approach, however, provides continuous tracking information under all circumstances. For example, GPS tracking requires a tracking device to have an unobstructed view of at least four GPS satellites at the same time, making GPS tracking in urban and indoor environments problematic. Dead reckoning may be used to supplement GPS locationing when GPS signals are unavailable or inaccurate (e.g., as a result of signal multipath error).
However, dead-reckoning navigation is limited by the rapid accumulation of errors and, requires a complex fusion process to integrate dead-reckoning navigation data with GPS navigation data. Map-matching techniques can improve locationing accuracy by identifying the most likely locations of a vehicle on a road network. However, the accuracy of map-matching techniques depends significantly on the accuracy of the position estimates for the mobile target being tracked and the fidelity of the spatial road map used to locate the mobile target in a geographic region.
DESCRIPTION OF DRAWINGS
[0003] FIGS. 1 A and 1 B respectively show diagrammatic side and bottom views of an example vehicle that includes a locationing system on the vehicle chassis.
[0004] FIG. 2 shows a diagrammatic view of an example locationing system.
[0005] FIG. 3 is a diagrammatic view of one or more light sources illuminating a surface and one or more imaging devices capturing images of the illuminated surface.
[0006] FIG. 4 shows components of an example of the locationing system shown in FIGS. 1A and 1 B. [0007] FIG. 5 is a flow diagram of an example image based locationing method.
[0008] FIG. 6A is a diagrammatic view of an example of an imaged region divided into a set of parallel tracks.
[0009] FIG. 6B is a diagrammatic view of an example of an imaged region divided into an array of tracks.
[0010] FIG. 7 is a flow diagram of an example process of estimating motion during a time step based on the positions determined for multiple tracks.
[0011] FIG. 8 shows one or more locationing systems located at different locations on an example vehicle.
[0012] FIG. 9 shows an example vehicle that includes multiple locationing systems located on the vehicle chassis.
[0013] FIGS. 10A and 10B respectively show examples of different systems for maintaining the performance of the locationing system.
DETAILED DESCRIPTION
[0014] In the following description, like reference numbers are used to identify like elements. Furthermore, the drawings are intended to illustrate major features of exemplary embodiments in a diagrammatic manner. The drawings are not intended to depict every feature of actual embodiments nor relative dimensions of the depicted elements, and are not drawn to scale.
[0015] This disclosure relates to systems and methods of obtaining accurate motion and orientation estimates for a vehicle traveling at high speed based on images of a road surface. A purpose of these systems and methods is to provide a
supplementary or alternative means of locating a vehicle on a map, particularly in cases where other locationing approaches (e.g., GPS) are unreliable or unavailable.
[0016] FIGS. 1A and 1 B show an example of a vehicle 10 that includes an example locationing system 12 on the chassis 14 of the vehicle 10. The locationing system 12 may be integrated with the vehicle 10 during manufacture or it may be an add-on component that is retrofitted to the vehicle 10. In the illustrated example, the locationing system 12 may be located anywhere on the chassis 14 that provides a vantage point from which the locationing system 12 can capture images of the road surface. In some examples, the locationing system 12 is configured to simultaneously capture images of both the road surface and at least one wheel (e.g., wheel 16) so that wheel slippage can be optically detected and incorporated into the locationing process.
[0017] In the example shown in FIG. 2, the locationing system 12 includes at least one light source 18, at least one image capture device 20, a data processing system 22, and a data storage system 24. In some examples, the data processing system 22 is implemented by or integrated with the central control system of the vehicle. In other examples, the data processing system 22 is implemented, at least in part, by a remote data processing system or service that communicates with the locationing system 12.
[0018] Referring to FIG. 3, the light source 18 typically is a high brightness illumination source, such as a light emitting diode (LED) that emits light 19 of a particular wavelength towards a road surface 25. In some examples, the light source 18 emits light in the infrared wavelength range (e.g., 750-1400 nm). In other examples, different wavelengths of light may be used.
[0019] The image capture device 20 may include one or more imaging devices 26 that are capable of capturing images of a road surface at a high rate (e.g., 750-1000 frames per second). In some examples, the images are captured as grayscale images. In some examples (see FIG. 4), the image capture device 20 includes two stereoscopic imaging devices 28, 30 that are configured to capture a respective frame that includes at least two simultaneous stereoscopic images 38 for each image capture event. In some examples, a trigger signal or a global shutter 32 synchronizes the capture of the image frames by the imaging devices 28, 30. The image capture device 20 also may include one or more lenses 34 (e.g., fixed lenses or automatic zoom lenses), one or more optical filters 36 at least one of which may be matched to pass a wavelength of light generated by the light source 18, and an automatic exposure control system. The various components of the light source 18 and the image capture device 20 may be integrated and configured to optimize one or more parameters of the captured images (e.g., contrast).
[0020] The data processing system 22 processes the images captured by the image capture device 20 to determine an image-based trajectory of the vehicle along a road. In some examples, the image capture device 20 consists of a monocular camera with an optional ranging sensor that enables vehicle motion and depth to be calculated in terms of pixel motion across successive images. In preferred examples, however, the image capture device 20 includes two or more cameras that are fixed in position in a known geometry, which allows vehicle motion and depth to be calculated in real world dimensions (e.g., meters).
[0021] Referring to FIG. 5, in one example, the data processing system 22 processes a sequence of frames of stereoscopic image pairs that are captured at different times (e.g., at equal time intervals) to determine respective estimates of three- dimensional points on the road surface.
[0022] The data processing system 22 rectifies each stereo image pair to a common epipolar plane (FIG. 5, block 40). The data processing system 22 may use any of a wide variety of methods to rectify the stereo image pair in each frame.
Examples of such methods include planar rectification, cylindrical rectification, and polar rectification. In some examples, the resulting rectified images have epipolar lines that are parallel to the horizontal axis and corresponding points have identical vertical coordinates.
[0023] If the images of the stereo pair are not grayscale images, the images optionally may be converted to a grayscale format (FIG. 5, block 42).
[0024] The data processing system 22 detects features in each of the rectified stereo pair of images in each successive frame (FIG. 5, block 44). In general, features are points in an image that can be uniquely identified based on comparisons of colors, highlights, and other features in the pair of images. Examples include points that have high contrast with its neighbors in a local region of an image.
[0025] The data processing system 22 matches corresponding features in the rectified stereo pair of images (FIG. 5, block 46). In this process, the processing system 22 matches points (e.g., pixels) or features in one image of the stereo pair with corresponding interest points or features in the other image of the stereo pair. The result of this process is a feature disparity map, where the disparity is the pixel displacement between the matched features.
[0026] In some examples, the depth at each feature point is calculated to obtain a sparse three-dimensional surface profile (FIG. 5, block 48). In some examples, the data processing system 22 produces a depth map by calculating an elevation value for each feature in the disparity map using triangulation.
[0027] The data processing system 22 determines the motion of the vehicle (e.g., orientation and translation) between successive frames (FIG. 5, block 50). In some examples, an optical flow process is used to determine the motion and orientation of the vehicle between successive stereoscopic image frames.
[0028] In the examples described above, a stereoscopic image set in a frame is used to estimate the current trajectory of the vehicle. In order to obtain a denser and more robust dataset that is less sensitive to outliers, the imaged region of the road surface under the vehicle is divided into multiple tracks that can be processed to obtain multiple independent estimates of the vehicle’s trajectory. Multiple of these estimates can be combined or otherwise used to determine estimates of the vehicle’s motion and orientation between successive stereoscopic image frames. In some examples, the multiple independent sets of estimates can be used to identify and reject as outliers motion and orientation estimates that are inconsistent with the majority of other estimates.
[0029] Referring to FIG. 6A, in one example, an imaged road region 60 is divided into three parallel tracks 62, 64, 66, each of which is processed as an
independent stereoscopic image frame. Each track 62, 64, 66 is processed
independently of the other tracks 62, 64, 66 to obtain a respective set of interest points or features for the stereoscopic images in the frame for each track 62, 64, 66. From this information, the data processing system 22 can determine a respective set of disparity and depth maps for each track 62, 64, 66.
[0030] FIG. 6B shows another example, in which the imaged road region 68 is divided into a rectangular array of six tracks 68, 70, 72, 74, 76, 78, each of which can be processed to obtain respective independent sets of estimates of motion and orientation along the vehicle’s trajectory.
[0031] In some examples, the imaged region 60 is divided into the tracks 62- 66 and 68-78 by capturing a stereoscopic image set and windowing or bucketing the frame into regions that correspond to the sets of tracks 62-66 and 68-78. In other examples, the imaged region 60 is divided into the sets of tracks 62-66 and 68-78 using lenses or diffractive optical elements that are configured to direct the light reflecting from the road surface to respective ones of the sets of tracks 62-68 and 68-78. In still other examples the imaged region 60 is divided into the sets of tracks 62-66 and 68-78 using respective pairs of stereoscopic image devices 28, 30 (e.g., cameras) arrayed according to the desired track layout. [0032] FIG. 7 shows an example method of determining estimates of the motion and orientation of a vehicle from successive stereoscopic image frames. In this example, for each track, the data processing system 22 estimates the motion and orientation between corresponding features in a respective stereoscopic pair images (FIG. 7, block 80). The data processing system 22 removes inconsistent motion and orientation estimates as outliers (FIG. 7, block 82). Estimates of the motion and orientation are determined based on an aggregation of the consistent motion and orientation estimates (FIG. 7, block 84). In some examples, the consistent motion and orientation estimates are averaged to determine the motion and orientation estimates.
[0033] FIG. 8 shows an example of a vehicle 85, in which one or more image- based locationing systems 86, 88 are mounted on the front and/or rear ends of the vehicle 10 and oriented to obtain images of the road surface.
[0034] FIG. 9 shows an example of a vehicle 96 in which first and second imaging devices 90, 92 and a light source 94 are located on the chassis of the vehicle 96 at different locations along the longitudinal dimension of the vehicle 10. Each of the image capture devices 90, 92 is operable to obtain a series of successive stereoscopic image frames of the light reflecting from the road surface. The data processing system 22 is operable to determine the vehicle’s direction of travel based on the locations of features detected in the images captured by the first imaging device 90 and the locations of matching features detected in the images captured by the second imaging device 92. By increasing the distance between the complementary images, this approach increases the accuracy of measured changes in direction. In another example, the first image capture device 90 is located under the vehicle 96 on the chassis and the second image capture device 92 is located on the roof of the vehicle 96.
[0035] FIG. 10A shows an example of a vehicle 98 that includes a cleaning system 100 for the locationing system 12. In this example, the cleaning system 100 includes a funnel that includes a large input end 102 and a small output end 104.
Movement of the vehicle 98 over the road causes air 106 to flow into the input end 102 of the funnel 100 at a first rate that depends at least in part on the speed of the vehicle 98. The optical cleaning system 100 is designed so that , under typical driving conditions, the input air is ejected from the output end 104 of the funnel 100 at pressure that is sufficient to remove dust and other debris from the optical and other components of the locationing system 12.
[0036] FIG. 10B shows an example of a vehicle that includes a cleaning system 108 for the locationing system 12. In this example, the cleaning system 108 includes an electrically powered cylindrical brush that is configured to remove dust and other debris from the optical and other components of the locationing system 12. In some examples, the rotating brush is configured to move into and out of contact with the locationing system components either on a regular basis (e.g., each time the vehicle is turned on) or on demand (e.g., when a sensor detects that the optical and/or other components require cleaning).
[0037] Referring back to FIG. 2, in some examples, the data processing system 22 also integrates data received from one or more auxiliary locationing sensors 38 into the locationing process. Example auxiliary locationing sensors 38 include a wheel encoder, an inertial measurement unit (IMU), an inertial navigation system (INS), a global positioning system (GPS), a sound navigation and ranging (SONAR) system, a LIDAR system, and a radar system.
[0038] Examples of the subject matter described herein, including the disclosed systems, methods, processes, functional operations, and logic flows, can be implemented in data processing apparatus (e.g., computer hardware and digital electronic circuitry) operable to perform functions by operating on input and generating output. Examples of the subject matter described herein also can be tangibly embodied in software or firmware, as one or more sets of computer instructions encoded on one or more tangible non-transitory carrier media (e.g., a machine readable storage device, substrate, or sequential access memory device) for execution by data processing apparatus.
[0039] The details of specific implementations described herein may be specific to particular embodiments of particular inventions and should not be construed as limitations on the scope of any claimed invention. For example, features that are described in connection with separate embodiments may also be incorporated into a single embodiment, and features that are described in connection with a single embodiment may also be implemented in multiple separate embodiments. In addition, the disclosure of steps, tasks, operations, or processes being performed in a particular order does not necessarily require that those steps, tasks, operations, or processes be performed in the particular order; instead, in some cases, one or more of the disclosed steps, tasks, operations, and processes may be performed in a different order or in accordance with a multi-tasking schedule or in parallel.
[0040] Other embodiments are within the scope of the claims.

Claims

1. A method of processing a sequence of images to determine an image based trajectory of a vehicle (10) along a road, comprising by a data processing system (22):
rectifying each stereoscopic pair of images (38) to a common epipolar plane (40); detecting features in each of the rectified stereoscopic pair of images (38) in each successive frame (44);
matching corresponding features in each rectified stereoscopic pair of images (38), wherein the matching comprises matching points in one image of the rectified stereoscopic pair with corresponding features in the other image of the stereoscopic image pair to produce a feature disparity map (46);
calculating a depth at each location point to obtain a sparse three-dimensional depth map of the road (48); and
determining motion and orientation of the vehicle (10) between successive stereoscopic image frames based on execution of an optical flow process by the data processing system (22) that determines estimates of motion and orientation between successive stereoscopic pairs of images (50).
2. The method of claim 1 , comprising by the data processing system (22), converting the image of the stereoscopic pair of images (38) to a grayscale format.
3. The method of claim 1 , wherein the disparity map measures pixel displacement between the matched features.
4. The method of claim 1 , further comprising estimating a current trajectory of the vehicle (10) based on the estimates of motion and orientation between successive stereoscopic pairs of images (38).
5. The method of claim 1 , further comprising by the data processing system (22), dividing an image region above a road surface into multiple tracks (62, 64, 66), and processing the multiple tracks to obtain multiple independent estimates of a trajectory of the vehicle (10).
6. The method of claim 5, further comprising by the data processing system (22), combining multiple ones of the estimates to determine estimates of motion and orientation of the vehicle (10) between successive stereoscopic frames.
7. The method of claim 6, further comprising by the data processing system (22), identifying and rejecting as outliers motion and orientation estimates that are inconsistent with a majority of other motion and orientation estimates.
8. The method of claim 5, further comprising by the data processing system (22), processing each of the tracks independently of other tracks to obtain a respective set of features for the stereographic image in the frame for each track.
9. The method of claim 8, further comprising by the data processing system (22), determining a respective set of disparity and depth maps for each track.
10. The method of claim 9, further comprising by the data processing system (22), dividing an imaged road region into a rectangular array multiple tracks wide and multiple tracks long (68, 70, 72, 74, 76, 78), and processing the tracks to obtain respective independent sets of estimates of motion and orientation along a trajectory of the vehicle (10).
11. The method of claim 10, wherein the dividing comprises dividing the imaged road region into the rectangular array by capturing a stereographic image set and windowing the imaged road surface into the sets of tracks.
12. The method of claim 10, wherein the dividing comprises dividing the imaged road region into the rectangular array using optical elements configured to direct light reflecting from a surface of the road to respective ones of the sets of tracks.
13. The method of claim 10, wherein the dividing comprises dividing the imaged road region into the rectangular array using respective pairs of stereoscopic image devices array according to the track layout.
14. A method of processing a sequence of successive stereographic image frames to determine estimates of motion and orientation of a vehicle (10), the method comprising:
dividing an image region above a road surface into multiple tracks;
for each track, by the data processing system (22), estimating motion and orientation between corresponding features in respective stereographic image pairs; removing, by the data processing system (22), inconsistent motion and orientation estimates as outliers;
determining, by the data processing system (22), estimates of motion and orientation based on an aggregation of consistent motion and orientation estimates; and processing the multiple tracks to obtain multiple independent estimates of a trajectory of the vehicle (10).
15. The method of claim 14, wherein consistent motion and orientation estimates are averaged to determine the motion and orientation estimates.
16. A computer program product for execution by a computer system and comprising at least one non-transitory computer-readable medium having computer readable program code portions embodied therein to process a sequence of images to determine an image based trajectory of a vehicle (10) along a road, the computer- readable program code portions comprising:
an executable code portion to rectify each stereoscopic pair of images (38) to a common epipolar plane;
an executable code portion to detect features in each of the rectified stereoscopic pair of images (38) in each successive frame;
an executable code portion to match corresponding features in each rectified stereoscopic pair of images (38), wherein the matching comprises matching points in one image of the rectified stereoscopic pair with corresponding features in the other image of the stereoscopic image pair to produce a feature disparity map;
an executable code portion to calculate a depth at each location point to obtain a sparse three-dimensional depth map of the road; and an executable code portion to determine motion and orientation of the vehicle (10) between successive stereoscopic image frames based on execution of an optical flow process by the data processing system (22) that determines estimates of motion and orientation between successive stereoscopic pairs of images.
PCT/US2020/039362 2019-06-22 2020-06-24 Image based locationing WO2020263982A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201962865192P 2019-06-22 2019-06-22
US62/865,192 2019-06-22

Publications (1)

Publication Number Publication Date
WO2020263982A1 true WO2020263982A1 (en) 2020-12-30

Family

ID=74039297

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2020/039362 WO2020263982A1 (en) 2019-06-22 2020-06-24 Image based locationing

Country Status (2)

Country Link
US (1) US11620832B2 (en)
WO (1) WO2020263982A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11681297B2 (en) 2021-03-02 2023-06-20 Southwest Research Institute Autonomous vehicle and self-location estimating method in autonomous vehicle
CN113191550B (en) * 2021-04-29 2024-04-09 北京百度网讯科技有限公司 Map matching method and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040221790A1 (en) * 2003-05-02 2004-11-11 Sinclair Kenneth H. Method and apparatus for optical odometry
US20090279741A1 (en) * 2008-05-06 2009-11-12 Honeywell Method and apparatus for vision based motion determination
US20160014394A1 (en) * 2014-07-09 2016-01-14 Hyundai Mobis Co., Ltd. Driving assistant apparatus of vehicle and operating method thereof
US20170045889A1 (en) * 2014-02-24 2017-02-16 Nissan Motor Co., Ltd. Self-Position Calculating Apparatus and Self-Position Calculating Method
US20170064287A1 (en) * 2015-08-24 2017-03-02 Itseez3D, Inc. Fast algorithm for online calibration of rgb-d camera

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2678156C (en) * 2007-02-16 2013-10-08 Mitsubishi Electric Corporation Measurement apparatus, measurement method, and feature identification apparatus
US8923558B2 (en) * 2012-05-15 2014-12-30 Raytheon Company Moving object detection using stereo rectified images acquired from a moving aircraft
TWI554103B (en) * 2014-11-13 2016-10-11 聚晶半導體股份有限公司 Image capturing device and digital zooming method thereof
US10248124B2 (en) * 2016-07-21 2019-04-02 Mobileye Vision Technologies, Inc. Localizing vehicle navigation using lane measurements
US10067513B2 (en) * 2017-01-23 2018-09-04 Hangzhou Zero Zero Technology Co., Ltd Multi-camera system and method of use
KR20180087944A (en) * 2017-01-26 2018-08-03 삼성전자주식회사 Method and apparatus for stereo matching
US11067995B2 (en) * 2017-03-20 2021-07-20 Mobileye Vision Technologies Ltd. Navigation by augmented path prediction
US10864855B2 (en) * 2017-03-31 2020-12-15 Sony Semiconductor Solutions Corporation Imaging control apparatus, method for controlling imaging control apparatus, and mobile body
US10832439B1 (en) * 2019-04-19 2020-11-10 Zoox, Inc. Locating entities in a mapped environment

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040221790A1 (en) * 2003-05-02 2004-11-11 Sinclair Kenneth H. Method and apparatus for optical odometry
US20090279741A1 (en) * 2008-05-06 2009-11-12 Honeywell Method and apparatus for vision based motion determination
US20170045889A1 (en) * 2014-02-24 2017-02-16 Nissan Motor Co., Ltd. Self-Position Calculating Apparatus and Self-Position Calculating Method
US20160014394A1 (en) * 2014-07-09 2016-01-14 Hyundai Mobis Co., Ltd. Driving assistant apparatus of vehicle and operating method thereof
US20170064287A1 (en) * 2015-08-24 2017-03-02 Itseez3D, Inc. Fast algorithm for online calibration of rgb-d camera

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
MOZILLA AGRAWAL , KURT KONOLIGE: "Real-time Localization in Outdoor Environments using Stereo Vision and Inexpensive GPS", 18TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR'06), 3 November 2020 (2020-11-03), pages 1063 - 1068, XP055670058, ISSN: 1051-4651, DOI: 10.1109/ICPR.2006.962 *
NICOLA KROMBACH; DAVID DROESCHEL; SEBASTIAN HOUBEN; SVEN BEHNKE: "Feature-based Visual Odometry Prior for Real-time Semi-dense Stereo SLAM", ROBOTICS AND AUTONOMOUS SYSTEMS, 19 October 2018 (2018-10-19), pages 1 - 46, XP080924920, DOI: 10.1016/j.robot.2018.08.002 *
PAUL TIMOTHY FURGALE: "Extensions to the Visual Odometry Pipeline for the Exploration of Planetary Surfaces", THESES, 1 November 2011 (2011-11-01), pages 1 - 156, XP055780244 *

Also Published As

Publication number Publication date
US11620832B2 (en) 2023-04-04
US20200401819A1 (en) 2020-12-24

Similar Documents

Publication Publication Date Title
TWI695181B (en) Methods and systems for color point cloud generation
CN109313031B (en) Vehicle-mounted processing device
JP6821712B2 (en) Calibration of integrated sensor in natural scene
CN108692719B (en) Object detection device
JP7082545B2 (en) Information processing methods, information processing equipment and programs
JP5157067B2 (en) Automatic travel map creation device and automatic travel device.
US10872246B2 (en) Vehicle lane detection system
CN113997931B (en) Overhead image generation device, overhead image generation system, and automatic parking device
JP6758160B2 (en) Vehicle position detection device, vehicle position detection method and computer program for vehicle position detection
US10832428B2 (en) Method and apparatus for estimating a range of a moving object
KR101709317B1 (en) Method for calculating an object's coordinates in an image using single camera and gps
JP6552448B2 (en) Vehicle position detection device, vehicle position detection method, and computer program for vehicle position detection
US11620832B2 (en) Image based locationing
Liu et al. Semalign: Annotation-free camera-lidar calibration with semantic alignment loss
CN111353453A (en) Obstacle detection method and apparatus for vehicle
Hanel et al. Towards the influence of a car windshield on depth calculation with a stereo camera system
JP7179687B2 (en) Obstacle detector
AU2018102199A4 (en) Methods and systems for color point cloud generation
KR20220050386A (en) Method of generating map and visual localization system using the map
WO2021056185A1 (en) Systems and methods for partially updating high-definition map based on sensor data matching
JP2019132701A (en) Map information creation method
WO2019151109A1 (en) Road surface information acquisition method
JP2023178490A (en) object recognition system
JP5962331B2 (en) Speed estimation apparatus and speed estimation method
JP2020077297A (en) Position and posture estimation device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20831530

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20831530

Country of ref document: EP

Kind code of ref document: A1