US9626599B2 - Reconfigurable clear path detection system - Google Patents

Reconfigurable clear path detection system Download PDF

Info

Publication number
US9626599B2
US9626599B2 US13/441,962 US201213441962A US9626599B2 US 9626599 B2 US9626599 B2 US 9626599B2 US 201213441962 A US201213441962 A US 201213441962A US 9626599 B2 US9626599 B2 US 9626599B2
Authority
US
United States
Prior art keywords
clear path
path detection
input image
module
road
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/441,962
Other versions
US20130265424A1 (en
Inventor
Wende Zhang
Jinsong Wang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
GM Global Technology Operations LLC
Original Assignee
GM Global Technology Operations LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by GM Global Technology Operations LLC filed Critical GM Global Technology Operations LLC
Priority to US13/441,962 priority Critical patent/US9626599B2/en
Assigned to GM Global Technology Operations LLC reassignment GM Global Technology Operations LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: WANG, JINSONG, ZHANG, WENDE
Priority to DE102013205952.3A priority patent/DE102013205952B4/en
Priority to CN201310120741.9A priority patent/CN103366154B/en
Assigned to WILMINGTON TRUST COMPANY reassignment WILMINGTON TRUST COMPANY SECURITY AGREEMENT Assignors: GM Global Technology Operations LLC
Publication of US20130265424A1 publication Critical patent/US20130265424A1/en
Assigned to GM Global Technology Operations LLC reassignment GM Global Technology Operations LLC RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: WILMINGTON TRUST COMPANY
Application granted granted Critical
Publication of US9626599B2 publication Critical patent/US9626599B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06K9/6292
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/588Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/254Fusion techniques of classification results, e.g. of results related to same input data
    • G06K9/00798
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/809Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of classification results, e.g. where the classifiers operate on the same input data
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W40/00Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
    • B60W40/02Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to ambient conditions
    • B60W40/06Road conditions
    • B60W40/072Curvature of the road

Definitions

  • An embodiment relates generally to road recognition.
  • Vision-imaging systems are used in vehicles for enhancing sensing applications within the vehicle such as clear path detection systems, object detection systems, and other vision/positioning systems. Such systems may be used to autonomously or semi-autonomously control the vehicle through steering systems, throttle control, braking control, or even utilized for lane departure warning systems.
  • each respective technique has attributes which provide advantages in certain respects. Utilizing more than one technique would offer enhanced detection of the clear path, but would also increase the processing time for determining the clear path.
  • An embodiment contemplates a method of detecting a clear path of a road of travel for a vehicle. Images of a scene in a path of travel are captured by an image capture device. A clear path in an input image is determined by a primary clear path detection module. The primary clear path detection module analyzes the input image from the captured image device. The primary clear path detection module segments the input image into a plurality of patches. Probability values are determined for each of the patches representing a likelihood of whether a respective patch is a clear path of travel. Feature data of the input image is associated with the segmented patches. A trained classifier is applied for identifying a clear path of travel in the input image based on the feature data and the corresponding probability values.
  • a plurality of secondary clear path detection modules is provided for independently identifying a respective clear path of the travel in the input image.
  • One or more of the secondary clear path detection modules are selectively enabled for identifying the clear path of the travel. Only the selectively enabled secondary clear path detection modules are used for identifying the clear path of the road of travel in the input image.
  • a fusion module collectively analyzes the clear path detection results of the primary clear path detection module and analyzes the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.
  • An embodiment contemplates a reconfigurable clear path detection system for a vehicle.
  • An image capture device captures images of a scene in a path of travel.
  • a primary clear path detection module determines a clear path in an input image captured from the image capture device.
  • a plurality of secondary clear path detection modules where each secondary clear path detection module is configured to independently assist in identifying a respective clear path of the road of travel in the input image.
  • One or more of the secondary clear path detection modules are selectively enabled for assisting in identifying the clear path of the road of travel wherein only the selectively enabled secondary clear path detection modules are used to identify the clear path of the road of travel in the input image.
  • a fusion module collectively analyzes the clear path detection results of the primary clear path detection module and the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.
  • FIG. 1 is an illustration of a vehicle integrating a vehicular clear path detection system.
  • FIG. 2 is a block diagram of the reconfigurable clear path detection system.
  • FIG. 1 a vehicle integrating a reconfigurable clear path detection system for a vehicle 12 .
  • the reconfigurable clear path detection system is equipped with an image capture device 14 mounted to the vehicle 12 .
  • the image capture device 14 is in communication with a processing unit 16 .
  • FIG. 2 illustrates a block diagram of the reconfigurable clear path detection system 10 .
  • the reconfigurable clear path detection system 10 may further include a memory storage device 18 to store or buffer captured images, a primary clear path detection module 20 , a plurality of secondary clear path detection modules, a first fusion module 22 , and a second fusion module 24 .
  • the image capture device 14 may include a camera or other imaging device.
  • the image capture device 14 captures images forward of the vehicle that are used to identify a clear path of travel.
  • the image capture device 14 is configured for performing continuous image capture.
  • the images captured by the image capture device 14 may be stored in a memory of the image capture device 14 , transferred to an off-board memory device, or may be transferred to a processing unit for analysis.
  • the processing unit may include the various clear path detection modules and fusion module or may be independent and retrieve the information from a respective off-board module.
  • more than one image capture device may be used in order to capture images rearward of the vehicle. Images captured rearward of the vehicle may be used in the analysis for identifying the clear path of travel forward of the vehicle.
  • the captured images are preferably a two-dimensional image of known pixel dimensions.
  • the image contains a plurality of identifiable pixels.
  • Each of the pixels includes a set of bits that correspond to a color on a predetermined color map that represents a color intensity value.
  • the processing unit is a single unit; however, it should be understood that functions of the processing unit may be performed by one or more devices implemented in software, hardware, and/or application-specific integrated circuitry.
  • the processing unit may be a general purpose processor, digital computer including a central microprocessor or central processing unit, ASIC, or other processing module having non-volatile memory, read only memory, programmable read only memory, RAM, A-to-D circuitry, D-to-A circuitry, input/output circuitry, buffering capability and appropriate signal conditioning such as digital signal processing.
  • the processing unit 16 is devised to execute algorithms utilizing inputs from the devices described herein as well as other devices or systems within the vehicle.
  • the processing unit determines a clear path in the captured image. The clear path is utilized in guiding the vehicle the along the road of travel.
  • the processing unit may perform the guidance functions as described above or may transfer the results to a secondary application that performs the guidance functions.
  • the primary clear path detection module 20 utilizes a segmentation-based technique for detecting a clear path of the traveled road.
  • the primary clear path detection module 20 analyzes an input image captured from the image capture device 14 .
  • the primary clear path detection module 20 segments the input image into a plurality of patches and determines probability values for each of the patches.
  • the determined probability values of each patch represent a likelihood of whether a respective patch is a clear path of the traveled road.
  • Feature data of the input image is associated with the segmented patches and is provided to a classifier for identifying a clear path of the traveled road based on the feature data and the corresponding probability values.
  • Clear path detection utilizing the segmentation-based technique is described in co-pending application having a Serial No. 12/581,742 filed on Oct. 19, 2009, entitled “Clear Path Detection Using Segmentation-Based Method” which is incorporated by reference in its entirety.
  • the plurality of secondary clear path detection modules are selectively used in cooperation with the results from the primary clear path detection module 20 for determining a clear path for the road of travel in the captured image.
  • the secondary clear path detection modules include, but are not limited to, an on-line similar scene detection module 26 utilizing previous reliable clear path patches, an adaptation module 28 , a top-down view classification module 30 , a road structure identification module 32 and a temporal coherence module 34 .
  • the on-line similar scene detection module 26 is a technique that requires low data processing. While the vehicle is running, an image is captured in real-time and is immediately compared to various dataset image samples that include a clear path that are previously labeled and identified. That is, prior sample images stored in the memory are already classified as clear paths. A cursory comparison is performed between the current image and the samples stored in memory, such as a database, for determining whether a match is present. Since the comparison requires low data processing, a quick assessment may be utilized. This process is referred to as a pass-through since the image is not buffered for intense analyzation. If the road of travel is new and not a previously traveled road, then the on-line similar scene detection module 26 may not be utilized since there will be no familiarity with the road of travel.
  • the adaptation module 28 utilizes adaptive machine learning techniques that include at least two classifiers. Images obtained by the image capture device 14 are segmented into patches. Characteristic features are extracted from the images. Attributes identifiable from the captured images include color and texture. The attributes are analyzed by a first classifier, such as a support vector machine, which has been previously trained to identify clear path regions in an image. The various regions are identified by a confidence value that pertains to a likelihood of whether a respective patch is part of the clear path. The patches that have a confidence value below a predetermined threshold are analyzed by an additional classifier that is adaptively trained utilizing real-world test samples previously classified as having a high confidence level which indicates a clear path of travel.
  • a first classifier such as a support vector machine
  • the data results from the first classifier and the additional adaptively-updated classifier are then used to make a cooperative decision relating to a clear path existence in patches of subsequent captured images. Details for clear path detection utilizing the adaptative technique is described in co-pending application having a Serial No. 12/963,426 filed on Dec. 8, 2010 entitled “Adaptation For Clear Path Detection With Additional Classifiers”, and co-pending application having Ser. No. 12/963,404 filed on Dec. 8, 2010, entitled “Adaptation For Clear Path Detection Using Reliable Local Model Updating”, both which are incorporated by reference in their entirety.
  • the top-down view classification module 30 utilizes a top-down view generator that converts one or more images obtained from the image capture device 14 into a top-down view of the road of travel.
  • One or more images are captured by the image capture device 14 .
  • An image-warping conversion technique is applied to the image for converting the image from a front-view road of travel to a top-down view road of travel.
  • the converted top-down image is segmented into patches. Characteristic features are extracted from the patches in the top-down view. Attributes identifiable from the converted top-down view images include color and texture.
  • the extracted features are analyzed by a classifier, such as a support vector machine, for identifying which patches are clear path regions for the road of travel.
  • Each of the plurality of secondary clear path detection modules described above may identify or assist in identifying a clear path of the road of travel for the input image.
  • clear path detection results from any of the secondary clear path detection modules are provided to the first fusion module 22 where the results are processed in cooperation with results from the primary clear path detection module 20 .
  • Each of the results provided to the first fusion module 22 are selectively combined for improving the accuracy of identifying the clear path from the input image.
  • Each of the secondary clear path detection modules may also be enabled or disabled when determining the clear path of the traveled road or may be selectively utilized, including selectively weighting the probabilities of the secondary clear path detection.
  • the clear path detection system can select a trade-off between accuracy and the processing time in identifying the clear path. That is, a clear path detection system may be reconfigured as a trade-off based on the number of modules selected which relates directly to the accuracy and time of performance of the system as a whole.
  • Identifying a region that the vehicle typically travels may assist in determining which clear path detection techniques may be best suited for identifying the clear path as certain techniques may provide no additional benefit with respect to accuracy in identifying the clear path of the road of travel. For example, if the vehicle primarily travels in an urban environment, then a secondary clear path detection module such as the road structure tracking technique utilizing vanish point-line detection may not increase the accuracy of the results as a vanishing point in an urban environment may be difficult to detect. Therefore, the clear path detection system can be reconfigured to either not execute a respective secondary clear path detection module or apply a weight of zero to the results from that respective technique. As a result, only those secondary clear path detection techniques that are beneficial from either an accuracy perspective or a processing speed perspective in identifying the clear path may be selected.
  • the results from each of the clear path detection modules are weighted for determining a degree as to which each of the results will be respectively relied on and utilized.
  • the following equation represents a weighted formula for applying each of the respective results from the clear path detection results as described above.
  • the formula is represented as follows:
  • P fuse ⁇ ( c ) w offline ⁇ P offline ⁇ ( c ) + w online ⁇ P online ⁇ ( c ) + w topview ⁇ P topview ⁇ ( c ) + w adapt ⁇ P adapt ⁇ ( c ) w offline + w online + w topview + w adapt
  • the fused clear-path likelihood P fuse (c) is determined as a function of a weighted linear combination of offline SVM classification probability P offline (c), an online similar scene classification probability P online (c), a top-down view classification probability P topview (c), and the adaption likelihood P adapt (c).
  • the weights of the four module outputs w offline , w online , w topview , and w adapt may be equal (e.g., 1 ⁇ 4) or a regression method can be utilized to generate an optimal weight for each module.
  • the reconfigurable clear path detection system can utilize only the primary clear path detection module 20 or can selectively include any combination of the other three modules.
  • the results from the first fusion module 22 are then applied to the second fusion module 24 .
  • the second fusion module 24 utilizes the results of the first fusion module 22 and factors in road structure constraints. Weighting is used to determine a degree as to which the road structure constraints are used in the second fusion module 24 . Weighting is based on a probabilistic framework. A confidence map is generated based on the road structure detection result. The confidence map is applied to the probability map of the clear path classification for refining the clear path detection result.
  • results from the first fusion module 22 are provided to second fusion module 24 .
  • Results from the road structure identification module 32 are in cooperation with the results from the first fusion module 22 to gain enhanced confidence of identifying the clear path.
  • the road structure identification module 32 vanishing point and potential road edges are detected for identifying boundaries of the road which is used to identify the clear path.
  • the road structure identification module 32 obtains images from the image capture device 14 or memory 18 . Images are provided to a line clustering and vanish-point detection sub-module.
  • the sub-module utilizes an illumination intensity image and a yellow image transformed from the captured input image and performs edge analysis on the illumination intensity image and the yellow image for detecting line candidates.
  • Edge filters are applied on the illumination intensity image to obtain vertical and horizontal gradients, which are then used to determine gradient magnitude and gradient angle.
  • the gradient magnitude of each pixel is compared with a threshold to extract edge points which have a gradient magnitude larger than a predetermined threshold.
  • the edge analysis identifies pixels associated with edges in the image data and then utilizes the edge points in the image to determine a potential line candidate.
  • the line candidates are detected by clustering (grouping) potential pixels associated with a potential lane marker or road curb.
  • Line clustering involves scanning the edge pixels one by one. A connectivity of pixels is identified for clustering a respective set of pixels based on a similar gradient between the neighboring pixels. For each edge pixel, the surrounding area of a respective pixel is searched for other edge pixels which have similar gradient angle to the respective pixel. The pixels having similar gradient angles are grouped together and assigned a same line index. After scanning and grouping each of the pixels, line groups which are longer than a respective length and vertical direction spreading are extracted and identified as potential line clusters for further processing. As a result, potential line-clusters are identified based on those clusters having gradient angle similarity and local connectivity.
  • Line clusters identified in the illumination intensity image and the yellow image are merged and potential line candidate detection is performed on the merged line clusters.
  • Potential line candidates may be identified from any of the bright line strips, dark line strips, and elongated line strips (i.e., lines that extend a substantial distance in the image).
  • Vanishing point detection is applied based on the potential line candidate results.
  • a vanishing point is estimated based on the identified bright strips and the long dark strips.
  • other dark strips and single long lines are identified for refining the vanishing point. Lines terminating that are in close proximity to the estimated initial vanishing point are selected for refining the vanishing point.
  • the road structure is identified based on the detected lines as described herein.
  • lines in close proximity to the estimated vanishing points are identified. For each of the identified lines in close proximity to the estimated vanishing points, features are extracted from the image. Classification analysis is performed on the candidate features of categorized left and right line clusters that represent reliable road edges.
  • a confidence map is generated for which high confidence is assigned to the region or regions between the detected road edges detected by the road structure identification module 32 and low confidence value is assigned to the region or regions outside of the detected road edges. For those segmentation patches that fall in the region outside of the identified road edges, the likelihood of that segmentation patch being part of the clear path is decreased. This is performed by fusing their associated probability as determined in the first fusion module 22 with the confidence map generated from the road structure identification module 32 .
  • the fusion module as described is reconfigurable, which means that the fused clear path likelihood can be utilized with or without considering road edge constraints.
  • the results from the second fusion module 24 are provided to the temporal coherence module 34 .
  • the temporal coherence module 34 applies temporal smoothing to the fused output data.
  • the temporal coherence technique generates a fused clear-path likelihood from previous image frame data and current image frame data.
  • the temporal coherence module 34 identifies patches within each of the time-displaced captured images. Features associated with patches are extracted from each image. A probability value that indicates a likelihood of a patch being part of the clear path of the traveled road is determined for each extracted feature.
  • Each of the time-displaced images is compared and features having substantially similar appearance properties between the time-displaced images are identified. In response to the comparison of the features having substantially similar appearance properties and their associated probability values, the features indicating the clear path are determined.
  • This technique can be represented utilizing the following formula:
  • u ⁇ ( t ) , v ⁇ ( t ) ) ) ⁇ P t ( u ⁇ ( t ) , ( v ⁇ ( t ) ) c 0 + ⁇ t 1 n ⁇ ⁇ c t ⁇ ⁇ ⁇ ( m t ⁇ ( u 0 , v 0
  • a pixel location (u t v t ) in a previous frame which corresponds to (u 0 v 0 ) in the current frame is calculated utilizing factors such as vehicle motion and on-the-ground (clear path) assumptions.
  • u t v t )) indicates whether or not the pixel (u t v t ) in the t-th previous frame contributes to temporal smoothing, which is based on the similarity to the pixel (u 0 v 0 ) in the current frame.
  • the weights can be fixed equal weights which mean equal distribution among each frame, or may decay with time t which means that the previous video frame contributes less when from an earlier time.
  • the clear path results produced by the first fusion module 22 and the second fusion module 24 are provided to an output device 36 .
  • the output device 36 may include a display for displaying the road of travel and the clear path to the driver of the vehicle, such as the display utilized by a backup camera.
  • the output device 36 may include a vehicle application, such as an object detection system that scans the identified clear path for detecting objects in the scene of the traveled road.
  • the table below illustrates exemplary configurations based on accuracy and speed of using secondary clear path modules in cooperation with the primary clear path module.
  • the baseline as shown in the table below represents the primary clear path detection module.
  • speed data and accuracy data may be different dependent upon environment, hardware, and software utilized. As is shown, as the accuracy increases, processing time to identify the clear path utilizing the selected clear path modules increases. Alternatively, as less clear path detection modules are utilized, the processing time decreases as does the accuracy.
  • the vehicle may be serviced by a technician at a service center or other trained personnel that are able to access and reconfigure the software for executing the clear path detection system. If the results of the secondary clear path detection modules are to be weighted, then the system automatically determines the weights based equal distribution of weighting between the secondary clear path detection modules or weighting may be based on a regression technique.

Abstract

A reconfigurable clear path detection system includes an image capture device and a primary clear path detection module for determining corresponding probability values of identified patches within a captured image representing a likelihood of whether a respective patch is a clear path of the road. A plurality of secondary clear path detection modules each are used to assist in identifying a respective clear path of the traveled road in the input image. One or more of the secondary clear path detection modules are selectively enabled for identifying the clear path. The selectively enabled secondary clear path detection modules are used to identify the clear path of the road of travel in the input image. A fusion module collectively analyzes the clear path detection results of the primary clear path detection module and the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.

Description

BACKGROUND OF INVENTION
An embodiment relates generally to road recognition.
Vision-imaging systems are used in vehicles for enhancing sensing applications within the vehicle such as clear path detection systems, object detection systems, and other vision/positioning systems. Such systems may be used to autonomously or semi-autonomously control the vehicle through steering systems, throttle control, braking control, or even utilized for lane departure warning systems.
Various techniques may be utilized for identifying a clear path for object detection. Typically, each respective technique has attributes which provide advantages in certain respects. Utilizing more than one technique would offer enhanced detection of the clear path, but would also increase the processing time for determining the clear path.
SUMMARY OF INVENTION
An embodiment contemplates a method of detecting a clear path of a road of travel for a vehicle. Images of a scene in a path of travel are captured by an image capture device. A clear path in an input image is determined by a primary clear path detection module. The primary clear path detection module analyzes the input image from the captured image device. The primary clear path detection module segments the input image into a plurality of patches. Probability values are determined for each of the patches representing a likelihood of whether a respective patch is a clear path of travel. Feature data of the input image is associated with the segmented patches. A trained classifier is applied for identifying a clear path of travel in the input image based on the feature data and the corresponding probability values. A plurality of secondary clear path detection modules is provided for independently identifying a respective clear path of the travel in the input image. One or more of the secondary clear path detection modules are selectively enabled for identifying the clear path of the travel. Only the selectively enabled secondary clear path detection modules are used for identifying the clear path of the road of travel in the input image. A fusion module collectively analyzes the clear path detection results of the primary clear path detection module and analyzes the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.
An embodiment contemplates a reconfigurable clear path detection system for a vehicle. An image capture device captures images of a scene in a path of travel. A primary clear path detection module determines a clear path in an input image captured from the image capture device. A plurality of secondary clear path detection modules where each secondary clear path detection module is configured to independently assist in identifying a respective clear path of the road of travel in the input image. One or more of the secondary clear path detection modules are selectively enabled for assisting in identifying the clear path of the road of travel wherein only the selectively enabled secondary clear path detection modules are used to identify the clear path of the road of travel in the input image. A fusion module collectively analyzes the clear path detection results of the primary clear path detection module and the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.
BRIEF DESCRIPTION OF DRAWINGS
FIG. 1 is an illustration of a vehicle integrating a vehicular clear path detection system.
FIG. 2 is a block diagram of the reconfigurable clear path detection system.
DETAILED DESCRIPTION
There is shown in FIG. 1 a vehicle integrating a reconfigurable clear path detection system for a vehicle 12. The reconfigurable clear path detection system is equipped with an image capture device 14 mounted to the vehicle 12. The image capture device 14 is in communication with a processing unit 16.
FIG. 2 illustrates a block diagram of the reconfigurable clear path detection system 10. The reconfigurable clear path detection system 10 may further include a memory storage device 18 to store or buffer captured images, a primary clear path detection module 20, a plurality of secondary clear path detection modules, a first fusion module 22, and a second fusion module 24.
The image capture device 14 may include a camera or other imaging device. The image capture device 14 captures images forward of the vehicle that are used to identify a clear path of travel. Preferably, the image capture device 14 is configured for performing continuous image capture. The images captured by the image capture device 14 may be stored in a memory of the image capture device 14, transferred to an off-board memory device, or may be transferred to a processing unit for analysis. The processing unit may include the various clear path detection modules and fusion module or may be independent and retrieve the information from a respective off-board module. Moreover, more than one image capture device may be used in order to capture images rearward of the vehicle. Images captured rearward of the vehicle may be used in the analysis for identifying the clear path of travel forward of the vehicle.
The captured images are preferably a two-dimensional image of known pixel dimensions. The image contains a plurality of identifiable pixels. Each of the pixels includes a set of bits that correspond to a color on a predetermined color map that represents a color intensity value.
The processing unit is a single unit; however, it should be understood that functions of the processing unit may be performed by one or more devices implemented in software, hardware, and/or application-specific integrated circuitry. The processing unit, may be a general purpose processor, digital computer including a central microprocessor or central processing unit, ASIC, or other processing module having non-volatile memory, read only memory, programmable read only memory, RAM, A-to-D circuitry, D-to-A circuitry, input/output circuitry, buffering capability and appropriate signal conditioning such as digital signal processing. The processing unit 16 is devised to execute algorithms utilizing inputs from the devices described herein as well as other devices or systems within the vehicle. The processing unit determines a clear path in the captured image. The clear path is utilized in guiding the vehicle the along the road of travel. The processing unit may perform the guidance functions as described above or may transfer the results to a secondary application that performs the guidance functions.
The primary clear path detection module 20 utilizes a segmentation-based technique for detecting a clear path of the traveled road. The primary clear path detection module 20 analyzes an input image captured from the image capture device 14. The primary clear path detection module 20 segments the input image into a plurality of patches and determines probability values for each of the patches. The determined probability values of each patch represent a likelihood of whether a respective patch is a clear path of the traveled road. Feature data of the input image is associated with the segmented patches and is provided to a classifier for identifying a clear path of the traveled road based on the feature data and the corresponding probability values. Clear path detection utilizing the segmentation-based technique is described in co-pending application having a Serial No. 12/581,742 filed on Oct. 19, 2009, entitled “Clear Path Detection Using Segmentation-Based Method” which is incorporated by reference in its entirety.
The plurality of secondary clear path detection modules are selectively used in cooperation with the results from the primary clear path detection module 20 for determining a clear path for the road of travel in the captured image. The secondary clear path detection modules include, but are not limited to, an on-line similar scene detection module 26 utilizing previous reliable clear path patches, an adaptation module 28, a top-down view classification module 30, a road structure identification module 32 and a temporal coherence module 34.
The on-line similar scene detection module 26 is a technique that requires low data processing. While the vehicle is running, an image is captured in real-time and is immediately compared to various dataset image samples that include a clear path that are previously labeled and identified. That is, prior sample images stored in the memory are already classified as clear paths. A cursory comparison is performed between the current image and the samples stored in memory, such as a database, for determining whether a match is present. Since the comparison requires low data processing, a quick assessment may be utilized. This process is referred to as a pass-through since the image is not buffered for intense analyzation. If the road of travel is new and not a previously traveled road, then the on-line similar scene detection module 26 may not be utilized since there will be no familiarity with the road of travel.
The adaptation module 28 utilizes adaptive machine learning techniques that include at least two classifiers. Images obtained by the image capture device 14 are segmented into patches. Characteristic features are extracted from the images. Attributes identifiable from the captured images include color and texture. The attributes are analyzed by a first classifier, such as a support vector machine, which has been previously trained to identify clear path regions in an image. The various regions are identified by a confidence value that pertains to a likelihood of whether a respective patch is part of the clear path. The patches that have a confidence value below a predetermined threshold are analyzed by an additional classifier that is adaptively trained utilizing real-world test samples previously classified as having a high confidence level which indicates a clear path of travel. The data results from the first classifier and the additional adaptively-updated classifier are then used to make a cooperative decision relating to a clear path existence in patches of subsequent captured images. Details for clear path detection utilizing the adaptative technique is described in co-pending application having a Serial No. 12/963,426 filed on Dec. 8, 2010 entitled “Adaptation For Clear Path Detection With Additional Classifiers”, and co-pending application having Ser. No. 12/963,404 filed on Dec. 8, 2010, entitled “Adaptation For Clear Path Detection Using Reliable Local Model Updating”, both which are incorporated by reference in their entirety.
The top-down view classification module 30 utilizes a top-down view generator that converts one or more images obtained from the image capture device 14 into a top-down view of the road of travel. One or more images are captured by the image capture device 14. An image-warping conversion technique is applied to the image for converting the image from a front-view road of travel to a top-down view road of travel. The converted top-down image is segmented into patches. Characteristic features are extracted from the patches in the top-down view. Attributes identifiable from the converted top-down view images include color and texture. The extracted features are analyzed by a classifier, such as a support vector machine, for identifying which patches are clear path regions for the road of travel.
Each of the plurality of secondary clear path detection modules described above may identify or assist in identifying a clear path of the road of travel for the input image. In FIG. 2, clear path detection results from any of the secondary clear path detection modules are provided to the first fusion module 22 where the results are processed in cooperation with results from the primary clear path detection module 20. Each of the results provided to the first fusion module 22 are selectively combined for improving the accuracy of identifying the clear path from the input image.
Each of the secondary clear path detection modules may also be enabled or disabled when determining the clear path of the traveled road or may be selectively utilized, including selectively weighting the probabilities of the secondary clear path detection. In utilizing only a selected group of secondary clear path detection modules, the clear path detection system can select a trade-off between accuracy and the processing time in identifying the clear path. That is, a clear path detection system may be reconfigured as a trade-off based on the number of modules selected which relates directly to the accuracy and time of performance of the system as a whole.
Increasing the number of clear path detection techniques that are executed in cooperation with the primary clear path detection module 12 increases the confidence level that the clear path of travel has been properly identified. However, while each additional second clear path detection technique increases the confidence level accurately identifying the clear path, the processing time required for executing each of the selected modules increases. Therefore, only a selected number of secondary clear path detection techniques may be executed for optimizing a trade-off between accuracy and performance time.
Identifying a region that the vehicle typically travels may assist in determining which clear path detection techniques may be best suited for identifying the clear path as certain techniques may provide no additional benefit with respect to accuracy in identifying the clear path of the road of travel. For example, if the vehicle primarily travels in an urban environment, then a secondary clear path detection module such as the road structure tracking technique utilizing vanish point-line detection may not increase the accuracy of the results as a vanishing point in an urban environment may be difficult to detect. Therefore, the clear path detection system can be reconfigured to either not execute a respective secondary clear path detection module or apply a weight of zero to the results from that respective technique. As a result, only those secondary clear path detection techniques that are beneficial from either an accuracy perspective or a processing speed perspective in identifying the clear path may be selected.
In the first fusion module 22, the results from each of the clear path detection modules are weighted for determining a degree as to which each of the results will be respectively relied on and utilized. The following equation represents a weighted formula for applying each of the respective results from the clear path detection results as described above. The formula is represented as follows:
P fuse ( c ) = w offline · P offline ( c ) + w online · P online ( c ) + w topview · P topview ( c ) + w adapt · P adapt ( c ) w offline + w online + w topview + w adapt
The fused clear-path likelihood Pfuse(c) is determined as a function of a weighted linear combination of offline SVM classification probability Poffline(c), an online similar scene classification probability Ponline(c), a top-down view classification probability Ptopview(c), and the adaption likelihood Padapt(c). The weights of the four module outputs woffline, wonline, wtopview, and wadapt, may be equal (e.g., ¼) or a regression method can be utilized to generate an optimal weight for each module. As a result, the reconfigurable clear path detection system can utilize only the primary clear path detection module 20 or can selectively include any combination of the other three modules.
The results from the first fusion module 22 are then applied to the second fusion module 24. The second fusion module 24 utilizes the results of the first fusion module 22 and factors in road structure constraints. Weighting is used to determine a degree as to which the road structure constraints are used in the second fusion module 24. Weighting is based on a probabilistic framework. A confidence map is generated based on the road structure detection result. The confidence map is applied to the probability map of the clear path classification for refining the clear path detection result.
In block 24, the results from the first fusion module 22 are provided to second fusion module 24. Results from the road structure identification module 32 are in cooperation with the results from the first fusion module 22 to gain enhanced confidence of identifying the clear path.
In the road structure identification module 32, vanishing point and potential road edges are detected for identifying boundaries of the road which is used to identify the clear path. The road structure identification module 32 obtains images from the image capture device 14 or memory 18. Images are provided to a line clustering and vanish-point detection sub-module. The sub-module utilizes an illumination intensity image and a yellow image transformed from the captured input image and performs edge analysis on the illumination intensity image and the yellow image for detecting line candidates. Edge filters are applied on the illumination intensity image to obtain vertical and horizontal gradients, which are then used to determine gradient magnitude and gradient angle. The gradient magnitude of each pixel is compared with a threshold to extract edge points which have a gradient magnitude larger than a predetermined threshold. The edge analysis identifies pixels associated with edges in the image data and then utilizes the edge points in the image to determine a potential line candidate.
The line candidates are detected by clustering (grouping) potential pixels associated with a potential lane marker or road curb. Line clustering involves scanning the edge pixels one by one. A connectivity of pixels is identified for clustering a respective set of pixels based on a similar gradient between the neighboring pixels. For each edge pixel, the surrounding area of a respective pixel is searched for other edge pixels which have similar gradient angle to the respective pixel. The pixels having similar gradient angles are grouped together and assigned a same line index. After scanning and grouping each of the pixels, line groups which are longer than a respective length and vertical direction spreading are extracted and identified as potential line clusters for further processing. As a result, potential line-clusters are identified based on those clusters having gradient angle similarity and local connectivity.
Line clusters identified in the illumination intensity image and the yellow image are merged and potential line candidate detection is performed on the merged line clusters. Potential line candidates may be identified from any of the bright line strips, dark line strips, and elongated line strips (i.e., lines that extend a substantial distance in the image).
Vanishing point detection is applied based on the potential line candidate results. A vanishing point is estimated based on the identified bright strips and the long dark strips. In addition, other dark strips and single long lines are identified for refining the vanishing point. Lines terminating that are in close proximity to the estimated initial vanishing point are selected for refining the vanishing point. As a result, the road structure is identified based on the detected lines as described herein.
After the vanishing point is determined, lines in close proximity to the estimated vanishing points are identified. For each of the identified lines in close proximity to the estimated vanishing points, features are extracted from the image. Classification analysis is performed on the candidate features of categorized left and right line clusters that represent reliable road edges.
In the second fusion module 24, a confidence map is generated for which high confidence is assigned to the region or regions between the detected road edges detected by the road structure identification module 32 and low confidence value is assigned to the region or regions outside of the detected road edges. For those segmentation patches that fall in the region outside of the identified road edges, the likelihood of that segmentation patch being part of the clear path is decreased. This is performed by fusing their associated probability as determined in the first fusion module 22 with the confidence map generated from the road structure identification module 32. The formula for determining the probability of the clear path as a function to the identified road structure is represented by the following formula:
P rdstr _ fuse(c)=Confrd _ edge(cP fuse(c)
where Pfuse(c) is the probability of the clear path for a respective segmented patch generated by the first fusion module 22, and Confrd _ edge is the confidence value assigned to the region inside or outside of the road edge.
The fusion module as described is reconfigurable, which means that the fused clear path likelihood can be utilized with or without considering road edge constraints.
The results from the second fusion module 24 are provided to the temporal coherence module 34. The temporal coherence module 34 applies temporal smoothing to the fused output data. The temporal coherence technique generates a fused clear-path likelihood from previous image frame data and current image frame data. The temporal coherence module 34 identifies patches within each of the time-displaced captured images. Features associated with patches are extracted from each image. A probability value that indicates a likelihood of a patch being part of the clear path of the traveled road is determined for each extracted feature. Each of the time-displaced images is compared and features having substantially similar appearance properties between the time-displaced images are identified. In response to the comparison of the features having substantially similar appearance properties and their associated probability values, the features indicating the clear path are determined. This technique can be represented utilizing the following formula:
p TS ( u 0 v 0 ) = c 0 · P 0 ( u 0 v 0 ) + t = 1 n c t · δ ( m t ( u 0 , v 0 | u ( t ) , v ( t ) ) ) · P t ( u ( t ) , ( v ( t ) ) c 0 + t = 1 n c t · δ ( m t ( u 0 , v 0 | u ( t ) , v ( t ) ) ) .
A pixel location (utvt) in a previous frame which corresponds to (u0v0) in the current frame is calculated utilizing factors such as vehicle motion and on-the-ground (clear path) assumptions. A delta function δ(mt(u0v0|utvt)) indicates whether or not the pixel (utvt) in the t-th previous frame contributes to temporal smoothing, which is based on the similarity to the pixel (u0v0) in the current frame. The weights ct,t=0,1,2, . . . indicate the contribution weight of each frame. The weights can be fixed equal weights which mean equal distribution among each frame, or may decay with time t which means that the previous video frame contributes less when from an earlier time.
The clear path results produced by the first fusion module 22 and the second fusion module 24 are provided to an output device 36. The output device 36 may include a display for displaying the road of travel and the clear path to the driver of the vehicle, such as the display utilized by a backup camera. The output device 36 may include a vehicle application, such as an object detection system that scans the identified clear path for detecting objects in the scene of the traveled road.
The table below illustrates exemplary configurations based on accuracy and speed of using secondary clear path modules in cooperation with the primary clear path module. The baseline as shown in the table below represents the primary clear path detection module.
Configuration Accuracy Speed (sec)
Baseline 93.93% 0.736
Baseline + Adaptation Detection 93.38% 0.774
Baseline + Road Structure 93.94% 1.411
Detection
Baseline + Temporal Coherence 91.51% 2.21
Detection
Baseline + Temporal Coherence + 94.63% 2.25
Adaptation Detection
Baseline + Temporal Coherence + 94.64% 2.759
Adaptation + Road Structure
Detection

It should be understood that the configurations and values shown herein are exemplary and that the table can include more or less information than what is shown. For example, various configurations may utilize that top-down view techniques and/or on-line techniques in addition to other known techniques without deviating form the scope of the invention. In addition, speed data and accuracy data may be different dependent upon environment, hardware, and software utilized. As is shown, as the accuracy increases, processing time to identify the clear path utilizing the selected clear path modules increases. Alternatively, as less clear path detection modules are utilized, the processing time decreases as does the accuracy.
To reconfigure the clear path detection system so that only the selected secondary clear path detection modules are utilized, the vehicle may be serviced by a technician at a service center or other trained personnel that are able to access and reconfigure the software for executing the clear path detection system. If the results of the secondary clear path detection modules are to be weighted, then the system automatically determines the weights based equal distribution of weighting between the secondary clear path detection modules or weighting may be based on a regression technique.
While certain embodiments of the present invention have been described in detail, those familiar with the art to which this invention relates will recognize various alternative designs and embodiments for practicing the invention as defined by the following claims.

Claims (26)

What is claimed is:
1. A reconfigurable clear path detection system for a vehicle comprising:
an image capture device for capturing images of a scene in a path of travel;
a primary clear path detection module determining a clear path in an input image captured from the image capture device;
a plurality of secondary clear path detection modules, each secondary clear path detection module configured to independently assist in identifying a respective clear path of the road of travel in the input image, wherein one or more of the secondary clear path detection modules are selectively enabled for assisting in identifying the clear path of the road of travel, wherein only the selectively enabled secondary clear path detection modules are used to identify the clear path of the road of travel in the input image; and
a fusion module for collectively analyzing the clear path detection results of the primary clear path detection module and the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.
2. The reconfigurable clear path detection system of claim 1 wherein the plurality of secondary clear path detection modules includes a temporal coherence module, wherein a clear path detection probability from a previous time step frame is utilized to update a clear path detection probability in a current time step frame, wherein a probability update is based on a weighted inter-frame matching technique which is derived from vehicle motion and an on-ground plane assumption.
3. The reconfigurable clear path detection system of claim 1 wherein the plurality of secondary clear path detection modules includes a top-down view classification module, wherein a top-down view image is generated from the input image, and wherein features are extracted from the top-down view image for identifying the clear path.
4. The reconfigurable clear path detection system of claim 1 wherein the plurality of secondary clear path detection modules includes a road structure detection module, wherein vanishing point and pass-vanishing point line tracking is performed and learning-based road edge detection is applied for identifying road edges and road structure in the input image.
5. The reconfigurable clear path detection system of claim 1 wherein the plurality of secondary clear path detection modules includes an on-line similar scene module, wherein the on-line similar scene module utilizes an image captured in real-time and compares the captured image to various images retrieved from a database that are previously labeled or identified.
6. The reconfigurable clear path detection system of claim 1 wherein a combination of the primary clear path detection module and the selectively enabled clear path detection modules are collectively analyzed by weighting each of the results of the primary clear path detection module and the selectively enabled secondary clear path detection modules.
7. The reconfigurable clear path detection system of claim 6 wherein weighting each of the results of the primary clear path detection module and the selectively enabled secondary clear path detection modules are represented by the formula as follows:
P fuse ( c ) = w offline · P offline ( c ) + w online · P online ( c ) + w topview · P topview ( c ) + w adapt · P adapt ( c ) w offline + w online + w topview + w adapt
where Pfuse(c) is a fused clear-path likelihood, Poffline(c) is a primary clear path detection probability classified by an offline trained SVM model, Ponline(c) is an online similar scene classification probability, Ptopview(c) is a top-down view classification probability, and Padapt(c)is the adaption probability, and woffline, wonline, wtopview, and wadapt are weight values.
8. The reconfigurable clear path detection system of claim 7 wherein the clear path is further determined as a function of the identified road structure and is represented by the following formula:

P rdstr _ fuse(c)=Confrd _ edge(cP fuse(c)
where—Prdstr _ fuse(c) is a probability of the clear path as a function of the identified road structure, Pfuse(c) is the probability of the clear path for a respective segmented patch generated by a first fusion module, and Confrd _ edge is a confidence map with regions inside and outside of the identified road edges being assigned different confidence values.
9. The reconfigurable clear path detection system of claim 1 wherein the selectively enabled secondary clear path detection modules are selected based on a trade-off between a processing time and accuracy of identifying the clear path in the input image.
10. The reconfigurable clear path detection system of claim 1 wherein the selectively enabled secondary clear path detection modules utilize time-displaced captured images of the traveled road for identifying the clear path in the input image.
11. The reconfigurable clear path detection system of claim 10 wherein the time-displaced captured images of the traveled road are captured within a predetermined frame rate of capturing the input image.
12. The reconfigurable clear path detection system of claim 1 further comprising an output device for displaying the clear path of road to a driver of a vehicle.
13. The reconfigurable clear path detection system of claim 1 further comprising an object detection system, the object detection system utilizing the identified clear path for detecting objects in the input image.
14. A method for detecting a clear path of a road of travel for a vehicle, the method comprising the steps of:
capturing images of a scene in a path of travel by an image capture device;
determining a clear path in an input image by a primary clear path detection module, the primary clear path detection module analyzing the input image from the captured image device, the primary clear path detection module segmenting the input image into a plurality of patches, determining probability values for each of the patches representing a likelihood of whether a respective patch is a clear path of travel, associating feature data of the input image with the segmented patches, and applying a trained classifier for identifying a clear path of travel in the input image based on the feature data and the corresponding probability values;
providing a plurality of secondary clear path detection modules for independently identifying a respective clear path of the travel in the input image;
selectively enabling one or more of the secondary clear path detection modules for identifying the clear path of the travel, wherein only the selectively enabled secondary clear path detection modules are used for identifying the clear path of the road of travel in the input image;
a fusion module collectively analyzing the clear path detection results of the primary clear path detection module and the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.
15. The method of claim 14 wherein the plurality of secondary clear path detection modules includes a temporal coherence module, wherein a clear path detection probability from a previous time step frame is utilized to update a clear path detection probability in a current time step frame, wherein a probability update is based on a weighted inter-frame matching technique which is derived from vehicle motion and an on-ground plane assumption.
16. The method of claim 14 wherein the plurality of secondary clear path detection modules includes a top-down view classification module, wherein a top-down view image of the captured image is generated from one or more captured input images exterior of the vehicle, and wherein features are extracted from the top-down view image, and wherein extracted features from the top-down view are applied to a classifier for identifying the clear path in the input image.
17. The method of claim 14 wherein the plurality of secondary clear path detection modules includes a road structure detection module, wherein vanishing point and pass-vanishing point line tracking is performed and learning-based road edge detection is applied for identifying road edges and road structure in the input image.
18. The method of claim 14 wherein the plurality of secondary clear path detection modules includes an on-line similar scene module, wherein the on-line similar scene module utilizes an image captured in real-time and compares the captured image to various images retrieved from a database that are previously labeled or identified.
19. The method of claim 14 wherein the fusion module weights the result of the primary clear path detection module and the results of the selectively enabled secondary clear path detection modules for identifying the clear path in the input image.
20. The method of claim 19 wherein weighting each of the results of the primary clear path detection module and the selectively enabled secondary clear path detection modules are represented by the formula as follows:
P fuse ( c ) = w offline · P offline ( c ) + w online · P online ( c ) + w topview · P topview ( c ) + w adapt · P adapt ( c ) w offline + w online + w topview + w adapt
where Pfuse(c) is a fused clear-path likelihood, Poffline(c) is a primary clear path detection probability classified by an offline trained SVM model, Ponline(c) is an online similar scene classification probability, Ptopview(c) is a top-down view classification probability, and Padapt(c)is the adaption probability, and woffline, wonline, wtopview, and wadapt are weight values.
21. The method of claim 20 wherein the clear path is further determined as a function of the identified road structure and is represented by the following formula:

P rdstr fuse(c)=Confrd _ edge(cP fuse(c)
where—Prdstr _ fuse(c) is a probability of the clear path as a function of the identified road structure, Pfuse(c) is the probability of the clear path for a respective segmented patch generated by a first fusion module, and Confrd _ edge is a confidence map with regions inside and outside of the identified road edges being assigned different confidence values.
22. The method of claim 14 wherein the enabled secondary clear path detection modules are selected based on trade-off between a processing time and accuracy of identifying the clear path in the input image.
23. The method of claim 14 wherein the selectively enabled secondary clear path detection modules utilize time-displaced captured images of the path of travel for identification of the clear path in the input image.
24. The method of claim 14 wherein the time-displaced captured images of the path of travel include images captured within a previous predetermined time period.
25. The method of claim 14 further comprising the step of displaying the clear path of travel to a driver of the vehicle using an output display device.
26. The method of claim 14 further comprising the step of providing the identified clear path of travel to an object detection system for detecting objects in the input image.
US13/441,962 2012-04-09 2012-04-09 Reconfigurable clear path detection system Active 2035-05-03 US9626599B2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US13/441,962 US9626599B2 (en) 2012-04-09 2012-04-09 Reconfigurable clear path detection system
DE102013205952.3A DE102013205952B4 (en) 2012-04-09 2013-04-04 Reconfigurable system and method for the detection of a free path
CN201310120741.9A CN103366154B (en) 2012-04-09 2013-04-09 Reconfigurable clear path detection system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/441,962 US9626599B2 (en) 2012-04-09 2012-04-09 Reconfigurable clear path detection system

Publications (2)

Publication Number Publication Date
US20130265424A1 US20130265424A1 (en) 2013-10-10
US9626599B2 true US9626599B2 (en) 2017-04-18

Family

ID=49290316

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/441,962 Active 2035-05-03 US9626599B2 (en) 2012-04-09 2012-04-09 Reconfigurable clear path detection system

Country Status (3)

Country Link
US (1) US9626599B2 (en)
CN (1) CN103366154B (en)
DE (1) DE102013205952B4 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11334605B2 (en) 2015-06-04 2022-05-17 Here Global B.V. Incremental update of compressed navigational databases

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8750567B2 (en) * 2012-04-09 2014-06-10 GM Global Technology Operations LLC Road structure detection and tracking
US10229363B2 (en) * 2015-10-19 2019-03-12 Ford Global Technologies, Llc Probabilistic inference using weighted-integrals-and-sums-by-hashing for object tracking
US10267908B2 (en) * 2015-10-21 2019-04-23 Waymo Llc Methods and systems for clearing sensor occlusions
CN105825173B (en) * 2016-03-11 2019-07-19 福州华鹰重工机械有限公司 General road and lane detection system and method
US9881230B2 (en) 2016-05-11 2018-01-30 International Business Machines Corporation System and method for automated road identification in distant traffic camera images
US11222438B2 (en) * 2016-05-27 2022-01-11 Kabushiki Kaisha Toshiba Information processing apparatus, vehicle, and information processing method for presence probability of object
US11132611B2 (en) * 2016-05-27 2021-09-28 Kabushiki Kaisha Toshiba Information processing apparatus and information processing method for determining presence probability of object
EP3252658B1 (en) * 2016-05-30 2021-08-11 Kabushiki Kaisha Toshiba Information processing apparatus and information processing method
US11204610B2 (en) * 2016-05-30 2021-12-21 Kabushiki Kaisha Toshiba Information processing apparatus, vehicle, and information processing method using correlation between attributes
US10421452B2 (en) * 2017-03-06 2019-09-24 GM Global Technology Operations LLC Soft track maintenance
US10670725B2 (en) 2017-07-25 2020-06-02 Waymo Llc Determining yaw error from map data, lasers, and cameras
JP6918425B2 (en) * 2017-11-09 2021-08-11 株式会社Mobility Technologies Devices, methods, and programs for detecting boundaries contained in captured images
KR20210103865A (en) * 2020-02-14 2021-08-24 삼성전자주식회사 Vanishing point extraction device and method of extracting vanishing point
US11380093B2 (en) * 2020-07-30 2022-07-05 GM Global Technology Operations LLC Detecting road edges by fusing aerial image and telemetry evidences

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040247157A1 (en) * 2001-06-15 2004-12-09 Ulrich Lages Method for preparing image information
US20100046840A1 (en) * 2008-08-20 2010-02-25 Canon Kabushiki Kaisha Image processing apparatus and image processing method for evaluating a plurality of image recognition processing units
US20100104199A1 (en) * 2008-04-24 2010-04-29 Gm Global Technology Operations, Inc. Method for detecting a clear path of travel for a vehicle enhanced by object detection
US20110274315A1 (en) * 2010-05-07 2011-11-10 Fan Shengyin Method, device, and computer-readable medium of object detection
US20120105638A1 (en) * 2006-11-07 2012-05-03 Rosco Inc. Camera system for large vehicles

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7983835B2 (en) * 2004-11-03 2011-07-19 Lagassey Paul J Modular intelligent transportation system
US8478480B2 (en) * 2006-10-27 2013-07-02 International Electronic Machines Corp. Vehicle evaluation using infrared data
CN101016053A (en) * 2007-01-25 2007-08-15 吉林大学 Warning method and system for preventing collision for vehicle on high standard highway
CN101975951B (en) * 2010-06-09 2013-03-20 北京理工大学 Field environment barrier detection method fusing distance and image information
JP5206752B2 (en) * 2010-08-30 2013-06-12 株式会社デンソー Driving environment recognition device
DE102011016770B4 (en) * 2011-04-12 2021-02-04 Daimler Ag Method for assisting a driver of a vehicle when changing lanes and device for carrying out the method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040247157A1 (en) * 2001-06-15 2004-12-09 Ulrich Lages Method for preparing image information
US20120105638A1 (en) * 2006-11-07 2012-05-03 Rosco Inc. Camera system for large vehicles
US20100104199A1 (en) * 2008-04-24 2010-04-29 Gm Global Technology Operations, Inc. Method for detecting a clear path of travel for a vehicle enhanced by object detection
US20100046840A1 (en) * 2008-08-20 2010-02-25 Canon Kabushiki Kaisha Image processing apparatus and image processing method for evaluating a plurality of image recognition processing units
US20110274315A1 (en) * 2010-05-07 2011-11-10 Fan Shengyin Method, device, and computer-readable medium of object detection

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11334605B2 (en) 2015-06-04 2022-05-17 Here Global B.V. Incremental update of compressed navigational databases

Also Published As

Publication number Publication date
CN103366154B (en) 2019-04-02
US20130265424A1 (en) 2013-10-10
DE102013205952A1 (en) 2013-10-24
DE102013205952B4 (en) 2020-03-19
CN103366154A (en) 2013-10-23

Similar Documents

Publication Publication Date Title
US9626599B2 (en) Reconfigurable clear path detection system
CN110175576B (en) Driving vehicle visual detection method combining laser point cloud data
US8750567B2 (en) Road structure detection and tracking
Keller et al. The benefits of dense stereo for pedestrian detection
US8447139B2 (en) Object recognition using Haar features and histograms of oriented gradients
US8890951B2 (en) Clear path detection with patch smoothing approach
US8634593B2 (en) Pixel-based texture-less clear path detection
JP5124592B2 (en) System and method for detecting and tracking a vehicle
US20170032676A1 (en) System for detecting pedestrians by fusing color and depth information
US20100097455A1 (en) Clear path detection using a vanishing point
US10867403B2 (en) Vehicle external recognition apparatus
Ohgushi et al. Road obstacle detection method based on an autoencoder with semantic segmentation
US20100098297A1 (en) Clear path detection using segmentation-based method
CN102609720A (en) Pedestrian detection method based on position correction model
US8559727B1 (en) Temporal coherence in clear path detection
CN111274926B (en) Image data screening method, device, computer equipment and storage medium
CN104615986A (en) Method for utilizing multiple detectors to conduct pedestrian detection on video images of scene change
JP5931662B2 (en) Road condition monitoring apparatus and road condition monitoring method
Gabb et al. High-performance on-road vehicle detection in monocular images
Álvarez et al. Perception advances in outdoor vehicle detection for automatic cruise control
Kurnianggoro et al. Camera and laser range finder fusion for real-time car detection
Morales Rosales et al. On-road obstacle detection video system for traffic accident prevention
WO2017077261A1 (en) A monocular camera cognitive imaging system for a vehicle
CN110765877B (en) Pedestrian detection method and system based on thermal imager and binocular camera
KR101936108B1 (en) Method and apparatus for detecting traffic sign

Legal Events

Date Code Title Description
AS Assignment

Owner name: GM GLOBAL TECHNOLOGY OPERATIONS LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHANG, WENDE;WANG, JINSONG;REEL/FRAME:028010/0139

Effective date: 20120403

AS Assignment

Owner name: WILMINGTON TRUST COMPANY, DELAWARE

Free format text: SECURITY AGREEMENT;ASSIGNOR:GM GLOBAL TECHNOLOGY OPERATIONS LLC;REEL/FRAME:030694/0500

Effective date: 20101027

AS Assignment

Owner name: GM GLOBAL TECHNOLOGY OPERATIONS LLC, MICHIGAN

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WILMINGTON TRUST COMPANY;REEL/FRAME:034287/0415

Effective date: 20141017

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4