US20180082131A1 - Vehicle detection, tracking and localization based on enhanced anti-perspective transformation - Google Patents

Vehicle detection, tracking and localization based on enhanced anti-perspective transformation Download PDF

Info

Publication number
US20180082131A1
US20180082131A1 US15/267,875 US201615267875A US2018082131A1 US 20180082131 A1 US20180082131 A1 US 20180082131A1 US 201615267875 A US201615267875 A US 201615267875A US 2018082131 A1 US2018082131 A1 US 2018082131A1
Authority
US
United States
Prior art keywords
vehicle
perspective
images
image
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/267,875
Other versions
US9928426B1 (en
Inventor
Hailiang Li
Zhi Bin LEI
Yang Liu
Man Yau CHIU
Kangheng Wu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hong Kong Applied Science and Technology Research Institute ASTRI
Original Assignee
Hong Kong Applied Science and Technology Research Institute ASTRI
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hong Kong Applied Science and Technology Research Institute ASTRI filed Critical Hong Kong Applied Science and Technology Research Institute ASTRI
Priority to US15/267,875 priority Critical patent/US9928426B1/en
Assigned to Hong Kong Applied Science and Technology Research Institute Company Limited reassignment Hong Kong Applied Science and Technology Research Institute Company Limited ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHIU, MAN YAU, LEI, ZHI BIN, LI, Hailiang, LIU, YANG, WU, KANGHENG
Priority to CN201680001279.2A priority patent/CN106462762B/en
Priority to PCT/CN2016/101702 priority patent/WO2018049704A1/en
Publication of US20180082131A1 publication Critical patent/US20180082131A1/en
Application granted granted Critical
Publication of US9928426B1 publication Critical patent/US9928426B1/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06K9/00785
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/54Surveillance or monitoring of activities, e.g. for recognising suspicious objects of traffic, e.g. cars on the road, trains or boats
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • G06K9/00771
    • G06K9/325
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/32Normalisation of the pattern dimensions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/247Aligning, centring, orientation detection or correction of the image by affine transforms, e.g. correction due to perspective effects; Quadrilaterals, e.g. trapezoids
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/625License plates

Definitions

  • aspects of the present invention relate to traffic information, and in particular, vehicle detection, tracking and localization of a distorted object in a real time situation.
  • Traffic video surveillance is an application of image acquisition through the use of a camera, transmission, control, display and other equipment and control software for fixed area monitoring, tracking and information processing traffic management information systems, including video transmission and video processing.
  • Such surveillance based on peer-to-peer (P2P) transmission is being considered an important aspect in modern smart city planning.
  • the P2P network structure ensures the transmission efficiency and robustness.
  • Vehicle detection is the preparative operation for the application later, such as vehicle recognition or classification, and vehicle speed estimation.
  • vehicle recognition or classification When capturing vehicles traveling on a highway, normally images are obtained in a perspective form, which may generate a distorted result for the vehicles. Distorted vehicle images will challenge subsequent detection, localization, and classification or recognition operations.
  • the detection efficiency is another issue to deal with since the size of each vehicle changes in relation to the image perspective. This can/will create misclassification during the vehicle detection or tracking process. Noting that vehicle tracking is a critical operation for vehicle counting and other operations, an object tracking algorithm results in the generation of a large scalable problem, such as a vehicle appearing to be more than ten times its actual size when traveling past a camera.
  • Embodiments of the present invention have been made to solve the above and/or other problems, and an exemplary aspect of the invention is to normalize the shape and size of a vehicle wherever it is in the camera view, and to transform the camera image into an enhanced anti-perspective domain to achieve real time detection with robust tracking performance.
  • FIG. 1 is a diagram of an intelligent traffic surveillance system according to an embodiment
  • FIG. 2 is a flowchart of an illustrating an embodiment of a method to generate enhanced anti-perspective images for vehicle detection, tracking and localization;
  • FIG. 3 is a perspective view of an image from a camera of traffic flow on a highway
  • FIG. 4 is a view illustrating an embodiment of a method for scale transform
  • FIG. 5 illustrates the benefit generating enhanced anti-perspective images for detection of vehicles on the highway
  • FIG. 6 illustrates the benefit of generating enhanced anti-perspective images for tracking of vehicles on the highway
  • FIGS. 7 a -7 b illustrate an example of vehicle recognition by using a landmark localization method
  • FIG. 8 illustrates an embodiment of a method for landmark localization of a vehicle by using a feature mapping function and a linear projection
  • FIG. 9 illustrates an example of locating license plates of vehicles on a highway by using the landmark localization method.
  • FIG. 10 is a view illustrating an embodiment of a method for generative model of super-resolution in the identification of a vehicle's license plate.
  • FIG. 11 shows a two stage scheme based on a vehicle plate detection and a vehicle-plate color extraction algorithm.
  • FIG. 1 shows an intelligent traffic surveillance system 10 according to an embodiment to monitor vehicle traffic on roads, such as highways.
  • An IP camera 12 captures video (which can be video or a series of still images) of vehicles traveling in a particular direction along a highway.
  • the highway can have 1, 3, 6 or even more lanes, or any number of lanes in between.
  • the IP camera 12 is connected to a communications network 14 , such as an internet network or other type of linked network.
  • Images (video) are sent to a processing center 16 through the communications network 14 , where algorithms are performed, such a vehicle detection, vehicle-face landmark detection, vehicle-plate detection, and other algorithms, to be described later on in this specification.
  • the processing center 16 may be a monitor and management center comprising one or more computers/computing devices.
  • the processing center 16 is connected to a data center 18 .
  • a control center 20 is operated by individuals to control the intelligent surveillance system 10 and is connected to the processing center 16 and the data center 18 .
  • a vehicle detector 32 , a vehicle-face landmark detector 34 , and a vehicle plate detector 36 are loaded onto the one or more computers of the processing center 16 , to perform corresponding algorithms as described later in the specification.
  • code having a set of instructions for performing the operations of the vehicle detector 32 , the vehicle-face landmark detector 34 , the vehicle plate detector 36 , or any other device performing the above-described algorithms may be stored on a non-transitory computer readable medium usable with any compatible hardware element.
  • These elements comprise various units, including an anti-perspective transform unit, a scale factor determining unit, an enhanced anti-perspective transformation unit, a vehicle detector, a vehicle tracker, a vehicle localizer, and a color recognizer.
  • FIG. 2 is a flowchart of a method to generate an enhanced anti-perspective image for vehicle detection, tracking and localization.
  • a method 100 of transforming a perspective image 108 to an enhanced anti-perspective image 112 of the video can be performed in the processing center 16 according to three main operations, namely “Operation 102 : Anti-perspective Transform”, “Operation 104 : Determine Scale Factor” and “Operation 106 : Scale Transform”.
  • the noise in an initial distribution of the detection box size 114 can be reduced.
  • FIG. 1 is a flowchart of a method to generate an enhanced anti-perspective image for vehicle detection, tracking and localization.
  • the vehicle detector 32 (which is the same as a face detector in traditional face detection) can detect vehicles among the many different possible vehicle sizes (i.e., different bounding box sizes of the detected vehicles) which is with a distribution (e.g., from 24-120 pixel size).
  • the vehicle detector 32 detects vehicles in an anti-perspective (transform) image 110 in a relatively narrow range, (e.g., from 60- 110 pixel size), which means the distribution of the size range is narrowed to half. Based on the distribution of the size range, it is possible to set an optimal detection range, which may speed up the detection more than two (2) times, (e.g., from 14.86 frames/second to 31.64 frames/second).
  • the detection size range setting and size distribution obtaining is a chicken-egg situation, so in the beginning, (i.e., in the beginning on getting the first 500 samples of detected car sizes, a relatively larger range is set, e.g., 20 ⁇ 150 pixel size.
  • a relatively larger range is set, e.g., 20 ⁇ 150 pixel size.
  • the new (relatively narrow) range e.g., 60 ⁇ 110 pixel size.
  • a conservative estimate range can be set to 55 ⁇ 115 (60 ⁇ 5, 110+5) pixel size. Based on the new range, new samples can be collected, and then the range can be updated by new samples.
  • an anti-perspective transform is a combination of a 2D scale, a rotation, and a translation transformation which can be described as:
  • [g h] is employed for translation and [c f] T is employed to generate a perspective transform.
  • This linear system can be solved using Gaussian elimination or other methods for the forward mapping coefficients a-h.
  • the determination of a scale factor includes a radial scale factor (RSF) and a tangential scale factor (TSF) relative to the traffic direction.
  • RSF radial scale factor
  • TSF tangential scale factor
  • only RSF is used since the camera 12 is installed on the radial direction of traffic flow. If the camera is not installed on the radial or tangential direction of traffic flow (but a perspective direction), as shown in FIG. 3 , both RSF and TSF should be used.
  • the methods to calculate RSF and TSF are the same, but are provided the corresponding directions.
  • FIG. 4 illustrates a method for RSF calculation.
  • RSF and TSF are calculated independently. They can be set by experience or by calculation.
  • the scale factor calculation is based on the anti-perspective transform 102 output, which is the anti-perspective image 110 .
  • the anti-perspective image 110 after the anti-perspective transform 102 , is divided into n parts (P1, P2. . . P3.
  • n is an even number
  • RSF n/2 ⁇ 1 1.
  • RSF (n+1)/2 1.
  • the size of n parts can equal each other or not.
  • the scale transform 106 receives the anti-perspective image 110 and outputs the enhanced anti perspective image 112 , and the scale transform 106 includes the radial scale transform and the tangential scale transform according to the vehicle shooting direction.
  • the methods to perform the radial scale transform and the tangential scale transform are the same as noted above.
  • Pi is extended in the radial direction by RSF i .
  • FIG. 4 illustrates a method 120 for the scale transform.
  • RSF will be used for illustration in the following descriptions.
  • the image after the anti-perspective transform operation 102 is divided into n parts (P 1 , P 2 . . . P n ).
  • n is an even number
  • RSF n/2 ⁇ 1 1
  • RSF (n+1)/2 1.
  • the size of n parts can be equal to each other or not RSF of the middle strip is always set to 1. But for the other parts, the RSF values are different from 1 according to the calculations (as shown in our current example) or according to manual configurations based on experience.
  • RSF 1 is equal to 0.5; RSF 2 is equal to 0.6; RSF 3 is equal to 0.8; RSF 4 is equal to 1; RSF 5 is equal to 1.2; RSF 6 is equal to 1.8; and RSF 7 is equal to 2.2.
  • P i is extended in the radial direction by RSF i .
  • FIG. 5 illustrates a benefit of generating enhanced anti-perspective images for detection of vehicles on the highway.
  • a detection box size 116 see FIGS. 2 and 5
  • the noise in an initial distribution of the detection box size 114 can be reduced.
  • FIG. 6 illustrates a benefit of generating enhanced anti-perspective images 172 from original perspective images 170 for tracking of vehicles on the highway. With increased tracking time, more chances of recognition of the vehicle type and details become available.
  • FIGS. 7 a -7 b illustrate an example of vehicle recognition by using a landmark localization method based upon input from the vehicle-face landmark detector 34 in which the features of the detection object will be extracted for analysis.
  • features such as a vehicle's license plate 180 and a vehicle's nameplate position based upon input from the vehicle-face landmark detector 34 can be extracted in a more accurate way. This approach can assist with the vehicle recognition process as described in the following paragraph.
  • FIG. 8 illustrates an embodiment of a method for landmark localization 130 of a vehicle by using a feature mapping function and a linear projection.
  • the current detection and localization algorithms have difficulty handling heavy deformations.
  • some fast localization methods such as a random forests based regression algorithm with local features can work efficiently on vehicle landmark localization, which may simplify vehicle classification or recognition in later stages.
  • FIG. 8 given a vehicle image I, there is a ground truth shape S 130 corresponding to this vehicle image.
  • t forests are trained (with several binary trees) to get a feature mapping function ⁇ t (l, S t ⁇ 1 ) 132 for each tree to generate local features 134 .
  • t also means that t stages are trained, and n forests are trained (here n forests are equal to n experts) in each stage.
  • the shape error ⁇ S is relative to shape-indexed features. The shape-indexed features are used for construction of a binary tree and then local features are generated from the leaves of a binary.
  • a linear projection matrix W t 136 is trained to get the feature mapping between the local binary feature and shape error ⁇ S by linear regression.
  • the learned linear projection (regression) matrix W t is described as the visualized matrix (element 138 ) in FIG. 8 . All the elements in the visualized matrix are the learned weights from linear regression processing. From the visualized matrix, the estimated shape S t ⁇ 1 and ground truth shape S
  • FIG. 9 illustrates an example 160 of locating license plates of vehicles in highway by using the landmark localization method to determine a vehicle plate localization area 162 . It is a significant challenge to detect a vehicle's license plate where the vehicles are on a highway moving at normal speeds and a camera captures images of the vehicles across more than six lanes. Besides, it is a time consuming process to locate small and blurred vehicles' license plates 164 in large images. In light of the landmark localization 130 as shown in FIG. 8 , the computation time can be reduced to 1/10 of that in prior art methods. On the other hand, creating a false positive 166 is another challenge in the process of locating vehicles' license plates.
  • the landmark localization for the vehicle's license plate area can give a more robust criterion for locating the vehicle's license plate, which means among all the detected potential objects, the one which is closer to the landmark localized area will be the best one.
  • FIG. 10 illustrates a method 180 for a generative model of super-resolution in the identification of a vehicle's license plate.
  • the objects such as vehicle license plates
  • the objects can be determined quickly and normalized into a similar size, which is helpful for object registration for later vehicle license plate image super-resolution.
  • a number of low-resolution images can differ in geometric transformations, lighting (photometric) transformations, camera blur (point-spread function) and image quantization and noise, it is still possible to generate an estimated high-resolution vehicle license plate, which can be used for identification and other purposes.
  • the low resolution image x 1 182 can be treated by an image processing method w 1 for example, registrations, uniform lighting, deburr, etc.
  • the low resolution image x 2 184 which can be treated by an image processing method w 2
  • the low resolution image x 3 186 can be treated by an image processing method w 3
  • the low resolution image x 4 188 can be treated by an image processing method w 2 .
  • the high resolution image y can be generated by ⁇ i ⁇ 1 4 wixi.
  • FIG. 11 shows a two stage scheme based on a vehicle plate detection and a vehicle-plate color extraction algorithm.
  • Input a captured video frame image (1920 ⁇ 1080 RGB) from the traffic surveillance system, an Adaboost+HOG trained vehicle detector D vehicle 32 , also an Adaboost+HOG trained vehicle-plate detector D vehicle-plate 36 , a random forests trained vehicle face landmark localizer L vehicle-face ;
  • 3 For each detected vehicle, the vehicle-plate detector D vehicle-plate 36 is used to detect any potential vehicle-plates in the bottom half part of vehicle area since it is with high probability that a vehicle-plate is located there; 4: For each detected vehicle, a vehicle face landmark localizer L vehicle-face is able to extract the landmark points, in which the potential vehicle-plate position can be obtained;
  • vehicle detection, tracking and localization are improved, particularly in environments where vehicles are high in volume and traveling at high speeds, such as on a highway.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Medical Informatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Traffic Control Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

A method for a vehicle management system with a perspective view camera generating perspective images of vehicles in a traffic direction, including performing an anti-perspective transform on the perspective images; separating each anti-perspective image into sub-images along one of a radial direction and a tangential direction relative to the traffic direction; determining a scale factor for each sub-image based upon measuring a scale of each vehicle at plural positions in one image using a frame difference method performing a scale transform for each sub-image using the corresponding scale factors for the sub-image; combining each of the scale transformed sub-images for each vehicle into corresponding enhanced anti-perspective images; performing vehicle detection for each vehicle based a combination of the enhanced anti-perspective images in which tracking is enhanced with an optimized detection box size range determined by the enhanced anti-perspective images; and performing vehicle tracking for each vehicle based on the combination of the enhanced anti-perspective images in which detecting is enhanced with the optimized detection box size range determined by the enhanced anti-perspective images.

Description

    BACKGROUND OF THE INVENTION Field of the Invention
  • Aspects of the present invention relate to traffic information, and in particular, vehicle detection, tracking and localization of a distorted object in a real time situation.
  • Description of the Related Art
  • Traffic video surveillance is an application of image acquisition through the use of a camera, transmission, control, display and other equipment and control software for fixed area monitoring, tracking and information processing traffic management information systems, including video transmission and video processing. Such surveillance based on peer-to-peer (P2P) transmission is being considered an important aspect in modern smart city planning. The P2P network structure ensures the transmission efficiency and robustness.
  • Vehicle detection is the preparative operation for the application later, such as vehicle recognition or classification, and vehicle speed estimation. When capturing vehicles traveling on a highway, normally images are obtained in a perspective form, which may generate a distorted result for the vehicles. Distorted vehicle images will challenge subsequent detection, localization, and classification or recognition operations. In order to detect distorted vehicle images in a perspective view, it is necessary to train a detector detecting a warped vehicle image to detect profile faces while a lot of training samples on profile faces should be collected, which may cause the detector to be bigger and more complex than the current ones.
  • The detection efficiency is another issue to deal with since the size of each vehicle changes in relation to the image perspective. This can/will create misclassification during the vehicle detection or tracking process. Noting that vehicle tracking is a critical operation for vehicle counting and other operations, an object tracking algorithm results in the generation of a large scalable problem, such as a vehicle appearing to be more than ten times its actual size when traveling past a camera.
  • Most recently, it is a common method to capture a target vehicle's license plate by using the camera for identification purposes, especially at a vehicle's entry or exit point to a highway, parking lot or other location. In this situation, the target vehicle is usually moving at a low speed such that the captured image is clear enough for identification and checking. In another situation, it is a difficult task to recognize the figures (such as the numerals and letters) on a vehicle's license plate on the highway where vehicles move at high speeds, particularly with relatively small vehicle size. The captured images, whether video or distinct still images (collectively called images), taken by the camera over the highway, tends to become either small or blurred. Thus, it is difficult to see the video or still images accurately.
  • SUMMARY OF THE INVENTION
  • Embodiments of the present invention have been made to solve the above and/or other problems, and an exemplary aspect of the invention is to normalize the shape and size of a vehicle wherever it is in the camera view, and to transform the camera image into an enhanced anti-perspective domain to achieve real time detection with robust tracking performance.
  • Additional aspects and/or advantages of the invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • These and/or other aspects and advantages of the invention will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
  • FIG. 1 is a diagram of an intelligent traffic surveillance system according to an embodiment;
  • FIG. 2 is a flowchart of an illustrating an embodiment of a method to generate enhanced anti-perspective images for vehicle detection, tracking and localization;
  • FIG. 3 is a perspective view of an image from a camera of traffic flow on a highway;
  • FIG. 4 is a view illustrating an embodiment of a method for scale transform;
  • FIG. 5 illustrates the benefit generating enhanced anti-perspective images for detection of vehicles on the highway;
  • FIG. 6 illustrates the benefit of generating enhanced anti-perspective images for tracking of vehicles on the highway;
  • FIGS. 7a-7b illustrate an example of vehicle recognition by using a landmark localization method;
  • FIG. 8 illustrates an embodiment of a method for landmark localization of a vehicle by using a feature mapping function and a linear projection;
  • FIG. 9 illustrates an example of locating license plates of vehicles on a highway by using the landmark localization method; and
  • FIG. 10 is a view illustrating an embodiment of a method for generative model of super-resolution in the identification of a vehicle's license plate; and
  • FIG. 11 shows a two stage scheme based on a vehicle plate detection and a vehicle-plate color extraction algorithm.
  • DETAILED DESCRIPTION OF THE EMBODIMENTS
  • According to the embodiments of the invention, it is possible to perform detection or tracking of vehicles on highways, traveling at normal highway speeds, even when the issue of distorted objects effects arises.
  • FIG. 1 shows an intelligent traffic surveillance system 10 according to an embodiment to monitor vehicle traffic on roads, such as highways. An IP camera 12 captures video (which can be video or a series of still images) of vehicles traveling in a particular direction along a highway. The highway can have 1, 3, 6 or even more lanes, or any number of lanes in between. The IP camera 12 is connected to a communications network 14, such as an internet network or other type of linked network. Images (video) are sent to a processing center 16 through the communications network 14, where algorithms are performed, such a vehicle detection, vehicle-face landmark detection, vehicle-plate detection, and other algorithms, to be described later on in this specification. The processing center 16 may be a monitor and management center comprising one or more computers/computing devices. The processing center 16 is connected to a data center 18. A control center 20 is operated by individuals to control the intelligent surveillance system 10 and is connected to the processing center 16 and the data center 18.
  • A vehicle detector 32, a vehicle-face landmark detector 34, and a vehicle plate detector 36 are loaded onto the one or more computers of the processing center 16, to perform corresponding algorithms as described later in the specification.
  • In an embodiment, code having a set of instructions for performing the operations of the vehicle detector 32, the vehicle-face landmark detector 34, the vehicle plate detector 36, or any other device performing the above-described algorithms may be stored on a non-transitory computer readable medium usable with any compatible hardware element. These elements comprise various units, including an anti-perspective transform unit, a scale factor determining unit, an enhanced anti-perspective transformation unit, a vehicle detector, a vehicle tracker, a vehicle localizer, and a color recognizer.
  • FIG. 2 is a flowchart of a method to generate an enhanced anti-perspective image for vehicle detection, tracking and localization. Referring to FIG. 2, a method 100 of transforming a perspective image 108 to an enhanced anti-perspective image 112 of the video can be performed in the processing center 16 according to three main operations, namely “Operation 102: Anti-perspective Transform”, “Operation 104: Determine Scale Factor” and “Operation 106: Scale Transform”. As shown in an optimized distribution of a detection box size 116, the noise in an initial distribution of the detection box size 114 can be reduced. As is shown in FIG. 2, the vehicle detector 32 (which is the same as a face detector in traditional face detection) can detect vehicles among the many different possible vehicle sizes (i.e., different bounding box sizes of the detected vehicles) which is with a distribution (e.g., from 24-120 pixel size). After the anti-perspective transform 102, the vehicle detector 32 detects vehicles in an anti-perspective (transform) image 110 in a relatively narrow range, (e.g., from 60-110 pixel size), which means the distribution of the size range is narrowed to half. Based on the distribution of the size range, it is possible to set an optimal detection range, which may speed up the detection more than two (2) times, (e.g., from 14.86 frames/second to 31.64 frames/second). The detection size range setting and size distribution obtaining is a chicken-egg situation, so in the beginning, (i.e., in the beginning on getting the first 500 samples of detected car sizes, a relatively larger range is set, e.g., 20˜150 pixel size. Once 500 samples are obtained, based on the distribution on these 500 samples, it is possible to get the new (relatively narrow) range, e.g., 60˜110 pixel size. Then, a conservative estimate range can be set to 55˜115 (60−5, 110+5) pixel size. Based on the new range, new samples can be collected, and then the range can be updated by new samples.
  • For operation 102 in FIG. 2, for the images of the video, an anti-perspective transform is a combination of a 2D scale, a rotation, and a translation transformation which can be described as:
  • ( x y w ) = ( a b c d e f g h i ) ( u v q )
  • Here the matrix coefficients can be separated as four parts.
  • ( a b d e )
  • is used for linear transformation, such as scaling, shearing and rotation. [g h] is employed for translation and [c f]T is employed to generate a perspective transform.
  • If w≠1, then q=1 can be chosen. The general form of a projective mapping is a rational linear mapping:
  • x = au + bv + c gu + hv + i y = du + ev + f gu + hv + i
  • Begin with a correspondence map (uk; vk)Tto (xk; yk)T for vertices numbered cyclically k=0; 1; 2; 3. All coordinates are assumed to be real (finite). To compute the forward mapping matrix, assuming that i=1, eight equations in the eight unknowns a-h can be generated:
  • x k = au k + bv k + c gu k + hv k + 1 au k + bv k + c - gu k x k - hv k x k = x k y k = du k + ev k + f gu k + hv k + 1 du k + ev k + f - gu k y k - hv k y k = y k
  • for k=0; 1; 2; 3. This can be rewritten as an 8×8 system:
  • ( u 0 v 0 1 0 0 0 - u 0 x 0 - v 0 x 0 u 1 v 1 1 0 0 0 - u 1 x 1 - v 1 x 1 u 2 v 2 1 0 0 0 - u 2 x 2 - v 2 x 2 u 3 v 3 1 0 0 0 - u 3 x 3 - v 3 x 3 0 0 0 u 0 v 0 1 - u 0 y 0 - v 0 y 0 0 0 0 u 1 v 1 1 - u 1 y 1 - v 1 y 1 0 0 0 u 2 v 2 1 - u 2 y 2 - v 2 y 2 0 0 0 u 3 v 3 1 - u 3 y 3 - v 3 y 3 ) ( a b c d e f g h ) = ( x 0 x 1 x 2 x 3 y 0 y 1 y 2 y 3 ) .
  • This linear system can be solved using Gaussian elimination or other methods for the forward mapping coefficients a-h.
  • For operation 104 in FIG. 2, the determination of a scale factor (SF) includes a radial scale factor (RSF) and a tangential scale factor (TSF) relative to the traffic direction. In this example, only RSF is used since the camera 12 is installed on the radial direction of traffic flow. If the camera is not installed on the radial or tangential direction of traffic flow (but a perspective direction), as shown in FIG. 3, both RSF and TSF should be used. The methods to calculate RSF and TSF are the same, but are provided the corresponding directions. FIG. 4 illustrates a method for RSF calculation. RSF and TSF are calculated independently. They can be set by experience or by calculation. The scale factor calculation is based on the anti-perspective transform 102 output, which is the anti-perspective image 110.
  • The anti-perspective image 110, after the anti-perspective transform 102, is divided into n parts (P1, P2. . . P3. When n is an even number, RSFn/2−1=1. When n is an odd number, RSF(n+1)/2=1. The size of n parts can equal each other or not.
  • For operation 106 in FIG. 2, the scale transform 106 receives the anti-perspective image 110 and outputs the enhanced anti perspective image 112, and the scale transform 106 includes the radial scale transform and the tangential scale transform according to the vehicle shooting direction. The methods to perform the radial scale transform and the tangential scale transform are the same as noted above. For each sub-image Pi, (I=1 . . . , 7 in this example), Pi is extended in the radial direction by RSFi. In this case, no tangential scale transform is used (i.e., TSFi=1 1 for i=1 . . . , 7)
  • FIG. 4 illustrates a method 120 for the scale transform. Referring to FIG. 4, since the traffic direction in the example is only radial, RSF will be used for illustration in the following descriptions. The image after the anti-perspective transform operation 102 is divided into n parts (P1, P2. . . Pn). When n is an even number, RSFn/2−1=1, and when n is an odd number, RSF(n+1)/2=1. The size of n parts can be equal to each other or not RSF of the middle strip is always set to 1. But for the other parts, the RSF values are different from 1 according to the calculations (as shown in our current example) or according to manual configurations based on experience.
  • Take the frame difference result when there is only one object moving in the vehicle view. The vehicle (object) length Li is defined by the object length measured by pixels when the rear of the object is inside Pi. If n is an even number, when n+1>i>n/2, RSFi=Li/Ln/2−1, and when n/2>i>0, RSFi=Ln/2−/Li. If n is an odd number, when n+1>i>(n+1)/2, RSFi=Li/L(n+1)/2, and when (n+1)/2>i>0, RSFi=L(n+1)/2/Li. In this example, as shown in FIG. 4, RSF1 is equal to 0.5; RSF2 is equal to 0.6; RSF3 is equal to 0.8; RSF4 is equal to 1; RSF5 is equal to 1.2; RSF6 is equal to 1.8; and RSF7 is equal to 2.2. For each sub-image Pi, (i=1 . . . , 7 in this example), Pi is extended in the radial direction by RSFi. In this example, no tangential scale transform is used (i.e., TSFi=1 for i=1, . . . , 7).
  • FIG. 5 illustrates a benefit of generating enhanced anti-perspective images for detection of vehicles on the highway. As can be seen in the optimized distribution of a detection box size 116 (see FIGS. 2 and 5) can be narrowed to half of an original range, the noise in an initial distribution of the detection box size 114 can be reduced.
  • FIG. 6 illustrates a benefit of generating enhanced anti-perspective images 172 from original perspective images 170 for tracking of vehicles on the highway. With increased tracking time, more chances of recognition of the vehicle type and details become available.
  • FIGS. 7a-7b illustrate an example of vehicle recognition by using a landmark localization method based upon input from the vehicle-face landmark detector 34 in which the features of the detection object will be extracted for analysis. Referring to FIGS. 7a-7b , aided by the landmark localization method, features, such as a vehicle's license plate 180 and a vehicle's nameplate position based upon input from the vehicle-face landmark detector 34 can be extracted in a more accurate way. This approach can assist with the vehicle recognition process as described in the following paragraph.
  • FIG. 8 illustrates an embodiment of a method for landmark localization 130 of a vehicle by using a feature mapping function and a linear projection. Mainly due to the perspective image 108, the current detection and localization algorithms have difficulty handling heavy deformations. Once the image 108 of the vehicle is transformed an into anti-perspective image, some fast localization methods, such as a random forests based regression algorithm with local features can work efficiently on vehicle landmark localization, which may simplify vehicle classification or recognition in later stages. Referring to FIG. 8, given a vehicle image I, there is a ground truth shape S 130 corresponding to this vehicle image. t forests are trained (with several binary trees) to get a feature mapping function φt (l, St−1) 132 for each tree to generate local features 134. Here t also means that t stages are trained, and n forests are trained (here n forests are equal to n experts) in each stage. The error of the last stage ΔSt is calculated from ground truth and previous stage estimated shape St−1, and also defined as the target shape increments {ΔSt=S−St−1}. The shape error ΔS is relative to shape-indexed features. The shape-indexed features are used for construction of a binary tree and then local features are generated from the leaves of a binary. A linear projection matrix W t 136 is trained to get the feature mapping between the local binary feature and shape error ΔS by linear regression. The learned linear projection (regression) matrix Wt is described as the visualized matrix (element 138) in FIG. 8. All the elements in the visualized matrix are the learned weights from linear regression processing. From the visualized matrix, the estimated shape St−1 and ground truth shape S
  • FIG. 9 illustrates an example 160 of locating license plates of vehicles in highway by using the landmark localization method to determine a vehicle plate localization area 162. It is a significant challenge to detect a vehicle's license plate where the vehicles are on a highway moving at normal speeds and a camera captures images of the vehicles across more than six lanes. Besides, it is a time consuming process to locate small and blurred vehicles' license plates 164 in large images. In light of the landmark localization 130 as shown in FIG. 8, the computation time can be reduced to 1/10 of that in prior art methods. On the other hand, creating a false positive 166 is another challenge in the process of locating vehicles' license plates. For instance, a vehicle's fan is near the vehicle's license plate and therefore may trigger the false positive 166. Therefore, the landmark localization for the vehicle's license plate area can give a more robust criterion for locating the vehicle's license plate, which means among all the detected potential objects, the one which is closer to the landmark localized area will be the best one.
  • FIG. 10 illustrates a method 180 for a generative model of super-resolution in the identification of a vehicle's license plate. In an anti-perspective domain, the objects, such as vehicle license plates, can be determined quickly and normalized into a similar size, which is helpful for object registration for later vehicle license plate image super-resolution. Given that a number of low-resolution images can differ in geometric transformations, lighting (photometric) transformations, camera blur (point-spread function) and image quantization and noise, it is still possible to generate an estimated high-resolution vehicle license plate, which can be used for identification and other purposes.
  • Referring to FIG. 10, for the purpose of illustration, assuming a vehicle's license plate can be separated into four sections, the low resolution image x1 182 can be treated by an image processing method w1 for example, registrations, uniform lighting, deburr, etc. The same holds true for the low resolution image x2 184 which can be treated by an image processing method w2, the low resolution image x3 186 can be treated by an image processing method w3, and the low resolution image x4 188 can be treated by an image processing method w2. Then, the high resolution image y can be generated by Σi−1 4 wixi.
  • FIG. 11 shows a two stage scheme based on a vehicle plate detection and a vehicle-plate color extraction algorithm. 1: Input: a captured video frame image (1920×1080 RGB) from the traffic surveillance system, an Adaboost+HOG trained vehicle detector Dvehicle 32, also an Adaboost+HOG trained vehicle-plate detector Dvehicle-plate 36, a random forests trained vehicle face landmark localizer Lvehicle-face; 2: For each input image, the vehicle detector Dvehicle 32 is employed to detect any potential vehicles in the image; 3: For each detected vehicle, the vehicle-plate detector Dvehicle-plate 36 is used to detect any potential vehicle-plates in the bottom half part of vehicle area since it is with high probability that a vehicle-plate is located there; 4: For each detected vehicle, a vehicle face landmark localizer Lvehicle-face is able to extract the landmark points, in which the potential vehicle-plate position can be obtained; 5: Compare all the vehicle-plate candidates obtained by the vehicle-plate detector Dvehicle-plate 36, the one which obtains the max value of an overlapping area will be the most approximated one; 6: For the extracted blurred vehicle-plate, a cartoon texture decomposition algorithm is used to extract the cartoon part, in which we can extract the RGB color histogram to determine the color type by a trained SVM classifier; and 7: Output: the color type of an extracted blurred vehicle-plate.
  • As a result, of the foregoing system and algorithms, vehicle detection, tracking and localization are improved, particularly in environments where vehicles are high in volume and traveling at high speeds, such as on a highway.
  • Although a few embodiments of the present invention have been shown and described, it would be appreciated by those skilled in the art that changes may be made in this embodiment without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.

Claims (17)

What is claimed is:
1. A method for a vehicle management system with a perspective view camera generating perspective images of vehicles in a traffic direction, comprising:
performing an anti-perspective transform on the perspective images;
separating each anti-perspective image into sub-images along one of a radial direction and a tangential direction relative to the traffic direction;
determining a scale factor for each sub-image based upon measuring a scale of each vehicle at plural positions in one image using a frame difference method performing a scale transform for each sub-image using the corresponding scale factors for the sub-image;
combining each of the scale transformed sub-images for each vehicle into corresponding enhanced anti-perspective images;
performing vehicle detection for each vehicle based a combination of the enhanced anti-perspective images in which detection is enhanced with an optimized detection box size range determined by the enhanced anti-perspective images; and
performing vehicle tracking for each vehicle based on the combination of the enhanced anti-perspective images in which detection is enhanced with the optimized detection box size range determined by the enhanced anti-perspective images.
2. The method of claim 1, further comprising;
performing vehicle localization for each vehicle by integrating each of the detected and/or tracked results from a same one of the vehicles.
3. The method of claim 1, further comprising:
performing landmark localization to localize a vehicle plate area for each vehicle based upon at least one of the detected and tracked results; and
locating a vehicle license plate around the localized plate area of each vehicle based upon the landmark localization.
4. The method of claim 2, further comprising:
performing landmark localization to localize a vehicle plate area for each vehicle based upon at least one of the detected and tracked results; and
locating a vehicle license plate around the localized plate area of each vehicle based upon the landmark localization.
5. The method of claim 3, further comprising generating higher resolution images based on images of the located vehicle license plates.
6. The method of claim 4, further comprising generating higher resolution images based on images of the located vehicle license plates.
7. The method of claim 1, wherein the performing of the anti-perspective transform for the perspective images comprises:
performing an anti-perspective transform which is a combination of a 2D scale, a rotation, and a translation transformation which is described as:
( x y w ) = ( a b c d e f g h i ) ( u v q )
the matrix coefficients separated as four parts.
( a b d e )
is used for linear transformation, including scaling, shearing and rotation; [g h] is employed for translation; and [c f]T is employed to generate a perspective transform
wherein:
if w≠1, then q=1 is chosen, a general form of a projective mapping is a rational linear mapping:
x = au + bv + c gu + hv + i y = du + ev + f gu + hv + i
a correspondence map (Uk; Vk)T to (Xk; Yk)T is for vertices numbered cyclically k=0; 1; 2; 3;
all coordinates are assumed to be real;
to compute the forward mapping matrix, assuming that i=1, eight equations in the eight unknowns a-h are generated:
x k = au k + bv k + c gu k + hv k + 1 au k + bv k + c - gu k x k - hv k x k = x k y k = du k + ev k + f gu k + hv k + 1 du k + ev k + f - gu k y k - hv k y k = y k
for k=0; 1; 2; 3. This can be rewritten as an 8×8 system:
( u 0 v 0 1 0 0 0 - u 0 x 0 - v 0 x 0 u 1 v 1 1 0 0 0 - u 1 x 1 - v 1 x 1 u 2 v 2 1 0 0 0 - u 2 x 2 - v 2 x 2 u 3 v 3 1 0 0 0 - u 3 x 3 - v 3 x 3 0 0 0 u 0 v 0 1 - u 0 y 0 - v 0 y 0 0 0 0 u 1 v 1 1 - u 1 y 1 - v 1 y 1 0 0 0 u 2 v 2 1 - u 2 y 2 - v 2 y 2 0 0 0 u 3 v 3 1 - u 3 y 3 - v 3 y 3 ) ( a b c d e f g h ) = ( x 0 x 1 x 2 x 3 y 0 y 1 y 2 y 3 )
the linear system being solved using Gaussian elimination or other methods for the forward mapping coefficients a-h.
8. The method of claim 1, wherein the performing of the scale transform for each sub-image comprises:
dividing each anti-perspective image into n parts (P1, P2. . . Pn); wherein when n is an even number, RSFn/2−1=1, and when n is an odd number, RSF(n+1)/2=1; a size of n parts can be equal to each other or not;
taking the frame difference result when there is only one vehicle moving in camera view, wherein vehicle length Li is defined by the vehicle length measured by pixels when the rear of the vehicle is inside Pi; wherein if n is an even number, when n+1>i>n/2, RSFi =Li/Ln/2−1, and when n/2>i>0, RSFi=Ln/2−1/Li; and if n is an odd number, when n+1>i>(n+1)/2, RSFi=Li/L(n+1)/2, and when (n+1)/2>i>0, RSFi =L(n+1)/2/Li.
9. The method of claim 7, wherein the performing of the scale transform for each sub-image comprises:
dividing each anti-perspective image into n parts (P1, P2. . . Pn); wherein when n is an even number, RSFn/2−1=1, and when n is an odd number, RSF(n+1)/2=1; a size of n parts can be equal to each other or not;
taking the frame difference result when there is only one vehicle moving in camera view, wherein vehicle length Li is defined by the vehicle length measured by pixels when the rear of the vehicle is inside Pi; wherein if n is an even number, when n+1>i>n/2, RSFi =Li/Ln/2−1, and when n/2>i>0, RSFi=Ln/2−1/Li; and if n is an odd number, when n+1>i>(n+1)/2, RSFi=Li/L(n+1)/2, and when (n+1)/2>i>0, RSFi=L(n+1)/2/Li.
10. The method of claim 2, wherein the performing of the vehicle localization for each vehicle comprises:
given a vehicle image I, there is a ground truth shape S corresponding to this vehicle image, training t forests to get a feature mapping function φt (l, St-1) for each tree to generate local features, wherein t indicates that t stages are trained, and n forests are trained (here n forests are equal n experts) in each stage;
calculating the error of the last stage ΔSt from ground truth and previous stage estimated shape St−1, and also defined as the target shape increments {ΔSt=S−St−1}, wherein the shape error ΔS is relative to shape-indexed features; the shape-indexed features being used for construction of a binary tree and then local features are generated from the leaves of a binary; and
training a linear projection matrix Wt to get the feature mapping between the local binary feature and shape error ΔS by linear regression.
11. The method of claim 4, wherein the performing of the vehicle localization for each vehicle comprises:
given a vehicle image I, there is a ground truth shape S corresponding to this vehicle image, training t forests to get a feature mapping function φt (l, St−1) for each tree to generate local features, wherein t indicates that t stages are trained, and n forests are trained (here n forests are equal n experts) in each stage;
calculating the error of the last stage ΔSt from ground truth and previous stage estimated shape St−1, and also defined as the target shape increments {ΔS=S−St−1}, wherein the shape error ΔS is relative to shape-indexed features; the shape-indexed features being used for construction of a binary tree and then local features are generated from the leaves of a binary; and
training a linear projection matrix Wt to get the feature mapping between the local binary feature and shape error ΔS by linear regression.
12. The method of claim 3, further comprising recognizing colors of the located vehicle license plates based on images of the located vehicle license plates.
13. The method of claim 4, further comprising recognizing colors of the located vehicle license plates based on images of the located vehicle license plates.
14. An intelligent traffic surveillance system for vehicle management with a perspective view camera generating perspective images of vehicles in a traffic direction, comprising:
an anti-perspective transform unit to perform an anti-perspective transform on the perspective images, and separating each anti-perspective image into sub-images along one of a radial direction and a tangential direction relative to the traffic direction;
a scale factor determining unit to determine a scale factor for each sub-image based upon measuring a scale of each vehicle at plural in one image using a frame difference method performing a scale transform for each sub-image using the corresponding scale factors for the sub-image;
an enhanced anti-perspective transformation unit combining each of the scale transformed sub-images for each vehicle into corresponding enhanced anti-perspective images;
a vehicle detector performing vehicle detection for each vehicle based a combination of the enhanced anti-perspective images in which detecting is enhanced with an optimized detection box size range determined by the enhanced anti-perspective images; and
a vehicle tracker performing vehicle tracking for each vehicle based on the combination of the enhanced anti-perspective images in which tracking is enhanced with the optimized detection box size range determined by the enhanced anti-perspective images.
15. The traffic surveillance system of claim 14, further comprising;
a vehicle localizer performing vehicle localization for each vehicle by integrating each of the detected and/or tracked results from a same one of the vehicles.
16. The traffic surveillance system of claim 15, wherein the vehicle localizer:
performs landmark localization to localize a vehicle plate area for each vehicle based upon at least one of the detected and tracked results; and
locates a vehicle license plate around the localized plate area of each vehicle based upon the landmark localization.
17. The traffic surveillance system of claim 16, further comprising a color recognizer recognizing colors of the located vehicle license plates based on images of the located vehicle license plates.
US15/267,875 2016-09-16 2016-09-16 Vehicle detection, tracking and localization based on enhanced anti-perspective transformation Active 2036-10-12 US9928426B1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US15/267,875 US9928426B1 (en) 2016-09-16 2016-09-16 Vehicle detection, tracking and localization based on enhanced anti-perspective transformation
CN201680001279.2A CN106462762B (en) 2016-09-16 2016-10-10 Based on vehicle detection, tracking and the positioning for enhancing anti-perspective transform
PCT/CN2016/101702 WO2018049704A1 (en) 2016-09-16 2016-10-10 Vehicle detection, tracking and localization based on enhanced anti-perspective transformation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/267,875 US9928426B1 (en) 2016-09-16 2016-09-16 Vehicle detection, tracking and localization based on enhanced anti-perspective transformation

Publications (2)

Publication Number Publication Date
US20180082131A1 true US20180082131A1 (en) 2018-03-22
US9928426B1 US9928426B1 (en) 2018-03-27

Family

ID=61619821

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/267,875 Active 2036-10-12 US9928426B1 (en) 2016-09-16 2016-09-16 Vehicle detection, tracking and localization based on enhanced anti-perspective transformation

Country Status (2)

Country Link
US (1) US9928426B1 (en)
WO (1) WO2018049704A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109543598A (en) * 2018-11-20 2019-03-29 哈尔滨工程大学 A kind of highway accident response and warning system and method based on image recognition
CN110110608A (en) * 2019-04-12 2019-08-09 国网浙江省电力有限公司嘉兴供电公司 The fork truck speed monitoring method and system of view-based access control model under a kind of overall view monitoring
US11087119B2 (en) * 2018-05-16 2021-08-10 Gatekeeper Security, Inc. Facial detection and recognition for pedestrian traffic
WO2022046908A1 (en) * 2020-08-25 2022-03-03 Axon Enterprise, Inc. Systems and methods for rapid license plate reading
US20220092785A1 (en) * 2018-12-18 2022-03-24 Agfa Nv Method of decomposing a radiographic image into sub-images of different types
US20220101017A1 (en) * 2020-09-30 2022-03-31 Rekor Systems, Inc. Systems and methods for efficient information extraction in traffic monitoring
US11501541B2 (en) 2019-07-10 2022-11-15 Gatekeeper Inc. Imaging systems for facial detection, license plate reading, vehicle overview and vehicle make, model and color detection
US20220368810A1 (en) * 2021-05-14 2022-11-17 Denso Ten Limited Image processing device, image processing method, and computer readable medium
US11538257B2 (en) 2017-12-08 2022-12-27 Gatekeeper Inc. Detection, counting and identification of occupants in vehicles
US11736663B2 (en) 2019-10-25 2023-08-22 Gatekeeper Inc. Image artifact mitigation in scanners for entry control systems

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7639841B2 (en) * 2004-12-20 2009-12-29 Siemens Corporation System and method for on-road detection of a vehicle using knowledge fusion
KR101075615B1 (en) * 2006-07-06 2011-10-21 포항공과대학교 산학협력단 Apparatus and method for generating a auxiliary information of moving vehicles for driver
US8760472B2 (en) 2008-04-01 2014-06-24 Apple Inc. Pixel transforms
JP2012212282A (en) * 2011-03-31 2012-11-01 Honda Elesys Co Ltd Road surface state detection device, road surface state detection method, and road surface state detection program
CN102254318B (en) 2011-04-08 2013-01-09 上海交通大学 Method for measuring speed through vehicle road traffic videos based on image perspective projection transformation
US8611598B2 (en) * 2011-07-26 2013-12-17 Harman International (China) Holdings Co., Ltd. Vehicle obstacle detection system
DE102011053052B3 (en) 2011-08-26 2013-02-28 Jenoptik Robot Gmbh Method and device for identifying motor vehicles for traffic monitoring
IN2015KN00315A (en) * 2012-07-27 2015-07-10 Nissan Motor
CN102902945B (en) 2012-09-28 2015-04-15 南京汇兴博业数字设备有限公司 Distortion correction method of outer contour based on quick response matrix code
PL402304A1 (en) 2012-12-31 2014-07-07 Instytut Badawczy Dróg i Mostów Vehicle identification method and vehicle identification system
US9412031B2 (en) * 2013-10-16 2016-08-09 Xerox Corporation Delayed vehicle identification for privacy enforcement
US9165361B1 (en) * 2014-03-13 2015-10-20 Raytheon Company Video tracking with jitter, slewing, or zoom
US20160019698A1 (en) 2014-07-21 2016-01-21 Florida Atlantic University Systems and methods for people counting in sequential images
JP6440411B2 (en) * 2014-08-26 2018-12-19 日立オートモティブシステムズ株式会社 Object detection device
US9811743B2 (en) * 2015-06-29 2017-11-07 Sharp Laboratories Of America, Inc. Tracking road boundaries
US10055651B2 (en) * 2016-03-08 2018-08-21 Magna Electronics Inc. Vehicle vision system with enhanced lane tracking

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11538257B2 (en) 2017-12-08 2022-12-27 Gatekeeper Inc. Detection, counting and identification of occupants in vehicles
US11087119B2 (en) * 2018-05-16 2021-08-10 Gatekeeper Security, Inc. Facial detection and recognition for pedestrian traffic
CN109543598A (en) * 2018-11-20 2019-03-29 哈尔滨工程大学 A kind of highway accident response and warning system and method based on image recognition
US20220092785A1 (en) * 2018-12-18 2022-03-24 Agfa Nv Method of decomposing a radiographic image into sub-images of different types
CN110110608A (en) * 2019-04-12 2019-08-09 国网浙江省电力有限公司嘉兴供电公司 The fork truck speed monitoring method and system of view-based access control model under a kind of overall view monitoring
US11501541B2 (en) 2019-07-10 2022-11-15 Gatekeeper Inc. Imaging systems for facial detection, license plate reading, vehicle overview and vehicle make, model and color detection
US11736663B2 (en) 2019-10-25 2023-08-22 Gatekeeper Inc. Image artifact mitigation in scanners for entry control systems
WO2022046908A1 (en) * 2020-08-25 2022-03-03 Axon Enterprise, Inc. Systems and methods for rapid license plate reading
US11978260B2 (en) 2020-08-25 2024-05-07 Axon Enterprise, Inc. Systems and methods for rapid license plate reading
US20220101017A1 (en) * 2020-09-30 2022-03-31 Rekor Systems, Inc. Systems and methods for efficient information extraction in traffic monitoring
US11804046B2 (en) * 2020-09-30 2023-10-31 Rekor Systems, Inc. Systems and methods for efficient information extraction in traffic monitoring
US20220368810A1 (en) * 2021-05-14 2022-11-17 Denso Ten Limited Image processing device, image processing method, and computer readable medium
US11902487B2 (en) * 2021-05-14 2024-02-13 Denso Ten Limited Image processing device, image processing method, and computer readable medium

Also Published As

Publication number Publication date
US9928426B1 (en) 2018-03-27
WO2018049704A1 (en) 2018-03-22

Similar Documents

Publication Publication Date Title
US9928426B1 (en) Vehicle detection, tracking and localization based on enhanced anti-perspective transformation
US11643076B2 (en) Forward collision control method and apparatus, electronic device, program, and medium
JP6670071B2 (en) Vehicle image recognition system and corresponding method
Hu et al. Moving object detection and tracking from video captured by moving camera
JP6095018B2 (en) Detection and tracking of moving objects
CN103325112B (en) Moving target method for quick in dynamic scene
Liu et al. Combining statistical hough transform and particle filter for robust lane detection and tracking
US9607228B2 (en) Parts based object tracking method and apparatus
Hoang et al. Enhanced detection and recognition of road markings based on adaptive region of interest and deep learning
US20120121166A1 (en) Method and apparatus for three dimensional parallel object segmentation
CN105654031B (en) System and method for object detection
EP3182370B1 (en) Method and device for generating binary descriptors in video frames
JP5936561B2 (en) Object classification based on appearance and context in images
Yang et al. Multiple object tracking with kernelized correlation filters in urban mixed traffic
Ji et al. RGB-D SLAM using vanishing point and door plate information in corridor environment
Kachach et al. Hybrid three-dimensional and support vector machine approach for automatic vehicle tracking and classification using a single camera
Zaarane et al. Vehicle to vehicle distance measurement for self-driving systems
CN112837404B (en) Method and device for constructing three-dimensional information of planar object
Yang et al. Edge-based moving object tracking algorithm for an embedded system
Tasson et al. FPGA-based pedestrian detection under strong distortions
Umamaheswaran et al. Stereo vision based speed estimation for autonomous driving
Date et al. Vision based lane detection and departure warning system
CN106462762A (en) Detection, tracking and positioning of vehicle based on enhanced inverse perspective mapping
Płaczek et al. Model based vehicle extraction and tracking for road traffic control
Mandlik et al. A review on lane detection and tracking techniques

Legal Events

Date Code Title Description
AS Assignment

Owner name: HONG KONG APPLIED SCIENCE AND TECHNOLOGY RESEARCH

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LI, HAILIANG;LEI, ZHI BIN;LIU, YANG;AND OTHERS;REEL/FRAME:039767/0559

Effective date: 20160915

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4