WO2017068397A1 - A moving object detection method - Google Patents

A moving object detection method Download PDF

Info

Publication number
WO2017068397A1
WO2017068397A1 PCT/IB2015/058188 IB2015058188W WO2017068397A1 WO 2017068397 A1 WO2017068397 A1 WO 2017068397A1 IB 2015058188 W IB2015058188 W IB 2015058188W WO 2017068397 A1 WO2017068397 A1 WO 2017068397A1
Authority
WO
WIPO (PCT)
Prior art keywords
moving object
object detection
image
detection method
storage unit
Prior art date
Application number
PCT/IB2015/058188
Other languages
French (fr)
Inventor
Murat GEVREKCI
Mehmet Umut Demircin
Erkan OKUYAN
Original Assignee
Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi filed Critical Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi
Priority to PCT/IB2015/058188 priority Critical patent/WO2017068397A1/en
Publication of WO2017068397A1 publication Critical patent/WO2017068397A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/269Analysis of motion using gradient-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/94Hardware or software architectures specially adapted for image or video understanding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Definitions

  • the present invention relates to moving object detection methods for surveillance systems.
  • Moving object detection application is critical for both airborne and land-based surveillance systems with a fixed and a rotating imaging sensor.
  • the aim of such systems is to report moving objects to an operator for awareness against possible threats, or to track multiple moving targets automatically.
  • Moving object detection registers a sequence of images acquired by an imaging sensor and extracts possible moving regions by taking difference of this geometrically registered image with respect to selected reference.
  • Image residuals (difference with respect to the geometrically registered reference) of an aligned sequence include both moving and non-moving objects. The major reasons of false alarms are parallax effect and contrast changes among consecutive images.
  • False alarms also occur due to misalignment of high contrast image structures. Gradient based suppression of residuals dampens such false alarms. False alarms manifest itself by misalignment of thin line structures among consecutive images such as road sides, roof edges.
  • US2008278584 discloses a moving object detection apparatus and method by using optical flow analysis.
  • the apparatus includes four modules of image capturing, image aligning, pixel matching, and moving object detection.
  • Plural images are successively inputted under a camera.
  • frame relationship on the neighbouring images is estimated.
  • a set of warping parameter is further estimated.
  • the background areas of the neighbouring images are aligned to obtain an aligned previous image.
  • a corresponding motion vector for each pixel on the neighbouring images is traced.
  • Optical flow technique used in US2008278584 is utilized for registration purposes.
  • US2011116682 discloses an object detection method and an object detection system, suitable for detecting moving object information of a video stream having a plurality of images, are provided.
  • the method performs a moving object foreground detection on each of the images, so as to obtain a first foreground detection image comprising a plurality of moving objects.
  • the method also performs a texture object foreground detection on each of the images, so as to obtain a second foreground detection image comprising a plurality of texture objects.
  • the moving objects in the first foreground detection image and the texture objects in the second foreground detection image are selected and filtered, and then the remaining moving objects or texture objects after the filtering are output as real moving object information.
  • US2009052740 discloses a moving object detecting device measures a congestion degree of a space and utilizes the congestion degree for tracking.
  • a direction measured by a laser range sensor is heavily weighted when the congestion degree is low.
  • a sensor fusion is performed by heavily weighting a direction measured by an image processing on a captured image to obtain a moving object estimating direction, and obtains a distance by the laser range sensor in the moving object estimating direction.
  • the United States patent document numbered US20150030202 discloses a method for an intelligent video processing system based on object detection.
  • the method includes receiving an input video sequence corresponding to a video program, obtaining a plurality of frames of the input video sequence, and obtaining a computational constraint and a temporal rate constraint.
  • the method also includes determining one or more regions of interest (ROIs) of the plurality of frames based on the computational constraint and temporal rate constraint, and selecting a desired set of frames from the plurality of frames based on the ROIs such that the desired set of frames substantially represent a view path of the plurality of frames.
  • the method includes detecting object occurrences from the desired set of frames based on the selected desired set of frames such that a computational cost and a number of frames for detecting the object occurrences are under the computational constraint and temporal rate constraint.
  • the objective of the present invention is to realize a moving object detection method which reduces the false alarms resulting from parallax effect and contrast changes.
  • Another objective of the present invention is to realize a moving object detection method which utilizes optical flow to suppress residuals, without any relation to registration, in accumulative manner using the camera motion.
  • Figure 1 is the flowchart of the method.
  • Figure 2 is a single image taken from a video sequence
  • Figure 3 is Residual history image (RHI) accumulated through time of a video sequence
  • Figure 4 is optical flow vectors
  • step 106 Suppressing accumulated residuals calculated in step 103, by using the stored optical flow projections,
  • Residual storage unit 1020 Residual storage unit
  • acquired images during surveillance are recorded to an image storage unit (1010), in step 101.
  • the image storage unit (1010) is a circular buffer.
  • Images in the image storage unit (1010) are geometrically registered onto a current reference image in step 102.
  • Geometric registration is the processing block which aligns an input image onto a given reference spatially.
  • Geometric registration parameters of the images, which are stored in the image storage unit (1010) are stored in a geometric transformations storage unit (1021). Difference between registered images and the reference frame are also calculated in step 102 and sum of the residuals is stored in residual storage unit (1020).
  • the residual storage unit (1020), and geometric transformation storage unit are circular buffers. Residuals are defined in detail in following paragraphs. Residuals are accumulated in step 103 and result is written to the Residual History Image (RHI).
  • RHI Residual History Image
  • P_ x denotes the geometric projection computed from time ( ⁇ -1) to ( ⁇ ).
  • D ' is the temporal decay to penalize pixels with no motion.
  • D(x,y,x,k) is residual at location (x,y) at time instance ( ⁇ ) with a time gap of (k) (i.e. residual is computed between image ⁇ and ⁇ +k). Residual is the difference of an image acquired at time ⁇ and geometrically registered image that is acquired at time ⁇ +k, where "k” may be positive or negative.
  • T is the difference threshold to declare that there is putative motion among consecutive images.
  • / is the image acquired, whose time instance is denoted by 1 , and a pixel coordinate is represented by x,y.
  • H F (x, y, r ) is the pixel value of a Motion history image (MHI) at pixel (x,y) computed at time instance ( 1).
  • MHI equation in [Collins, R. 2006, Moving Object Localization in Thermal Imagery by Forward-backward MHI] is slightly modified to incorporate multiple input images to be used for computing residual of several images captured at different times.
  • W ⁇ is the weight assigned to each residual for MHI computation.
  • a linearly decreasing weighting function is used in the preferred embodiment of the invention.
  • the reference image is selected from the image storage unit (1010). There is an inherent tradeoff over the selection of reference image index. Selecting the reference image index as the most current index produces less false alarms, however has high probability of misdetecting the slowly moving targets. Selecting the reference image acquired at distant time instances enables catching slowly moving targets at the expense of extreme false alarm due to parallax.
  • reference image index has to be selected depending on application requirements. Using every image in MHI as proposed overcomes the presented challenges. Optical flow is calculated for the most recently acquired image with respect to a previous image with a specified time delay in step 104 and result is stored in Flow History Storage Unit (1040). An error constant is needed to calculate optical flows (u & v)
  • ⁇ and ⁇ are the image gradients in horizontal and vertical dimensions, respectively.
  • £ 3 ⁇ 4 is the temporal image difference with respect to the previous frame.
  • ⁇ 1 jj [ 2 e c 2 + ⁇ )dxdy
  • Optical flows are projected to the direction of camera motion in step 105 and these projected flow vectors (2D vector arrays) are stored in Flow history storage unit (1040) after being warped by the inverse of the estimated camera motion.
  • a pixel in Flow history image (FHI) obtained after the process in step 105, accumulates flows corresponding to the same 3D coordinate throughout time.
  • Flow history image (FHI) holds the optical flow vector (u,v) for each pixel of the image.
  • Optical flow vector of each pixel is alpha-filtered temporally and stored on Flow history storage unit (1040) for temporal consistency.
  • optical flow vector tuple (u,v) at each spatial location is projected onto the direction of the camera motion using ( P _ x ).
  • OC denotes the temporal weighting for each optical flow component projected to the camera direction.
  • FHI u (x, y, T) FHI u (x, y, ⁇ - 1) * (1 - a) + p;_ lU (x, y) * a
  • FHI V (x, y, T) FHI v (x, y, ⁇ - 1) * (1 - a) + P ⁇ vix, y) * a
  • Residual history image (RHI) obtained after the process in step 103 is suppressed using Flow history image (FHI) in step 106. Residuals are penalized by the suppression operation depending on the flow history magnitude. Higher optical flow magnitude results in further suppression of residual in exponential manner. In a preferred embodiment, residuals can be reduced exponentially.
  • Suppressed residual is stored in Supressed Residual History Image (SRI). Suppressed residuals are used in Motion History Image ( H f (X, ⁇ , ⁇ ) ) computation. Blobs within this Motion History Image are extracted using connected component analysis in step 107. Connected component analysis labels binary pixels which have neighborhood as a component and enumerates each component with a unique number.
  • Results of step 107 are stored in Connected Component Storage Unit (1070). Centroids of extracted blobs are given to deferred decision logic tracking module to declare moving targets in step 108. Deferred decision logic analyzes motion of each blob throughout time and assigns scores to blobs depending on the expected kinematic. Blobs with consistent motion have higher scores, and thus declared as a moving object.
  • moving object detection method 100

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to moving object detection methods for surveillance systems. The objective of the present invention is to realize a moving object detection method which reduces the false alarms resulting from parallax effect and contrast changes comprising the steps of storing acquired images in an image storage unit (1010); registering stored images onto a reference image and take difference; accumulating residuals using motion history imaging of differences; calculating optical flows of an acquired image with respect to the reference image; projecting optical flow vectors onto the direction of camera motion; suppressing accumulated residuals, by using the stored optical flow projections; extracting blobs within suppressed residual image (SRI); declaring moving targets by using a joint deferred decision logic tracker.

Description

DESCRIPTION
A MOVING OBJECT DETECTION METHOD
Field of the Invention
The present invention relates to moving object detection methods for surveillance systems.
Background of the Invention
Moving object detection application is critical for both airborne and land-based surveillance systems with a fixed and a rotating imaging sensor. The aim of such systems is to report moving objects to an operator for awareness against possible threats, or to track multiple moving targets automatically.
Moving object detection registers a sequence of images acquired by an imaging sensor and extracts possible moving regions by taking difference of this geometrically registered image with respect to selected reference. Image residuals (difference with respect to the geometrically registered reference) of an aligned sequence include both moving and non-moving objects. The major reasons of false alarms are parallax effect and contrast changes among consecutive images.
Parallax effect occurs on objects with varying depths with respect to the camera used for surveillance. Buildings and trees are the major source of false alarms due to parallax. Such static structures produce optical flow fields consistently since their depth deviation causes an optical misalignment when warped by the estimated global geometric transformation (perspective, or affine).
False alarms also occur due to misalignment of high contrast image structures. Gradient based suppression of residuals dampens such false alarms. False alarms manifest itself by misalignment of thin line structures among consecutive images such as road sides, roof edges.
The United States patent document numbered US2008278584 discloses a moving object detection apparatus and method by using optical flow analysis. The apparatus includes four modules of image capturing, image aligning, pixel matching, and moving object detection. Plural images are successively inputted under a camera. Based on neighbouring images, frame relationship on the neighbouring images is estimated. With the frame relationship, a set of warping parameter is further estimated. Based on the wrapping parameter, the background areas of the neighbouring images are aligned to obtain an aligned previous image. After the alignment, a corresponding motion vector for each pixel on the neighbouring images is traced. By analysing all the information generated from the optical flow, registration of the images with respect to the reference can be carried out for a moving object detection application. Optical flow technique used in US2008278584 is utilized for registration purposes.
The United States patent document numbered US2011116682 discloses an object detection method and an object detection system, suitable for detecting moving object information of a video stream having a plurality of images, are provided. The method performs a moving object foreground detection on each of the images, so as to obtain a first foreground detection image comprising a plurality of moving objects. The method also performs a texture object foreground detection on each of the images, so as to obtain a second foreground detection image comprising a plurality of texture objects. The moving objects in the first foreground detection image and the texture objects in the second foreground detection image are selected and filtered, and then the remaining moving objects or texture objects after the filtering are output as real moving object information. The United States patent document numbered US2009052740 discloses a moving object detecting device measures a congestion degree of a space and utilizes the congestion degree for tracking. In performing the tracking, a direction measured by a laser range sensor is heavily weighted when the congestion degree is low. When the congestion degree is high, a sensor fusion is performed by heavily weighting a direction measured by an image processing on a captured image to obtain a moving object estimating direction, and obtains a distance by the laser range sensor in the moving object estimating direction.
The United States patent document numbered US20150030202 discloses a method is provided for an intelligent video processing system based on object detection. The method includes receiving an input video sequence corresponding to a video program, obtaining a plurality of frames of the input video sequence, and obtaining a computational constraint and a temporal rate constraint. The method also includes determining one or more regions of interest (ROIs) of the plurality of frames based on the computational constraint and temporal rate constraint, and selecting a desired set of frames from the plurality of frames based on the ROIs such that the desired set of frames substantially represent a view path of the plurality of frames. Further, the method includes detecting object occurrences from the desired set of frames based on the selected desired set of frames such that a computational cost and a number of frames for detecting the object occurrences are under the computational constraint and temporal rate constraint.
Existing methods may not produce satisfactory results in terms of false alarms. Thus, further methodology is necessary to improve moving object detection performance in terms of false alarms. Optical flow based moving object detections was investigated earlier, where flow vectors were only utilized for registration purposes. However, this usage of optical flow does not necessarily reduce false alarms. Summary of the Invention
The objective of the present invention is to realize a moving object detection method which reduces the false alarms resulting from parallax effect and contrast changes.
Another objective of the present invention is to realize a moving object detection method which utilizes optical flow to suppress residuals, without any relation to registration, in accumulative manner using the camera motion.
Detailed Description of the Invention
A method realized to fulfill the objective of the present invention is illustrated in the accompanying figures, in which:
Figure 1 is the flowchart of the method.
Figure 2 is a single image taken from a video sequence
Figure 3 is Residual history image (RHI) accumulated through time of a video sequence
Figure 4 is optical flow vectors
Figure 5 is optical flow magnitudes
Figure 6 is suppressed residual image The steps and parts illustrated in the figures are individually numbered where the numbers refer to the following:
100. Moving object detection method
101. Storing acquired images in an image storage unit (1010),
102. Registering stored images onto a reference image and taking difference between the stored images and the reference image,
103. Accumulating residuals using motion history imaging of differences, 104. Calculating optical flows of an acquired image with respect to the reference image,
105. Projecting optical flow vectors onto the direction of camera motion,
106. Suppressing accumulated residuals calculated in step 103, by using the stored optical flow projections,
107. Extracting blobs within suppressed residual image,
108. Declaring moving targets by using a joint deferred decision logic tracker. 1010. Image storage unit
1020. Residual storage unit
1021. Transformations storage unit
1040. Flow history storage unit
1070. Connected component storage unit
RHI. Residual history image
FHI. Flow history image
SRI. Suppressed residual image
In the method (100), acquired images during surveillance are recorded to an image storage unit (1010), in step 101. In the preferred embodiment of the invention, the image storage unit (1010) is a circular buffer.
Images in the image storage unit (1010) are geometrically registered onto a current reference image in step 102. Geometric registration is the processing block which aligns an input image onto a given reference spatially. Geometric registration parameters of the images, which are stored in the image storage unit (1010), are stored in a geometric transformations storage unit (1021). Difference between registered images and the reference frame are also calculated in step 102 and sum of the residuals is stored in residual storage unit (1020). In the preferred embodiment of the invention, the residual storage unit (1020), and geometric transformation storage unit are circular buffers. Residuals are defined in detail in following paragraphs. Residuals are accumulated in step 103 and result is written to the Residual History Image (RHI). Here, pixels which do not include any motion are penalized as given in [Collins, R. 2006, Moving Object Localization in Thermal Imagery by Forward-backward MHI] ;
Figure imgf000007_0001
Here P_x denotes the geometric projection computed from time (τ -1) to (τ). "d' is the temporal decay to penalize pixels with no motion. D(x,y,x,k) is residual at location (x,y) at time instance (τ) with a time gap of (k) (i.e. residual is computed between image τ and τ+k). Residual is the difference of an image acquired at time τ and geometrically registered image that is acquired at time τ +k, where "k" may be positive or negative. "T" is the difference threshold to declare that there is putative motion among consecutive images. P+kI(x, y,T+k) is the geometric projection that aligns image acquired at time 1 onto image acquired at time T + k D(x, y, T,k) = I(x, y, T) -P+kI(x, y, T+k) . /is the image acquired, whose time instance is denoted by 1 , and a pixel coordinate is represented by x,y. H F (x, y, r ) is the pixel value of a Motion history image (MHI) at pixel (x,y) computed at time instance ( 1).
x, y, T, i) < T
Figure imgf000007_0002
Here MHI equation in [Collins, R. 2006, Moving Object Localization in Thermal Imagery by Forward-backward MHI] is slightly modified to incorporate multiple input images to be used for computing residual of several images captured at different times. W{ is the weight assigned to each residual for MHI computation. A linearly decreasing weighting function is used in the preferred embodiment of the invention. The reference image is selected from the image storage unit (1010). There is an inherent tradeoff over the selection of reference image index. Selecting the reference image index as the most current index produces less false alarms, however has high probability of misdetecting the slowly moving targets. Selecting the reference image acquired at distant time instances enables catching slowly moving targets at the expense of extreme false alarm due to parallax. In the preferred embodiment of the invention, reference image index has to be selected depending on application requirements. Using every image in MHI as proposed overcomes the presented challenges. Optical flow is calculated for the most recently acquired image with respect to a previous image with a specified time delay in step 104 and result is stored in Flow History Storage Unit (1040). An error constant is needed to calculate optical flows (u & v)
Where
£b = Exu + Eyv + Et (brightness constancy equation) is the rate of brightness change
and
£c 2 = (u— uf + (v— v)2 is the measure of departure from smoothness in the velocity field, u and v are the final averaged optical flow magnitudes. In one of the preferred embodiments of the invention u and v can be calculated with Horn- Schunck methods. However, computation of u and v is not limited to mentioned methods here.
^and ^ are the image gradients in horizontal and vertical dimensions, respectively. £¾ is the temporal image difference with respect to the previous frame.
Total error constant to be minimized to compute optical flows (u & v) is calculated as follows;
ε1 = jj [ 2ec 2 + ε )dxdy
1 Optical flows are projected to the direction of camera motion in step 105 and these projected flow vectors (2D vector arrays) are stored in Flow history storage unit (1040) after being warped by the inverse of the estimated camera motion. This way, a pixel in Flow history image (FHI) obtained after the process in step 105, accumulates flows corresponding to the same 3D coordinate throughout time. Flow history image (FHI) holds the optical flow vector (u,v) for each pixel of the image. Optical flow vector of each pixel is alpha-filtered temporally and stored on Flow history storage unit (1040) for temporal consistency. Note that optical flow vector tuple (u,v) at each spatial location is projected onto the direction of the camera motion using ( P _x ). OC denotes the temporal weighting for each optical flow component projected to the camera direction.
FHI u (x, y, T) = FHIu (x, y, τ - 1) * (1 - a) + p;_lU(x, y) * a
FHIV (x, y, T) = FHIv(x, y, τ - 1) * (1 - a) + P^vix, y) * a
Residual history image (RHI) obtained after the process in step 103, is suppressed using Flow history image (FHI) in step 106. Residuals are penalized by the suppression operation depending on the flow history magnitude. Higher optical flow magnitude results in further suppression of residual in exponential manner. In a preferred embodiment, residuals can be reduced exponentially. Suppressed residual is stored in Supressed Residual History Image (SRI). Suppressed residuals are used in Motion History Image ( H f (X, γ, τ) ) computation. Blobs within this Motion History Image are extracted using connected component analysis in step 107. Connected component analysis labels binary pixels which have neighborhood as a component and enumerates each component with a unique number. Results of step 107 are stored in Connected Component Storage Unit (1070). Centroids of extracted blobs are given to deferred decision logic tracking module to declare moving targets in step 108. Deferred decision logic analyzes motion of each blob throughout time and assigns scores to blobs depending on the expected kinematic. Blobs with consistent motion have higher scores, and thus declared as a moving object. Within the scope of these basic concepts, it is possible to develop a wide variety of embodiments of the inventive "moving object detection method (100)". The invention cannot be limited to the examples described herein; it is essentially according to the claims.

Claims

1. A moving object detection method (100) comprises the steps of
- storing acquired images in an image storage unit (1010) (101),
- registering stored images onto a reference image and taking difference between them (102),
- extracting blobs within suppressed residual image (SRI) (107),
- declaring moving targets by using a joint deferred decision logic tracker (108) and characterized by the steps of
- accumulating weighted residuals using motion history imaging of differences (103),
- computing optical flows of an acquired image with respect to the reference image (104),
- projecting optical flow vectors onto the direction of camera motion (105),
- suppressing accumulated residuals calculated in step 103, by using the stored optical flow projections (106).
2. A moving object detection method (100) according to claim 1 wherein a circular buffer is used as image storage unit (1010).
3. A moving object detection method (100) according to claim 1 wherein a circular buffer is used as residual storage unit (1020).
4. A moving object detection method (100) according to claim 1 wherein a circular buffer is used as Flow history storage unit (1040) (1010).
5. A moving object detection method (100) according to claim 1 wherein a circular buffer is used as geometric transformations storage unit (1021).
6. A moving object detection method (100) according to claim 1 wherein multiple input images are incorporated to be used for computing residual of several images captured at different times with a assigned weight to each residual by using the formula;
Figure imgf000012_0001
7. A moving object detection method (100) according to claim 1 wherein an error constant is calculated by using brightness change and the measure of departure from smoothness with the formula ε2 = J J {p 2ec 2 + ε )dxdy where £b = Exu + Eyv + Et represents brightness change and €^ = (u—u )2 + (v— )2 represents the measure of departure from smoothness, in the step 104.
8. A moving object detection method (100) according to claim 1 wherein optical flow vectors projected onto camera direction are temporally weighted
, FHIu (x, y, T) = FHIu (x, y, T - l) * - ) + P;_lU(x, y) * by using the formula
FHIV (x, y, T) = FHIv(x, y, τ - 1) * (1 - a) + Ρτ τ_ ν{χ, y) * a where Οί is the temporal weighting, in step 105.
9. A moving object detection method (100) according to claim 1 wherein residuals are reduced exponentially, in step 106.
PCT/IB2015/058188 2015-10-23 2015-10-23 A moving object detection method WO2017068397A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/IB2015/058188 WO2017068397A1 (en) 2015-10-23 2015-10-23 A moving object detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/IB2015/058188 WO2017068397A1 (en) 2015-10-23 2015-10-23 A moving object detection method

Publications (1)

Publication Number Publication Date
WO2017068397A1 true WO2017068397A1 (en) 2017-04-27

Family

ID=54697617

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2015/058188 WO2017068397A1 (en) 2015-10-23 2015-10-23 A moving object detection method

Country Status (1)

Country Link
WO (1) WO2017068397A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110533692A (en) * 2019-08-21 2019-12-03 深圳新视达视讯工程有限公司 A kind of automatic tracking method towards target mobile in unmanned plane video
CN110889833A (en) * 2019-11-18 2020-03-17 山东大学 Deep sea plankton detection method and system based on gradient optical flow method

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003077552A1 (en) * 2002-02-13 2003-09-18 Reify Corporation Method and apparatus for acquisition, compression, and characterization of spatiotemporal signals
US20060285724A1 (en) * 2005-06-20 2006-12-21 Ying-Li Tian Salient motion detection system, method and program product therefor
US20080278584A1 (en) 2007-05-11 2008-11-13 Ming-Yu Shih Moving Object Detection Apparatus And Method By Using Optical Flow Analysis
US20090052740A1 (en) 2007-08-24 2009-02-26 Kabushiki Kaisha Toshiba Moving object detecting device and mobile robot
US20100315505A1 (en) * 2009-05-29 2010-12-16 Honda Research Institute Europe Gmbh Object motion detection system based on combining 3d warping techniques and a proper object motion detection
US20110116682A1 (en) 2009-11-19 2011-05-19 Industrial Technology Research Institute Object detection method and system
US20150030202A1 (en) 2013-07-23 2015-01-29 TCL Research America Inc. System and method for moving object detection and processing

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003077552A1 (en) * 2002-02-13 2003-09-18 Reify Corporation Method and apparatus for acquisition, compression, and characterization of spatiotemporal signals
US20060285724A1 (en) * 2005-06-20 2006-12-21 Ying-Li Tian Salient motion detection system, method and program product therefor
US20080278584A1 (en) 2007-05-11 2008-11-13 Ming-Yu Shih Moving Object Detection Apparatus And Method By Using Optical Flow Analysis
US20090052740A1 (en) 2007-08-24 2009-02-26 Kabushiki Kaisha Toshiba Moving object detecting device and mobile robot
US20100315505A1 (en) * 2009-05-29 2010-12-16 Honda Research Institute Europe Gmbh Object motion detection system based on combining 3d warping techniques and a proper object motion detection
US20110116682A1 (en) 2009-11-19 2011-05-19 Industrial Technology Research Institute Object detection method and system
US20150030202A1 (en) 2013-07-23 2015-01-29 TCL Research America Inc. System and method for moving object detection and processing

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
COLLINS, R., MOVING OBJECT LOCALIZATION IN THERMAL IMAGERY BY FORWARD-BACKWARD MHI, 2006
TSAI DU-MING ET AL: "Optical flow-motion history image (OF-MHI) for action recognition", SIGNAL, IMAGE AND VIDEO PROCESSING, SPRINGER LONDON, LONDON, vol. 9, no. 8, 24 July 2014 (2014-07-24), pages 1897 - 1906, XP035537090, ISSN: 1863-1703, [retrieved on 20140724], DOI: 10.1007/S11760-014-0677-9 *
ZHAOZHENG YIN ET AL: "Moving Object Localization in Thermal Imagery by Forward-backward MHI", COMPUTER VISION AND PATTERN RECOGNITION WORKSHOP, 2006 CONFERENCE ON NEW YORK, NY, USA 17-22 JUNE 2006, PISCATAWAY, NJ, USA,IEEE, PISCATAWAY, NJ, USA, 17 June 2006 (2006-06-17), pages 133 - 133, XP010922949, ISBN: 978-0-7695-2646-1, DOI: 10.1109/CVPRW.2006.131 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110533692A (en) * 2019-08-21 2019-12-03 深圳新视达视讯工程有限公司 A kind of automatic tracking method towards target mobile in unmanned plane video
CN110533692B (en) * 2019-08-21 2022-11-11 深圳新视达视讯工程有限公司 Automatic tracking method for moving target in aerial video of unmanned aerial vehicle
CN110889833A (en) * 2019-11-18 2020-03-17 山东大学 Deep sea plankton detection method and system based on gradient optical flow method
CN110889833B (en) * 2019-11-18 2022-04-19 山东大学 Deep sea plankton detection method and system based on gradient optical flow method

Similar Documents

Publication Publication Date Title
EP2632160B1 (en) Method and apparatus for image processing
Ayazoglu et al. Dynamic subspace-based coordinated multicamera tracking
CN105427345B (en) Three-dimensional stream of people&#39;s method of motion analysis based on camera projection matrix
JP2009514066A (en) &#34;Camera frame matching model generation graphics image for each frame and pixel computer video&#34;
CA2670021A1 (en) System and method for estimating characteristics of persons or things
Okuma et al. Automatic rectification of long image sequences
Shukla et al. Speed determination of moving vehicles using Lucas-Kanade algorithm
US11410459B2 (en) Face detection and recognition method using light field camera system
WO2017068397A1 (en) A moving object detection method
KR100994722B1 (en) Method for tracking moving object on multiple cameras using probabilistic camera hand-off
Patel et al. Comparison of optical flow algorithms for speed determination of moving objects
Sincan et al. Moving object detection by a mounted moving camera
CN104616320A (en) Method for detecting vehicle in low-altitude aerial video based on gradient inhibition and epipolar constraint
Yu et al. Accurate motion detection in dynamic scenes based on ego-motion estimation and optical flow segmentation combined method
JP6504711B2 (en) Image processing device
Wong et al. Shared-use motion vector algorithm for moving objects detection for automobiles
Son et al. Tiny drone tracking framework using multiple trackers and Kalman-based predictor
Yousaf et al. Real time video stabilization methods in IR domain for UAVs—A review
Joshi Vehicle speed determination using image processing
Lamża et al. Depth estimation in image sequences in single-camera video surveillance systems
Fonseca et al. Design and implementation of an optical flow-based autonomous video surveillance system
JP2017068375A (en) Device, method and program for tracking person using plural cameras
Veeraraghavan et al. Motion based correspondence for 3D tracking of multiple dim objects
Castelli et al. Moving object detection for unconstrained low-altitude aerial videos, a pose-independant detector based on artificial flow
Havasi et al. A statistical method for object localization in multi-camera tracking

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15798563

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15798563

Country of ref document: EP

Kind code of ref document: A1