WO2014092550A2 - Method for camera motion estimation with presence of moving object - Google Patents
Method for camera motion estimation with presence of moving object Download PDFInfo
- Publication number
- WO2014092550A2 WO2014092550A2 PCT/MY2013/000264 MY2013000264W WO2014092550A2 WO 2014092550 A2 WO2014092550 A2 WO 2014092550A2 MY 2013000264 W MY2013000264 W MY 2013000264W WO 2014092550 A2 WO2014092550 A2 WO 2014092550A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- feature points
- image
- frame
- correspond
- matched
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/285—Analysis of motion using a sequence of stereo image pairs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/75—Determining position or orientation of objects or cameras using feature-based methods involving models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
- G06T2207/10021—Stereoscopic video; Stereoscopic image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30232—Surveillance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30244—Camera pose
Definitions
- the present invention relates to image processing. More particularly, the present invention relates to a method for motion estimation on a camera with presence of moving object.
- RANSAC technique can be used to remove matched feature points that do not correspond to the actual motion.
- a basic assumption used in the technique to remove matched feature points is that the data includes "inliers" whereby the data whose distribution can be explained by sets of model parameters, and "outliers” whereby the data that do not fit the model.
- the data can be subjected to noise. Some of the outliers may come from extreme values of the noise or from erroneous measurements or incorrect hypotheses about the interpretation of data.
- RANSAC technique also assumes that outlier are usually found in a small amount, i.e. most feature points shall belongs to background, which can be used to estimate the parameters of a model that optimally explains or fits this data.
- the method comprises extracting feature points from a current frame of the sequence of the image frames; estimating disparity of the feature points; identifying an object from the current frame that occupies a large area of the current frame through the disparity of the feature points; matching feature points between the current image frame and a previous image frame to determine presence of a moving object; filtering out the matched feature points that correspond to the moving object to obtain matched feature points that correspond to background; and estimating motion model based on the feature points that correspond to the background.
- the imaging device is a stereo camera adapted for capturing the sequence of image frames, each image frame has a left image and a right image that form a stereo image.
- extracting feature points and estimating disparity of the feature points from the current frame comprises extracting feature points of the left image and the right image; and matching feature points between the left image and the right image.
- identifying the object that occupies a large area of the current frame comprises computing a percentage of feature points with disparity value less than a predefined threshold; estimating a two-dimensional (2D) Gaussion parameters to represent the spatial distribution of matched feature points when the percentage is more than the predefined threshold; concluding that a huge object is present when size of the 2D Gaussian parameters fitted the distribution is more than a size threshold of the image frame size. Otherwise, it can be concluded that there is no huge object present in the current frame.
- determining presence of moving object comprises splitting the matched feature points into groups of conform motion model vectors by recursively filter the matched feature points using RANSAC technique; and concluding that potential moving object is present in the current image, when at least two major groups of motion vectors exist, otherwise, concluding that no potential moving object present.
- Data associated to the matched feature points are divided into inliers and outliers through a RANSAC technique, the RANSAC technique is carried out recursively to further divide outliers until no valid outliers are to be divided.
- determining matched feature points correspond to background area comprises clustering pixels of the current image frame into their corresponding group of conformed matched feature point; computing the percentage of image area correspond to each group cluster; and group of conformed matched feature point with the largest area in the image is deemed as the background points.
- FIG. 1A illustrates a process for camera motion estimation in accordance with one embodiment of the present invention
- FIGs. IB and 1C exemplify the image processing through the method provided the present invention
- FIG. 2 illustrates the process for determining presence of huge object in accordance with one embodiment of the present invention
- FIG. 3 illustrates a process of determining a presence of moving object in frame sequence in accordance with one embodiment of the present invention.
- FIG. 4 illustrates the process for determining matched feature points correspond to background area in accordance with one embodiment of the present invention.
- the present invention provides a method for eliminating unwanted matched feature points (i.e. noise) before geometric or affine transformation or homography matrix is computed in camera motion model estimation.
- unwanted matched feature points are typically caused by the presence of moving object in the image.
- the method includes checking for present of moving/huge object(s) in image frames through distribution of disparity information and recursive RANSAC or Maximum Likelihood Estimation Sample Consensus (MLESAC); selecting sets of matching points that correspond to background with high confidence instead of moving object by integrating area-based and feature-based in finding correspondences between subsequent image frames to define potential feature points on moving objects or background; and eliminating huge object in the image using disparity information.
- the present invention provides a method for obtaining correct camera motion model estimation by filtering unwanted matched feature points used in estimation due to huge objects and moving objects in the image frames.
- the motion estimation includes computing geometric or affine transformation or homography matrix between subsequent frames to define how each frame is related to another.
- FIG. IB exemplifies the step of computing affine transformation between two frames to identify the relationship between the two consecutive frames.
- the camera motion model estimation includes extracting feature points, matching feature points, filtering matched feature points through RANSAC/MLESAC to remove the matched feature points that do not correspond to the actual motion object, i.e. outliers, and using matched feature points to compute transformation or homography matrix that defines the model of the motion of one frame with respect to the other.
- FIG. 1A illustrates a process for camera motion estimation in accordance with one embodiment of the present invention.
- the process can be implemented on a system that acquires video or image sequences from an imaging device.
- the imaging device may be a stereo camera adapted for capturing stereo images.
- the stereo is able to capture video that contains sequence of image frames, each image frame contains a stereo image that made up of at least a left and a right image.
- the process acquires a sequence of frames of stereo images and processes the same in accordance with the embodiment of the present invention, and outputs a homography matrix or an affine transformation matrix that defines how the imaging device is moved from previous to current state.
- this matrix will be referred as transformation matrix for simplicity.
- the process comprises extracting feature points in stereo image of a current frame at step 100; estimating feature points' disparity through matching the same between the left and right images of the stereo image at step 101; determining presence of huge object in the current frame at step 102; matching feature points between the current frame and a previous frame at step 103; determining presence of moving object in the current frame at step 104; determining matched feature points that correspond to background of the current frame at step 105; filtering matched feature points at step 106; and estimating motion model of the imaging device at step 107.
- the feature point extraction is carried on both left and right images of the current frame.
- the process estimate the feature point disparity on each feature point by matching between the left and right images of the current frame in the step 101.
- the feature point disparity estimation also determines a distance between the matched feature points and the imaging device in 3D space.
- the process determines the presence of huge object in current frame, if any.
- the process further matches feature points between the current frame and the previous frame in the step 103. In the step, all feature points from the right or left image can be used.
- the matched feature points between the current frame and the previous frame is then used as the basis for determining presence of moving object in sequence of frames in the step 104. Subsequently, the matched feature points that correspond to the background can be determined by taking the remaining matched feature points that are not regarded as the moving object as the background feature points in the step 105. The matched feature points that belong to the moving object and huge object are further filtered out in the step 106 for estimating the imaging device motion in the step 107.
- FIG. 2 illustrates the process for determining presence of huge object in accordance with one embodiment of the present invention.
- the process can be adapted in the step 102 of FIG 1.
- the feature points of the left and right images and their respective disparity information are already extracted for processing herein.
- the huge object presents in the image sequence, most of image area is covered by the huge object, instead of background.
- an object located too close to the capturing area of the imaging device will also appear as huge object in the frame sequence.
- the process may choose to ignore this frame, thus no transformation matrix will be computed.
- the process computes the percentage of matched feature points with disparity value less than a preset threshold.
- a preset threshold is set at 50%. This threshold can be determined and preset before the process is carried out. Depending on a nearest acceptable distance of object from the imaging device, such threshold value can be changed accordingly when desired.
- the process continues to estimate two- dimensional (2D) Gaussian parameters that represent spatial distributions of the matched feature points with disparity less than threshold in step 202.
- the 2D Gaussian parameters include mean x and y; (x0,y0), which indicate the center of the distribution; and the ⁇ and ay which indicate how the distribution is spreading in x and y directions respectively. These parameters shall then be used to define how much has the huge object occupies the image or relative size of the object in the image.
- a size threshold can be preset as 50% for example. When the size of the object is determined to be more than the size threshold, the process concludes that the object is a huge object in step 204, which otherwise (i.e. less than 50%), it shall conclude that the object present in the image of the frame is not a huge object in the step 205.
- FIG. 3 illustrates a process of determining a presence of moving object in frame sequence in accordance with one embodiment of the present invention.
- This process can be adapted in the step 104 of FIG. 1.
- This process is provided to determine whether there are any moving objects present in a current frame.
- the process monitors the number of groups of similar motion vectors in the frame. Group of similar motion vectors may belong to a same object or a same background. Thus, when more than one group exist, it can be concluded that there is a likelihood of moving object(s) present in the frame.
- the matched feature points that correspond to these moving objects will be disregarded to obtain the features point that correspond to background.
- the process comprises splitting the matched feature points into groups of conformed motion vectors by recursively filtering the matched feature points in step 300.
- This can be achieved by using a RANSAC- based technique
- RANSAC technique can be used to remove the matched feature points that do not correspond to an actual motion. These matched points are herein referred as the outlier.
- the actual motion contains a group of motion vectors that correspond to a majority. Accordingly, perform recursive the RANSAC-based technique to delineate matched points to their conform motion.
- the RANSAC-based technique includes dividing the matched feature points into two groups, i.e. inliers and outliers.
- the RANSAC-based technique will be carried out to process the outliers, which are derived through the previous ANSAC process.
- the outliers will then be further divided into inlier and outliers through the subsequent RANSAC process.
- the recursive steps will be repeated until no valid outliers are to be divided.
- Outliers are considered valid to be divided if the numbers is at least more than 20% of total number of matched points.
- the process determines the number of major group of motion vectors at step 301.
- a group is considered as a major group when the number of matched feature points in the group is more than 20% of total matched feature points or at least 8 matched points, whichever is more. If at least 2 major groups exist in step 302, then it is concluded that there is potential moving objects in the current frame. If not, it is concluded otherwise in step 304.
- FIG. 4 illustrates the process for determining matched feature points correspond to background area in accordance with one embodiment of the present invention.
- This process can be adapted in the step 105 of FIG. 1.
- This process is provided for delineating matched points that correspond to background from moving objects, so that only these points are considered in transformation matrix computation to determine the imaging device's movement. This process may be required only if it is concluded that there is a potential moving object in the image.
- image pixels of the image are clustered into corresponding group of conformed matched feature points.
- any known clustering techniques can be adapted herein.
- the major group's properties e.g. appearance, centroid, average motion direction etc.
- step 401 the percentage of image area correspond to each group segment is computed in step 401. This can be computed by dividing total number of pixels correspond to the group divided by total number of pixels in the image.
- the group with largest area will be regarded as background points in step 402. Accordingly, all other matched points not correspond to background can be filtered as the step 106 of FIG. 1, leaving only matched points correspond to background to be used in computing the camera motion model as the step 107.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The present invention provides a method of processing sequence of image frames captured through an imaging device. The method comprises extracting feature points from a current frame of the sequence of the image frames; estimating disparity of the feature points; identifying an object from the current frame that occupies a large area of the current frame through the disparity of the feature points; matching feature points between the current image frame and a previous image frame to determine presence of a moving object; filtering out the matched feature points that correspond to the moving object to obtain matched feature points that correspond to background; estimating motion model based on the feature points that correspond to the background.
Description
Method For Camera Motion Estimation With Presence of Moving
Object
Field of the Invention
[0001] The present invention relates to image processing. More particularly, the present invention relates to a method for motion estimation on a camera with presence of moving object.
Background
[0002] In digital image processing, it is well known in the art that camera motion model estimation can be achieved through computing geometric/affme transformation between frames to define how each frames are related to each other in the sequence. Accordingly, it is possible to extract feature points from sequence of frames taken, then match the extracted feature points, then filter the matched feature points that do not correspond the actual motion (outliers) through Random Sample Consensus (RANSAC) technique, and finally use filtered matched feature points to compute transformation or homography matrix that defines the model of the motion of one frame with respect to the other.
[0003] In the aforesaid mentioned method, RANSAC technique can be used to remove matched feature points that do not correspond to the actual motion. A basic assumption used in the technique to remove matched feature points is that the data includes "inliers" whereby the data whose distribution can be explained by sets of model parameters, and "outliers" whereby the data that do not fit the model.
[0004] In addition to this, the data can be subjected to noise. Some of the outliers may come from extreme values of the noise or from erroneous measurements or incorrect hypotheses about the interpretation of data. Further, RANSAC technique also assumes that outlier are usually found in a small amount, i.e. most feature points shall belongs to background, which can be used to estimate the parameters of a model that optimally explains or fits this data.
[0005] In the case when majority of the matched feature points are determined belongs to moving objects, the camera motion may be incorrect. It follows that any image warping will fail, which will in turn fail subsequent process of extracting moving object from the image.
[0006] The problem arises when the majority of matched points are contributed by correct matched points yet incorrect to be used to define the scene motion because they belong to moving objects in the scene, instead of background. This scenario may happen when less texture in background area is lacking of texture and/or a huge object occupies the frames.
[0007] Further, there is also a question on how to check whether there is possible moving object in the image so that the elimination process has to be considered.
Summary [0008] In one aspect of the present invention, there is provided method of processing sequence of image frames captured through an imaging device. The method comprises extracting feature points from a current frame of the sequence of
the image frames; estimating disparity of the feature points; identifying an object from the current frame that occupies a large area of the current frame through the disparity of the feature points; matching feature points between the current image frame and a previous image frame to determine presence of a moving object; filtering out the matched feature points that correspond to the moving object to obtain matched feature points that correspond to background; and estimating motion model based on the feature points that correspond to the background.
[0009] In one embodiment, the imaging device is a stereo camera adapted for capturing the sequence of image frames, each image frame has a left image and a right image that form a stereo image. In this case, extracting feature points and estimating disparity of the feature points from the current frame comprises extracting feature points of the left image and the right image; and matching feature points between the left image and the right image.
[0010] In another embodiment, identifying the object that occupies a large area of the current frame comprises computing a percentage of feature points with disparity value less than a predefined threshold; estimating a two-dimensional (2D) Gaussion parameters to represent the spatial distribution of matched feature points when the percentage is more than the predefined threshold; concluding that a huge object is present when size of the 2D Gaussian parameters fitted the distribution is more than a size threshold of the image frame size. Otherwise, it can be concluded that there is no huge object present in the current frame.
[0011] In yet another embodiment, determining presence of moving object comprises splitting the matched feature points into groups of conform motion model
vectors by recursively filter the matched feature points using RANSAC technique; and concluding that potential moving object is present in the current image, when at least two major groups of motion vectors exist, otherwise, concluding that no potential moving object present. Data associated to the matched feature points are divided into inliers and outliers through a RANSAC technique, the RANSAC technique is carried out recursively to further divide outliers until no valid outliers are to be divided.
[0012] In yet a further embodiment, determining matched feature points correspond to background area comprises clustering pixels of the current image frame into their corresponding group of conformed matched feature point; computing the percentage of image area correspond to each group cluster; and group of conformed matched feature point with the largest area in the image is deemed as the background points.
Brief Description of the Drawings [0013] Preferred embodiments according to the present invention will now be described with reference to the figures accompanied herein, in which like reference numerals denote like elements;
[0014] FIG. 1A illustrates a process for camera motion estimation in accordance with one embodiment of the present invention; [0015] FIGs. IB and 1C exemplify the image processing through the method provided the present invention;
[0016] FIG. 2 illustrates the process for determining presence of huge object in accordance with one embodiment of the present invention;
[0017] FIG. 3 illustrates a process of determining a presence of moving object in frame sequence in accordance with one embodiment of the present invention; and
[0018] FIG. 4 illustrates the process for determining matched feature points correspond to background area in accordance with one embodiment of the present invention.
Detailed Description [0019] Embodiments of the present invention shall now be described in detail, with reference to the attached drawings. It is to be understood that no limitation of the scope of the invention is thereby intended, such alterations and further modifications in the illustrated device, and such further applications of the principles of the invention as illustrated therein being contemplated as would normally occur to one skilled in the art to which the invention relates.
[0020] The present invention provides a method for eliminating unwanted matched feature points (i.e. noise) before geometric or affine transformation or homography matrix is computed in camera motion model estimation. Such unwanted matched feature points are typically caused by the presence of moving object in the image. The method includes checking for present of moving/huge object(s) in image frames through distribution of disparity information and recursive RANSAC or Maximum Likelihood Estimation Sample Consensus (MLESAC);
selecting sets of matching points that correspond to background with high confidence instead of moving object by integrating area-based and feature-based in finding correspondences between subsequent image frames to define potential feature points on moving objects or background; and eliminating huge object in the image using disparity information.
[0021] In another aspect, the present invention provides a method for obtaining correct camera motion model estimation by filtering unwanted matched feature points used in estimation due to huge objects and moving objects in the image frames. The motion estimation includes computing geometric or affine transformation or homography matrix between subsequent frames to define how each frame is related to another. FIG. IB exemplifies the step of computing affine transformation between two frames to identify the relationship between the two consecutive frames. The camera motion model estimation includes extracting feature points, matching feature points, filtering matched feature points through RANSAC/MLESAC to remove the matched feature points that do not correspond to the actual motion object, i.e. outliers, and using matched feature points to compute transformation or homography matrix that defines the model of the motion of one frame with respect to the other.
[0022] FIG. 1A illustrates a process for camera motion estimation in accordance with one embodiment of the present invention. The process can be implemented on a system that acquires video or image sequences from an imaging device. The imaging device may be a stereo camera adapted for capturing stereo images. In particular, the stereo is able to capture video that contains sequence of
image frames, each image frame contains a stereo image that made up of at least a left and a right image. The process acquires a sequence of frames of stereo images and processes the same in accordance with the embodiment of the present invention, and outputs a homography matrix or an affine transformation matrix that defines how the imaging device is moved from previous to current state. In this document, this matrix will be referred as transformation matrix for simplicity. To obtain the transformation matrix, it is important to get matching feature points that belong to background on the frames. Unwanted matched feature points pairs (of the stereo images) that belong to objects, such as huge objects, based on disparity information of current frames shall be remove.
[0023] Accordingly, the process comprises extracting feature points in stereo image of a current frame at step 100; estimating feature points' disparity through matching the same between the left and right images of the stereo image at step 101; determining presence of huge object in the current frame at step 102; matching feature points between the current frame and a previous frame at step 103; determining presence of moving object in the current frame at step 104; determining matched feature points that correspond to background of the current frame at step 105; filtering matched feature points at step 106; and estimating motion model of the imaging device at step 107. [0024] Returning to the step 100, the feature point extraction is carried on both left and right images of the current frame. It is understood to a skilled person that any feature point extraction techniques or methods can be adapted herein and therefore, no further information is provided herewith. Once the feature points are
extracted, the process estimate the feature point disparity on each feature point by matching between the left and right images of the current frame in the step 101. The feature point disparity estimation also determines a distance between the matched feature points and the imaging device in 3D space. Based on the feature point disparity information, in the step 102, the process determines the presence of huge object in current frame, if any. Following that, the process further matches feature points between the current frame and the previous frame in the step 103. In the step, all feature points from the right or left image can be used. The matched feature points between the current frame and the previous frame is then used as the basis for determining presence of moving object in sequence of frames in the step 104. Subsequently, the matched feature points that correspond to the background can be determined by taking the remaining matched feature points that are not regarded as the moving object as the background feature points in the step 105. The matched feature points that belong to the moving object and huge object are further filtered out in the step 106 for estimating the imaging device motion in the step 107.
[0025] FIG. 2 illustrates the process for determining presence of huge object in accordance with one embodiment of the present invention. The process can be adapted in the step 102 of FIG 1. As provided in the process of FIG. 1, the feature points of the left and right images and their respective disparity information are already extracted for processing herein. Basically when huge object presents in the image sequence, most of image area is covered by the huge object, instead of background. Besides the object's physical size, an object located too close to the capturing area of the imaging device will also appear as huge object in the frame sequence. When the object occupies more than half of the image of the frame, the
background area is hidden behind thus very little features can be extracted from the background area. In that case, the process may choose to ignore this frame, thus no transformation matrix will be computed. The detected object(s) occupies less half of the image of the frame, a transformation matrix will be computed once the matched feature points that correspond to the detected object will be removed. Based on the above, in the step 200, the process computes the percentage of matched feature points with disparity value less than a preset threshold. As mentioned, this step is used to determine whether the object is too near to the imaging device or physically too big by itself. In FIG. 2, the preset threshold is set at 50%. This threshold can be determined and preset before the process is carried out. Depending on a nearest acceptable distance of object from the imaging device, such threshold value can be changed accordingly when desired. When the percentage is less than 50% (or not larger than 50%) at the process concludes that there is no huge object in 205. When the percentage is larger than 50%, the process continues to estimate two- dimensional (2D) Gaussian parameters that represent spatial distributions of the matched feature points with disparity less than threshold in step 202. The 2D Gaussian parameters include mean x and y; (x0,y0), which indicate the center of the distribution; and the σχ and ay which indicate how the distribution is spreading in x and y directions respectively. These parameters shall then be used to define how much has the huge object occupies the image or relative size of the object in the image. Similarly, a size threshold can be preset as 50% for example. When the size of the object is determined to be more than the size threshold, the process concludes that the object is a huge object in step 204, which otherwise (i.e. less than 50%), it
shall conclude that the object present in the image of the frame is not a huge object in the step 205.
[0026] FIG. 3 illustrates a process of determining a presence of moving object in frame sequence in accordance with one embodiment of the present invention. This process can be adapted in the step 104 of FIG. 1. This process is provided to determine whether there are any moving objects present in a current frame. The process monitors the number of groups of similar motion vectors in the frame. Group of similar motion vectors may belong to a same object or a same background. Thus, when more than one group exist, it can be concluded that there is a likelihood of moving object(s) present in the frame. In the further step (i.e. the step 105 of FIG. 1), the matched feature points that correspond to these moving objects will be disregarded to obtain the features point that correspond to background.
[0027] As shown in FIG. 3, the process comprises splitting the matched feature points into groups of conformed motion vectors by recursively filtering the matched feature points in step 300. This can be achieved by using a RANSAC- based technique It is known that RANSAC technique can be used to remove the matched feature points that do not correspond to an actual motion. These matched points are herein referred as the outlier. The actual motion contains a group of motion vectors that correspond to a majority. Accordingly, perform recursive the RANSAC-based technique to delineate matched points to their conform motion. The RANSAC-based technique includes dividing the matched feature points into two groups, i.e. inliers and outliers. The RANSAC-based technique will be carried
out to process the outliers, which are derived through the previous ANSAC process. The outliers will then be further divided into inlier and outliers through the subsequent RANSAC process. The recursive steps will be repeated until no valid outliers are to be divided. Outliers are considered valid to be divided if the numbers is at least more than 20% of total number of matched points. Then, the process determines the number of major group of motion vectors at step 301. A group is considered as a major group when the number of matched feature points in the group is more than 20% of total matched feature points or at least 8 matched points, whichever is more. If at least 2 major groups exist in step 302, then it is concluded that there is potential moving objects in the current frame. If not, it is concluded otherwise in step 304.
[0028] FIG. 4 illustrates the process for determining matched feature points correspond to background area in accordance with one embodiment of the present invention. This process can be adapted in the step 105 of FIG. 1. This process is provided for delineating matched points that correspond to background from moving objects, so that only these points are considered in transformation matrix computation to determine the imaging device's movement. This process may be required only if it is concluded that there is a potential moving object in the image. In the step 400, image pixels of the image are clustered into corresponding group of conformed matched feature points. In this case, any known clustering techniques can be adapted herein. The major group's properties (e.g. appearance, centroid, average motion direction etc.) can be used for indicating whether other pixels belong to their group. Then the percentage of image area correspond to each group segment is computed in step 401. This can be computed by dividing total number of pixels
correspond to the group divided by total number of pixels in the image. The group with largest area will be regarded as background points in step 402. Accordingly, all other matched points not correspond to background can be filtered as the step 106 of FIG. 1, leaving only matched points correspond to background to be used in computing the camera motion model as the step 107.
[0029] While specific embodiments have been described and illustrated, it is understood that many changes, modifications, variations, and combinations thereof could be made to the present invention without departing from the scope of the invention.
Claims
1. A method of processing sequence of image frames captured through an imaging device, the method comprising:
extracting feature points from a current frame of the sequence of the image frames;
estimating disparity of the feature points;
identifying an object from the current frame that occupies a large area of the current frame through the disparity of the feature points;
matching feature points between the current image frame and a previous image frame to determine presence of a moving object;
filtering out the matched feature points that correspond to the moving object to obtain matched feature points that correspond to background;
estimating motion model based on the feature points that correspond to the background.
2. The method according to claim 1, wherein the imaging device is a stereo camera adapted for capturing the sequence of image frames, each image frame has a left image and a right image that form a stereo image.
3. The method according to claim 2, wherein extracting feature points and estimating disparity of the feature points from the current frame comprising:
extracting feature points of the left image and the right image; and matching feature points between the left image and the right image.
4. The method according to claim 1, wherein identifying the object that occupies a large area of the current comprises:
computing a percentage of feature points with disparity value less than a predefined threshold;
estimating a two-dimensional (2D) Gaussion parameters to represent the spatial distribution of matched feature points when the percentage is more than the predefined threshold;
concluding that a huge object is present when size of the 2D Gaussian parameters fitted the distribution is more than a size threshold of the image frame size.
otherwise, concluding that there is no huge object present in the current frame.
5. The method according to claim 4, further comprising processing the next frame when a huge object is present in the current frame.
6. The method according to claim 1, wherein determining presence of moving object comprises
splitting the matched feature points into groups of conform motion model vectors by recursively filter the matched feature points using RANSAC-based technique; concluding that potential moving object is present in the current image, when at least two major groups of motion vectors exist, otherwise, concluding that no potential moving object present.
7. The method according to claim 6, wherein data associated to the matched feature points are divided into inliers and outliers through a RANSAC-based technique, the RANSAC technique is carried out recursively to further divide outliers until no valid outliers are to be divided.
8. The method according to claim 1, wherein determining matched feature points correspond to background area comprising clustering pixels of the current image frame into their corresponding group of conformed matched feature point;
computing the percentage of image area correspond to each group cluster; and group of conformed matched feature point with the largest area in the image is deemed as the background points.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
MYPI2012005333 | 2012-12-10 | ||
MYPI2012005333A MY188908A (en) | 2012-12-10 | 2012-12-10 | Method for camera motion estimation with presence of moving object |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2014092550A2 true WO2014092550A2 (en) | 2014-06-19 |
WO2014092550A3 WO2014092550A3 (en) | 2014-09-12 |
Family
ID=50179899
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/MY2013/000264 WO2014092550A2 (en) | 2012-12-10 | 2013-12-10 | Method for camera motion estimation with presence of moving object |
Country Status (2)
Country | Link |
---|---|
MY (1) | MY188908A (en) |
WO (1) | WO2014092550A2 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105354578A (en) * | 2015-10-27 | 2016-02-24 | 安徽大学 | Multi-target object image matching method |
US9449397B2 (en) | 2014-10-15 | 2016-09-20 | Caterpillar Inc. | Real-time visual odometry system for determining motion of a machine with a range detection unit |
CN107257980A (en) * | 2015-03-18 | 2017-10-17 | 英特尔公司 | Local change in video detects |
EP3362988A4 (en) * | 2015-10-15 | 2018-09-12 | AGT International GmbH | A method and system for stabilizing video frames |
CN110689554A (en) * | 2019-09-25 | 2020-01-14 | 深圳大学 | Background motion estimation method and device for infrared image sequence and storage medium |
CN116385480A (en) * | 2023-02-03 | 2023-07-04 | 腾晖科技建筑智能(深圳)有限公司 | Detection method and system for moving object below tower crane |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105654061A (en) * | 2016-01-05 | 2016-06-08 | 安阳师范学院 | 3D face dynamic reconstruction method based on estimation compensation |
-
2012
- 2012-12-10 MY MYPI2012005333A patent/MY188908A/en unknown
-
2013
- 2013-12-10 WO PCT/MY2013/000264 patent/WO2014092550A2/en active Application Filing
Non-Patent Citations (3)
Title |
---|
ALPER YILMAZ ET AL: "Object tracking", ACM COMPUTING SURVEYS, vol. 38, no. 4, 25 December 2006 (2006-12-25), pages 13-es, XP055053833, ISSN: 0360-0300, DOI: 10.1145/1177352.1177355 * |
EVELAND C ET AL: "Background modeling for segmentation of video-rate stereo sequences", COMPUTER VISION AND PATTERN RECOGNITION, 1998. PROCEEDINGS. 1998 IEEE COMPUTER SOCIETY CONFERENCE ON SANTA BARBARA, CA, USA 23-25 JUNE 1998, LOS ALAMITOS, CA, USA,IEEE COMPUT. SOC, US, 23 June 1998 (1998-06-23), pages 266-271, XP010291609, DOI: 10.1109/CVPR.1998.698619 ISBN: 978-0-8186-8497-5 * |
Yingdi Xie: "Research on a Modified RANSAC and its Applications to Ellipse Detection from a Static Image and Motion Detection from Active Stereo Video Sequences", , 1 February 2010 (2010-02-01), pages 1-132, XP055124755, Retrieved from the Internet: URL:https://dspace.wul.waseda.ac.jp/dspace /bitstream/2065/34981/3/Honbun-5358.pdf [retrieved on 2014-06-23] * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9449397B2 (en) | 2014-10-15 | 2016-09-20 | Caterpillar Inc. | Real-time visual odometry system for determining motion of a machine with a range detection unit |
CN107257980A (en) * | 2015-03-18 | 2017-10-17 | 英特尔公司 | Local change in video detects |
EP3271867A4 (en) * | 2015-03-18 | 2018-08-15 | Intel Corporation | Local change detection in video |
EP3362988A4 (en) * | 2015-10-15 | 2018-09-12 | AGT International GmbH | A method and system for stabilizing video frames |
CN105354578A (en) * | 2015-10-27 | 2016-02-24 | 安徽大学 | Multi-target object image matching method |
CN110689554A (en) * | 2019-09-25 | 2020-01-14 | 深圳大学 | Background motion estimation method and device for infrared image sequence and storage medium |
CN110689554B (en) * | 2019-09-25 | 2022-04-12 | 深圳大学 | Background motion estimation method and device for infrared image sequence and storage medium |
CN116385480A (en) * | 2023-02-03 | 2023-07-04 | 腾晖科技建筑智能(深圳)有限公司 | Detection method and system for moving object below tower crane |
CN116385480B (en) * | 2023-02-03 | 2023-10-20 | 腾晖科技建筑智能(深圳)有限公司 | Detection method and system for moving object below tower crane |
Also Published As
Publication number | Publication date |
---|---|
WO2014092550A3 (en) | 2014-09-12 |
MY188908A (en) | 2022-01-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014092550A2 (en) | Method for camera motion estimation with presence of moving object | |
CN106056089B (en) | A kind of 3 d pose recognition methods and system | |
CN106503671B (en) | The method and apparatus for determining human face posture | |
CN107316326B (en) | Edge-based disparity map calculation method and device applied to binocular stereo vision | |
US9338437B2 (en) | Apparatus and method for reconstructing high density three-dimensional image | |
CN109784130B (en) | Pedestrian re-identification method, device and equipment thereof | |
CN107798702B (en) | Real-time image superposition method and device for augmented reality | |
KR101524548B1 (en) | Apparatus and method for alignment of images | |
CN107798704B (en) | Real-time image superposition method and device for augmented reality | |
KR101139389B1 (en) | Video Analysing Apparatus and Method Using Stereo Cameras | |
CN101923641A (en) | Improved human face recognition method | |
CA2757681A1 (en) | Method and an apparatus for performing a cross-calculation | |
Gálai et al. | Feature selection for Lidar-based gait recognition | |
CN110505398B (en) | Image processing method and device, electronic equipment and storage medium | |
US9323989B2 (en) | Tracking device | |
CN110243390A (en) | The determination method, apparatus and odometer of pose | |
US20080226159A1 (en) | Method and System For Calculating Depth Information of Object in Image | |
CN110800020B (en) | Image information acquisition method, image processing equipment and computer storage medium | |
CN111382607B (en) | Living body detection method, living body detection device and face authentication system | |
US20150178927A1 (en) | Method and system for determining a transformation associated with a capturing device | |
CN110223219B (en) | 3D image generation method and device | |
CN105913395B (en) | A kind of observation of moving target and smear restoration method | |
CN107680083B (en) | Parallax determination method and parallax determination device | |
KR101501531B1 (en) | Stereo Vision-based Pedestrian Detection System and the method of | |
JP2017512398A (en) | Method and apparatus for presenting video |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13831951 Country of ref document: EP Kind code of ref document: A2 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 13831951 Country of ref document: EP Kind code of ref document: A2 |