US20090310822A1 - Feedback object detection method and system - Google Patents

Feedback object detection method and system Download PDF

Info

Publication number
US20090310822A1
US20090310822A1 US12/456,186 US45618609A US2009310822A1 US 20090310822 A1 US20090310822 A1 US 20090310822A1 US 45618609 A US45618609 A US 45618609A US 2009310822 A1 US2009310822 A1 US 2009310822A1
Authority
US
United States
Prior art keywords
pixel
image
feedback
pixels
object detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/456,186
Inventor
Chih-Hao Chang
Zhong-Lan Yang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vatics Inc
Original Assignee
Vatics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vatics Inc filed Critical Vatics Inc
Assigned to VATICS INC. reassignment VATICS INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHANG, CHIH-HAO, YANG, Zhong-lan
Publication of US20090310822A1 publication Critical patent/US20090310822A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/187Segmentation; Edge detection involving region growing; involving region merging; involving connected component labelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • G06T2207/20012Locally adaptive
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/62Extraction of image or video features relating to a temporal dimension, e.g. time-based feature extraction; Pattern tracking

Definitions

  • Object detection is one rapidly developed subject of the technological fields, and capable of getting a lot of information from images.
  • the most important concept of object detection is to extract object from images to be analyzed, and then track the changes in appearances or positions of the objects.
  • applications such as intelligent video surveillance system, computer vision, man-machine communication interface, image compression, it is of vital importance.
  • FIG. 1 a functional block diagram illustrating a conventional object detection system.
  • the conventional object detection system basically includes three elements—object segmentation element 102 , object acquisition element 104 and object tracking element 106 .
  • the images are firstly inputted to the object segmentation element 102 to obtain a binary mask in which the foreground pixels are extracted from the image.
  • the binary mask is processed by the object acquisition element 104 to collect the features of the foreground pixels and grouping related foreground pixels into objects.
  • a typical method to acquire objects is connected component labeling algorithm.
  • the objects in different images are tracked by the object tracking element 106 to realize their changes in appearances or positions.
  • the analysis results are outputted and the object information such as object speed, object category and object interaction is thus received.
  • This approach compares the pixel information including color and brightness of each pixel in the current image with that of the previous image. If the difference is greater than a predetermined threshold, the corresponding pixel is considered as a foreground pixel.
  • the threshold value affects the sensitivity of the segmentation.
  • the calculation of this approach is relatively simple.
  • One drawback of this approach is that the foreground object cannot be segmented from the image if it is not moving.
  • pixels are compared with the nearby pixels to calculate the similarity. After a certain calculation, pixels having similar properties are merged and segmented from the image.
  • the threshold value or sensitivity affects the similarity variation tolerance in the region. No background model is required for this approach. The calculation is more difficult than the frame difference approach.
  • One drawback of this approach is that only object having homogenous feature can be segmented from the image. Further, an object is often composed of several different parts with different features.
  • This approach establishes a background model based on historical images. By subtracting the background model from the current image, the foreground object is obtained.
  • This approach has the highest reliability among the three approaches and is suitable for analyzing images having dynamic background. However, it is necessary to maintain the background model frequently.
  • False alarm is an annoying problem for the above-described object segmentation methods since only pixel connection or pixel change is considered. Local change such as flash or shadow affects the object segmentation very much. Besides, noise is probably considered as a foreground object. These accidental factors trigger and increase false alarms. These problems are sometimes overcome by adjusting the threshold value or sensitivity. The determination of the threshold value or sensitivity is always in a dilemma. If the threshold value is too high, the foreground pixels cannot be segmented from the image when the foreground pixels are somewhat similar to the background pixels. Hence, a single object may be separated into more than one part in the object segmentation procedure if some pixels within the object share similar properties with the background pixels. On the other hand, if the threshold value is too low, noise and brightness variation are identified as foreground objects. Hence, the fixed threshold value does not satisfy the accuracy requirement for the object segmentation.
  • controllable threshold values and sensitivities may be considered to achieve smart object detection.
  • the present invention provides a feedback object detection method to increase accuracy in object segmentation.
  • the object is extracted from an image based on prediction information of the object. Then, the extracted object is tracked to generate motion information such as moving speed and moving direction of the object. From the motion information, another prediction information is derived for the analysis of the next image.
  • the threshold value for each pixel in the extracting step is adjustable. If one pixel is a predicted foreground pixel, the threshold value of the pixel decreases. On the contrary, if one pixel is a predicted background pixel, the threshold value of the pixel increases.
  • a feedback object detection system is also provided.
  • the system includes an object segmentation element, an object tracking element and an object prediction element.
  • the object segmentation element extracts the object from the first image according to prediction information of the object provided by the object prediction element. Then, the object tracking element tracks the extracted object to generate motion information of the object such as moving speed and moving function.
  • the object prediction element generates the prediction information of the object according to the motion information. In an embodiment, the prediction information indicates the possible position and size of the object to facilitate the object segmentation.
  • the system further includes an object acquisition element for calculating object information of the extracted object by performing a connected component labeling algorithm on the foreground pixels.
  • the object information may be color distribution, center of mass or size of the object. Then, the object tracking element tracks the motion of the object according to the object information derived from different images.
  • FIG. 1 is a functional block diagram illustrating the conventional object detection system
  • FIGS. 2A ⁇ 2C illustrate three types of known object segmentation procedures applied to the object segmentation element of FIG. 1 ;
  • FIG. 3 is a functional block diagram illustrating a preferred embodiment of a feedback object detection system according to the present invention
  • FIG. 4 is a flowchart illustrating an object segmentation procedure according to the present invention.
  • FIG. 5 is a flowchart illustrating an object prediction procedure according to the present invention.
  • the feedback object detection system includes one element, object prediction element 308 , more than the conventional object detection system.
  • the object prediction element 308 generates prediction information of objects to indicate the possible positions and sizes of the objects in the next image. Accordingly, the object segmentation element 302 obtains a binary mask by considering the current image and the prediction information of the known objects. If one pixel is located in the predicted regions of the objects, the object segmentation element 302 increases the probability that the pixel is determined as a foreground pixel in the current image.
  • the pixels in the current image may be assigned with different segmentation sensitivities to obtain a proper binary mask which accurately distinguishes the foreground pixels from the background pixels.
  • the binary mask is processed by the object acquisition element 304 to collect the features of the foreground pixels and group related foreground pixels into objects.
  • a typical method for acquiring objects is connected component labeling algorithm.
  • the feature of each segmented object for example color distribution, center of mass and size, is calculated.
  • the objects in different images are tracked by the object tracking element 306 by comparing the acquired features of corresponding objects in sequential images to realize their changes in appearances and positions.
  • the analysis results are outputted and the object information such as object speed, object category and object interaction is thus received.
  • the analysis results are also processed by the object prediction element 308 to get the prediction information for the segmentation of the next image.
  • the sensitivity and the threshold value for object segmentation according to the present invention become variable in the entire image. If the pixel is supposed to be a foreground pixel, the threshold value for this pixel is decreased to raise the sensitivity of the segmentation procedure. Otherwise, if the pixel is supposed to be a background pixel, the threshold value for this pixel is increased to lower the sensitivity of the segmentation procedure.
  • FIG. 4 is a flowchart illustrating the object segmentation procedure for one pixel using the variable threshold value (sensitivity) and the later two of these approaches.
  • the variable threshold value (sensitivity) may be applied to only background subtraction without the other two.
  • the prediction information is inputted to the object segmentation element.
  • the current pixel is preliminarily determined as a predicted foreground pixel or a predicted background pixel (step 404 ). If it is supposed that the current pixel is a foreground pixel, the threshold value of the pixel is decreased to raise the sensitivity. On the other hand, if it is supposed that the current pixel is a background pixel, the threshold value is increased to lower the sensitivity (step 406 ).
  • Steps 410 ⁇ 416 correspond to region merge approach.
  • the current pixel is compared with nearby pixels (step 412 ).
  • the similarity variation between the current pixel and the nearby pixels is obtained after a certain calculation (step 414 ).
  • the similarity variation is compared with the adjusted threshold value to find out a first probability of that the current pixel is the foreground pixel (step 416 ). Accordingly, this path from step 410 to step 416 is a spatial based segmentation.
  • Steps 420 ⁇ 428 correspond to background subtraction approach.
  • Historical images are analyzed to establish a background model (steps 420 and 422 ).
  • the background model may be selected from a still model, a probability distribution model, and a mixed Gaussian distribution model according to the requirements.
  • the established background model is then subtracted from the current image to get the difference at current pixel (steps 424 and 426 ).
  • the difference is compared with the adjusted threshold value to find out a second probability of that the current pixel is the foreground pixel (step 428 ). Accordingly, this path from step 420 to step 428 is a temporal based segmentation.
  • the procedure determines at step 430 whether the current pixel is a foreground pixel by considering the probabilities obtained at steps 416 and 428 .
  • the adjustable threshold value obtained at step 406 significantly increases the accuracy in the last determination.
  • the procedure repeats for all pixels till the current image is completely analyzed to obtain a binary mask for the object acquisition element.
  • the object segmentation procedure can solve the problems incurred by the prior arts.
  • the object is not segmented into multiple parts even some pixels within the object has similar feature as the background.
  • the decrease of threshold value of these pixels can compensate this phenomenon.
  • the reflected light or shadow does not force the background pixels to be segmented as foreground pixels since the increase of threshold value reduce the probability of misclassifying them as foreground pixels.
  • one object is not moving, it is still considered as a foreground object rather than be learnt into the background model.
  • the object prediction information may include object motion information, object category information, environment information, object depth information, interaction information, etc.
  • Object motion information includes speed and position of the object. It is basic information associated with other object prediction information.
  • Object category information indicates the categories of the object, for example a car, a bike or a human. It is apparent that the predicted speed is from fast to slow in this order. Furthermore, a human usually has more irregular moving track than a car. Hence, for a human, more historical images are required to analyze and predict the position in the next image.
  • Environment information indicates where the object is located. If the object is moving down a hill, the acceleration results in an increasing speed. If the object is moving toward a nearby exit, it may predict that the object disappear in the next image and no predict position is provided for the object segmentation element.
  • Object depth information indicates a distance between the object and the video camera. If the object is moving toward the video camera, the size of the object becomes bigger and bigger in the following images. On the contrary, if the object is moving away from the video camera, the object is of smaller and smaller size.
  • Interaction information is high-level and more complicated information. For example, one person is moving behind a pillar. The person temporarily disappears in the images. The object prediction element can predict the moving after he appears again according to the historical images before his walking behind the pillar.
  • the object motion information is taken as an example for further description.
  • the position and motion vector of object k at time t is respectively expressed as Pos(Obj(k), t) and MV(Obj(k), t).
  • a motion prediction function MP(Obj(k), t) is defined as:
  • the predicted position of the object Predict_pos(Obj(k), t+1) may be obtained by adding the motion prediction function to the current position as the following equation:
  • Predict_pos(Obj( k ), t+ 1) Pos(Obj( k ), t )+ MP (Obj( k ), t ) (3)
  • pixels within the prediction region of the object are preliminarily considered as foreground pixels.
  • FIG. 5 a flowchart illustrating a simple object prediction used for obtaining object motion information as explained above.
  • information of a specific object in the current image and previous image, provided by the object tracking element is inputted (steps 602 and 606 ).
  • the current object position Pos(Obj(k), t) and the previous object position Pos(Obj(k), t ⁇ 1) are picked form the inputted information (steps 604 and 608 ).
  • the procedure calculates the current object motion MV(Obj(k), t) (step 610 ).
  • the term “motion” indicates a motion vector consisting of moving speed and moving direction.
  • the object motion in the current and historical images is collected (step 612 ).
  • motion prediction function MP(Obj(k), t) is obtained by the calculation related to the object motion MV(Obj(k), t) and earlier object motion MV(Obj(k), t ⁇ 1), MV(Obj(k), t ⁇ 2), . . . (step 614 ).
  • the procedure successfully predicts the object position Predict_pos(Obj(k), t+1) in the next image (step 618 ).
  • the present feedback object detection method utilizes the prediction information of objects to facilitate the segmentation determination of the pixels.
  • the variable threshold value flexibly adjusts the segmentation sensitivities along the entire image so as to increase the accuracy of object segmentation.
  • the dilemma of neglecting noise or extracting all existing objects in the image resulted from fixed threshold value is thus solved. It is applicable to take advantage of this feedback object detection method in many fields including intelligent video surveillance system, computer vision, man-machine communication interface and image compression because of the high-level segmentation and detection ability.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

A feedback object detection method and system. The system includes an object segmentation element, an object tracking element and an object prediction element. The object segmentation element extracts the object from an image according to prediction information of the object provided by the object prediction element. Then, the object tracking element tracks the extracted object to generate motion information of the object like moving speed and moving direction. The object prediction element generates the prediction information such as predicted position and predicted size of the object according to the motion information. The feedback of the prediction information to the object segmentation element facilitates accurately extracting foreground pixels from the image.

Description

    FIELD OF THE INVENTION
  • The present invention relates to an object detection method and system, and more particularly to an object detection method and system using feedback mechanism in object segmentation.
  • BACKGROUND OF THE INVENTION
  • Nowadays, image processing is applied to many systems. It covers many technological fields. Object detection is one rapidly developed subject of the technological fields, and capable of getting a lot of information from images. The most important concept of object detection is to extract object from images to be analyzed, and then track the changes in appearances or positions of the objects. For many applications such as intelligent video surveillance system, computer vision, man-machine communication interface, image compression, it is of vital importance.
  • Compared with the conventional video surveillance system, the intelligent video surveillance systems adopting object detection may economize manpower for the purpose of monitoring the systems every moment. The requirement of accuracy of object detection tends to increase to improve the monitor efficiency. If the accuracy can reach a satisfying level, many events, for example dangerous article left over in public place or suspicious character loitering around guarded region, can be detected, recorded and alarmed automatically.
  • Please refer to FIG. 1, a functional block diagram illustrating a conventional object detection system. The conventional object detection system basically includes three elements—object segmentation element 102, object acquisition element 104 and object tracking element 106. The images are firstly inputted to the object segmentation element 102 to obtain a binary mask in which the foreground pixels are extracted from the image. Then, the binary mask is processed by the object acquisition element 104 to collect the features of the foreground pixels and grouping related foreground pixels into objects. A typical method to acquire objects is connected component labeling algorithm. At last, the objects in different images are tracked by the object tracking element 106 to realize their changes in appearances or positions. The analysis results are outputted and the object information such as object speed, object category and object interaction is thus received.
  • There are some approaches proposed for the object segmentation. FIGS. 2A˜2C illustrate three of these approaches: frame difference, region merge, and background subtraction, respectively.
  • (1) Frame Difference (FIG. 2A):
  • This approach compares the pixel information including color and brightness of each pixel in the current image with that of the previous image. If the difference is greater than a predetermined threshold, the corresponding pixel is considered as a foreground pixel. The threshold value affects the sensitivity of the segmentation. The calculation of this approach is relatively simple. One drawback of this approach is that the foreground object cannot be segmented from the image if it is not moving.
  • (2) Region Merge (FIG. 2B):
  • In this approach, pixels are compared with the nearby pixels to calculate the similarity. After a certain calculation, pixels having similar properties are merged and segmented from the image. The threshold value or sensitivity affects the similarity variation tolerance in the region. No background model is required for this approach. The calculation is more difficult than the frame difference approach. One drawback of this approach is that only object having homogenous feature can be segmented from the image. Further, an object is often composed of several different parts with different features.
  • (3) Background Subtraction (FIG. 2C):
  • This approach establishes a background model based on historical images. By subtracting the background model from the current image, the foreground object is obtained. This approach has the highest reliability among the three approaches and is suitable for analyzing images having dynamic background. However, it is necessary to maintain the background model frequently.
  • False alarm is an annoying problem for the above-described object segmentation methods since only pixel connection or pixel change is considered. Local change such as flash or shadow affects the object segmentation very much. Besides, noise is probably considered as a foreground object. These accidental factors trigger and increase false alarms. These problems are sometimes overcome by adjusting the threshold value or sensitivity. The determination of the threshold value or sensitivity is always in a dilemma. If the threshold value is too high, the foreground pixels cannot be segmented from the image when the foreground pixels are somewhat similar to the background pixels. Hence, a single object may be separated into more than one part in the object segmentation procedure if some pixels within the object share similar properties with the background pixels. On the other hand, if the threshold value is too low, noise and brightness variation are identified as foreground objects. Hence, the fixed threshold value does not satisfy the accuracy requirement for the object segmentation.
  • Therefore, there is a need of providing an efficient object detection method and system to reduce the frequency of false alarm. In particular, controllable threshold values and sensitivities may be considered to achieve smart object detection.
  • SUMMARY OF THE INVENTION
  • The present invention provides a feedback object detection method to increase accuracy in object segmentation. According to the feedback object detection method, the object is extracted from an image based on prediction information of the object. Then, the extracted object is tracked to generate motion information such as moving speed and moving direction of the object. From the motion information, another prediction information is derived for the analysis of the next image.
  • In an embodiment, the threshold value for each pixel in the extracting step is adjustable. If one pixel is a predicted foreground pixel, the threshold value of the pixel decreases. On the contrary, if one pixel is a predicted background pixel, the threshold value of the pixel increases.
  • A feedback object detection system is also provided. The system includes an object segmentation element, an object tracking element and an object prediction element. The object segmentation element extracts the object from the first image according to prediction information of the object provided by the object prediction element. Then, the object tracking element tracks the extracted object to generate motion information of the object such as moving speed and moving function. The object prediction element generates the prediction information of the object according to the motion information. In an embodiment, the prediction information indicates the possible position and size of the object to facilitate the object segmentation.
  • In an embodiment, the system further includes an object acquisition element for calculating object information of the extracted object by performing a connected component labeling algorithm on the foreground pixels. The object information may be color distribution, center of mass or size of the object. Then, the object tracking element tracks the motion of the object according to the object information derived from different images.
  • An object segmentation method is further provided to analyze an image consisting of a plurality of pixels, a portion of which constitutes an object. Prediction information of the object such as predicted position and predicted size is provided, and segmentation sensitivity for each pixel is adjusted according to the prediction information. Each pixel is determined to be a foreground pixel or a background pixel according to its property and the corresponding segmentation sensitivity.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above contents of the present invention will become more readily apparent to those ordinarily skilled in the art after reviewing the following detailed description and accompanying drawings, in which:
  • FIG. 1 is a functional block diagram illustrating the conventional object detection system;
  • FIGS. 2A˜2C illustrate three types of known object segmentation procedures applied to the object segmentation element of FIG. 1;
  • FIG. 3 is a functional block diagram illustrating a preferred embodiment of a feedback object detection system according to the present invention;
  • FIG. 4 is a flowchart illustrating an object segmentation procedure according to the present invention; and
  • FIG. 5 is a flowchart illustrating an object prediction procedure according to the present invention.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
  • The present invention will now be described more specifically with reference to the following embodiments. It is to be noted that the following descriptions of preferred embodiments of this invention are presented herein for purpose of illustration and description only. It is not intended to be exhaustive or to be limited to the precise form disclosed.
  • Please refer to FIG. 3, a functional block diagram illustrating a feedback object detection system according to the present invention. The feedback object detection system includes one element, object prediction element 308, more than the conventional object detection system. The object prediction element 308 generates prediction information of objects to indicate the possible positions and sizes of the objects in the next image. Accordingly, the object segmentation element 302 obtains a binary mask by considering the current image and the prediction information of the known objects. If one pixel is located in the predicted regions of the objects, the object segmentation element 302 increases the probability that the pixel is determined as a foreground pixel in the current image. The pixels in the current image may be assigned with different segmentation sensitivities to obtain a proper binary mask which accurately distinguishes the foreground pixels from the background pixels.
  • Then, the binary mask is processed by the object acquisition element 304 to collect the features of the foreground pixels and group related foreground pixels into objects. A typical method for acquiring objects is connected component labeling algorithm. At this stage, the feature of each segmented object, for example color distribution, center of mass and size, is calculated. At last, the objects in different images are tracked by the object tracking element 306 by comparing the acquired features of corresponding objects in sequential images to realize their changes in appearances and positions. The analysis results are outputted and the object information such as object speed, object category and object interaction is thus received. The analysis results are also processed by the object prediction element 308 to get the prediction information for the segmentation of the next image.
  • Compared with the conventional object segmentation procedure, the sensitivity and the threshold value for object segmentation according to the present invention become variable in the entire image. If the pixel is supposed to be a foreground pixel, the threshold value for this pixel is decreased to raise the sensitivity of the segmentation procedure. Otherwise, if the pixel is supposed to be a background pixel, the threshold value for this pixel is increased to lower the sensitivity of the segmentation procedure.
  • As mentioned above, there are three known approaches for the object segmentation, including frame difference, region merge, and background subtraction. The variable threshold value and sensitivity of the present invention can be used with one, all, or combination of these approaches. FIG. 4 is a flowchart illustrating the object segmentation procedure for one pixel using the variable threshold value (sensitivity) and the later two of these approaches. This embodiment is just for description, but not limiting the scope of the invention. For example, the variable threshold value (sensitivity) may be applied to only background subtraction without the other two.
  • At step 402, the prediction information is inputted to the object segmentation element. According to the prediction information such as object positions and object sizes, the current pixel is preliminarily determined as a predicted foreground pixel or a predicted background pixel (step 404). If it is supposed that the current pixel is a foreground pixel, the threshold value of the pixel is decreased to raise the sensitivity. On the other hand, if it is supposed that the current pixel is a background pixel, the threshold value is increased to lower the sensitivity (step 406).
  • Steps 410˜416 correspond to region merge approach. After the input of the current image (step 410), the current pixel is compared with nearby pixels (step 412). The similarity variation between the current pixel and the nearby pixels is obtained after a certain calculation (step 414). Then, the similarity variation is compared with the adjusted threshold value to find out a first probability of that the current pixel is the foreground pixel (step 416). Accordingly, this path from step 410 to step 416 is a spatial based segmentation.
  • Steps 420˜428 correspond to background subtraction approach. Historical images are analyzed to establish a background model (steps 420 and 422). The background model may be selected from a still model, a probability distribution model, and a mixed Gaussian distribution model according to the requirements. The established background model is then subtracted from the current image to get the difference at current pixel (steps 424 and 426). The difference is compared with the adjusted threshold value to find out a second probability of that the current pixel is the foreground pixel (step 428). Accordingly, this path from step 420 to step 428 is a temporal based segmentation.
  • At last, the procedure determines at step 430 whether the current pixel is a foreground pixel by considering the probabilities obtained at steps 416 and 428. The adjustable threshold value obtained at step 406 significantly increases the accuracy in the last determination. The procedure repeats for all pixels till the current image is completely analyzed to obtain a binary mask for the object acquisition element.
  • According to the present invention, the object segmentation procedure can solve the problems incurred by the prior arts. First of all, the object is not segmented into multiple parts even some pixels within the object has similar feature as the background. The decrease of threshold value of these pixels can compensate this phenomenon. Secondly, the reflected light or shadow does not force the background pixels to be segmented as foreground pixels since the increase of threshold value reduce the probability of misclassifying them as foreground pixels. Finally, if one object is not moving, it is still considered as a foreground object rather than be learnt into the background model.
  • From the above description, the object prediction information fed back to the object segmentation element affects the controllable threshold value very much. Some object prediction information is explained herein. The object prediction information may include object motion information, object category information, environment information, object depth information, interaction information, etc.
  • Object motion information includes speed and position of the object. It is basic information associated with other object prediction information.
  • Object category information indicates the categories of the object, for example a car, a bike or a human. It is apparent that the predicted speed is from fast to slow in this order. Furthermore, a human usually has more irregular moving track than a car. Hence, for a human, more historical images are required to analyze and predict the position in the next image.
  • Environment information indicates where the object is located. If the object is moving down a hill, the acceleration results in an increasing speed. If the object is moving toward a nearby exit, it may predict that the object disappear in the next image and no predict position is provided for the object segmentation element.
  • Object depth information indicates a distance between the object and the video camera. If the object is moving toward the video camera, the size of the object becomes bigger and bigger in the following images. On the contrary, if the object is moving away from the video camera, the object is of smaller and smaller size.
  • Interaction information is high-level and more complicated information. For example, one person is moving behind a pillar. The person temporarily disappears in the images. The object prediction element can predict the moving after he appears again according to the historical images before his walking behind the pillar.
  • The object motion information is taken as an example for further description. The position and motion vector of object k at time t is respectively expressed as Pos(Obj(k), t) and MV(Obj(k), t).

  • MV(Obj(k), t)=Pos(Obj(k), t)−Pos(Obj(k), t−1)   (1)
  • A motion prediction function MP(Obj(k), t) is defined as:

  • MP(Obj(k), t)=(MV(Obj(k), t)+MV(Obj(k), t−1)+MV(Obj(k), t−2)+ . . . )low pass   (2)
  • A low pass filter is used in the above equation to filter out the possible irregular motion. Accordingly, the predicted position of the object Predict_pos(Obj(k), t+1) may be obtained by adding the motion prediction function to the current position as the following equation:

  • Predict_pos(Obj(k), t+1)=Pos(Obj(k), t)+MP(Obj(k), t)   (3)
  • Thus, pixels within the prediction region of the object are preliminarily considered as foreground pixels.
  • Please refer to FIG. 5, a flowchart illustrating a simple object prediction used for obtaining object motion information as explained above. At first, information of a specific object in the current image and previous image, provided by the object tracking element, is inputted (steps 602 and 606). The current object position Pos(Obj(k), t) and the previous object position Pos(Obj(k), t−1) are picked form the inputted information (steps 604 and 608). By comparing the two positions, the procedure calculates the current object motion MV(Obj(k), t) (step 610). In this embodiment, the term “motion” indicates a motion vector consisting of moving speed and moving direction. The object motion in the current and historical images is collected (step 612). Then, motion prediction function MP(Obj(k), t) is obtained by the calculation related to the object motion MV(Obj(k), t) and earlier object motion MV(Obj(k), t−1), MV(Obj(k), t−2), . . . (step 614). By adding the motion prediction function MP(Obj(k), t) to the current object position Pos(Obj(k), t), the procedure successfully predicts the object position Predict_pos(Obj(k), t+1) in the next image (step 618). These steps repeat till all the objects have the corresponding prediction information assisting in object segmentation as described above.
  • From the above description, the present feedback object detection method utilizes the prediction information of objects to facilitate the segmentation determination of the pixels. The variable threshold value flexibly adjusts the segmentation sensitivities along the entire image so as to increase the accuracy of object segmentation. The dilemma of neglecting noise or extracting all existing objects in the image resulted from fixed threshold value is thus solved. It is applicable to take advantage of this feedback object detection method in many fields including intelligent video surveillance system, computer vision, man-machine communication interface and image compression because of the high-level segmentation and detection ability.
  • While the invention has been described in terms of what is presently considered to be the most practical and preferred embodiments, it is to be understood that the invention needs not to be limited to the disclosed embodiment. On the contrary, it is intended to cover various modifications and similar arrangements included within the spirit and scope of the appended claims which are to be accorded with the broadest interpretation so as to encompass all such modifications and similar structures.

Claims (20)

1. A feedback object detection method, comprising steps of:
receiving a first image comprising an object;
receiving prediction information of the object
extracting the object from the first image according to the prediction information;
tracking the extracted object to generate motion information of the object; and
generating prediction information of the object corresponding to a second image later than the first image according to the motion information.
2. The feedback object detection method according to claim 1 wherein the prediction information indicates that a portion of pixels in the first image are predicted foreground pixels.
3. The feedback object detection method according to claim 2 wherein the extracting step further comprises a step of adjusting a threshold value for each pixel according to the prediction information to determine whether a selected pixel in the first image is a foreground pixel or a background pixel.
4. The feedback object detection method according to claim 3 wherein the adjusting step comprises steps of:
decreasing the threshold value when the selected pixel is one of the predicted foreground pixels; and
increasing the threshold value when the selected pixel is not one of the predicted foreground pixels.
5. The feedback object detection method according to claim 3 wherein the extracting step further comprises steps of:
comparing the first image with a background model to get a first difference for the selected pixel; and
determining the selected pixel is the foreground pixel when the first difference is greater than the threshold value.
6. The feedback object detection method according to claim 5 wherein the background model is a mixed Gaussian background model, a probability distribution background model, or a still background model.
7. The feedback object detection method according to claim 3 wherein the extracting step further comprises steps of:
comparing the selected pixel with the nearby pixels to get a second difference; and
determining the selected pixel is the foreground pixel when the second difference is greater than the threshold value.
8. The feedback object detection method according to claim 1, further comprising a step of calculating object information of the extracted object.
9. The feedback object detection method according to claim 8 wherein the object information is one selected from a group consisting of color distribution, center of mass, size and a combination thereof.
10. The feedback object detection method according to claim 9 wherein the extracted object is tracked according to the similarity of the object information between the first image and a third image earlier than the first image.
11. The feedback object detection method according to claim 1 wherein the motion information includes moving speed and moving direction of the tracked object.
12. A feedback object detection system for detecting an object in an image, comprising:
an object segmentation element for extracting the object from the first image according to prediction information;
an object tracking element for tracking the extracted object to generate motion information of the object; and
an object prediction element for generating the prediction information of the object according to the motion information.
13. The feedback object detection system according to claim 12 wherein the prediction information indicates that a portion of pixels in the image are predicted foreground pixels.
14. The feedback object detection system according to claim 13 wherein the object segmentation element adjusts a threshold value for each pixel according to the prediction information to determine whether a selected pixel in the image is a foreground pixel or a background pixel.
15. The feedback object detection system according to claim 14 wherein the object segmentation element decreases the threshold value when the selected pixel is one of the predicted foreground pixels, and increases the threshold value when the selected pixel is not one of the predicted foreground pixels.
16. The feedback object detection system according to claim 14 wherein the selected pixel is determined to be one of the foreground pixel and the background pixel according to a property of the selected pixel and the threshold value.
17. The feedback object detection system according to claim 12, further comprising an object acquisition element for calculating object information of the extracted object.
18. The feedback object detection system according to claim 17 wherein the object acquisition element performs a connected component labeling algorithm on the pixels determined as the foreground pixels to obtain the object information.
19. An object segmentation method for analyzing an image comprising a plurality of pixels, a portion of the pixels constituting an object, comprising steps of:
receiving prediction information of the object;
adjusting a segmentation sensitivity for each pixel according to the prediction information; and
for each pixel, determining whether the pixel is a foreground pixel or a background pixel according to a property of the pixel by considering the segmentation sensitivity corresponding to the pixel.
20. The object segmentation method according to claim 19 wherein the prediction information indicates that a first portion of the pixels are predicted foreground pixels, wherein the segmentation sensitivity of a selected pixel increases when the selected pixel is one of the predicted foreground pixels.
US12/456,186 2008-06-11 2009-06-11 Feedback object detection method and system Abandoned US20090310822A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
TW097121629A TWI420401B (en) 2008-06-11 2008-06-11 Algorithm for feedback type object detection
TW097121629 2008-06-11

Publications (1)

Publication Number Publication Date
US20090310822A1 true US20090310822A1 (en) 2009-12-17

Family

ID=41414828

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/456,186 Abandoned US20090310822A1 (en) 2008-06-11 2009-06-11 Feedback object detection method and system

Country Status (2)

Country Link
US (1) US20090310822A1 (en)
TW (1) TWI420401B (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110280478A1 (en) * 2010-05-13 2011-11-17 Hon Hai Precision Industry Co., Ltd. Object monitoring system and method
US20110280442A1 (en) * 2010-05-13 2011-11-17 Hon Hai Precision Industry Co., Ltd. Object monitoring system and method
US20120121191A1 (en) * 2010-11-16 2012-05-17 Electronics And Telecommunications Research Institute Image separation apparatus and method
US20140301604A1 (en) * 2011-11-01 2014-10-09 Canon Kabushiki Kaisha Method and system for luminance adjustment of images in an image sequence
CN104658007A (en) * 2013-11-25 2015-05-27 华为技术有限公司 Identifying method and device for actual moving targets
WO2016186649A1 (en) * 2015-05-19 2016-11-24 Hewlett Packard Enterprise Development Lp Database comparison operation to identify an object
US9530215B2 (en) * 2015-03-20 2016-12-27 Qualcomm Incorporated Systems and methods for enhanced depth map retrieval for moving objects using active sensing technology
US9635339B2 (en) 2015-08-14 2017-04-25 Qualcomm Incorporated Memory-efficient coded light error correction
US9846943B2 (en) 2015-08-31 2017-12-19 Qualcomm Incorporated Code domain power control for structured light
US9948920B2 (en) 2015-02-27 2018-04-17 Qualcomm Incorporated Systems and methods for error correction in structured light
US10068338B2 (en) 2015-03-12 2018-09-04 Qualcomm Incorporated Active sensing spatial resolution improvement through multiple receivers and code reuse
US10205953B2 (en) * 2012-01-26 2019-02-12 Apple Inc. Object detection informed encoding
US10310087B2 (en) * 2017-05-31 2019-06-04 Uber Technologies, Inc. Range-view LIDAR-based object detection
EP4156098A1 (en) 2021-09-22 2023-03-29 Axis AB A segmentation method
DE102019107103B4 (en) 2018-03-20 2023-08-17 Logitech Europe S.A. METHOD AND SYSTEM FOR OBJECT SEGMENTATION IN A MIXED REALITY ENVIRONMENT
US11885910B2 (en) 2017-05-31 2024-01-30 Uatc, Llc Hybrid-view LIDAR-based object detection

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI518601B (en) * 2014-05-28 2016-01-21 廣達電腦股份有限公司 Devices and methods of information extraction
TWI656507B (en) * 2017-08-21 2019-04-11 瑞昱半導體股份有限公司 Electronic device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6075875A (en) * 1996-09-30 2000-06-13 Microsoft Corporation Segmentation of image features using hierarchical analysis of multi-valued image data and weighted averaging of segmentation results
US6141433A (en) * 1997-06-19 2000-10-31 Ncr Corporation System and method for segmenting image regions from a scene likely to represent particular objects in the scene
US6999620B1 (en) * 2001-12-10 2006-02-14 Hewlett-Packard Development Company, L.P. Segmenting video input using high-level feedback
US20060170769A1 (en) * 2005-01-31 2006-08-03 Jianpeng Zhou Human and object recognition in digital video
US20070273765A1 (en) * 2004-06-14 2007-11-29 Agency For Science, Technology And Research Method for Detecting Desired Objects in a Highly Dynamic Environment by a Monitoring System
US20090110236A1 (en) * 2007-10-29 2009-04-30 Ching-Chun Huang Method And System For Object Detection And Tracking

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TW200733004A (en) * 2006-02-22 2007-09-01 Huper Lab Co Ltd Method for video object segmentation
US8340185B2 (en) * 2006-06-27 2012-12-25 Marvell World Trade Ltd. Systems and methods for a motion compensated picture rate converter

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6075875A (en) * 1996-09-30 2000-06-13 Microsoft Corporation Segmentation of image features using hierarchical analysis of multi-valued image data and weighted averaging of segmentation results
US6141433A (en) * 1997-06-19 2000-10-31 Ncr Corporation System and method for segmenting image regions from a scene likely to represent particular objects in the scene
US6999620B1 (en) * 2001-12-10 2006-02-14 Hewlett-Packard Development Company, L.P. Segmenting video input using high-level feedback
US20070273765A1 (en) * 2004-06-14 2007-11-29 Agency For Science, Technology And Research Method for Detecting Desired Objects in a Highly Dynamic Environment by a Monitoring System
US20060170769A1 (en) * 2005-01-31 2006-08-03 Jianpeng Zhou Human and object recognition in digital video
US20090110236A1 (en) * 2007-10-29 2009-04-30 Ching-Chun Huang Method And System For Object Detection And Tracking

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110280442A1 (en) * 2010-05-13 2011-11-17 Hon Hai Precision Industry Co., Ltd. Object monitoring system and method
US20110280478A1 (en) * 2010-05-13 2011-11-17 Hon Hai Precision Industry Co., Ltd. Object monitoring system and method
US20120121191A1 (en) * 2010-11-16 2012-05-17 Electronics And Telecommunications Research Institute Image separation apparatus and method
US9609233B2 (en) * 2011-11-01 2017-03-28 Canon Kabushiki Kaisha Method and system for luminance adjustment of images in an image sequence
US20140301604A1 (en) * 2011-11-01 2014-10-09 Canon Kabushiki Kaisha Method and system for luminance adjustment of images in an image sequence
US10205953B2 (en) * 2012-01-26 2019-02-12 Apple Inc. Object detection informed encoding
CN104658007A (en) * 2013-11-25 2015-05-27 华为技术有限公司 Identifying method and device for actual moving targets
US9948920B2 (en) 2015-02-27 2018-04-17 Qualcomm Incorporated Systems and methods for error correction in structured light
US10068338B2 (en) 2015-03-12 2018-09-04 Qualcomm Incorporated Active sensing spatial resolution improvement through multiple receivers and code reuse
US9530215B2 (en) * 2015-03-20 2016-12-27 Qualcomm Incorporated Systems and methods for enhanced depth map retrieval for moving objects using active sensing technology
WO2016186649A1 (en) * 2015-05-19 2016-11-24 Hewlett Packard Enterprise Development Lp Database comparison operation to identify an object
US10956493B2 (en) 2015-05-19 2021-03-23 Micro Focus Llc Database comparison operation to identify an object
US9635339B2 (en) 2015-08-14 2017-04-25 Qualcomm Incorporated Memory-efficient coded light error correction
US9846943B2 (en) 2015-08-31 2017-12-19 Qualcomm Incorporated Code domain power control for structured light
US10223801B2 (en) 2015-08-31 2019-03-05 Qualcomm Incorporated Code domain power control for structured light
US10310087B2 (en) * 2017-05-31 2019-06-04 Uber Technologies, Inc. Range-view LIDAR-based object detection
US11885910B2 (en) 2017-05-31 2024-01-30 Uatc, Llc Hybrid-view LIDAR-based object detection
DE102019107103B4 (en) 2018-03-20 2023-08-17 Logitech Europe S.A. METHOD AND SYSTEM FOR OBJECT SEGMENTATION IN A MIXED REALITY ENVIRONMENT
EP4156098A1 (en) 2021-09-22 2023-03-29 Axis AB A segmentation method

Also Published As

Publication number Publication date
TWI420401B (en) 2013-12-21
TW200951829A (en) 2009-12-16

Similar Documents

Publication Publication Date Title
US20090310822A1 (en) Feedback object detection method and system
US7916944B2 (en) System and method for feature level foreground segmentation
US7778445B2 (en) Method and system for the detection of removed objects in video images
Cucchiara et al. Detecting moving objects, ghosts, and shadows in video streams
EP2801078B1 (en) Context aware moving object detection
US7639840B2 (en) Method and apparatus for improved video surveillance through classification of detected objects
Lei et al. Real-time outdoor video surveillance with robust foreground extraction and object tracking via multi-state transition management
US9977970B2 (en) Method and system for detecting the occurrence of an interaction event via trajectory-based analysis
US9965687B2 (en) System and method for detecting potential mugging event via trajectory-based analysis
US20100310122A1 (en) Method and Device for Detecting Stationary Targets
KR101868103B1 (en) A video surveillance apparatus for identification and tracking multiple moving objects and method thereof
US10210392B2 (en) System and method for detecting potential drive-up drug deal activity via trajectory-based analysis
WO2018179202A1 (en) Information processing device, control method, and program
KR101646000B1 (en) Surveillance System And Method To Object And Area
Zaidi et al. Video anomaly detection and classification for human activity recognition
KR102159954B1 (en) Method for establishing region of interest in intelligent video analytics and video analysis apparatus using the same
Huang et al. Rapid detection of camera tampering and abnormal disturbance for video surveillance system
Lacabex et al. Lightweight tracking-by-detection system for multiple pedestrian targets
Monteiro et al. Robust segmentation for outdoor traffic surveillance
Sehairi et al. Real-time implementation of human action recognition system based on motion analysis
Foggia et al. A Method for Detecting Long Term Left Baggage based on Heat Map.
Borkar et al. Real time abandoned bag detection using OpenCV
Sehairi et al. A Real-Time Implementation of Moving Object Action Recognition System Based on Motion Analysis
Alexandropoulos et al. Real-time change detection for surveillance in public transportation
WO2020139071A1 (en) System and method for detecting aggressive behaviour activity

Legal Events

Date Code Title Description
AS Assignment

Owner name: VATICS INC., TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHANG, CHIH-HAO;YANG, ZHONG-LAN;REEL/FRAME:022879/0910

Effective date: 20090521

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION