US9082019B2 - Method of establishing adjustable-block background model for detecting real-time image object - Google Patents

Method of establishing adjustable-block background model for detecting real-time image object Download PDF

Info

Publication number
US9082019B2
US9082019B2 US13/918,513 US201313918513A US9082019B2 US 9082019 B2 US9082019 B2 US 9082019B2 US 201313918513 A US201313918513 A US 201313918513A US 9082019 B2 US9082019 B2 US 9082019B2
Authority
US
United States
Prior art keywords
color
block
pixel
image
blocks
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/918,513
Other versions
US20140369552A1 (en
Inventor
Ming-Hwa Sheu
Wen-Kai Tsai
Jian-Hui Chen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
National Yunlin University of Science and Technology
Original Assignee
National Yunlin University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by National Yunlin University of Science and Technology filed Critical National Yunlin University of Science and Technology
Priority to US13/918,513 priority Critical patent/US9082019B2/en
Assigned to NATIONAL YUNLIN UNIVERSITY OF SCIENCE AND TECHNOLOGY reassignment NATIONAL YUNLIN UNIVERSITY OF SCIENCE AND TECHNOLOGY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SHEU, MING-HWA, CHEN, Jian-hui, TSAI, WEN-KAI
Publication of US20140369552A1 publication Critical patent/US20140369552A1/en
Application granted granted Critical
Publication of US9082019B2 publication Critical patent/US9082019B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06K9/00771
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/457Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by analysing connectivity, e.g. edge linking, connected component analysis or slices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20224Image subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Definitions

  • the present invention relates to an image surveillance technique, and particularly to a method of establishing an adjustable-block background model for detecting a real-time image object.
  • a surveillance system essentially involves a moving object detection technique, which is intended to allow the system to more accurately and rapidly detect a moving object. Subsequent operations such as tracking, identification and analysis procedures on the moving object are highly dependent on the accuracy of moving object detection. The accuracy of moving object detection may even be decisive on the reliability and accuracy of the entire system, and is thus a crucial factor for evaluating the quality of a surveillance system.
  • optical flow method a motion vector in an image in consecutive frames is identified and characteristic matching is then performed, and is applicable to motion detection and moving object segmentation.
  • the optical flow method is extremely effective in applications including pattern recognition and computer vision as well as other image processing applications, the optical flow method falls short in providing a real-time effect due to a high sensitivity on noises in a scene and a huge computation amount resulted by complex algorithm.
  • every two pixel values in consecutive frames are subtracted to obtain a difference.
  • the pixel is considered as the background when the difference is smaller than a threshold, or the pixel is considered as a part of a moving object when a large variance exists in the pixel and the difference is greater than the threshold.
  • This method being quite simple and fast, however frequently obtains only borders of a moving object rather than a complete object.
  • a background model is first established.
  • the background model is compared with a new image to further obtain a moving object.
  • This method is both simple and fast.
  • the concept of the background subtraction method may seem easy, the background subtraction method suffers from many challenging issues. For example, results of segmentation are liable to errors including fragments, high noises and segmentation failures.
  • updating and display capabilities of the background need to be reinforced, which may on the other hand lead to a tremendous load on the system memory and computation performance.
  • the primary object of the present invention is to satisfy requirements of a high background display capability, a low memory capacity and a low system computation load for the background subtraction method.
  • a method of establishing an adjustable-block background model for detecting a real-time image object obtains a surveillance image by a surveillance apparatus, and comprises the following steps.
  • S 1 The surveillance image is segmented into a plurality of blocks, each of which comprising a first pixel and at least one second pixel.
  • a color of the first pixel is selected as a major color, and a difference is determined by comparing the color of the at least one second pixel and the color of the first pixel by an operation processing unit. The color of the at least one second pixel is added to the major color when the difference is greater than a threshold.
  • One of the blocks is defined as an initial block, and the initial block and the blocks adjacent to the initial block are jointly defined as a search region.
  • the number of the major color of the initial block is smaller than a block threshold, the number and color information of the major color in the search region are reorganized and integrated by the operation processing unit.
  • S 7 A new surveillance image is inputted from the surveillance apparatus to perform color comparison between the new surveillance image and the block background model, and a moving image object is determined and obtained when a difference of the color comparison is too large.
  • the present invention offers the advantages below.
  • the major color is utilized as a basis for mutual comparison to reduce system memory capacity load.
  • different-sized block background models are established to effectively further reduce the utilization amount of the memory.
  • the computer is allowed to more efficiently perform data computation to enhance computation performance.
  • FIG. 1 is a flowchart on the process of the present invention.
  • FIG. 2 is a system block diagram of the present invention.
  • FIG. 3 is a schematic diagram of a block in a surveillance image of the present invention.
  • FIG. 4 is a schematic diagram of blocks of an image in a surveillance image of the present invention.
  • FIG. 5A is a schematic diagram showing number of a major color in blocks of the present invention.
  • FIG. 5B is a schematic diagram of merging blocks of the present invention.
  • FIG. 5C is a schematic diagram of a block background model of the present invention.
  • FIG. 6A is a schematic diagram of a border of a moving object image of the present invention.
  • FIG. 6B is a schematic diagram of updating a moving object image of the present invention.
  • a method of establishing an adjustable-block background model for detecting a real-time image object which obtains a surveillance image 11 by a surveillance apparatus 10 and comprises the following steps.
  • each of the blocks 12 includes a first pixel 121 and at least one second pixel 122 .
  • each block 12 may include 16 pixels to form a matrix in 4 ⁇ 4 dimensions. One of the 16 pixels may be selected as the first pixel 121 , while the remaining pixels may be the second pixels 122 .
  • a major color in the blocks 12 is selected and calculated.
  • the color of the first pixel 121 is selected as a major color, and the colors of the at least one second pixel 122 and the first pixel 121 are compared to determine a difference between them by an operation processing unit 20 .
  • the color of the at least one second pixel 122 is added into the major color when the difference is greater than a threshold.
  • the difference is obtained according to a square root of a sum of squares of the three primary colors, i.e., red, blue and green, and can be represented as equation (1) below:
  • R Mn , G Mn and B Mn respectively represent color values of the red, green and blue primary colors in the major color
  • R Ixy , G Ixy and B Ixy respectively represent color values of the red, green and blue primary colors in the at least one second pixel 122 , so as to obtain the difference as a reference index of a color difference.
  • the difference may be obtained according to a sum of absolute values of the color differences in the red, green and blue primary colors.
  • equation (2) a sum of absolute values of color differences is obtained, and can be calculated with simpler mathematical calculation.
  • step S 3 the numbers and color information of the major color in the blocks 12 are integrated and stored by the operation processing unit 20 .
  • the color information represents information associated with colors of the major color, e.g., red, blue and green, etc.
  • a threshold of each block 12 is determined.
  • a region having a matrix dimension of 32 ⁇ 32 pixels is selected for illustration and analysis. In the present invention, every 16 pixels are defined as one block 12 .
  • the region selected in FIG. 4 includes a matrix formed by 8 ⁇ 8 blocks 12 .
  • one of the blocks 12 is selected and is defined as an initial block 13 , and the initial block 13 and the blocks 12 adjacent to the initial block 13 are jointly defined as a search region 40 .
  • the block 12 at the upper-left corner is selected as the initial block 13
  • the initial block 13 and the three blocks 12 adjacent to the initial block 13 are jointly defined as the search region 40 .
  • Every four blocks 12 are similarly defined as one search region 40 to altogether form 16 search regions 40 .
  • the operation processing unit 20 reorganizes and integrates the number and color information of the major color in the search region 40 , In FIG. 5A , the number in each of the blocks 12 represents the number of the major color therein. More specifically, in the reorganization and integration performed by the operation processing unit 20 , the operation processing unit 20 obtains the number and color information of the major color in the search region 40 by again performing the determination process in steps S 2 and S 3 as the search region 40 is a new unit.
  • the color of the major color of each of the search regions 40 may differ from one another.
  • the number of the major color in the search region 40 may also change after the reorganization and integration performed by the operation processing unit 20 , and so the reorganization and integration are required.
  • the comparison and determination process for a next search region 40 is directly performed. That is to say, when the number of the major color in the initial block 13 is large enough, it means that the color difference in the initial block 13 is quite high. As such, no reorganization and integration are needed to be performed, and so the comparison and determination for the next search region 40 can be directly performed.
  • step S 5 the blocks 12 are merged.
  • the search region 40 when the number of the major color in the search region 40 is smaller than a region threshold, the search region 40 is merged into a large block 14 .
  • the numbers of the major color in the upper-row search regions 40 are both 1, indicating that the color differences in pixels in these regions are not large.
  • the search regions 40 are merged into one large block 14 to reduce a load on the memory.
  • the search region 40 is not merged, and the comparison and determination for a next search region 40 is directly performed.
  • step S 6 the analysis on the surveillance image 11 is completed, and steps S 4 and S 5 are sequentially repeated to obtain a block background model 30 of the surveillance image 11 .
  • the numbers of the major color in the four lower-left regions 40 are also 1, and so, in addition to respectively merging the four regions into large blocks 14 , the four large blocks 14 are merged into a mega block 15 .
  • FIG. 5C shows a final block background model 30 .
  • step S 7 detection on a moving object image is performed.
  • the surveillance apparatus 10 inputs a new surveillance image to perform color comparison between the new surveillance image and the block background model 30 .
  • a difference of the color comparison is large, a moving object image is determined and obtained.
  • the operation processing unit 20 compares the image corresponding to the newly inputted surveillance image.
  • the new surveillance image matches the block background model 30 , the new surveillance image is determined as the background, which is then displayed in black after being converted and outputted to surveillance personnel.
  • the color comparison result of the new surveillance image indicates a large color difference in the colors between the new surveillance image and the block background model 30 , it is determined that the new surveillance image is a foreground, which is then displayed in white or in colors.
  • step S 8 the moving object image is updated.
  • a pixel in each of the blocks 12 in the large block 14 is extracted, and is compared with a corresponding position of the pixel in the new surveillance image. When the comparison results do not match, data in the blocks 12 is updated.
  • the blocks 12 are updated in a unit of the blocks 12 of the moving object image, patterns in the foreground displayed (the moving object image) are also updated in a unit of blocks.
  • the large block 14 it is frequent that image patterns are incorrectly or incompletely segmented to cause an incorrect actual image displayed in the block 12 . Therefore, in this step, one pixel in each of the blocks 12 in the large block 14 is randomly extracted for comparison.
  • the comparison process for a next block 12 is performed.
  • the color information of the block 12 is updated according to the new surveillance image.
  • the large block 14 can be included, and is verified and updated by an omitted moving object image, thereby reinforcing the capability of correctly displaying the moving object image.
  • step S 9 a refinement treatment is performed on the moving object image.
  • white blocks 51 in FIG. 6A represent pixels of the moving object image
  • shaded blocks 52 represent background pixels.
  • border regions 53 of the moving object image are compared with corresponding positions in the new surveillance image, and the pixels are updated for the refinement treatment.
  • the Codebook method occupies 1.11 MB in memory, whereas the present invention occupies only 0.13 MB in memory, which is only about 1/10 of the Codebook method.
  • the Codebook method occupies 17.7 MB in memory, whereas the present invention occupies only 0.3 MB in memory, which is far reduced to only 1/60 of the Codebook method.
  • the major color is utilized as a basis for mutual comparison to reduce the system memory capacity load.
  • different-sized block background models are established to effectively further reduce the utilization amount of the memory.
  • the computer is allowed to more efficiently perform data computation to enhance computation performance.
  • the pixels in the large block are randomly extracted and compared to update the moving object image, such that erroneous images resulted by incorrect or incomplete image segmentation are mitigated.
  • the image display capability is reinforced by the refinement treatment to enhance image recognition.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

A method of establishing an adjustable-block background model for detecting a real-time image object is provided to obtain a surveillance image by a surveillance apparatus. The surveillance image has a plurality of pixels. The method includes steps of: segmenting the surveillance image into a plurality of blocks each having a first pixel and at least one second pixel; defining the first pixel as a major color and comparing the first pixel with the at least one second pixel to determine a number and color information of the major color in the block; merging the blocks having the same major color into a large block to obtain a block background model; and performing image comparison to identify a moving object image. With the establishment of the block background model, a required memory space is effectively reduced while outstanding image display performance is still maintained.

Description

FIELD OF THE INVENTION
The present invention relates to an image surveillance technique, and particularly to a method of establishing an adjustable-block background model for detecting a real-time image object.
BACKGROUND OF THE INVENTION
A surveillance system essentially involves a moving object detection technique, which is intended to allow the system to more accurately and rapidly detect a moving object. Subsequent operations such as tracking, identification and analysis procedures on the moving object are highly dependent on the accuracy of moving object detection. The accuracy of moving object detection may even be decisive on the reliability and accuracy of the entire system, and is thus a crucial factor for evaluating the quality of a surveillance system.
Numerous publications on the object detection realm have been disclosed. Among the researches, three methods, namely an optical flow method, a frame difference method and a background subtraction method, are most prevalent. In the optical flow method, a motion vector in an image in consecutive frames is identified and characteristic matching is then performed, and is applicable to motion detection and moving object segmentation. Although the optical flow method is extremely effective in applications including pattern recognition and computer vision as well as other image processing applications, the optical flow method falls short in providing a real-time effect due to a high sensitivity on noises in a scene and a huge computation amount resulted by complex algorithm.
In the frame subtraction method, every two pixel values in consecutive frames are subtracted to obtain a difference. The pixel is considered as the background when the difference is smaller than a threshold, or the pixel is considered as a part of a moving object when a large variance exists in the pixel and the difference is greater than the threshold. This method, being quite simple and fast, however frequently obtains only borders of a moving object rather than a complete object.
In the background subtraction method, a background model is first established. The background model is compared with a new image to further obtain a moving object. This method is both simple and fast. Although the concept of the background subtraction method may seem easy, the background subtraction method suffers from many challenging issues. For example, results of segmentation are liable to errors including fragments, high noises and segmentation failures. To prevent the above errors, updating and display capabilities of the background need to be reinforced, which may on the other hand lead to a tremendous load on the system memory and computation performance.
SUMMARY OF THE INVENTION
Therefore the primary object of the present invention is to satisfy requirements of a high background display capability, a low memory capacity and a low system computation load for the background subtraction method.
To achieve the above object, a method of establishing an adjustable-block background model for detecting a real-time image object is provided. It obtains a surveillance image by a surveillance apparatus, and comprises the following steps.
S1: The surveillance image is segmented into a plurality of blocks, each of which comprising a first pixel and at least one second pixel.
S2: A color of the first pixel is selected as a major color, and a difference is determined by comparing the color of the at least one second pixel and the color of the first pixel by an operation processing unit. The color of the at least one second pixel is added to the major color when the difference is greater than a threshold.
S3: The number and color information of the major color of the blocks are determined by the operation processing unit, and stored.
S4: One of the blocks is defined as an initial block, and the initial block and the blocks adjacent to the initial block are jointly defined as a search region. When the number of the major color of the initial block is smaller than a block threshold, the number and color information of the major color in the search region are reorganized and integrated by the operation processing unit.
S5: When the number of major color in the search region is smaller than a region threshold, the search region is merged into a large block.
S6: The analysis on the surveillance image is completed, and steps S4 and S5 are sequentially repeated to obtain a block background model of the surveillance image.
S7: A new surveillance image is inputted from the surveillance apparatus to perform color comparison between the new surveillance image and the block background model, and a moving image object is determined and obtained when a difference of the color comparison is too large.
As seen from the above descriptions, the present invention offers the advantages below. First of all, the major color is utilized as a basis for mutual comparison to reduce system memory capacity load. Secondly, with the determination mechanism of the block threshold and the region threshold, different-sized block background models are established to effectively further reduce the utilization amount of the memory. Moreover, as the load on the memory system is reduced, the computer is allowed to more efficiently perform data computation to enhance computation performance.
The foregoing, as well as additional objects, features and advantages of the invention will be more readily apparent from the following detailed description, which proceeds with reference to the accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
FIG. 1 is a flowchart on the process of the present invention.
FIG. 2 is a system block diagram of the present invention.
FIG. 3 is a schematic diagram of a block in a surveillance image of the present invention.
FIG. 4 is a schematic diagram of blocks of an image in a surveillance image of the present invention.
FIG. 5A is a schematic diagram showing number of a major color in blocks of the present invention.
FIG. 5B is a schematic diagram of merging blocks of the present invention.
FIG. 5C is a schematic diagram of a block background model of the present invention.
FIG. 6A is a schematic diagram of a border of a moving object image of the present invention.
FIG. 6B is a schematic diagram of updating a moving object image of the present invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Referring to FIGS. 1, 2 and 3, a method of establishing an adjustable-block background model for detecting a real-time image object is provided, which obtains a surveillance image 11 by a surveillance apparatus 10 and comprises the following steps.
In step S1, the surveillance image 11 is segmented into a plurality of blocks 12. Referring to FIG. 3, each of the blocks 12 includes a first pixel 121 and at least one second pixel 122. In the embodiment, each block 12 may include 16 pixels to form a matrix in 4×4 dimensions. One of the 16 pixels may be selected as the first pixel 121, while the remaining pixels may be the second pixels 122.
In step S2, a major color in the blocks 12 is selected and calculated. The color of the first pixel 121 is selected as a major color, and the colors of the at least one second pixel 122 and the first pixel 121 are compared to determine a difference between them by an operation processing unit 20. The color of the at least one second pixel 122 is added into the major color when the difference is greater than a threshold. For example, the difference is obtained according to a square root of a sum of squares of the three primary colors, i.e., red, blue and green, and can be represented as equation (1) below:
distance ( M n , I x , y ) = ( M n - I x , y ) 2 = ( R M n - R I x , y ) 2 + ( G M n - G I x , y ) 2 + ( B M n - B I x , y ) 2 ( 1 )
In equation (1), RMn, GMn and BMn respectively represent color values of the red, green and blue primary colors in the major color, and RIxy, GIxy and BIxy respectively represent color values of the red, green and blue primary colors in the at least one second pixel 122, so as to obtain the difference as a reference index of a color difference. Alternatively, the difference may be obtained according to a sum of absolute values of the color differences in the red, green and blue primary colors. By comparing the amount of the difference, equation (2) is obtained:
distance ( M n , I x , y ) = ( M n - I x , y ) = R M n - R I x , y + G M n - G I x , y + B M n - B I x , y ( 2 )
In equation (2), a sum of absolute values of color differences is obtained, and can be calculated with simpler mathematical calculation.
In step S3, the numbers and color information of the major color in the blocks 12 are integrated and stored by the operation processing unit 20. The color information represents information associated with colors of the major color, e.g., red, blue and green, etc.
In step S4, a threshold of each block 12 is determined. Referring to FIGS. 4 and 5A, a region having a matrix dimension of 32×32 pixels is selected for illustration and analysis. In the present invention, every 16 pixels are defined as one block 12. Thus, the region selected in FIG. 4 includes a matrix formed by 8×8 blocks 12. Referring to FIG. 5A, one of the blocks 12 is selected and is defined as an initial block 13, and the initial block 13 and the blocks 12 adjacent to the initial block 13 are jointly defined as a search region 40. In the embodiment, the block 12 at the upper-left corner is selected as the initial block 13, and the initial block 13 and the three blocks 12 adjacent to the initial block 13 are jointly defined as the search region 40. Every four blocks 12 are similarly defined as one search region 40 to altogether form 16 search regions 40. When the number of the major color in the initial block 13 is smaller than a block threshold, the operation processing unit 20 reorganizes and integrates the number and color information of the major color in the search region 40, In FIG. 5A, the number in each of the blocks 12 represents the number of the major color therein. More specifically, in the reorganization and integration performed by the operation processing unit 20, the operation processing unit 20 obtains the number and color information of the major color in the search region 40 by again performing the determination process in steps S2 and S3 as the search region 40 is a new unit. The color of the major color of each of the search regions 40 may differ from one another. Therefore, the number of the major color in the search region 40 may also change after the reorganization and integration performed by the operation processing unit 20, and so the reorganization and integration are required. When the number of the major color in the initial block 13 is greater than the block threshold, the comparison and determination process for a next search region 40 is directly performed. That is to say, when the number of the major color in the initial block 13 is large enough, it means that the color difference in the initial block 13 is quite high. As such, no reorganization and integration are needed to be performed, and so the comparison and determination for the next search region 40 can be directly performed.
In step S5, the blocks 12 are merged. Referring to FIG. 5B, when the number of the major color in the search region 40 is smaller than a region threshold, the search region 40 is merged into a large block 14. As shown in FIG. 5B, the numbers of the major color in the upper-row search regions 40 are both 1, indicating that the color differences in pixels in these regions are not large. Thus, the search regions 40 are merged into one large block 14 to reduce a load on the memory. When the number of the major color in the search region 40 is greater than the region threshold, the search region 40 is not merged, and the comparison and determination for a next search region 40 is directly performed.
In step S6, the analysis on the surveillance image 11 is completed, and steps S4 and S5 are sequentially repeated to obtain a block background model 30 of the surveillance image 11. The numbers of the major color in the four lower-left regions 40 are also 1, and so, in addition to respectively merging the four regions into large blocks 14, the four large blocks 14 are merged into a mega block 15. FIG. 5C shows a final block background model 30.
In step S7, detection on a moving object image is performed. Referring to FIG. 2, the surveillance apparatus 10 inputs a new surveillance image to perform color comparison between the new surveillance image and the block background model 30. When a difference of the color comparison is large, a moving object image is determined and obtained. More specifically, according to the data of the block background model 30, the operation processing unit 20 compares the image corresponding to the newly inputted surveillance image. When the color comparison result of the new surveillance image matches the block background model 30, the new surveillance image is determined as the background, which is then displayed in black after being converted and outputted to surveillance personnel. When the color comparison result of the new surveillance image indicates a large color difference in the colors between the new surveillance image and the block background model 30, it is determined that the new surveillance image is a foreground, which is then displayed in white or in colors.
In step S8, the moving object image is updated. A pixel in each of the blocks 12 in the large block 14 is extracted, and is compared with a corresponding position of the pixel in the new surveillance image. When the comparison results do not match, data in the blocks 12 is updated. As the blocks 12 are updated in a unit of the blocks 12 of the moving object image, patterns in the foreground displayed (the moving object image) are also updated in a unit of blocks. Thus, in the large block 14, it is frequent that image patterns are incorrectly or incompletely segmented to cause an incorrect actual image displayed in the block 12. Therefore, in this step, one pixel in each of the blocks 12 in the large block 14 is randomly extracted for comparison. When the comparison result matches the new surveillance image, the comparison process for a next block 12 is performed. When the comparison results do not match, the color information of the block 12 is updated according to the new surveillance image. As such, the large block 14 can be included, and is verified and updated by an omitted moving object image, thereby reinforcing the capability of correctly displaying the moving object image.
In step S9, a refinement treatment is performed on the moving object image. Referring to FIGS. 6A and 6B, white blocks 51 in FIG. 6A represent pixels of the moving object image, and shaded blocks 52 represent background pixels. As shown in FIG. 6B, in the refinement treatment, border regions 53 of the moving object image are compared with corresponding positions in the new surveillance image, and the pixels are updated for the refinement treatment.
It is discovered from comparisons on experimental results that, compared to “Real-Time Foreground-Background Segmentation using Codebook Model” (to he referred to as the Codebook method) disclosed in Real-Time Imaging, pp. 172-185, 2005 by Kyungnam Kim et al., for 160×120 pixels, 91% similarity to an original image is achieved by utilizing the Codebook method, whereas 92% similarity is achieved by the method disclosed by the present invention. For 768×576 pixels, 63% similarity to an original image is achieved by utilizing the Codebook method, whereas 69.7% similarity is achieved by the method disclosed by the present invention. Therefore, with respect to image analysis capabilities, the method disclosed by the present invention is preferred over the Codebook method of the prior art.
By comparing the memory utilization rate, for 160×120 pixels, the Codebook method occupies 1.11 MB in memory, whereas the present invention occupies only 0.13 MB in memory, which is only about 1/10 of the Codebook method. For 768×576 pixels, the Codebook method occupies 17.7 MB in memory, whereas the present invention occupies only 0.3 MB in memory, which is far reduced to only 1/60 of the Codebook method.
In conclusion, the present invention offers the advantages below.
First of all, the major color is utilized as a basis for mutual comparison to reduce the system memory capacity load. Secondly, with the determination mechanism of the block threshold and the region threshold, different-sized block background models are established to effectively further reduce the utilization amount of the memory. Thirdly, as the load on the memory system is reduced, the computer is allowed to more efficiently perform data computation to enhance computation performance. Further, the pixels in the large block are randomly extracted and compared to update the moving object image, such that erroneous images resulted by incorrect or incomplete image segmentation are mitigated. Moreover, the image display capability is reinforced by the refinement treatment to enhance image recognition.

Claims (6)

What is claimed is:
1. A method of establishing an adjustable-block background model for detecting a real-time image object, which obtains a surveillance image by a surveillance apparatus, the method comprising steps of:
S1) segmenting the surveillance image into a plurality of blocks, each of which comprising a first pixel and at least one second pixel;
S2) selecting a color of the first pixel as a major color, and determining a difference between color of the at least one second pixel and the color of the first pixel by an operation processing unit, and adding the color of the at least one second pixel into the major color when the difference is greater than a threshold;
S3) determining a number and color information of the major color in each of the blocks by the operation processing unit, and storing the determined number and color information;
S4) selecting and defining one of the blocks as an initial block, and defining the initial block and the blocks adjacent to the initial block as a search region; when the number of the major color in the initial block is smaller than a block threshold, reorganizing and integrating the number and the color information of the major color in the search region by the operation processing unit;
S5) when the number of the major color in the search region is smaller than a region threshold, merging the search region into a large block;
S6) completing analysis on the surveillance image, and sequentially repeating the steps S4 and S5 to obtain a block background model of the surveillance image;
S7) inputting a new surveillance image by the surveillance apparatus, and performing color comparison between the new surveillance image and the block background model; and
obtaining a moving object image when a color comparison result indicates a large difference; and
S8) extracting a pixel in each of the blocks in the large block and comparing with a corresponding position in the new surveillance image; and updating data of the blocks when a pixel comparison result does not match.
2. The method of claim 1, wherein the pixel in each of the blocks of the large block is randomly extracted and compared; a comparison process for a next block is performed when a comparison result matches the new surveillance image, or the color information of the block is updated when the comparison result does not match.
3. The method of claim 1, wherein in the step S4, one search region includes four blocks which are sequentially determined with the block threshold.
4. The method of claim 1, wherein in the step S2, the difference is obtained according to a square root of a sum of squares of red, green and blue primary colors.
5. The method of claim 1, wherein in step S2, the difference is obtained according to a square root of a sum of squares of color differences between red, green and blue primary colors.
6. The method of claim 1, after the step S7, further comprising:
S9) performing a refinement treatment on the moving object image, and comparing a border region of the moving object image with a corresponding position in the new surveillance image to complete the refinement treatment.
US13/918,513 2013-06-14 2013-06-14 Method of establishing adjustable-block background model for detecting real-time image object Expired - Fee Related US9082019B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/918,513 US9082019B2 (en) 2013-06-14 2013-06-14 Method of establishing adjustable-block background model for detecting real-time image object

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/918,513 US9082019B2 (en) 2013-06-14 2013-06-14 Method of establishing adjustable-block background model for detecting real-time image object

Publications (2)

Publication Number Publication Date
US20140369552A1 US20140369552A1 (en) 2014-12-18
US9082019B2 true US9082019B2 (en) 2015-07-14

Family

ID=52019253

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/918,513 Expired - Fee Related US9082019B2 (en) 2013-06-14 2013-06-14 Method of establishing adjustable-block background model for detecting real-time image object

Country Status (1)

Country Link
US (1) US9082019B2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106651860A (en) * 2016-11-02 2017-05-10 帝麦克斯(苏州)医疗科技有限公司 Image segmenting method and apparatus
US20180075605A1 (en) * 2016-09-13 2018-03-15 Intelligent Fusion Technology, Inc Method and system for detecting multiple moving objects from real-time aerial images
US10249047B2 (en) 2016-09-13 2019-04-02 Intelligent Fusion Technology, Inc. System and method for detecting and tracking multiple moving targets based on wide-area motion imagery

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017106846A2 (en) * 2015-12-18 2017-06-22 Iris Automation, Inc. Real-time visual situational awareness system
CN105488542B (en) * 2015-12-24 2019-04-23 福建星网物联信息系统有限公司 A kind of foreground object detection method and equipment
US10121093B2 (en) * 2017-04-11 2018-11-06 Sony Corporation System and method for background subtraction in video content
CN108648210B (en) * 2018-05-09 2022-06-14 西安天和防务技术股份有限公司 Rapid multi-target detection method and device under static complex scene

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Wen-kai Tsai; Ming-hwa Sheu; Chung-chi Lin, "Block-Based Major Color Method for Foreground Object Detection on Embedded SoC Platforms," Embedded Systems Letters, IEEE , vol. 4, No. 2, pp. 49,52, May 7, 2012. *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180075605A1 (en) * 2016-09-13 2018-03-15 Intelligent Fusion Technology, Inc Method and system for detecting multiple moving objects from real-time aerial images
US9940724B2 (en) * 2016-09-13 2018-04-10 Intelligent Fusion Technology, Inc. Method and system for detecting multiple moving objects from real-time aerial images
US10249047B2 (en) 2016-09-13 2019-04-02 Intelligent Fusion Technology, Inc. System and method for detecting and tracking multiple moving targets based on wide-area motion imagery
CN106651860A (en) * 2016-11-02 2017-05-10 帝麦克斯(苏州)医疗科技有限公司 Image segmenting method and apparatus
CN106651860B (en) * 2016-11-02 2019-07-09 帝麦克斯(苏州)医疗科技有限公司 A kind of image partition method and device

Also Published As

Publication number Publication date
US20140369552A1 (en) 2014-12-18

Similar Documents

Publication Publication Date Title
US9082019B2 (en) Method of establishing adjustable-block background model for detecting real-time image object
EP3540637B1 (en) Neural network model training method, device and storage medium for image processing
US9430704B2 (en) Image processing system with layout analysis and method of operation thereof
US20120121166A1 (en) Method and apparatus for three dimensional parallel object segmentation
US20060029276A1 (en) Object image detecting apparatus, face image detecting program and face image detecting method
US20140169667A1 (en) Removing an object from an image
EP3073443B1 (en) 3d saliency map
EP3340075B1 (en) Video abstract using signed foreground extraction and fusion
CN111860414A (en) Method for detecting Deepfake video based on multi-feature fusion
CN110390327B (en) Foreground extraction method and device, computer equipment and storage medium
CN105095835A (en) Pedestrian detection method and system
CN110598795A (en) Image difference detection method and device, storage medium and terminal
CN111242026A (en) Remote sensing image target detection method based on spatial hierarchy perception module and metric learning
CN110910445A (en) Object size detection method and device, detection equipment and storage medium
Recky et al. Window detection in complex facades
US9196025B2 (en) Image processing apparatus, image processing method and image processing program
CN110909568A (en) Image detection method, apparatus, electronic device, and medium for face recognition
CN112396594A (en) Change detection model acquisition method and device, change detection method, computer device and readable storage medium
CN111667419A (en) Moving target ghost eliminating method and system based on Vibe algorithm
CN114639013A (en) Remote sensing image airplane target detection and identification method based on improved Orient RCNN model
CN110490877B (en) Target segmentation method for binocular stereo image based on Graph Cuts
US10839252B2 (en) System and method for color invariant feature matching
JP2017027363A (en) Video processing device, video processing method and program
CN113516609A (en) Split screen video detection method and device, computer equipment and storage medium
KR20100009451A (en) Method for determining ground line

Legal Events

Date Code Title Description
AS Assignment

Owner name: NATIONAL YUNLIN UNIVERSITY OF SCIENCE AND TECHNOLO

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SHEU, MING-HWA;TSAI, WEN-KAI;CHEN, JIAN-HUI;SIGNING DATES FROM 20130222 TO 20130228;REEL/FRAME:030637/0171

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20190714