USRE42367E1 - Method for illumination independent change detection in a pair of registered gray images - Google Patents
Method for illumination independent change detection in a pair of registered gray images Download PDFInfo
- Publication number
- USRE42367E1 USRE42367E1 US12/188,654 US18865408A USRE42367E US RE42367 E1 USRE42367 E1 US RE42367E1 US 18865408 A US18865408 A US 18865408A US RE42367 E USRE42367 E US RE42367E
- Authority
- US
- United States
- Prior art keywords
- blobs
- blob
- image
- images
- gray
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime, expires
Links
- 230000008859 change Effects 0.000 title claims abstract description 121
- 238000000034 method Methods 0.000 title claims abstract description 110
- 238000001514 detection method Methods 0.000 title claims abstract description 62
- 238000005286 illumination Methods 0.000 title claims abstract description 30
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 54
- 238000004458 analytical method Methods 0.000 claims abstract description 15
- 238000009826 distribution Methods 0.000 claims abstract description 15
- 238000012360 testing method Methods 0.000 claims description 18
- 238000000605 extraction Methods 0.000 abstract description 22
- 230000008569 process Effects 0.000 description 11
- 238000012545 processing Methods 0.000 description 7
- 230000000694 effects Effects 0.000 description 5
- 238000013459 approach Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000010276 construction Methods 0.000 description 2
- 238000009472 formulation Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000007619 statistical method Methods 0.000 description 2
- 238000013179 statistical model Methods 0.000 description 2
- 240000005475 Abelmoschus moschatus Species 0.000 description 1
- 101100168642 Arabidopsis thaliana CRN gene Proteins 0.000 description 1
- 101100045632 Arabidopsis thaliana TCX2 gene Proteins 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 101150103732 sol2 gene Proteins 0.000 description 1
- 238000000528 statistical test Methods 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 238000010998 test method Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/254—Analysis of motion involving subtraction of images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/28—Quantising the image, e.g. histogram thresholding for discrimination between background and foreground patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10152—Varying illumination
Definitions
- blob extraction includes extraction of parts or the entire object.
- object extraction is also covered by the term “blob extraction” as used herein.
- FIG. 1 shows a pair of images used as inputs to the change detection method. Each image contains a blob that is not contained in the other (e.g. a vehicle 102 in (a) and 104 in (b)). Notice that the images contain the same scenery, under different illumination.
- the known techniques for change detection can be classified into the following categories:
- Pixel-level change detection 1.1 Pixel-level change detection. 1.2 Surface modeling. 1.3 Comparison among derivative images. 1.4 Contrast invariant representation. 1.5 Region based comparison of first or higher order statistics.
- Change between two images or frames can be detected by comparing the differences in intensity values of corresponding pixels in the two frames.
- An algorithm counts the number of the changed pixels, and a camera break is declared if the percentage of the total number of pixels changed exceeds a certain threshold [R. Kasturi and R. Jain, “Dynamic Vision”, Computer Vision: Principles, Eds. R. Kasturi, R. Jain, IEEE Computer Society Press, Washington, pp. 469-480, 1991 (hereinafter KAS91); A. Nagasaka, and Y. Tanaka, “Automatic Video Indexing and Full-Video Search for Blob Appearances”, Visual Database Systems, II, Eds. E. Knuth, and L. M. Wegner, Elsevier Science Publishers B.
- DP i ⁇ ( x , y ) ⁇ 1 0 ⁇ ⁇ if ⁇ ⁇ ⁇ F i ⁇ ( x , y ) - F i + 1 ⁇ ( x , y ) ⁇ ⁇ > t ( 1 )
- F i (x,y) is the intensity value of the pixel in frame i at the coordinates (x,y). If the difference between the corresponding pixels in the two consecutive frames is above a certain minimum intensity value, then DP i (x,y), the difference picture, is set to one. In Eq. 2, the percentage difference between the pixels in the two frames is calculated by summing the difference picture and dividing by the total number of pixels in a frame. If this percentage is above a certain threshold T, a camera break is declared.
- Camera movement e.g., pan or zoom
- Fast moving blobs also have the same effect. If the mean intensity values of the pixels and their connected pixels are compared [ZHA93], then the effects of the camera and blob motion are reduced.
- Hsu Y. Z., Nagel H. H, and Rekers G., “New likelihood test methods for change detection in image sequences”, Computer Vision Graphics Image Processing, vol. 26, pp. 73-106, 1984 (hereinafter HSU84)] model the gray-level surface by patches of a second order bivariate polynomial in the pixel coordinates.
- R 1 (x 0 ,y 0 ) in the image I (1) and R 2 (x 0 ,y 0 ) in I (2) represent each region by a set of seven parameters—the six coefficients of the quadratic polynomial patch, and the sum of square differences between the polynomial patch and the gray-levels.
- the approximating patch represents the gray-level surface up to uncorrected noise errors.
- R 1 (x 0 ,y 0 ) and R 2 (x 0 ,y 0 ) come from the same gray-value distribution.
- R 1 (x 0 ,y 0 ) and R 2 (x 0 ,y 0 ) come from different gray-value distributions.
- the input images are divided into regions, usually squares of m ⁇ m pixels.
- R 1 (x 0 ,y 0 ) the square in the image I (1) that its center is the pixel with coordinates (x 0 ,y 0 ), and similarly denote by R 2 (x 0 ,y 0 ) the corresponding square in the image I (2) .
- the gray-levels in the region R 1 (x 0 ,y 0 ) are normalized such that the mean gray-level and the variance of the gray-levels of R 1 (x 0 ,y 0 ) are the same as the mean and variance gray-level of R 2 (x 0 ,y 0 ).
- the image I (2) is compare to the image I (1) .
- the normalization process of this statistical method is supposed to be a variation of illumination correction.
- Each gray-level is basically the product of two components: (1) the amount of source light incident on the scene and (2) the amount of light reflected by the blobs in the scene.
- the amount of source light incident on a small region of the scene is approximately uniform, but the reflected light of two adjacent blobs may be different.
- i(x 0 ,y 0 ) the amount of source light incident on point (x 0 ,y 0 ) in the scene
- r(x 0 ,y 0 ) the amount of reflected light from the point (x 0 ,y 0 ) in the scene.
- I (1) and I (2) be two images with corresponding functions, i 1 (x,y), i 2 (x,y), r 1 (x,y) and r 2 (x,y). If at pixel (x 0 ,y 0 ) both images contain the same blob, then the following is satisfied:
- F Def I ( 1 ) I ( 2 ) ( 4 ) where F is assumed to have accuracy of real numbers.
- R F (x 0 ,y 0 ) be a small neighborhood around the point (x 0 ,y 0 ) in the image F.
- the surface patch that is composed of the values in the region R F (x 0 ,y 0 ) is expected to be a smooth and slow varying surface, since the change of the illumination in a small region is slow-varying.
- the surface patch that is composed of the values in R F (x 0 ,y 0 ) is expected to be much less smooth, since the region R F (x 0 ,y 0 ) can include a transition from one blob to another.
- the method in [SKI89] proposes to examine the variance in each pixel (x 0 ,y 0 ) of the region R F (x 0 ,y 0 ). If the variance is higher than some pre-specified threshold, then the pixel is considered as belonging to a region of change.
- the change detection mask of this method is defined for each pixel by the following formula:
- ⁇ i is the average value of the ratio of intensities
- E is the expectation
- N is the size of the image
- “ ⁇ ” is a 5 ⁇ 5 region.
- this method and the statistical method that will be introduced in the next paragraph, are the only ones that address directly the problem of change in illumination.
- This method is based on the assumption that the division of the images cancels the difference in the illumination between the two images, which does not always hold in practice.
- ⁇ k ⁇ ( x , y ) ⁇ ( x ′ , y ′ ) ⁇ w ⁇ ( x , y ) ⁇ ⁇ d k 2 ⁇ ( x ′ , y ′ ) ⁇ 2 ( 8 )
- w(x,y) is a window of observation centered at (x,y).
- N Gaussian distribution
- the local sum ⁇ k (x,y) follows a x 2 distribution with N degrees of freedom, N being the number of pixels within the windows w(x,y).
- a decision rule for each pixel can obtained by a significance test on ⁇ k (x,y).
- T ⁇ the probability distribution of ⁇ k (x,y)
- the significance level ⁇ is in fact the false alarm rate associate with the statistical test. The higher the value of ⁇ , the more likely is the classification of unchanged pixels as change. It is obvious that the significance test depends on the noise variance ⁇ 2 . Thus, an accurate estimate of the noise variance is crucial for performance of the test. To ensure that, the variance is estimated only within the background region of the current frame, to remove the influence change region. The background regions are determined according to the tracked mask of the previous frame.
- One of the problems of this concept is the initial step when the background regions are not yet known: it requires a heuristics method that is strongly based on a threshold for estimating the background region.
- the likelihood ratio approach is suggested based on the assumption of uniform second-order statistics over a region [KAS91; N. H. Nagel, “Formulation of a blob concept by analysis of systematic time variation in the optically perceptible environment”, Computer Graphics and Image Processing, Vol. 7. pp. 149-194, 1978 (hereinafter NAG78); ZHA93].
- the frames can be subdivided into blocks, and then the blocks are compared on the basis of the statistical characteristics of their intensity levels.
- Eq. (10) represents the formula that calculates the likelihood function. Let ⁇ i and ⁇ i+1 be the mean intensity values for a given region in two consecutive frames, and ⁇ i and ⁇ i+1 be the corresponding variances.
- the number of the blocks that exceed a certain threshold t are counted. If the number of blocks exceeds a certain value (dependent on the number of blocks), a segment is declared. A subset of the blocks can be used to detect the difference between the images so as to expedite the process of block matching.
- SZE98 also propose a statistical decision rule to cope with the effects of noise.
- the change detection problem can be treated as one of hypothesis testing. Critical values are determined according to the desired level of significance. This does not perform change detection well, and there are many “false alarms”.
- a method for illumination-independent change detection between a pair of registered images comprising: a) providing a first original gray-level image, a second original gray-level image, a first negative image related to the first original image and a second negative images related to the second original image, b) extracting respective pluralities of blobs from each of the first and second original images and each of the first and second negative images, c) matching each extracted blob in the first original and negative images with each extracted blob in the second original and negative images to obtain matched and unmatched blobs, and d) testing all the unmatched blobs to identify blobs of change, whereby the blobs of change indicate a change detected by a method that is exact, fast, robust, illumination-insensitive and has low time-complexity.
- a method for change detection in images comprising: a) providing a pair of first and second registered gray level images, b) extracting respective first and second pluralities of blobs from each of the images using a modified connectivity along gray levels (CAG) analysis, c) locating at least one unmatched blob in at least one of the images, and d) identifying at least one blob of change related to the images by applying a test on the at least one unmatched blob.
- CAG modified connectivity along gray levels
- FIG. 1 shows a pair of images used as input to the change detection method.
- the images contain the same scenery, under different illumination, and each image contains a blob that is not contained in the other;
- FIG. 2 shows (a) the original input image and (b) extracted blobs bounded by white curves;
- FIG. 3 shows (a) an exemplary gray-level image, (b) its binary image I 120 , (c) its binary image I 140 and (d) its binary image I 190 ;
- FIG. 4 illustrates a weight function that corresponds to blob 304 in FIG. 3 .
- FIG. 5 is a block diagram showing a preferred embodiment of the method for change detection of the present invention.
- FIG. 6 shows the outputs of the algorithm for extracted blobs of the present invention, for the images in FIG. 1 .
- FIG. 7 is an illustration of a problem of a lack of matching between blobs, as demonstrated by synthetic blobs that might not be matched in the first step of the matching process of the change detection method
- FIG. 8 shows an exemplary pair I (1) and I (2) of infrared input images.
- FIG. 9 shows the absolute differences between the images I (1) and I (2) of FIG. 8 .
- FIG. 10 shows the blobs that belong to the SOL 1 and SOL 2 lists, drawn on the FIG. 8 .
- FIG. 11 shows the boundary of a blob in FIG. 10 , marked by a dotted line, laid over the gradient magnitudes of I (2) .
- FIG. 12 shows the result of the change detection method after comparison between I (1) and I (2) from FIG. 8 .
- FIG. 13 shows another example illustrating the effect of different illumination between two images.
- FIG. 14 shows a difference image
- FIG. 15 shows the outputs SOL 1 and SOL 2 relating to FIG. 13 marked on I (1) and I (2) , respectively;
- FIG. 16 shows the output from the change detection method after its application on FIG. 15 ;
- FIG. 17 shows another example of two input images I (1) and I (2) ;
- FIG. 18 shows on the x-axis the index of a blob in SOL 2 and on the y-axis the ratio between the fitness measures of this blob in image I (2) and in image I (1) of FIG. 17 ;
- FIG. 19 shows the output of the change detection method applied on the images in FIG. 17 ;
- FIG. 20 shows a comparison between results of change detection using the method of the present invention, and results of the Shading Model algorithm.
- the present invention is of a method to extract blobs that appear in only one of two images of any registered pair of images.
- the present invention can be used for illumination-independent change detection in a pair of gray images based on connectivity analysis along gray-levels.
- the principles and operation of a method to extract blobs that appear in only one of two images according to the present invention may be better understood with reference to the drawings and the accompanying description.
- the first step (the blob extraction step) of the present invention is preferably based on algorithm for blobs extraction based on connectivity analysis along gray-levels (CAG) as shown in Pikaz Arie, “Connectivity Analysis in Digital Gray-Levels Images and Its Applications”, Ph.D. thesis, Tel-Aviv University, Israel, February 1998 (hereinafter PIK98).
- the original CAG algorithm is preferably used herein as a starting point for detecting, with an appropriate set of parameters, all visually conspicuous blobs. It is worthwhile pointing out that the CAG algorithm does not deal with change detection. Thus, all the steps of the present method beyond the blob extraction step are novel.
- a perfect match between corresponding blobs is not expected. More than that, several blobs from one image may be connected or united into a big blob in the other image (thus, even if no change occurred, the values N 1 and N 2 might be different).
- the coordinates of the corresponding blobs should be the same in both images. If sufficient number of pixels of both blobs has the same coordinates (“sufficient” is determined by a pre-defined parameter), the examined pair of blobs is considered as a match. If not, then the blob from one image is “marked” in the second image. The image gradients along its boundary are examined. A measure of saliency is defined according to the distribution of the magnitudes of these gradients. If this saliency measure is sufficiently high, then the blob is classified as existing in both images. Otherwise, it represents a change. The proposed change detection method is very efficient and robust, and it is adequate for real-time applications.
- I t denotes a binary image that is the result of thresholding image I with a threshold t.
- a segment is defined as a set of black pixels in which there exists a 4-connected path of black pixels between each two pixels of the set. It is clear that a binary image can be represented by the set of all segments that it contains.
- FIG. 2 shows (a) an original input image and (b) extracted blobs bound by white curves.
- blobs which are the outputs of the CAG algorithm, significant blobs.
- a blob in a gray-level image is visually conspicuous if one or more of the following exist:
- C i (t) be the i th connected-segment in the binary image I t .
- C j ⁇ tilde over (t) ⁇
- Ci Ci
- I 140 corresponds to a value of t ⁇ 140, for which I t best represents the two significant blobs 302 and 304 from FIG. 3(a) .
- FIG. 3(d) presents I 190 that contains segments that are larger than the “real” blobs 302 and 304 .
- the threshold value t that corresponds to each significant blob has to be detected automatically.
- the weight is a function of the threshold parameter t. It is denoted by w c (t), where C is the relevant segment.
- the function w c (t) is defined as follows: in the binary image I t there exists at most a single segment C′ that satisfies C′ ⁇ C ⁇ .
- the value of w c (t) is defined as the average value of the gradient magnitudes along the boundary of the segment C′.
- the weight of a blob is defined as the average weight of the pixels along the blob boundary. This weight is expected to be proportional to the blob saliency, defined hereinbelow.
- C (t) be a segment that corresponds to a “real blob”. Its weight is expected to be the maximal weight among the weights of all the clusters that are not disjoint to C ( ⁇ tilde over (t) ⁇ ) .
- the weight function w c (t) is expected to have local maxima at values of t that correspond to the binary image I t that contains the significant blob. For illustration, the weight function that corresponds to blob 304 in FIG. 3(a) is presented in FIG. 4 .
- a weight is attached to each pixel in the input gray-level image I.
- the weight that is attached to a pixel is a measure of edge saliency.
- a pixel that resides on an edge gets assigned a higher weight than a non-edge pixel.
- a reasonable choice for the image of weights is the magnitudes of the gradients of I.
- w ⁇ ( C ) ⁇ def ⁇ 1 ⁇ ⁇ C ⁇ ⁇ ⁇ q _ ⁇ ⁇ C ⁇ ⁇ w ⁇ ( q _ ) ( 12 )
- ⁇ C is the set of boundary pixels of the segment C and
- is the size of this set.
- a pixel q is defined as a boundary pixel of the segment C if it belongs to C and at least one of its four nearest neighbors does not belong to C.
- the definition in Eq. 13 has the following convenient property.
- the weight of the union of a segment C with a segment that is composed of a single pixel can be computed in a constant number of operations.
- p be a pixel that is united with the segment C.
- C′ be the result of the union between the segment C and ⁇ p ⁇ .
- the weight of C′ satisfies
- w ⁇ ( C ′ ) w ⁇ ( C ) ⁇ s ⁇ ( C ) + w ⁇ ( p _ ) - ⁇ q _ ⁇ ⁇ C ⁇ ⁇ and ⁇ ⁇ q _ ⁇ ⁇ C ′ ⁇ w ⁇ ( q _ ) s ⁇ ( C ) + 1 ( 13 )
- s(C) is the number of pixels that are contained in segment C.
- an examined blob O corresponds to a connected segment C in a binary image I ⁇ tilde over (t) ⁇ , then its gray-levels must differ from the gray-levels of its local background. From the definition of the function w c (t) we conclude that if a local maximum of w c (t) exists at point ⁇ tilde over (t) ⁇ , then the blob O is salient related to its local background. Thus, the combination of connectivity and gradients along the boundary of the connected segments is a powerful measure of the significance of a given blob.
- the change detection starts with a novel and much more efficient method and algorithm of the present invention disclose a much more efficient way to extract blobs than the original CAG algorithm in PIK98.
- the present invention significantly adds to, and substantially enhances the capabilities of the original CAG algorithm with the following features: 1) the CAG of the present invention selects the conspicuous blobs based on local considerations; 2) local considerations are added to the original algorithm so that the analysis of the image is more reliable; 3) the present algorithm is more robust, this being achieved by changing the formulae and computations of local weights. 4) the computation of connected components is based herein on lower and upper bounds, which increases the accuracy of the detected blobs.
- the lower bound is computed by increasing the value of the threshold t from 0. This yields a binary image according to the threshold t. Then a weight is computed according to Eqs. 11 and 12. As t increases, one gets a monotonically increasing function until t reaches the maximum. This is illustrated in FIG. 4 . This maximum is the value of the final threshold, but may be wrong because it is based on gray level values to which it is very sensitive.
- the modified blob extraction algorithm is first applied on two input images I 1 and I 2 (herein “original” images), and on their negatives I 1 and I 2 ,
- the outputs are four lists of blobs, SOL 1 , SOL 1 , SOL 2 and SOL 2 .
- the upper bar means “negative”.
- SOL 1 is a first unified list that contains the union SOL 1 ⁇ SOL 1 (i.e. all extracted blobs in SOL 1 and SOL 1 )
- SOL 2 is a second unified list that contains the union SOL 2 ⁇ SOL 2 .
- the lists SOL 1 and SOL 2 contain all the candidate blobs of change that exist in images I (1) and I (2) , respectively.
- the idea is to find for each blob in SOL 1 a matched blob in SOL 2 , and then for each blob in SOL 2 a matched blob in SOL 1 .
- the method for change detection of the present invention is presented first in general steps in a block diagram in FIG. 5 , with a detailed description of each step given later.
- FIG. 5 shows a block diagram of a preferred embodiment of the method of the present invention.
- the change detection is performed between two images I 1 and I 2 . Therefore, the inputs are four images 20 : I 1 , I 2 and their negatives I 1 and I 2 , respectively.
- a blob extraction step 22 is separately applied on each of the input images I 1 , I 2 and their negatives I 1 , and I 2 .
- Each application produces as output a list of extracted significant blobs 24 .
- SOL 1 ⁇ C i (1) ⁇ i ⁇ 1 n 1
- SOL 1 ⁇ C i (1) ⁇ i ⁇ 1 m 1
- SOL 2 ⁇ C i (2) ⁇ i ⁇ 1 n 2
- each of the four lists are disjoint, but there might be a pair of connected segments from SOL 1 and SOL 1 (and similarly, from SOL 2 and SOL 2 with non-empty intersection.
- the assumption is that each blob of change exists in one of the four lists, SOL 1 , SOL 1 , SOL 2 and SOL 2 .
- Each list of extracted blobs contains information, for example geometrical information in the form of pixel locations, on each blob in the list. An example is shown in FIG. 6 .
- FIG. 6 shows the four output lists (“outputs”) of the blob extraction step for the pair of images of FIG. 1 .
- the extracted blobs are bound by white curves.
- r is the radius of the search
- s min and s max are the minimum and maximum of the size of the blob, respectively
- w is the weight of the blob computed by Eq. 13.
- the algorithm is stable and insensitive to the exact choice of the parameters:
- FIG. 6 shows in (a) all the extracted blobs from the right image of FIG.
- each blob O i (1) in SOL 1 is trial matched with a blob O j (2) in SOL 2 . If blobs O i (1) and O j (2) overlap (have the same coordinates) by at least ⁇ % of pixels (where ⁇ ranges typically from 75% to 100%, and preferably between 90-100%, i.e. where preferably the lower bound of ⁇ is about 90%) then a logical check step 32 checks that blob O i (1) also exists in SOL 2 . A logical yes answer in checking step 32 leads to a “no change” step 34 that determines that this is not a blob of change.
- step 30 the algorithm proceeds to another novel step, unique to the present invention: a fitness-measuring step 36 , which computes a fitness measure fm o 1 (I 1 ,I 2 ) of blob O i (1) in SOL 2 .
- the fitness measure is then compared to a pre-defined parameter (threshold) ⁇ in a comparison step 38 . If fm o 1 (I 1 ,I 2 ) ⁇ , then blob O i (1) is determined as existing in both images (yes), i.e.
- step 30 the algorithm proceeds back to step 30 to process another blob. If fm o 1 (i t I 2 ) > ⁇ (no) then the blob is declared in a step 40 as a blob that exists in one list and not in the other. Therefore, this is a blob of change, and the algorithm returns to process another blob in step 30 .
- the parameter ⁇ is preferably between 0 and 1, and most preferably about 0.6.
- Phase 1 (Steps 30 and 32 in FIG. 1 )—O i (1) has a Corresponding Blob in SOL 2
- each pixel of O j (2) has a corresponding pixel in O i (1) with identical coordinates, but usually this is not the case.
- the two blobs O i (2) and O i (1) match if the coordinates of at least ⁇ % of the pixels from both blobs are identical.
- a is typically chosen to be 90.
- the blob O i (1) has no matching blob in SOL 2 but it has a corresponding blob in the image I (2) .
- An example for such a case is shown in FIG. 7 that presents a pair of images with synthetic blobs in (a) and (b), in which there are no “blobs of change”.
- FIG. 7 is an illustration of the problem as demonstrated by synthetic blobs that might not be matched in the first step of the matching process of the change detection algorithm (step 30 ). As can be seen, no change exists in the given pair of gray-level images, but the images are not exactly the same.
- the output of the blob extraction algorithm for the left image (a) will contain three blobs—a circle 702 , an ellipse 704 and a background 706 .
- the output of the blob extraction algorithm for the right image (b) will contain only two blobs—one blob 708 composed of the circle, the ellipse and the very thin line that connects them, and another blob 710 representing the background.
- blob 702 from the right image has no matching blob, according to matching step 30 of the matching process.
- the saliency measure of a blob is a function of the magnitudes of the gradients of its boundary pixels.
- the saliency measure of a blob O with boundary ⁇ O was defined as:
- the saliency measure of each of the two blobs 702 and 704 in the left image (a) will be also high in right image (b). Then, the value sal(O 1 (1) ;I (1 is expected to be close to sal(O 1 (1) ;I (2) ), and similarly, the value sal(O 2 (1) ; I (1) ) is expected to be close to sal(O 1 (2) ;I (2) ).
- E m this maximal entropy value, that is:
- dm o 1 ( I 1 , I 2 ) ⁇ dst o 1 ( I 1 ) dst o 1 ( I 2 ) sm o ⁇ 1 ( I 1 , I 2 ) ⁇ ⁇ 1 else ( 17 ) where the ratio in dm o 1 (I 1 I 2 ) (Eq. 17) along the gradients boundary pixels is considered only if sm o 1 (I 1 ,I 2 ) (Eq. 15) is less than ⁇ and ⁇ [0,1]. We preferably choose ⁇ to be 0.6. 3.5 The Fitness Measure
- SOL (out) are the list of the output blobs, that is, the final list of “blobs of change”. SOL (out) is initialized to be an empty list.
- the worst-case complexity of this pass is linear in the number of pixels of the blob.
- the overall time complexity of the algorithm is almost linear in the image size, n. Specifically, it is O(n ⁇ (n,n)), where ⁇ (n,n) is the inverse of the Ackermman function [COR90, chapter 22.4], which is almost a constant.
- ⁇ (n,n) is the inverse of the Ackermman function [COR90, chapter 22.4]
- each of the four SOL lists takes O(n ⁇ (n,n)) operations in the worst-case, where ⁇ (n,n) is the inverse of the Ackermman function. Therefore, the worst-case time complexity for the creation of SOL 1 and SOL 2 is O(n ⁇ (n,n)).
- section 5.1 presents the complete process of the change detection algorithm.
- section 5.2 we demonstrate the robustness of the algorithm, and its insensitivity to change in the illumination.
- section 5.3 we focus on step 36 , FIG. 5 of the matching procedure.
- section 5.4 the proposed method is compared to the “Shading Model” method, which is one of the methods that explicitly deal with significant changes in the luminance, as reviewed above.
- FIG. 8 shows a pair of InfraRed registered input images.
- the left image I (1) in (a) contains two blobs (a bus 802 at the bottom and a vehicle 804 at the top) that are not contained in the right image I (2) in (b).
- Image I (2) contains two blobs (a vehicle 806 at the top and a vehicle 808 in the middle) that are not contained in I (1) ; These are four potential “blobs of change”.
- FIG. 9 shows a difference image of images I (1) and I (2) from FIG. 8 in which it is seen that the change in the illumination between I (1) and I (2) is significant
- FIG. 10 shows the blobs that belong to the lists SOL 1 and SOL 2 , drawn on FIG. 8 .
- the two vehicles of I (1) 802 ′ and 804 ′) and the top vehicle of I (2) ( 806 ′) are pointed by the arrows.
- the list SOL 1 also contains a blob 1002 .
- We can see in the same location in I (2) the same blob (marked as blob 810 ).
- blob 810 was not found by the blob extraction that was applied on I (2) .
- the blob corresponding to blob 1002 was not detected at all by the blob-extraction algorithm, even though it appears in I (2) .
- Blobs in SOL 1 that have no corresponding blobs in SOL 2 are passed to step 36 in FIG. 5 .
- the same is done with blobs in SOL 2 that have no corresponding blobs in SOL 1 .
- saliency and fitness measures for each blob in SOL 1 are computed in I (2) , and vice-versa. If the saliency measure of any blob from I (1) is sufficiently high in I (2) , (e.g.
- this saliency measure is represented in the image of gradient magnitudes of I (2) by a dotted contour, e.g. a dotted contour 1102 in FIG. 11 , which corresponds to blob 810 in FIG. 10 .
- FIG. 12 shows the result of the change detection algorithm that represents the changes in I (1) relatively to I (2) .
- the two blobs in I (1) which do not exist in I (2) , are marked 1202 and 1204 .
- FIG. 13 Two generated images I (1) (a) and I (2) (b) having extreme differences between their illuminations are shown in FIG. 13 .
- a capital letter “A” is indicated by an arrow 1302 in (b).
- Capital letter “A” was omitted from I (1) (an arrow 1004 in (a)) in order to make it the “blob of change”.
- FIG. 14 shows a difference image
- the subtraction between the two input images I (1) and I (2) demonstrates a significant difference in the illumination.
- Step 24 ( FIG. 5 ) of the change detection algorithm outputs the extracted blobs.
- FIG. 15 shows the outputs (a) SOL 1 and (b) SOL 2 relating to FIG. 13 , marked on I (1) (a) and I (2) (b), respectively. Letters 1502 that compose the word “DISK 2/2” in I (1) were deliberately omitted from SOL 1 in order to demonstrate the robustness of the change detection algorithm.
- each blob in SOL 1 is searched for a corresponding blob in SOL 2 , and vice versa. Seven blobs that correspond to the letters “DISK 2/2”, and the blob for capital letter “A” in SOL 2 ( FIG. 15b ) have no corresponding blobs found in SOL 1 ( FIG. 15a ). Each of these eight blobs is checked again in step 36 . In this step, the saliency and the fitness measures in I (1) for each “non-matched” blob in SOL2 is computed. The only blob in SOL 2 that has saliency and fitness measures in I (1) with a higher value than ⁇ (section 5.1), is the capital letter “A”.
- the output of the change detection is the blob composed of the pixels of the letter “A”. This is shown in FIG. 16 , which shows the capital letter “A” as a blob of change 1602 .
- FIG. 17 shows two input images I (1) in (a) and I (2) in (b).
- the blobs of change are a vehicle 1702 in I (1) , and three vehicles 1704 , 1706 and 1708 in I (2) .
- the outputs SOL 1 and SOL 2 (not shown) of each image contain 116 blobs.
- the fitness measure for blobs in SOL 2 that also appear in SOL 1 is much higher than the fitness measure in SOL 1 of blobs that appear only in SOL 2 .
- FIG. 18 shows on the x-axis the ordinal number of the blob (index i) in SOL 2 and on the y-axis the ratio between the fitness measure of this blob in image I (2) and in image I (1) of FIG. 17 , the ratio computed using Eq. 18.
- the three peaks 1802 , 1804 and 1806 correspond to the three vehicles 1704 , 1706 and 1708 that appear in I (2) and not in I (1) ;
- FIG. 19 shows the output of the change detection algorithm applied on the images in FIG. 17 .
- the three extracted “blobs of change” from I (2) ( 1704 , 1706 and 1708 ) are marked by curves 1902 , 1904 and 1906 .
- FIG. 20 shows a comparison between the results obtained with the two methods on a pair of images: in (a) a blob of change 2002 that appears in I (1) and not in I (2) after the application present method; in (b) a blob of change 2004 that appears in I (2) and not in I (1) after the application of present method; in (c) the image of variances, which is the output of the Shading Model method, and in (d) the binary image of (c) derived by choosing the most appropriate threshold to extract blob of change 2002 . From FIGS.
- the Shading Model method fails to detect the change because there too many blobs that do not have a match. This happened because of the abrupt and extreme changes in illumination.
- the “Shading Model” method computes the variance over a window with pre-defined size. In case of a blob of change that is considerably larger than the window size, only part of the blob boundary will be detected.
- the present invention introduces an efficient and robust method that provides a novel algorithm for performing change detection between a pair of registered gray-level images, under severe differences in the illumination of the images.
- the output of the algorithm is a set of connected components, where each component describes a “blob of change”, which is a blob that exists in only one of the two images.
- the time complexity of the change detection algorithm is almost linear in the image size. Therefore, it is suitable for real-time applications.
- the examples detailed above demonstrate its robustness even when extreme changes in the illumination exist.
- the main advantages of the method disclosed herein include:
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
Description
1.1 | Pixel-level change detection. |
1.2 | Surface modeling. |
1.3 | Comparison among derivative images. |
1.4 | Contrast invariant representation. |
1.5 | Region based comparison of first or higher order statistics. |
-
- otherwise
In Eq. 1, Fi(x,y) is the intensity value of the pixel in frame i at the coordinates (x,y). If the difference between the corresponding pixels in the two consecutive frames is above a certain minimum intensity value, then DPi(x,y), the difference picture, is set to one. In Eq. 2, the percentage difference between the pixels in the two frames is calculated by summing the difference picture and dividing by the total number of pixels in a frame. If this percentage is above a certain threshold T, a camera break is declared.
1.3. Comparison among Derivative Images
since the amount of reflected light from point (x0,y0) depends on the blob itself. Let F be the image of real numbers that is the result of the division of the two images, I(1) and I(2), that is:
where F is assumed to have accuracy of real numbers. Let RF(x0,y0) be a small neighborhood around the point (x0,y0) in the image F. Then, for a point (x0,y0) that belongs to the same blob in both images, I(1) and I(2), the surface patch that is composed of the values in the region RF(x0,y0) is expected to be a smooth and slow varying surface, since the change of the illumination in a small region is slow-varying. On the other hand, for a pixel (x0,y0) that belongs to a different blob in each image, the surface patch that is composed of the values in RF(x0,y0) is expected to be much less smooth, since the region RF(x0,y0) can include a transition from one blob to another. The method in [SKI89] proposes to examine the variance in each pixel (x0,y0) of the region RF(x0,y0). If the variance is higher than some pre-specified threshold, then the pixel is considered as belonging to a region of change. The change detection mask of this method is defined for each pixel by the following formula:
where μi is the average value of the ratio of intensities, E is the expectation, N is the size of the image, and “Λ” is a 5×5 region. Among all the other reviewed methods, this method, and the statistical method that will be introduced in the next paragraph, are the only ones that address directly the problem of change in illumination. This method is based on the assumption that the division of the images cancels the difference in the illumination between the two images, which does not always hold in practice. Moreover, the variance inside a region RF(x0,y0), whose size is not based on the image content, adds inaccuracies of its own.
dk=(x, y)=Ik+1(x, y)+Ik(x, y) (6)
denote the image of gray level differences between frames I(k) and I(k+1). Under the
hypothesis than no changes occurred at position (x,y) (the null hypothesis H0), the corresponding difference dk(x,y) follows a zero-mean Gaussian distribution
where the noise variance σ2 is equal to twice the variance of the camera noise, assuming that the camera noise is white. Rather than performing the significance test on the values dk(x,y), it is better to evaluate a local sum of normalized differences:
where w(x,y) is a window of observation centered at (x,y). Under the assumption that no changes occur within the windows, the normalized differences dk/σ obey a Gaussian distribution N(0,1) and are spatially uncorrelated. Thus, the local sum Δk(x,y) follows a x2 distribution with N degrees of freedom, N being the number of pixels within the windows w(x,y). With the distribution p(Δk(x,y)) known, a decision rule for each pixel can obtained by a significance test on Δk(x,y). For a specific level a one can compute a corresponding threshold Tα using:
α=Pr{Δk(x, y)>Tα|H 0} (9)
The significance level α is in fact the false alarm rate associate with the statistical test. The higher the value of α, the more likely is the classification of unchanged pixels as change. It is obvious that the significance test depends on the noise variance α2. Thus, an accurate estimate of the noise variance is crucial for performance of the test. To ensure that, the variance is estimated only within the background region of the current frame, to remove the influence change region. The background regions are determined according to the tracked mask of the previous frame. One of the problems of this concept is the initial step when the background regions are not yet known: it requires a heuristics method that is strongly based on a threshold for estimating the background region.
a camera break is declared. This approach increases the tolerance against noise associated with camera and blob movement. It is possible that even though the two corresponding blocks are different, they can have the same density function. In such cases no change is detected.
- 1. The gray-levels inside the blob are considerably different from the gray-levels of the local background of the blob.
- 2. Most of the magnitudes of the gradients that correspond to pixels along the blob boundary, are higher than the magnitudes of the gradients that correspond to pixels in the local background.
- 3. The texture inside the blob is considerably different from the texture of the local background of the blob.
The CAG algorithm deals with blobs that satisfy the first two conditions. It is assumed that for each significant blob there exists a threshold value t such that the blob is a connected segment in It.
- 1. If t′<t″ then C(t′) ⊂C(t″)
- 2. There exists a value {tilde over (t)}≧t0 such that O⊂C({tilde over (t)}).
where ƒ(u,v) is the gray-level of the pixel at column u and row v in the image I. Denote the image of the magnitudes of the gradients by I(MAG). For a given pixel p i=(ui, vi) let w(p i) be the intensity value at column ui and now vi in the image I(MAG). The weight of a given segment C is defined by
where ∂C is the set of boundary pixels of the segment C and |∂C| is the size of this set. A pixel q is defined as a boundary pixel of the segment C if it belongs to C and at least one of its four nearest neighbors does not belong to C. The definition in Eq. 13 has the following convenient property. The weight of the union of a segment C with a segment that is composed of a single pixel can be computed in a constant number of operations. Let p be a pixel that is united with the segment C. Let C′ be the result of the union between the segment C and {p}. The weight of C′ satisfies
where s(C) is the number of pixels that are contained in segment C. It is clear that the set {q|qε∂C and qε∂C′} is composed only of pixels that are nearest-neighbors of the pixel p. Therefore, only a constant number of operations is required to compute
The same is true for w(C′).
where |∂O| is the number of boundary pixels, ∇I(x,y) is the gradient vector of the image I at pixel (x,y) and |∇I(x,y)| is the magnitude of the gradient. The saliency measure of a blob Oi (1) in an image I(2) is computed according to the gradient values of the pixels (x,y)ε∂Oi (1) in image I(2). In the example of
is sufficiently high (above some pre-defined threshold), the blob Oi (1) is declared as an “blob of change”. Otherwise, analysis of a second order statistic is required.
3.4 The use of Gradient Distribution Vector for the Matching Process
It is clear that for each k,
Denote by E the entropy of p1, . . . , Pm, that is:
The maximal entropy is achieved when pk are uniformly distributed: pk=1/m. Denote by Em this maximal entropy value, that is:
When there are more pk s that are uniformly distributed we get higher value for the gradients distribution and it is denoted by:
m/n>γ where γε[0,1]. We preferably choose γ to be 0.6. Eq. 15 enables to determine whether the blob is a change when the smo (I
where the ratio in dmo
3.5 The Fitness Measure
fmo
fm enables to decide whether blobs exist in both images or only in one of them. According to the definition of (Eq. 17), when the value of sm (Eq. 15) is higher than βfm will be based only on the ratio of the smo (I
4. Implementation of the Change Detection Algorithm
4.1 The Pseudo-code
- 1. Apply the blob extraction algorithm (section 3) on images I(1) and
I(1) in order to get the output lists of significant blobs, SOL1 andSOL1 respectively. Denote the unified list SOL1∪SOL1 by SOL1. Similarly, construct unified list SOL2. - 2. For each blob Oi (1) in SOL1 do:
- 2.1. Let p=(xi,yi) be a representative pixel of Oi (1). Assume, without loss of generality, that Oi (1) was extracted from I(1) (and not from I(1)) Let Oj (2) be the blob in I(2) that contains the pixel p.
- 2.2. If the blobs Oi (1) and Oj (2) are overlapped by at least α % of pixels (chose preferably α=90) then blob Oi (1) also exists in Oj (2).
- 2.3. Else, compute fmo
1 (I1 ,I2 ) (the fitness measure of blob Oi (1) in image I(2)). The values of sal(Oi (1);I(2)) and sal(Oi 1;I(1)) are computed by using the contour following algorithm [Jain Anil K., Fundamentals of Digital Image Processing, Prentice-Hall, 1989, chapter 9.5 (hereinafter JAI89)], starting with a boundary pixel of Oi (1) and Oi (2)). - If fmo
1 (I1 ,I2 )<γ, where γ is a pre-defined parameter valued between 0 and 1, and preferably about 0.6, then blob Oi (1) exists in both images. - Else (that is, if fmo
1 (I1 ,I2 )≧γ) mark blob Oi (1) as a blob that appears in image I(1) and not in image I(2). Then, insert Oi (1) into SOL(out).
- 3.
Repeat step 2 for each blob Oi (2) in SOL2, while replacing the roles of image I(1) with image I(2).
-
- 1.1. Given a pixel p, the blob that contains it can be found in O(1) operations by simply keeping an array of n entries such that entry i points to the pixel who is the head of the class.
- 1.2. In order to find the percentage of matching pixels between two blobs, a single pass on both of them is required. This pass is linear in the number of pixels of the blob.
- 1.3. The boundary of each blob is extracted by a single pass on the boundary pixels, using the contour-following algorithm [JAI89, chapter 9.5). All that is required is a pixel that is known to reside on the boundary. Such a pixel is attached in advance to each blob, as part of the output of the blob extraction algorithm. The worst-case time complexity of computing the saliency/fitness measure of blob Oi (1) in the image I(2) is linear in the number of boundary pixels.
- 2. As in
part 2, the worst-case time complexity is O(n). - 5. Experimental Results
-
- a. Exact detection of the change. The method works also for noisy inputs with very small “blobs of change” (ca. 30 pixels).
- b. The input images can contain several “blobs of change” with a considerable difference in their sizes. This is a consequence of the fact that the disclosed method does not use a window with pre-defined size, but works directly on the extracted blobs.
- c. The detection of change is robust and insensitive to noise as long as the change is a connected blob.
Claims (40)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/188,654 USRE42367E1 (en) | 2001-10-31 | 2008-08-08 | Method for illumination independent change detection in a pair of registered gray images |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/984,756 US7088863B2 (en) | 2001-10-31 | 2001-10-31 | Method for illumination independent change detection in a pair of registered gray images |
US12/188,654 USRE42367E1 (en) | 2001-10-31 | 2008-08-08 | Method for illumination independent change detection in a pair of registered gray images |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/984,756 Reissue US7088863B2 (en) | 2001-10-31 | 2001-10-31 | Method for illumination independent change detection in a pair of registered gray images |
Publications (1)
Publication Number | Publication Date |
---|---|
USRE42367E1 true USRE42367E1 (en) | 2011-05-17 |
Family
ID=28675896
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/984,756 Ceased US7088863B2 (en) | 2001-10-31 | 2001-10-31 | Method for illumination independent change detection in a pair of registered gray images |
US12/188,654 Expired - Lifetime USRE42367E1 (en) | 2001-10-31 | 2008-08-08 | Method for illumination independent change detection in a pair of registered gray images |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/984,756 Ceased US7088863B2 (en) | 2001-10-31 | 2001-10-31 | Method for illumination independent change detection in a pair of registered gray images |
Country Status (1)
Country | Link |
---|---|
US (2) | US7088863B2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100127704A1 (en) * | 2008-11-27 | 2010-05-27 | Marcel Warntjes | Method of Visualizing Segmented MR Images |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7983446B2 (en) * | 2003-07-18 | 2011-07-19 | Lockheed Martin Corporation | Method and apparatus for automatic object identification |
US7499570B2 (en) * | 2004-03-02 | 2009-03-03 | Siemens Corporate Research, Inc. | Illumination invariant change detection |
JP4534594B2 (en) * | 2004-05-19 | 2010-09-01 | ソニー株式会社 | Image processing apparatus, image processing method, program for image processing method, and recording medium recording program for image processing method |
US9837013B2 (en) * | 2008-07-09 | 2017-12-05 | Sharp Laboratories Of America, Inc. | Methods and systems for display correction |
US8446463B2 (en) * | 2008-08-22 | 2013-05-21 | Genprime, Inc. | Apparatus, method and article to perform assays using assay strips |
CN101799434B (en) * | 2010-03-15 | 2011-06-29 | 深圳市中钞科信金融科技有限公司 | Printing image defect detection method |
CN103049902A (en) * | 2012-10-23 | 2013-04-17 | 中国人民解放军空军装备研究院侦察情报装备研究所 | Image change detection method and device |
CN106295699A (en) * | 2016-08-11 | 2017-01-04 | 北京师范大学 | A kind of earthquake Damage assessment method and apparatus based on high-definition remote sensing data |
US10576475B2 (en) | 2016-09-15 | 2020-03-03 | Genprime, Inc. | Diagnostic assay strip cassette |
CN107274357B (en) * | 2017-05-23 | 2020-09-15 | 浙江大学 | Gray level image enhancement processing system with optimal parameters |
CN108765465B (en) * | 2018-05-31 | 2020-07-10 | 西安电子科技大学 | Unsupervised SAR image change detection method |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5627907A (en) * | 1994-12-01 | 1997-05-06 | University Of Pittsburgh | Computerized detection of masses and microcalcifications in digital mammograms |
US6028970A (en) | 1997-10-14 | 2000-02-22 | At&T Corp | Method and apparatus for enhancing optical character recognition |
US6691126B1 (en) | 2000-06-14 | 2004-02-10 | International Business Machines Corporation | Method and apparatus for locating multi-region objects in an image or video database |
US6701007B1 (en) * | 1999-03-11 | 2004-03-02 | Fuji Photo Film Co., Ltd. | Image processing device, image processing method, and recording medium for correcting color balance of output images |
US6751345B2 (en) * | 2001-02-12 | 2004-06-15 | Koninklijke Philips Electronics N.V. | Method and apparatus for improving object boundaries extracted from stereoscopic images |
US6829371B1 (en) | 2000-04-29 | 2004-12-07 | Cognex Corporation | Auto-setup of a video safety curtain system |
-
2001
- 2001-10-31 US US09/984,756 patent/US7088863B2/en not_active Ceased
-
2008
- 2008-08-08 US US12/188,654 patent/USRE42367E1/en not_active Expired - Lifetime
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5627907A (en) * | 1994-12-01 | 1997-05-06 | University Of Pittsburgh | Computerized detection of masses and microcalcifications in digital mammograms |
US6028970A (en) | 1997-10-14 | 2000-02-22 | At&T Corp | Method and apparatus for enhancing optical character recognition |
US6701007B1 (en) * | 1999-03-11 | 2004-03-02 | Fuji Photo Film Co., Ltd. | Image processing device, image processing method, and recording medium for correcting color balance of output images |
US6829371B1 (en) | 2000-04-29 | 2004-12-07 | Cognex Corporation | Auto-setup of a video safety curtain system |
US6691126B1 (en) | 2000-06-14 | 2004-02-10 | International Business Machines Corporation | Method and apparatus for locating multi-region objects in an image or video database |
US6751345B2 (en) * | 2001-02-12 | 2004-06-15 | Koninklijke Philips Electronics N.V. | Method and apparatus for improving object boundaries extracted from stereoscopic images |
Non-Patent Citations (2)
Title |
---|
Kurt et al. "Illumination independent change detection for real world image sequences", Computer Vision, Graphics, and Image Processing, vol. 46, pp. 387-399, 1989. * |
Teasdale et al. "Functional MRI Study of the Cognitive Generation of Affect", American Journal of Psychiatry (GPSI), v156 n2, p. 209-215, Feb. 7, 1999. * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100127704A1 (en) * | 2008-11-27 | 2010-05-27 | Marcel Warntjes | Method of Visualizing Segmented MR Images |
US8643363B2 (en) * | 2008-11-27 | 2014-02-04 | Syntheticmr Ab | Method of visualizing segmented MR images with absolute-scale values independent of MR scanner settings |
Also Published As
Publication number | Publication date |
---|---|
US7088863B2 (en) | 2006-08-08 |
US20030190075A1 (en) | 2003-10-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
USRE42367E1 (en) | Method for illumination independent change detection in a pair of registered gray images | |
Sajid et al. | Universal multimode background subtraction | |
US9898686B2 (en) | Object re-identification using self-dissimilarity | |
US8600105B2 (en) | Combining multiple cues in a visual object detection system | |
Sanin et al. | Shadow detection: A survey and comparative evaluation of recent methods | |
Zhang et al. | Image segmentation based on 2D Otsu method with histogram analysis | |
Kittler et al. | Threshold selection based on a simple image statistic | |
Sari et al. | Road crack detection using support vector machine (SVM) and OTSU algorithm | |
US9547800B2 (en) | System and a method for the detection of multiple number-plates of moving cars in a series of 2-D images | |
US8401292B2 (en) | Identifying high saliency regions in digital images | |
US7639878B2 (en) | Shadow detection in images | |
Sharma | Performance analysis of vehicle number plate recognition system using template matching techniques | |
US7623683B2 (en) | Combining multiple exposure images to increase dynamic range | |
US6792134B2 (en) | Multi-mode digital image processing method for detecting eyes | |
US7136524B1 (en) | Robust perceptual color identification | |
US20050147298A1 (en) | Detection of sky in digital color images | |
US20140056519A1 (en) | Method, apparatus and system for segmenting an image in an image sequence | |
US8094971B2 (en) | Method and system for automatically determining the orientation of a digital image | |
CN111783665A (en) | Action recognition method and device, storage medium and electronic equipment | |
US20210027202A1 (en) | Method, system, and computer readable medium for performance modeling of crowd estimation techniques | |
Lecca et al. | Comprehensive evaluation of image enhancement for unsupervised image description and matching | |
US7403636B2 (en) | Method and apparatus for processing an image | |
CN111402185B (en) | Image detection method and device | |
Fang et al. | 1-D barcode localization in complex background | |
CN117218672A (en) | Deep learning-based medical records text recognition method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: PONSET MGMT. LIMITED LIABILITY COMPANY, DELAWARE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INFOWRAP SYSTEMS, LTD.;REEL/FRAME:022187/0936 Effective date: 20080526 |
|
CC | Certificate of correction | ||
FPAY | Fee payment |
Year of fee payment: 8 |
|
AS | Assignment |
Owner name: INTELLECTUAL VENTURES FUND 83 LLC, NEVADA Free format text: MERGER;ASSIGNOR:PONSET MGMT. LIMITED LIABILITY COMPANY;REEL/FRAME:037274/0785 Effective date: 20150826 |
|
AS | Assignment |
Owner name: CHEMTRON RESEARCH LLC, DELAWARE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTELLECTUAL VENTURES FUND 83 LLC;REEL/FRAME:041469/0177 Effective date: 20161230 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553) Year of fee payment: 12 |